OpenAI Integration with Instructor¶
OpenAI is the primary integration for Instructor, offering robust support for structured outputs with GPT-3.5, GPT-4, and future models. This guide covers everything you need to know about using OpenAI with Instructor for type-safe, validated responses.
Quick Start¶
Instructor comes with support for OpenAI out of the box, so you don't need to install anything extra.
⚠️ Important: You must set your OpenAI API key before using the client. You can do this in two ways:
- Set the environment variable:
- Or provide it directly to the client:
Simple User Example (Sync)¶
import os
from openai import OpenAI
import instructor
from pydantic import BaseModel
# Initialize with API key
client = OpenAI(api_key=os.getenv('OPENAI_API_KEY'))
# Enable instructor patches for OpenAI client
client = instructor.from_openai(client)
class User(BaseModel):
name: str
age: int
# Create structured output
user = client.chat.completions.create(
model="gpt-4o-mini",
messages=[
{"role": "user", "content": "Extract: Jason is 25 years old"},
],
response_model=User,
)
print(user)
#> User(name='Jason', age=25)
Simple User Example (Async)¶
import os
from openai import AsyncOpenAI
import instructor
from pydantic import BaseModel
import asyncio
# Initialize with API key
client = AsyncOpenAI(api_key=os.getenv('OPENAI_API_KEY'))
# Enable instructor patches for async OpenAI client
client = instructor.from_openai(client)
class User(BaseModel):
name: str
age: int
async def extract_user():
user = await client.chat.completions.create(
model="gpt-4-turbo-preview",
messages=[
{"role": "user", "content": "Extract: Jason is 25 years old"},
],
response_model=User,
)
return user
# Run async function
user = asyncio.run(extract_user())
print(user)
#> User(name='Jason', age=25)
Nested Example¶
from pydantic import BaseModel
from typing import List
import os
from openai import OpenAI
import instructor
from pydantic import BaseModel
class Address(BaseModel):
street: str
city: str
country: str
class User(BaseModel):
name: str
age: int
addresses: List[Address]
# Initialize with API key
client = OpenAI(api_key=os.getenv('OPENAI_API_KEY'))
# Enable instructor patches for OpenAI client
client = instructor.from_openai(client)
# Create structured output with nested objects
user = client.chat.completions.create(
model="gpt-4o-mini",
messages=[
{"role": "user", "content": """
Extract: Jason is 25 years old.
He lives at 123 Main St, New York, USA
and has a summer house at 456 Beach Rd, Miami, USA
"""},
],
response_model=User,
)
print(user)
#> {
#> 'name': 'Jason',
#> 'age': 25,
#> 'addresses': [
#> {
#> 'street': '123 Main St',
#> 'city': 'New York',
#> 'country': 'USA'
#> },
#> {
#> 'street': '456 Beach Rd',
#> 'city': 'Miami',
#> 'country': 'USA'
#> }
#> ]
#> }
Streaming Support¶
Instructor has two main ways that you can use to stream responses out
- Iterables: These are useful when you'd like to stream a list of objects of the same type (Eg. use structured outputs to extract multiple users)
- Partial Streaming: This is useful when you'd like to stream a single object and you'd like to immediately start processing the response as it comes in.
Partials¶
from instructor import from_openai
import openai
from pydantic import BaseModel
client = from_openai(openai.OpenAI())
class User(BaseModel):
name: str
age: int
bio: str
user = client.chat.completions.create_partial(
model="gpt-4o-mini",
messages=[
{"role": "user", "content": "Create a user profile for Jason, age 25"},
],
response_model=User,
)
for user_partial in user:
print(user_partial)
# > name='Jason' age=None bio='None'
# > name='Jason' age=25 bio='A tech'
# > name='Jason' age=25 bio='A tech enthusiast'
# > name='Jason' age=25 bio='A tech enthusiast who loves coding, gaming, and exploring new'
# > name='Jason' age=25 bio='A tech enthusiast who loves coding, gaming, and exploring new technologies'
Iterable Example¶
import os
from openai import OpenAI
import instructor
from pydantic import BaseModel
class User(BaseModel):
name: str
age: int
# Extract multiple users from text
users = client.chat.completions.create_iterable(
model="gpt-4o-mini",
messages=[
{"role": "user", "content": """
Extract users:
1. Jason is 25 years old
2. Sarah is 30 years old
3. Mike is 28 years old
"""},
],
response_model=User,
)
for user in users:
print(user)
#> name='Jason' age=25
#> name='Sarah' age=30
#> name='Mike' age=28
Instructor Modes¶
We provide several modes to make it easy to work with the different response models that OpenAI supports
instructor.Mode.TOOLS
: This uses the tool calling API to return structured outputs to the clientinstructor.Mode.JSON
: This forces the model to return JSON by using OpenAI's JSON mode.instructor.Mode.FUNCTIONS
: This uses OpenAI's function calling API to return structured outputs and will be deprecated in the future.instructor.Mode.PARALLEL_TOOLS
: This uses the parallel tool calling API to return structured outputs to the client. This allows the model to generate multiple calls in a single response.instructor.Mode.MD_JSON
: This makes a simple call to the OpenAI chat completion API and parses the raw response as JSON.instructor.Mode.TOOLS_STRICT
: This uses the new Open AI structured outputs API to return structured outputs to the client using constrained grammar sampling. This restricts users to a subset of the JSON schema.instructor.Mode.JSON_O1
: This is a mode for theO1
model. We created a new mode becauseO1
doesn't support any system messages, tool calling or streaming so you need to use this mode to use Instructor withO1
.
In general, we recommend using Mode.Tools
because it's the most flexible and future-proof mode. It has the largest set of features that you can specify your schema in and makes things significantly easier to work with.
Batch API¶
We also support batching requests using the create_batch
method. This is helpful if your request is not time sensitive because you'll get a 50% discount on the token cost.
Read more about how to use it here
Best Practices¶
-
Model Selection : We recommend using gpt-4o-mini for simpler use cases because it's cheap and works well with a clearly defined objective for structured outputs. When the task is more ambigious, consider upgrading to
4o
or evenO1
depending on your needs -
Performance Optimization : Streaming a response model is faster and should be done from the get-go. This is especially true if you're using a simple response model.
Common Use Cases¶
- Data Extraction
- Form Parsing
- API Response Structuring
- Document Analysis
- Configuration Generation
Related Resources¶
Updates and Compatibility¶
Instructor maintains compatibility with the latest OpenAI API versions and models. Check the changelog for updates.