Facebook iconWhy the Instructor Beats OpenAI for Structured JSON Output
Blogs/AI

Why the Instructor Beats OpenAI for Structured JSON Output

May 27, 202510 Min Read
Written by Krishna Purwar
Why the Instructor Beats OpenAI for Structured JSON Output Hero

Integrating LLMs in our code and workflow is surely exciting, but it can get tiresome quickly as we need our outputs to follow proper structure/schema, and need to validate them along the way. This is where Instructor shines, let’s go through it one step at a time, see how it performs compared to OpenAI and much more.

First, let’s understand what Structured Output means!

Structured Output means getting output in a particular schema, widely used as JSON. Strictly adhering to JSON output enables us to use LLM in traditional or AI-enabled services. 

Instructor’s library enforces strict JSON schemas via Pydantic integration, enabling immediate validation and deep-nesting support out of the box. In contrast, OpenAI’s native structured-output feature works well for small, simple schemas but quickly breaks on complex or nested data, forcing manual workarounds. 

The key takeaway: The Instructor transforms schema complexity from a blocker into a one-line response_model call, streamlining development and reducing error handling.

Why Structured Output Matters?

  • LLMs generate probabilistic texts which can sometimes be different from what we expect, and hence they need to follow a strict schema.
  • Structured Output provides us with that predictability, lets us validate the code and helps in integrating the code into pipelines due to its ease and reliability.

Overview of Instructor Library

Started by Jason Liu, now with 200+ contributors and more than 2.5M downloads per month,  Instructor ensures that we get structured output that strictly adheres to our schema.Core features: Pydantic integration, vendor-agnostic support, complexity handling

Install the core library via pip:

pip install instructor

Basic Usage Snippet:

import instructor
from openai import OpenAI
from pydantic import BaseModel

# 1. Define the shape of the data you want
class UserInfo(BaseModel):
    name: str
    age: int

# 2. Create an Instructor client for your LLM provider
client = instructor.from_openai(OpenAI())

# 3. Call the chat completions API with your Pydantic model
user = client.chat.completions.create(
    model="gpt-4.1-mini",
    response_model=UserInfo,
    messages=[{"role": "user", "content": "Alice is 28 years old."}],
)

# 4. Access validated, structured fields
print(user.name)  # Alice
print(user.age)   # 28

Overview of OpenAI’s Built-in Structured Output

  • Easy to use as it is built in, it works well with simple and small schema that only requires JSON output.
  • Limitations: Pydantic validation issues, schema complexity breakdown

Side-by-Side Comparison of Instructor Library and OpenAI Structured Output

FeatureInstructor LibraryOpenAI Structured Output

Vendor lock-in

Works with any supported LLM

Tied to OpenAI’s API

Validation support

Native Pydantic validators

Fails on complex or nested schemas

Schema complexity handling

Handles deep nesting seamlessly

Breaks on large/complex schemas

Ease of use

Single-step response_model integration

Requires manual crafting, parsing, and error handling

Error messaging

Clear Pydantic-style errors

Cryptic JSON schema errors

Vendor lock-in

Instructor Library

Works with any supported LLM

OpenAI Structured Output

Tied to OpenAI’s API

1 of 5

Simple Example: Species Extraction

In this example, we would focus on the validation of our schema and compare how both approaches perform. 

Partner with Us for Success

Experience seamless collaboration and exceptional results.

Instructor version 

Code:

# Import required libraries
from openai import OpenAI  # OpenAI's official Python client
import instructor  # Library for structured outputs from LLMs
from pydantic import BaseModel, field_validator  # For data validation
from dotenv import load_dotenv  # To load environment variables from .env file

# Load environment variables (like API keys) from .env file
load_dotenv()

# Define a data model for species information using Pydantic
# This ensures the data follows a specific structure and validation rules
class Species(BaseModel):
    name: str  # Species name as a string
    habitat: str  # Where the species lives
    average_lifespan: int  # Average lifespan in years

    # Custom validator to ensure species name is in capital letters
    @field_validator("name")
    def name_must_be_capital(cls, v: str) -> str:
        if not v.isupper():
            raise ValueError("Species name must be in capital letters.")
        return v

# Initialize the OpenAI client with instructor wrapper
# This allows us to get structured JSON responses from the model
client = instructor.from_openai(OpenAI(), mode=instructor.Mode.JSON)

# Sample text containing information about a species
paragraph = """
The African elephant is one of the largest land animals. It typically lives in savannas and forests. 
These elephants can live up to 70 years in the wild.
"""

# Make an API call to GPT-4 to extract structured information
# The response will be automatically converted to our Species model
response = client.chat.completions.create(
    model="gpt-4.1-mini-2025-04-14",  # Specify which model to use
    response_model=Species,  # Tell the model to format response as Species object
    messages=[
        {
            "role": "user",
            "content": f"Extract species information (name, habitat, average lifespan) from the following text:\n{paragraph}",
        }
    ],
)

# Print the structured response
print(response)

Output: 

name='AFRICAN ELEPHANT' habitat='savannas and forests' average_lifespan=70

The output is exactly what we want.

OpenAI version

Code:

# Import required libraries
from openai import OpenAI  # OpenAI's official Python client
from pydantic import BaseModel, field_validator, ValidationError  # For data validation and error handling
from dotenv import load_dotenv  # To load environment variables from .env file

# Load environment variables (like API keys) from .env file
load_dotenv()

# Define a data model for species information using Pydantic
# This ensures the data follows a specific structure and validation rules
class Species(BaseModel):
    name: str  # Species name as a string
    habitat: str  # Where the species lives
    average_lifespan: int  # Average lifespan in years

    # Custom validator to ensure species name is in capital letters
    # This will cause validation errors if the model returns lowercase names
    @field_validator("name")
    def name_must_be_capital(cls, v: str) -> str:
        if not v.isupper():
            raise ValueError("Species name must be in capital letters.")
        return v

# Initialize the basic OpenAI client
# Note: This version doesn't use instructor, which means we'll need to handle
# response parsing and validation manually
client = OpenAI()

# Sample text containing information about a species
# Note: The text contains lowercase "african" which will cause validation errors
# This demonstrates why instructor's automatic formatting is helpful
paragraph = """
The african elephant is one of the largest land animals. It typically lives in savannas and forests. 
These elephants can live up to 70 years in the wild.
"""

# Try to parse the response and handle potential errors
try:
    # Make an API call to GPT-4 to extract structured information
    # Note: Without instructor, we need to manually parse and validate the response
    response = client.responses.parse(
        model="gpt-4.1-mini-2025-04-14",  # Specify which model to use
        input=[
            {
                "role": "user",
                "content": f"Extract species information (name, habitat, average lifespan) from the following text:\n{paragraph}",
            }
        ],
        text_format=Species,  # Attempt to format response as Species object
    )

    # Print the parsed response if successful
    print(response.output_parsed)

except ValidationError as ve:
    # Handle validation errors (e.g., when species name isn't in capital letters)
    print("Validation Error:")
    for error in ve.errors():
        print(f"Field: {error['loc'][0]} - Error: {error['msg']}")

except Exception as e:
    # Handle any other unexpected errors
    print("An error occurred while processing the response:")
    print(str(e))

# This version is more prone to errors because:
# 1. It doesn't automatically format the response
# 2. It requires manual error handling
# 3. The model might return data in an unexpected format
# 4. We need to handle validation ourselves

Output:

Validation Error:
Field: name - Error: Value error, Species name must be in capital letters.

It fails upon validation and throws an error. On top of that, to make sure that the code doesn’t explode, I had to write extra try & except, which was not the case with the Instructor. 

Key takeaway: Instructor succeeds out of the box.

Complex Example: Financial Analysis

Model definition (nested Pydantic classes)

main.py:

from typing import Dict, List, Optional, Union, Literal
from datetime import datetime
from pydantic import BaseModel, Field

class FinancialAnalysisModel(BaseModel):
    """
    Base model for financial data analysis with nested classes
    for comparing structured output approaches.
    
    This model demonstrates a complex nested structure that can be used to:
    1. Validate financial data
    2. Ensure type safety
    3. Provide structured output for financial analysis
    4. Compare different approaches to data extraction
    """
    
    class TransactionData(BaseModel):
        """
        Nested class for detailed transaction information.
        This class enforces strict typing and validation for financial transactions.
        """
        # Enforce specific transaction types using Literal type
        transaction_type: Literal["deposit", "withdrawal", "transfer", "investment"] = Field(
            ..., description="The type of financial transaction"
        )
        amount: float = Field(..., description="Transaction amount in specified currency")
        currency_code: str = Field(..., description="Three-letter currency code")
        timestamp: datetime = Field(..., description="ISO format timestamp of the transaction")
        status: str = Field(..., description="Current status of the transaction")
        # Optional field that can be None
        notes: Optional[str] = Field(None, description="Additional transaction notes")
        # Dictionary for flexible metadata storage
        metadata: Dict[str, Union[str, float, bool]] = Field(
            default_factory=dict, description="Additional contextual information"
        )

    class AccountSummary(BaseModel):
        """
        Nested class for account summary information.
        Provides a structured way to store account-level financial data.
        """
        account_id: str = Field(..., description="Unique identifier for the account")
        account_type: str = Field(..., description="Type of financial account")
        balance: float = Field(..., description="Current balance in base currency")
        risk_level: Literal["low", "medium", "high"] = Field(
            ..., description="Risk assessment of the account"
        )
        # Dictionary to store various performance metrics
        performance_metrics: Dict[str, float] = Field(
            default_factory=dict, description="Key performance indicators for the account"
        )
        last_updated: str = Field(..., description="When the data was last refreshed")

    # Top-level fields for the financial analysis
    client_id: str = Field(..., description="Unique identifier for the client")
    analysis_date: str = Field(..., description="Date of financial analysis")
    # Dictionary mapping account IDs to their summaries
    accounts: Dict[str, AccountSummary] = Field(
        default_factory=dict, description="Map of account IDs to account summaries"
    )
    # List of transactions for the analysis period
    transactions: List[TransactionData] = Field(
        default_factory=list, description="List of relevant transactions in analysis period"
    )
    # Flexible dictionary for analysis results and recommendations
    result: Dict[str, Union[float, str, bool]] = Field(
        default_factory=dict, description="Analysis results and recommendations"
    )

# Sample text demonstrating the kind of financial data we'll be parsing
# This text contains various pieces of information that need to be extracted
# and structured according to our model
SAMPLE_TEXT = """
    Client C123456's financial analysis from May 15, 2025 shows one investment account (A001) with a balance of usd 158,432.50. 
    The account has medium risk, with YTD return of 8.3%, one-year return of 12.1%, and volatility of 6.7%. 
    The account was last updated on May 14, 2025 at 11:00 PM UTC.
    
    A recent transaction shows an investment of usd 10,000 completed on May 1, 2025 at 2:30 PM UTC, 
    noted as "Monthly portfolio contribution" in the "scheduled" category for an "index_fund".
    
    The analysis projects a 10.5% annual return, with a risk-adjusted score of 7.8. 
    Rebalancing is recommended, and tax efficiency is moderate.
"""

Instructor workflow

structured_output_instructor.py:

# Import required libraries
from openai import OpenAI  # OpenAI's official Python client
import instructor  # Library for structured outputs from LLMs
from main import FinancialAnalysisModel, SAMPLE_TEXT  # Import our data model and sample text
from dotenv import load_dotenv  # To load environment variables from .env file

# Load environment variables (like API keys) from .env file
load_dotenv()

# Initialize the OpenAI client with instructor wrapper
# This enables automatic structured output parsing and validation
client = instructor.from_openai(OpenAI(), mode=instructor.Mode.JSON)

def analyze_financial_data(text):
    """
    Analyze financial data using OpenAI's chat completions API with instructor.
    
    This function demonstrates how instructor simplifies the process of:
    1. Extracting structured data from unstructured text
    2. Validating the extracted data against our model
    3. Handling complex nested structures
    4. Ensuring type safety
    
    Args:
        text: The financial text to analyze (unstructured text containing financial information)
        
    Returns:
        Extracted financial analysis data matching FinancialAnalysisModel structure
        Returns None if an error occurs during processing
    """
    try:
        # Make API call to GPT-4 with instructor's structured output handling
        # The response will be automatically parsed and validated against FinancialAnalysisModel
        response = client.chat.completions.create(
            response_model=FinancialAnalysisModel,  # Specify our model for structured output
            messages=[
                {
                    "role": "user",
                    "content": text,  # The financial text to analyze
                }
            ],
            model="gpt-4.1-mini-2025-04-14",  # Specify which model to use
        )

        return response
        
    except Exception as e:
        # Handle any errors that occur during processing
        print(f"Error occurred: {e}")
        return None

# When running this script directly, analyze the sample text
if __name__ == "__main__":
    # Process the sample financial text and print the structured result
    result = analyze_financial_data(SAMPLE_TEXT)
    print(result)

Output:

python3 structured_outputs_instructor.py
client_id='C123456' analysis_date='2025-05-15' accounts={'A001': AccountSummary(account_id='A001', account_type='investment', balance=158432.5, risk_level='medium', performance_m
etrics={'YTD_return': 8.3, 'one_year_return': 12.1, 'volatility': 6.7}, last_updated='2025-05-14T23:00:00Z')} transactions=[TransactionData(transaction_type='investment', amount=10000.0, currency_code='USD', timestamp=datetime.datetime(2025, 5, 1, 14, 30, tzinfo=TzInfo(UTC)), status='completed', notes='Monthly portfolio contribution', metadata={'category': 'scheduled', 'type': 'index_fund'})] result={'projected_annual_return': 10.5, 'risk_adjusted_score': 7.8, 'rebalancing_recommended': True, 'tax_efficiency': 'moderate'}

The output is a mix of strings, lists and dictionaries, as defined in the FinancialAnalysisModel class, which is gracefully handled by the Instructor. 

Partner with Us for Success

Experience seamless collaboration and exceptional results.

OpenAI workflow (standard parse fails)

structured_outputs_openai.py:

# Import required libraries
from openai import OpenAI  # OpenAI's official Python client
from main import FinancialAnalysisModel, SAMPLE_TEXT  # Import our data model and sample text
import os  # For environment variable access
from dotenv import load_dotenv  # To load environment variables from .env file

# Load environment variables (like API keys) from .env file
load_dotenv()

# Initialize basic OpenAI client without instructor wrapper
# Note: This version lacks automatic structured output handling
client = OpenAI(api_key=os.getenv('OPENAI_API_KEY'))

def analyze_financial_data(text):
    """
    Analyze financial data using OpenAI's basic chat completions API.
    
    This implementation demonstrates the limitations of using OpenAI's basic client:
    1. No automatic handling of complex nested structures
    2. Manual parsing required for nested objects
    3. No built-in validation against our model
    4. More prone to errors with complex data types
    
    Args:
        text: The financial text to analyze (unstructured text containing financial information)
        
    Returns:
        Extracted financial analysis data matching FinancialAnalysisModel structure
        Returns None if an error occurs during processing
    """
    try:
        # Call the chat completions API with JSON response format
        # Note: This approach struggles with nested structures in FinancialAnalysisModel
        response = client.responses.parse(
            model="gpt-4o-2024-08-06",  # Specify which model to use
            input=[
                {
                    "role": "system",
                    "content": "You are a financial analysis assistant. Extract structured data from the financial information provided."
                },
                {
                    "role": "user",
                    "content": text  # The financial text to analyze
                }
            ],
            text_format=FinancialAnalysisModel,  # Attempt to format as our model
        )
        
        # Extract the parsed output
        # Note: This might fail for complex nested structures
        result = response.output_parsed

        return result
        
    except Exception as e:
        # Handle any errors that occur during processing
        # Common errors include:
        # - Nested structure parsing failures
        # - Type validation errors
        # - Missing required fields
        print(f"Error occurred: {e}")
        return None

# When running this script directly, analyze the sample text
if __name__ == "__main__":
    # Process the sample financial text and print the structured result
    # Note: This is likely to fail due to the complexity of our model
    result = analyze_financial_data(SAMPLE_TEXT)
    print(result)

Output:

python3 structured_outputs_openai.py
Error occurred: Error code: 400 - {'error': {'message': "Invalid schema for response_format 'FinancialAnalysisModel': In context=(), 'required' is required to be supplied and to be an ar
ray including every key in properties. Extra required key 'performance_metrics' supplied.", 'type': 'invalid_request_error', 'param': 'text.format.schema', 'code': 'invalid_json_schema'}}                                                                                                                                                                                         None

As we can see, it did not work when the schema got complex; to get the same result with OpenAI, we would have to do a complex manoeuvre, write a schema for every part and then parse it, increasing the complexity and burden, which doesn’t make any sense.

Key takeaway: The Instructor handles complexity with minimal code against this monstrosity we have to do by defining our schema again in JSON format. 

Conclusion

Instructor’s model-first design turns structured output into a predictable, maintainable process. By integrating Pydantic validators and embracing vendor-agnostic support, it ensures that any response aligns exactly with your data models, no extra parsing, no mysterious JSON errors. 

Meanwhile, relying solely on OpenAI’s built-in JSON formatting quickly becomes untenable as schema complexity grows, leaving developers to build elaborate manual workarounds. 

Ultimately, Instructor empowers teams to adopt LLM-powered pipelines without sacrificing reliability or debugging clarity.

Author-Krishna Purwar
Krishna Purwar

You can find me exploring niche topics, learning quirky things and enjoying 0 n 1s until qbits are not here-

Phone

Next for you

MCP or Function Calling: Everything You Need To Know Cover

AI

May 29, 20254 min read

MCP or Function Calling: Everything You Need To Know

Are you tired of your AI assistant giving you outdated information or saying "I can't access that" when you need real-time data? You've tried asking AI about your business data it doesn't know. You want it to check your emails or calendar it can't. You need it to look up information from your company systems which is impossible. Here's the problem: Most AI tools are disconnected from everything else you use. It's like having a really smart assistant who's locked in a separate room with no pho

What are Temperature, Top_p, and Top_k in AI? Cover

AI

May 27, 20256 min read

What are Temperature, Top_p, and Top_k in AI?

LLMs work their wonders by crafting text that feels just like human writing, predicting what word comes next in a perfect flow. The real charm happens behind the curtain, where it's all about the game of probabilities and tokens!  Let’s control this magic by fine-tuning specific parameters: temperature, top_p (nucleus sampling), and top_k sampling, making us the magicians of this magic. It's going to be an exciting exploration! What is Temperature in AI?  Range: 0 to 2 (in practice) Tempera

7 Best AI Code Editors in 2025 Cover

AI

May 9, 20258 min read

7 Best AI Code Editors in 2025

AI is everywhere these days, right? For us developers, having AI in our code editors is kinda like having a super experienced senior dev buddy who's always there to help with questions, fixing bugs, and even explaining stuff. Using AI code editors has definitely sped things up, though sometimes those fancy tools can make debugging a bit tricky. Still, it's pretty cool how they can understand entire codebases, write functions independently, fix our mistakes, and even think through some seriously