Test a ReAct agent with Pytest/Vitest and LangSmith
This tutorial will show you how to use LangSmith's integrations with popular testing tools Pytest and Vitest/Jest to evaluate your LLM application. We will create a ReAct agent that answers questions about publicly traded stocks and write a comprehensive test suite for it.
Setup
This tutorial uses LangGraph for agent orchestration, OpenAI's GPT-4o, Tavily for search, E2B's code interpreter, and Polygon to retrieve stock data but it can be adapted for other frameworks, models and tools with minor modifications. Tavily, E2B and Polygon are free to sign up for.
Installation
First, install the packages required for making the agent:
- Python
- TypeScript
pip install -U langgraph langchain langchain-openai langchain-community e2b-code-interpreter
yarn add @langchain/openai @langchain/community @langchain/langgraph @langchain/core @e2b/code-interpreter @polygon.io/client-js openai zod
Next, install the testing framework:
- Pytest
- Vitest
- Jest
# Make sure you have langsmith>=0.3
pip install -U "langsmith[pytest]"
yarn add -D langsmith vitest
yarn add -D langsmith jest
Environment Variables
Set the following environment variables:
export LANGSMITH_TRACING=true
export LANGSMITH_API_KEY=<YOUR_LANGSMITH_API_KEY>
export OPENAI_API_KEY=<YOUR_OPENAI_API_KEY>
export TAVILY_API_KEY=<YOUR_TAVILY_API_KEY>
export E2B_API_KEY=<YOUR_E2B_API_KEY>
export POLYGON_API_KEY=<YOUR_POLYGON_API_KEY>
Create your app
To define our React agent, we will use LangGraph/LangGraph.js for the orchestation and LangChain for the LLM and tools.
Define tools
First we are going to define the tools we are going to use in our agent. There are going to be 3 tools:
- A search tool using Tavily
- A code interpreter tool using E2B
- A stock information tool using Polygon
Search Tool
We will use Tavily for the search tool.
- Python
- TypeScript
from langchain_community.tools import TavilySearchResults
search_tool = TavilySearchResults(
max_results=5,
include_raw_content=True,
)
import { TavilySearchResults } from "@langchain/community/tools/tavily_search";
const searchTool = new TavilySearchResults({
maxResults: 5,
});
Code Interpreter Tool
We will use E2B for the code interpreter tool.
- Python
- TypeScript
from e2b_code_interpreter import Sandbox
def code_tool(code: str) -> str:
"""Execute python code and return the result."""
sbx = Sandbox()
execution = sbx.run_code(code)
if execution.error:
return f"Error: {execution.error}"
return f"Results: {execution.results}, Logs: {execution.logs}"
import { Sandbox } from "@e2b/code-interpreter";
import { tool } from "@langchain/core/tools";
import { z } from "zod";
const codeTool = tool(
async (input): Promise<string> => {
const sbx = await Sandbox.create();
const execution = await sbx.runCode(input.code);
if (execution.error) {
return `Error: ${execution.error}`;
}
return `Results: ${execution.results}, Logs: ${execution.logs}`;
},
{
name: "code",
description: "Execute python code and return the result.",
schema: z.object({
code: z.string().describe("The python code to execute"),
}),
}
);
Stock Information Tool
We will use Polygon to get real-time data about stocks.
- Python
- TypeScript
from langchain_community.tools import PolygonAggregates
from langchain_community.utilities.polygon import PolygonAPIWrapper
polygon_aggregates = PolygonAggregates(api_wrapper=PolygonAPIWrapper())
import { restClient } from "@polygon.io/client-js";
import { tool } from "@langchain/core/tools";
import { z } from "zod";
const TickerToolInputSchema = z.object({
ticker: z.string().describe("The ticker symbol of the stock"),
timespan: z.enum(["minute", "hour", "day", "week", "month", "quarter", "year"]).describe("The size of the time window."),
timespan_multiplier: z.number().describe("The multiplier for the time window"),
from_date: z
.string()
.describe("The date to start pulling data from, YYYY-MM-DD format - ONLY include the year, month, and day"),
to_date: z
.string()
.describe("The date to stop pulling data, YYYY-MM-DD format - ONLY include the year, month, and day"),
});
const rest = restClient(process.env.POLYGON_API_KEY);
const tickerTool = tool(async (query): Promise<string> =>
{
const parsed = TickerToolInputSchema.parse(query);
const result = await rest.stocks.aggregates(
parsed.ticker,
parsed.timespan_multiplier,
parsed.timespan,
parsed.from_date,
parsed.to_date
);
return JSON.stringify(result);
},
{
name: "ticker",
description: "Pull data for the ticker",
schema: TickerToolInputSchema,
}
);
Define agent
Now that we have defined all of our tools, we can use LangGraph's create_react_agent
/createReactAgent
to create our agent.
- Python
- TypeScript
from typing import Optional
from typing_extensions import Annotated, TypedDict
from langgraph.prebuilt import create_react_agent
class AgentOutputFormat(TypedDict):
numeric_answer: Annotated[Optional[float], ..., "The numeric answer, if the user asked for one"]
text_answer: Annotated[Optional[str], ..., "The text answer, if the user asked for one"]
reasoning: Annotated[str, ..., "The reasoning behind the answer"]
agent = create_react_agent(
model="openai:gpt-4o-mini",
tools=[code_tool, search_tool, polygon_aggregates],
response_format=AgentOutputFormat,
state_modifier="You are a financial expert. Respond to the users query accurately",
)
import { z } from "zod";
import { ChatOpenAI } from "@langchain/openai";
import { createReactAgent } from "@langchain/langgraph/prebuilt";
const AgentOutputFormatSchema = z.object({
numeric_answer: z.number().optional().describe("The numeric answer, if the user asked for one"),
text_answer: z.string().optional().describe("The text answer, if the user asked for one"),
reasoning: z.string().describe("The reasoning behind the answer"),
})
const tools = [codeTool, searchTool, tickerTool];
const agent = createReactAgent({
llm: new ChatOpenAI({ model: "gpt-4o" }),
tools: tools,
responseFormat: AgentOutputFormatSchema,
stateModifier: "You are a financial expert. Respond to the users query accurately",
});
export default agent;
Write tests
Now that we have defined our agent, let's write a few tests to ensure basic functionality. In this tutorial we are going to test whether the agent's tool calling abilities are working, whether the agent knows to ignore irrelevant questions, and whether it is able to answer complex questions that involve using all of the tools.
Setup test file
We need to first set up a test file and add the imports needed at the top of the file.
- Pytest
- Vitest
- Jest
Create a tests/test_agent.py
file.
# from app import agent, polygon_aggregates, search_tool # import from wherever your agent is defined
import pytest
from langsmith import testing as t
Name your test file agent.vitest.eval.ts
import { expect } from "vitest";
import * as ls from "langsmith/vitest";
import agent from "../../src/agent"; // import from wherever your agent is defined
Name your test file agent.jest.eval.ts
import { expect } from "@jest/globals";
import * as ls from "langsmith/jest";
import agent from "../../src/agent"; // import from wherever your agent is defined
Test 1: Ignore off-topic questions
The first test will be a simple check that the agent does not use tools on irrelevant queries.
- Pytest
- Vitest
- Jest
@pytest.mark.langsmith
@pytest.mark.parametrize( # <-- Can still use all normal pytest markers
"query",
["Hello!", "How are you doing?"],
)
def test_no_tools_on_offtopic_query(query: str) -> None:
"""Test that the agent does not use tools on offtopic queries."""
# Log the test example
t.log_inputs({"query": query})
expected = []
t.log_reference_outputs({"tool_calls": expected})
# Call the agent's model node directly instead of running the ReACT loop.
result = agent.nodes['agent'].invoke(
{"messages": [{"role": "user", "content": query}]}
)
actual = result['messages'][1].tool_calls
t.log_outputs({"tool_calls": actual})
# Check that no tool calls were made.
assert actual == expected
Remember to place each individual test inside a describe block
`ls.describe("Agent Tests", () => {
// PLACE TESTS Here
})`
ls.test.each([
{ inputs: { query: "Hello!" }, expected: { numMessages: 2 } },
{ inputs: { query: "How are you doing?" }, expected: { numMessages: 2 } },
])(
"should not use tools on offtopic query: %s",
async ({ inputs: { query }, expected: { numMessages } }) => {
const result = await agent.invoke({ messages: [{ role: "user", content: query }] });
ls.logOutputs(result);
// Check that the flow was HUMAN -> AI FINAL RESPONSE (no tools called)
expect(result.messages).toHaveLength(numMessages);
}
);
Remember to place each individual test inside a describe block
`ls.describe("Agent Tests", () => {
// PLACE TESTS Here
})`
ls.test.each([
{ inputs: { query: "Hello!" }, expected: { numMessages: 2 } },
{ inputs: { query: "How are you doing?" }, expected: { numMessages: 2 } },
])(
"should not use tools on offtopic query: %s",
async ({ inputs: { query }, expected: { numMessages } }) => {
const result = await agent.invoke({ messages: [{ role: "user", content: query }] });
ls.logOutputs(result);
// Check that the flow was HUMAN -> AI FINAL RESPONSE (no tools called)
expect(result.messages).toHaveLength(numMessages);
}
);
Test 2: Simple tool calling
For tool calling, we are going to verify that the agent calls the correct tool with the correct parameters.
- Pytest
- Vitest
- Jest
@pytest.mark.langsmith
def test_searches_for_correct_ticker() -> None:
"""Test that the model looks up the correct ticker on simple query.""
# Log the test example
query = "What is the price of Aple?"
t.log_inputs({"query": query})
expected = "AAPL"
t.log_reference_outputs({"ticker": expected})
# Call the agent's model node directly instead of running the full ReACT loop.
result = agent.nodes['agent'].invoke(
{"messages": [{"role": "user", "content": query}]}
)
tool_calls = result["messages"][1].tool_calls
if tool_calls[0]['name'] == polygon_aggregates.name:
actual = tool_calls[0]['args']['ticker']
else:
actual = None
t.log_outputs({"ticker": actual})
# Check that the right ticker was queried
assert actual == expected
Remember to place each individual test inside a describe block
`ls.describe("Agent Tests", () => {
// PLACE TESTS Here
})`
ls.test(
"should search for correct ticker",
{
inputs: { query: "What is the price of Aple?" },
expected: { numMessages: 4 },
},
async ({ inputs: { query }, expected: { numMessages } }) => {
const result = await agent.invoke({
messages: [{ role: "user", content: query }],
});
ls.logOutputs(result);
// The agent should have made a single tool call to the ticker tool
const toolCalls = (result.messages[1] as AIMessage).tool_calls || [];
const tickerQuery = JSON.parse(toolCalls[0].function.arguments).query.ticker;
// Check that the right ticker was queried
expect(tickerQuery).toBe("AAPL");
// Check that the flow was HUMAN -> AI -> TOOL -> AI FINAL RESPONSE
expect(result.messages).toHaveLength(numMessages);
}
);
Remember to place each individual test inside a describe block
`ls.describe("Agent Tests", () => {
// PLACE TESTS Here
})`
ls.test(
"should search for correct ticker",
{
inputs: { query: "What is the price of Aple?" },
expected: { numMessages: 4 },
},
async ({ inputs: { query }, expected: { numMessages } }) => {
const result = await agent.invoke({
messages: [{ role: "user", content: query }],
});
ls.logOutputs(result);
// The agent should have made a single tool call to the ticker tool
const toolCalls = (result.messages[1] as AIMessage).tool_calls || [];
const tickerQuery = JSON.parse(toolCalls[0].function.arguments).query.ticker;
// Check that the right ticker was queried
expect(tickerQuery).toBe("AAPL");
// Check that the flow was HUMAN -> AI -> TOOL -> AI FINAL RESPONSE
expect(result.messages).toHaveLength(numMessages);
}
);
Test 3: Complex tool calling
Some tool calls are easier to test than others. With the ticker lookup, we can assert that the correct ticker is searched. With the coding tool, the inputs and outputs of the tool are much less constrained, and there are lots of ways to get to the right answer. In this case, it's simpler to test that the tool is used correctly by running the full agent and asserting that it both calls the coding tool and that it ends up with the right answer.
- Pytest
- Vitest
- Jest
@pytest.mark.langsmith
def test_executes_code_when_needed() -> None:
query = "What was the average return rate for FAANG stock in 2024?"
t.log_inputs({"query": query})
expected = 53
t.log_reference_outputs({"response": expected})
# Test that the agent executes code when needed
result = agent.invoke({"messages": [("user", user_query)]})
t.log_outputs({"result": result['structured_response'].get("numeric_answer")})
# Grab all the tool calls made by the LLM
tool_calls = [tc['name'] for msg in result['messages'] for tc in getattr(msg, "tool_calls", [])]
# This will log the number of steps taken by the agent, which is useful for
# determining how efficiently the agent gets to an answer.
t.log_feedback(key="num_steps", value=len(result['messages']) - 1)
# Assert that the code tool was used
assert "code_tools" in tool_calls
# Assert that a numeric answer was provided:
assert result['structured_response'].get('numeric_answer') is not None
# Assert that the answer is within 1 of the expected value
assert abs(result['structured_response']['numeric_answer'] - 53) <= 1
Remember to place each individual test inside a describe block
`ls.describe("Agent Tests", () => {
// PLACE TESTS Here
})`
ls.test(
"should execute code when needed",
{
inputs: { query: "What was the average return rate for FAANG stock in 2024?" },
expected: { answer: 53 },
},
async ({ inputs: { query }, expected: { answer } }) => {
const result = await agent.invoke({
messages: [{ role: "user", content: query }],
});
ls.logOutputs(result);
// Grab all the tool calls made by the LLM
const toolCalls = result.messages
.filter(m => (m as AIMessage).tool_calls)
.flatMap(m => (m as AIMessage).tool_calls?.map(tc => tc.name));
// This will log the number of steps taken by the LLM, which we can track over time to measure performance
ls.logFeedback({
key: "num_steps",
score: result.messages.length - 1, // The first message is the user message
});
// Assert that the tool calls include the "code_tool" function
expect(toolCalls).toContain("code_tool");
// Assert that the answer is within 1 of the expected answer
expect(Math.abs(result.structured_response.numeric_answer - answer)).toBeLessThanOrEqual(1);
}
);
Remember to place each individual test inside a describe block
`ls.describe("Agent Tests", () => {
// PLACE TESTS Here
})`
ls.test(
"should execute code when needed",
{
inputs: { query: "What was the average return rate for FAANG stock in 2024?" },
expected: { answer: 53 },
},
async ({ inputs: { query }, expected: { answer } }) => {
const result = await agent.invoke({
messages: [{ role: "user", content: query }],
});
ls.logOutputs(result);
// Grab all the tool calls made by the LLM
const toolCalls = result.messages
.filter(m => (m as AIMessage).tool_calls)
.flatMap(m => (m as AIMessage).tool_calls?.map(tc => tc.name));
// This will log the number of steps taken by the LLM, which we can track over time to measure performance
ls.logFeedback({
key: "num_steps",
score: result.messages.length - 1, // The first message is the user message
});
// Assert that the tool calls include the "code_tool" function
expect(toolCalls).toContain("code_tool");
// Assert that the answer is within 1 of the expected answer
expect(Math.abs(result.structured_response.numeric_answer - answer)).toBeLessThanOrEqual(1);
}
);
Test 4: LLM-as-a-judge
We are going to ensure that the agent's answer is grounded in the search results by running an LLM-as-a-judge evaluation.
In order to trace the LLM as a judge call separately from our agent, we will use the LangSmith provided trace_feedback
context manager
in Python and wrapEvaluator
function in JS/TS.
- Pytest
- Vitest
- Jest
from typing_extensions import Annotated, TypedDict
class Grade(TypedDict):
"""Evaluate the groundedness of an answer in source documents."""
score: Annotated[bool, ..., "Return True if the answer is fully frounded in the source documents, otherwise False."]
judge_llm = init_chat_model("gpt-4o").with_structured_output(Grade)
@pytest.mark.langsmith
def test_grounded_in_source_info() -> None:
"""Test that response is grounded in the tool outputs."""
query = "How did Nvidia stock do in 2024 according to analysts?"
t.log_inputs({"query": query})
result = agent.invoke({"messages": [{"role": "user", "content": query})]})
# Grab all the search calls made by the LLM
search_results = "\n\n".join(
msg.content
for msg in result["messages"]
if msg.type == "tool" and msg.name == search_tool.name
)
t.log_outputs({
"response": result["structured_response"].get("text_answer"),
"search_results": search_results
})
# Trace the feedback LLM run separately from the agent run.
with t.trace_feedback():
# Instructions for the LLM judge
instructions = (
"Grade the following ANSWER. "
"The ANSWER should be fully grounded in (i.e. supported by) the source DOCUMENTS. "
"Return True if the ANSWER is fully grounded in the DOCUMENTS. "
"Return False if the ANSWER is not grounded in the DOCUMENTS."
)
answer_and_docs = (
f"ANSWER: {result['structured_response'].get("text_answer", "")}\n"
f"DOCUMENTS:\n{search_results}"
)
# Run the judge LLM
grade = judge_llm.invoke([
{"role": "system", "content": instructions},
{"role": "user", "content": answer_and_docs},
])
t.log_feedback(key="groundedness", score=grade.score)
assert score
// THIS CODE GOES OUTSIDE THE TEST - IT IS JUST A HELPER FUNCTION
const judgeLLM = new ChatOpenAI({ model: "gpt-4o" });
const groundedEvaluator = async (params: {
answer: string;
referenceDocuments: string,
}) => {
// Instructions for the LLM judge
const instructions = [
"Return 1 if the ANSWER is grounded in the DOCUMENTS",
"Return 0 if the ANSWER is not grounded in the DOCUMENTS",
].join("\n");
// Run the judge LLM
const grade = await judgeLLM.invoke([
{ role: "system", content: instructions },
{ role: "user", content: `ANSWER: ${params.answer}\nDOCUMENTS: ${params.referenceDocuments}` },
]);
const score = parseInt(grade.content.toString());
return { key: "groundedness", score };
};
// THIS CODE GOES INSIDE THE TEST
ls.test(
"grounded in the source",
{
inputs: { query: "How did Nvidia stock do in 2024 according to analysts?" },
},
async ({ inputs: { query } }) => {
const result = await agent.invoke({
messages: [{ role: "user", content: query }],
});
const wrappedEvaluator = ls.wrapEvaluator(groundedEvaluator);
await wrappedEvaluator({
answer: result.structuredResponse.text_answer ?? "",
referenceDocuments: result.structuredResponse.reasoning,
})
ls.logOutputs(result);
});
// THIS CODE GOES OUTSIDE THE TEST - IT IS JUST A HELPER FUNCTION
const judgeLLM = new ChatOpenAI({ model: "gpt-4o" });
const groundedEvaluator = async (params: {
answer: string;
referenceDocuments: string,
}) => {
const instructions = [
"Return 1 if the ANSWER is grounded in the DOCUMENTS",
"Return 0 if the ANSWER is not grounded in the DOCUMENTS",
].join("\n");
const grade = await judgeLLM.invoke([
{ role: "system", content: instructions },
{ role: "user", content: `ANSWER: ${params.answer}\nDOCUMENTS: ${params.referenceDocuments}` },
]);
const score = parseInt(grade.content.toString());
return { key: "groundedness", score };
};
// THIS CODE GOES INSIDE THE TEST
ls.test(
"grounded in the source",
{
inputs: { query: "How did Nvidia stock do in 2024 according to analysts?" },
},
async ({ inputs: { query } }) => {
const result = await agent.invoke({
messages: [{ role: "user", content: query }],
});
const wrappedEvaluator = ls.wrapEvaluator(groundedEvaluator);
await wrappedEvaluator({
answer: result.structuredResponse.text_answer ?? "",
referenceDocuments: result.structuredResponse.reasoning,
})
ls.logOutputs(result);
});
Run tests
Once you have setup your config files (if you are using Vitest or Jest), you can run your tests using the following commands:
Config files for Vitest/Jest
- Vitest
- Jest
import { defineConfig } from "vitest/config";
export default defineConfig({
test: {
include: ["**/*.eval.?(c|m)[jt]s"],
reporters: ["langsmith/vitest/reporter"],
setupFiles: ["dotenv/config"],
},
});
require('dotenv').config();
module.exports = {
preset: 'ts-jest',
testEnvironment: 'node',
testMatch: [
'<rootDir>/tests/jest/**/*.jest.eval.ts'
],
testPathIgnorePatterns: [
'<rootDir>/tests/vitest/.*.vitest.eval.ts$'
],
reporters: ["langsmith/jest/reporter"],
};
- Pytest
- Vitest
- Jest
pytest --output='ls' tests
yarn vitest --config ls.vitest.config.ts
yarn jest --config ls.jest.config.ts
Reference code
Remember to also add the config files for Vitest and Jest to your project.
Agent
Agent code
- Pytest
- TypeScript
from langchain_openai import ChatOpenAI
from langgraph.prebuilt import create_react_agent
from langchain_community.tools import TavilySearchResults
from e2b_code_interpreter import Sandbox
from langchain_community.tools.polygon.aggregates import PolygonAggregates
from langchain_community.utilities.polygon import PolygonAPIWrapper
from typing_extensions import Annotated, TypedDict, Optional, Literal
def code_tool(code: str) -> str:
"""Execute python code and return the result."""
sbx = Sandbox()
execution = sbx.run_code(code)
if execution.error:
return f"Error: {execution.error}"
return f"Results: {execution.results}, Logs: {execution.logs}"
class TickerToolInput(TypedDict):
"""Input format for the ticker tool.
The tool will pull data in aggregate blocks (timespan_multiplier * timespan) from the from_date to the to_date
"""
ticker: Annotated[str, ..., "The ticker symbol of the stock"]
timespan: Annotated[Literal["minute", "hour", "day", "week", "month", "quarter", "year"], ..., "The size of the time window."]
timespan_multiplier: Annotated[int, ..., "The multiplier for the time window"]
from_date: Annotated[str, ..., "The date to start pulling data from, YYYY-MM-DD format - ONLY include the year month and day"]
to_date: Annotated[str, ..., "The date to stop pulling data, YYYY-MM-DD format - ONLY include the year month and day"]
api_wrapper = PolygonAPIWrapper()
polygon_aggregate = PolygonAggregates(api_wrapper=api_wrapper)
def ticker_tool(query: TickerToolInput) -> str:
"""Pull data for the ticker."""
return polygon_aggregate.invoke(query)
search_tool = TavilySearchResults(
max_results=5,
include_raw_content=True,
)
class AgentOutputFormat(TypedDict):
numeric_answer: Annotated[Optional[float], ..., "The numeric answer, if the user asked for one"]
text_answer: Annotated[Optional[str], ..., "The text answer, if the user asked for one"]
reasoning: Annotated[str, ..., "The reasoning behind the answer"]
tools = [code_tool, search_tool, polygon_aggregates]
agent = create_react_agent(
model=ChatOpenAI(model="gpt-4o-mini"),
tools=tools,
response_format=AgentOutputFormat,
state_modifier="You are a financial expert. Respond to the users query accurately",
)
import { ChatOpenAI } from "@langchain/openai";
import { createReactAgent } from "@langchain/langgraph/prebuilt";
import { TavilySearchResults } from "@langchain/community/tools/tavily_search";
import { Sandbox } from '@e2b/code-interpreter'
import { restClient } from '@polygon.io/client-js';
import { tool } from "@langchain/core/tools";
import { z } from "zod";
const codeTool = tool(
async (input): Promise<string> => {
const sbx = await Sandbox.create();
const execution = await sbx.runCode(input.code);
if (execution.error) {
return `Error: ${execution.error}`;
}
return `Results: ${execution.results}, Logs: ${execution.logs}`;
},
{
name: "code",
description: "Execute python code and return the result.",
schema: z.object({
code: z.string().describe("The python code to execute"),
}),
}
);
const TickerToolInputSchema = z.object({
ticker: z.string().describe("The ticker symbol of the stock"),
timespan: z.enum(["minute", "hour", "day", "week", "month", "quarter", "year"]).describe("The size of the time window."),
timespan_multiplier: z.number().describe("The multiplier for the time window"),
from_date: z
.string()
.describe("The date to start pulling data from, YYYY-MM-DD format - ONLY include the year, month, and day"),
to_date: z
.string()
.describe("The date to stop pulling data, YYYY-MM-DD format - ONLY include the year, month, and day"),
});
const rest = restClient(process.env.POLYGON_API_KEY);
const tickerTool = tool(async (query): Promise<string> =>
{
const parsed = TickerToolInputSchema.parse(query);
const result = await rest.stocks.aggregates(
parsed.ticker,
parsed.timespan_multiplier,
parsed.timespan,
parsed.from_date,
parsed.to_date
);
return JSON.stringify(result);
},
{
name: "ticker",
description: "Pull data for the ticker",
schema: TickerToolInputSchema,
}
);
const searchTool = new TavilySearchResults({
maxResults: 5,
});
const AgentOutputFormatSchema = z.object({
numeric_answer: z.number().optional().describe("The numeric answer, if the user asked for one"),
text_answer: z.string().optional().describe("The text answer, if the user asked for one"),
reasoning: z.string().describe("The reasoning behind the answer"),
})
const tools = [codeTool, searchTool, tickerTool];
const agent = createReactAgent({
llm: new ChatOpenAI({ model: "gpt-4o" }),
tools: tools,
responseFormat: AgentOutputFormatSchema,
stateModifier: "You are a financial expert. Respond to the users query accurately",
});
export default agent;
Tests
Test code
- Pytest
- Vitest
- Jest
from agent import agent
import pytest
from langsmith import testing as t
from langchain import init_chat_model
@pytest.mark.langsmith
@pytest.mark.parametrize(
"question", ["Hello!", "How are you doing?"],
)
def test_no_tools_on_offtopic_query(question: str) -> None:
"""Test that the agent does not use tools on offtopic queries."""
result = agent.invoke({"messages": [("user", question)]})
# first human message and then the AI response
assert len(result['messages']) == 2
@pytest.mark.langsmith
def test_looks_up_correct_ticker():
"""Test that the agent gets the price of the correct ticker."""
question = "What is the price of Aple?"
t.log_inputs({"question": question})
expected = "AAPL"
t.log_reference_outputs({"ticker": expected})
result = agent.invoke({"messages": [{"role": "user", "content": question}]})
for msg in result['messages'][1:]:
if hasattr(msg, "tool_calls") and any(tc['name'] == )
actual = result['messages'][1].tool_calls[0]['args']["query"]["ticker"]
t.log_outputs({"ticker": actual})
t.log_feedback(key="num_messages", score=len(result['messages']))
assert actual == expected
@pytest.mark.langsmith
def test_executes_code_when_needed():
# Test that the agent executes code when needed
result = agent.invoke({"messages": [("user", "What was the average return rate for FAANG stock in 2024?")]})
tool_calls = [tc['function']['name'] for m in result['messages'] if m.additional_kwargs and 'tool_calls' in m.additional_kwargs for tc in m.additional_kwargs['tool_calls'] ]
t.log_feedback(key="num_steps", score=len(result['messages']) - 1)
assert "code_tool" in tool_calls
assert abs(result['structured_response']['numeric_answer'] - 53) <= 1
judge_llm = init_chat_model("gpt-4o")
@pytest.mark.langsmith
def test_grounded_in_the_source():
# Test that the agent is grounded in the source
result = agent.invoke({"messages": [("user", "How did Nvidia stock do in 2024 according to analysts?")]})
tavily_tool_message_indices = []
for i, m in enumerate(result['messages']):
if m.additional_kwargs and 'tool_calls' in m.additional_kwargs and m.additional_kwargs['tool_calls'][0]['function']['name'] == 'tavily_search_results_json':
tavily_tool_message_indices.append(i+1)
documents = "\n".join([m.content for i,m in enumerate(result['messages']) if i in tavily_tool_message_indices])
with t.trace_feedback():
instructions = (
"Return 1 if the ANSWER is grounded in the DOCUMENTS"
"Return 0 if the ANSWER is not grounded in the DOCUMENTS"
)
grade = judge_llm.invoke([
{"role": "system", "content": instructions},
{"role": "user", "content": f"ANSWER: {result['structured_response']['text_answer']}\nDOCUMENTS: {documents}"},
])
score = int(grade.content)
t.log_feedback(key="groundedness", score=score)
assert score
import { expect } from "vitest";
import * as ls from "langsmith/vitest";
import agent from "../../src/agent";
import { AIMessage, ToolMessage } from "@langchain/core/messages";
import { ChatOpenAI } from "@langchain/openai";
const judgeLLM = new ChatOpenAI({ model: "gpt-4o" });
const groundedEvaluator = async (params: {
answer: string;
referenceDocuments: string,
}) => {
const instructions = [
"Return 1 if the ANSWER is grounded in the DOCUMENTS",
"Return 0 if the ANSWER is not grounded in the DOCUMENTS",
].join("\n");
const grade = await judgeLLM.invoke([
{ role: "system", content: instructions },
{ role: "user", content: `ANSWER: ${params.answer}\nDOCUMENTS: ${params.referenceDocuments}` },
]);
const score = parseInt(grade.content.toString());
return { key: "groundedness", score };
}
ls.describe("Agent Tests", () => {
ls.test.each([
{ inputs: { query: "Hello!" }, referenceOutputs: { numMessages: 2 } },
{ inputs: { query: "How are you doing?" }, referenceOutputs: { numMessages: 2 } },
])(
"should not use tools on offtopic query: %s",
async ({ inputs: { query }, referenceOutputs: { numMessages } }) => {
const result = await agent.invoke({
messages: [{ role: "user", content: query }],
});
ls.logOutputs(result);
expect(result.messages).toHaveLength(numMessages);
}
);
ls.test(
"should search for correct ticker",
{
inputs: { query: "What is the price of Aple?" },
referenceOutputs: { numMessages: 4 },
},
async ({ inputs: { query }, referenceOutputs: { numMessages } }) => {
const result = await agent.invoke({
messages: [{ role: "user", content: query }],
});
const toolCalls = (result.messages[1] as AIMessage).tool_calls || [];
const tickerQuery = toolCalls[0].args.ticker;
ls.logOutputs(result);
expect(tickerQuery).toBe("AAPL");
expect(result.messages).toHaveLength(numMessages);
}
);
ls.test(
"should execute code when needed",
{
inputs: { query: "What was the average return rate for FAANG stock in 2024?" },
referenceOutputs: { answer: 53 },
},
async ({ inputs: { query }, referenceOutputs: { answer } }) => {
const result = await agent.invoke({
messages: [{ role: "user", content: query }],
});
const toolCalls = result.messages
.filter(m => (m as AIMessage).tool_calls)
.flatMap(m => (m as AIMessage).tool_calls?.map(tc => tc.name));
ls.logFeedback({
key: "num_steps",
score: result.messages.length - 1,
});
ls.logOutputs(result);
expect(toolCalls).toContain("code_tool");
expect(Math.abs((result.structuredResponse.numeric_answer ?? 0 - answer) - answer)).toBeLessThanOrEqual(1);
}
);
ls.test(
"grounded in the source",
{
inputs: { query: "How did Nvidia stock do in 2024?" },
referenceOutputs: {},
},
async ({ inputs: { query }, referenceOutputs: {} }) => {
const result = await agent.invoke({
messages: [{ role: "user", content: query }],
});
const referenceDocuments = result.messages
.filter((m): m is ToolMessage => m.name?.includes('tavily_search_results_json') ?? false)
.map(m => m.content)
.join('\n');
const wrappedEvaluator = ls.wrapEvaluator(groundedEvaluator);
await wrappedEvaluator({
answer: result.structuredResponse.text_answer ?? "",
referenceDocuments: referenceDocuments,
})
ls.logOutputs(result);
});
});
import { expect } from "@jest/globals";
import * as ls from "langsmith/jest";
import agent from "../../src/agent";
import { AIMessage } from "@langchain/core/messages";
import { ChatOpenAI } from "@langchain/openai";
const judgeLLM = new ChatOpenAI({ model: "gpt-4o" });
const groundedEvaluator = async (params: {
answer: string;
referenceDocuments: string,
}) => {
const instructions = [
"Return 1 if the ANSWER is grounded in the DOCUMENTS",
"Return 0 if the ANSWER is not grounded in the DOCUMENTS",
].join("\n");
const grade = await judgeLLM.invoke([
{ role: "system", content: instructions },
{ role: "user", content: `ANSWER: ${params.answer}\nDOCUMENTS: ${params.referenceDocuments}` },
]);
const score = parseInt(grade.content.toString());
return { key: "groundedness", score };
}
ls.describe("Agent Tests", () => {
ls.test.each([
{ inputs: { query: "Hello!" }, referenceOutputs: { numMessages: 2 } },
{ inputs: { query: "How are you doing?" }, referenceOutputs: { numMessages: 2 } },
])(
"should not use tools on offtopic query: %s",
async ({ inputs: { query }, referenceOutputs: { numMessages } }) => {
const result = await agent.invoke({
messages: [{ role: "user", content: query }],
});
ls.logOutputs(result);
expect(result.messages).toHaveLength(numMessages);
}
);
ls.test(
"should search for correct ticker",
{
inputs: { query: "What is the price of Aple?" },
referenceOutputs: { numMessages: 4 },
},
async ({ inputs: { query }, referenceOutputs: { numMessages } }) => {
const result = await agent.invoke({
messages: [{ role: "user", content: query }],
});
const toolCalls = (result.messages[1] as AIMessage).tool_calls || [];
const tickerQuery = toolCalls[0].args.ticker;
ls.logOutputs(result);
expect(tickerQuery).toBe("AAPL");
expect(result.messages).toHaveLength(numMessages);
}
);
ls.test(
"should execute code when needed",
{
inputs: { query: "What was the average return rate for FAANG stock in 2024?" },
referenceOutputs: { answer: 53 },
},
async ({ inputs: { query }, referenceOutputs: { answer } }) => {
const result = await agent.invoke({
messages: [{ role: "user", content: query }],
});
const toolCalls = result.messages
.filter(m => (m as AIMessage).tool_calls)
.flatMap(m => (m as AIMessage).tool_calls?.map(tc => tc.name));
ls.logFeedback({
key: "num_steps",
score: result.messages.length - 1,
});
ls.logOutputs(result);
expect(toolCalls).toContain("code_tool");
expect(Math.abs(result.structuredResponse.numeric_answer ?? 0 - answer)).toBeLessThanOrEqual(1);
}
);
ls.test(
"grounded in the source",
{
inputs: { query: "How did Nvidia stock do in 2024 according to analysts?" },
referenceOutputs: {},
},
async ({ inputs: { query }, referenceOutputs: {} }) => {
const result = await agent.invoke({
messages: [{ role: "user", content: query }],
});
const wrappedEvaluator = ls.wrapEvaluator(groundedEvaluator);
await wrappedEvaluator({
answer: result.structuredResponse.text_answer ?? "",
referenceDocuments: result.structuredResponse.reasoning,
})
ls.logOutputs(result);
});
});