File tree Expand file tree Collapse file tree 5 files changed +38
-6
lines changed
docs/source/notebooks/tool_usage Expand file tree Collapse file tree 5 files changed +38
-6
lines changed Original file line number Diff line number Diff line change 300300 ],
301301 "source" : [
302302 " import uuid\n " ,
303+ " \n " ,
303304 " from langsmith.client import Client\n " ,
304305 " \n " ,
305306 " from langchain_benchmarks.tool_usage import get_eval_config\n " ,
306307 " \n " ,
307- " \n " ,
308308 " experiment_uuid = uuid.uuid4().hex[:4]\n " ,
309309 " \n " ,
310310 " client = Client()\n " ,
Original file line number Diff line number Diff line change 1+ from typing import Optional
2+
3+ from langchain .chat_models import ChatOpenAI
14from langchain .chat_models .base import BaseChatModel
25from langchain .smith import RunEvalConfig
36
47
5- def get_eval_config (eval_llm : BaseChatModel ) -> RunEvalConfig :
8+ def get_eval_config (eval_llm : Optional [BaseChatModel ] = None ) -> RunEvalConfig :
9+ eval_llm = eval_llm or ChatOpenAI (
10+ model = "gpt-4" ,
11+ temperature = 0 ,
12+ model_kwargs = {"seed" : 42 },
13+ max_retries = 1 ,
14+ request_timeout = 60 ,
15+ )
616 """Get the evaluation configuration for the email task."""
717 return RunEvalConfig (
818 evaluators = [
Original file line number Diff line number Diff line change @@ -61,7 +61,13 @@ def run_on_dataset(
6161 kwargs: Additional arguments to pass to the client.
6262 """
6363 client = Client ()
64- eval_llm = ChatOpenAI (model = "gpt-4" , temperature = 0.0 , model_kwargs = {"seed" : 42 })
64+ eval_llm = ChatOpenAI (
65+ model = "gpt-4" ,
66+ temperature = 0.0 ,
67+ model_kwargs = {"seed" : 42 },
68+ max_retries = 1 ,
69+ request_timeout = 60 ,
70+ )
6571 return client .run_on_dataset (
6672 dataset_name = task .name ,
6773 llm_or_chain_factory = create_openai_function_based_extractor (
Original file line number Diff line number Diff line change @@ -84,10 +84,20 @@ def evaluate_run(
8484
8585def get_eval_config () -> RunEvalConfig :
8686 """Returns the evaluator for the environment."""
87- eval_llm = ChatOpenAI (model = "gpt-4" , temperature = 0.0 , model_kwargs = {"seed" : 42 })
87+ eval_llm = ChatOpenAI (
88+ model = "gpt-4" ,
89+ temperature = 0.0 ,
90+ model_kwargs = {"seed" : 42 },
91+ max_retries = 1 ,
92+ request_timeout = 60 ,
93+ )
8894 # Use a longer-context LLM to check documents
8995 faithfulness_eval_llm = ChatOpenAI (
90- model = "gpt-4-1106-preview" , temperature = 0.0 , model_kwargs = {"seed" : 42 }
96+ model = "gpt-4-1106-preview" ,
97+ temperature = 0.0 ,
98+ model_kwargs = {"seed" : 42 },
99+ max_retries = 1 ,
100+ request_timeout = 60 ,
91101 )
92102
93103 return RunEvalConfig (
Original file line number Diff line number Diff line change @@ -99,7 +99,13 @@ class AgentTrajectoryEvaluator(RunEvaluator):
9999
100100 def __init__ (self ) -> None :
101101 """Initialize the evaluator."""
102- eval_llm = ChatOpenAI (model = "gpt-4" , temperature = 0 , model_kwargs = {"seed" : 42 })
102+ eval_llm = ChatOpenAI (
103+ model = "gpt-4" ,
104+ temperature = 0 ,
105+ model_kwargs = {"seed" : 42 },
106+ max_retries = 1 ,
107+ request_timeout = 60 ,
108+ )
103109 self .qa_evaluator = load_evaluator (EvaluatorType .QA , llm = eval_llm )
104110
105111 def evaluate_run (
You can’t perform that action at this time.
0 commit comments