Skip to content

Commit 6c3dcb6

Browse files
authored
Limit Retries in OpenAI (#93)
1 parent f912841 commit 6c3dcb6

File tree

5 files changed

+38
-6
lines changed

5 files changed

+38
-6
lines changed

docs/source/notebooks/tool_usage/multiverse_math.ipynb

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -300,11 +300,11 @@
300300
],
301301
"source": [
302302
"import uuid\n",
303+
"\n",
303304
"from langsmith.client import Client\n",
304305
"\n",
305306
"from langchain_benchmarks.tool_usage import get_eval_config\n",
306307
"\n",
307-
"\n",
308308
"experiment_uuid = uuid.uuid4().hex[:4]\n",
309309
"\n",
310310
"client = Client()\n",

langchain_benchmarks/extraction/evaluators.py

Lines changed: 11 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,8 +1,18 @@
1+
from typing import Optional
2+
3+
from langchain.chat_models import ChatOpenAI
14
from langchain.chat_models.base import BaseChatModel
25
from langchain.smith import RunEvalConfig
36

47

5-
def get_eval_config(eval_llm: BaseChatModel) -> RunEvalConfig:
8+
def get_eval_config(eval_llm: Optional[BaseChatModel] = None) -> RunEvalConfig:
9+
eval_llm = eval_llm or ChatOpenAI(
10+
model="gpt-4",
11+
temperature=0,
12+
model_kwargs={"seed": 42},
13+
max_retries=1,
14+
request_timeout=60,
15+
)
616
"""Get the evaluation configuration for the email task."""
717
return RunEvalConfig(
818
evaluators=[

langchain_benchmarks/extraction/implementations.py

Lines changed: 7 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -61,7 +61,13 @@ def run_on_dataset(
6161
kwargs: Additional arguments to pass to the client.
6262
"""
6363
client = Client()
64-
eval_llm = ChatOpenAI(model="gpt-4", temperature=0.0, model_kwargs={"seed": 42})
64+
eval_llm = ChatOpenAI(
65+
model="gpt-4",
66+
temperature=0.0,
67+
model_kwargs={"seed": 42},
68+
max_retries=1,
69+
request_timeout=60,
70+
)
6571
return client.run_on_dataset(
6672
dataset_name=task.name,
6773
llm_or_chain_factory=create_openai_function_based_extractor(

langchain_benchmarks/rag/evaluators.py

Lines changed: 12 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -84,10 +84,20 @@ def evaluate_run(
8484

8585
def get_eval_config() -> RunEvalConfig:
8686
"""Returns the evaluator for the environment."""
87-
eval_llm = ChatOpenAI(model="gpt-4", temperature=0.0, model_kwargs={"seed": 42})
87+
eval_llm = ChatOpenAI(
88+
model="gpt-4",
89+
temperature=0.0,
90+
model_kwargs={"seed": 42},
91+
max_retries=1,
92+
request_timeout=60,
93+
)
8894
# Use a longer-context LLM to check documents
8995
faithfulness_eval_llm = ChatOpenAI(
90-
model="gpt-4-1106-preview", temperature=0.0, model_kwargs={"seed": 42}
96+
model="gpt-4-1106-preview",
97+
temperature=0.0,
98+
model_kwargs={"seed": 42},
99+
max_retries=1,
100+
request_timeout=60,
91101
)
92102

93103
return RunEvalConfig(

langchain_benchmarks/tool_usage/evaluators.py

Lines changed: 7 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -99,7 +99,13 @@ class AgentTrajectoryEvaluator(RunEvaluator):
9999

100100
def __init__(self) -> None:
101101
"""Initialize the evaluator."""
102-
eval_llm = ChatOpenAI(model="gpt-4", temperature=0, model_kwargs={"seed": 42})
102+
eval_llm = ChatOpenAI(
103+
model="gpt-4",
104+
temperature=0,
105+
model_kwargs={"seed": 42},
106+
max_retries=1,
107+
request_timeout=60,
108+
)
103109
self.qa_evaluator = load_evaluator(EvaluatorType.QA, llm=eval_llm)
104110

105111
def evaluate_run(

0 commit comments

Comments
 (0)