Skip to content

Commit

Permalink
[evals] added eval_model flag to modelgraded eval (openai#519)
Browse files Browse the repository at this point in the history
  • Loading branch information
rlbayes committed Mar 30, 2023
1 parent f097a2f commit 2486f7e
Showing 1 changed file with 2 additions and 3 deletions.
5 changes: 2 additions & 3 deletions evals/elsuite/modelgraded/classify.py
Original file line number Diff line number Diff line change
Expand Up @@ -100,6 +100,7 @@ def __init__(
multicomp_temperature: float = 0.4,
samples_renamings: Optional[dict[str, str]] = None,
eval_type: Optional[str] = None,
eval_model: str = "gpt-3.5-turbo",
metaeval: bool = False,
modelgraded_spec_args: Optional[dict[str, dict[str, str]]] = None,
**kwargs,
Expand Down Expand Up @@ -130,9 +131,7 @@ def __init__(
if self.model_spec.name == "dummy-completion" or self.model_spec.name == "dummy-chat":
self.eval_modelspec = self.model_spec
else:
self.eval_modelspec = ModelSpec(
name="gpt-3.5-turbo", model="gpt-3.5-turbo", is_chat=True
)
self.eval_modelspec = ModelSpec(name=eval_model, model=eval_model, is_chat=True)

"""import prompt and set attributes"""
modelgraded_specs = self.registry.get_modelgraded_spec(modelgraded_spec)
Expand Down

0 comments on commit 2486f7e

Please sign in to comment.