Skip to content

Commit

Permalink
fix mypy issue
Browse files Browse the repository at this point in the history
  • Loading branch information
bugsz committed Jun 26, 2024
1 parent aefb489 commit af5f725
Showing 1 changed file with 3 additions and 3 deletions.
6 changes: 3 additions & 3 deletions sotopia/cli/benchmark/benchmark.py
Original file line number Diff line number Diff line change
Expand Up @@ -360,7 +360,7 @@ def benchmark_all(
batch_size: int = typer.Option(10, help="The batch size you want to use."),
task: str = typer.Option("hard", help="The task id you want to benchmark."),
print_logs: bool = typer.Option(False, help="Print logs."),
):
) -> None:
for model in model_list:
benchmark(
model=model,
Expand Down Expand Up @@ -388,7 +388,7 @@ def benchmark_display(
"gpt-4o", help="The evaluator model you want to use."
),
task: str = typer.Option("hard", help="The task id you want to benchmark."),
):
) -> None:
"""
Usage: sotopia benchmark-display --model-list gpt-4o --model-list together_ai/meta-llama-Llama-3-70b-chat-hf
Aggregate all the results for the benchmark, as described in https://github.com/sotopia-lab/sotopia-space/blob/main/data_dir/models_vs_gpt35.jsonl
Expand All @@ -402,7 +402,7 @@ def benchmark_display(
if len(episodes) == 0:
print(f"No episodes found for {model}")
continue
avg_rewards = get_avg_reward(episodes, model)
avg_rewards = get_avg_reward(episodes, model) # type: ignore
model_rewards_dict[model] = avg_rewards
print(f"Model: {model}, episodes: {len(episodes)}, Avg Rewards: {avg_rewards}")

Expand Down

0 comments on commit af5f725

Please sign in to comment.