diff --git a/llm_perf/utils.py b/llm_perf/utils.py index 6665536c..a982759f 100644 --- a/llm_perf/utils.py +++ b/llm_perf/utils.py @@ -44,8 +44,8 @@ # model for model in PRETRAINED_OPEN_LLM_LIST if model.split("/")[0] in CANONICAL_ORGANIZATIONS # ] CANONICAL_PRETRAINED_OPEN_LLM_LIST = [ - "01-ai/Yi-6B", - "01-ai/Yi-34B", + # "01-ai/Yi-6B", + # "01-ai/Yi-34B", "Deci/DeciLM-7B", "Deci/DeciCoder-1b", "EleutherAI/gpt-j-6b", diff --git a/optimum_benchmark/backends/config.py b/optimum_benchmark/backends/config.py index f03d3571..e413a4f8 100644 --- a/optimum_benchmark/backends/config.py +++ b/optimum_benchmark/backends/config.py @@ -73,6 +73,7 @@ def __post_init__(self): self.library, revision=self.model_kwargs.get("revision", None), token=self.model_kwargs.get("token", None), + trust_remote_code=self.model_kwargs.get("trust_remote_code", False), ) if self.device is None: diff --git a/optimum_benchmark/task_utils.py b/optimum_benchmark/task_utils.py index 74773faf..337e835e 100644 --- a/optimum_benchmark/task_utils.py +++ b/optimum_benchmark/task_utils.py @@ -190,6 +190,7 @@ def infer_model_type_from_model_name_or_path( library_name: Optional[str] = None, revision: Optional[str] = None, token: Optional[str] = None, + trust_remote_code: bool = False, ) -> str: if library_name is None: library_name = infer_library_from_model_name_or_path(model_name_or_path, revision=revision, token=token) @@ -216,7 +217,9 @@ def infer_model_type_from_model_name_or_path( break else: - transformers_config = get_transformers_pretrained_config(model_name_or_path, revision=revision, token=token) + transformers_config = get_transformers_pretrained_config( + model_name_or_path, revision=revision, token=token, trust_remote_code=trust_remote_code + ) inferred_model_type = transformers_config.model_type if inferred_model_type is None: