diff --git a/lm_eval/models/ibm_watsonx_ai.py b/lm_eval/models/ibm_watsonx_ai.py index 7017184ffc..6f6a09c170 100644 --- a/lm_eval/models/ibm_watsonx_ai.py +++ b/lm_eval/models/ibm_watsonx_ai.py @@ -158,7 +158,7 @@ def __init__( project_id = watsonx_credentials.get("project_id", None) deployment_id = watsonx_credentials.get("deployment_id", None) client.set.default_project(project_id) - self.generate_params = generate_params + self.generate_params = generate_params or {} self.model = ModelInference( model_id=model_id, deployment_id=deployment_id, @@ -167,12 +167,6 @@ def __init__( ) self._model_id = model_id - def dump_parameters(self): - """ - Dumps the model's parameters into a serializable format. - """ - return self._parameters.model_dump() - @staticmethod def _has_stop_token(response_tokens: List[str], context_tokens: List[str]) -> bool: """