From 101d32c543c81318beab7bbd754c222f3af72c8a Mon Sep 17 00:00:00 2001 From: Shawn Hurley Date: Tue, 25 Feb 2025 16:11:45 -0500 Subject: [PATCH] removing monkey patch as no longer needed for openshift-ai Signed-off-by: Shawn Hurley --- kai/llm_interfacing/model_provider.py | 30 --------------------------- 1 file changed, 30 deletions(-) diff --git a/kai/llm_interfacing/model_provider.py b/kai/llm_interfacing/model_provider.py index c1c5520d..8b3712fc 100644 --- a/kai/llm_interfacing/model_provider.py +++ b/kai/llm_interfacing/model_provider.py @@ -272,36 +272,6 @@ def prepare_model_args( model_args = deep_update(defaults, config_args) model_id = model_args["model"] - # NOTE(JonahSussman): This is a hack to prevent `max_tokens` - # from getting converted to `max_completion_tokens` for every - # model, except for the o1 and o3 family of models. - - @property # type: ignore[misc] - def _default_params(self: ChatOpenAI) -> dict[str, Any]: - return super(ChatOpenAI, self)._default_params - - def _get_request_payload( - self: ChatOpenAI, - input_: LanguageModelInput, - *, - stop: list[str] | None = None, - **kwargs: Any, - ) -> dict: # type: ignore[type-arg] - return super(ChatOpenAI, self)._get_request_payload( - input_, stop=stop, **kwargs - ) - - if not ( - model_args["model"].startswith("o1") or model_args["model"].startswith("o3") - ): - ChatOpenAI._default_params = _default_params # type: ignore[method-assign] - ChatOpenAI._get_request_payload = _get_request_payload # type: ignore[method-assign] - else: - if "streaming" in model_args: - del model_args["streaming"] - if "temperature" in model_args: - del model_args["temperature"] - return model_args, model_id