diff --git a/skyvern/forge/sdk/api/llm/api_handler_factory.py b/skyvern/forge/sdk/api/llm/api_handler_factory.py index f21279f6e..0585988de 100644 --- a/skyvern/forge/sdk/api/llm/api_handler_factory.py +++ b/skyvern/forge/sdk/api/llm/api_handler_factory.py @@ -186,6 +186,7 @@ async def llm_api_handler( # TODO (kerem): add a timeout to this call # TODO (kerem): add a retry mechanism to this call (acompletion_with_retries) # TODO (kerem): use litellm fallbacks? https://litellm.vercel.app/docs/tutorials/fallbacks#how-does-completion_with_fallbacks-work + LOG.info("Calling LLM API", llm_key=llm_key, model=llm_config.model_name) response = await litellm.acompletion( model=llm_config.model_name, messages=messages, diff --git a/skyvern/forge/sdk/api/llm/config_registry.py b/skyvern/forge/sdk/api/llm/config_registry.py index 3096d3aef..0f6651a68 100644 --- a/skyvern/forge/sdk/api/llm/config_registry.py +++ b/skyvern/forge/sdk/api/llm/config_registry.py @@ -149,7 +149,7 @@ def get_config(cls, llm_key: str) -> LLMRouterConfig | LLMConfig: if SettingsManager.get_settings().ENABLE_AZURE: LLMConfigRegistry.register_config( - "AZURE_OPENAI_GPT4V", + "AZURE_OPENAI", LLMConfig( f"azure/{SettingsManager.get_settings().AZURE_DEPLOYMENT}", [