diff --git a/libs/agno/agno/models/defaults.py b/libs/agno/agno/models/defaults.py index d68d23efdc..0f9f26ed23 100644 --- a/libs/agno/agno/models/defaults.py +++ b/libs/agno/agno/models/defaults.py @@ -1 +1 @@ -DEFAULT_OPENAI_MODEL_ID: str = "gpt-4o" +DEFAULT_OPENAI_MODEL_ID: str = "gpt-5.4-mini" diff --git a/libs/agno/agno/models/fireworks/fireworks.py b/libs/agno/agno/models/fireworks/fireworks.py index 8e3a0b8fee..837cae9840 100644 --- a/libs/agno/agno/models/fireworks/fireworks.py +++ b/libs/agno/agno/models/fireworks/fireworks.py @@ -12,14 +12,14 @@ class Fireworks(OpenAILike): A class for interacting with models hosted on Fireworks. Attributes: - id (str): The model name to use. Defaults to "accounts/fireworks/models/llama-v3p1-405b-instruct". + id (str): The model name to use. Defaults to "accounts/fireworks/models/gpt-oss-120b". name (str): The model name to use. Defaults to "Fireworks". provider (str): The provider to use. Defaults to "Fireworks". api_key (Optional[str]): The API key to use. base_url (str): The base URL to use. Defaults to "https://api.fireworks.ai/inference/v1". """ - id: str = "accounts/fireworks/models/llama-v3p1-405b-instruct" + id: str = "accounts/fireworks/models/gpt-oss-120b" name: str = "Fireworks" provider: str = "Fireworks" diff --git a/libs/agno/agno/models/google/gemini.py b/libs/agno/agno/models/google/gemini.py index fcf52080cf..4616fff47f 100644 --- a/libs/agno/agno/models/google/gemini.py +++ b/libs/agno/agno/models/google/gemini.py @@ -78,7 +78,7 @@ class Gemini(Model): Based on https://googleapis.github.io/python-genai/ """ - id: str = "gemini-2.0-flash-001" + id: str = "gemini-flash-latest" name: str = "Gemini" provider: str = "Google" diff --git a/libs/agno/agno/models/moonshot/moonshot.py b/libs/agno/agno/models/moonshot/moonshot.py index f0152516b6..1419417357 100644 --- a/libs/agno/agno/models/moonshot/moonshot.py +++ b/libs/agno/agno/models/moonshot/moonshot.py @@ -12,14 +12,14 @@ class MoonShot(OpenAILike): A class for interacting with MoonShot models. Attributes: - id (str): The model id. Defaults to "kimi-k2-thinking". + id (str): The model id. Defaults to "kimi-k2.5". name (str): The model name. Defaults to "Moonshot". provider (str): The provider name. Defaults to "Moonshot". api_key (Optional[str]): The API key. base_url (str): The base URL. Defaults to "https://api.moonshot.ai/v1". """ - id: str = "kimi-k2-thinking" + id: str = "kimi-k2.5" name: str = "Moonshot" provider: str = "Moonshot" diff --git a/libs/agno/agno/models/openai/chat.py b/libs/agno/agno/models/openai/chat.py index dc4aa198cf..111e129f45 100644 --- a/libs/agno/agno/models/openai/chat.py +++ b/libs/agno/agno/models/openai/chat.py @@ -38,7 +38,7 @@ class OpenAIChat(Model): For more information, see: https://platform.openai.com/docs/api-reference/chat/create """ - id: str = "gpt-4o" + id: str = "gpt-5.4-mini" name: str = "OpenAIChat" provider: str = "OpenAI" supports_native_structured_outputs: bool = True diff --git a/libs/agno/agno/models/openai/responses.py b/libs/agno/agno/models/openai/responses.py index 5e6f197b18..86eae73aa7 100644 --- a/libs/agno/agno/models/openai/responses.py +++ b/libs/agno/agno/models/openai/responses.py @@ -35,7 +35,7 @@ class OpenAIResponses(Model): For more information, see: https://platform.openai.com/docs/api-reference/responses """ - id: str = "gpt-4o" + id: str = "gpt-5.4-mini" name: str = "OpenAIResponses" provider: str = "OpenAI" supports_native_structured_outputs: bool = True diff --git a/libs/agno/agno/models/openrouter/openrouter.py b/libs/agno/agno/models/openrouter/openrouter.py index be163a3b68..c7652e6280 100644 --- a/libs/agno/agno/models/openrouter/openrouter.py +++ b/libs/agno/agno/models/openrouter/openrouter.py @@ -18,7 +18,7 @@ class OpenRouter(OpenAILike): A class for using models hosted on OpenRouter. Attributes: - id (str): The model id. Defaults to "gpt-4o". + id (str): The model id. Defaults to "gpt-5.4-mini". name (str): The model name. Defaults to "OpenRouter". provider (str): The provider name. Defaults to "OpenRouter". api_key (Optional[str]): The API key. @@ -29,7 +29,7 @@ class OpenRouter(OpenAILike): these models in order. Example: ["anthropic/claude-sonnet-4", "deepseek/deepseek-r1"] """ - id: str = "gpt-4o" + id: str = "gpt-5.4-mini" name: str = "OpenRouter" provider: str = "OpenRouter" diff --git a/libs/agno/agno/models/together/together.py b/libs/agno/agno/models/together/together.py index c08dc7a2c3..f64d2a7bb9 100644 --- a/libs/agno/agno/models/together/together.py +++ b/libs/agno/agno/models/together/together.py @@ -12,14 +12,14 @@ class Together(OpenAILike): A class for interacting with Together API. Attributes: - id (str): The id of the Together model to use. Default is "mistralai/Mixtral-8x7B-Instruct-v0.1". + id (str): The id of the Together model to use. Default is "MiniMaxAI/MiniMax-M2.7". name (str): The name of this chat model instance. Default is "Together" provider (str): The provider of the model. Default is "Together". api_key (str): The api key to authorize request to Together. base_url (str): The base url to which the requests are sent. Defaults to "https://api.together.xyz/v1". """ - id: str = "mistralai/Mixtral-8x7B-Instruct-v0.1" + id: str = "MiniMaxAI/MiniMax-M2.7" name: str = "Together" provider: str = "Together" api_key: Optional[str] = None diff --git a/libs/agno/agno/models/xai/xai.py b/libs/agno/agno/models/xai/xai.py index e8c4c554f8..b0190ab364 100644 --- a/libs/agno/agno/models/xai/xai.py +++ b/libs/agno/agno/models/xai/xai.py @@ -23,7 +23,7 @@ class xAI(OpenAILike): Class for interacting with the xAI API. Attributes: - id (str): The ID of the language model. Defaults to "grok-beta". + id (str): The ID of the language model. Defaults to "grok-4-1-fast-non-reasoning-latest". name (str): The name of the API. Defaults to "xAI". provider (str): The provider of the API. Defaults to "xAI". api_key (Optional[str]): The API key for the xAI API. @@ -31,7 +31,7 @@ class xAI(OpenAILike): search_parameters (Optional[Dict[str, Any]]): Search parameters for enabling live search. """ - id: str = "grok-beta" + id: str = "grok-4-1-fast-non-reasoning-latest" name: str = "xAI" provider: str = "xAI" diff --git a/libs/agno/tests/integration/agent/test_parser_model.py b/libs/agno/tests/integration/agent/test_parser_model.py index 573a6740de..1a56ac6108 100644 --- a/libs/agno/tests/integration/agent/test_parser_model.py +++ b/libs/agno/tests/integration/agent/test_parser_model.py @@ -70,7 +70,7 @@ def test_openai_with_claude_parser_model(): def test_gemini_with_openai_parser_model(): park_agent = Agent( - model=Gemini(id="gemini-2.0-flash-001"), # Main model to generate the content + model=Gemini(id="gemini-flash-latest"), # Main model to generate the content description="You are an expert on national parks and provide concise guides.", output_schema=ParkGuide, parser_model=OpenAIChat(id="gpt-4o"), # Model to parse the output diff --git a/libs/agno/tests/integration/knowledge/filters/test_agentic_filtering.py b/libs/agno/tests/integration/knowledge/filters/test_agentic_filtering.py index fdaa1835b6..ec8a67a268 100644 --- a/libs/agno/tests/integration/knowledge/filters/test_agentic_filtering.py +++ b/libs/agno/tests/integration/knowledge/filters/test_agentic_filtering.py @@ -173,7 +173,7 @@ async def test_agentic_filtering_openai_with_output_schema(knowledge_base): @pytest.mark.skipif(not os.environ.get("GOOGLE_API_KEY"), reason="GOOGLE_API_KEY not set") async def test_agentic_filtering_gemini(knowledge_base): - agent = Agent(model=Gemini("gemini-2.0-flash-001"), knowledge=knowledge_base, enable_agentic_knowledge_filters=True) + agent = Agent(model=Gemini("gemini-flash-latest"), knowledge=knowledge_base, enable_agentic_knowledge_filters=True) response = await agent.arun( "Tell me about revenue performance and top selling products in the region north_america and data_type sales", markdown=True, diff --git a/libs/agno/tests/integration/models/fireworks/test_basic.py b/libs/agno/tests/integration/models/fireworks/test_basic.py index 8d9be543c7..0aa1298946 100644 --- a/libs/agno/tests/integration/models/fireworks/test_basic.py +++ b/libs/agno/tests/integration/models/fireworks/test_basic.py @@ -117,7 +117,7 @@ class MovieScript(BaseModel): plot: str = Field(..., description="Brief plot summary") agent = Agent( - model=Fireworks(id="accounts/fireworks/models/llama-v3p1-405b-instruct"), + model=Fireworks(id="accounts/fireworks/models/gpt-oss-120b"), output_schema=MovieScript, telemetry=False, ) @@ -138,7 +138,7 @@ class MovieScript(BaseModel): plot: str = Field(..., description="Brief plot summary") agent = Agent( - model=Fireworks(id="accounts/fireworks/models/llama-v3p1-405b-instruct"), + model=Fireworks(id="accounts/fireworks/models/gpt-oss-120b"), output_schema=MovieScript, use_json_mode=True, telemetry=False, diff --git a/libs/agno/tests/integration/models/fireworks/test_structured_response.py b/libs/agno/tests/integration/models/fireworks/test_structured_response.py index f9cc15ae42..dffd050fd9 100644 --- a/libs/agno/tests/integration/models/fireworks/test_structured_response.py +++ b/libs/agno/tests/integration/models/fireworks/test_structured_response.py @@ -24,7 +24,7 @@ class MovieScript(BaseModel): def test_structured_response(): structured_output_agent = Agent( - model=Fireworks(id="accounts/fireworks/models/llama-v3p1-405b-instruct"), + model=Fireworks(id="accounts/fireworks/models/gpt-oss-120b"), description="You help people write movie scripts.", output_schema=MovieScript, ) @@ -52,7 +52,7 @@ class Recipe(BaseModel): rating: Grade structured_output_agent = Agent( - model=Fireworks(id="accounts/fireworks/models/llama-v3p1-405b-instruct"), + model=Fireworks(id="accounts/fireworks/models/gpt-oss-120b"), description="You help generate recipe names and ratings.", output_schema=Recipe, ) diff --git a/libs/agno/tests/integration/models/fireworks/test_tool_use.py b/libs/agno/tests/integration/models/fireworks/test_tool_use.py index c9ee786f60..0ab3cc1eea 100644 --- a/libs/agno/tests/integration/models/fireworks/test_tool_use.py +++ b/libs/agno/tests/integration/models/fireworks/test_tool_use.py @@ -11,7 +11,7 @@ def test_tool_use(): agent = Agent( - model=Fireworks(id="accounts/fireworks/models/llama-v3p1-405b-instruct"), + model=Fireworks(id="accounts/fireworks/models/gpt-oss-120b"), tools=[YFinanceTools(cache_results=True)], markdown=True, telemetry=False, @@ -28,7 +28,7 @@ def test_tool_use(): def test_tool_use_stream(): agent = Agent( - model=Fireworks(id="accounts/fireworks/models/llama-v3p1-405b-instruct"), + model=Fireworks(id="accounts/fireworks/models/gpt-oss-120b"), tools=[YFinanceTools(cache_results=True)], markdown=True, telemetry=False, @@ -55,7 +55,7 @@ def test_tool_use_stream(): @pytest.mark.asyncio async def test_async_tool_use(): agent = Agent( - model=Fireworks(id="accounts/fireworks/models/llama-v3p1-405b-instruct"), + model=Fireworks(id="accounts/fireworks/models/gpt-oss-120b"), tools=[YFinanceTools(cache_results=True)], markdown=True, telemetry=False, @@ -73,7 +73,7 @@ async def test_async_tool_use(): @pytest.mark.asyncio async def test_async_tool_use_stream(): agent = Agent( - model=Fireworks(id="accounts/fireworks/models/llama-v3p1-405b-instruct"), + model=Fireworks(id="accounts/fireworks/models/gpt-oss-120b"), tools=[YFinanceTools(cache_results=True)], markdown=True, telemetry=False, @@ -92,7 +92,7 @@ async def test_async_tool_use_stream(): def test_parallel_tool_calls(): agent = Agent( - model=Fireworks(id="accounts/fireworks/models/llama-v3p1-405b-instruct"), + model=Fireworks(id="accounts/fireworks/models/gpt-oss-120b"), tools=[YFinanceTools(cache_results=True)], markdown=True, telemetry=False, @@ -113,7 +113,7 @@ def test_parallel_tool_calls(): def test_multiple_tool_calls(): agent = Agent( - model=Fireworks(id="accounts/fireworks/models/llama-v3p1-405b-instruct"), + model=Fireworks(id="accounts/fireworks/models/gpt-oss-120b"), tools=[YFinanceTools(cache_results=True), WebSearchTools(cache_results=True)], markdown=True, telemetry=False, @@ -140,7 +140,7 @@ def get_the_weather_in_tokyo(): return "It is currently 70 degrees and cloudy in Tokyo" agent = Agent( - model=Fireworks(id="accounts/fireworks/models/llama-v3p1-405b-instruct"), + model=Fireworks(id="accounts/fireworks/models/gpt-oss-120b"), tools=[get_the_weather_in_tokyo], markdown=True, telemetry=False, @@ -169,7 +169,7 @@ def get_the_weather(city: Optional[str] = None): return f"It is currently 70 degrees and cloudy in {city}" agent = Agent( - model=Fireworks(id="accounts/fireworks/models/llama-v3p1-405b-instruct"), + model=Fireworks(id="accounts/fireworks/models/gpt-oss-120b"), tools=[get_the_weather], markdown=True, telemetry=False, @@ -186,7 +186,7 @@ def get_the_weather(city: Optional[str] = None): def test_tool_call_list_parameters(): agent = Agent( - model=Fireworks(id="accounts/fireworks/models/llama-v3p1-405b-instruct"), + model=Fireworks(id="accounts/fireworks/models/gpt-oss-120b"), tools=[ExaTools()], instructions="Use a single tool call if possible", markdown=True, diff --git a/libs/agno/tests/integration/models/google/test_multimodal.py b/libs/agno/tests/integration/models/google/test_multimodal.py index fbaef75ca3..4a10895948 100644 --- a/libs/agno/tests/integration/models/google/test_multimodal.py +++ b/libs/agno/tests/integration/models/google/test_multimodal.py @@ -12,7 +12,7 @@ def test_image_input(image_path): agent = Agent( - model=Gemini(id="gemini-2.0-flash-001"), + model=Gemini(id="gemini-flash-latest"), exponential_backoff=True, delay_between_retries=5, markdown=True, @@ -37,7 +37,7 @@ def test_audio_input_bytes(): # Provide the agent with the audio file and get result as text agent = Agent( - model=Gemini(id="gemini-2.0-flash-001"), + model=Gemini(id="gemini-flash-latest"), exponential_backoff=True, delay_between_retries=5, markdown=True, @@ -50,7 +50,7 @@ def test_audio_input_bytes(): def test_audio_input_url(): agent = Agent( - model=Gemini(id="gemini-2.0-flash-001"), + model=Gemini(id="gemini-flash-latest"), exponential_backoff=True, delay_between_retries=5, markdown=True, @@ -67,7 +67,7 @@ def test_audio_input_url(): def test_video_input_bytes(): agent = Agent( - model=Gemini(id="gemini-2.0-flash-001"), + model=Gemini(id="gemini-flash-latest"), exponential_backoff=True, delay_between_retries=5, markdown=True, @@ -255,7 +255,7 @@ def test_combined_text_and_image_generation(): def test_file_input_bytes(): agent = Agent( - model=Gemini(id="gemini-2.0-flash-001"), + model=Gemini(id="gemini-flash-latest"), exponential_backoff=True, delay_between_retries=5, markdown=True, @@ -276,7 +276,7 @@ def test_file_input_bytes(): def test_file_input_with_text_prompt(): agent = Agent( - model=Gemini(id="gemini-2.0-flash-001"), + model=Gemini(id="gemini-flash-latest"), exponential_backoff=True, delay_between_retries=5, markdown=True, diff --git a/libs/agno/tests/integration/models/google/test_retryable_exceptions.py b/libs/agno/tests/integration/models/google/test_retryable_exceptions.py index 33f957b838..1fee505003 100644 --- a/libs/agno/tests/integration/models/google/test_retryable_exceptions.py +++ b/libs/agno/tests/integration/models/google/test_retryable_exceptions.py @@ -15,7 +15,7 @@ @pytest.fixture def model(): """Fixture to create a Gemini model.""" - return Gemini(id="gemini-2.0-flash-001") + return Gemini(id="gemini-flash-latest") def create_mock_response(finish_reason: str = "STOP", content: str = "Test response"): diff --git a/libs/agno/tests/integration/models/google/test_tool_use.py b/libs/agno/tests/integration/models/google/test_tool_use.py index c4072af48b..5ea2a2dad4 100644 --- a/libs/agno/tests/integration/models/google/test_tool_use.py +++ b/libs/agno/tests/integration/models/google/test_tool_use.py @@ -119,7 +119,7 @@ def get_weather(city: str) -> str: return f"The weather in {city} is sunny." agent = Agent( - model=Gemini(id="gemini-2.0-flash-001"), + model=Gemini(id="gemini-flash-latest"), tools=[get_weather], tool_choice="none", markdown=True, @@ -142,7 +142,7 @@ def get_weather(city: str) -> str: return f"The weather in {city} is sunny." agent = Agent( - model=Gemini(id="gemini-2.0-flash-001"), + model=Gemini(id="gemini-flash-latest"), tools=[get_weather], tool_choice="auto", markdown=True, @@ -185,7 +185,7 @@ class StockPrice(BaseModel): currency: str = Field(..., description="The currency of the stock") agent = Agent( - model=Gemini(id="gemini-2.0-flash-001"), + model=Gemini(id="gemini-flash-latest"), tools=[YFinanceTools(cache_results=True)], exponential_backoff=True, delay_between_retries=5, @@ -250,7 +250,7 @@ def test_multiple_tool_calls(): def test_grounding(): agent = Agent( - model=Gemini(id="gemini-2.0-flash-001", grounding=True), + model=Gemini(id="gemini-flash-latest", grounding=True), exponential_backoff=True, delay_between_retries=5, telemetry=False, @@ -267,7 +267,7 @@ def test_grounding(): def test_grounding_stream(): agent = Agent( - model=Gemini(id="gemini-2.0-flash-001", grounding=True), + model=Gemini(id="gemini-flash-latest", grounding=True), exponential_backoff=True, delay_between_retries=5, telemetry=False, @@ -289,7 +289,7 @@ def test_grounding_stream(): def test_search_stream(): agent = Agent( - model=Gemini(id="gemini-2.0-flash-001", search=True), + model=Gemini(id="gemini-flash-latest", search=True), exponential_backoff=True, delay_between_retries=5, telemetry=False, diff --git a/libs/agno/tests/integration/models/openrouter/test_basic.py b/libs/agno/tests/integration/models/openrouter/test_basic.py index 5f0a64c5e1..ab1b4f0e49 100644 --- a/libs/agno/tests/integration/models/openrouter/test_basic.py +++ b/libs/agno/tests/integration/models/openrouter/test_basic.py @@ -22,7 +22,7 @@ def _assert_metrics(response: RunOutput): def test_basic(): - agent = Agent(model=OpenRouter(id="gpt-4o"), markdown=True, telemetry=False) + agent = Agent(model=OpenRouter(id="gpt-5.4-mini"), markdown=True, telemetry=False) response: RunOutput = agent.run("Share a 2 sentence horror story") @@ -35,7 +35,7 @@ def test_basic(): def test_basic_stream(): - agent = Agent(model=OpenRouter(id="gpt-4o"), markdown=True, telemetry=False) + agent = Agent(model=OpenRouter(id="gpt-5.4-mini"), markdown=True, telemetry=False) response_stream = agent.run("Share a 2 sentence horror story", stream=True) @@ -50,7 +50,7 @@ def test_basic_stream(): @pytest.mark.asyncio async def test_async_basic(): - agent = Agent(model=OpenRouter(id="gpt-4o"), markdown=True, telemetry=False) + agent = Agent(model=OpenRouter(id="gpt-5.4-mini"), markdown=True, telemetry=False) response = await agent.arun("Share a 2 sentence horror story") @@ -63,7 +63,7 @@ async def test_async_basic(): @pytest.mark.asyncio async def test_async_basic_stream(): - agent = Agent(model=OpenRouter(id="gpt-4o"), markdown=True, telemetry=False) + agent = Agent(model=OpenRouter(id="gpt-5.4-mini"), markdown=True, telemetry=False) async for response in agent.arun("Share a 2 sentence horror story", stream=True): assert response.content is not None @@ -71,7 +71,7 @@ async def test_async_basic_stream(): def test_with_memory(): agent = Agent( - model=OpenRouter(id="gpt-4o"), + model=OpenRouter(id="gpt-5.4-mini"), db=InMemoryDb(), add_history_to_context=True, markdown=True, @@ -103,7 +103,7 @@ class MovieScript(BaseModel): plot: str = Field(..., description="Brief plot summary") agent = Agent( - model=OpenRouter(id="gpt-4o"), + model=OpenRouter(id="gpt-5.4-mini"), markdown=True, telemetry=False, output_schema=MovieScript, @@ -125,7 +125,7 @@ class MovieScript(BaseModel): plot: str = Field(..., description="Brief plot summary") agent = Agent( - model=OpenRouter(id="gpt-4o"), + model=OpenRouter(id="gpt-5.4-mini"), use_json_mode=True, telemetry=False, output_schema=MovieScript, @@ -142,7 +142,7 @@ class MovieScript(BaseModel): def test_history(): agent = Agent( - model=OpenRouter(id="gpt-4o"), + model=OpenRouter(id="gpt-5.4-mini"), db=SqliteDb(db_file="tmp/openrouter/test_basic.db"), add_history_to_context=True, store_history_messages=True, diff --git a/libs/agno/tests/integration/models/openrouter/test_structured_response.py b/libs/agno/tests/integration/models/openrouter/test_structured_response.py index 21e080c67e..333f60670b 100644 --- a/libs/agno/tests/integration/models/openrouter/test_structured_response.py +++ b/libs/agno/tests/integration/models/openrouter/test_structured_response.py @@ -24,7 +24,7 @@ class MovieScript(BaseModel): def test_structured_response(): structured_output_agent = Agent( - model=OpenRouter(id="gpt-4o"), + model=OpenRouter(id="gpt-5.4-mini"), description="You help people write movie scripts.", output_schema=MovieScript, ) @@ -52,7 +52,7 @@ class Recipe(BaseModel): rating: Grade structured_output_agent = Agent( - model=OpenRouter(id="gpt-4o"), + model=OpenRouter(id="gpt-5.4-mini"), description="You help generate recipe names and ratings.", output_schema=Recipe, ) @@ -65,7 +65,7 @@ class Recipe(BaseModel): def test_structured_response_strict_output_false(): """Test structured response with strict_output=False (guided mode)""" guided_output_agent = Agent( - model=OpenRouter(id="gpt-4o", strict_output=False), + model=OpenRouter(id="gpt-5.4-mini", strict_output=False), description="You write movie scripts.", output_schema=MovieScript, ) diff --git a/libs/agno/tests/integration/models/openrouter/test_tool_use.py b/libs/agno/tests/integration/models/openrouter/test_tool_use.py index 56d5a6d086..3f9c19b9af 100644 --- a/libs/agno/tests/integration/models/openrouter/test_tool_use.py +++ b/libs/agno/tests/integration/models/openrouter/test_tool_use.py @@ -11,7 +11,7 @@ def test_tool_use(): agent = Agent( - model=OpenRouter(id="gpt-4o"), + model=OpenRouter(id="gpt-5.4-mini"), tools=[YFinanceTools(cache_results=True)], markdown=True, telemetry=False, @@ -28,7 +28,7 @@ def test_tool_use(): def test_tool_use_stream(): agent = Agent( - model=OpenRouter(id="gpt-4o"), + model=OpenRouter(id="gpt-5.4-mini"), tools=[YFinanceTools(cache_results=True)], markdown=True, telemetry=False, @@ -53,7 +53,7 @@ def test_tool_use_stream(): @pytest.mark.asyncio async def test_async_tool_use(): agent = Agent( - model=OpenRouter(id="gpt-4o"), + model=OpenRouter(id="gpt-5.4-mini"), tools=[YFinanceTools(cache_results=True)], markdown=True, telemetry=False, @@ -71,7 +71,7 @@ async def test_async_tool_use(): @pytest.mark.asyncio async def test_async_tool_use_stream(): agent = Agent( - model=OpenRouter(id="gpt-4o"), + model=OpenRouter(id="gpt-5.4-mini"), tools=[YFinanceTools(cache_results=True)], markdown=True, telemetry=False, @@ -95,7 +95,7 @@ async def test_async_tool_use_stream(): def test_multiple_tool_calls(): agent = Agent( - model=OpenRouter(id="gpt-4o"), + model=OpenRouter(id="gpt-5.4-mini"), tools=[YFinanceTools(cache_results=True), WebSearchTools(cache_results=True)], markdown=True, telemetry=False, @@ -122,7 +122,7 @@ def get_the_weather_in_tokyo(): return "It is currently 70 degrees and cloudy in Tokyo" agent = Agent( - model=OpenRouter(id="gpt-4o"), + model=OpenRouter(id="gpt-5.4-mini"), tools=[get_the_weather_in_tokyo], markdown=True, telemetry=False, @@ -151,7 +151,7 @@ def get_the_weather(city: Optional[str] = None): return f"It is currently 70 degrees and cloudy in {city}" agent = Agent( - model=OpenRouter(id="gpt-4o"), + model=OpenRouter(id="gpt-5.4-mini"), tools=[get_the_weather], markdown=True, telemetry=False, @@ -168,7 +168,7 @@ def get_the_weather(city: Optional[str] = None): def test_tool_call_list_parameters(): agent = Agent( - model=OpenRouter(id="gpt-4o"), + model=OpenRouter(id="gpt-5.4-mini"), tools=[ExaTools()], instructions="Use a single tool call if possible", markdown=True, diff --git a/libs/agno/tests/unit/models/moonshot/test_moonshot.py b/libs/agno/tests/unit/models/moonshot/test_moonshot.py index 057f14e253..ebed0c6f1b 100644 --- a/libs/agno/tests/unit/models/moonshot/test_moonshot.py +++ b/libs/agno/tests/unit/models/moonshot/test_moonshot.py @@ -8,15 +8,15 @@ def test_moonshot_initialization_with_api_key(): - model = MoonShot(id="kimi-k2-thinking", api_key="test-api-key") - assert model.id == "kimi-k2-thinking" + model = MoonShot(id="kimi-k2.5", api_key="test-api-key") + assert model.id == "kimi-k2.5" assert model.api_key == "test-api-key" assert model.base_url == "https://api.moonshot.ai/v1" def test_moonshot_initialization_without_api_key(): with patch.dict(os.environ, {}, clear=True): - model = MoonShot(id="kimi-k2-thinking") + model = MoonShot(id="kimi-k2.5") client_params = None with pytest.raises(ModelAuthenticationError): client_params = model._get_client_params() @@ -25,12 +25,12 @@ def test_moonshot_initialization_without_api_key(): def test_moonshot_initialization_with_env_api_key(): with patch.dict(os.environ, {"MOONSHOT_API_KEY": "env-api-key"}): - model = MoonShot(id="kimi-k2-thinking") + model = MoonShot(id="kimi-k2.5") assert model.api_key == "env-api-key" def test_moonshot_client_params(): - model = MoonShot(id="kimi-k2-thinking", api_key="test-api-key") + model = MoonShot(id="kimi-k2.5", api_key="test-api-key") client_params = model._get_client_params() assert client_params["api_key"] == "test-api-key" assert client_params["base_url"] == "https://api.moonshot.ai/v1"