cookbook/11_models/litellm/async_basic_stream.py
import asyncio
from agno.agent import Agent
from agno.models.litellm import LiteLLM
openai_agent = Agent(
model=LiteLLM(
id="gpt-5-mini",
name="LiteLLM",
),
markdown=True,
)
# Print the response in the terminal
asyncio.run(
openai_agent.aprint_response("Share a 2 sentence horror story", stream=True)
)