-
Notifications
You must be signed in to change notification settings - Fork 159
Open
Labels
bugSomething isn't workingSomething isn't working
Description
LLM spans from autogen-agentchat don't have a model name.
Run the following code snippet with the following requirements:
autogen autogen-agentchat autogen-core autogen-ext ag2 openai openinference-instrumentation-autogen-agentchat
import asyncio
from autogen_agentchat.agents import AssistantAgent
from autogen_agentchat.ui import Console
from autogen_ext.models.openai import OpenAIChatCompletionClient
import asyncio
from phoenix.otel import register
import os
import openai
tracer_provider = register(project_name="autogen-test", auto_instrument=True)
model_client = OpenAIChatCompletionClient(model="gpt-4o-mini", api_key=os.getenv("OPENAI_API_KEY"))
async def get_weather(city: str) -> str:
"""Get the weather for a given city."""
return f"The weather in {city} is 73 degrees and Sunny."
agent = AssistantAgent(
name="weather_agent",
model_client=model_client,
tools=[get_weather],
system_message="You are a helpful assistant.",
reflect_on_tool_use=True,
model_client_stream=True, # Enable streaming tokens from the model client.
)
# Run the agent and stream the messages to the console.
async def main() -> None:
await Console(agent.run_stream(task="What is the weather in New York?"))
# Close the connection to the model client.
await model_client.close()
asyncio.run(main())dosubot
Metadata
Metadata
Assignees
Labels
bugSomething isn't workingSomething isn't working
Type
Projects
Status
In Review