autogen/python/packages/autogen-agentchat
Eric Zhu 225eb9d0b2
feat: introduce ModelClientStreamingChunkEvent for streaming model output and update handling in agents and console (#5208)
Resolves #3983

* introduce `model_client_stream` parameter in `AssistantAgent` to
enable token-level streaming output.
* introduce `ModelClientStreamingChunkEvent` as a type of `AgentEvent`
to pass the streaming chunks to the application via `run_stream` and
`on_messages_stream`. Although this will not affect the inner messages
list in the final `Response` or `TaskResult`.
* handle this new message type in `Console`.
2025-01-29 02:49:02 +00:00
..
src/autogen_agentchat feat: introduce ModelClientStreamingChunkEvent for streaming model output and update handling in agents and console (#5208) 2025-01-29 02:49:02 +00:00
tests feat: introduce ModelClientStreamingChunkEvent for streaming model output and update handling in agents and console (#5208) 2025-01-29 02:49:02 +00:00
LICENSE-CODE Include license file in package (#3703) 2024-10-09 15:01:09 -04:00
README.md Make package readmes slightly less empty (#4961) 2025-01-09 10:44:13 -08:00
pyproject.toml update versions to 0.4.4 and m1 cli to 0.2.3 (#5229) 2025-01-28 17:59:14 +00:00

README.md

AutoGen AgentChat

AgentChat is a high-level API for building multi-agent applications. It is built on top of the autogen-core package. For beginner users, AgentChat is the recommended starting point. For advanced users, autogen-core's event-driven programming model provides more flexibility and control over the underlying components.

AgentChat provides intuitive defaults, such as Agents with preset behaviors and Teams with predefined multi-agent design patterns.