Comparisons / AutoGen vs Semantic Kernel

AutoGen vs Semantic Kernel: Which Agent Framework to Use?

AutoGen autogen by microsoft models agents as conversableagents that chat with each other. Semantic Kernel semantic kernel is microsoft's enterprise sdk for building ai agents. Here is how they compare — and what the same patterns look like in plain Python.

By the numbers

AutoGen

GitHub Stars

56.7k

Forks

8.5k

Language

Python

License

CC-BY-4.0

Created

2023-08-18

Created by

Microsoft Research

github.com/microsoft/autogen

Semantic Kernel

GitHub Stars

27.6k

Forks

4.5k

Language

C#

License

MIT

Created

2023-02-27

Created by

Microsoft

github.com/microsoft/semantic-kernel

GitHub stats as of April 2026. Stars indicate community interest, not necessarily quality or fit for your use case.

ConceptAutoGenSemantic KernelPlain Python
AgentConversableAgent with system_message, llm_configChatCompletionAgent with Kernel, instructions, and service configA function with a system prompt that POSTs to the LLM API
Toolsregister_for_llm() and register_for_execution()A dict of callables + JSON schema descriptions
ConversationTwo-agent chat with initiate_chat(), message historyA messages array that grows with each turn
Multi-AgentGroupChat with GroupChatManager, speaker selectionMultiple agent functions called in sequence on shared messages
Nested Chatsregister_nested_chats() for sub-task handlingA task queue (BFS) — agent schedules follow-ups via a tool
Terminationis_termination_msg callback, max_consecutive_auto_replyThe while loop exits when no tool_calls or max_turns reached
Tools / PluginsKernelPlugin with @kernel_function decorators, typed parametersA dict of callables: tools = {"search": lambda q: ...}
PlanningStepwisePlanner, HandlebarsPlanner for multi-step decompositionA system prompt that says 'break this into steps' — the LLM plans natively
MemorySemanticTextMemory with embeddings and vector storesA dict injected into the system prompt, or a list searched with embeddings
OrchestrationKernel.invoke() with plugin resolution and filter pipelineA while loop: call LLM, check for tool_calls, dispatch, repeat
Multi-LanguageC#, Python, Java SDKs with shared abstractionsThe HTTP API is the same in every language — just POST JSON

What both do in plain Python

Every concept in the table above — agent, tools, loop, memory, state — maps to a handful of Python primitives: a function, a dict, a list, and a while loop. Both AutoGen and Semantic Kernel wrap these primitives in their own class hierarchies and APIs. The underlying pattern is the same ~60 lines of code. The difference is how much ceremony each framework adds on top.

When to use AutoGen

AutoGen excels at complex multi-agent workflows where agents need to debate or collaborate. For single-agent use cases or simple tool-calling agents, the plain Python version is significantly simpler.

What AutoGen does

AutoGen's core abstraction is the ConversableAgent — an agent that can send and receive messages. Two agents chat by alternating turns on a shared message history. GroupChat extends this to N agents, with a GroupChatManager that selects the next speaker (round-robin, random, or LLM-based selection). Nested chats allow an agent to spin up a sub-conversation to handle a complex subtask before returning to the main thread. AutoGen also provides code execution sandboxes, letting agents write and run code as part of their conversation. The framework thinks in terms of conversations, not chains or graphs. This makes it natural for workflows where agents need to debate, critique, or iteratively refine outputs together.

The plain Python equivalent

A ConversableAgent is a function that takes a messages array, calls the LLM with a system prompt, and returns the assistant message. Two-agent chat is a while loop where you alternate between calling agent_a(messages) and agent_b(messages), appending each response. GroupChat is the same loop but with a speaker selection step — either rotate through a list or ask the LLM "who should speak next?" and call that agent function. Nested chats are a function call within the loop: pause the main conversation, run a sub-loop with different agents, and inject the result back. Tool registration is adding functions to a tools dict with their JSON schemas. The conversation-as-primitive model is just messages arrays passed between functions.

Full AutoGen comparison →

When to use Semantic Kernel

Semantic Kernel earns its complexity in enterprise environments with Azure OpenAI, .NET backends, and existing Microsoft infrastructure. But the core agent pattern — LLM call, tool dispatch, loop — is identical to what you can build in 60 lines of Python.

What Semantic Kernel does

Semantic Kernel is Microsoft's SDK for building AI-powered applications. The central object is the Kernel — it holds your AI service connections, plugins, and configuration. Plugins are collections of KernelFunctions (decorated Python/C# methods) that the LLM can call as tools. Planners like StepwisePlanner break complex goals into multi-step plans, choosing which plugins to invoke at each step. The SDK provides deep integration with Azure OpenAI, including managed identity auth, content filtering, and deployment management. It also ships memory connectors for vector stores (Azure AI Search, Qdrant, Pinecone) and supports filters — middleware that runs before and after each function invocation. For teams already on Azure with .NET backends, it fits naturally into the existing stack.

The plain Python equivalent

The Kernel is a config object that holds your API key and a dict of tools. A KernelFunction is a regular function in that dict. The Planner is a system prompt instruction — tell the LLM to break the task into steps and it will, no planner class needed. Memory is a list of strings you embed and search, or just a dict you inject into the prompt. Orchestration is the same while loop every agent uses: call the LLM, check if the response has tool_calls, look up the function in your tools dict, call it, append the result, repeat. The filter pipeline is a try/except around your function calls. The entire agent — including plugin dispatch, planning, and memory — is about 60 lines. No Kernel object, no plugin registry, no planner hierarchy.

Full Semantic Kernel comparison →

Or build your own in 60 lines

Both AutoGen and Semantic Kernel implement the same 8 patterns. An agent is a function. Tools are a dict. The loop is a while loop. The whole thing composes in ~60 lines of Python.

No framework. No dependencies. No opinions. Just the code.

Build it from scratch →