Lesson 519 of 2116
Kimi as an Agent: Browsing, Tools, and Multi-Step Tasks
Kimi isn't just a chat model — its newer variants act on tools, browse the web, and chain steps. Here is what the platform actually offers and where the rough edges are.
Lesson map
What this lesson covers
Learning path
The main moves in order
- 1Beyond a chat box
- 2function calling
- 3tool use
- 4browsing
Concept cluster
Terms to connect while reading
Section 1
Beyond a chat box
Modern Kimi variants ship with first-class tool calling, web browsing inside the chat product, and structured output modes. The platform is moving in the same direction as OpenAI and Anthropic: less 'answer the question' and more 'do the task end to end'. The patterns you already know from those ecosystems carry over almost perfectly.
Function calling on Moonshot's API
Kimi's function-calling shape mirrors OpenAI's. The same tool schema usually works on both.
tools = [
{
"type": "function",
"function": {
"name": "lookup_invoice",
"description": "Find an invoice by ID in the internal billing system.",
"parameters": {
"type": "object",
"properties": {
"invoice_id": {"type": "string"}
},
"required": ["invoice_id"],
},
},
}
]
resp = client.chat.completions.create(
model="<long-context-model-id>",
messages=messages,
tools=tools,
tool_choice="auto",
)Compare the options
| Capability | Kimi consumer chat | Kimi API | Claude / GPT-5 |
|---|---|---|---|
| Web browsing | Built in | Limited / via tools | Built in |
| Function calling | Implicit | Explicit, OpenAI-compatible | Mature |
| File upload + analysis | Excellent | Possible via tool | Excellent |
| Long-running task / agent loop | Improving | DIY orchestration | Mature |
Where Kimi's agents shine
- Research assignments that require reading a stack of uploaded files plus a few web sources
- Bilingual research tasks where Chinese-language sources matter
- Tasks where the long context lets the agent skip a retrieval step entirely
Where it still feels rough
- Long agent loops sometimes drift in tone after many turns — anchor with explicit goal restatement
- Tool-error handling is less battle-tested than mature Western SDKs
- Observability tooling around the API is thinner; build your own logging early
Apply this
- 1Define a small two-tool agent task (lookup + email-draft, for example)
- 2Implement it on Moonshot's function-calling API and on one Western API
- 3Compare reliability, latency, and how each handles a deliberately broken tool response
Key terms in this lesson
The big idea: Kimi is now an agent platform, not just a chat model. The patterns you already know about agent reliability port directly — and the rough edges are mostly around tooling, not the model itself.
End-of-lesson quiz
Check what stuck
15 questions · Score saves to your progress.
Tutor
Curious about “Kimi as an Agent: Browsing, Tools, and Multi-Step Tasks”?
Ask anything about this lesson. I’ll answer using just what you’re reading — short, friendly, grounded.
Progress saved locally in this browser. Sign in to sync across devices.
Related lessons
Keep going
Creators · 10 min
Hermes For Function Calling: Tool-Use Without OpenAI
Hermes ships with a documented function-calling format. That makes it one of the few open-weight models you can wire into agent frameworks without months of prompting hacks.
Creators · 9 min
MiniMax For Agentic Tasks: Strengths And Gaps
MiniMax models can drive agents, but their tool-use shape, refusal patterns, and ecosystem differ from Western frontier. Plan for it.
Creators · 40 min
Tool Use Quality Across Claude, GPT, Gemini, Llama
Compare native tool-calling reliability and patterns across model families.
