MCP Servers in Claude Code
What is MCP
MCP is defined as "USB-C for AI applications." MCP provides a consistent protocol for linking AI models with external tools.
With MCP, AI applications like Claude or ChatGPT aren’t limited to just generating text. They can connect directly to data sources (such as local files or databases), tools (like search engines or calculators), and structured workflows (specialized prompts or automations).
When MCP is not yet available.
When there is no protocol like MCP, each AI application has to integrate with every external tool separately. This makes the process very complex, time-consuming, and costly

When we have multiple AI applications and many tools, the number of required integrations becomes extremely large
When MCP is available
MCP solves the M × N integration problem by transforming it into an M + N model through a standardized connection protocol. Each AI application integrates once on the MCP client side, and each tool or data source integrates once on the MCP server side.

Key Terminology of MCP
Components
Similar to the client–server model in the HTTP protocol, MCP also follows a client–server architecture.
- Host: The environment where the user directly interacts with the AI application (e.g., Claude Desktop, Cursor).
- Client: A component within the Host responsible for establishing and managing the connection to the MCP Server.
- Server: An external application or service that provides capabilities (such as tools, resources, and prompts) through the MCP protocol.

Capabilities
Although MCP can connect to many different tools, there are common tools that are shared across multiple AI applications. Below are the main categories of tools commonly used across AI systems:
- Tools: Executable functions that an AI model can call to perform actions or computations (e.g., a calculate_summary tool).
- Resources: Read-only data sources that provide contextual information without requiring significant computation (e.g., company documentation pages).
- Prompts: Templates or predefined workflows that guide interactions between users, AI models, and external tools.
MCP Architecture Components
After understanding the key concepts and terminology of MCP, we can now look at its architecture.

The Model Context Protocol (MCP) is built on a client–server architecture that enables AI models to interact with external tools and services.
Host
The Host is the environment where end users directly interact with the AI application (e.g., Claude Desktop, Cursor).
The Host is responsible for:
- Managing user interactions and permissions
- Initiating connections to MCP Servers through MCP Clients
- Processing user requests and routing them to appropriate external tools
- Returning results back to the user
Client
The Client is a component inside the Host that manages the connection to a specific MCP Server.
Key characteristics:
- Each Client maintains a 1:1 connection with a single Server
- Handles MCP protocol-level communication
- Acts as an intermediary between the Host and the Server
Server
The Server is an external program or service that provides capabilities to the AI model via the MCP protocol.
The Server is responsible for:
- Providing access to external tools, data sources, or services
- Running either locally (on the same machine as the Host) or remotely (over the network)
- Exposing standardized interfaces so Clients can interact with its capabilities
Building the simple MCP Server
Install fastmcp
pip install fastmcp
Basic MCP Server: Weather Tool

That’s it! FastMCP handles everything, including:
- JSON-RPC protocol
- Tool registration
- Type validation
- Error handling
Connecting to Claude Code
claude mcp add weather --command python --args /full/path/to/get_weather.py
Restart Claude Code → the MCP servers will automatically connect.
Now you can ask Claude things like:
- “What’s the temperature in Japan?”
- “Read the file at ~/documents/report.txt”
Claude will automatically invoke the tools from your MCP servers seamlessly.
Setting up MCP with Claude Code
Basedon the Claude docs (https://code.claude.com/docs/en/mcp), setting up
MCPin Claude Code is pretty straightforward — just run
claude mcp add
andit handles the configuration for you automatically
Tolist and verify all configured MCP servers in Claude Code, try to run:
claude mcp list
Claude Code & MCP: Best Practices
1. Serena MCP
Link: https://github.com/oraios/serena
I’ve been experimenting with an AI-driven workflow and plugged Serena MCP straight in (just using Sonet 4.5).
Honestly, it feels kind of “wow.” Instead of dumping a bunch of files on the AI and hoping it figures things out, it actually reads the codebase like a senior dev on the team.
Why it works so well?
- RAG: It indexes the whole codebase, uses semantic search to pull only the most relevant parts, and feeds clean context to the model → less noise, better answers.
- Built on Language Server Protocol, so it understands code structurally, not just as raw text.
- Deep memory: It remembers the indexed codebase, so you don’t have to keep reloading tokens every time.
- Semantic search is — ask “Where is authentication handled?” and it finds related functions/classes even if the naming is weird. On big projects, that’s a lifesaver.
Overall: fewer tokens, cleaner context, much deeper code understanding.
If you’re building coding agents, you should try it. It really starts to feel like AI is your teammate.
2. Sequential Thinking MCP
Link: https://github.com/modelcontextprotocol/servers/tree/main/src/sequentialthinking
Key Highlights
- Structured reasoning: Claude Code solves complex problems using logical, step-by-step thinking.
- Handles complex tasks: Optimized for multi-stage tasks such as system design or architectural refactoring.
- Scalability: Supports step-by-step planning and analysis for large-scale codebases.
Use Cases
- Refactoring microservice architectures
- Phase-based task planning for large projects
- Optimizing system design and debugging workflows
3. Using Specialized Sub-Agents
Link: https://github.com/wshobson/agents
This is a comprehensive, production-ready system designed to integrate with Claude Code and significantly extend its capabilities.
It combines:
- 112 specialized AI agents
- 16 multi-agent workflow orchestrators
- 146 agent skills
- 79 development tools
- Organized into 72 focused, single-purpose plugins for Claude Code
Each agent has a clearly defined role — such as backend architecture design, frontend development, cloud infrastructure optimization, automated testing, MLOps, and more — all configured following modern best practices.
Installation:
git clone https://github.com/wshobson/agents.git ~/.claude/agents
Built for Speed: ~10ms Latency, Even Under Load
Blazingly fast way to build, track and deploy your models!
- Handles 350+ RPS on just 1 vCPU — no tuning needed
- Production-ready with full enterprise support
TrueFoundry AI Gateway delivers ~3–4 ms latency, handles 350+ RPS on 1 vCPU, scales horizontally with ease, and is production-ready, while LiteLLM suffers from high latency, struggles beyond moderate RPS, lacks built-in scaling, and is best for light or prototype workloads.



.jpeg)





