
An open, standardized framework that enables Large Language Models (LLMs) to connect with external tools and data sources, allowing AI agents to perform complex tasks beyond their original training data. Developed by Anthropic, MCP provides a standardized, two-way communication channel, similar to an API, that allows AI systems to access real-time information and execute actions by interacting with services, applications, and databases.
MCP adds a level of logic to the LLMs, through which the LLMs become capable of executing deterministic problems. Instead of generating code or randomly fetching resources, the LLMs will just analyse your prompts and run the commands you have written for code execution or for context.
The MCP primarily has 3 layers:
The MCP Host is the chat application that is user facing, this can be GitHub Copilot in VS Code or Claude Code Desktop app. You can think of an MCP host as an orchestrator. It creates and manages the MCP clients. MCP Client maintains a channel for an MCP server and gets the context/metadata from MCP server for the Host. MCP Server is the place where the logic is defined and run. MCP server provides this metadata to client as a context to LLMs.
There are 3 main constructs currently stable in MCP Servers:
Whenever we are talking about MCP, We primarily focus on the MCP Servers. And in MCP server tools are the hot topic. Tools give the LLMs a deterministic capability to execute programs that can be tested and verified.
For more info: What is Model Context Protocol?
In simple terms you can think of MCP + LLMs as an AI wrapper around a REST API. The AI augments the execution of the logic that you define, and provides output in natural language.
Here's how it works when a user prompts in a Chat Host:
In a recent news, Anthropic is Donating the Model Context Protocol and establishing the Agentic AI Foundation. This makes it a very exciting space currently for MCP. This article is the basic mental model to think about MCPs, please let me know for any suggestions.