Join our VAR & VAD ecosystem — deliver enterprise AI governance across LLMs, MCPs & Agents. Become a Partner →

MCP Security Risks & Best Practices in 2026

Por Ashish Dubey

Actualizado: May 4, 2026

TrueFoundry effectively handles MCP security risks
Resumir con
Metallic silver knot design with interlocking loops and circular shape forming a decorative pattern.
Blurry black butterfly or moth icon with outstretched wings on white background.
Blurry red snowflake on white background, symmetrical frosty design with soft edges and abstract shape.

When Anthropic released the Model Context Protocol in late 2024, it addressed a real problem. Instead of writing custom integration code for every tool an AI agent needed to use, teams could connect agents to databases, APIs, and internal systems through a standardized interface. Adoption was fast. By 2025, tens of thousands of MCP servers had been deployed across enterprise environments.

But the speed of adoption exposed a gap that most organizations were not ready for: . MCP was designed to make AI agents more capable. It was not designed with enterprise security as a first principle. The original specification shipped without a mandatory authentication framework. The implicit trust model assumed that MCP servers were benign. And the tools those servers exposed were treated as the outputs of well-behaved systems, not as a new attack surface.

Security researchers documented the consequences quickly. CVE-2025-49596, with a CVSS score of 9.4, showed that unauthenticated MCP Inspector instances could be exploited to execute arbitrary commands. Invariant Labs demonstrated that a malicious MCP server could silently exfiltrate an entire WhatsApp message history. The Supabase Cursor agent incident in June 2025 showed how a privileged agent processing user-supplied support tickets could be tricked into leaking integration tokens through prompt injection.

These are not edge cases from poorly built experiments. They happened in real production systems because the architecture of MCP creates attack surfaces that traditional security tools were not built to see. 

This article covers the specific MCP security risks enterprises face, the controls that address them, and why the approach you use to enforce those controls matters as much as the controls themselves.

Your AI Agents Have Access to Everything, Does Your Security?

TrueFoundry enforces identity-aware execution and granular RBAC inside your private cloud environment.

The Top 5 MCP Security Risks Facing Enterprises

These are the five MCP security risks most commonly encountered in MCP deployments.

Over-Privileged Agent Access

When a developer first connects an AI agent to enterprise systems, they typically reach for the credentials they already have access to, often admin-level API keys or service accounts with broad permissions. That is because they want the agent to work without hitting authorization errors during development. The problem is that those broad permissions rarely get scoped down before the agent goes into production.

The OWASP MCP Top 10 project, launched specifically to track MCP vulnerabilities, identifies over-privileged access as a foundational risk. When a customer support agent carries the same credentials as a database administrator, a single successful attack against that agent does not expose a support queue. It exposes the database. The blast radius of any compromise scales directly with how much access the agent was given.

This is not a theoretical concern. The Supabase incident involved an agent running with privileged service-role access processing inputs that contained attacker-controlled content. The privilege level is what made the attack possible. An agent scoped to read-only access on support data would not have had the credentials to exfiltrate integration tokens.

Prompt Injection Through External Content

Prompt injection is ranked the number one vulnerability in the OWASP Top 10 for LLM Applications 2025, and it takes on a different character in MCP environments than in simple chatbot deployments. When an agent has the ability to execute actions, a successful injection does not just produce a bad output. It can trigger real operations: sending emails, calling APIs, modifying records, or forwarding data to external addresses.

The attack pattern is indirect. The attacker does not need access to your systems. They embed malicious instructions inside content the agent will process as part of its normal work: a PDF it is asked to summarize, a customer email it is routing, a web page it is reading for research. The agent reads the content, encounters the embedded instructions, and because there is no hard boundary between data and instructions in how language models process context, it interprets those instructions as legitimate and acts on them.

Security researcher Simon Willison, who has tracked this problem since 2022, wrote: "The curse of prompt injection continues to be that we've known about the issue for more than two and a half years and we still don't have convincing mitigations."

The MCP specification itself acknowledges the risk, noting that there should always be a human in the loop.

Tool Poisoning and the Rug Pull Attack

MCP agents trust tool metadata. When an agent connects to an MCP server and requests the list of available tools, the server responds with tool names, descriptions, and parameter schemas. The agent uses that metadata to decide which tools to call and how to call them. Tool poisoning exploits that trust relationship.

In a tool poisoning attack, an attacker crafts or compromises a tool's metadata to contain hidden instructions. The tool description appears normal in a visual inspection, but carries embedded directives that cause the agent to execute harmful actions, exfiltrate data, or escalate its own permissions. Because the malicious content is inside the tool definition rather than the user's input, it bypasses input validation layers that only check user-facing content.

Invariant Labs demonstrated a variant of this in 2025, showing how a malicious MCP server in the same agent context as a legitimate WhatsApp MCP server could use tool poisoning to silently read and export a user's entire message history. The attack required no user error and no network-level exploit.

A related attack is the rug pull, where a tool behaves legitimately at installation and then silently changes its behavior after the user has granted it permissions. MCP servers can update tool definitions without notifying the client, and most clients do not flag or detect those changes. 

Credential Sprawl Across Ungoverned MCP Servers

In environments without centralized credential management, each MCP server maintains its own authentication configuration. Developers store API keys in environment variables, hardcode OAuth tokens in configuration files, and use long-lived service account credentials because they are easier to manage than short-lived scoped tokens. Research tracking MCP deployments found 492 publicly exposed MCP servers lacking basic authentication or encryption.

CVE-2025-6514, a critical OS command injection vulnerability in mcp-remote, an OAuth proxy with over 437,000 downloads used in integrations from Cloudflare, Hugging Face, and Auth0, showed the supply chain dimension of this risk. 

A malicious MCP endpoint could send a crafted authorization URL that mcp-remote passed directly to the system shell, achieving remote code execution on the client machine and providing access to API keys, cloud credentials, local files, and SSH keys stored on that machine.

Audit Blind Spots That Break Compliance

Standard MCP servers do not log execution context in a structured, compliance-ready format. They may log that a tool was called. They typically do not log which human user the request originated from, what the agent's reasoning was at the point of the call, what data was returned, or whether an access policy was consulted before the operation was permitted.

For enterprises operating under SOC 2, HIPAA, or GDPR, this is not a minor gap. SOC 2 requires evidence of data management and processing controls. HIPAA requires audit trails of every system access that touches protected health information, attributed to an identifiable user or service. GDPR requires the ability to demonstrate what data was processed, when, and by whom. An MCP deployment without structured audit logging fails all of these requirements simultaneously.

A 2025 survey of AI system usage in enterprise environments found that fewer than 30% of AI systems had structured audit trails of agent tool access. Fewer than 15% could reconstruct the entire decision path for an agent action.

When a security incident occurs in that environment, the investigation relies on forensic reconstruction from incomplete logs, adding weeks to incident response time and making it nearly impossible to produce auditor-ready evidence.

 MCP security issues risks

Essential MCP Security Best Practices

The MCP security risks above share a common thread: they all stem from the absence of a controlled boundary between AI agents and the systems they access. Each best practice below is designed to establish and enforce that boundary at the layer where the attack occurs, not after the fact.

Enforce Identity-Aware Execution

The single most important architectural decision in any MCP deployment is how AI agent identity is managed. Shared service accounts create accountability gaps that no amount of logging can close. If five agents share one set of credentials, there is no way to attribute a specific action to a specific agent, let alone to the human user who triggered it.

The correct approach is On-Behalf-Of (OBO) authentication, where each AI agent request carries and exercises the exact permissions of the human user who initiated it. When a developer queries data sources through an MCP-connected agent, the operation happens with the developer's access level, not with admin-level credentials that happen to be configured on the server. The access is bounded by what that person is authorized to do everywhere else in the organization.

This requires integration with your enterprise identity provider, whether that is Okta, Azure AD, or a custom SSO setup. It means dynamic client registration and authorization code flows where tokens are user-scoped, short-lived, and automatically refreshed rather than long-lived static keys that persist until someone manually revokes them.

Apply Granular Role-Based Access Control at the Tool Level

Server-level access control is too coarse for production MCP deployments. Controlling whether a user or agent can reach an MCP server is a starting point. What you actually need is the ability to control which specific tools within that server each role can invoke.

The practical requirement is this: a customer support agent should be able to invoke read operations on a CRM server, but not delete operations. A finance agent should be able to query payment records, but not trigger outbound transfers. These distinctions cannot be expressed through server-level access policies. They require tool-level RBAC, where each tool within a server carries its own authorization requirements and those requirements are evaluated for every invocation.

The MCP specification's 2026 update introduced incremental scope consent, allowing clients to request only the minimum access needed for each operation rather than requesting all permissions upfront. Implementing this requires an authorization layer that understands tool-level semantics, not just network-level routing.

Require Human Approval for Irreversible Actions

Agents should not be able to execute destructive or high-risk operations without a human checkpoint. Database deletions, external data transfers, financial transactions, bulk record modifications, and outbound communications are all operations where a mistaken or injected instruction can cause damage that is difficult or impossible to reverse.

The MCP specification acknowledges this explicitly: there should always be a human in the loop with the ability to deny tool invocations. Implementing this in practice means annotating MCP tools with risk classifications and enforcing approval flows for operations tagged as destructive or irreversible. The agent presents the intended action with its arguments before executing, and waits for explicit human authorization.

This is not just a safety mechanism. It is also a prompt injection mitigation. An injected instruction that cannot execute without human approval cannot silently exfiltrate data. The approval step breaks the attack chain at the execution layer, even when detection at the input layer fails.

Consolidate Tool Connections Into a Single Governed Registry

Point-to-point connections between individual agents and individual MCP servers are what create the visibility gaps and governance failures described in the risks section. When each team manages its own server connections, there is no central place to see what tools exist, who has access to them, or what is being called in production right now.

A governed central registry changes this. Tools are registered once. Access policies are defined once and applied consistently. When a tool is updated or retired, that change is reflected in one place and agents pick it up automatically rather than continuing to call deprecated endpoints or breaking because they were not told about the change. Security teams get a single inventory instead of a spreadsheet they are always behind on.

The registry also mitigates the rug pull risk. When tool definitions are managed centrally, changes to tool metadata are versioned and auditable. A tool that silently mutates its own description after installation is detectable because the registry maintains a history of what each tool looked like at registration versus what it looks like now.

Point-to-point connection vs centralized MCP Gateway

Maintain Comprehensive, Structured Audit Logs

Every authentication event, token issuance, tool invocation, and policy decision should be logged with structured metadata that includes who made the request, which agent executed it, which tool was called, what arguments were passed, what was returned, and whether any policy applied or was overridden.

The distinction between having logs and having compliance-ready logs is significant. Audit logs that record tool names and timestamps satisfy a narrow technical requirement. Logs that record user identity, agent identity, full request context, applied policies, and output content in a structured format that can be queried and exported on demand satisfy SOC 2, HIPAA, and GDPR requirements. These are different things, and the gap between them is where most audit failures occur.

Logs should be retained within your own environment. Logs that live in a third-party SaaS system are not fully under your control, cannot be produced on-demand with complete certainty, and may themselves be subject to data handling restrictions that conflict with what your compliance framework requires.

The Flaws in Current Market Solutions

There is no shortage of tools marketed as MCP security solutions. The question is whether they address the MCP security issues at the layer where those security vulnerabilities actually occur, or whether they apply general-purpose controls to a problem that requires something more specific.

  • Legacy API gateways cannot read agent intent. Traditional API gateways were built for HTTP traffic between services. They inspect headers, enforce rate limiting, and apply MCP authentication at the transport layer. An AI agent task that triggers tool calls to 20 different MCP tools generates traffic that, at the network level, appears as a sequence of legitimate, authenticated requests. There is nothing in the HTTP envelope that reveals whether those requests were driven by a user action or by an injected instruction buried in a document the agent read three steps earlier. Gateways that cannot parse MCP protocol semantics cannot detect or stop semantic-layer attack vectors against AI applications.
  • Managed SaaS platforms route your data through infrastructure you do not control. Cloud-hosted AI platforms solve the deployment problem but create a data residency problem. When an AI agent tool discovery requests and MCP interactions route through a third-party cloud, sensitive information leaves your network boundary on every interaction. For organizations with HIPAA, GDPR, or financial services data residency obligations, this creates AI compliance exposure that exists independent of whether the SaaS vendor's own security posture is adequate. You cannot satisfy a data residency requirement by trusting someone else's infrastructure. This is an acute security concern for teams running remote MCP servers with access to user data.
  • DIY open-source configurations break down at scale. Building a custom proxy from open-source components avoids vendor fees and gives engineering teams maximum control, but it creates maintenance obligations that grow faster than the teams managing them. Tool registration through static YAML or JSON files means every new integration requires a manual edit, a code review, and a deployment. RBAC policies written in configuration files cannot be changed in real time. Audit logging requires custom instrumentation for every server. When you are managing a handful of agents, this is manageable. When you are managing dozens, it becomes a full-time platform engineering problem.

Legacy Gateways Cannot Read Agent Intent, Your MCP Security Layer Must

Get semantic-aware governance, credential vaulting, and SOC 2 audit trails built in with TrueFoundry.

Securing Agent Workflows with TrueFoundry

TrueFoundry's MCP Gateway is built around the premise that the controls enterprises need to run MCP safely in production cannot be bolted onto existing infrastructure. They need to be part of the infrastructure.

The gateway deploys entirely within the customer's AWS, GCP, or Azure environment. Every tool request, every discovery call, every prompt-response pair stays inside the organization's network boundary. There is no SaaS routing layer or third-party infrastructure in the data path. No external dependency that creates data egress risk. This matters not just for compliance but for the threat model: data that never leaves your environment cannot be intercepted in transit.

  • OAuth 2.0 identity injection with On-Behalf-Of flows: Agents act under the exact permissions of the requesting user, inherited through the organization's existing identity provider. Okta, Azure AD, and custom SSO setups are supported out of the box. There are no shared service accounts. Every tool call is attributed to a specific human identity, producing an accountability chain that survives any compliance review.
  • Tool-level RBAC enforced at the gateway layer: Access policies are applied before requests reach any model or tool. The registry exposes only the tools each role or team is authorized to use, not the full surface area of every connected server. A support agent sees CRM lookup tools. It does not see database administration commands. The distinction is enforced at the platform level, not by developer convention.
  • Virtual MCP Server abstraction for supply chain protection: Backend tool implementations sit behind a virtual abstraction layer in the registry. When a tool's underlying implementation changes, or when a compromised tool needs to be replaced, the change happens in the registry without touching any of the agents that depend on it. This also insulates agents from rug pull attacks: tool definitions are version-controlled through the registry, and changes to tool metadata are tracked and auditable.
  • Human-in-the-loop approval flows for high-risk operations: Tools can be annotated with risk classifications. Operations tagged as destructive or irreversible require explicit human authorization before execution. This is configurable per tool and per role, so teams can require approval for database deletions while allowing reads to proceed without interruption.
  • Structured audit logs retained inside your own environment: Every request is logged with full metadata: user identity, agent identity, model, tool, arguments, response, applied policies, latency, and cost. Logs are retained in the customer's own infrastructure and can be exported in structured JSON format for integration with Grafana, Splunk, Datadog, or any existing observability pipeline. Innovaccer, processing approximately 17 million AI inference requests per month under HIPAA in AWS GovCloud, uses TrueFoundry's structured telemetry through OpenTelemetry and views it in Grafana, with no sensitive data ever leaving their cloud boundary.

Want to experience TrueFoundry’s AI Gateway? Book a demo today!

Conclusion: Making MCP Safe for Production

MCP solved the integration problem. It did not solve the governance problem. The tens of thousands of MCP servers deployed in enterprise environments in 2025 created an enormous amount of new capability, and an equally significant expansion of attack surface that most organizations were not equipped to govern.

The incidents that have already occurred including, Supabase, Asana, Atlassian, the mcp-remote CVE, were not caused by poor engineering. They were caused by an architecture that did not account for the security requirements of enterprise production environments. The base protocol was not built with those requirements in mind.

Addressing MCP security risks in a way that actually works in production means enforcing controls at the infrastructure layer: identity-aware execution, tool-level access control, centralized credential management, human approval for destructive operations, and structured audit logs retained in your own environment. These are not optional features for regulated industries. They are the baseline requirements for running AI agents that touch business-critical systems.

TrueFoundry deploys all of those controls inside your own cloud. Your data stays where your governance applies. Your audit logs are yours to produce, query, and export on your schedule.

La forma más rápida de crear, gobernar y escalar su IA

Inscríbase
Tabla de contenido

Controle, implemente y rastree la IA en su propia infraestructura

Reserva 30 minutos con nuestro Experto en IA

Reserve una demostración

La forma más rápida de crear, gobernar y escalar su IA

Demo del libro

Descubra más

No se ha encontrado ningún artículo.
TrueFoundry effectively handles MCP security risks
May 4, 2026
|
5 minutos de lectura

MCP Security Risks & Best Practices in 2026

No se ha encontrado ningún artículo.
May 4, 2026
|
5 minutos de lectura

AI Gateway On Premise: todo lo que necesita saber

No se ha encontrado ningún artículo.
May 4, 2026
|
5 minutos de lectura

Controlling Claude Code in the Enterprise: MCP Tool Scoping and Audit Trails

No se ha encontrado ningún artículo.
May 2, 2026
|
5 minutos de lectura

The Agentic Token Explosion: Cost Attribution & Budgets for Claude Code in CI/CD

No se ha encontrado ningún artículo.
No se ha encontrado ningún artículo.

Blogs recientes

Black left pointing arrow symbol on white background, directional indicator.
Black left pointing arrow symbol on white background, directional indicator.
Realice un recorrido rápido por el producto
Comience el recorrido por el producto
Visita guiada por el producto
-->