With innovation comes risk. As organizations race to build AI-first infrastructure, security is struggling to keep pace. Multi-Agentic Systems – those built on Large Language Models (LLMs) and Multi-Component Protocols (MCP) - bring immense potential, but also novel vulnerabilities that traditional tools weren’t designed to handle.
At Wallarm, we’re closely following emerging guidance around these new attack surfaces, including the recently released OWASP Multi-Agentic System Threat Modeling Guide v1.0. Drawing from those recommendations, we’ve compiled a practical MCP Security Checklist that outlines real-world risks and defensive best practices for securing AI-native environments.
Click on the following links to navigate yourself through the checklist
In AI-native systems, agents aren’t just responding to queries; they’re generating and executing code. That opens the door to subtle and dangerous prompt injection. If an attacker injects malicious instructions like OS commands into a prompt, the LLM can unknowingly embed and execute them. Here’s how you can mitigate it:
When an LLM is unstable, it can cause the MCP client to send inconsistent or erratic requests to the MCP server, disrupting operations. For example, a model might attempt a task like building an SQL query but encounter an error. Instead of stopping, it retries the same malformed instruction repeatedly, generating excessive traffic to the backend. This instability can spiral into a denial-of-service issue. To prevent it:
Remember: models don’t always know when they’re wrong, so your infrastructure needs to.
Agents operating in MCP environments can access tool APIs to perform calculations, conversions, or business operations. But without proper validation, they can be tricked into using the wrong tool for the task, altering outputs or undermining trust. To mitigate this risk:
In a distributed agent ecosystem, identity is everything. If a malicious actor obtains valid credentials, or if the system fails to verify an agent identity correctly, they can access privileged tools or data. To prevent this risk:
Poorly implemented communication protocols like JSCON-RPC and SSE may allow attackers to intercept, modify, or inject data in transit. For example, requests and responses between the MCP client and server are sent over unencrypted HTTP. An attacker manipulates responses, leading to invalid LLM outputs. Mitigating this risk involves:
Attackers may overload AI systems with CPU, memory, or service requests to degrade performance or cause denial of service. For example, a malicious actor may trigger a large file read operation 100 times or send a pg_sleep(600) command to freeze database operations. You can mitigate this risk by:
MCP service accounts may be misconfigured, allowing credential exposure through tools, logs, or file access. For example, an LLM connected to a tool that reads environment variables might inadvertently leak AWS credentials. You can prevent this by:
Attackers can manipulate logging behavior to hide malicious activity or overwhelm systems with noise. For example, they could change log levels to DEBUG or TRACE to hide attacks, overload the log system with useless events, or disable logging functionality via configuration manipulation. Mitigation strategies include:
Attackers can escalate privileges through misconfiguration, prompt injection, or manipulation of access logic. For example, a prompt could convince the LLM to assign admin privileges to the attacker, or a malicious user could downgrade admin group privileges to match user access levels. To mitigate this:
If an MCP server has excessive access to host resources or the network, compromise can escalate across systems. For example, a vulnerable tool on the MCP server allows path traversal, exposing secrets from other services or MCP environments. To prevent this:
An attacker could deploy a malicious MCP server posing as a legitimate one. Agents connecting to this rogue server are then compromised. This is an ecosystem-level attack targeting the trust model of MCP. To mitigate this risk:
Without robust security models, the flexibility of MCP and multi-agent systems becomes fragile. By aligning with OWASP’s threat modelling guide and following the recommendations above, organizations can build AI infrastructure that’s as secure as it is intelligent.
Need help evaluating your risk exposure or building safer agent ecosystems? Request a demo today.
Your board wants AI. Your developers are building with it. Your budget committee is asking…
AI systems are no longer just isolated models responding to human prompts. In modern production…
Broken authorization is one of the most widely known API vulnerabilities. It features in the…
The shadow technology problem is getting worse. Over the past few years, organizations have scaled…
API security has been a growing concern for years. However, while it was always seen…
It’s an unusually cold winter morning in Houston, and Craig Riddell is settling into his…