API Security

Reducing False Positives in API Security: Advanced Techniques Using Machine Learning

False positives in API security are a serious problem, often resulting in wasted results and time, missing real threats, alert fatigue, and operational disruption. Fortunately, however, emerging technologies like machine learning (ML) can help organizations minimize false positives and streamline the protection of their APIs. Let's examine how.  

What are the Risks Associated with False Positives in API Security?

Before discussing how machine learning can reduce false positives in API security, it's essential to understand their risks, which fall into two categories: business and security. 

  • Business Risks: False positives can cause financial losses by blocking legitimate API calls, leading clients to cancel contracts, and damaging the service provider's reputation. These issues increase support costs and impact multiple teams.
  • Security Risks: False positives can distort service analytics, leading to poor decision-making and weakened threat responses. Security teams might disable protective measures to reduce alerts, introducing vulnerabilities. Additionally, blocked API requests can corrupt data sets, compromising decision-making and increasing exposure to attacks.

How Can Machine Learning Minimize False Positives in API Security?

ML technologies enhance API security by considering the context of each request—such as a user's action history, geolocation, and time of access—to make better-informed decisions and reduce false positives that might occur due to context-insensitive rules.  

This process works by ML algorithms analyzing historical API traffic, such as typical request rates, commonly accessed endpoints, usual data payloads, and user interaction patterns, to establish a baseline of normal behavior. The algorithm compares ongoing behaviors to this baseline and alerts security teams to any deviations. Over time, the algorithm learns more about legitimate behavior patterns, thus reducing the chance of flagging them as threats.

 Machine learning-powered techniques improve on traditional, static methods, as the latter does not account for the dynamic nature of API usage. ML models continuously learn and evolve with data, providing more accurate detection and minimizing disruptions caused by false alarms. This adaptability ensures that security measures remain effective while maintaining a seamless user experience. 

For example, a common API attack prone to false positives is Account Takeover, where attackers gain unauthorized access using compromised credentials. This attack can mimic legitimate user behavior, such as multiple login attempts or using a VPN, which traditional rule-based systems flag as suspicious. Wallarm’s ML-based API Abuse Prevention solution handles this by analyzing various indicators like authentication time, location, IP reputation, and number of login attempts. They adapt to the specific behavior of users and the protected API, allowing for legitimate actions (e.g., password errors) without unnecessary blocking, reducing false positives effectively.

Wallarm has proven the value of AI to improve detection since BSides 2019 in San Francisco. You can watch Wallarm's CEO, Ivan Novikov, discussing how to reduce false positive events, and you can also get access to the respective open-source code in GitHub.

Challenges in Using Machine Learning to Reduce False Positives

That said, using machine learning technology to reduce false positives is challenging. As an API security provider, Wallarm is all too aware of this.  

Developing Approaches for Multiple Customers

First and foremost, it can be difficult to develop approaches that work for multiple, diverse customers, as each customer has their own APIs with unique features. While it's possible to build an ML model that works well for a specific API, creating one that works for all APIs can be a serious challenge. To overcome this, Wallarm identifies the clients for whom the model does not work as well as it should and adapts detectors to the specific characteristics of the customer's API. 

Insufficient or Poor-Quality Training Data

Moreover, accurate results rely on a large volume of real traffic on which the model can be trained, and some customers may not have the necessary traffic or, if the customer is trialing the solution, use synthetic traffic. Other customers may have highly diverse traffic whereby all users behave differently, or, conversely, the traffic is homogeneous, and users are all but indistinguishable from one another. In all cases, the solution is to use Wallarm's special detectors that combine the traditional, signature-based approach to threat detection with threat intelligence feeds.  

Changes in Traffic Volume or APIs

Another potential problem is changes in traffic volume or APIs. For example, during sales events, the traffic volume can increase sharply, and users may start behaving much more actively than before. In this case, the appearance of more active users may be perceived as an attack. To solve this problem, Wallarm updates the models in our detectors before each traffic analysis using new traffic from the customer. The thresholds in the detectors are adjusted to account for the changes in traffic if such behavior is present in most sessions, ensuring we do not block new, more active users. 

Minimizing False Positives Without Missing Genuine Threats

When attempting to minimize false positives in API security, it's crucial to ensure that solutions don't miss genuine threats. Wallarm's API Abuse Prevention solution uses several methods to address this: 

  • Filtering System: A rule-based system filters out false positives by accounting for specific behaviors, like health checks or non-standard automation, which may be detected as suspicious but are essential for the customer.
  • Weighted Voting System: This system assigns different weights to multiple detectors, enabling them to focus on specific signs of automated threats for more accurate detection.
  • Regular Evaluation: Automated and manual quality checks ensure the accuracy of results, using tools to visualize API activity and confirm whether a session is genuinely malicious or business-related.
  • Hybrid Approach: The Wallarm solution combines signature-based methods with machine learning. Depending on customer needs, the approach adjusts to prioritize either accuracy or broader threat detection, balancing false positives and negatives.

Balancing Machine Learning Tools with Human Oversight and Expertise

It's also important to consider that human oversight is absolutely necessary when using AI technologies like ML tools. For organizations deploying their own ML tools for API security, consider the following:

  • Set Clear Goals: Define your objectives for using ML, like reducing false positives or detecting advanced threats. If the problem can be solved without ML, use more straightforward methods.
  • Educate Your Team: Ensure your security team understands ML concepts, especially those specific to your tools. Utilize public resources, courses, and thorough documentation review.
  • Data Quality: Train ML models with accurate and relevant data to ensure optimal performance.
  • Transparency: Make sure security analysts can interpret and control ML decisions and results.
  • Clear Protocols: Establish protocols for escalating alerts from ML tools to human analysts, specifying timelines and decision-making roles.

For Wallarm's models, customer support and regular result verification enhance tool quality and user benefits.  Book a demo of our cutting-edge solutions to see how Wallarm can help your organization.

Recent Posts

The CISO’s Dilemma: How To Scale AI Securely

Your board wants AI. Your developers are building with it. Your budget committee is asking…

2 days ago

Agent-to-Agent Attacks Are Coming: What API Security Teaches Us About Securing AI Systems

AI systems are no longer just isolated models responding to human prompts.  In modern production…

3 days ago

Everyone Knows About Broken Authorization – So Why Does It Still Work for Attackers?

Broken authorization is one of the most widely known API vulnerabilities.  It features in the…

2 weeks ago

From Shadow APIs to Shadow AI: How the API Threat Model Is Expanding Faster Than Most Defenses

The shadow technology problem is getting worse.  Over the past few years, organizations have scaled…

3 weeks ago

Inside Modern API Attacks: What We Learn from the 2026 API ThreatStats Report

API security has been a growing concern for years. However, while it was always seen…

4 weeks ago

CISO Spotlight: Craig Riddell on Curiosity, Translation, and Why API Security is the New Business Imperative

It’s an unusually cold winter morning in Houston, and Craig Riddell is settling into his…

1 month ago