The EU AI Act’s most consequential deadline is less than five months away. Starting August 2, 2026, organizations that provide or deploy high-risk AI systems must demonstrate full compliance with requirements covering risk management, data governance, record-keeping, transparency, and human oversight. The penalties for falling short are severe — up to €35 million or 7% of global annual turnover, whichever is higher.
If your organization exposes AI capabilities through APIs — whether that’s a model inference endpoint, an AI-powered recommendation service, or an agent that calls tools via the Model Context Protocol — every one of those API interactions is a compliance touchpoint. And the enforcement mechanism best positioned to address these requirements is the one you may already have in place: your API gateway.
This guide maps specific EU AI Act requirements to concrete API gateway capabilities, gives you a compliance checklist you can start working through today, and explains how Zuplo’s policy-based approach makes implementation straightforward.
- The August 2026 deadline explained
- How the EU AI Act applies to APIs
- API gateway capabilities that support compliance
- Compliance checklist for API teams
- MCP and agent governance under the EU AI Act
- Getting started with Zuplo
The August 2026 deadline explained
The EU AI Act entered into force on August 1, 2024 and follows a phased implementation timeline. Prohibited AI practices were banned in February 2025. General-purpose AI (GPAI) model obligations took effect in August 2025. But the biggest wave of requirements — those covering high-risk AI systems listed in Annex III — takes full effect on August 2, 2026. (Note: The European Commission’s Digital Omnibus proposal, currently in trilogues, may extend this deadline for certain Annex III obligations. Organizations should plan for the original date until any extension is formally adopted.)
By that date, providers of high-risk AI systems must have:
- Completed conformity assessments proving their systems meet the Act’s requirements
- Finalized technical documentation demonstrating compliance
- Affixed CE marking to approved systems
- Registered systems in the EU database
- Implemented ongoing post-market monitoring
Who’s affected
The EU AI Act has extraterritorial scope, mirroring the GDPR’s reach. You’re subject to the regulation if:
- You provide or deploy AI systems within the EU
- Your AI system’s output is used in the EU, even if the system itself runs elsewhere
- You offer AI-enabled services to individuals in the EU, regardless of where your company is headquartered
A US-based company running AI models on AWS that serves European customers through an API? In scope. A startup in Singapore whose AI recommendation engine is consumed by an EU-based e-commerce platform? Also in scope.
The penalty structure
The EU AI Act’s fines exceed even the GDPR:
- €35 million or 7% of global turnover for prohibited AI practices
- €15 million or 3% of global turnover for violating high-risk AI system requirements
- €7.5 million or 1% of global turnover for supplying incorrect information to authorities
For SMEs and startups, the fines are capped at the lower of the fixed amount or the percentage, but even the reduced amounts are substantial enough to threaten business viability.
How the EU AI Act applies to APIs
If your AI system is accessible via an API, the Act’s requirements apply to the entire chain — from the model itself to the infrastructure that serves it. Here’s how key articles translate to API-level concerns:
Article 12: Record-keeping
High-risk AI systems must support automatic recording of events (logs) throughout their lifecycle. These logs must capture enough detail to:
- Identify situations where the AI system may present a risk
- Facilitate post-market monitoring
- Enable monitoring of the system’s operation by deployers
For API-exposed AI systems, this means every request and response flowing through your AI endpoints should be logged with sufficient detail — who made the request, what data was sent, what the system returned, and when it happened. Under Article 19 and Article 26(6), these logs must be retained for at least six months.
Article 9: Risk management
The Act requires a continuous risk management system that identifies and mitigates risks throughout the AI system’s lifecycle. At the API level, this includes implementing rate limiting to prevent misuse, monitoring for anomalous usage patterns, and having the ability to quickly restrict or shut down access when risks are identified.
Article 10: Data governance
Data governance requirements mandate strict controls over who can access AI systems and the data they process. Special categories of personal data require “strict controls and documentation of the access, to avoid misuse and ensure that only authorised persons have access.” This maps directly to authentication and authorization at the API layer.
Article 14: Human oversight
High-risk AI systems must be designed so humans can effectively oversee their operation. For API-accessible systems, this means building in mechanisms to monitor real-time usage, intervene when needed (e.g., disable an endpoint or revoke access), and review the system’s decisions through comprehensive audit trails.
API gateway capabilities that support compliance
An API gateway sits at the intersection of every requirement above. It processes every request before it reaches your AI system and every response before it reaches the consumer. That position makes it a natural enforcement point for EU AI Act compliance.
Logging and audit trails → Article 12
The Act’s record-keeping requirements demand comprehensive, structured logging of all AI system interactions. Your API gateway can capture:
- Request metadata: Timestamp, source IP, authenticated user identity, request path, and HTTP method
- Request payloads: The actual prompts or input data sent to AI systems
- Response metadata: Status codes, latency, response size
- Response payloads: The AI system’s output, critical for auditing decisions made by high-risk systems
Zuplo provides
built-in structured logging with
every request automatically tagged with fields like requestId, environment,
and buildId for full traceability. You can
log custom request and response data
by adding inbound and outbound policies that capture headers, query parameters,
and body content — with the ability to redact sensitive fields like
Authorization headers.
For long-term retention and compliance reporting, Zuplo integrates with nine major logging platforms including Datadog, Splunk, AWS CloudWatch, Dynatrace, and Google Cloud Logging. This means you can route your AI endpoint audit logs directly into the systems your compliance team already uses.
Authentication and access control → Articles 9 and 10
The Act requires strict controls over who can access AI systems. Your API gateway enforces this through:
- API key authentication: Issue unique keys to each consumer with Zuplo’s API key management, creating a clear record of which consumers access which AI endpoints
- JWT validation: Validate tokens from identity providers like Auth0, Clerk, or any OpenID Connect provider to enforce identity-based access
- Multiple auth methods: Combine API keys and JWT tokens on the same route using multiple authentication policies to support different consumer types while maintaining consistent access logs
Every authenticated request creates an audit record tying a specific identity to a specific AI system interaction — exactly what Article 10’s data governance provisions demand.
Rate limiting → Article 9
Rate limiting is a risk management control. By capping how many requests a consumer can make to your AI endpoints, you:
- Prevent misuse: Stop consumers from overwhelming your AI system in ways that could produce unreliable or harmful outputs
- Enforce fair usage: Ensure no single consumer monopolizes system resources
- Create usage baselines: Establish normal consumption patterns that make anomalies visible
Zuplo’s rate limiting policy supports limiting by IP address, authenticated user, API key, or custom attributes. You can set different limits for different consumer tiers — giving trusted, audited consumers higher limits while keeping tighter controls on others.
For AI-specific use cases, Zuplo’s AI Gateway adds cost controls that let you set spending limits per application and per team — preventing runaway usage that could indicate misuse or a compromised credential.
Observability and monitoring → Article 14
Human oversight requires the ability to monitor AI system operations in real time. Your API gateway provides:
- Real-time dashboards: Track request volumes, error rates, and latency across your AI endpoints
- Anomaly detection: Identify unusual patterns like sudden traffic spikes or unexpected error rates that may indicate the AI system is behaving outside its intended parameters
- Usage analytics: Understand which consumers are using which AI capabilities and how usage patterns change over time
Zuplo’s API analytics give your team visibility into every AI endpoint interaction, while integrations with monitoring platforms like Datadog and New Relic let you build custom dashboards and alerts tailored to your compliance requirements.
Compliance checklist for API teams
Here’s a practical checklist you can work through before August 2026. Steps 1 and 2 should be completed first to establish your baseline. Steps 3 through 6 can be tackled in parallel depending on your team’s capacity.
1. Audit your AI-exposed endpoints
- Inventory every API endpoint that exposes an AI system or AI-powered feature
- Classify each endpoint’s risk level based on the EU AI Act’s categories
- Document the intended purpose and potential impact of each AI endpoint
- Identify which endpoints fall under Annex III high-risk categories
2. Implement comprehensive request and response logging
- Enable structured logging on all AI-related endpoints
- Capture request payloads (with PII redaction where appropriate)
- Log response data for auditability of AI system outputs
- Ensure logs are tagged with consumer identity, timestamps, and request IDs
- Configure log retention for a minimum of six months, as required by the Act
- Route logs to a centralized platform your compliance team can query
3. Enforce authentication on all AI endpoints
- Require API key or JWT authentication on every AI-related route — no unauthenticated access
- Implement role-based access control to restrict which consumers can access which AI capabilities
- Maintain an up-to-date registry of all authorized consumers and their permission levels
- Set up automated key rotation and revocation procedures
4. Configure rate limiting for AI consumers
- Set per-user and per-key rate limits on all AI endpoints
- Implement differentiated limits based on consumer trust levels
- Monitor for rate limit violations as potential indicators of misuse
- Document your rate limiting rationale as part of your risk management system
5. Establish audit trail exports
- Configure automated log exports to your compliance and legal teams’ preferred systems
- Build or deploy dashboards that answer compliance-critical questions: who accessed what AI system, when, and what was the output?
- Test your ability to produce a complete audit trail for any given AI system interaction within a reasonable timeframe
- Verify your log pipeline can handle the retention requirements (minimum six months)
6. Document everything
- Maintain technical documentation mapping each AI endpoint to its compliance controls
- Record your risk management decisions — why you chose specific rate limits, what access control model you use, and how you handle incident response
- Keep this documentation current; the Act requires ongoing compliance, not a one-time assessment
MCP and agent governance under the EU AI Act
The rise of AI agents that autonomously call tools and APIs adds another layer of compliance complexity. When an AI agent uses the Model Context Protocol (MCP) to discover and invoke tools, each tool call is an API interaction subject to the same EU AI Act requirements.
Consider the compliance implications:
- Which tools can an agent access? The Act’s risk management requirements (Article 9) demand that you control what an AI system can do. Unrestricted tool access is a compliance risk.
- What data does the agent process through each tool? Article 10’s data governance requirements apply to every piece of data that flows through an agent’s tool interactions.
- Is every tool call logged? Article 12’s record-keeping requirements extend to the full chain of an agent’s actions, not just its final output.
- Can a human intervene? Article 14 requires the ability to monitor and override an agent’s behavior, which means visibility into its tool usage in real time.
Zuplo’s MCP Gateway addresses these requirements directly:
- Virtual MCP servers let you curate which tools each team or application can access, implementing the principle of least privilege at the tool level
- Granular access control enforces who can use which MCP servers, with permissions configurable at the team, user, or application level
- Centralized audit trails capture every tool call across all MCP interactions, providing the record-keeping Article 12 requires
- Security policies including PII detection and prompt injection protection help meet the Act’s cybersecurity and data governance requirements
For organizations deploying AI agents that serve EU users, MCP governance isn’t optional — it’s a compliance requirement.
Getting started with Zuplo
Zuplo’s policy-based architecture makes it straightforward to implement the compliance controls described in this guide. Instead of writing custom middleware for every requirement, you configure declarative policies that apply to your routes:
- Add authentication — Apply an API key or JWT validation policy to your AI endpoints in minutes
- Enable rate limiting — Configure the rate limiting policy with per-user limits appropriate for your AI workloads
- Set up logging — Use Zuplo’s built-in logging with custom request and response data capture, then route logs to your compliance platform
- Govern MCP servers — If you’re running AI agents, use the MCP Gateway to centralize tool access control and audit logging
Every policy is version-controlled through Zuplo’s GitOps workflow, meaning your compliance configuration is auditable and reproducible — another requirement the Act implicitly demands through its documentation and traceability provisions.
The August 2026 deadline will arrive faster than most teams expect. The organizations that start mapping their API gateway capabilities to EU AI Act requirements today will be the ones that pass conformity assessments without scrambling. Start with the checklist above, and explore Zuplo to see how policy-based API governance turns compliance requirements into configuration.