# rate-limiting
14 articlestagged with “rate-limiting”
Rate Limiting and Abuse Assessment
Assessment of rate limiting bypass techniques, cost-based attacks, and billing abuse in AI services.
Rate Limiting and Abuse Prevention
Implementing rate limiting and abuse prevention for LLM API endpoints and applications.
Rate Limiting, Sandboxing & Execution Controls
Rate limiting strategies for AI APIs, sandboxing code execution with E2B and Docker, tool call approval workflows, and the principle of least privilege for AI agents.
AI API Abuse Detection
Detecting and mitigating API abuse patterns targeting AI inference endpoints including prompt extraction and model theft.
LLM API Security Testing
Security testing methodology for LLM APIs, covering authentication, rate limiting, input validation, output filtering, and LLM-specific API vulnerabilities.
Advanced Rate Limiting Strategies for LLM API Endpoints
Designing, attacking, and defending rate limiting systems for LLM inference APIs to prevent abuse, model extraction, and resource exhaustion
Lab: Rate Limit Enumeration and Bypass
Enumerate API rate limits and test common bypass techniques including header manipulation and request distribution.
Basic Rate Limit Abuse Patterns
Test common rate-limit bypass patterns including header manipulation and endpoint discovery.
AI API Reverse Engineering
Techniques for reverse engineering AI APIs including mapping undocumented endpoints, parameter discovery, rate limit profiling, and extracting implementation details from API behavior.
API Rate Limit Bypass
Techniques to bypass API rate limiting on LLM services, including header manipulation, distributed requests, authentication rotation, and endpoint discovery.
Rate Limiting and Abuse Prevention for LLM APIs
Walkthrough for implementing rate limiting and abuse prevention systems for LLM API endpoints, covering token bucket algorithms, per-user quotas, cost-based limiting, anomaly detection, and graduated enforcement.
AI Rate Limiting Walkthrough
Step-by-step walkthrough for implementing token-aware rate limiting for AI applications: request-level limiting, token budget enforcement, sliding window algorithms, abuse detection, and production deployment.
Adaptive Rate Limiting for LLM APIs
Implement adaptive rate limiting that adjusts thresholds based on detected attack patterns and user behavior.
AI API Red Team Engagement
Complete walkthrough for testing AI APIs: endpoint enumeration, authentication bypass, rate limit evasion, input validation testing, output data leakage, and model fingerprinting through API behavior.