Requesty is a unified LLM gateway and router that provides access to 400+ models across 20+ providers through a single OpenAI-compatible API endpoint. Based in London, the company positions itself as 'Cloudflare for AI' — an infrastructure layer that sits between applications and LLM providers, handling intelligent routing, automatic failover, cost optimization, and enterprise governance.
Founded in 2023, Requesty pivoted from data analytics to the LLM gateway space in early 2025. The company raised a GBP 2M (~$3M) seed round led by 20VC (Harry Stebbings) with participation from Tapestry VC and others. At the time of funding, Requesty had reached $1.5M ARR and 25,000+ developers, growing to 50,000+ developers by early 2026. Notable enterprise clients include Shopify, Amadeus, Chargebee, and Pfizer.
Requesty differentiates through its transparent pricing model (flat 5% markup on base model costs), smart routing that automatically selects the optimal model per request, semantic caching that delivers up to 40% cost reduction, and enterprise governance features including PII detection, spending controls, and EU data residency. The platform claims 99.99% uptime SLA with failover in under 20ms.
Free trial available
Enterprise AI teams needing governed LLM access
Requesty and Respan operate at adjacent layers of the AI infrastructure stack. Requesty handles model routing and gateway management, while Respan provides deep observability, evaluation, and optimization of the LLM calls flowing through gateways like Requesty.
Top companies in LLM Gateways you can use instead of Requesty.
Companies from adjacent layers in the AI stack that work well with Requesty.
Last verified: March 27, 2026