The Fastest Video API.
Measured.
Every number on this page comes from production Prometheus metrics on c5.xlarge EC2 instances. No marketing fluff. No synthetic benchmarks. Real latency, measured at p99.
Every number on this page comes from production Prometheus metrics on c5.xlarge EC2 instances. No marketing fluff. No synthetic benchmarks. Real latency, measured at p99.
The time from TCP connection to the first byte of your handler executing. Every millisecond here multiplies across every API call your application makes.
Time from "join call" to peer connection established. This is the latency your users feel when they click "Join Meeting" and wait for video to appear.
Time to process a 60-second video through the pipeline: decode, transcode, assemble, and export. V100's Rust orchestration layer eliminates the overhead that makes other platforms slow.
Every API call hits rate limiting and authentication before reaching business logic. This overhead is invisible to most developers but adds up across millions of requests.
Rate limiting and auth checks happen on every single API call. At 10,000 requests/second, the difference between 0.1ms and 10ms is the difference between 1 second and 100 seconds of cumulative overhead per second.
The performance advantage nobody talks about. When competitors need 3-5 vendor round-trips to do what V100 does in a single API call, the latency difference is massive.
Infrastructure decisions compound. The language, runtime, and architecture choices made at the foundation level determine every performance ceiling above them.
| V100 | Mux | Shotstack | Cloudinary | Descript | |
|---|---|---|---|---|---|
| Language | Rust | Go | Node.js | Java/Scala | Python |
| Gateway overhead | <5ms | ~20ms | ~50ms | ~100ms | N/A |
| GC pauses | None | ~1ms | ~10ms | ~50ms | ~100ms |
| Binary size | ~11MB | N/A | N/A | N/A | N/A |
| Cold start | 0ms | N/A | ~500ms | ~2s | N/A |
| Architecture | 16 Rust services | Microservices | Monolith | Monolith | Monolith |
We publish our methodology because we want you to verify these numbers, not just trust them.
Don't take our word for it. Get an API key, point your load tester at our endpoints, and measure the latency yourself.
# Measure gateway latency
curl -w "connect: %{time_connect}s\nttfb: %{time_starttransfer}s\ntotal: %{time_total}s\n" \
-H "Authorization: Bearer YOUR_API_KEY" \
https://api.v100.ai/v1/health
# Typical response:
connect: 0.002s
ttfb: 0.004s
total: 0.005s