FC

⚡ API Response Time Calculator

Analyse API performance metrics, calculate throughput, and grade response times against industry benchmarks.

Total Response Time

200ms

Good

TTFB

80ms

Transfer

120ms

Throughput

0.24 MB/s

Req/sec

5.0

Response Time Benchmarks

< 100msExcellent - users perceive as instant
100-300msGood - acceptable for most APIs
300ms-1sNoticeable - optimisation recommended
> 1sPoor - users will abandon
What is the difference between latency and throughput?

Latency is the time for a single request to complete — from the client sending the request to receiving the full response. Throughput is the number of requests completed per unit time. A system can have high throughput but high latency (many requests processed per second, but each individual request takes a long time) or low latency but low throughput (each request is fast, but the system can only handle a few at once). For user-facing APIs, latency is the primary concern — a 2-second response feels slow regardless of how many other users are being served simultaneously.

What are realistic API latency targets for different scenarios?

P50/P99 latency targets: User-facing read APIs (product listings, search): P50 under 100ms, P99 under 500ms. Write APIs (order placement, form submission): P50 under 200ms, P99 under 1000ms. Internal microservice calls: P50 under 20ms, P99 under 100ms. Database queries (simple): under 10ms. Database queries (complex joins): under 50ms. Third-party API calls: P50 under 500ms, P99 under 2000ms (with retry logic). Real-time features (WebSocket messages): under 100ms round-trip for good UX. P99 (the 99th percentile — 1% of requests are slower) is more important than average/P50 for user experience because slow outliers disproportionately affect perceived performance.

What causes high API latency?

Common culprits: N+1 query problem (making one database query per item in a loop instead of one batch query). Missing database indexes (table scans instead of index lookups). Synchronous blocking calls (waiting for external services in series rather than parallel). Large response payloads (serialising unnecessary data). Network distance (geographic distance between client and server adds ~1ms per 100km). Cold starts (serverless functions that have not run recently). Garbage collection pauses (common in Java/JVM languages). The first two (N+1 queries and missing indexes) account for the majority of unexpected API latency in typical web applications.

What is the 95th vs 99th percentile and why does it matter?

Percentile metrics measure the latency that X% of requests complete within. P95 = 95% of requests complete within this time. P99 = 99% complete within this time. The gap between P50 and P99 reveals tail latency — how badly the slowest requests perform. A P50 of 50ms and P99 of 2000ms means 1% of users wait over 2 seconds. At 1000 requests per second, that is 10 users per second experiencing 2s latency. High tail latency is often caused by garbage collection, lock contention, or resource exhaustion that affects only some requests. Optimising P99 is harder but more impactful for user experience than optimising the average.

How do I set a realistic SLA (Service Level Agreement) for an API?

Measure current performance first — run a load test to get real P50, P95, P99 numbers at your expected traffic volume. Then set targets 20-30% better than current P99 to give yourself a buffer. Common SLA structure: Availability (uptime): 99.9% (8.7 hours downtime/year) for standard services, 99.99% (52 minutes/year) for critical services. Latency: P95 under 500ms, P99 under 1000ms for user-facing APIs. Error rate: under 0.1% for 5xx errors. Define these with your team before launch — retrofitting SLA compliance into an existing system is significantly harder.

Is my data sent to a server?

No — all calculations happen in your browser. Nothing is transmitted.

What other API and network tools are on this site?

The HTTP Status Codes reference explains the status codes your API returns. The Curl Builder generates test commands for your endpoints. The JSON Formatter formats API response payloads for inspection. The Bandwidth Calculator estimates data transfer requirements. All are in the Dev Tools section.

Complete Guide

📊 Key Data Points

<200ms TTFB

Target for a well-optimized API endpoint — Google recommends under 200ms for server response

Performance API

Browser-native timing API — same data shown in DevTools Network tab, programmatically

TTFB = server + RTT

Time to first byte includes server processing time plus one network round-trip

API Response Time Tester -- Complete USA Guide 2026

API latency has multiple components — DNS resolution, TLS handshake, connection setup, server processing, and download. Breaking down the total response time by phase tells you where to optimize rather than guessing.

This tool measures API response time from your browser, broken down by phase. Runs in your browser — no proxy, no server.

**Long-tail searches answered here:** measure API response time broken down by phase online, test endpoint latency from browser free, how to check if API is slow DNS or TTFB.

For building test requests, pair with curl Builder.

🔬 How This Calculator Works

Uses the browser Performance API (PerformanceResourceTiming) to measure timing breakdowns for requests made from your browser. This gives you: DNS lookup time, TCP connection time, TLS handshake time, time-to-first-byte (TTFB), and content download time.

TTFB is the most important metric: it is the time from request sent to first byte received — representing server processing time plus network round-trip. High TTFB means the server is slow to process; high download time with low TTFB means the response is large.

✅ What You Can Calculate

Phase breakdown

Separates DNS lookup, TCP connection, TLS handshake, TTFB, and download time. Tells you exactly where latency is coming from.

Time to first byte (TTFB)

TTFB is the key server performance metric. Under 200ms is good; over 500ms indicates server-side processing issues. This tool surfaces TTFB without needing DevTools.

Multiple run averaging

Run the same request 5-10 times and get average latency. Single measurements are noisy — averages reveal the real baseline performance.

Header inspection

Shows response headers alongside timing data — useful for verifying that caching headers are correctly set for your API endpoints.

🎯 Real Scenarios & Use Cases

Pre-launch performance baseline

Before launching a feature, measure your critical API endpoint response times. Document the baseline so you can detect regressions after future deployments.

CDN configuration verification

After adding a CDN, compare TTFB before and after. A correctly configured CDN should reduce TTFB for cached responses from 200ms to under 20ms.

Third-party API SLA verification

Your payment or auth provider promises under 200ms p95 response time. Test from here to get an independent measurement from your deployment region.

Database vs application bottleneck

High TTFB (>500ms) with fast download indicates server processing time is the problem. Use this data to justify profiling your database queries or application logic.

💡 Pro Tips for Accurate Results

Run 5+ times for reliable data. A single measurement is affected by local DNS cache state, network jitter, and server-side variability. Run 5-10 times and use the median.

High TTFB = server problem. If TTFB is consistently over 500ms, the issue is server-side: database queries, application processing, or server resources.

DNS cache warms on subsequent runs. The first request often shows higher DNS time than subsequent ones. Run once to warm the cache, then measure.

CORS and timing info. Browsers hide timing details for cross-origin requests unless the server sends Timing-Allow-Origin: * header.

🔗 Use These Together

🏁 Bottom Line

API latency has multiple components and fixing the wrong one wastes engineering time. This tool separates DNS, connection, TLS, TTFB, and download time so you can optimize the actual bottleneck.

For complete API debugging: measure response times here, build test requests with curl Builder, and analyze response headers with HTTP Headers Analyzer.