tickerr

Tickerr / Limits / Groq

Groq rate limits, context window & usage caps by plan

Rate limits, context window, message limits, file upload caps and image generation limits — all plans compared.

Compare free tiers across all AI tools →
Context window explained: Groq's context window is 128000 tokens — that's ~96K wordsof text the model can "see" at once, including your conversation history.
Rpm30 requests/minFree tier
Tpm6000 tokens/minLlama 3.3 70B
Rpd14400 requests/dayFree tier daily cap
Context Window128000 tokens(~96K words)Llama 3.3 70B
Concurrent Requests5 requestsSimultaneous
Rpm1000 requests/minPaid tier
Tpm500000 tokens/minHigher throughput
Context Window128000 tokens(~96K words)All supported models
Concurrent Requests50 requestsSimultaneous
Audio Hours Per Hour7200 seconds/hourWhisper transcription

About Groq limits

Groq enforces usage limits to manage server load and ensure fair access across all users. Limits vary significantly by plan tier — free plans are the most restricted while paid plans offer higher caps or unlimited access. Limits shown here are updated manually when Groq announces changes.

Check the Groq live status and outage history to see if current limits are being affected by an outage.