Groq is an ultra-fast AI inference platform that leverages custom-designed LPU (Language Processing Unit) hardware to deliver unprecedented inference speeds for open-source LLMs. The platform provides free access to popular models like Llama 2, Mixtral, and Gemma through an OpenAI-compatible API, making it easy for developers to integrate blazing-fast AI capabilities into their applications. Groq's custom hardware enables token generation speeds up to 10x faster than traditional GPUs, with a generous free tier and competitive pay-per-use pricing for production workloads requiring maximum performance.
Review freshness
Mar 13, 2026
Last checked Feb 2, 2026
Next review Apr 3, 2026 ยท overdue
Confidence
65%
Completeness: partial
Update method
official baseline
Official-source-first profile maintenance
Cadence every 21 days
Sources
5
Baseline: complete
Live verification
Partially verified
Verified required types: 2
Last check Apr 15, 2026
ๅฎๆน็ฝ็ซไผๅ ็ปไฝ ็ป่ฎบ๏ผๅ็ปไฝ ๅฝไปค๏ผไธๅๆ provider ่ฏฆๆ ้กตๅฝๆๆง CLI ๆไปคๅ ใ
่ฟ่ฅๆจกๅผ
Cloud platform access
ๆฅๅ ฅๅปบ่ฎฎ
Recommended
ๅ ่ฎฉ CCJK ๆฃๆต็ฏๅข๏ผๅ็ปๅๆฌ้กต็่ฟ่ฅๆจกๅผๅๆฅๅ ฅๅปบ่ฎฎๅณๅฎๆฏๅฆๅฏ็จ่ฏฅ providerใ
npx ccjkCIใๆน้ๆบๅจๆๆ ๅๅ้จ็ฝฒๆถไฝฟ็จ้้ปๆจกๅผ๏ผไฟๆๅๅฎๆน README ไธ่ดใ
export ANTHROPIC_API_KEY="sk-ant-..." && npx ccjk init --silentๅฎๆๅๅงๅๅๅๅฏ็จ็ฏๅขไผๅไธๅผๅๆ้ preset๏ผไธๆ provider ๆฅๅ ฅๅๅบ็ก็ฏๅข่ฆๅๅจไธ่ตทใ
npx ccjk boost && ccjk zc --preset devๅ ่ฏปๅฎๆฌ้กต็ๅฎๆนๆฅๆบใ่ฟ่ฅๆจกๅผๅๆฅๅ ฅ็ป่ฎบ๏ผๅๅฏผๅบ้ ็ฝฎใๅฏน relayใ่ๅๅๆ่ฝฌๅฎ้้๏ผๅปบ่ฎฎไฝฟ็จ้็ฆป API keyใ็ฌ็ซ้ข็ฎๅๅ็ฌ็ฏๅขใ
ๅฏผๅบ็ฏๅขๅ้ๆจกๆฟ
curl -fsSL "https://www.claudehome.cn/api/providers/groq/config?format=env"ๅฏผๅบ curl ็คบไพ
curl -fsSL "https://www.claudehome.cn/api/providers/groq/config?format=curl"ๅฏผๅบ OpenAI ๅ ผๅฎน้ ็ฝฎ
curl -fsSL "https://www.claudehome.cn/api/providers/groq/config?format=newapi"Groq is an ultra-fast AI inference platform that leverages custom-designed LPU (Language Processing Unit) hardware to deliver unprecedented inference speeds for open-source LLMs. The platform provides free access to popular models like Llama 2, Mixtral, and Gemma through an OpenAI-compatible API, making it easy for developers to integrate blazing-fast AI capabilities into their applications. Groq's custom hardware enables token generation speeds up to 10x faster than traditional GPUs, with a generous free tier and competitive pay-per-use pricing for production workloads requiring maximum performance.
Classification
Cloud platform access
Groq operates its own managed inference platform and commercial layer rather than acting as a simple pass-through relay.
Procurement
Platform account with usage billing and enterprise sales options
Integration Pattern
GroqCloud managed inference API
Risk Signal
medium
Good production candidate when low-latency managed inference on GroqCloud matters more than direct control over every open model host. Live verification is currently partial because some required official source types are blocked from this environment: documentation, pricing.
openaiVerified required types
Blocked or broken types
Completeness
complete
Groq provides a complete public baseline across GroqCloud docs, pricing, support, and terms.
Covered types
Missing types