Skip to main contentGroq provides ultra-fast AI inference through custom LPU™ (Language Processing Unit) architecture. Hosts open-source models from OpenAI, Meta, DeepSeek, and others.
Website: https://groq.com/
Getting an API Key
- Go to Groq Console and sign in
- Navigate to API Keys section
- Create a new API key and name it (e.g., “CodinIT”)
- Copy the key immediately - you won’t see it again
Configuration
- Click the settings icon (⚙️) in CodinIT
- Select “Groq” as the API Provider
- Paste your API key
- Choose your model
Supported Models
llama-3.3-70b-versatile (Meta) - 131K context
openai/gpt-oss-120b (OpenAI) - 131K context
moonshotai/kimi-k2-instruct - 1T parameters with caching
deepseek-r1-distill-llama-70b - Reasoning optimized
qwen/qwen3-32b (Alibaba) - Q&A enhanced
meta-llama/llama-4-maverick-17b-128e-instruct
Key Features
- Ultra-fast inference: Sub-millisecond latency with LPU architecture
- Large context: Up to 131K tokens
- Prompt caching: Available on select models
- Vision support: Available on select models
Learn more about LPU architecture.
Notes
- Speed: Optimized for single-request latency
- Pricing: See Groq Pricing