Skip to main content
LM Studio provides a user-friendly way to run AI models locally with privacy, speed, and offline capabilities. Website: https://lmstudio.ai

Setup

  1. Download: Visit lmstudio.ai and download for your OS
  2. Install and launch: Open LM Studio
  3. Download a model: Go to “Discover” tab and download a model
    • Recommended: Qwen3 Coder 30B A3B Instruct for best CodinIT experience
  4. Start server: Go to “Developer” tab and toggle server to “Running” (runs at http://localhost:51732)
  5. Configure model settings:
    • Context Length: Set to 262,144 (maximum)
    • KV Cache Quantization: Leave unchecked (critical for performance)
    • Flash Attention: Enable if available

Configuration in CodinIT

  1. Click the settings icon (⚙️) in CodinIT
  2. Select “LM Studio” as the API Provider
  3. Set server URL to http://localhost:51732
  4. Choose your model

Quantization Guide

Choose based on available RAM:
  • 32GB RAM: 4-bit quantization (~17GB download)
  • 64GB RAM: 8-bit quantization (~32GB download)
  • 128GB+ RAM: Full precision or larger models

Model Format

  • Mac (Apple Silicon): Use MLX format
  • Windows/Linux: Use GGUF format

Features

  • Complete privacy: All data stays on your device
  • No API costs: Unlimited free usage
  • Offline operation: Works without internet
  • Hardware flexibility: Runs on any modern computer

System Requirements

Minimum:
  • 8GB RAM (16GB recommended)
  • 10GB free storage
  • Modern multi-core CPU
Recommended:
  • 32GB+ RAM for large models
  • NVIDIA GPU with 8GB+ VRAM (optional)
  • SSD with 50GB+ free space

Troubleshooting

If CodinIT can’t connect:
  1. Verify LM Studio server is running
  2. Ensure a model is loaded
  3. Check system meets hardware requirements
  4. Confirm server URL matches in CodinIT settings

Notes

  • Start LM Studio before using with CodinIT
  • Keep LM Studio running in background
  • First model download may take several minutes
  • Models are stored locally after download