Compare Free LLM APIs
Select up to 6 models to compare capabilities, limits, and performance side by side.
0/6 selected
Pick models from the gallery below
Side-by-Side Comparison
Select 2-4 models from the table above to compare them side by side.
How to Compare Free LLM APIs — What Matters
Not all specs matter equally. Here's what to focus on when comparing free models:
- Context window — Most important for coding tools (Claude Code, Cursor). 128K is the practical minimum for processing a real codebase. For long documents or multi-file refactors, look for 256K+.
- Rate limit (RPM) — 30+ RPM is comfortable for solo developers (one request every 2 seconds). Below 10 RPM, you'll feel the throttle. Check both RPM and daily caps (RPD).
- OpenAI compatibility — If the model's API is OpenAI-compatible, it works with virtually every AI tool (Cursor, Codex, OpenCode, LangChain). Non-compatible APIs need custom integration.
- Tool calling support — Essential for AI coding agents. If the model can't call tools, it can generate code but can't run commands or search your codebase.
- Credit card requirement — Some "free" tiers still require card verification. If you want zero friction, filter for "No Card" models on the models page.
- Max output tokens — For coding, 8K output is the minimum (one full source file). 16K+ lets the model generate entire modules. Reasoning models need extra output for chain-of-thought.
After comparing specs, test your top picks in the Playground with a real API key — latency and output quality can't be judged from numbers alone.