AI Model Comparison
42 models compared side-by-side. LLMs, image generators, code models, embeddings, speech, and video — all in one place.
Last updated: April 19, 2026
Best for Coding
Claude 3.5 Sonnet
92% HumanEval, $3/$15 per 1M tokens
Best for Writing
Claude 3 Opus
Nuanced, long-form, deeply reasoned output
Best Open-Source
Llama 3.1 405B
Matches GPT-4o, fully open weights
Best Value
DeepSeek V3
GPT-4o quality at $0.27/$1.10 per 1M tokens
Sort by:
| Model | Provider | Params | Context | Input | Output | Type | Released |
|---|---|---|---|---|---|---|---|
|
Open
Closed
Multimodal
Key Strengths Description Benchmarks
MMLU
HumanEval
MATH
No benchmark data available for this model type. |
|||||||
Params:
Context:
In:
Open
Closed
Multimodal
Key Strengths
Description
Benchmarks
MMLU
HumanEval
MATH
Released:
Output:
No models match your search.
models shown
17 open-weight
17 multimodal
23 providers