Large Language Models
docAnalyzer.ai keeps pace with releases from Anthropic, DeepSeek, Google, Meta, Moonshot AI, OpenAI, Zhipu AI, and xAI, adding new models within days so you can pick the right fit for every workflow instead of being locked to a single AI source. We currently surface 22 models across 8 creators, each benchmarked for quality and speed so you can choose confidently. Explore the latest line-up below.
| Model | Quality | Speed | Latency* |
|---|---|---|---|
Gemini 3 Pro Preview Google | 73 | 4.5 token/s | 110 seconds |
Claude Opus 4.5 Unknown | 70 | 10 token/s | 57 seconds |
GPT-5 OpenAI | 68 | 3.44 token/s | 99 seconds |
Grok 4 xAI | 65 | 6 token/s | 35 seconds |
GPT-5 mini OpenAI | 64 | 0.69 token/s | 73 seconds |
Claude Sonnet 4.5 Anthropic | 63 | 6 token/s | 69 seconds |
GPT OSS 120b OpenAI | 61 | 0.26 token/s | 340 seconds |
Gemini 2.5 Pro Google | 60 | 3.44 token/s | 41 seconds |
Grok 4 Fast Reasoning xAI | 60 | 0.28 token/s | 224 seconds |
DeepSeek V3.2 (Thinking) DeepSeek | 57 | 0.32 token/s | 29 seconds |
Z.ai GLM 4.6 Zhipu AI | 56 | 1 token/s | 99 seconds |
Claude Haiku 4.5 Anthropic | 55 | 2 token/s | 84 seconds |
Gemini 2.5 Flash Google | 54 | 0.85 token/s | 135 seconds |
GPT OSS 20b OpenAI | 52 | 0.1 token/s | 237 seconds |
GPT-5 nano OpenAI | 51 | 0.14 token/s | 135 seconds |
Kimi K2 0905 Moonshot AI | 50 | 1.2 token/s | 92 seconds |
Gemini 2.5 Flash-Lite Google | 48 | 0.17 token/s | 720 seconds |
DeepSeek V3.2 (Non thinking) DeepSeek | 46 | 0.32 token/s | 28 seconds |
GPT-5.1 OpenAI | 43 | 3.44 token/s | 88 seconds |
Grok 4 Fast xAI | 39 | 0.28 token/s | 184 seconds |
Llama 4 Maverick Meta | 36 | 0.42 token/s | 131 seconds |
Llama 4 Scout Meta | 28 | 0.24 token/s | 128 seconds |
* Latency progress bars use a logarithmic scale, where lower latency values indicate better performance.