Anthropic: Claude Opus 4 vs Google: Gemini 3.1 Pro Preview: Which AI Model Is Better?

Updated March 24, 2026· Based on independent benchmark data

Quick Verdict

Google: Gemini 3.1 Pro Preview leads in intelligence with a score of 57.2 vs 46.5. Google: Gemini 3.1 Pro Preview is 7.5x cheaper at $2.00/1M tokens vs $15/1M. For speed, Google: Gemini 3.1 Pro Preview wins at 117 tok/s vs 44 tok/s.

Head-to-Head Comparison

MetricAnthropic: Claude Opus 4Google: Gemini 3.1 Pro Preview
Intelligence Score46.557.2
Coding Score47.655.5
Math ScoreN/AN/A
Speed (tok/s)44 tok/s117 tok/s
Latency (TTFT)1.94s21.91s
Input Price / 1M tokens$15$2.00
Output Price / 1M tokens$75$12
Context Window200K1.0M
Max Output Tokens32K66K
Input ModalitiesImage + Text + FileAudio + File + Image + Text + Video
Output ModalitiesTextText
Free TierNoNo

Detailed Analysis

Intelligence & Quality

Google: Gemini 3.1 Pro Preview outperforms Anthropic: Claude Opus 4 on the Artificial Analysis intelligence index with a score of 57.2 compared to 46.5. For coding tasks, Google: Gemini 3.1 Pro Preview has the edge with a coding score of 55.5 vs 47.6.

Speed & Latency

Google: Gemini 3.1 Pro Preview generates output significantly faster at 117 tok/s compared to Anthropic: Claude Opus 4's 44 tok/s, making it 2.6x faster for streaming responses. Time to first token is 1.94s for Anthropic: Claude Opus 4 vs 21.91s for Google: Gemini 3.1 Pro Preview, which affects perceived responsiveness in interactive applications.

Pricing

Google: Gemini 3.1 Pro Preview is more affordable at $2.00/1M input tokens ($12/1M output), while Anthropic: Claude Opus 4 costs $15/1M input ($75/1M output). That makes Anthropic: Claude Opus 4 7.5x more expensive per token, which can add up significantly at scale. For a typical workload of 100 requests per day at 2,000 tokens each, Anthropic: Claude Opus 4 would cost approximately $90.00/month vs $12.00/month for Google: Gemini 3.1 Pro Preview in input costs alone.

Context Window

Google: Gemini 3.1 Pro Preview offers a larger context window at 1.0M tokens compared to Anthropic: Claude Opus 4's 200K. This means Google: Gemini 3.1 Pro Preview can process roughly 524 pages of text in a single request vs 100 pages for Anthropic: Claude Opus 4. For output length, Google: Gemini 3.1 Pro Preview can generate up to 66K tokens per response vs 32K for Anthropic: Claude Opus 4.

Best Use Cases

Choose Google: Gemini 3.1 Pro Preview when you need higher intelligence (57.2), stronger coding performance (55.5), faster output (117 tok/s), lower cost, larger context window (1.0M).

Choose Anthropic: Claude Opus 4 if:

  • You want lower latency (1.94s vs 21.91s TTFT)

Choose Google: Gemini 3.1 Pro Preview if:

  • You need higher intelligence (score: 57.2 vs 46.5)
  • You prioritize coding performance (score: 55.5 vs 47.6)
  • You need faster throughput (117 tok/s vs 44 tok/s)
  • Budget is a concern ($2.00/1M vs $15/1M)
  • You need a larger context window (1.0M vs 200K)

Frequently Asked Questions

Is Anthropic: Claude Opus 4 better than Google: Gemini 3.1 Pro Preview for coding?

Google: Gemini 3.1 Pro Preview scores higher on coding benchmarks (55.5 vs 47.6), making it the better choice for programming tasks.

Which is cheaper, Anthropic: Claude Opus 4 or Google: Gemini 3.1 Pro Preview?

Google: Gemini 3.1 Pro Preview is cheaper at $2.00/1M input tokens vs $15/1M for Anthropic: Claude Opus 4.

Is Anthropic: Claude Opus 4 faster than Google: Gemini 3.1 Pro Preview?

Google: Gemini 3.1 Pro Preview is faster, producing output at 117 tok/s compared to Anthropic: Claude Opus 4's 44 tok/s.

Can Anthropic: Claude Opus 4 process images?

Yes, Anthropic: Claude Opus 4 supports image input. Google: Gemini 3.1 Pro Preview also supports images.

Which has a larger context window, Anthropic: Claude Opus 4 or Google: Gemini 3.1 Pro Preview?

Google: Gemini 3.1 Pro Preview has a larger context window at 1.0M compared to Anthropic: Claude Opus 4's 200K.

Should I use Anthropic: Claude Opus 4 or Google: Gemini 3.1 Pro Preview?

It depends on your priorities. Google: Gemini 3.1 Pro Preview scores higher on intelligence (57.2), but Anthropic: Claude Opus 4 may be better for specific use cases like budget-conscious projects or speed-critical applications.

Related Comparisons

Benchmark data by Artificial Analysis

Data last synced: March 24, 2026