How many parameters does the new possibly-SOTA large language model, Claude 3 Opus, have?
โž•
Plus
23
แน€941
2026
1.2%
Less than 250 billion
1.3%
Between 250 and 500 billion
13%
Between 500 billion and 1 trillion
49%
Between 1 trillion and 1.5 trillion
26%
Between 1.5 trillion and 2 trillion
10%
More than 2 trillion

Anthropic released their new Claude 3 model today, which apparently outperforms GPT-4, the previous state of the art LLM, on various benchmarks: https://www.anthropic.com/news/claude-3-family

GPT-4 is rumored to have 1.8 trillion parameters. How many does Claude 3 have?

Resolves N/A if there is no information on this by 2026. There does not need to be official confirmation of parameter count, just media reports which don't contradict each other.

Get แน€1,000 play money
Sort by:

https://x.com/aidan_mclau/status/1849607168607002674

This random Twitter user claims to have heard that Claude Opus is smaller than GPT-4. Not sure how credible they are

Deleted comment