
Is code-davinci-002 just the largest non-GPT-4 model in the GPT-4 scaling law experiment?
Mini
3
Ṁ35Feb 17
44%
chance
1D
1W
1M
ALL
Get Ṁ1,000 play money
Related questions
Related questions
Will GPT-5 have over 1 trillion parameters?
87% chance
OpenAI releases a model unquestionably named "GPT-4.5" or "GPT-4.5x" (where X varies) by mid 2025, hardcore legalism arc
85% chance
Will GPT-4 be trained (roughly) compute-optimally using the best-known scaling laws at the time?
30% chance
What will be true about GPT-5? (See description)
GPT-4 #5: Will GPT-4 be a dense model?
1% chance
What hardware will GPT-5 be trained on?
GPT-5 trained with >=24k GPUs?
82% chance
What will be true about GPT-5?
How many parameters does GPT4o have?
Will the performance jump from GPT4->GPT5 be less than the one from GPT3->GPT4?
72% chance