Will jose luis ricon's brain still calculate 10^24 or more FLOPS in 2040?
2
Ṁ72040
47%
chance
1D
1W
1M
ALL
Get Ṁ1,000 play money
Related questions
Related questions
At least one of the most powerful neural nets at end of 2026 will be trained using 10^26 FLOPs
98% chance
At least one of the most powerful neural nets at end of 2030 will be trained using 10^26 FLOPs
98% chance
Will Jose Luis Ricon run more than 864 quintillion floating point ops on a NVIDIA h100/b200 (or better) by EOY2026
45% chance
In 2020, Joe Carlsmith estimated that 10^15 FLOPS is "enough" to perform human functionality. The next post estimating this number will believe it is >=10^15
32% chance
Will an AI model use more than 1e28 FLOPS in training before 2026?
10% chance
Will there be an announcement of a model with a training compute of over 1e30 FLOPs by the end of 2025?
5% chance
Will any supercomputer reach 1 zettaFLOP before 2035?
82% chance
Will a machine learning training run exceed 10^26 FLOP in China before 2026?
52% chance
At least one of the most powerful neural nets at end of 2030 will be trained using 10^27 FLOPs
93% chance
Will a machine learning training run exceed 10^27 FLOP in China before 2030?
77% chance