This in contrast to the $191 million Google spent on Gemini Ultra sounds like a bargain! ๐ฐ
Gemini Ultra required 50 billion petaFLOPS (one petaFLOP equals one quadrillion FLOPs). ๐ค
Compared to OpenAIโs GPT-4, which required 21 billion petaFLOPS, at a cost of $78 million. ๐ก
2017: Original Transformer Model: $930 [@Google ] ๐ป
2018: BERT-Large: $3,288 [@Google] ๐
2019: RoBERTa Large: 160k [@Meta] ๐
2020: GPT-3(175B): $4.32M [@OpenAI] ๐ง
2023: Llama 2 70B: $3.93M [@Meta] ๐
2023: GPT-4: $78.35M [@OpenAI] ๐
Now, Gemini Ultra: $191.4M [@Google] ๐
This forms an exponential curve! ๐คฏ
But, why? ๐ค
Compute, data, and expertise. All three come at a great cost! โ๏ธ๐๐ก
Google recently made Gemini-1.5-Flash fine-tuning free, as it's almost impossible for regular businesses to justify an in-house trained foundational model! ๐
This barrier of cost is going to result in fewer new foundational models/less competition and more fine-tunes! ๐๐
Data [Stanford Universityโs 2024 AI Index Report]: https://aiindex.stanford.edu/report/
Graphic: https://voronoiapp.com/technology/Googles-Gemini-Ultra-Cost-191M-to-Develop--1088
Many thanks to everyone spending tons of resources and open-sourcing the models! ๐ค