Staggering Costs, Slow Progress of the GPT-5 Project
Analysts previously predicted that tech giants could invest $1 trillion in artificial intelligence projects in the coming years. There are also estimates showing that a 6-month training of GPT-5 would cost about $500 million just in computing costs. OpenAI CEO Sam Altman said that the cost of future AI models is expected to exceed $1 billion. However, people familiar with the project said:In October this year, OpenAI's $157 billion valuation was largely based on Altman's forecast, in which he previously said that GPT-5 would be a "major breakthrough" and that GPT-4 performed like a smart high school student, but the eventual GPT-5 would actually be more like a PhD in some tasks. The report said that GPT-5 should be able to unlock new scientific discoveries and complete everyday human tasks such as scheduling or flights. Researchers hope it will make fewer mistakes than existing AI, or at least acknowledge "doubts", as current models may hallucinate. However, there is no fixed standard for "when it will become smart enough AI", it is more a matter of feeling. So far, the GPT-5 under development still does not feel strong enough. Altman said in November that "no product called GPT-5 will be released by 2024"."Although Orion's performance is somewhat better than OpenAI's current products, it is not enough to justify the huge operating costs."
Data Shortage as the Main Bottleneck
To avoid wasting huge investments, researchers are trying to minimize the chances of failure through small-scale trials. However, the GPT-5 plan seems to have had problems from the start. In mid-2023, OpenAI began a training run, which was also a test of the proposed new design for Orion. But the process has been slow, indicating that larger-scale training may take a very long time, which in turn will make the cost extremely high. OpenAI's researchers decided to make some technical adjustments to enhance Orion, and they also found that to make Orion smarter, they need more high-quality, diverse data. Model testing is an ongoing process, and large-scale training runs may take months, with trillions of tokens being "fed" to the model. However, data such as news articles, social media posts, and scientific papers on the public internet is no longer sufficient. As Datology AI CEO Ari Morcos said:To solve this problem, OpenAI has chosen to create data from scratch. They hire software engineers, mathematicians and other professionals to write new code or solve math problems, which are then used as training data. The company also collaborates with experts in fields such as theoretical physics to explain how they would handle the most challenging problems in their field, but this process is very slow, and the training of GPT-4 used about 13 trillion tokens. Even with 1,000 people writing 5,000 words a day, it would only produce 1 billion tokens in a few months. OpenAI has also begun to develop "synthetic data", using AI-generated data to train Orion, and believes it can avoid failures by using data generated by its other AI model o1."It's becoming very expensive, and it's hard to find more data of the same high quality."