‘GPT-Next’ is coming, and it’s 100x better than GPT-4. Says CEO of OpenAI Japan

2 mins read

Category:

  • Optical Character Recognition (OCR)

Voluptates voluptatibus omnis qui porro sed

The CEO of OpenAI Japan has announced that the company plans to release “GPT-Next” later this year, boasting a computational load that is 100 times greater than GPT-4. The next-generation model, referred to as “GPT-4 NEXT,” is expected to be trained using a smaller version of OpenAI’s latest training infrastructure, “Strawberry,” but with computational resources equivalent to those used for GPT-4. This immense leap in effective computational load, which also benefits from architectural improvements and enhanced learning efficiency, is expected to significantly increase the model’s capabilities.

Unlike traditional software, AI evolves exponentially, and “GPT-Next” is expected to deliver performance growth nearly 100-fold based on its predecessors. The slide presented by the CEO emphasized that this increase in computational power does not merely reflect scaling in computing resources but rather the combination of improved architecture and enhanced training methods, pushing the boundaries of what AI models can achieve.

Additionally, OpenAI’s “Orion” project, which has gained significant attention, was trained over several months using the equivalent of 10,000 H100 GPUs. This represents a tenfold increase in computational resources compared to GPT-4, adding three orders of magnitude (OOMs) in terms of its performance. Orion is anticipated for release next year, further highlighting the rapid advancements in AI technology expected from OpenAI.


Jane Smith

Editor

Jane Smith has been the Editor-in-Chief at Urban Transport News for a decade, providing in-depth analysis and reporting on urban transportation systems and smart city initiatives. His work focuses on the intersection of technology and urban infrastructure.