Home Artificial Intelligence Former Tesla AI Director reproduces GPT-2 in 24 hours for only $672 — GPT-4 costs $100 million to train

Former Tesla AI Director reproduces GPT-2 in 24 hours for only $672 — GPT-4 costs $100 million to train

OpenAI launched GPT-2 in 2019, reportedly costing $256 per hour to train. However, it’s been five years since then, and we’re already at GPT-4o. Advancements in hardware, software, and data mean that training the same model will take less time and less money, as Andrej Karpathy, the developer behind the project to reproduce GPT-2 in llm.c, has proven.

The primary driver of cost savings is using a single 8XH100 node to do the training, which dropped the cost to just $28 an hour — almost 90% off in as little as five years. Nvidia launched the H100 in 2023, so OpenAI likely used hardware with much less power when it started working on GPT-2. However, the number of hours it took to train for GPT -2 is unknown. In comparison, the cost of training GPT-4 was more than $100 million.

 

Reference

Denial of responsibility! TechCodex is an automatic aggregator of Global media. In each content, the hyperlink to the primary source is specified. All trademarks belong to their rightful owners, and all materials to their authors. For any complaint, please reach us at – [email protected]. We will take necessary action within 24 hours.
DMCA compliant image

Leave a Comment