Friday, October 6th 2023
OpenAI Could Make Custom Chips to Power Next-Generation AI Models
OpenAI, the company behind ChatGPT and the GPT-4 large language model, is reportedly exploring the possibility of creating custom silicon to power its next-generation AI models. According to Reuters, Insider sources have even alluded to the firm evaluating potential acquisitions of chip design firms. While a final decision is yet to be cemented, conversations from as early as last year highlighted OpenAI's struggle with the growing scarcity and escalating costs of AI chips, with NVIDIA being its primary supplier. The CEO of OpenAI, Sam Altman, has been rather vocal about the shortage of GPUs, a sector predominantly monopolized by NVIDIA, which holds control over an astounding 80% of the global market for AI-optimized chips.
Back in 2020, OpenAI banked on a colossal supercomputer crafted by Microsoft, a significant investor in OpenAI, which harnesses the power of 10,000 NVIDIA GPUs. This setup is instrumental in driving the operations of ChatGPT, which, as per Bernstein's analyst Stacy Rasgon, comes with its own hefty price tag. Each interaction with ChatGPT is estimated to cost around 4 cents. Drawing a comparative scale with Google search, if ChatGPT queries ever burgeoned to a mere tenth of Google's search volume, the initial GPU investment would skyrocket to an overwhelming $48.1 billion, with a recurring annual expenditure of approximately $16 billion for sustained operations. For an invitation to comment, OpenAI declined to provide any statements. The potential entry into the world of custom silicon signals a strategic move towards greater self-reliance and cost optimization so further development of AI can be sustained.
Source:
Reuters
Back in 2020, OpenAI banked on a colossal supercomputer crafted by Microsoft, a significant investor in OpenAI, which harnesses the power of 10,000 NVIDIA GPUs. This setup is instrumental in driving the operations of ChatGPT, which, as per Bernstein's analyst Stacy Rasgon, comes with its own hefty price tag. Each interaction with ChatGPT is estimated to cost around 4 cents. Drawing a comparative scale with Google search, if ChatGPT queries ever burgeoned to a mere tenth of Google's search volume, the initial GPU investment would skyrocket to an overwhelming $48.1 billion, with a recurring annual expenditure of approximately $16 billion for sustained operations. For an invitation to comment, OpenAI declined to provide any statements. The potential entry into the world of custom silicon signals a strategic move towards greater self-reliance and cost optimization so further development of AI can be sustained.
41 Comments on OpenAI Could Make Custom Chips to Power Next-Generation AI Models
Price your products into the stratosphere nGreedia, and even the corps won't stick with you.
It was doing 10million queries per day
So 3/300= 1cent that is still very expensive for a query but not 4
That hardware is bought and used over a period of years it's not that part of it is used up every query and that it needs to be replaced after a billion.
That cluster will be happy running all sorts of stuff a couple of years on.
Also power, yeah microsoft has a power bill in the billions but they are doing a shit ton more than just doing 10m queries per day.
Conclusion someone was adding a heck of a lot of fudge to the numbers.
Personally I prefer popcorn. It's cheaper, takes longer to eat, and fits better when you are watching movies.
All of the costs are distributed in much larger quantities, and there is no way that rubbish 150 g chips pack would cost more than 2£.
Electricity cost is towards 0, salt is towards 0.
1.17£ is 1 kg potatos. Then, 150 g is only 0.17£ raw material.
The other problem is that, because nobody outside Tesla knows how Dojo works, anyone they recruit to work with it is going to need to be trained on its intricacies - which increases the cost of hiring and time for a new dev to ramp up. It also decreases the pool of talent willing to work for you, because Dojo isn't a transferable skill, so you are basically shackling yourself to Tesla if you choose to work there - and people experienced in the software development industry are simply too smart to do that. Which kinda sucks, since those are the people you most need.
Of course, maybe Tesla spent part of those 5 years perfecting a CUDA-to-Dojo transpiler and none of the above is relevant. But my experience with big companies is that they love building walled gardens - if not to trap users, then to trap developers. This is why open and shared standards are important.