Dune 2
“He who controls the spice, controls the universe”
That quote from the famous book (and movies) by Frank Herbert told of the rare, but extremely valuable spice that was only found on the planet Arrakis. This spice was the cause of wars, base of currency and separator of classes. Those who had it, had all the power. Those who didn’t, well…
Wikipedia: In Dune, Arrakis is the most important planet in the universe, as it is the only source of the drug melange. Melange (or, "the spice") is the most essential and valuable commodity in the universe, as it extends life and makes safe interstellar travel possible (among other uses).
Replace spice with anything rare and critical and the rest of the quote applies. In today’s Gen AI world, the spice is GPUs.
GPUs (Graphical Processing Unit) - the cousin of the CPU, but designed to scale at floating point computation and vector calculations was primarily hardware for graphics intensive applications (game play, game design, video and graphics, etc.). But the advances in AI and Gen AI require significant data and compute. Data storage is relatively cheap and commoditized, but specialized compute is not.
There are only a handful of companies that make GPUs - NVIDIA being the most popular. Apple’s proprietary silicon Metal, Google’s TPUs and AMD’s Radeon are also strong competitors. NVIDIA is now a trillion dollar valued company (almost overnight) thanks to the sudden rise in demand from Gen AI!
Large Language Models (LLMs), image generation, audio/video generation all require heavy compute workloads and hardware and software engineering to take advantage of the compute. Massive data centers and deep wallets are needed to build and train these models.
And so, only major tech companies can afford to buy the quantities and configuration of GPUs and see the competitive value in doing so. Therefore, almost all advances in Gen AI will be dictated and controlled by these companies (Amazon, Google, Meta, Apple, Tesla, etc.). The conclusion is - making better models means having more even GPUs to train on even larger datasets to build humongous neural networks.
<aside> 🚘 Tesla just received an order of 10,000 NVIDIA H100 GPUs to power their $300MM supercomputer (Dojo). This is an investment in AI that will pay off for Tesla as AI is their competitive advantage to enable driverless cars.
</aside>
Eventually, software and hardware optimization will catch up as it always has to enable these complex workloads on traditional consumer-grade hardware (CPU or GPU). You can already run quantized versions of LLMs and Stable Diffusion models on a MacBook Pro (ex: llama.cpp). And there are prototypes running on an iPhone even! The open source models are enabling this innovation while also leveling the playing field. But even the open source models are developed by well financed and resourced companies / entities (ex: LLaMA is from Meta, Falcon from the UAE).
You too can be a spice trader!
Here are some options for how to participate in the revolution without a $100 million investment fund.
Mileage will vary based on use case and architecture, but these are great starter options.
Because GPUs are highly specialized, they are not needed for every type of compute or all the time. That’s important when you’re thinking of how to manage this limited and expensive resource.