Recent developments in the tech industry have unveiled noteworthy initiatives by six prominent technology companies: Microsoft, OpenAI, Tesla, Google, Amazon, and Meta. These companies are currently engaged in the research and development of their own AI chips, signaling potential competition with Nvidia's flagship microprocessor, the H100 GPU. These endeavors hold the promise of not only reducing costs but also diminishing reliance on Nvidia's AI chips.
Nvidia's dominance in the AI computing sphere, particularly with its A100/H100 series chips at the pinnacle, has been accompanied by considerable cost implications for end-users. According to analyses conducted by Bernstein, the cost of a single ChatGPT query stands at approximately 4 cents, with scaling up to one-tenth of Google's search volume necessitating an annual expenditure of approximately $48.1 billion on GPUs and approximately $16 billion on chips to maintain operations. These financial considerations have motivated tech industry leaders to embark on the journey of self-developed AI chips.
Microsoft is set to unveil its inaugural AI-designed chip at its forthcoming annual developer conference scheduled for the next month. Building on previous reports by The Information, Microsoft has been diligently working on a dedicated chip, codenamed "Athena," since 2019, designed to power large-scale language models. This chip, slated for production by TSMC using an advanced 5nm manufacturing process, is anticipated to launch as early as the next calendar year. Analysts speculate that the development of chips akin to Athena could entail an annual cost of approximately $100 million. If Athena can rival Nvidia's products, it holds the potential to reduce chip costs by as much as one-third.
OpenAI, too, is in the pursuit of manufacturing its own AI chips, with ongoing evaluations of potential acquisition targets. It has been reported that OpenAI has been exploring a variety of solutions to address the shortage of AI chips since at least the past year. Furthermore, OpenAI has fostered closer collaborations with chip manufacturers, including Nvidia, and is actively diversifying its supplier base beyond the confines of a single provider.
Tesla, a prominent electric vehicle manufacturer, has strategically ventured into the realm of intelligent driving and has already introduced two proprietary chips. One is dedicated to Full Self-Driving (FSD) capabilities within Tesla vehicles, while the other, known as Dojo D1, is employed in Tesla's supercomputing infrastructure to expedite the training and enhancement of the company's autonomous driving system.
Google embarked on a covert development endeavor in 2013, focusing on a chip designed to accommodate AI machine learning algorithms. This chip was intended to replace Nvidia's GPUs in Google's internal cloud computing data centers. The result was the "TPU," which remained undisclosed until its public debut in 2016. The TPU is adept at conducting large-scale matrix operations essential for deep learning models employed in natural language processing, computer vision, and recommendation systems. Although Google deployed the TPU v4 within its data centers in 2020, it only recently revealed technical specifics in April of this year.
Amazon, a trailblazer among cloud providers, has been at the forefront of self-developed chip technology since introducing its inaugural Nitro1 chip in 2013. Amazon Web Services (AWS) boasts an array of self-developed chip product lines, encompassing network chips, server chips, and AI machine learning chips. In early 2023, AWS unveiled Inferentia 2 (Inf2), a chip meticulously engineered for AI applications, boasting a threefold increase in computational prowess and a 25% increment in accelerator total memory. It facilitates distributed inference through direct, high-speed chip-to-chip connections and can accommodate up to a staggering 175 billion parameters, positioning it as a formidable contender for large-scale model inference.
Meta, formerly known as Facebook, relied primarily on CPUs and custom chip amalgamations tailored to expedite AI algorithms for its AI workloads until 2022. However, the efficiency of CPUs often lagged behind that of GPUs. In a strategic shift, Meta abandoned its plans for an extensive rollout of custom chips in 2022 and opted for Nvidia GPUs valued at several billion dollars. To reverse this trajectory, Meta has embarked on the development of in-house chips, unveiling an AI training and inference chip project on May 19th. This chip boasts a mere 25-watt power consumption, constituting a fraction of the power draw associated with chips from market-leading suppliers like Nvidia, and employs the open-source RISC-V architecture, heralded as the fifth generation of reduced instruction set processors.