Microsoft is Reportedly Working on AI Chips to Train LLMs

Microsoft is Reportedly Working on AI Chips to Train LLMs
Published on

Microsoft is reportedly working on its own AI chips that can be used to train LLMs

After investing billions in the firm that created ChatGPT, OpenAI, Microsoft is reportedly working on its own AI chips that can be used to train large language models or LLMs. Microsoft is making waves in the artificial intelligence (AI) space.

According to The Information, which cited two sources with firsthand knowledge of the project, the software behemoth has been working on the microprocessor since as early as 2019. A limited number of Microsoft and OpenAI personnel who are evaluating the technology reportedly already have access to it under the code name Athena.

Microsoft hopes that the chip would outperform those it now purchases from other suppliers, saving it time and money on its expensive AI projects. According to the study, other well-known tech firms like Amazon, Google, and Facebook also produce their own AI processors.

For the AI research firm OpenAI, Microsoft has already constructed a supercomputer that can train vast numbers of models. To support ChatGPT and the Bing AI chatbot, the business uses tens of thousands of Nvidia A100 graphics chips connected to a network for the supercomputer. A "massive, cutting-edge supercomputer" was going to be built, thus it invested US$1 billion in OpenAI in 2019.

Microsoft created this supercomputer to have the processing capacity necessary to train and retrain an expanding number of AI models over extended periods using massive amounts of data.

Nidhi Chappell, Microsoft's head of product for Azure high-performance computing and AI, stated, "One of the things we had learned from research is that the larger the model, the more data you have, and the longer you can train, the greater the accuracy of the model is.

Google's TPU AI Chip

Google revealed that it has created an AI chip dubbed the Tensor Processing Unit (TPU) that was made exclusively for machine learning activities last year. According to claims, the TPU uses little power and can do billions of processes per second.

TensorFlow, Google's open-source machine learning software framework, is intended to be used with the tensor processing unit.

Join our WhatsApp Channel to get the latest news, exclusives and videos on WhatsApp

                                                                                                       _____________                                             

Disclaimer: Analytics Insight does not provide financial advice or guidance. Also note that the cryptocurrencies mentioned/listed on the website could potentially be scams, i.e. designed to induce you to invest financial resources that may be lost forever and not be recoverable once investments are made. You are responsible for conducting your own research (DYOR) before making any investments. Read more here.

Related Stories

No stories found.
logo
Analytics Insight
www.analyticsinsight.net