AI & ML


GPU demand set to increase dramatically with AI arms race

29 March 2023 AI & ML

The recent arrival of ChatGPT has generated a lot of buzz across the industry sectors related to cloud computing and artificial intelligence (AI). Tech giants such as Microsoft, Google, and Baidu have all built products and services derived from ‘generative AI’ technologies. This new wave of interest is expected to bring benefits to the participants across the supply chain for GPUs and AI chips. These participants include NVIDIA, TSMC, Unimicron, and AIChip.

However, there are challenges pertaining to the adoption and function-related optimisation of products and services that are powered by generative AI. Furthermore, user experience is at the core of an AI-based technology and involves the protection of personal information and the accuracy of the responses to content requests. Therefore, regulatory issues will likely emerge as generative AI moves to the next phase of its development.

TrendForce says generative AI represents an integration of different types of algorithms, pre-trained models, and multimodal machine learning. Notable ones include Generative Adversarial Network (GAN), Contrast Language-Image Pre-Training (CLIP), Transformer, and Diffusion. Generative AI searches for patterns in the existing data or batches of information and efficiently outputs content that is suitable for scenarios such as data collection and analysis, social interactions, copywriting, etc. There are already many apps powered by generative AI in the market right now, and the most common kinds of output from them include texts, images, music, and software codes.

Data, computing power, and algorithms are the three indispensable factors that drive the development of generative AI. Also, while AI-based products and services are relatively easy to build, optimising them is much more difficult. In this respect, the major cloud companies are in a more advantageous position since they possess huge amounts of the essential resources.

From the perspective of the developers of these products, the existing chat robots such as ChatGPT are able to not only converse with users in natural language but also somewhat meet the demand for ‘comprehending’ users’ input. Hence, having a better capability to understand what users need or desire can, in turn, provide further suggestions to users’ enquiries and responses. And since using an internet search engine is pretty much a habit for the majority of people worldwide, the most urgent task of the major cloud companies is to keep optimising their own search engines.

TrendForce’s latest investigation finds that Google remains the absolute leader in the global market for internet search engines, with a market share of more than 90%. Microsoft, with its Bing, now has a market share of just 3% and will unlikely pose a significant threat in the short term. However, Bing is gaining more users that can contribute to its data feedback and model optimisation cycle. Therefore, Google needs to be on guard against the chance of Microsoft creating differentiation in search-related services and perhaps capture certain kinds of opportunities in the area of online advertising.

Generative AI requires a huge amount of data for training, so deploying a large number of high-performance GPUs helps shorten the training time. In the case of the Generative Pre-Trained Transformer (GPT) that underlays ChatGPT, the number of training parameters used in the development of this autoregressive language model rose from around 120 million in 2018 to almost 180 billion in 2020. According to TrendForce’s estimation, the number of GPUs that the GPT model needed to process training data in 2020 came to around 20 000. Going forward, the number of GPUs that will be needed for the commercialisation of the GPT model (or ChatGPT) is projected to reach above 30 000.

Hence, with generative AI becoming a trend, demand is expected to rise significantly for GPUs, and thereby benefit the participants in the related supply chain. NVIDIA, for instance, will probably gain the most from the development of generative AI. Its DGX A100, which is a universal system for AI-related workloads, delivers 5 petaFLOPS and has nearly become the top choice for big data analysis and AI acceleration.

For more information https://insider.trendforce.com




Share this article:
Share via emailShare via LinkedInPrint this page

Further reading:

The dream of Edge AI
Altron Arrow Editor's Choice AI & ML
AI technology carries a great promise – the idea that machines can make decisions based on the world around them, processing information like a human might. But the promise of AI is currently only being fulfilled by big machines.

Read more...
MAX78000 neural network accelerator chip
Altron Arrow AI & ML
The hardware-based convolutional neural network accelerator enables even battery-powered applications to execute AI inferences.

Read more...
Nanomaterials to build next-gen AI hardware?
AI & ML
From improving scientific analyses and imaging capabilities, to predictive maintenance and monitoring operations in industrial settings, artificial intelligence is becoming ever more present in modern-day society.

Read more...
Microchip launches MPLAB ML development suite
AI & ML
Microchip’s unique solution is first to support 8-, 16- and 32-bit MCUs and 32-MPUs for machine learning at the edge.

Read more...
ToF sensor enables AI applications
Altron Arrow AI & ML
The VL53L7CH from STMicroelectronics is the perfect Time-of-Flight sensor enabling AI applications, with ultrawide 90° diagonal FoV and low power consumption.

Read more...
Analogue compute platform to accelerate Edge AI
Altron Arrow Editor's Choice AI & ML
Microchip has teamed up with Intelligent Hardware Korea to develop an analogue compute platform to accelerate Edge AI/ML inferencing using Microchip’s memBrain non-volatile in-memory compute technology.

Read more...
World’s most powerful open LLM
AI & ML
With a staggering 180 billion parameters, and trained on 3,5 trillion tokens, Falcon 180B has soared to the top of the Hugging Face Leaderboard for pretrained LLMs.

Read more...
Advancing quality control
Avnet Silica AI & ML
As manufacturing processes continue to become more sophisticated, the importance and effectiveness of advanced DVI solutions escalate, presenting opportunities for improved quality control.

Read more...
Give your edge AI model a performance boost
AI & ML
Join this webinar from STMicroelectronics to learn how to create an edge AI application easily on an STM32 MCU using the NVIDIA TAO toolkit.

Read more...
Three reasons why AI, ML add value for SMMEs only if the basics are in place
AI & ML
There is much chatter around artificial intelligence (AI) and the subfield of machine learning (ML), which can be confusing for SMME owners who may believe that they need to climb on the bandwagon.

Read more...