AI & ML


GPU demand set to increase dramatically with AI arms race

29 March 2023 AI & ML

The recent arrival of ChatGPT has generated a lot of buzz across the industry sectors related to cloud computing and artificial intelligence (AI). Tech giants such as Microsoft, Google, and Baidu have all built products and services derived from ‘generative AI’ technologies. This new wave of interest is expected to bring benefits to the participants across the supply chain for GPUs and AI chips. These participants include NVIDIA, TSMC, Unimicron, and AIChip.

However, there are challenges pertaining to the adoption and function-related optimisation of products and services that are powered by generative AI. Furthermore, user experience is at the core of an AI-based technology and involves the protection of personal information and the accuracy of the responses to content requests. Therefore, regulatory issues will likely emerge as generative AI moves to the next phase of its development.

TrendForce says generative AI represents an integration of different types of algorithms, pre-trained models, and multimodal machine learning. Notable ones include Generative Adversarial Network (GAN), Contrast Language-Image Pre-Training (CLIP), Transformer, and Diffusion. Generative AI searches for patterns in the existing data or batches of information and efficiently outputs content that is suitable for scenarios such as data collection and analysis, social interactions, copywriting, etc. There are already many apps powered by generative AI in the market right now, and the most common kinds of output from them include texts, images, music, and software codes.

Data, computing power, and algorithms are the three indispensable factors that drive the development of generative AI. Also, while AI-based products and services are relatively easy to build, optimising them is much more difficult. In this respect, the major cloud companies are in a more advantageous position since they possess huge amounts of the essential resources.

From the perspective of the developers of these products, the existing chat robots such as ChatGPT are able to not only converse with users in natural language but also somewhat meet the demand for ‘comprehending’ users’ input. Hence, having a better capability to understand what users need or desire can, in turn, provide further suggestions to users’ enquiries and responses. And since using an internet search engine is pretty much a habit for the majority of people worldwide, the most urgent task of the major cloud companies is to keep optimising their own search engines.

TrendForce’s latest investigation finds that Google remains the absolute leader in the global market for internet search engines, with a market share of more than 90%. Microsoft, with its Bing, now has a market share of just 3% and will unlikely pose a significant threat in the short term. However, Bing is gaining more users that can contribute to its data feedback and model optimisation cycle. Therefore, Google needs to be on guard against the chance of Microsoft creating differentiation in search-related services and perhaps capture certain kinds of opportunities in the area of online advertising.

Generative AI requires a huge amount of data for training, so deploying a large number of high-performance GPUs helps shorten the training time. In the case of the Generative Pre-Trained Transformer (GPT) that underlays ChatGPT, the number of training parameters used in the development of this autoregressive language model rose from around 120 million in 2018 to almost 180 billion in 2020. According to TrendForce’s estimation, the number of GPUs that the GPT model needed to process training data in 2020 came to around 20 000. Going forward, the number of GPUs that will be needed for the commercialisation of the GPT model (or ChatGPT) is projected to reach above 30 000.

Hence, with generative AI becoming a trend, demand is expected to rise significantly for GPUs, and thereby benefit the participants in the related supply chain. NVIDIA, for instance, will probably gain the most from the development of generative AI. Its DGX A100, which is a universal system for AI-related workloads, delivers 5 petaFLOPS and has nearly become the top choice for big data analysis and AI acceleration.




Share this article:
Share via emailShare via LinkedInPrint this page

Further reading:

World’s most powerful open LLM
AI & ML
With a staggering 180 billion parameters, and trained on 3,5 trillion tokens, Falcon 180B has soared to the top of the Hugging Face Leaderboard for pretrained LLMs.

Read more...
Bridging the gap between MCUs and MPUs
Future Electronics Editor's Choice AI & ML
The Renesas RA8 series microcontrollers feature Arm Helium technology, which boosts the performance of DSP functions and of AI and machine learning algorithms.

Read more...
Hardware architectural options for artificial intelligence systems
NuVision Electronics Editor's Choice AI & ML
With smart sensors creating data at an ever-increasing rate, it is becoming exponentially more difficult to consume and make sense of the data to extract relevant insight. This is providing the impetus behind the rapidly developing field of artificial intelligence.

Read more...
xG26 sets new standard in multiprotocol wireless device performance
Altron Arrow AI & ML
Silicon Labs has announced its new xG26 family of Wireless SoCs and MCUs, which consists of the multiprotocol MG26 SoC, the Bluetooth LE BG26 SoC, and the PG26 MCU.

Read more...
SolidRun unveils new SoM
Altron Arrow AI & ML
SolidRun and Hailo has unveiled a game-changer for engineers and AI product developers with the launch of their market-ready SoM, which packs the cutting-edge capabilities of the Hailo-15H SoC.

Read more...
Banana Pi with NPU
CST Electronics AI & ML
The latest Banana Pi SBC, the BPI-M7, is powered by Rockchip’s latest flagship RK3588 octa-core 64-bit processor, with a maximum frequency of 2,4 GHz.

Read more...
ESP32-P4 high-performance MCU
iCorp Technologies AI & ML
Powered by a dual-core RISC-V CPU running up to 400 MHz, ESP32-P4 also supports single-precision FPU and AI extensions, thus providing all the necessary computational resources.

Read more...
AI-native IoT platform launched
EBV Electrolink AI & ML
These highly-integrated Linux and Android SoCs from Synaptics are optimised for consumer, enterprise, and industrial applications and deliver an ‘out-of-the-box’ edge AI experience.

Read more...
Flash for AI
EBV Electrolink AI & ML
SCM offers a midway latency point between DRAM and SSDs, and when coupled with the introduction of CXL, low-latency flash, such as XL-FLASH, is well-positioned to deliver improvements in price, system performance, and power consumption to everything from servers to edge devices deploying the power of AI.

Read more...
Speeding up the rollout of renewable energy with AI
AI & ML
Understanding that AI, particularly within the renewables space, will not take away jobs, but rather create them, is key to leveraging the immense power of this technology to drive South Africa forward.

Read more...