Development of the Tensor SoC is making continuous progress, breaking Samsung's integrated wafer foundry and packaging testing model.
Since the first Tensor SoC was equipped in the Pixel 6 series in 2021, Google has used Samsung's foundry-produced chips as the core of its mobile phones. However, next year's tenth-generation Pixel will usher in a significant change, with the Tensor G5 expected to become the first Pixel series dedicated chip produced by TSMC.
The Information reported in July last year that Google has reached an agreement with TSMC to produce fully customized Tensor SoCs for Pixel devices. If Google retains the existing naming method, this chip may be called the Tensor G5. Since the foreign media revealed this, the development of the Tensor SoC has been making continuous progress, including the news that the test order was won by Kyun Yuan Electronics, breaking Samsung's integrated wafer foundry and packaging testing model.
Another media outlet, Business Korea, recently reported that Google will use TSMC's 3-nanometer process for the Tensor G5 to be launched next year, which is expected to greatly enhance the performance level of the Pixel series. The currently available Pixel 8 series uses the Tensor G3, which is built with Samsung's 4-nanometer process. By the second half of 2025, switching to a 3-nanometer process is inevitable.
This is not surprising, as Apple started using a 3-nanometer process from the iPhone 15 Pro series last year. More importantly, it is expected that Qualcomm and MediaTek's next-generation chips will follow suit, and the non-Apple camp's Tensor G5 will not have a unique process advantage.
In addition, the report from Business Korea also discussed that Samsung is working hard to solve the issues of yield and power consumption. The upcoming Exynos 2500 chip claims to have power consumption and heat dissipation performance about 10% to 20% lower than TSMC's 3-nanometer process.
Apple has been equipped with its self-developed A-series chips since the iPhone 4 and has expanded its customized M-series chips to the entire Mac series. It has been nearly a year since the development of 3-nanometer process chips for the iPhone and Mac. Android camp competitors have just started to get involved in this technology, and the Tensor SoC's transfer to TSMC production is expected to bring a perceptible upgrade to the new Pixel devices.Google's New Generation Cloud AI Chip
Previously, Google introduced the TPUv5p, which is its latest generation of cloud AI chips and also the most powerful and cost-effective chip to date. Each TPUv5p Pod contains up to 8,960 chips, interconnected through high-bandwidth chip-to-chip connections to achieve fast data transfer and optimal performance.
The new generation TPUv5p excels in AI performance, offering 459 teraFLOPS of bfloat16 performance or 918 teraOPS of Int8 performance, equipped with 95GB of high-bandwidth memory and a data transfer rate of 2.76TB/s. Compared to the previous TPUv4, the TPUv5p has doubled the number of floating-point operations and tripled the high memory bandwidth, which has attracted widespread attention in the field of artificial intelligence.
In addition, the TPUv5p has also improved the training speed of large language models (LLM) by 2.8 times, which is about 50% faster than the previous TPUv5e. Google has also increased the computational power, making the scalability of TPUv5p four times higher than TPUv4. Overall, the TPUv5p has the following improvements compared to TPUv4: the number of floating-point operations has increased by 2 times, memory capacity has increased by 3 times, LLM training speed has improved by 2.8 times, embedding-intensive model training speed has improved by 1.9 times, bandwidth has increased by 2.25 times, and chip-to-chip interconnect bandwidth has doubled.
Google has achieved significant success in the field of AI and attributes it to excellent hardware and software resources. Google's cloud AI supercomputer is a set of elements that work together to achieve modern AI workloads. Google integrates performance-optimized computing, optimal storage, and liquid cooling to fully utilize its huge computing power, thus achieving industry-leading performance.
In terms of software, Google has strengthened support for popular machine learning frameworks (such as JAX, TensorFlow, and PyTorch) and provided some powerful tools and compilers. These tools and compilers can optimize distributed architectures, making the development and training of complex models on different hardware platforms more efficient and easier to use. Google has also developed multi-chip training and multi-host inference software to simplify the management of scaling, training, and serving workloads.
Google's revolutionary approach to artificial intelligence is strongly supported by hardware and software elements that will break various limitations in the industry. The newly released cloud AI chip TPUv5p and Google's AI supercomputer will bring more possibilities and opportunities for ongoing AI development. It is foreseeable that these advanced technologies will further intensify competition and promote the development of the field of artificial intelligence.
Google's new generation cloud AI chip TPUv5p has excellent performance. Compared with the previous TPUv4, TPUv5p has made significant improvements in many aspects.
Firstly, the number of floating-point operations of TPUv5p has doubled. It can provide 459 teraFLOPS of bfloat16 performance or 918 teraOPS of Int8 performance, greatly increasing the speed of computation. This is very helpful for handling complex computing tasks and large-scale machine learning models.
Secondly, the memory capacity of TPUv5p is three times higher than that of TPUv4. It is equipped with 95GB of high-bandwidth memory, which can access and store data faster. This is crucial for handling large-scale datasets and complex model training.Thirdly, the TPUv5p has demonstrated a 2.8-fold generational improvement in the training speed of large language models (LLMs). This is very helpful for tasks such as natural language processing and machine translation, as it can accelerate the training speed of models and enhance work efficiency.
In addition, the TPUv5p has also made significant progress in the training speed of embedding-dense models, achieving a 1.9-fold increase. This is very beneficial for handling deep learning models and neural network models, as it can improve the training efficiency and accuracy of the models.
Finally, the TPUv5p has also seen a significant improvement in bandwidth and inter-chip interconnect speed. Its bandwidth has increased by 2.25 times, reaching a transfer speed of 2765GB per second. The inter-chip interconnect bandwidth is twice the original, reaching a transfer speed of 4800Gbps per chip. This can improve the efficiency and speed of data transmission, thereby enhancing the overall performance.
Comments
Share your experience
Related Articles
Mind-Blowing: 10 Countries Where Average Lifespan Is Under 50—Here’s the Chilling Reason
200,000: The shock China's new energy vehicles give to the world
This year, the explosive popularity of the Beijing Auto Show has continued to the Chongqing Auto Show. At the bustling a...
Automotive PMIC applications, who is the strong player?
The market size of automotive Power Management IC (PMIC) chips is entering a period of rapid growth.PMIC stands for Powe...
NB-IoT and LoRa become the kings of the Internet of Things
By the end of 2023, China still accounts for nearly 90% of global NB-IoT connections.Omdia's latest research finds that ...
Mitsui Chemicals will mass-produce a new photoresist film to support ASML's next
According to reports, this CNT film can achieve a high EUV transmission rate of over 92% and a light resistance capabili...
Understanding CPU benchmark testing
Benchmarks do not always accurately reflect real-world performance.Upgrading to a new CPU is both exciting and frustrati...
Contract price increases boost DRAM's first-quarter revenue by 5.1%
The shipment of DRAM suppliers is expected to see a seasonal rebound in the second quarter.TrendForce points out that in...
Wolfspeed postpones the plan to build a factory in Germany, and the start time i
On Wall Street, the majority do not have high hopes for Wolfspeed.Wolfspeed has postponed its plans to build a $3 billio...
Obstacles encountered in the PCIe 6.0 and 7.0 standards
The adoption of new technologies may face some delays.Earlier this week, the organization responsible for developing the...
Unbelievable! These Are the 10 Most Painful Conditions on Earth (#5 Still Exists in the US!)
Shock! Market value evaporates 200 billion
Can the company's market value stabilize above 3 trillion US dollars?This Friday, the AI chip leader Nvidia suffered a f...
Taiwan's semiconductor output value will break through the 5 trillion mark! Behi
Concerns for Taiwan's Semiconductor Industry in China.The 2024 Taipei International Computer Show in China came to a suc...
The hotel location shared by my best friend is the same one my husband often stays at-7
The price war of large models has started, can multi-core hybrid become a breakt
Recently, several major model companies in China have successively lowered the prices of their related products.Starting...
The World’s 10 Most Dangerous Cities: Crime, War, and Terror Lurk Everywhere
Intensified competition in network technology, Nvidia may not beat Broadcom
Broadcom's financial report highlights three major positives: wired communication, software revenue, and stock split.Aft...
The rise of storage power in the AI era
Under the tide of the digital economy, data has become a new type of production material.Currently, there are three majo...
Hua Hai Qingke delivers the first 12-inch equipment
China's domestic semiconductor equipment is gradually reaching its prime.On June 12th, Huahai Qingke announced that its ...
The 10 Craziest World Records You’d NEVER Want to Break(Spoiler: It’s Brutal)
TSMC researches advanced chip packaging technology: rectangles replace round waf
As chip sizes increase, the capacity of 12-inch wafers is gradually becoming insufficient.TSMC is researching a new adva...
The future of AI chips may not be GPUs
In the layout of artificial intelligence (AI) computing architectures, the collaborative working model between the Centr...
Fighting back against "big girl" Micron
Recently, South Korea's lawsuit against Micron has been a shock to many. For American companies, South Korea has always ...
$3.34 trillion! Nvidia's market value tops the world
When calculated in Chinese Yuan, Nvidia's market value is approximately equivalent to 12.96 times the market value of Kw...
The hotel location shared by my best friend is the same one my husband often stays at-13
Best AI chip companies: Don't underestimate these Nvidia competitors
Nvidia Dominates the AI Chip Market, but Competitors are Emerging.The development of integrated circuits and the upgradi...
TSMC's Nanjing factory receives an unlimited exemption permit from the United St
This move heralds that TSMC will continue to play a significant role in the global chip market competition in the future...
Will application-specific integrated circuits become ubiquitous?
How Changes in the End Market and Shrinking Device Sizes are Altering Chip DesignThe evolution of the end market and the...
Ruthenium interconnect technology is coming
Has the Era of Copper Come to an End?With the continuous decline in the efficiency of copper interconnect technology, ch...
TOP 10 Giants That Actually Existed: #4 Will Defy Everything You Knew.
Dialogue with GAC Aion's Gu Huinan: discussing going overseas, technology, and i
In 2024, one of the key words in the automotive industry is undoubtedly "intense competition." The market competition is...