AMD Launches New AI Chips to Take on Leader Nvidia 

Lisa Su, chairwoman and CEO of Advanced Micro Devices (AMD), delivers the opening keynote speech at Computex 2024, Taiwan's premier tech expo, in Taipei on June 3, 2024. (AFP)
Lisa Su, chairwoman and CEO of Advanced Micro Devices (AMD), delivers the opening keynote speech at Computex 2024, Taiwan's premier tech expo, in Taipei on June 3, 2024. (AFP)
TT

AMD Launches New AI Chips to Take on Leader Nvidia 

Lisa Su, chairwoman and CEO of Advanced Micro Devices (AMD), delivers the opening keynote speech at Computex 2024, Taiwan's premier tech expo, in Taipei on June 3, 2024. (AFP)
Lisa Su, chairwoman and CEO of Advanced Micro Devices (AMD), delivers the opening keynote speech at Computex 2024, Taiwan's premier tech expo, in Taipei on June 3, 2024. (AFP)

Advanced Micro Devices unveiled its latest artificial intelligence processors on Monday and detailed its plan to develop AI chips over the next two years in a bid to challenge industry leader Nvidia.

At the Computex technology trade show in Taipei, AMD CEO Lisa Su introduced the MI325X accelerator, which is set to be made available in the fourth quarter of 2024.

The race to develop generative artificial intelligence programs has led to towering demand for the advanced chips used in AI data centers able to support these complex applications.

Santa Clara, California-based AMD has been vying to compete against Nvidia, which currently dominates the lucrative market for AI semiconductors and commands about 80% of its share.

Since last year, Nvidia has made it clear to investors that it plans to shorten its release cycle to annually, and now AMD has followed suit.

"AI is clearly our number one priority as a company and we have really harnessed all of the development capability within the company to do that," Su told reporters.

"This annual cadence is something that is there because the market requires newer products and newer capabilities... Every year we have the next big thing such that we always have the most competitive portfolio."

AMD also introduced an upcoming series of chips titled MI350, which is expected to be available in 2025 and will be based on new chip architecture.

Compared to the currently available MI300 series of AI chips, AMD said it expects the MI350 to perform 35 times better in inference - the process of computing generative AI responses. Additionally, AMD revealed the MI400 series, which will arrive in 2026 and will be based on an architecture called "Next".

Investors who have poured billions of dollars into Wall Street's picks-and-shovels trade have been seeking longer-term updates from chip firms, as they evaluate the longevity of the booming genAI rally, which so far has shown no signs of slowing down.

AMD's shares have more than doubled since the start of 2023. This surge still pales in comparison to the more than seven-fold rise in Nvidia's shares in the same time period.

AMD is aiming at an AI chip product cycle of one year. Similarly, Nvidia said it plans to release a new family of AI chips every year.

AMD's Su said in April that the company expects AI chip sales of roughly $4 billion for 2024, an increase of $500 million from its prior estimate.

At the Computex event, AMD also said its latest generation of central processor units (CPUs) will likely be available in the second half of 2024.

While businesses generally prioritize spending on AI chips in data centers, some of AMD's CPUs are used in conjunction with graphics processor units, though the ratio is skewed in favor of GPUs.

AMD detailed architecture for its new neural processing units (NPUs), which are dedicated to handling on-device AI tasks in AI PCs.

Chipmakers have been banking on added AI capabilities to drive growth in the PC market as it emerges from a years-long slump.

PC providers such as HP and Lenovo will release devices which include AMD's AI PC chips. AMD said its processors exceed Microsoft's Copilot+ PC requirements.



Samsung Elec and AMD Sign MoU on AI Memory, Explore Foundry Partnership

FILE PHOTO: The logo of Samsung Electronics is seen at the company's store in Seoul, South Korea, April 15, 2025. REUTERS/Kim Hong-Ji/File Photo
FILE PHOTO: The logo of Samsung Electronics is seen at the company's store in Seoul, South Korea, April 15, 2025. REUTERS/Kim Hong-Ji/File Photo
TT

Samsung Elec and AMD Sign MoU on AI Memory, Explore Foundry Partnership

FILE PHOTO: The logo of Samsung Electronics is seen at the company's store in Seoul, South Korea, April 15, 2025. REUTERS/Kim Hong-Ji/File Photo
FILE PHOTO: The logo of Samsung Electronics is seen at the company's store in Seoul, South Korea, April 15, 2025. REUTERS/Kim Hong-Ji/File Photo

Samsung Electronics and Advanced Micro Devices (AMD) signed a memorandum of understanding to expand their strategic partnership on memory chip supplies for artificial intelligence infrastructure, the companies said on Wednesday.

The agreement will focus on supplying Samsung's next-generation high-bandwidth memory (HBM4) for AMD's upcoming Instinct MI455X AI accelerators, as well as optimized DDR5 memory for AMD's sixth-generation EPYC processors, they said in a statement.

The companies will also discuss opportunities for a foundry partnership, under which Samsung could provide contract chip manufacturing services ⁠for next-generation AMD ⁠products.

Under the agreement, Samsung will position itself as a key HBM4 supplier for AMD's next-generation AI GPUs, Reuters reported. The South Korean firm has already been a primary HBM supplier for AMD, supplying HBM3E chips used in AMD's MI350X and MI355X accelerators.

The ⁠agreement comes during the week of Nvidia's annual developer conference GTC, where CEO Jensen Huang on Monday announced a foundry partnership with the Korean firm and praised its HBM4 chips.

The tie-up highlights a broader race among global chipmakers to lock in long-term supply partnerships for advanced memory, as AI-driven demand reshapes the semiconductor industry and tightens supply of HBM chips.

Last month, AMD said it had agreed ⁠to sell ⁠up to $60 billion worth of AI chips to Meta Platforms over five years, a deal that allows the Facebook owner to purchase as much as 10% of the chips. AMD signed a similar deal with OpenAI last year.

Samsung, the world's largest memory chipmaker, has been seeking to narrow the gap with rivals in the fast-growing HBM segment. It holds about a 22% share of the global HBM market, compared with market leader SK Hynix's 57%, according to Counterpoint.


Nvidia Making AI Module for Outer Space

Nvidia CEO Jensen Huang says artificial intelligence powered by the company's graphics processing units is quickly infusing nearly everything from Disney character robots to data centers that may one day be orbiting the planet. JOSH EDELSON / AFP
Nvidia CEO Jensen Huang says artificial intelligence powered by the company's graphics processing units is quickly infusing nearly everything from Disney character robots to data centers that may one day be orbiting the planet. JOSH EDELSON / AFP
TT

Nvidia Making AI Module for Outer Space

Nvidia CEO Jensen Huang says artificial intelligence powered by the company's graphics processing units is quickly infusing nearly everything from Disney character robots to data centers that may one day be orbiting the planet. JOSH EDELSON / AFP
Nvidia CEO Jensen Huang says artificial intelligence powered by the company's graphics processing units is quickly infusing nearly everything from Disney character robots to data centers that may one day be orbiting the planet. JOSH EDELSON / AFP

Nvidia chief Jensen Huang on Monday said the leading artificial intelligence chip maker is heading for space with a goal of powering orbiting data centers.

An Nvidia graphics processing unit (GPU) was launched into space late last year by startup Starcloud in what was touted as an off-planet debut for the technology, but now Nvidia is creating a module intended as a building block for data centers there.

"We're working with our partners on a new computer called Vera Rubin Space One," Huang said as he kicked off the GPU-maker's annual developers conference in Silicon Valley.

"It's going to go out to space and start data centers."

Partners in the project include Starcloud, which is planning a November satellite launch that will mark the "cosmic debut" of the new Nvidia module.

A Starcloud-1 satellite, about the size of a small refrigerator, is expected to be packed with 100 times more computing power than any previous space-based operation.

"In 10 years, nearly all new data centers will be being built in outer space," predicted Starcloud co-founder and chief Philip Johnston.

The startup explained that it plans to power Google AI with the Nvidia GPUs to show that large language models can run in outer space.

Nvidia described the Vera Rubin module as being optimized for AI, enabling real-time sensing, decision making, and autonomous functioning.

"Space computing, the final frontier, has arrived," Huang said.

"With our partners, we're extending Nvidia beyond our planet - boldly taking intelligence where it's never gone before."

Tech firms are floating the idea of building data centers in space and tapping into the sun's energy to meet out-of-this-world power demands in a fierce artificial intelligence race.

More than a dozen startups, aerospace leaders, and major tech firms are involved in the development, testing, or planning of space-based data centers.

The big draw of space for data centers is power supply, with the option of synchronizing satellites to the sun's orbit to ensure constant light beaming onto solar panels.

Building in space also avoids the challenges of acquiring land and meeting local regulations or community resistance to projects.

Critical technical aspects of such operations need to be resolved, however, particularly damage to the orbiting data centers from high levels of radiation and extreme temperatures, and the danger of them being hit by space junk.


Samsung Shares Rise After Nvidia’s Huang Flags Tie-up on New AI Chips

10 September 2025, Bavaria, Munich: The Samsung logo can be seen at the Samsung stand during the International Motor Show (IAA Mobility). (dpa)
10 September 2025, Bavaria, Munich: The Samsung logo can be seen at the Samsung stand during the International Motor Show (IAA Mobility). (dpa)
TT

Samsung Shares Rise After Nvidia’s Huang Flags Tie-up on New AI Chips

10 September 2025, Bavaria, Munich: The Samsung logo can be seen at the Samsung stand during the International Motor Show (IAA Mobility). (dpa)
10 September 2025, Bavaria, Munich: The Samsung logo can be seen at the Samsung stand during the International Motor Show (IAA Mobility). (dpa)

Shares of Samsung Electronics rose as much as 5% on Tuesday after Nvidia CEO Jensen Huang said the South Korean company was producing Nvidia's new artificial intelligence chips.

The news fueled expectations that Samsung's foundry division, which makes logic chips for customers including Tesla, Apple and Samsung's phone division, may be able to turn around ‌as early ‌as next year after posting ‌billions ⁠of dollars in annual ⁠losses in recent years, analysts said.

At Nvidia's GTC developer conference in California on Monday, Huang unveiled Nvidia's new AI inference processor based on technology from chip startup Groq.

"I want to thank Samsung who manufactures the Groq LP30 ⁠chip for us and they're cranking as ‌hard as they ‌can," he said, adding the chips were in production, ‌and would be shipped in the second half ‌of this year.

Samsung also showcased the Nvidia chips made using its 4-nanometer manufacturing process at the GTC.

Samsung shares were up 4.3% at 196,800 won ‌as of 0252 GMT, after earlier reaching 198,000 won. The broader market was ⁠up 2.7%.

Sohn ⁠In-joon, an analyst at Heungkuk Securities, expected Samsung's foundry business would be able to reach breakeven later next year. But he said weak demand from mobile phones stemming from surging memory chip prices could weigh on foundry earnings.

Advanced Micro Devices' CEO Lisa Su will meet Samsung Electronics Chairman Jay Y. Lee in South Korea on Wednesday, media reports said, with eyes on whether the two would discuss cooperation in memory chips and logic semiconductors.