Nvidia Rivals Focus on Building a Different Kind of Chip to Power AI Products

The NVIDIA logo is seen near a computer motherboard in this illustration taken January 8, 2024. (Reuters)
The NVIDIA logo is seen near a computer motherboard in this illustration taken January 8, 2024. (Reuters)
TT

Nvidia Rivals Focus on Building a Different Kind of Chip to Power AI Products

The NVIDIA logo is seen near a computer motherboard in this illustration taken January 8, 2024. (Reuters)
The NVIDIA logo is seen near a computer motherboard in this illustration taken January 8, 2024. (Reuters)

Building the current crop of artificial intelligence chatbots has relied on specialized computer chips pioneered by Nvidia, which dominates market and made itself the poster child of the AI boom.

But the same qualities that make those graphics processor chips, or GPUs, so effective at creating powerful AI systems from scratch make them less efficient at putting AI products to work.

That's opened up the AI chip industry to rivals who think they can compete with Nvidia in selling so-called AI inference chips that are more attuned to the day-to-day running of AI tools and designed to reduce some of the huge computing costs of generative AI.

“These companies are seeing opportunity for that kind of specialized hardware,” said Jacob Feldgoise, an analyst at Georgetown University's Center for Security and Emerging Technology. “The broader the adoption of these models, the more compute will be needed for inference and the more demand there will be for inference chips.”

What is AI inference? It takes a lot of computing power to make an AI chatbot. It starts with a process called training or pretraining — the “P” in ChatGPT — that involves AI systems “learning” from the patterns of huge troves of data. GPUs are good at doing that work because they can run many calculations at a time on a network of devices in communication with each other.

However, once trained, a generative AI tool still needs chips to do the work — such as when you ask a chatbot to compose a document or generate an image. That's where inferencing comes in. A trained AI model must take in new information and make inferences from what it already knows to produce a response.

GPUs can do that work, too. But it can be a bit like taking a sledgehammer to crack a nut.

“With training, you’re doing a lot heavier, a lot more work. With inferencing, that’s a lighter weight,” said Forrester analyst Alvin Nguyen.

That's led startups like Cerebras, Groq and d-Matrix as well as Nvidia's traditional chipmaking rivals — such as AMD and Intel — to pitch more inference-friendly chips as Nvidia focuses on meeting the huge demand from bigger tech companies for its higher-end hardware.

Inside an AI inference chip lab D-Matrix, which is launching its first product this week, was founded in 2019 — a bit late to the AI chip game, as CEO Sid Sheth explained during a recent interview at the company’s headquarters in Santa Clara, California, the same Silicon Valley city that's also home to AMD, Intel and Nvidia.

“There were already 100-plus companies. So when we went out there, the first reaction we got was ‘you’re too late,’” he said. The pandemic's arrival six months later didn't help as the tech industry pivoted to a focus on software to serve remote work.

Now, however, Sheth sees a big market in AI inferencing, comparing that later stage of machine learning to how human beings apply the knowledge they acquired in school.

“We spent the first 20 years of our lives going to school, educating ourselves. That’s training, right?” he said. “And then the next 40 years of your life, you kind of go out there and apply that knowledge — and then you get rewarded for being efficient.”

The product, called Corsair, consists of two chips with four chiplets each, made by Taiwan Semiconductor Manufacturing Company — the same manufacturer of most of Nvidia's chips — and packaged together in a way that helps to keep them cool.

The chips are designed in Santa Clara, assembled in Taiwan and then tested back in California. Testing is a long process and can take six months — if anything is off, it can be sent back to Taiwan.

D-Matrix workers were doing final testing on the chips during a recent visit to a laboratory with blue metal desks covered with cables, motherboards and computers, with a cold server room next door.

Who wants AI inference chips? While tech giants like Amazon, Google, Meta and Microsoft have been gobbling up the supply of costly GPUs in a race to outdo each other in AI development, makers of AI inference chips are aiming for a broader clientele.

Forrester's Nguyen said that could include Fortune 500 companies that want to make use of new generative AI technology without having to build their own AI infrastructure. Sheth said he expects a strong interest in AI video generation.

“The dream of AI for a lot of these enterprise companies is you can use your own enterprise data,” Nguyen said. “Buying (AI inference chips) should be cheaper than buying the ultimate GPUs from Nvidia and others. But I think there’s going to be a learning curve in terms of integrating it.”

Feldgoise said that, unlike training-focused chips, AI inference work prioritizes how fast a person will get a chatbot's response.

He said another whole set of companies is developing AI hardware for inference that can run not just in big data centers but locally on desktop computers, laptops and phones.

Why does this matter? Better-designed chips could bring down the huge costs of running AI to businesses. That could also affect the environmental and energy costs for everyone else.

Sheth says the big concern right now is, “are we going to burn the planet down in our quest for what people call AGI — human-like intelligence?”

It’s still fuzzy when AI might get to the point of artificial general intelligence — predictions range from a few years to decades. But, Sheth notes, only a handful of tech giants are on that quest.

“But then what about the rest?” he said. “They cannot be put on the same path.”

The other set of companies don’t want to use very large AI models — it’s too costly and uses too much energy.

“I don’t know if people truly, really appreciate that inference is actually really going to be a much bigger opportunity than training. I don’t think they appreciate that. It’s still training that is really grabbing all the headlines,” Sheth said.



Google Says to Build New Subsea Cables from India in AI Push

A logo of Google is on display at Bharat Mandapam, one of the venues for AI Impact Summit, in New Delhi, India, February 17, 2026. REUTERS/Bhawika Chhabra
A logo of Google is on display at Bharat Mandapam, one of the venues for AI Impact Summit, in New Delhi, India, February 17, 2026. REUTERS/Bhawika Chhabra
TT

Google Says to Build New Subsea Cables from India in AI Push

A logo of Google is on display at Bharat Mandapam, one of the venues for AI Impact Summit, in New Delhi, India, February 17, 2026. REUTERS/Bhawika Chhabra
A logo of Google is on display at Bharat Mandapam, one of the venues for AI Impact Summit, in New Delhi, India, February 17, 2026. REUTERS/Bhawika Chhabra

Google announced Wednesday it would build new subsea cables from India and other locations as part of its existing $15 billion investment in the South Asian nation, which is hosting a major artificial intelligence summit this week.

The US tech giant said it would build "three subsea paths connecting India to Singapore, South Africa, and Australia; and four strategic fiber-optic routes that bolster network resilience and capacity between the United States, India, and multiple locations across the Southern Hemisphere".


Mark Zuckerberg Set to Testify in Watershed Social Media Trial 

Meta's CEO Mark Zuckerberg testifies during the Senate Judiciary Committee hearing on online child sexual exploitation at the US Capitol in Washington, US, January 31, 2024. (Reuters)
Meta's CEO Mark Zuckerberg testifies during the Senate Judiciary Committee hearing on online child sexual exploitation at the US Capitol in Washington, US, January 31, 2024. (Reuters)
TT

Mark Zuckerberg Set to Testify in Watershed Social Media Trial 

Meta's CEO Mark Zuckerberg testifies during the Senate Judiciary Committee hearing on online child sexual exploitation at the US Capitol in Washington, US, January 31, 2024. (Reuters)
Meta's CEO Mark Zuckerberg testifies during the Senate Judiciary Committee hearing on online child sexual exploitation at the US Capitol in Washington, US, January 31, 2024. (Reuters)

Mark Zuckerberg will testify in an unprecedented social media trial that questions whether Meta's platforms deliberately addict and harm children.

Meta's CEO is expected to answer tough questions on Wednesday from attorneys representing a now 20-year-old woman identified by the initials KGM, who claims her early use of social media addicted her to the technology and exacerbated depression and suicidal thoughts. Meta Platforms and Google’s YouTube are the two remaining defendants in the case, which TikTok and Snap have settled.

Zuckerberg has testified in other trials and answered questions from Congress about youth safety on Meta's platforms, and he apologized to families at that hearing whose lives had been upended by tragedies they believed were because of social media.

This trial, though, marks the first time Zuckerberg will answer similar questions in front of a jury. and, again, bereaved parents are expected to be in the limited courtroom seats available to the public.

The case, along with two others, has been selected as a bellwether trial, meaning its outcome could impact how thousands of similar lawsuits against social media companies would play out.

A Meta spokesperson said the company strongly disagrees with the allegations in the lawsuit and said they are “confident the evidence will show our longstanding commitment to supporting young people.”

One of Meta's attorneys, Paul Schmidt, said in his opening statement that the company is not disputing that KGM experienced mental health struggles, but rather that Instagram played a substantial factor in those struggles.

He pointed to medical records that showed a turbulent home life, and both he and an attorney representing YouTube argue she turned to their platforms as a coping mechanism or a means of escaping her mental health struggles.

Zuckerberg's testimony comes a week after that of Adam Mosseri, the head of Meta's Instagram, who said in the courtroom that he disagrees with the idea that people can be clinically addicted to social media platforms.

Mosseri maintained that Instagram works hard to protect young people using the service, and said it's “not good for the company, over the long run, to make decisions that profit for us but are poor for people’s well-being."

Much of Mosseri's questioning from the plaintiff's lawyer, Mark Lanier, centered on cosmetic filters on Instagram that changed people’s appearance — a topic that Lanier is sure to revisit with Zuckerberg.

He is also expected to face questions about Instagram’s algorithm, the infinite nature of Meta’ feeds and other features the plaintiffs argue are designed to get users hooked.


US Tech Giant Nvidia Announces India Deals at AI Summit

FILED - 04 February 2026, Bavaria, Munich: The NVIDIA logo is seen during a press conference at the opening of Telekom and NVIDIA's AI factory "Industrial AI Cloud". Photo: Sven Hoppe/dpa
FILED - 04 February 2026, Bavaria, Munich: The NVIDIA logo is seen during a press conference at the opening of Telekom and NVIDIA's AI factory "Industrial AI Cloud". Photo: Sven Hoppe/dpa
TT

US Tech Giant Nvidia Announces India Deals at AI Summit

FILED - 04 February 2026, Bavaria, Munich: The NVIDIA logo is seen during a press conference at the opening of Telekom and NVIDIA's AI factory "Industrial AI Cloud". Photo: Sven Hoppe/dpa
FILED - 04 February 2026, Bavaria, Munich: The NVIDIA logo is seen during a press conference at the opening of Telekom and NVIDIA's AI factory "Industrial AI Cloud". Photo: Sven Hoppe/dpa

US artificial intelligence chip titan Nvidia unveiled tie-ups with Indian computing firms on Wednesday as tech companies rushed to announce deals and investments at a global AI conference in New Delhi.

This week's AI Impact Summit is the fourth annual gathering to discuss how to govern the fast-evolving technology -- and also an opportunity to "define India's leadership in the AI decade ahead", organizers say.

Mumbai cloud and data center provider L&T said it was teaming up with Nvidia, the world's most valuable company, to build what it touted as "India's largest gigawatt-scale AI factory".

"We are laying the foundation for world-class AI infrastructure that will power India's growth," said Nvidia boss Jensen Huang in a statement that did not put a figure on the investment.

L&T said it would use Nvidia's powerful processors, which can train and run generative AI tech, to provide data center capacity of up to 30 megawatts in Chennai and 40 megawatts in Mumbai.

Nvidia said it was also working with other Indian AI infrastructure players such as Yotta, which will deploy more than 20,000 top-end Nvidia Blackwell processors as part of a $2 billion investment.

Dozens of world leaders and ministerial delegations have come to India for the summit to discuss the opportunities and threats, from job losses to misinformation, that AI poses.

Last year India leapt to third place -- overtaking South Korea and Japan -- in an annual global ranking of AI competitiveness calculated by Stanford University researchers.

But despite plans for large-scale infrastructure and grand ambitions for innovation, experts say the country has a long way to go before it can rival the United States and China.

The conference has also brought a flurry of deals, with IT minister Ashwini Vaishnaw saying Tuesday that India expects more than $200 billion in investments over the next two years, including roughly $90 billion already committed.

Separately, India's Adani Group said Tuesday it plans to invest $100 billion by 2035 to develop "hyperscale AI-ready data centers", a boost to New Delhi's push to become a global AI hub.

Microsoft said it was investing $50 billion this decade to boost AI adoption in developing countries, while US artificial intelligence startup Anthropic and Indian IT giant Infosys said they would work together to build AI agents for the telecoms industry.

Nvidia's Huang is not attending the AI summit but other top US tech figures joining include OpenAI's Sam Altman, Google DeepMind's Demis Hassabis and Microsoft founder Bill Gates.

Indian Prime Minister Narendra Modi and other world leaders including French President Emmanuel Macron and Brazil's Luiz Inacio Lula da Silva are expected to deliver a statement at the end of the week about how they plan to address concerns raised by AI technology.

But experts say that the broad focus of the event and vague promises made at previous global AI summits in France, South Korea and Britain mean that concrete commitments are unlikely.

Nick Patience, practice lead for AI at tech research group Futurum, told AFP that nonbinding declarations could still "set the tone for what acceptable AI governance looks like".

But "the largest AI companies deploy capabilities at a pace that makes 18-month legislative cycles look glacial," Patience said.

"So it's a case of whether governments can converge fast enough to create meaningful guardrails before de facto standards are set by the companies themselves."