Nvidia Rivals Focus on Building a Different Kind of Chip to Power AI Products

The NVIDIA logo is seen near a computer motherboard in this illustration taken January 8, 2024. (Reuters)
The NVIDIA logo is seen near a computer motherboard in this illustration taken January 8, 2024. (Reuters)
TT

Nvidia Rivals Focus on Building a Different Kind of Chip to Power AI Products

The NVIDIA logo is seen near a computer motherboard in this illustration taken January 8, 2024. (Reuters)
The NVIDIA logo is seen near a computer motherboard in this illustration taken January 8, 2024. (Reuters)

Building the current crop of artificial intelligence chatbots has relied on specialized computer chips pioneered by Nvidia, which dominates market and made itself the poster child of the AI boom.

But the same qualities that make those graphics processor chips, or GPUs, so effective at creating powerful AI systems from scratch make them less efficient at putting AI products to work.

That's opened up the AI chip industry to rivals who think they can compete with Nvidia in selling so-called AI inference chips that are more attuned to the day-to-day running of AI tools and designed to reduce some of the huge computing costs of generative AI.

“These companies are seeing opportunity for that kind of specialized hardware,” said Jacob Feldgoise, an analyst at Georgetown University's Center for Security and Emerging Technology. “The broader the adoption of these models, the more compute will be needed for inference and the more demand there will be for inference chips.”

What is AI inference? It takes a lot of computing power to make an AI chatbot. It starts with a process called training or pretraining — the “P” in ChatGPT — that involves AI systems “learning” from the patterns of huge troves of data. GPUs are good at doing that work because they can run many calculations at a time on a network of devices in communication with each other.

However, once trained, a generative AI tool still needs chips to do the work — such as when you ask a chatbot to compose a document or generate an image. That's where inferencing comes in. A trained AI model must take in new information and make inferences from what it already knows to produce a response.

GPUs can do that work, too. But it can be a bit like taking a sledgehammer to crack a nut.

“With training, you’re doing a lot heavier, a lot more work. With inferencing, that’s a lighter weight,” said Forrester analyst Alvin Nguyen.

That's led startups like Cerebras, Groq and d-Matrix as well as Nvidia's traditional chipmaking rivals — such as AMD and Intel — to pitch more inference-friendly chips as Nvidia focuses on meeting the huge demand from bigger tech companies for its higher-end hardware.

Inside an AI inference chip lab D-Matrix, which is launching its first product this week, was founded in 2019 — a bit late to the AI chip game, as CEO Sid Sheth explained during a recent interview at the company’s headquarters in Santa Clara, California, the same Silicon Valley city that's also home to AMD, Intel and Nvidia.

“There were already 100-plus companies. So when we went out there, the first reaction we got was ‘you’re too late,’” he said. The pandemic's arrival six months later didn't help as the tech industry pivoted to a focus on software to serve remote work.

Now, however, Sheth sees a big market in AI inferencing, comparing that later stage of machine learning to how human beings apply the knowledge they acquired in school.

“We spent the first 20 years of our lives going to school, educating ourselves. That’s training, right?” he said. “And then the next 40 years of your life, you kind of go out there and apply that knowledge — and then you get rewarded for being efficient.”

The product, called Corsair, consists of two chips with four chiplets each, made by Taiwan Semiconductor Manufacturing Company — the same manufacturer of most of Nvidia's chips — and packaged together in a way that helps to keep them cool.

The chips are designed in Santa Clara, assembled in Taiwan and then tested back in California. Testing is a long process and can take six months — if anything is off, it can be sent back to Taiwan.

D-Matrix workers were doing final testing on the chips during a recent visit to a laboratory with blue metal desks covered with cables, motherboards and computers, with a cold server room next door.

Who wants AI inference chips? While tech giants like Amazon, Google, Meta and Microsoft have been gobbling up the supply of costly GPUs in a race to outdo each other in AI development, makers of AI inference chips are aiming for a broader clientele.

Forrester's Nguyen said that could include Fortune 500 companies that want to make use of new generative AI technology without having to build their own AI infrastructure. Sheth said he expects a strong interest in AI video generation.

“The dream of AI for a lot of these enterprise companies is you can use your own enterprise data,” Nguyen said. “Buying (AI inference chips) should be cheaper than buying the ultimate GPUs from Nvidia and others. But I think there’s going to be a learning curve in terms of integrating it.”

Feldgoise said that, unlike training-focused chips, AI inference work prioritizes how fast a person will get a chatbot's response.

He said another whole set of companies is developing AI hardware for inference that can run not just in big data centers but locally on desktop computers, laptops and phones.

Why does this matter? Better-designed chips could bring down the huge costs of running AI to businesses. That could also affect the environmental and energy costs for everyone else.

Sheth says the big concern right now is, “are we going to burn the planet down in our quest for what people call AGI — human-like intelligence?”

It’s still fuzzy when AI might get to the point of artificial general intelligence — predictions range from a few years to decades. But, Sheth notes, only a handful of tech giants are on that quest.

“But then what about the rest?” he said. “They cannot be put on the same path.”

The other set of companies don’t want to use very large AI models — it’s too costly and uses too much energy.

“I don’t know if people truly, really appreciate that inference is actually really going to be a much bigger opportunity than training. I don’t think they appreciate that. It’s still training that is really grabbing all the headlines,” Sheth said.



Neuralink Plans ‘High-Volume’ Brain Implant Production by 2026, Musk Says

Elon Musk steps off Air Force One upon arrival at Morristown Municipal Airport in Morristown, New Jersey, US, March 22, 2025. (AFP)
Elon Musk steps off Air Force One upon arrival at Morristown Municipal Airport in Morristown, New Jersey, US, March 22, 2025. (AFP)
TT

Neuralink Plans ‘High-Volume’ Brain Implant Production by 2026, Musk Says

Elon Musk steps off Air Force One upon arrival at Morristown Municipal Airport in Morristown, New Jersey, US, March 22, 2025. (AFP)
Elon Musk steps off Air Force One upon arrival at Morristown Municipal Airport in Morristown, New Jersey, US, March 22, 2025. (AFP)

Elon Musk's brain implant company Neuralink will start "high-volume production" of brain-computer interface devices and move to an entirely automated surgical procedure in 2026, Musk said in a post on the social media platform X on ‌Wednesday.

Neuralink did ‌not immediately respond ‌to ⁠a Reuters ‌request for comment.

The implant is designed to help people with conditions such as a spinal cord injury. The first patient has used it to play video ⁠games, browse the internet, post on ‌social media, and ‍move a cursor ‍on a laptop.

The company began ‍human trials of its brain implant in 2024 after addressing safety concerns raised by the US Food and Drug Administration, which had initially rejected its application in ⁠2022.

Neuralink said in September that 12 people worldwide with severe paralysis have received its brain implants and were using them to control digital and physical tools through thought. It also secured $650 million in a June funding round.


Report: France Aims to Ban Under-15s from Social Media from September 2026

French President Emmanuel Macron holds a press conference during a European Union leaders' summit, in Brussels, Belgium December 19, 2025. (Reuters)
French President Emmanuel Macron holds a press conference during a European Union leaders' summit, in Brussels, Belgium December 19, 2025. (Reuters)
TT

Report: France Aims to Ban Under-15s from Social Media from September 2026

French President Emmanuel Macron holds a press conference during a European Union leaders' summit, in Brussels, Belgium December 19, 2025. (Reuters)
French President Emmanuel Macron holds a press conference during a European Union leaders' summit, in Brussels, Belgium December 19, 2025. (Reuters)

France plans to ban children under 15 from social media sites and to prohibit mobile phones in high schools from September 2026, local media reported on Wednesday, moves that underscore rising public angst over the impact of online harms on minors.

President Emmanuel Macron has often pointed to social media as one of the factors to blame for violence among young people and has signaled he wants France to follow Australia, whose world-first ‌ban for under-16s ‌on social media platforms including Facebook, Snapchat, TikTok ‌and ⁠YouTube came into force ‌in December.

Le Monde newspaper said Macron could announce the measures in his New Year's Eve national address, due to be broadcast at 1900 GMT. His government will submit draft legislation for legal checks in early January, Le Monde and France Info reported.

The Elysee and the prime minister's office did not immediately respond to a request for comment on the reports.

Mobile phones have been banned ⁠in French primary and middle schools since 2018 and the reported new changes would extend that ban ‌to high schools. Pupils aged 11 to ‍15 attend middle schools in the French ‍educational system.

France also passed a law in 2023 requiring social platforms to ‍obtain parental consent for under-15s to create accounts, though technical challenges have impeded its enforcement.

Macron said in June he would push for regulation at the level of the European Union to ban access to social media for all under-15s after a fatal stabbing at a school in eastern France shocked the nation.

The European Parliament in ⁠November urged the EU to set minimum ages for children to access social media to combat a rise in mental health problems among adolescents from excessive exposure, although it is member states which impose age limits. Various other countries have also taken steps to regulate children's access to social media.

Macron heads into the New Year with his domestic legacy in tatters after his gamble on parliamentary elections in 2024 led to a hung parliament, triggering France's worst political crisis in decades that has seen a succession of weak governments.

However, cracking down further on minors' access to social media could prove popular, according to opinion ‌polls. A Harris Interactive survey in 2024 showed 73% of those canvassed supporting a ban on social media access for under-15s.


Poland Urges Brussels to Probe TikTok Over AI-Generated Content

The TikTok logo is pictured outside the company's US head office in Culver City, California, US, September 15, 2020. (Reuters)
The TikTok logo is pictured outside the company's US head office in Culver City, California, US, September 15, 2020. (Reuters)
TT

Poland Urges Brussels to Probe TikTok Over AI-Generated Content

The TikTok logo is pictured outside the company's US head office in Culver City, California, US, September 15, 2020. (Reuters)
The TikTok logo is pictured outside the company's US head office in Culver City, California, US, September 15, 2020. (Reuters)

Poland has asked the European Commission to investigate TikTok after the social media platform hosted AI-generated content including calls for Poland to withdraw from the EU, it said on Tuesday, adding that the content was almost certainly Russian disinformation.

"The disclosed content poses a threat to public order, information security, and the integrity of democratic processes in Poland and across the European Union," Deputy Digitalization Minister Dariusz Standerski said in a letter sent to the Commission.

"The nature of ‌the narratives, ‌the manner in which they ‌are distributed, ⁠and the ‌use of synthetic audiovisual materials indicate that the platform is failing to comply with the obligations imposed on it as a Very Large Online Platform (VLOP)," he added.

A Polish government spokesperson said on Tuesday the content was undoubtedly Russian disinformation as the recordings contained Russian syntax.

TikTok, representatives ⁠of the Commission and of the Russian embassy in Warsaw did not ‌immediately respond to Reuters' requests for ‍comment.

EU countries are taking ‍measures to head off any foreign state attempts to ‍influence elections and local politics after warning of Russian-sponsored espionage and sabotage. Russia has repeatedly denied interfering in foreign elections.

Last year, the Commission opened formal proceedings against social media firm TikTok, owned by China's ByteDance, over its suspected failure to limit election interference, notably in ⁠the Romanian presidential vote in November 2024.

Poland called on the Commission to initiate proceedings in connection with suspected breaches of the bloc's sweeping Digital Services Act, which regulates how the world's biggest social media companies operate in Europe.

Under the Act, large internet platforms like X, Facebook, TikTok and others must moderate and remove harmful content like hate speech, racism or xenophobia. If they do not, the Commission can impose fines of up to 6% ‌of their worldwide annual turnover.