Nvidia Rivals Focus on Building a Different Kind of Chip to Power AI Products

The NVIDIA logo is seen near a computer motherboard in this illustration taken January 8, 2024. (Reuters)
The NVIDIA logo is seen near a computer motherboard in this illustration taken January 8, 2024. (Reuters)
TT

Nvidia Rivals Focus on Building a Different Kind of Chip to Power AI Products

The NVIDIA logo is seen near a computer motherboard in this illustration taken January 8, 2024. (Reuters)
The NVIDIA logo is seen near a computer motherboard in this illustration taken January 8, 2024. (Reuters)

Building the current crop of artificial intelligence chatbots has relied on specialized computer chips pioneered by Nvidia, which dominates market and made itself the poster child of the AI boom.

But the same qualities that make those graphics processor chips, or GPUs, so effective at creating powerful AI systems from scratch make them less efficient at putting AI products to work.

That's opened up the AI chip industry to rivals who think they can compete with Nvidia in selling so-called AI inference chips that are more attuned to the day-to-day running of AI tools and designed to reduce some of the huge computing costs of generative AI.

“These companies are seeing opportunity for that kind of specialized hardware,” said Jacob Feldgoise, an analyst at Georgetown University's Center for Security and Emerging Technology. “The broader the adoption of these models, the more compute will be needed for inference and the more demand there will be for inference chips.”

What is AI inference? It takes a lot of computing power to make an AI chatbot. It starts with a process called training or pretraining — the “P” in ChatGPT — that involves AI systems “learning” from the patterns of huge troves of data. GPUs are good at doing that work because they can run many calculations at a time on a network of devices in communication with each other.

However, once trained, a generative AI tool still needs chips to do the work — such as when you ask a chatbot to compose a document or generate an image. That's where inferencing comes in. A trained AI model must take in new information and make inferences from what it already knows to produce a response.

GPUs can do that work, too. But it can be a bit like taking a sledgehammer to crack a nut.

“With training, you’re doing a lot heavier, a lot more work. With inferencing, that’s a lighter weight,” said Forrester analyst Alvin Nguyen.

That's led startups like Cerebras, Groq and d-Matrix as well as Nvidia's traditional chipmaking rivals — such as AMD and Intel — to pitch more inference-friendly chips as Nvidia focuses on meeting the huge demand from bigger tech companies for its higher-end hardware.

Inside an AI inference chip lab D-Matrix, which is launching its first product this week, was founded in 2019 — a bit late to the AI chip game, as CEO Sid Sheth explained during a recent interview at the company’s headquarters in Santa Clara, California, the same Silicon Valley city that's also home to AMD, Intel and Nvidia.

“There were already 100-plus companies. So when we went out there, the first reaction we got was ‘you’re too late,’” he said. The pandemic's arrival six months later didn't help as the tech industry pivoted to a focus on software to serve remote work.

Now, however, Sheth sees a big market in AI inferencing, comparing that later stage of machine learning to how human beings apply the knowledge they acquired in school.

“We spent the first 20 years of our lives going to school, educating ourselves. That’s training, right?” he said. “And then the next 40 years of your life, you kind of go out there and apply that knowledge — and then you get rewarded for being efficient.”

The product, called Corsair, consists of two chips with four chiplets each, made by Taiwan Semiconductor Manufacturing Company — the same manufacturer of most of Nvidia's chips — and packaged together in a way that helps to keep them cool.

The chips are designed in Santa Clara, assembled in Taiwan and then tested back in California. Testing is a long process and can take six months — if anything is off, it can be sent back to Taiwan.

D-Matrix workers were doing final testing on the chips during a recent visit to a laboratory with blue metal desks covered with cables, motherboards and computers, with a cold server room next door.

Who wants AI inference chips? While tech giants like Amazon, Google, Meta and Microsoft have been gobbling up the supply of costly GPUs in a race to outdo each other in AI development, makers of AI inference chips are aiming for a broader clientele.

Forrester's Nguyen said that could include Fortune 500 companies that want to make use of new generative AI technology without having to build their own AI infrastructure. Sheth said he expects a strong interest in AI video generation.

“The dream of AI for a lot of these enterprise companies is you can use your own enterprise data,” Nguyen said. “Buying (AI inference chips) should be cheaper than buying the ultimate GPUs from Nvidia and others. But I think there’s going to be a learning curve in terms of integrating it.”

Feldgoise said that, unlike training-focused chips, AI inference work prioritizes how fast a person will get a chatbot's response.

He said another whole set of companies is developing AI hardware for inference that can run not just in big data centers but locally on desktop computers, laptops and phones.

Why does this matter? Better-designed chips could bring down the huge costs of running AI to businesses. That could also affect the environmental and energy costs for everyone else.

Sheth says the big concern right now is, “are we going to burn the planet down in our quest for what people call AGI — human-like intelligence?”

It’s still fuzzy when AI might get to the point of artificial general intelligence — predictions range from a few years to decades. But, Sheth notes, only a handful of tech giants are on that quest.

“But then what about the rest?” he said. “They cannot be put on the same path.”

The other set of companies don’t want to use very large AI models — it’s too costly and uses too much energy.

“I don’t know if people truly, really appreciate that inference is actually really going to be a much bigger opportunity than training. I don’t think they appreciate that. It’s still training that is really grabbing all the headlines,” Sheth said.



New Process for Stable, Long-Lasting Batteries

The image shows a test cell used to fabricate and test the all-solid-state battery developed at PSI. (Paul Scherrer Institute PSI/Mahir Dzambegovic) 
The image shows a test cell used to fabricate and test the all-solid-state battery developed at PSI. (Paul Scherrer Institute PSI/Mahir Dzambegovic) 
TT

New Process for Stable, Long-Lasting Batteries

The image shows a test cell used to fabricate and test the all-solid-state battery developed at PSI. (Paul Scherrer Institute PSI/Mahir Dzambegovic) 
The image shows a test cell used to fabricate and test the all-solid-state battery developed at PSI. (Paul Scherrer Institute PSI/Mahir Dzambegovic) 

Researchers at the Paul Scherrer Institute PSI have achieved a breakthrough on the path to practical application of lithium metal all-solid-state batteries.

The team expects the next generation of batteries to store more energy, are safer to operate, and charge faster than conventional lithium-ion batteries.

The team has reported these results in the journal Advanced Science.

All-solid-state batteries are considered a promising solution for electromobility, mobile electronics, and stationary energy storage – in part because they do not require flammable liquid electrolytes and therefore are inherently safer than conventional lithium-ion batteries.

Two key problems, however, stand in the way of market readiness: On the one hand, the formation of lithium dendrites at the anode remains a critical point.

On the other hand, an electrochemical instability – at the interface between the lithium metal anode and the solid electrolyte – can impair the battery’s long-term performance and reliability.

To overcome these two obstacles, the team led by Mario El Kazzi, head of the Battery Materials and Diagnostics group at the Paul Scherrer Institute PSI, developed a new production process:

“We combined two approaches that, together, both densify the electrolyte and stabilize the interface with the lithium,” the scientist explained.

Central to the PSI study is the argyrodite type LPSCl, a sulphide-based solid electrolyte made of lithium, phosphorus, and sulphur. The mineral exhibits high lithium-ion conductivity, enabling rapid ion transport within the battery – a crucial prerequisite for high performance and efficient charging processes.

To densify argyrodite into a homogeneous electrolyte, El Kazzi and his team did incorporate the temperature factor, but in a more careful way: Instead of the classic sintering process, they chose a gentler approach in which the mineral was compressed under moderate pressure and at a moderate temperature of only about 80 degrees Celsius.

The result is a compact, dense microstructure resistant to the penetration of lithium dendrites. Already, in this form, the solid electrolyte is ideally suited for rapid lithium-ion transport.

To ensure reliable operation even at high current densities, such as those encountered during rapid charging and discharging, the all-solid-state cell required further modification.

For this purpose, a coating of lithium fluoride (LiF), only 65 nanometres thick, was evaporated under vacuum and applied uniformly to the lithium surface – serving as a ultra-thin passivation layer at the interface between the anode and the solid electrolyte.

In laboratory tests with button cells, the battery demonstrated extraordinary performance under demanding conditions.

“Its cycle stability at high voltage was remarkable,” said doctoral candidate Jinsong Zhang, lead author of the study.

After 1,500 charge and discharge cycles, the cell still retained approximately 75% of its original capacity.

This means that three-quarters of the lithium ions were still migrating from the cathode to the anode. “An outstanding result. These values are among the best reported to date.”

Zhang therefore sees a good chance that all-solid-state batteries could soon surpass conventional lithium-ion batteries with liquid electrolyte in terms of energy density and durability.

Thus El Kazzi and his team have demonstrated for the first time that the combination of solid electrolyte mild sintering and a thin passivation layer on lithium anode effectively suppresses both dendrite formation and interfacial instability.

This combined solution marks an important advance for all-solid-state battery research – not least because it offers ecological and economic advantages: Due to the low temperatures, the process saves energy and therefore costs.

“Our approach is a practical solution for the industrial production of argyrodite-based all-solid-state batteries,” said El Kazzi. “A few more adjustments – and we could get started.”


Meta Urges Australia to Change Teen Social Media Ban

Meta has called for Australia's social media for under-16s to target app stores. Saeed KHAN / AFP
Meta has called for Australia's social media for under-16s to target app stores. Saeed KHAN / AFP
TT

Meta Urges Australia to Change Teen Social Media Ban

Meta has called for Australia's social media for under-16s to target app stores. Saeed KHAN / AFP
Meta has called for Australia's social media for under-16s to target app stores. Saeed KHAN / AFP

Tech giant Meta urged Australia on Monday to rethink its world-first social media ban for under-16s, while reporting that it has blocked more than 544,000 accounts under the new law.

Australia has required big platforms including Meta, TikTok and YouTube to stop underage users from holding accounts since the legislation came into force on December 10 last year.

Companies face fines of Aus $49.5 million (US$33 million) if they fail to take "reasonable steps" to comply.

Billionaire Mark Zuckerberg's Meta said it had removed 331,000 underage accounts from Instagram, 173,000 from Facebook, and 40,000 from Threads in the week to December 11.

The company said it was committed to complying with the law.

"That said, we call on the Australian government to engage with industry constructively to find a better way forward, such as incentivizing all of industry to raise the standard in providing safe, privacy-preserving, age appropriate experiences online, instead of blanket bans," it said in statement.

Meta renewed an earlier call for app stores to be required to verify people's ages and get parental approval before under-16s can download an app.

This was the only way to avoid a "whack-a-mole" race to stop teens migrating to new apps to avoid the ban, the company said.

The government said it was holding social media companies to account for the harm they cause young Australians.

"Platforms like Meta collect a huge amount of data on their users for commercial purposes. They can and must use that information to comply with Australian law and ensure people under 16 are not on their platforms," a government spokesperson said.

Meta said parents and experts were worried about the ban isolating young people from online communities, and driving some to less regulated apps and darker corners of the internet.

Initial impacts of the legislation "suggest it is not meeting its objectives of increasing the safety and well-being of young Australians", it said.

While raising concern over the lack of an industry standard for determining age online, Meta said its compliance with the Australian law would be a "multilayered process".

Since the ban, the California-based firm said it had helped found the OpenAge Initiative, a non-profit group that has launched age-verification tools called AgeKeys to be used with participating platforms.


China Is Closing in on US Technology Lead Despite Constraints, AI Researchers Say

 Visitors look at robots on display at robotics company Unitree's first retail store in Beijing in January 9, 2026. (AFP)
Visitors look at robots on display at robotics company Unitree's first retail store in Beijing in January 9, 2026. (AFP)
TT

China Is Closing in on US Technology Lead Despite Constraints, AI Researchers Say

 Visitors look at robots on display at robotics company Unitree's first retail store in Beijing in January 9, 2026. (AFP)
Visitors look at robots on display at robotics company Unitree's first retail store in Beijing in January 9, 2026. (AFP)

China can narrow its technological gap with the US driven by growing risk-taking and innovation, though the lack of advanced chipmaking tools is hobbling the sector, the country's leading artificial intelligence researchers said on Saturday.

China's so-called "AI tiger" startups MiniMax and Zhipu AI had strong debuts on the Hong Kong Stock Exchange this week, reflecting growing confidence in the sector as Beijing fast-tracks AI and chip listings to bolster domestic alternatives to advanced US technology.

Yao Shunyu, a former senior researcher at ChatGPT maker OpenAI ‌who was named ‌technology giant Tencent's chief AI scientist in December, ‌said ⁠there was a ‌high likelihood of a Chinese firm becoming the world's leading AI company in the next three to five years but said the lack of advanced chipmaking machines was the main technical hurdle.

"Currently, we have a significant advantage in electricity and infrastructure. The main bottlenecks are production capacity, including lithography machines, and the software ecosystem," Yao said at an AI conference in Beijing.

China has completed a working prototype of an extreme-ultraviolet lithography ⁠machine potentially capable of producing cutting-edge semiconductor chips that rival the West's, Reuters reported last month. However, the ‌machine has not yet produced working chips and may ‍not do so until 2030, people with ‍knowledge of the matter told Reuters.

MIND THE INVESTMENT GAP

Yao and other ‍Chinese industry leaders at the Beijing conference on Saturday also acknowledged that the US maintains an advantage in computing power due to its hefty investments in infrastructure.

"The US computer infrastructure is likely one to two orders of magnitude larger than ours. But I see that whether it's OpenAI or other platforms, they're investing heavily in next-generation research," said Lin Junyang, technical lead for Alibaba's flagship Qwen large language model.

"We, ⁠on the other hand, are relatively strapped for cash; delivery alone likely consumes the majority of our computer infrastructure," Lin said during a panel discussion at the AGI-Next Frontier Summit held by the Beijing Key Laboratory of Foundational Models at Tsinghua University.

Lin said China's limited resources have spurred its researchers to be innovative, particularly through algorithm-hardware co-design, which enables AI firms to run large models on smaller, inexpensive hardware.

Tang Jie, founder of Zhipu AI which raised HK$4.35 billion in its IPO, also highlighted the willingness of younger Chinese AI entrepreneurs to embrace high-risk ventures - a trait traditionally associated with Silicon Valley - as a positive development.

"I think if we can improve this environment, ‌allowing more time for these risk-taking, intelligent individuals to engage in innovative endeavors ... this is something our government and the country can help improve," said Tang.