Nvidia Rivals Focus on Building a Different Kind of Chip to Power AI Products

The NVIDIA logo is seen near a computer motherboard in this illustration taken January 8, 2024. (Reuters)
The NVIDIA logo is seen near a computer motherboard in this illustration taken January 8, 2024. (Reuters)
TT

Nvidia Rivals Focus on Building a Different Kind of Chip to Power AI Products

The NVIDIA logo is seen near a computer motherboard in this illustration taken January 8, 2024. (Reuters)
The NVIDIA logo is seen near a computer motherboard in this illustration taken January 8, 2024. (Reuters)

Building the current crop of artificial intelligence chatbots has relied on specialized computer chips pioneered by Nvidia, which dominates market and made itself the poster child of the AI boom.

But the same qualities that make those graphics processor chips, or GPUs, so effective at creating powerful AI systems from scratch make them less efficient at putting AI products to work.

That's opened up the AI chip industry to rivals who think they can compete with Nvidia in selling so-called AI inference chips that are more attuned to the day-to-day running of AI tools and designed to reduce some of the huge computing costs of generative AI.

“These companies are seeing opportunity for that kind of specialized hardware,” said Jacob Feldgoise, an analyst at Georgetown University's Center for Security and Emerging Technology. “The broader the adoption of these models, the more compute will be needed for inference and the more demand there will be for inference chips.”

What is AI inference? It takes a lot of computing power to make an AI chatbot. It starts with a process called training or pretraining — the “P” in ChatGPT — that involves AI systems “learning” from the patterns of huge troves of data. GPUs are good at doing that work because they can run many calculations at a time on a network of devices in communication with each other.

However, once trained, a generative AI tool still needs chips to do the work — such as when you ask a chatbot to compose a document or generate an image. That's where inferencing comes in. A trained AI model must take in new information and make inferences from what it already knows to produce a response.

GPUs can do that work, too. But it can be a bit like taking a sledgehammer to crack a nut.

“With training, you’re doing a lot heavier, a lot more work. With inferencing, that’s a lighter weight,” said Forrester analyst Alvin Nguyen.

That's led startups like Cerebras, Groq and d-Matrix as well as Nvidia's traditional chipmaking rivals — such as AMD and Intel — to pitch more inference-friendly chips as Nvidia focuses on meeting the huge demand from bigger tech companies for its higher-end hardware.

Inside an AI inference chip lab D-Matrix, which is launching its first product this week, was founded in 2019 — a bit late to the AI chip game, as CEO Sid Sheth explained during a recent interview at the company’s headquarters in Santa Clara, California, the same Silicon Valley city that's also home to AMD, Intel and Nvidia.

“There were already 100-plus companies. So when we went out there, the first reaction we got was ‘you’re too late,’” he said. The pandemic's arrival six months later didn't help as the tech industry pivoted to a focus on software to serve remote work.

Now, however, Sheth sees a big market in AI inferencing, comparing that later stage of machine learning to how human beings apply the knowledge they acquired in school.

“We spent the first 20 years of our lives going to school, educating ourselves. That’s training, right?” he said. “And then the next 40 years of your life, you kind of go out there and apply that knowledge — and then you get rewarded for being efficient.”

The product, called Corsair, consists of two chips with four chiplets each, made by Taiwan Semiconductor Manufacturing Company — the same manufacturer of most of Nvidia's chips — and packaged together in a way that helps to keep them cool.

The chips are designed in Santa Clara, assembled in Taiwan and then tested back in California. Testing is a long process and can take six months — if anything is off, it can be sent back to Taiwan.

D-Matrix workers were doing final testing on the chips during a recent visit to a laboratory with blue metal desks covered with cables, motherboards and computers, with a cold server room next door.

Who wants AI inference chips? While tech giants like Amazon, Google, Meta and Microsoft have been gobbling up the supply of costly GPUs in a race to outdo each other in AI development, makers of AI inference chips are aiming for a broader clientele.

Forrester's Nguyen said that could include Fortune 500 companies that want to make use of new generative AI technology without having to build their own AI infrastructure. Sheth said he expects a strong interest in AI video generation.

“The dream of AI for a lot of these enterprise companies is you can use your own enterprise data,” Nguyen said. “Buying (AI inference chips) should be cheaper than buying the ultimate GPUs from Nvidia and others. But I think there’s going to be a learning curve in terms of integrating it.”

Feldgoise said that, unlike training-focused chips, AI inference work prioritizes how fast a person will get a chatbot's response.

He said another whole set of companies is developing AI hardware for inference that can run not just in big data centers but locally on desktop computers, laptops and phones.

Why does this matter? Better-designed chips could bring down the huge costs of running AI to businesses. That could also affect the environmental and energy costs for everyone else.

Sheth says the big concern right now is, “are we going to burn the planet down in our quest for what people call AGI — human-like intelligence?”

It’s still fuzzy when AI might get to the point of artificial general intelligence — predictions range from a few years to decades. But, Sheth notes, only a handful of tech giants are on that quest.

“But then what about the rest?” he said. “They cannot be put on the same path.”

The other set of companies don’t want to use very large AI models — it’s too costly and uses too much energy.

“I don’t know if people truly, really appreciate that inference is actually really going to be a much bigger opportunity than training. I don’t think they appreciate that. It’s still training that is really grabbing all the headlines,” Sheth said.



India Eyes $200B in Data Center Investments as It Ramps Up Its AI Hub Ambitions

FILE -Google CEO Sundar Pichai, right, interacts with India's Minister for Information and Technology Ashwini Vaishnaw during Google for India 2022 event in New Delhi, Dec. 19, 2022. (AP Photo/Manish Swarup), File)
FILE -Google CEO Sundar Pichai, right, interacts with India's Minister for Information and Technology Ashwini Vaishnaw during Google for India 2022 event in New Delhi, Dec. 19, 2022. (AP Photo/Manish Swarup), File)
TT

India Eyes $200B in Data Center Investments as It Ramps Up Its AI Hub Ambitions

FILE -Google CEO Sundar Pichai, right, interacts with India's Minister for Information and Technology Ashwini Vaishnaw during Google for India 2022 event in New Delhi, Dec. 19, 2022. (AP Photo/Manish Swarup), File)
FILE -Google CEO Sundar Pichai, right, interacts with India's Minister for Information and Technology Ashwini Vaishnaw during Google for India 2022 event in New Delhi, Dec. 19, 2022. (AP Photo/Manish Swarup), File)

India is hoping to garner as much as $200 billion in investments for data centers over the next few years as it scales up its ambitions to become a hub for artificial intelligence, the country’s minister for electronics and information technology said Tuesday.

The investments underscore the reliance of tech titans on India as a key technology and talent base in the global race for AI dominance. For New Delhi, they bring in high-value infrastructure and foreign capital at a scale that can accelerate its digital transformation ambitions.

The push comes as governments worldwide race to harness AI's economic potential while grappling with job disruption, regulation and the growing concentration of computing power in a few rich countries and companies.

“Today, India is being seen as a trusted AI partner to the Global South nations seeking open, affordable and development-focused solutions,” Ashwini Vaishnaw told The Associated Press in an email interview, as New Delhi hosts a major AI Impact Summit this week drawing participation from at least 20 global leaders and a who’s who of the tech industry.

In October, Google announced a $15 billion investment plan in India over the next five years to establish its first artificial intelligence hub in the South Asian country. Microsoft followed two months later with its biggest-ever Asia investment announcement of $17.5 billion to advance India’s cloud and artificial intelligence infrastructure over the next four years.

Amazon too has committed $35 billion investment in India by 2030 to expand its business, specifically targeting AI-driven digitization. The cumulative investments are part of $200 billion in investments that are in the pipeline and New Delhi hopes would flow in.

Vaishnaw said India’s pitch is that artificial intelligence must deliver measurable impacts at scale rather than remain an elite technology.

“A trusted AI ecosystem will attract investment and accelerate adoption,” he said, adding that a central pillar of India’s strategy to capitalize on the use of AI is building infrastructure.

The government recently announced a long-term tax holiday for data centers as it hopes to provide policy certainty and attract global capital.

Vaishnaw said the government has already operationalized a shared computing facility with more than 38,000 graphics processing units, or GPUs, allowing startups, researchers and public institutions to access high-end computing without heavy upfront costs.

“AI must not become exclusive. It must remain widely accessible,” he said.

Alongside the infrastructure drive, India is backing the development of sovereign foundational AI models trained on Indian languages and local contexts. Some of these models meet global benchmarks and in certain tasks rival widely used large language models, Vaishnaw said.

India is also seeking a larger role in shaping how AI is built and deployed globally as the country doesn’t see itself strictly as a “rule maker or rule taker,” according to Vaishnaw, but an active participant in setting practical, workable norms while expanding its AI services footprint worldwide.

“India will become a major provider of AI services in the near future,” he said, describing a strategy that is “self-reliant yet globally integrated” across applications, models, chips, infrastructure and energy.

Investor confidence is another focus area for New Delhi as global tech funding becomes more cautious.

Vaishnaw said the technology’s push is backed by execution, pointing to the Indian government's AI Mission program which emphasizes sector specific solutions through public-private partnerships.

The government is also betting on reskilling its workforce as global concerns grow that AI could disrupt white collar and technology jobs. New Delhi is scaling AI education across universities, skilling programs and online platforms to build a large AI-ready talent pool, the minister said.

Widespread 5G connectivity across the country and a young, tech-savvy population are expected to help with the adoption of AI at a faster pace, he added.

Balancing innovation with safeguards remains a challenge though, as AI expands into sensitive sectors such as governance, health care and finance.

Vaishnaw outlined a fourfold strategy that includes implementable global frameworks, trusted AI infrastructure, regulation of harmful misinformation and stronger human and technical capacity to hedge the impact.

“The future of AI should be inclusive, distributed and development-focused,” he said.


Report: SpaceX Competing to Produce Autonomous Drone Tech for Pentagon 

The SpaceX logo is seen in this illustration taken, March 10, 2025. (Reuters)
The SpaceX logo is seen in this illustration taken, March 10, 2025. (Reuters)
TT

Report: SpaceX Competing to Produce Autonomous Drone Tech for Pentagon 

The SpaceX logo is seen in this illustration taken, March 10, 2025. (Reuters)
The SpaceX logo is seen in this illustration taken, March 10, 2025. (Reuters)

Elon Musk's SpaceX and its wholly-owned subsidiary xAI are competing in a secret new Pentagon contest to produce voice-controlled, autonomous drone swarming technology, Bloomberg News reported on Monday, citing people familiar with the matter.

SpaceX, xAI and the Pentagon's defense innovation unit did not immediately respond to requests for comment. Reuters could not independently verify the report.

Texas-based SpaceX recently acquired xAI in a deal that combined Musk's major space and defense contractor with the billionaire entrepreneur's artificial intelligence startup. It occurred ahead of SpaceX's planned initial public offering this year.

Musk's companies are reportedly among a select few chosen to participate in the $100 million prize challenge initiated in January, according to the Bloomberg report.

The six-month competition aims to produce advanced swarming technology that can translate voice commands into digital instructions and run multiple drones, the report said.

Musk was among a group of AI and robotics researchers who wrote an open letter in 2015 that advocated a global ban on “offensive autonomous weapons,” arguing against making “new tools for killing people.”

The US also has been seeking safe and cost-effective ways to neutralize drones, particularly around airports and large sporting events - a concern that has become more urgent ahead of the FIFA World Cup and America250 anniversary celebrations this summer.

The US military, along with its allies, is now racing to deploy the so-called “loyal wingman” drones, an AI-powered aircraft designed to integrate with manned aircraft and anti-drone systems to neutralize enemy drones.

In June 2025, US President Donald Trump issued the Executive Order (EO) “Unleashing American Drone Dominance” which accelerated the development and commercialization of drone and AI technologies.


SVC Develops AI Intelligence Platform to Strengthen Private Capital Ecosystem

The platform offers customizable analytical dashboards that deliver frequent updates and predictive insights- SPA
The platform offers customizable analytical dashboards that deliver frequent updates and predictive insights- SPA
TT

SVC Develops AI Intelligence Platform to Strengthen Private Capital Ecosystem

The platform offers customizable analytical dashboards that deliver frequent updates and predictive insights- SPA
The platform offers customizable analytical dashboards that deliver frequent updates and predictive insights- SPA

Saudi Venture Capital Company (SVC) announced the launch of its proprietary intelligence platform, Aian, developed in-house using Saudi national expertise to enhance its institutional role in developing the Kingdom’s private capital ecosystem and supporting its mandate as a market maker guided by data-driven growth principles.

According to a press release issued by the SVC today, Aian is a custom-built AI-powered market intelligence capability that transforms SVC’s accumulated institutional expertise and detailed private market data into structured, actionable insights on market dynamics, sector evolution, and capital formation. The platform converts institutional memory into compounding intelligence, enabling decisions that integrate both current market signals and long-term historical trends, SPA reported.

Deputy CEO and Chief Investment Officer Nora Alsarhan stated that as Saudi Arabia’s private capital market expands, clarity, transparency, and data integrity become as critical as capital itself. She noted that Aian represents a new layer of national market infrastructure, strengthening institutional confidence, enabling evidence-based decision-making, and supporting sustainable growth.

By transforming data into actionable intelligence, she said, the platform reinforces the Kingdom’s position as a leading regional private capital hub under Vision 2030.

She added that market making extends beyond capital deployment to shaping the conditions under which capital flows efficiently, emphasizing that the next phase of market development will be driven by intelligence and analytical insight alongside investment.

Through Aian, SVC is building the knowledge backbone of Saudi Arabia’s private capital ecosystem, enabling clearer visibility, greater precision in decision-making, and capital formation guided by insight rather than assumption.

Chief Strategy Officer Athary Almubarak said that in private capital markets, access to reliable insight increasingly represents the primary constraint, particularly in emerging and fast-scaling markets where disclosures vary and institutional knowledge is fragmented.

She explained that for development-focused investment institutions, inconsistent data presents a structural challenge that directly impacts capital allocation efficiency and the ability to crowd in private investment at scale.

She noted that SVC was established to address such market frictions and that, as a government-backed investor with an explicit market-making mandate, its role extends beyond financing to building the enabling environment in which private capital can grow sustainably.

By integrating SVC’s proprietary portfolio data with selected external market sources, Aian enables continuous consolidation and validation of market activity, producing a dynamic representation of capital deployment over time rather than relying solely on static reporting.

The platform offers customizable analytical dashboards that deliver frequent updates and predictive insights, enabling SVC to identify priority market gaps, recalibrate capital allocation, design targeted ecosystem interventions, and anchor policy dialogue in evidence.

The release added that Aian also features predictive analytics capabilities that anticipate upcoming funding activity, including projected investment rounds and estimated ticket sizes. In addition, it incorporates institutional benchmarking tools that enable structured comparisons across peers, sectors, and interventions, supporting more precise, data-driven ecosystem development.