Nvidia pays $1 billion for a stake in Nokia to collaborate on AI networking solutions

This is not only astonishing but unheard of:  the world’s largest and most popular fabless semiconductor company –Nvidia– taking a $1 billion stake in a telco/reinvented data center connectivity company-Nokia.

Indeed, GPU king Nvidia will pay $1 billion for a stake of 2.9% in Nokia as part of a deal focused on AI and data centers, the Finnish telecom equipment maker said on Tuesday as its shares hit their highest level in nearly a decade on hope for AI to lift their business revenue and profits. The nonexclusive partnership and the investment will make Nvidia the second-largest shareholder in Nokia. Nokia said it will issue 166,389,351 new shares for Nvidia, which the U.S. company will subscribe to at $6.01 per share.

Nokia said the companies will collaborate on artificial intelligence networking solutions and explore opportunities to include its data center communications products in Nvidia’s future AI infrastructure plans. Nokia and its Swedish rival Ericsson both make networking equipment for connectivity inside (intra-) data centers and between (inter-) data centers and have been benefiting from increased AI use.

Summary:

  • NVIDIA and Nokia to establish a strategic partnership to enable accelerated development and deployment of next generation AI native mobile networks and AI networking infrastructure.
  • NVIDIA introduces NVIDIA Arc Aerial RAN Computer, a 6G-ready telecommunications computing platform.
  • Nokia to expand its global access portfolio with new AI-RAN product based on NVIDIA platform.
  • T-Mobile U.S. is working with Nokia and NVIDIA to integrate AI-RAN technologies into its 6G development process.
  • Collaboration enables new AI services and improved consumer experiences to support explosive growth in mobile AI traffic.
  • Dell Technologies provides PowerEdge servers to power new AI-RAN solution.
  • Partnership marks turning point for the industry, paving the way to AI-native 6G by taking AI-RAN to innovation and commercialization at a global scale.

In some respects, this new partnership competes with Nvidia’s own data center connectivity solutions from its Mellanox Technologies division, which it acquired for $6.9 billion in 2019.  Meanwhile, Nokia now claims to have worked on a redesign to ensure its RAN software is compatible with Nvidia’s compute unified device architecture (CUDA) platform, meaning it can run on Nvidia’s GPUs. Nvidia has also modified its hardware offer, creating capacity cards that will slot directly into Nokia’s existing AirScale baseband units at mobile sites.

Having dethroned Intel several years ago, Nvidia now has a near-monopoly in supplying GPU chips for data centers and has partnered with companies ranging from OpenAI to Microsoft.  AMD is a distant second but is gaining ground in the data center GPU space as is ARM Ltd with its RISC CPU cores. Capital expenditure on data center infrastructure is expected to exceed $1.7 trillion by 2030, consulting firm McKinsey, largely because of the expansion of AI.

Nvidia CEO Jensen Huang said the deal would help make the U.S. the center of the next revolution in 6G. “Thank you for helping the United States bring telecommunication technology back to America,” Huang said in a speech in Washington, addressing Nokia CEO Justin Hotard (x-Intel). “The key thing here is it’s American technology delivering the base capability, which is the accelerated computing stack from Nvidia, now purpose-built for mobile,” Hotard told Reuters in an interview.  “Jensen and I have been talking for a little bit and I love the pace at which Nvidia moves,” Hotard said. “It’s a pace that I aspire for us to move at Nokia.”  He expects the new equipment to start contributing to revenue from 2027 as it goes into commercial deployment, first with 5G, followed by 6G after 2030.

Nvidia has been on a spending spree in recent weeks. The company in September pledged to invest $5 billion in beleaguered chip maker Intel. The investment pairs the world’s most valuable company, which has been a darling of the AI boom, with a chip maker that has almost completely fallen out of the AI conversation.

Later that month, Nvidia said it planned to invest up to $100 billion in OpenAI over an unspecified period that will likely span at least a few years. The partnership includes plans for an enormous data-center build-out and will allow OpenAI to build and deploy at least 10 gigawatts of Nvidia systems.

…………………………………………………………………………………………………………………………………………………………………………………………………………………………………

Tech Details:

Nokia uses Marvell Physical Layer (1) baseband chips for many of its products. Among other things, this ensured Nokia had a single software stack for all its virtual and purpose-built RAN products. Pallavi Mahajan, Nokia’s recently joined chief technology and AI officer recently told Light Reading that their software could easily adapt to run on Nvidia’s GPUs: “We built a hardware abstraction layer so that whether you are on Marvell, whether you are on any of the x86 servers or whether you are on GPUs, the abstraction takes away from that complexity, and the software is still the same.”

Fully independent software has been something of a Holy Grail for the entire industry. It would have ramifications for the whole market and its economics. Yet Nokia has conceivably been able to minimize the effort required to put its Layer 1 and specific higher-layer functions on a GPU. “There are going to be pieces of the software that are going to leverage on the accelerated compute,” said Mahajan. “That’s where we will bring in the CUDA integration pieces. But it’s not the entire software,” she added.  The appeal of Nvidia as an alternative was largely to be found in “the programmability pieces that you don’t have on any other general merchant silicon,” said Mahajan. “There’s also this entire ecosystem, the CUDA ecosystem, that comes in.” One of Nvidia’s most eye-catching recent moves is the decision to “open source” Aerial, its own CUDA-based RAN software framework, so that other developers can tinker, she says. “What it now enables is the entire ecosystem to go out and contribute.”

…………………………………………………………………………………………………………………………………………………………………………………………………………………………………

Quotes:

“Telecommunications is a critical national infrastructure — the digital nervous system of our economy and security,” said Jensen Huang, founder and CEO of NVIDIA. “Built on NVIDIA CUDA and AI, AI-RAN will revolutionize telecommunications — a generational platform shift that empowers the United States to regain global leadership in this vital infrastructure technology. Together with Nokia and America’s telecom ecosystem, we’re igniting this revolution, equipping operators to build intelligent, adaptive networks that will define the next generation of global connectivity.”

“The next leap in telecom isn’t just from 5G to 6G — it’s a fundamental redesign of the network to deliver AI-powered connectivity, capable of processing intelligence from the data center all the way to the edge. Our partnership with NVIDIA, and their investment in Nokia, will accelerate AI-RAN innovation to put an AI data center into everyone’s pocket,” said Justin Hotard, President and CEO of Nokia. “We’re proud to drive this industry transformation with NVIDIA, Dell Technologies, and T-Mobile U.S., our first AI-RAN deployments in T-Mobile’s network will ensure America leads in the advanced connectivity that AI needs.”

……………………………………………………………………………………………………………………………………………………………………………………

Editor’s Notes:

1.  In more advanced 5G networks, Physical Layer functions have demanded the support of custom silicon, or “accelerators.”  A technique known as “lookaside,” favored by Ericsson and Samsung, uses an accelerator for only a single problematic Layer 1 task – forward error correction – and keeps everything else on the CPU. Nokia prefers the “inline” approach, which puts the whole of Layer 1 on the accelerator.

2. The huge AI-RAN push that Nvidia started with the formation of the AI-RAN Alliance in early 2024 has not met with an enthusiastic telco response so far. Results from Nokia as well as Ericsson show wireless network operators are spending less on 5G rollouts than they were in the early 2020s. Telco numbers indicate the appetite for smartphone and other mobile data services has not produced any sales growth. As companies prioritize efficiency above all else, baseband units that include Marvell and Nvidia cards may seem too expensive.

……………………………………………………………………………………………………………………………………………………………………………………….

Other Opinions and Quotes:

Nvidia chips are likely to be more expensive, said Mads Rosendal, analyst at Danske Bank Credit Research, but the proposed partnership would be mutually beneficial, given Nvidia’s large share in the U.S. data center market.

“This is a strong endorsement of Nokia’s capabilities,” said PP Foresight analyst Paolo Pescatore. “Next-generation networks, such as 6G, will play a significant role in enabling new AI-powered experiences,” he added.

Iain Morris, International Editor, Light Reading: “Layer 1 control software runs on ARM RISC CPU cores in both Marvell and Nvidia technologies. The bigger differences tend to be in the hardware acceleration “kernels,” or central components, which have some unique demands. Yet Nokia has been working to put as much as it possibly can into a bucket of common software. Regardless, if Nvidia is effectively paying for all this with its $1 billion investment, the risks for Nokia may be small………….Nokia’s customers will in future have an AI-RAN choice that limits or even shrinks the floorspace for Marvell. The development also points to more subtle changes in Nokia’s thinking. The message earlier this year was that Nokia did not require a GPU to implement AI for RAN, whereby machine-generated algorithms help to improve network performance and efficiency. Marvell had that covered because it had incorporated AI and machine-learning technologies into the baseband chips used by Nokia.”

“If you start doing inline, you typically get much more locked into the hardware,” said Per Narvinger, the president of Ericsson’s mobile networks business group, on a recent analyst call. During its own trials with Nvidia, Ericsson said it was effectively able to redeploy virtual RAN software written for Intel’s x86 CPUs on the Grace CPU with minimal changes, leaving the GPU only as a possible option for the FEC accelerator.  Putting the entire Layer 1 on a GPU would mean “you probably also get more tightly into that specific implementation,” said Narvinger. “Where does it really benefit from having that kind of parallel compute system?”

………………………………………………………………………………………………………………………………………………….

Separately, Nokia and Nvidia will partner with T-Mobile U.S. to develop and test AI RAN technologies for developing 6G, Nokia said in its press release.  Trials are expected to begin in 2026, focused on field validation of performance and efficiency gains for customers.

References:

https://nvidianews.nvidia.com/news/nvidia-nokia-ai-telecommunications

https://www.reuters.com/world/europe/nvidia-make-1-billion-investment-finlands-nokia-2025-10-28/

https://www.lightreading.com/5g/nvidia-takes-1b-stake-in-nokia-which-promises-5g-and-6g-overhaul

https://www.wsj.com/business/telecom/nvidia-takes-1-billion-stake-in-nokia-69f75bb6

Highlights of Nokia’s Smart Factory in Oulu, Finland for 5G and 6G innovation

Nokia & Deutsche Bahn deploy world’s first 1900 MHz 5G radio network meeting FRMCS requirements

Will the wave of AI generated user-to/from-network traffic increase spectacularly as Cisco and Nokia predict?

Indosat Ooredoo Hutchison and Nokia use AI to reduce energy demand and emissions

Verizon partners with Nokia to deploy large private 5G network in the UK

184K global tech layoffs in 2025 to date; ~27.3% related to AI replacing workers

As of October, over 184,000 global tech jobs were cut in 2025, according to a report from Silicon Valley Business Journal.  50,184 were directly related to the implementation of artificial intelligence (AI) and automation tools by businesses. Silicon Valley’s AI boom has been pummeling headcounts across major companies in the region — and globally.  U.S. companies accounted for about 123,000 of the layoffs.

These are the 10 tech companies with the most significant mass layoffs since January 2025:

  • Intel: 33,900 layoffs. The company has cited the need to reduce costs and restructure its organization after years of technical and financial setbacks.
  • Microsoft: 19,215 layoffs. The tech giant has conducted multiple rounds of cuts throughout the year across various departments as it prioritizes AI investments.
  • TCS: 12,000 layoffs. As a major IT firm, Tata Consultancy Services’ cuts largely affected mid-level and senior positions, which are becoming redundant due to AI and evolving client demands.
  • Accenture: 11,000 layoffs. The consulting company reduced its headcount as it shifts toward greater automation and AI-driven services.
  • Panasonic: 10,000 layoffs. The Japanese manufacturer announced these job cuts as part of a strategy to improve efficiency and focus on core business areas.
  • IBM: 9,000 layoffs as part of a restructuring effort to shift some roles to India and align the workforce with areas like AI and hybrid cloud. The layoffs were reportedly concentrated in certain teams, including the Cloud Classic division, and impacted locations such as Raleigh, New York, Dallas, and California. 
  • Amazon: 5,555 layoffs. Cuts have impacted various areas, including the Amazon Web Services (AWS) cloud unit and the consumer retail business.
  • Salesforce: 5,000 layoffs. Many of these cuts impacted the customer service division, where AI agents now handle a significant portion of client interactions.
  • STMicro: 5,000 cuts in the next three years, including 2,800 job cuts announced earlier this year, its chief executive said on Wednesday. Around 2,000 employees will leave the Franco-Italian chipmaker due to attrition, bringing the total count with voluntary departures to 5,000, Jean-Marc Chery said at a June 4th event in Paris, hosted by BNP Paribas.
  • Meta: 3,720 layoffs. The company has made multiple rounds of cuts targeting “low-performers” and positions within its AI and virtual reality divisions.  More details below.

……………………………………………………………………………………………………………………………………………………………………..

Image Credit: simplehappyart via Getty Images

……………………………………………………………………………………………………………………………………………………………………..

In a direct contradiction in August, Cisco announced layoffs of 221 employees in the San Francisco Bay Area, affecting roles in Milpitas and San Francisco. This occurred despite strong financial results and the CEO’s previous statement that the company would not cut jobs in favor of AI. The cuts, which included software engineering roles, are part of the company’s broader strategy to streamline operations & focus on AI.

Only days after revealing a partnership with OpenAI, semiconductor maker Broadcom is cutting hundreds of staff in Palo Alto.  For Broadcom, the cuts follow its 2023 acquisition of VMware, which was accompanied by thousands of job cuts as part of a multiyear restructuring effort. Current reports indicate that the company is eliminating additional positions across its sales and account management teams.
“The wave of tech layoffs in 2025 keeps growing — and Broadcom has once again become one of the biggest names in the mix,” said RationalFX analyst Alan Cohen in a statement. “The broader industry climate isn’t helping: a squeeze from tariffs, trade tensions, and weakening demand has forced tech giants to slash costs just when AI automation was supposed to create new jobs — instead, it’s replacing more of them,” Cohen added.
………………………………………………………………………………………………………………………………………………………………..
Meta followed suit on October 22nd, announcing 600 job cuts within its AI division — both part of a widening wave of tech layoffs tied to automation and artificial intelligence. 700 additional jobs were cut by Meta after the report was published- 600 from its AI Division and 100 from its risk review organization. That group is largely staffed by employees responsible for making sure Meta’s products abide by an agreement with the Federal Trade Commission as well as privacy rules set by world-wide regulatory bodies.
………………………………………………………………………………………………………………………………………………………………..

About two-thirds of all job cuts — roughly 123,000 positions — came from U.S.-based companies, with the remainder spread across mainly Ireland, India and Japan. The report compiles data from WARN notices, TrueUp, TechCrunch and Layoffs.fyi through Oct. 21st.

Several trends are driving the ongoing reduction in tech jobs:
  • Shift to AI and automation: Many companies are restructuring their workforce to focus on AI-centric growth and are automating tasks previously done by human workers, particularly in customer service and quality assurance.
  • Economic headwinds: Ongoing economic uncertainty, inflation, and higher interest rates are prompting tech companies to cut costs and streamline operations.
  • Market corrections: Following a period of rapid over-hiring, many tech companies are now “right-sizing” their staff to become leaner and more efficient.

References:

https://www.bizjournals.com/sanjose/news/2025/10/22/tech-layoffs-ai-automation-broadcom-meta-intel.html

Report: Broadcom Announces Further Job Cuts as Global Tech Layoffs Approach 185,000 in 2025

 

The Tech Industry’s Workforce Crisis: 166,387 layoffs so far in 2025, projected to reach 235K by the end of the year

 

https://www.linkedin.com/posts/edmund-ho-1277b2125_180k-job-cuts-biggest-tech-company-layoffs-activity-7381201561152184320-N5ah/

 

Tech layoffs continue unabated: pink slip season in hard-hit SF Bay Area

HPE cost reduction campaign with more layoffs; 250 AI PoC trials or deployments

High Tech Layoffs Explained: The End of the Free Money Party

Massive layoffs and cost cutting will decimate Intel’s already tiny 5G network business

Big Tech post strong earnings and revenue growth, but cuts jobs along with Telecom Vendors

Telecom layoffs continue unabated as AT&T leads the pack – a growth engine with only 1% YoY growth?

Cisco restructuring plan will result in ~4100 layoffs; focus on security and cloud based products

Cloud Computing Giants Growth Slows; Recession Looms, Layoffs Begin

IBM and Groq Partner to Accelerate Enterprise AI Inference Capabilities

 IBM and Groq [1.] today announced a strategic market and technology partnership designed to give clients immediate access to Groq’s inference technology — GroqCloud, on watsonx Orchestrate – providing clients high-speed AI inference capabilities at a cost that helps accelerate agentic AI deployment. As part of the partnership, Groq and IBM plan to integrate and enhance RedHat open source vLLM technology with Groq’s LPU architecture. IBM Granite models are also planned to be supported on GroqCloud for IBM clients.

………………………………………………………………………………………………………………………………………………….

Note 1. Groq is a privately held company founded by Jonathan Ross in 2016. As a startup, its ownership is distributed among its founders, employees, and a variety of venture capital and institutional investors including BlackRock Private Equity PartnersGroq developed the LPU and GroqCloud to make compute faster and more affordable. The company says it is trusted by over two million developers and teams worldwide and is a core part of the American AI Stack.

NOTE that Grok, a conversational AI assistant developed by Elon Musk’s xAI is a completely different entity.

………………………………………………………………………………………………………………………………………………….

Enterprises moving AI agents from pilot to production still face challenges with speed, cost, and reliability, especially in mission-critical sectors like healthcare, finance, government, retail, and manufacturing. This partnership combines Groq’s inference speed, cost efficiency, and access to the latest open-source models with IBM’s agentic AI orchestration to deliver the infrastructure needed to help enterprises scale.

Powered by its custom LPU, GroqCloud delivers over 5X faster and more cost-efficient inference than traditional GPU systems. The result is consistently low latency and dependable performance, even as workloads scale globally. This is especially powerful for agentic AI in regulated industries.

For example, IBM’s healthcare clients receive thousands of complex patient questions simultaneously. With Groq, IBM’s AI agents can analyze information in real-time and deliver accurate answers immediately to enhance customer experiences and allow organizations to make faster, smarter decisions.

This technology is also being applied in non-regulated industries. IBM clients across retail and consumer packaged goods are using Groq for HR agents to help enhance automation of HR processes and increase employee productivity.

“Many large enterprise organizations have a range of options with AI inferencing when they’re experimenting, but when they want to go into production, they must ensure complex workflows can be deployed successfully to ensure high-quality experiences,” said Rob Thomas, SVP, Software and Chief Commercial Officer at IBM. “Our partnership with Groq underscores IBM’s commitment to providing clients with the most advanced technologies to achieve AI deployment and drive business value.”

“With Groq’s speed and IBM’s enterprise expertise, we’re making agentic AI real for business. Together, we’re enabling organizations to unlock the full potential of AI-driven responses with the performance needed to scale,” said Jonathan Ross, CEO & Founder at Groq. “Beyond speed and resilience, this partnership is about transforming how enterprises work with AI, moving from experimentation to enterprise-wide adoption with confidence, and opening the door to new patterns where AI can act instantly and learn continuously.”

IBM will offer access to GroqCloud’s capabilities starting immediately and the joint teams will focus on delivering the following capabilities to IBM clients, including:

  • High speed and high-performance inference that unlocks the full potential of AI models and agentic AI, powering use cases such as customer care, employee support and productivity enhancement.
  • Security and privacy-focused AI deployment designed to support the most stringent regulatory and security requirements, enabling effective execution of complex workflows.
  • Seamless integration  with IBM’s agentic product, watsonx Orchestrate, providing clients flexibility to adopt purpose-built agentic patterns tailored to diverse use cases.

The partnership also plans to integrate and enhance RedHat open source vLLM technology with Groq’s LPU architecture to offer different approaches to common AI challenges developers face during inference. The solution is expected to enable watsonx to leverage capabilities in a familiar way and let customers stay in their preferred tools while accelerating inference with GroqCloud. This integration will address key AI developer needs, including inference orchestration, load balancing, and hardware acceleration, ultimately streamlining the inference process.

Together, IBM and Groq provide enhanced access to the full potential of enterprise AI, one that is fast, intelligent, and built for real-world impact.

References:

https://www.prnewswire.com/news-releases/ibm-and-groq-partner-to-accelerate-enterprise-ai-deployment-with-speed-and-scale-302588893.html

FT: Scale of AI private company valuations dwarfs dot-com boom

AI adoption to accelerate growth in the $215 billion Data Center market

Big tech spending on AI data centers and infrastructure vs the fiber optic buildout during the dot-com boom (& bust)

Will billions of dollars big tech is spending on Gen AI data centers produce a decent ROI?

Can the debt fueling the new wave of AI infrastructure buildouts ever be repaid?

 

FT: Scale of AI private company valuations dwarfs dot-com boom

The Financial Times reports that ten loss­ mak­ing arti­fi­cial intel­li­gence (AI) start-ups have gained close to $1 trillion in private market valu­ation in the past 12 months, fuel­ling fears about a bubble in private mar­kets that is much greater than the dot com bubble at the end of the 20th century.  OpenAI leads the pack with a $500 billion valuation, but Anthropic and xAI have also seen their val­ues march higher amid a mad scramble to buy into emerging AI com­pan­ies. Smal­ler firms build­ing AI applic­a­tions have also surged, while more estab­lished busi­nesses, like Dat­ab­ricks, have soared after embra­cing the tech­no­logy.

U.S. venture capitalists (VCs) have poured $161 billion into artificial intelligence startups this year — roughly two-thirds of all venture spending, according to PitchBook — even as the technology’s commercial payoff remains elusive. VCs are on track to spend well over $200bn on AI companies this year.

Most of that money has gone to just 10 companies, including OpenAI, Anthropic, Databricks, xAI, Perplexity, Scale AI, and Figure AI, whose combined valuations have swelled by nearly $1 trillion, Financial Times calculations show.  Those AI start-ups are all burning cash with no profits forecasted for many years.

Start-ups with about $5mn in annual recurring revenue, a metric used by fast-growing young businesses to provide a snapshot of their earnings, are seeking valuations of more than $500mn, according to a senior Silicon Valley venture capitalist.

Valuing unproven businesses at 100 times their earnings or more dwarfs the excesses of 2021, he added: “Even during peak Zirp [zero-interest rate policies], these would have been $250mn-$300mn valuations.”

“The market is investing as if all these companies are outliers. That’s generally not the way it works out,” he said. VCs typically expect to lose money on most of their bets, but see one or two pay the rest off many times over.

There will be casualties. Just like there always will be, just like there always is in the tech industry,” said Marc Benioff, co-founder and chief executive of Salesforce, which has invested heavily in AI. He estimates $1tn of investment on AI might be wasted, but that the technology will ultimately yield 10 times that in new value.

“The only way we know how to build great technology is to throw as much against the wall as possible, see what sticks, and then focus on the winners,” he added.

Of course there’s a bubble,” said Hemant Taneja, chief executive of General Catalyst, which raised an $8 billion fund last year and has backed Anthropic and Mistral. “Bubbles align capital and talent around new trends. There’s always some destruction, but they also produce lasting innovation.”

Venture investors have weathered cycles of boom and bust before — from the dot-com crash in 2000 to the software downturn in 2022 — but the current wave of AI funding is unprecedented. In 2000, VCs invested $10.5 billion in internet startups; in 2021, they deployed $135 billion into software firms. This year, they are on pace to exceed $200 billion in AI. “We’ve gone from the doldrums to full-on FOMO,” said one investment executive.

OpenAI and its start-up peers are competing with Meta, Google, Microsoft, Amazon, IBM, and others in a hugely capital-intensive race to train ever-better models, meaning the path to profitability is also likely to be longer than for previous generations of start-ups.

Backers are betting that AI will open multi-trillion-dollar markets, from automated coding to AI friends or companionship. Yet some valuations are testing credulity. Startups generating about $5 million in annual recurring revenue are seeking valuations above $500 million, a Silicon Valley investor said — 100 times revenue, surpassing even the excesses of 2021. “The market is behaving as if every company will be an outlier,” he said. “That’s rarely how it works.”

The enthusiasm has spilled into public markets. Shares of Nvidia, AMD, Broadcom, and Oracle have collectively gained hundreds of billions in market value from their ties to OpenAI. But those gains could unwind quickly if questions about the startup’s mounting losses and financial sustainability persist.

Sebastian Mallaby, author of The Power Law, summed it up beautifully:

“The logic among investors is simple — if we get AGI (Artificial General Intelligence, which would match or exceed human thinking), it’s all worth it. If we don’t, it isn’t…. “It comes down to these articles of faith about Sam’s (Sam Altman of OpenAI) ability to work it out.”

References:

https://www.ft.com/content/59baba74-c039-4fa7-9d63-b14f8b2bb9e2

Big tech spending on AI data centers and infrastructure vs the fiber optic buildout during the dot-com boom (& bust)

Can the debt fueling the new wave of AI infrastructure buildouts ever be repaid?

Amazon’s Jeff Bezos at Italian Tech Week: “AI is a kind of industrial bubble”

Gartner: AI spending >$2 trillion in 2026 driven by hyperscalers data center investments

AI Data Center Boom Carries Huge Default and Demand Risks

Canalys & Gartner: AI investments drive growth in cloud infrastructure spending

 

Huawei to Double Output of Ascend AI chips in 2026; OpenAI orders HBM chips from SK Hynix & Samsung for Stargate UAE project

With sales of Nvidia AI chips restricted in China, Huawei Technologies Inc. plans to make about 600,000 of its 910C Ascend chips next year, roughly double this year’s output, people familiar with the matter told Bloomberg. The China tech behemoth will increase its Ascend product line in 2026 to as many as 1.6 million dies – the basic silicon component that’s packaged as a chip.

Huawei had struggled to get those products to potential customers for much of 2025, because of U.S. sanctions.  Yet if Huawei and its partner Semiconductor Manufacturing International Corp. (SMIC) can hit that ambitious AI chip manufacturing target, it suggest self sufficiency which will remove some of the bottlenecks that’ve hindered not just its AI business.

The projections for 2025 and 2026 include dies that Huawei has in inventory, as well as internal estimates of yields or the rate of failure during production, the people said. Shares in SMIC and rival chipmaker Hua Hong Semiconductor Ltd. gained more than 4% in Hong Kong Tuesday, while the broader market stayed largely unchanged.

Huawei Ascend branding at a trade show i China. Photographer: Ying Tang/Getty Images

Chinese AI companies from Alibaba Group Holding Ltd. to DeepSeek need millions of AI chips to develop and operate AI services. Nvidia alone was estimated to have sold a million H20 chips in 2024.

What Bloomberg Economics Says:

Huawei’s reported plan to double AI-chip output over the next year suggests China is making real progress in working around US export controls. Yet the plan also exposes the limitations imposed by US controls: Node development remains stuck at 7 nanometers, and Huawei will continue to rely on stockpiles of foreign high-bandwidth memory amid a lack of domestic production.

From Beijing’s perspective, Huawei’s production expansion represents another move in an ongoing back-and-forth with the West over semiconductor access and self-sufficiency. The priority remains accelerating indigenization of critical technologies while steadily pushing back against Western controls.

– Michael Deng, analyst

While Huawei’s new AI silicon promises massive performance gains it has several shortcomings, especially the lack of a developer community comparable to Nvidia’s CUDA ecosystem.  A Chinese tech executive said Nvidia’s biggest advantage wasn’t its advanced chips but the ecosystem built around CUDA, its parallel computing architecture and programming model. The exec called for the creation of a Chinese version of CUDA that can be used worldwide. 

Also, Huawei is playing catchup by progressively going open source. It announced last month that its Ascend and AI training toolkit CANN, its Mind development environment and Pangu models would all be open source by year-end.

Huawei chairman Eric Xu said in an interview the company had given the “ecosystem issue” a great deal of thought and regarded the transition to open source as a long-term project. “Why keep it hidden? If it’s widely used, an ecosystem will emerge; if it’s used less, the ecosystem will disappear,” he said.

………………………………………………………………………………………………………………………………………………………………………

At its customer event in Shanghai last month, Huawei revealed that it planned to spend 15 billion Chinese yuan (US$2.1 billion) annually over the next five years on ecosystem development and open source computing.

Xu announced a series of new Ascend chips – the 950, 960 and 970 – to be rolled out over the next three years.  He foreshadowed a new series of massive Atlas SuperPoD clusters – each one a single logical machine made up of multiple physical devices that can work together – and also announced Huawei’s unified bus interconnect protocol, which allows customers to stitch together compute power across multiple data centers. 

Xu acknowledged that Huawei’s single Ascend chips could not match Nvidia’s, but said the SuperPoDs were currently the world’s most powerful and will remain so “for years to come.” But the scale of its SuperPOD architecture points to its other shortcoming – the power consumption of these giant compute arrays. 

………………………………………………………………………………………………………………………………………………………………………….

Separately, OpenAI has made huge memory chip agreements with South Korea’s SK Hynix and Samsung, the world’s two biggest semiconductor memory manufacturers.  The partnership, aimed at locking up HBM ((High Bandwidth Memory) [1.] chip supply for the $400 billion Stargate AI infrastructure project, is estimated to be worth more than 100 trillion Korean won (US$71.3 billion) for the Korean chipmakers over the next four years. The two companies say they are targeting 900,000 DRAM wafer starts per month – more than double the current global HBM capacity.

Note 1. HBM is a specialized type of DRAM that uses a unique 3D vertical stacking architecture and Through-Silicon Via (TSV) technology to achieve significantly higher bandwidth and performance than traditional, flat DRAM configurations. HBM uses standard DRAM “dies” stacked vertically, connected by TSVs, to create a more densely packed, high-performance memory solution for demanding applications like AI and high-performance computing.

…………………………………………………………………………………………………………………………………………………………………………….

“These partnerships will focus on increasing the supply of advanced memory chips essential for next-generation AI and expanding data center capacity in Korea, positioning Samsung and SK as key contributors to global AI infrastructure and supporting Korea’s ambition to become a top-three global AI nation.” OpenAI said.

The announcement followed a meeting between President Lee Jae-myung, Samsung Electronics Executive Chairman Jay Y. Lee, SK Chairman Chey Tae-won, and OpenAI CEO Sam Altman at the Presidential Office in Seoul.

Through these partnerships, Samsung Electronics and SK hynix plan to scale up production of advanced memory chips, targeting 900,000 DRAM wafer starts per month at an accelerated capacity rollout, critical for powering OpenAI’s advanced AI models.

OpenAI also signed a series of agreements today to explore developing next-generation AI data centers in Korea. These include a Memorandum of Understanding (MoU) with the Korean Ministry of Science and ICT (MSIT) specifically to evaluate opportunities for building AI data centers outside the Seoul Metropolitan Area, supporting balanced regional economic growth and job creation across the country.

The agreements signed today also include a separate partnership with SK Telecom to explore building an AI data center in Korea, as well as an agreement with Samsung C&T, Samsung Heavy Industries, and Samsung SDS to assess opportunities for additional data center capacity in the country.

References:

https://www.bloomberg.com/news/articles/2025-09-29/huawei-to-double-output-of-top-ai-chip-as-nvidia-wavers-in-china

https://www.lightreading.com/ai-machine-learning/huawei-sets-itself-as-china-s-go-to-for-ai-tech

https://openai.com/index/samsung-and-sk-join-stargate/

OpenAI orders $71B in Korean memory chips

AI Data Center Boom Carries Huge Default and Demand Risks

Big tech spending on AI data centers and infrastructure vs the fiber optic buildout during the dot-com boom (& bust)

U.S. export controls on Nvidia H20 AI chips enables Huawei’s 910C GPU to be favored by AI tech giants in China

Huawei launches CloudMatrix 384 AI System to rival Nvidia’s most advanced AI system

China gaining on U.S. in AI technology arms race- silicon, models and research

Will billions of dollars big tech is spending on Gen AI data centers produce a decent ROI?

Can the debt fueling the new wave of AI infrastructure buildouts ever be repaid?

Despite U.S. sanctions, Huawei has come “roaring back,” due to massive China government support and policies

Qualcomm CEO: expect “pre-commercial” 6G devices by 2028

During his keynote speech at the 2025 Snapdragon Summit in Maui, Qualcomm CEO Cristiano Amon said:

“We have been very busy working on the next generation of connectivity…which is 6G. Designed to be the connection between the cloud and Edge devices, The difference between 5G and 6G, besides increasing the speeds, increasing broadband, increasing the amount of data, it’s also a network that has intelligence to have perception and sensor data.  We’re going to have completely new use cases for this network of intelligence — connecting the edge and the cloud.”

“We have been working on this (6G) for a while, and it’s sooner than you think. We are ready to have pre-commercial devices with 6G as early as 2028. And when we get that, we’re going to have context aware intelligence at scale.”

…………………………………………………………………………………………………………………………………………………………………………………………..

Analysis: Let’s examine that statement, in light of the ITU-R IMT 2030 recommendations not scheduled to be completed until the end of 2030:

pre-commercial devices” are not meant for general consumers while “as early as” leaves open the possibility that those 6G devices might not be available until after 2028.

…………………………………………………………………………………………………………………………………………………………………………………………..

Looking ahead at the future of devices, Amon noted that 6G would play a key role in the evolution of AI technology, with AI models becoming hybrid. This includes a combination of cloud and edge devices (like user interfaces, sensors, etc). According to Qualcomm, 6G will make this happen.  Anon envisions a  future where AI agents are a crucial part of our daily lives, upending the way we currently use our connected devices. He firmly believes that smartphones, laptops, cars, smart glasses, earbuds, and more will have a direct line of communication with these AI agents — facilitated by 6G connectivity.

Opinion: This sounds very much like the hype around 5G ushering a whole new set of ultra-low latency applications which never happened (because the 3GPP specs for URLLC were not completed in June 2020 when Release 16 was frozen).  Also, very few mobile operators deployed 5G SA core, without which there are no 5G features, like network slicing and security.

Separately, Nokia Bell Labs has said that in the coming 6G era, “new man-machine interfaces” controlled by voice and gesture input will gradually replace more traditional inputs, like typing on touchscreens. That’s easy to read as conjecture, but we’ll have to see if that really happens when the first commercial 5G networks are deployed in late 2030- early 2031.

We’re sure to see faster network speeds with higher amounts of data with 6G with AI in more devices, but standardized 6G is still at least five years from being a commercial reality.

References:

https://www.androidauthority.com/qualcomm-6g-2028-3600781/

https://www.nokia.com/6g/6g-explained/

https://telecom.economictimes.indiatimes.com/news/devices/qualcomm-to-unveil-pre-commercial-6g-devices-by-2028-cristiano-amon/124094776

ITU-R WP5D IMT 2030 Submission & Evaluation Guidelines vs 6G specs in 3GPP Release 20 & 21

ITU-R WP 5D reports on: IMT-2030 (“6G”) Minimum Technology Performance Requirements; Evaluation Criteria & Methodology

ITU-R: IMT-2030 (6G) Backgrounder and Envisioned Capabilities

Ericsson and e& (UAE) sign MoU for 6G collaboration vs ITU-R IMT-2030 framework

Highlights of 3GPP Stage 1 Workshop on IMT 2030 (6G) Use Cases

6th Digital China Summit: China to expand its 5G network; 6G R&D via the IMT-2030 (6G) Promotion Group

MediaTek overtakes Qualcomm in 5G smartphone chip market

 

Lumen: “We’re Building the Backbone for the AI Economy” – NaaS platform to be available to more customers

“Lumen is determined to lead the transformation of our industry to meet the demands of the AI economy,” said Lumen Technologies CEO Kate Johnson. “With ubiquitous reach and a digital-first platform, we are positioned to deliver next-gen connectivity, power enterprise innovation, and secure our own growth. This is how we build the trusted network for AI and deliver exceptional value to our customers and shareholders.”

Highlights included keynote remarks from Johnson, who outlined the three pillars of the company’s strategy:

  • Building the backbone for the AI economy with a physical network designed for scale, speed, and security – delivering connectivity anywhere and for everything customers want to do.
  • Cloudifying and agentifying telecom to reduce complexity and simplify the network for customers as an intelligent, on-demand, consumption-based digital platform.
  • Creating a connected ecosystem with partnerships that extend Lumen’s reach, accelerate customer-first, AI-driven innovation, and unlock new opportunities across industries.

Johnson noted how Lumen’s growth is powered by a set of unique enablers that turn the company’s network into a true digital platform. With near-term product launches like self-service digital portal Lumen Connect, a universal Fabric Port, and new innovations in development that extend intelligence into the network edge, Lumen is making connectivity programmable and effortless. Combined with the company’s Network-as-a-Service business model and a connected ecosystem of data centers, hyper-scalers and technology partners, these enablers give customers the speed, security, and simplicity they need to thrive in the AI economy.

Lumen Technologies CEO Kate Johnson spotlights the company’s bold strategy, financial progress, and early look at product roadmap to reimagine digital networking for the AI economy at a gathering of industry analysts.

………………………………………………………………………………………………………………………………………………………………………………………………………

Chief Financial Officer Chris Stansbury said 2026 is expected to mark an inflection point as new digital revenues, growth in IP and Wavelengths, and long-term hyper-scaler contracts begin to outpace legacy declines – setting up what he called a “trampoline moment” for expansion. Lumen projects business segment revenue growth in 2028 and a return to overall top-line growth in 2029, establishing a clear path from stabilization to value creation.

With a strengthened balance sheet and greater financial freedom, executives highlighted the bold investment in the company’s three strategic pillars, each designed to accelerate innovation and position Lumen for long-term industry leadership.

Lumen’s strategy begins with the physical network, which carries a significant portion of the world’s internet traffic. With construction underway coast-to-coast, the company is executing a multi-billion-dollar program to expand its intercity and metro fiber backbone:

  • Adding 34 million new fiber miles by the end of 2028 for a total of 47 million intercity and metro miles.
  • Connecting data centers, clouds, edge, and enterprise locations in any combination.
  • Delivering 400G today and plans to scale to 1.6 terabits in the future.

Lumen’s substantial investments to expand high-speed connectivity ensures customers have the network scale, speed, and reliability to confidently innovate and grow without constraints.

The rise of AI is driving unprecedented demands for a new, Cloud 2.0 architecture with distributed, low-latency, high-bandwidth networks that can move and process massive amounts of data across multi-cloud, edge, and enterprise locations. Lumen is meeting this challenge by cloudifying and agentifying telecom, turning its expansive fiber footprint into a programmable digital platform that strips away the complexity of legacy networking.

Lumen plans to make its network-as-a-service (NaaS) platform [1.] available to more customers, regardless of their existing internet connection. At the company’s Analyst Forum, The NaaS platform includes new innovations like Lumen Fabric Port (Q4 2025), Lumen Multi-Cloud Gateway (Q4 2025), and Lumen Connect (Q1 2026). Together, these technologies digitize the entire service lifecycle, so customers can provision, manage, and scale thousands of services across thousands of locations, within minutes.

Note 1. Network as a Service (NaaS) is a cloud-based model that allows businesses to rent networking services from a provider on a subscription or pay-per-use basis, instead of building and maintaining their own network infrastructure. NaaS provides scalable and flexible network capabilities, shifting the cost from a capital expense (CapEx) to an operational expense (OpEx). NaaS functions by using a virtualized, software-defined network, meaning the network capabilities are abstracted from the physical hardware. Businesses access and manage their network resources through a web-based interface or portal, and the NaaS provider manages the underlying infrastructure, including hardware, software, updates, and troubleshooting.

Lumen CTO Dave Ward unveiled “Project Berkeley,” a network interface device that essentially expands the company’s NaaS services, like on-demand internet, Ethernet and IP VPN, to off-net sites using any access type. Those access types can be 5G, fiber, copper, fixed wireless access, satellite and more.  Project Berkeley leverages digital twin technology, which lets Lumen have “a full replicate understanding of exactly what’s going on in this device running out of our cloud.”

Ward said on the company’s website:

“Lumen is taking the network out of its hardware box and transforming it into a true digital platform. Technology and Product Officer Dave Ward. “By cloudifying our fiber assets into software and disrupting cloud economics, we’re giving customers the ability to turn up services within minutes, scale as their AI workloads demand, and innovate at cloud speed. This is what the future of digital networking should deliver.”

Lumen has been growing its NaaS platform for some time. It launched its first offering in 2023 and now counts over 1,000 enterprise NaaS customers. The company now plans to bring its connectivity products to over 10 million off-net buildings, said Ward. The device will also allow hyper-scalers to integrate and sell these products in their respective marketplaces.

In closing the Analyst session, CEO Johnson underscored that Lumen’s strategies are the foundation of the company’s momentum today – transforming the industry with innovation to fuel growth, strengthening financial performance, and positioning the company as a critical enabler in the digital economy.

“We’re thrilled by the energy and engagement we’ve seen from the analyst community. The discussions around how Lumen is delivering an expansive network, digital platform, connected ecosystem and winning culture to meet the exponential enterprise demands of AI demonstrate the urgent need for innovation in our industry, and we’re proud to be at the forefront of that conversation.”

About Lumen Technologies:

Lumen is unleashing the world’s digital potential. We ignite business growth by connecting people, data, and applications – quickly, securely, and effortlessly. As the trusted network for AI, Lumen uses the scale of our network to help companies realize AI’s full potential. From metro connectivity to long-haul data transport to our edge cloud, security, managed service, and digital platform capabilities, we meet our customers’ needs today and as they build for tomorrow.

For news and insights visit news.lumen.com, LinkedIn: /lumentechnologies, X: lumentechco, Facebook: /lumentechnologies, Instagram: @lumentechnologies and YouTube: /lumentechnologies. Lumen and Lumen Technologies are registered trademarks of Lumen Technologies LLC in the United States. Lumen Technologies LLC is a wholly owned affiliate of Lumen Technologies, Inc.

References:

For a replay of the webcast, visit Lumen’s investor website

https://ir.lumen.com/news/news-details/2025/Lumen-Highlights-AI-Era-Transformation-and-Path-to-Growth-at-Analyst-Forum/default.aspx

https://www.fierce-network.com/broadband/lumen-says-its-taking-its-naas-new-level

Lumen deploys 400G on a routed optical network to meet AI & cloud bandwidth demands

Dell’Oro: Bright Future for Campus Network As A Service (NaaS) and Public Cloud Managed LAN

NaaS emerges as challenger to legacy network models; likely to grow rapidly along with SD WAN market

Verizon and WiPro in Network-as-a-Service (NaaS) partnership

ABI Research: Network-as-a-Service market to be over $150 billion by 2030

Cisco Plus: Network as a Service includes computing and storage too

Gartner: changes in WAN requirements, SD-WAN/SASE assumptions and magic quadrant for network services

Ericsson integrates agentic AI into its NetCloud platform for self healing and autonomous 5G private networks

Ericsson is integrating agentic AI into its NetCloud platform to create self-healing and autonomous 5G private (enterprise) networks. This initiative upgrades the existing NetCloud Assistant (ANA), a generative AI tool, into a strategic partner capable of managing complex workflows and orchestrating multiple AI agents.  The agentic AI agent aims to simplify private 5G adoption by reducing deployment complexity and the need for specialized administration.   This new agentic architecture allows the new Ericsson system to interpret high-level instructions and autonomously assign tasks to a team of specialized AI agents.

Key AI features include:

  • Agentic organizational hierarchy: ANA will be supported by multiple orchestrator and functional AI agents capable of planning and executing (with administrator direction). Orchestrator agents will be deployed in phases, starting with a troubleshooting agent planned in Q4 2025, followed by configuration, deployment, and policy agents planned in 2026. These orchestrators will connect with task, process, knowledge, and decision agents within an integrated agentic framework.
  • Automated troubleshooting: ANA’s troubleshooting orchestrator will include automated workflows that address the top issues identified by Ericsson support teams, partners, and customers, such as offline devices and poor signal quality. Planned to launch in Q4 2025, this feature is expected to reduce downtime and customer support cases by over 20 percent.
  • Multi-modal content generation: ANA can now generate dynamic graphs to visually represent trends and complex query results involving multiple data points.
  • Explainable AI: ANA displays real-time process feedback, revealing steps taken by AI agents in order to enhance transparency and trust.
  • Expanded AIOps insights: NetCloud AIOps will be expanded to provide isolation and correlation of fault, performance, configuration, and accounting anomalies for Wireless WAN and NetCloud SASE. For Ericsson Private 5G, NetCloud is expected to provide service health analytics including KPI monitoring and user equipment connectivity diagnostics. Planned availability Q4 2025.
Planned to be available Q4 2025, the integration of Ericsson Private 5G into the NetCloud platform brings powerful advantages to enterprise 5G customers, including access to AI features, real-time feature availability, simplified lifecycle management, greater agility across multisite deployments and better administrator controls with distinct user roles and permissions. NetCloud acts as a foundation for future agentic AI features focused on removing friction and adding value for the enterprise. These innovations directly address critical adoption barriers as more industrial enterprises leverage private 5G for business-critical connectivity. With this integration, Ericsson is empowering businesses to overcome these challenges and unlock the full potential of 5G in IT and OT environments.
Ericsson announces integration of new agentic AI technology into NetCloud
Ericsson says: “Agentic AI is the next wave of AI. It acts as a powerful force multiplier, characterized by multiple specialized agents working collaboratively to tackle complex problems and manage intricate workflows. These AI advisors serve as vigilant partners, providing continuous monitoring and intelligent assistance to maintain and optimize operational environments.”
Image Credit: Ericsson
…………………………………………………………………………………………………………………………………………………

Manish Tiwari, Head of Enterprise 5G, Ericsson Enterprise Wireless Solutions, adds: “With the integration of Ericsson Private 5G into the NetCloud platform, we’re taking a major step forward in making enterprise connectivity smarter, simpler, and adaptive. By building on powerful AI foundations, seamless lifecycle management, and the ability to scale securely across sites, we are providing flexibility to further accelerate digital transformation across industries. This is about more than connectivity: it is about giving enterprises the business-critical foundation they need to run IT and OT systems with confidence and unlock the next wave of innovation for their businesses.”

Pankaj Malhotra, Head of WWAN & Security, Ericsson Enterprise Wireless Solutions, says: “By introducing agentic AI into NetCloud, we’re enabling enterprises to simplify deployment and operations while also improving reliability, performance, and user experience. More importantly, it lays the foundation for our vision of fully autonomous, self-optimizing 5G enterprise networks, that can power the next generation of enterprise innovation.”

Ericsson is positioning itself as a leader in enterprise 5G by being the first major vendor to introduce agentic AI into network management. This move is seen as going beyond standard AIOps, aligning with the industry trend towards AI-native management systems.  Ericsson hopes it will increase revenues which grew at a tepid 2% year-over-year in the last quarter. The company had the largest sales (#1 vendor) of 5G network equipment outside of China last year.
References:

OpenAI and Broadcom in $10B deal to make custom AI chips

Overview:

Late last October, IEEE Techblog reported that “OpenAI the maker of ChatGPT, was working with Broadcom to develop a new artificial intelligence (AI) chip focused on running AI models after they’ve been trained.”  On Friday, the WSJ and FT (on-line subscriptions required) separately confirmed that OpenAI is working with Broadcom to develop custom AI chips, a move that could help alleviate the shortage of powerful processors needed to quickly train and release new versions of ChatGPT.  OpenAI plans to use the new AI chip internally, according to one person close to the project, rather than make them available to external customers.

………………………………………………………………………………………………………………………………………………………………………………………….

Broadcom:

During its earnings call on Thursday, Broadcom’s CEO Hock Tan said that it had signed up an undisclosed fourth major AI developer as a custom AI chip customer, and that this new customer had committed to $10bn in orders.  While Broadcom did not disclose the names of the new customer, people familiar with the matter confirmed OpenAI was the new client. Broadcom and OpenAI declined to comment, according to the FT.  Tan said the deal had lifted the company’s growth prospects by bringing “immediate and fairly substantial demand,” shipping chips for that customer “pretty strongly” starting next year. “The addition of a fourth customer with immediate and fairly substantial demand really changes our thinking of what 2026 would be starting to look like,” Tan added.

Image credit:  © Dado Ruvic/Reuters

HSBC analysts have recently noted that they expect to see a much higher growth rate from Broadcom’s custom chip business compared with Nvidia’s chip business in 2026. Nvidia continues to dominate the AI silicon market, with “hyperscalers” still representing the largest share of its customer base. While Nvidia doesn’t disclose specific customer names, recent filings show that a significant portion of their revenue comes from a small number of unidentified direct customers, which likely are large cloud providers like  Microsoft, Amazon, Alphabet (Google), and Meta Platforms.

In August, Broadcom launched its Jericho networking chip, which is designed to help speed up AI computing by connecting data centers as far as 60 miles apart.  By August, Broadcom’s market value had surpassed that of oil giant Saudi Aramco, making the chip firm the world’s seventh-largest publicly listed company.

……………………………………………………………………………………………………………………………………….

Open AI:

OpenAI CEO Sam Altman has been saying for months that a shortage of graphics processing units, or GPUs, has been slowing his company’s progress in releasing new versions of its flagship chatbot. In February, Altman wrote on X that ChatGPT-4.5, its then-newest large language model, was the closest the company had come to designing an AI model that behaved like a “thoughtful person,” but there were very high costs that came with developing it. “We will add tens of thousands of GPUs next week and roll it out to the plus tier then. (hundreds of thousands coming soon, and i’m pretty sure y’all will use every one we can rack up.)”

In recent years, OpenAI has relied heavily on so-called “off the shelf” GPUs produced by Nvidia, the biggest player in the chip-design space. But as demand from large AI firms looking to train increasingly sophisticated models has surged, chip makers and data-center operators have struggled to keep up. The company was one of the earliest customers for Nvidia’s AI chips and has since proven to be a voracious consumer of its AI silicon.

“If we’re talking about hyperscalers and gigantic AI factories, it’s very hard to get access to a high number of GPUs,” said Nikolay Filichkin, co-founder of Compute Labs, a startup that buys GPUs and offers investors a share in the rental income they produce. “It requires months of lead time and planning with the manufacturers.”

To solve this problem, OpenAI has been working with Broadcom for over a year to develop a custom chip for use in model training. Broadcom specializes in what it calls XPUs, a type of semiconductor that is designed with a particular application—such as training ChatGPT—in mind.

Last month, Altman said the company was prioritizing compute “in light of the increased demand from [OpenAI’s latest model] GPT-5” and planned to double its compute fleet “over the next 5 months.” OpenAI also recently struck a data-center deal with Oracle that calls for OpenAI to pay more than $30 billion a year to the cloud giant, and signed a smaller contract with Google earlier this year to alleviate computing shortages. It is also embarking on its own data-center construction project, Stargate, though that has gotten off to a slow start.

OpenAI’s move follows the strategy of tech giants such as Google, Amazon and Meta, which have designed their own specialized custom chips to run AI workloads. The industry has seen huge demand for the computing power to train and run AI models.

………………………………………………………………………………………………………………………………………………………………………………..

References:

https://www.ft.com/content/e8cc6d99-d06e-4e9b-a54f-29317fa68d6f

https://www.wsj.com/tech/ai/openai-broadcom-deal-ai-chips-5c7201d2

Reuters & Bloomberg: OpenAI to design “inference AI” chip with Broadcom and TSMC

Open AI raises $8.3B and is valued at $300B; AI speculative mania rivals Dot-com bubble

OpenAI announces new open weight, open source GPT models which Orange will deploy

OpenAI partners with G42 to build giant data center for Stargate UAE project

Generative AI Unicorns Rule the Startup Roost; OpenAI in the Spotlight

Will the wave of AI generated user-to/from-network traffic increase spectacularly as Cisco and Nokia predict?

Network operators are bracing themselves for a wave of AI traffic, partially based on a RtBrick survey, as well as forecasts by Cisco and Nokia, but that hasn’t happened yet.  The heavy AI traffic today is East to West (or vice-versa) within cloud resident AI data centers and for AI data center interconnects.

1.  Cisco believes that AI Inference agents will soon engage “continuously” with end-users, keeping traffic levels consistently high. has stated that AI will greatly increase network traffic, citing a shift toward new, more demanding traffic patterns driven by “agentic AI” and other applications. This perspective is a core part of Cisco’s business strategy, which is focused on selling the modernized infrastructure needed to handle the coming surge. Cisco identified three stages of AI-driven traffic growth, each with different network demands: 

  • Today’s generative AI models produce “spikey” traffic which spikes up when a user submits a query, but then returns to a low baseline. Current networks are largely handling this traffic without issues.
  • Persistent “agentic” AI traffic: The next phase will involve AI agents that constantly interact with end-users and other agents. Cisco CEO Chuck Robbins has stated that this will drive traffic “beyond the peaks of current chatbot interaction” and keep network levels “consistently high”.
  • Edge-based AI: A third wave of “physical AI” will require more computing and networking at the edge of the network to accommodate specialized use cases like industrial IoT. 

“As we move towards agentic AI and the demand for inferencing expands to the enterprise and end user networking environments, traffic on the network will reach unprecedented levels,” Cisco CEO Chuck Robbins said on the company’s recent earnings call. “Network traffic will not only increase beyond the peaks of current chatbot interaction, but will remain consistently high with agents in constant interaction.”

2. Nokia recently predicted that both direct and indirect AI traffic on mobile networks will grow at a faster pace than regular, non-AI traffic.

  • Direct AI traffic: This is generated by users or systems directly interacting with AI services and applications. Consumer examples: Using generative AI tools, interacting with AI-powered gaming, or experiencing extended reality (XR) environments. Enterprise examples: Employing predictive maintenance, autonomous operations, video and image analytics, or enhanced customer interactions.
  • Indirect AI traffic: This occurs when AI algorithms are used to influence user engagement with existing services, thereby increasing overall traffic. Examples: AI-driven personalized recommendations for video content on social media, streaming platforms, and online marketplaces, which can lead to longer user sessions and higher bandwidth consumption. 

The Finland based network equipment vendor warned that the AI wave could bring “a potential surge in uplink data traffic that could overwhelm our current network infrastructure if we’re not prepared,” noting that the rise of hybrid on-device and cloud tools will require much more than the 5-15 Mbps uplink available on today’s networks.  Nokia’s Global Network Traffic 2030 report forecasts that overall traffic could grow by 5 to 9 times current levels by 2033.  All told, Nokia said AI traffic is expected to hit 1088 exabytes (EB) per month by 2033. That means overall traffic will grow 5x in a best case scenario and 9x in a worse case.

To manage this anticipated traffic surge, Nokia advocates for radical changes to existing network infrastructure.

  • Cognitive networks: The company states that networks must become “cognitive,” leveraging AI and machine learning (ML) to handle the growing data demand.
  • Network-as-Code: As part of its Technology Strategy 2030, Nokia promotes a framework for more flexible and scalable networks that leverage AI and APIs.
  • 6G preparation: Nokia Bell Labs is already conducting research and field tests to prepare for 6G networks around 2030, with a focus on delivering the capacity needed for AI and other emerging technologies.
  • Optimizing the broadband edge: The company also highlights the need to empower the broadband network edge to handle the demands of AI applications, which require low latency and high reliability. 

Nokia’s Global Network Traffic 2030 report didn’t mention agentic AI, which are artificial intelligence systems designed to autonomously perceive, reason, and act in their environment to achieve complex goals with less human oversight. Unlike generative AI, which focuses on creating content, agentic AI specializes in workflow automation and independent problem-solving by making decisions, adapting plans, and executing tasks over extended periods to meet long-term objectives.

3.  Ericsson did point to traffic increases stemming from the use of AI-based assistants in its 2024 Mobility Report. In particular, it predicted the majority of traffic would be related to the use of consumer video AI assistants, rather than text-based applications and – outside the consumer realm – forecast increased traffic from “AI agents interacting with drones and droids. Accelerated consumer uptake of GenAI will cause a steady increase of traffic in addition to the baseline increase,” Ericsson said of its traffic growth scenario.

…………………………………………………………………………………………………………………………………………………………………………………..

Dissenting Views:

1.  UK Disruptive Analysis Founder Dean Bubley isn’t a proponent of huge AI traffic growth. “Many in the telecom industry and vendor community are trying to talk up AI as driving future access network traffic and therefore demand for investment, spectrum etc., but there is no evidence of this at present,” he told Fierce Network.

Bubley argues that AI agents won’t really create much traffic on access networks to homes or businesses. Instead, he said, they will drive traffic “inside corporate networks, and inside and between data centers on backbone networks and inside the cloud.  “There might be a bit more uplink traffic if video/images are sent to the cloud for AI purposes, but again that’s hypothetical,” he said.

2.  In a LinkedIn post, Ookla analyst Mike Dano said he was a bit suspicious about “Cisco predicting a big jump in network traffic due to AI agents constantly wandering around the Internet and doing things.”  While almost all of the comments agreed with Dano, it still is an open question whether the AI traffic Armageddon will actually materialize.

……………………………………………………………………………………………………………………………………………………………………………………….

References:

RtBrick survey: Telco leaders warn AI and streaming traffic to “crack networks” by 2030

https://www.fierce-network.com/cloud/will-ai-agents-really-raise-network-traffic-baseline

Q4FY25-Earnings-Slides.pdf

https://onestore.nokia.com/asset/213660

https://www.linkedin.com/posts/mikedano_it-looks-like-cisco-is-predicting-a-big-jump-activity-7363223007152017408-JiVS/

Page 1 of 12
1 2 3 12