Will “AI at the Edge” transform telecom or be yet another telco monetization failure?

New Telco Opportunity – AI at the Edge:

At MWC 2026 last week, there were a flurry of claims that “AI at the Edge” would transform the telecom industry.  One of many examples is an article titled, “The AI edge boom is giving telecom a new strategic role.”  In that piece, Jeff Aaron, vice president of product and solutions marketing at Hewlett Packard Enterprise (HPE) spoke with theCUBE’s John Furrier at MWC Barcelona, during an exclusive broadcast on theCUBE, SiliconANGLE Media’s livestreaming studio. They discussed telecom edge AI and why networking is becoming a strategic foundation for data-centric services.  Aaron said:

“A big reason for [reignited interest in routing] is AI workloads. They’re moving everywhere now. They have to move to the edge.  For them to move to the edge, you’ve got to get them outside of the factory and to all the locations. We’re right in the core of that, and it’s super exciting.”

As AI expands to the edge, data will need to move not only to local compute, but also between many distributed edge sites, making routing paramount. There are four ways AI infrastructure is scaling — inside data centers and across distributed edge locations, according to Aaron.

“There’s scale-out, scale-across, scale-up, and on-ramp. Two are within the data center — scale-out and scale-up — but scale-across and edge on-ramp basically mean you got to figure out how to connect to those areas, and those are just networking,” he added.

Scale-across refers to connecting distributed data centers and edge locations, while edge on-ramp brings remote sites such as factories or branch locations into the network to access AI services. Supporting those distributed environments creates an opportunity for HPE to bring networking and compute together into a more integrated infrastructure stack. At MWC 2026 Barcelona, those trends are clearly coming into focus, according to Aaron.

“Data is moving everywhere right now, and the network is back. The network isn’t just plumbing. The network is how you build a value-added service using an AI workload as a telco infrastructure,” he added.

Telecom carriers are now urgently trying to move from being “dumb data pipes” to becoming “AI performance platforms” by leveraging their geographically distributed infrastructure to host AI closer to the end user.  They urgently want to pivot from selling just bandwidth and connectivity to selling outcomes and intelligence with a heavy focus on industrial and enterprise-specific edge deployments.  They are considering the following services and business models:

  • Infrastructure as a Service (IaaS) & GPUaaS: Offering raw computing power, specifically GPUs, from edge data centers to enterprises that need low-latency processing without building their own facilities.
  • Sovereign AI Clouds: Providing AI services that guarantee data remains within national borders, appealing to government and highly regulated sectors like finance and healthcare.
  • API Monetization: Exposing real-time network data (e.g., location intelligence, predictive network quality, fraud risk scoring) via APIs that enterprises pay to integrate into their own applications.
  • Outcome-Based Pricing: Charging for specific business results, such as a “guaranteed video call quality” or “fraud loss reduction share,” rather than just data usage.
  • AI-as-a-Service (AIaaS): Bundling pre-trained models or specialized AI agents (e.g., for customer service or industrial monitoring) with connectivity

Major Carrier AI Edge Deployment Plans:

  • AT&T:
    • Launched Connected AI for Manufacturing in March 2026, which unifies 5G, IoT, and generative AI to provide real-time fault detection (claiming a 70% reduction in waste).
    • Deploying “Edge Zones” in major U.S. cities (Detroit, LA, Dallas) to allow developers to run low-latency, cloud-based software locally.
    • Partnering with AWS to link fiber and 5G directly into AWS environments for distributed AI workloads.
  • Verizon:
    • Unveiled Verizon AI Connect, a suite of products designed to manage resource-intensive AI workloads for hyperscalers like Google Cloud and Meta.
    • Trialing V2X (Vehicle-to-Everything) platforms to provide carmakers with standardized APIs for low-latency edge processing in autonomous driving.
    • Collaborating with NVIDIA to integrate GPUs into private 5G networks for on-premise AI inferencing in robotics and AR.
  • SK Telecom (SKT):
    • Announced an “AI Native” strategy at MWC 2026, including a roadmap for AI-RAN (Radio Access Network) that uses GPUs to optimize network performance and host user AI apps simultaneously.
    • Building a Manufacturing AI Cloud powered by over 2,000 NVIDIA RTX GPUs to support digital twin simulations and robotics.
    • Expanding AI Data Centers (AIDC) across South Korea and Southeast Asia (Vietnam, Malaysia) using energy-optimized LNG-powered facilities.
  • Orange & Deutsche Telekom:
    • Deploying AI-powered planning tools to cut fiber rollout costs and optimize site power consumption by up to 33% using AI “Deep Sleep” modes.
    • Focusing on Sovereign AI strategies to ensure data governance for European enterprise customers.
  • Vodafone:
    • Utilizing AI/ML applications for daily power reduction at 5G sites and testing autonomous network healing via AI agents
  • BT:
    • Offers 5G-connected VR for manufacturing design teams (e.g., Hyperbat) to collaborate on 3D models in real-time.  
……………………………………………………………………………………………………………..
Summary of Emerging AI Edge Products:
Product Category Primary Target Key Value Proposition
AI-RAN Industry 4.0 Seamless, ultra-low latency for robotics and sensing.
Connected AI Platforms Manufacturing Real-time predictive maintenance and waste reduction.
AI-as-a-Service (AIaaS) Developers/SMBs Access to GPU power and pre-trained models via telco edge nodes.
Network Slicing APIs App Developers Programmatic control over bandwidth for AR/VR and gaming.

…………………………………………………………………………………………………………………………………………………………………………………………..

A Dissenting View of “AI at the Edge”:

The global market for AI within the global telecommunications sector is valued at $6.69 billion in 2026, growing at a compound annual rate (CAGR) of 41.9% from 2025.   The broader edge AI market—including hardware, software, and services—is forecast to reach $29.98 billion in 2026, according to The Business Research Company We think those estimates are way too high.

The market research firm states:

………………………………………………………………………………………………………

Author’s Opinion:

Unless telcos change their corporate culture along with slowing the footprint growth of cloud service providers/hyperscalers, we think that AI at the Edge will be yet another telco monetization failure.  Just like their failure to monetize: 4G LTE apps, the telco cloud, 5G, multi-access edge computing (MEC), OpenRAN, LPWANs and other telecom technologies that never lived up to their promise and potential.

That’s largely because telcos are very weak: developing IT platforms, compute services, killer applications, and rapid execution of new services (e.g. 5G services require a 5G SA core network which telcos were very slow to deploy).  Telecom execs themselves cite cultural and speed‑of‑change issues: the industry is not organized like a software company, so it struggles to iterate products at AI/cloud pace. Also, telcos historically struggle with software. Managing distributed GPU clusters is vastly different from managing cell towers.

After spending billions on 5G with very  little or no ROI, investors are skeptical of the increased capex required for AI-grade edge servers which must be maintained by telcos.  Those servers will be expensive (especially if they contain clusters of Nvidia GPUs) and consume a lot of power, which is a critical issue at the edge of the carrier’s network.

Many network operators frame AI/edge as “network optimization” or “utilizing underused sites,” not as building monetizable AI platforms with APIs, SDKs, and ecosystems. This mirrors 5G, where huge RAN/core builds were not matched by a clear product and platform strategy, leaving value to OTTs and hyperscalers which are  extending their control planes and protocol stacks to the network edge (local zones, operator co‑lo, on‑premises stacks).

Telcos risk becoming “dumb pipes” for AI traffic if they can’t provide a superior developer ecosystem.  If they only sell space/power/connectivity, the cloud service providers will continue to own the developer and AI value chain.  Analysts warn that edge is a “right to participate, not a right to win.”  As such, value accrues to whoever owns the AI platform, tools, marketplace, and pricing power, not the entity that provides connectivity, PoP or cell towers.

Data fragmentation and weak “intelligence” layer:

  • AI monetization depends on high‑quality, cross‑domain data, but telco data is fragmented across OSS, BSS, probes, and partner systems; without unification, it is hard to expose compelling network/edge intelligence services.

  • Analysts emphasize that failure here reduces telcos to generic GPU landlords, while higher‑margin offers (real‑time quality, fraud, identity, mobility/context APIs) remain unrealized.

Narrow internal focus on cost savings:

  • Many operators’ early AI focus is inward (Opex reduction in assurance, planning, customer care) rather than building external, revenue‑generating products, echoing how early 5G was justified mainly on cost/efficiency.

  • Commentators warn that if AI/edge remains a “network efficiency” play, the commercial upside will go to cloud/AI natives that turn similar capabilities into products sold to enterprises.

What analysts say telcos must do differently:

  • Build “Sovereign AI factories” and edge AI clouds: GPU‑enabled sites with cloud‑like developer experience (APIs, self‑service portals, metering, SLAs) and clear sovereign/regional guarantees.

  • Combine differentiated connectivity with AI services (latency‑backed SLAs, AI‑on‑RAN, domain‑specific models for verticals) and use modern, flexible commercial models instead of just selling bandwidth or colocation.

Conclusions:

In summary, the main risk for telcos is to successfully transition from owning and maintaining network infrastructure to owning and operating AI platforms and products at software industry speed.  AI at the edge is less of a new service or product and more an architectural upgrade. The two ways telcos can benefit are from:

  1.  Internal cost reduction: If telcos use it to lower their own costs (fraud prevention, risk management, predictive maintenance, fault isolation, self-healing networks, etc.), it’s an automatic win but won’t increase the top line.
  2.  Revenue from new AI -Edge services, e.g. Verizon uses edge-based video analytics in warehouses to improve inventory turnover by up to 40%.   If they expect to charge a massive premium for “AI-enabled 5G,” they face the same monetization wall that has doomed them for the past 20 years!

References:

https://siliconangle.com/2026/03/04/telecom-edge-ai-makes-networking-strategic-mwc26/

https://www.nvidia.com/en-us/lp/ai/the-blueprint-for-ai-success-ebook/

How telcos can monetize AI beyond connectivity

https://www.thebusinessresearchcompany.com/report/generative-artificial-intelligence-ai-in-telecom-global-market-report

AT&T and AWS to deliver last mile connectivity for AI workloads; AT&T Geo Modeler™ AI simulation tool

Analysis: Edge AI and Qualcomm’s AI Program for Innovators 2026 – APAC for startups to lead in AI innovation

Ericsson goes with custom silicon (rather than Nvidia GPUs) for AI RAN

Private 5G networks move to include automation, autonomous systems, edge computing & AI operations

Dell’Oro: RAN Market Stabilized in 2025 with 1% CAG forecast over next 5 years; Opinion on AI RAN, 5G Advanced, 6G RAN/Core risks

Dell’Oro: Analysis of the Nokia-NVIDIA-partnership on AI RAN

Dell’Oro: AI RAN to account for 1/3 of RAN market by 2029; AI RAN Alliance membership increases but few telcos have joined

Dell’Oro: RAN revenue growth in 1Q2025; AI RAN is a conundrum

Nvidia AI-RAN survey results; AI inferencing as a reinvention of edge computing?

RAN silicon rethink – from purpose built products & ASICs to general purpose processors or GPUs for vRAN & AI RAN

CES 2025: Intel announces edge compute processors with AI inferencing capabilities

AT&T and AWS to deliver last mile connectivity for AI workloads; AT&T Geo Modeler™ AI simulation tool

AT&T is strategically re-architecting its infrastructure for the AI era through high-capacity network modernization and deep integration with hyperscale cloud providers.

In addition to its almost six year old deal to run its 5G SA core network in Microsoft Azure’s cloudAT&T announced at MWC 2026 that it’s now woring with Amazon Web Services (AWS) to extend 5G and fiber connectivity from business customers and locations directly into AWS environments, creating secure, resilient and reliable premises‑to‑cloud architectures for AI workloads. The collaboration is designed to reduce network complexity and latency while supporting real‑time analytics, machine learning, and agentic AI use cases.

This collaboration continues a long-standing relationship between AT&T and AWS and follows recent news outlining broader efforts to modernize the nation’s connectivity infrastructure by providing high-capacity fiber to AWS data centers, migrate AT&T workloads to AWS cloud capabilities and explore emerging satellite technologies.

AWS Interconnect – last mile embeds AT&T‑delivered connectivity directly into AWS workflows, designed to enable customers to provision and manage last‑mile connectivity within the AWS environment and lays the foundation for the use of AI agents to monitor and manage the AI experience from the user to the cloud. This streamlined, self‑managed approach helps enterprises reduce network complexity while maintaining control of their extended enterprise network, allowing businesses to move faster as they scale AI.

High level illustration of the planned AWS Interconnect – last mile architecture, showing how resilient interconnections and AT&T Fiber and fixed wireless access are intended to simplify private connectivity from customer locations into AWS environments. 

Diagram Source: AT&T

………………………………………………………………………………………………………

“AI does not just need more compute; it needs flatter networks and faster connections,” said Shawn Hakl, SVP & Head of Product, AT&T Business. “By bringing high‑capacity connectivity closer to cloud platforms, integrating the management of the networks directly into the cloud provisioning process and engineering for resiliency at the metro level, AT&T is helping enterprises streamline their networks, improve performance, security, and scale AI with confidence.”

AT&T says they are building an AI‑ready network (?) designed to scale performance by continuing ongoing network investment, including the growth of capacities up to 1.6Tbps across key metro and long‑haul routes.

AT&T also announced it would work with Nvidia, Microsoft and MicroAI through its Connected AI platform for “smart manufacturing.”

………………………………………………………………………………………………………………..

Finally, AT&T described  AT&T Geo Modeler which is able to better predict connectivity for emerging technologies like autonomous vehicles, drones, and robotics.

The Geo Modeler is an AI-powered simulation tool that helps predict, in near real time, how a wireless network will perform in the real world. Inspired by the video games Kounev played with his family growing up, the virtual model and simulation is “essentially like a giant video game of the United States” that, infused with AI tools, gives engineers a clearer picture of where potential weak spots may appear. Then issues can be addressed earlier and fixes can roll out faster. In essence, it creates virtual models, similar to the way video games are designed and developed.

“The Geo Modeler helps us see how the real world will shape coverage before we build, so we can deliver connectivity that’s ready for what’s next,” said AT&T scientist Velin Kounev.

Matt Harden, VP of Connected Solutions at AT&T, agrees. “The Geo Modeler is a foundational capability for the connected mobility era,” he said. “By marrying advanced geospatial simulation with AI-driven network orchestration, we can deliver predictable, high-performance connectivity that adapts with the environment. Whether it’s a hurricane, a packed stadium, or a city corridor full of autonomous vehicles, we will be prepared.”

References:

https://about.att.com/story/2026/aws-collaboration-scalable-business-ai.html

https://about.att.com/blogs/2026/150-years-of-connection.html

https://about.att.com/blogs/2025/geo-modeler.html

AT&T and Ericsson boost Cloud RAN performance with AI-native software running on Intel Xeon 6 SoC

AT&T deploys nationwide 5G SA while Verizon lags and T-Mobile leads

AT&T to buy spectrum licenses from EchoStar for $23 billion

AT&T’s convergence strategy is working as per its 3Q 2025 earnings report

Progress report: Moving AT&T’s 5G core network to Microsoft Azure Hybrid Cloud platform

AT&T 5G SA Core Network to run on Microsoft Azure cloud platform

 

Analysis & Economic Implications of AI adoption in China

Executive Summary:

Visible signs of artificial intelligence adoption in China are everywhere. Consumers interact seamlessly with chatbots, livestream hosts promote algorithmically selected products, and recommendation engines exhibit an almost anticipatory understanding of user preferences.  Yet, beyond these consumer-facing applications, a deeper and potentially more consequential transformation is unfolding. Across China’s retail and services sectors, AI is shifting from demand generation to cost optimization. Enterprises are deploying machine learning in logistics, inventory management, customer service, and fulfillment operations to reduce inefficiencies as revenue growth slows and pricing power tightens.

Highlights:

  • Chinese companies are increasingly using AI to control operational costs and improve efficiency in a low-growth economic environment.

  • AI is being deployed in logistics, inventory management, and customer service to reduce expenses rather than primarily drive demand.

  • This shift towards AI for cost reduction is leading to steadier cash flow and improved operating margins for consumer companies.

China’s Consumer Sector: AI Powers Efficiency Over Growth:

As China’s economy adjusts to structural deceleration—marked by subdued household confidence, persistent real-estate overhang, and maturing market saturation—consumer companies face an unfamiliar imperative: prioritize resilience over expansion. With pricing power eroded and cost inflation persistent, traditional growth levers have lost potency. Leading platforms are responding by reorienting AI investments toward operational efficiency, transforming algorithms from engagement engines into margin-defense mechanisms. For investors, this evolution signals a new phase of earnings potential—one where incremental productivity gains could prove more durable than cyclical demand recovery.

“In a low-growth environment, incremental efficiency gains matter more than top-line expansion,” notes Zhao Ming, senior analyst for China internet companies at Hongyuan Capital. “AI has become a strategic lever for margin preservation.”

China’s consumer sector entered 2026 navigating familiar structural headwinds: cautious household sentiment, a fading property-wealth effect, and fierce price competition. Unlike in previous cycles, companies are finding it increasingly difficult to pass rising costs on to consumers. The result has been a strategic realignment. Where past growth phases emphasized volume and engagement, today’s market is rewarding operational discipline. That shift has sharpened the appeal of AI—not as a marketing showcase, but as a core instrument of productivity and cost control.

“In a slower-growth environment, leading Chinese consumer companies are using AI primarily to improve productivity and reduce operating costs rather than to drive incremental demand,” McKinsey said in a recent analysis of AI adoption across China’s retail and services sectors.

From Growth Catalyst to Cost Lever:

The center of gravity for AI investment has shifted from customer-facing innovation to operational optimization. E-commerce platforms and logistics operators have been among the earliest to integrate AI into mission-critical workflows. Demand-forecasting models are helping warehouses fine-tune inventory levels and reduce exposure to slow-moving goods. Routing algorithms are compressing last-mile delivery times and cutting fuel consumption. Automated customer-service systems are deflecting an ever-larger share of inquiries typically handled by human agents.

On their own, each of these applications may appear incremental. Taken together, they represent a meaningful improvement in margin resilience at a time when top-line expansion remains constrained. In an environment where minor percentage-point gains in efficiency can significantly affect earnings quality, AI is emerging as a quiet but potent differentiator.

Logistics as a Testbed for Scalable Efficiency:

The operational impact of AI is most visible in the logistics ecosystem, a sector that remains one of the largest cost centers in China’s consumer economy. Machine-learning systems are now proficient at forecasting order density by neighborhood and time of day, enabling fulfillment centers to position inventory closer to anticipated demand. In dense urban markets, adaptive algorithms continually adjust delivery routes in response to evolving conditions—from traffic and weather to cancellations and reorders—reducing both transit times and redundancy.

For investors, the value proposition is compelling: logistics efficiency scales. Once AI models are trained and stress-tested, they can be deployed across regions at low incremental cost, generating operating leverage even in periods of stagnant demand. Crucially, incumbents benefit from data scale. Years of transaction and delivery records translate into more accurate predictive models, reinforcing competitive moats and raising barriers to entry. This dynamic is reshaping industry structure even as consumer-facing platform features converge toward commoditization.

AI Extends Gains to Physical Retail:

Beyond e-commerce, brick-and-mortar retail—long considered a laggard in China’s digital transformation—is also seeing measurable efficiency dividends. Smart shelving, computer-vision inventory systems, and automated stock monitoring are cutting labor intensity while increasing inventory turnover. Grocery and convenience chains now rely on AI to optimize product assortments at the store level, calibrating selections to localized consumption patterns instead of applying national averages. The effect is twofold: reduced waste and fewer markdowns, both of which have historically weighed on profitability. The outcomes may not register as eye-catching innovation, but they align closely with investor priorities—stabler cash flows and predictable margins.

Labor Efficiency as a Strategic Imperative:

AI-enhanced customer service represents another underappreciated margin driver. Major consumer platforms report that routine customer interactions—order tracking, returns, product troubleshooting—are now predominantly handled through automated systems. This transition is particularly relevant in a labor market where wage growth continues to outpace consumption. Limiting headcount growth while maintaining response times and service quality has become a key operational goal.

“AI doesn’t replace customer service,” says Li Wenyuan, chief technology officer at retail software firm Qimeng Tech. “It filters it, so humans deal only with the expensive problems.” That filtering function is transforming customer operations from cost centers into scalable service platforms, balancing efficiency with user satisfaction.

Economic Implications:

For investors, the impact of China’s second-wave AI adoption will likely manifest less in headline growth metrics and more in incremental financial performance indicators. Key areas to watch include:

  • Operating margin expansion driven by process automation

  • Reduced fulfillment and logistics costs as a share of revenue

  • Improved capital-expenditure efficiency through data-driven asset utilization

The first chapter of China’s AI consumer story was about differentiation—using algorithms to personalize experiences, boost engagement, and drive sales. The next chapter is about discipline. As growth normalizes, companies are deploying AI to do more with less: compress costs, stabilize earnings, and build leaner, more adaptive operating models. In a market where scale alone no longer guarantees profitability, AI has become not just a tool for innovation—but a mechanism for survival.

References:

https://www.barrons.com/articles/china-ai-boom-commerce-warehouses-b1ad55f1

China’s open source AI models to capture a larger share of 2026 global AI market

China’s telecom industry rapid growth in 2025 eludes Nokia and Ericsson as sales collapse

China ITU filing to put ~200K satellites in low earth orbit while FCC authorizes 7.5K additional Starlink LEO satellites

China gaining on U.S. in AI technology arms race- silicon, models and research

U.S. export controls on Nvidia H20 AI chips enables Huawei’s 910C GPU to be favored by AI tech giants in China

Bloomberg: China Lures Billionaires Into Race to Catch U.S. in AI

 

 

 

Nvidia CEO Huang: AI is the largest infrastructure buildout in human history; AI Data Center CAPEX will generate new revenue streams for operators

Executive Summary:

In a February 6, 2026 CNBC interview with with Scott Wapner, Nvidia CEO Jensen Huang [1.] characterized the current AI build‑out as “the largest infrastructure buildout in human history,” driven by exceptionally high demand for compute from hyperscalers and AI companies. “Through the roof” is how he described AI infrastructure spending.  It’s a “once-in-a-generation infrastructure buildout,” specifically highlighting that demand for Nvidia’s Blackwell chips and the upcoming Vera Rubin platform is “sky-high.” He emphasized that the shift from experimental AI to AI as a fundamental utility has reached a definitive inflection point for every major industry.

Jensen forecasts aa roughly 7–to- 8‑year AI investment cycle lies ahead, with capital intensity justified because deployed AI infrastructure is already generating rising cash flows for operators.  He maintains that the widely cited ~$660 billion AI data center capex pipeline is sustainable, on the grounds that GPUs and surrounding systems are revenue‑generating assets, not speculative overbuild. In his view, as long as customers can monetize AI workloads profitably, they will “keep multiplying their investments,” which underpins continued multi‑year GPU demand, including for prior‑generation parts that remain fully leased.

Note 1.  Being the undisputed leader of AI hardware (GPU chips and networking equipment via its Mellanox acquisition), Nvidia MUST ALWAYS MAKE POSITIVE REMARKS AND FORECASTS related to the AI build out boom.  Reader discretion is advised regarding Huang’s extremely bullish, “all-in on AI” remarks.

Huang reiterated that AI will “fundamentally change how we compute everything,” shifting data centers from general‑purpose CPU‑centric architectures to accelerated computing built around GPUs and dense networking. He emphasizes Nvidia’s positioning as a full‑stack infrastructure and computing platform provider—chips, systems, networking, and software—rather than a standalone chip vendor.  He accuratedly stated that Nvidia designs “all components of AI infrastructure” so that system‑level optimization (GPU, NIC, interconnect, software stack) can deliver performance gains that outpace what is possible with a single chip under a slowing Moore’s Law. The installed base is presented as productive: even six‑year‑old A100‑class GPUs are described as fully utilized through leasing, underscoring persistent elasticity of AI compute demand across generations.

AI Poster Childs – OpenAI and Anthropic:

Huang praised OpenAI and Anthropic, the two leading artificial intelligence labs, which both use Nvidia chips through cloud providers. Nvidia invested $10 billion in Anthropic last year, and Huang said earlier this week that the chipmaker will invest heavily in OpenAI’s next fundraising round.

“Anthropic is making great money. Open AI is making great money,” Huang said. “If they could have twice as much compute, the revenues would go up four times as much.”

He said that all the graphics processing units that Nvidia has sold in the past — even six-year old chips such as the A100 — are currently being rented, reflecting sustained demand for AI computing power.

“To the extent that people continue to pay for the AI and the AI companies are able to generate a profit from that, they’re going to keep on doubling, doubling, doubling, doubling,” Huang said.

Economics, utilization, and returns:

On economics, Huang’s central claim is that AI capex converts into recurring, growing revenue streams for cloud providers and AI platforms, which differentiates this cycle from prior overbuilds. He highlights very high utilization: GPUs from multiple generations remain in service, with cloud operators effectively turning them into yield‑bearing infrastructure.

This utilization and monetization profile underlies his view that the capex “arms race” is rational: when AI services are profitable, incremental racks of GPUs, network fabric, and storage can be modeled as NPV‑positive infrastructure projects rather than speculative capacity. He implies that concerns about a near‑term capex cliff are misplaced so long as end‑market AI adoption continues to inflect.

Competitive and geopolitical context:

Huang acknowledges intensifying global competition in AI chips and infrastructure, including from Chinese vendors such as Huawei, especially under U.S. export controls that have reduced Nvidia’s China revenue share to roughly half of pre‑control levels. He frames Nvidia’s strategy as maintaining an innovation lead so that developers worldwide depend on its leading‑edge AI platforms, which he sees as key to U.S. leadership in the AI race.

He also ties AI infrastructure to national‑scale priorities in energy and industrial policy, suggesting that AI data centers are becoming a foundational layer of economic productivity, analogous to past buildouts in electricity and the internet.

Implications for hyperscalers and chips:

Hyperscalers (and also Nvidia customers) Meta , Amazon, Google/Alphabet and Microsoft recently stated that they plan to dramatically increase spending on AI infrastructure in the years ahead. In total, these hyperscalers could spend $660 billion on capital expenditures in 2026 [2.] , with much of that spending going toward buying Nvidia’s chips. Huang’s message to them is that AI data centers are evolving into “AI factories” where each gigawatt of capacity represents tens of billions of dollars of investment spanning land, compute, and networking. He suggests that the hyperscaler industry—roughly a $2.5 trillion sector with about $500 billion in annual capex transitioning from CPU to GPU‑centric generative AI—still has substantial room to run.

Note 2.  An understated point is that while these hyperscalers are spending hundered of billions of dollars on AI data centers and Nvidia chips/equipment they are simultaneously laying off tens of thousands of employees.  For example, Amazon recently announced 16,000 job cuts this year after 14,000 layoffs last October.

From a chip‑level perspective, he argues that Nvidia’s competitive moat stems from tightly integrated hardware, networking, and software ecosystems rather than any single component, positioning the company as the systems architect of AI infrastructure rather than just a merchant GPU vendor.

References:

https://www.cnbc.com/2026/02/06/nvidia-rises-7percent-as-ceo-says-660-billion-capex-buildout-is-sustainable.html

Big tech spending on AI data centers and infrastructure vs the fiber optic buildout during the dot-com boom (& bust)

Analysis: Cisco, HPE/Juniper, and Nvidia network equipment for AI data centers

Networking chips and modules for AI data centers: Infiniband, Ultra Ethernet, Optical Connections

Will billions of dollars big tech is spending on Gen AI data centers produce a decent ROI?

Superclusters of Nvidia GPU/AI chips combined with end-to-end network platforms to create next generation data centers

184K global tech layoffs in 2025 to date; ~27.3% related to AI replacing workers

 

 

Analysis: Edge AI and Qualcomm’s AI Program for Innovators 2026 – APAC for startups to lead in AI innovation

Qualcomm is a strong believer in Edge AI as an enabler of faster, more secure, and energy-efficient processing directly on devices—rather than the cloud—unlocking real-time intelligence for industries like robotics and smart cities.

In support of that vision, the fabless SoC company announced the official launch of its Qualcomm AI Program for Innovators (QAIPI) 2026 – APAC, a regional startup incubation initiative that supports startups across Japan, Singapore, and South Korea in advancing the development and commercialization of innovative edge AI solutions.

Building on Qualcomm’s commitment to edge AI innovation, the second edition of QAIPI-APAC invites startups to develop intelligent solutions across a broad range of edge-AI applications using Qualcomm Dragonwing™ and Snapdragon® platforms, together with the new Arduino® UNO Q development board, strengthening their pathway toward global commercialization.

Startups gain comprehensive support and resources, including access to Qualcomm Dragonwing™ and Snapdragon® platforms, the Arduino® UNO Q development board, technical guidance and mentorship, a grant of up to US$10,000, and eligibility for up to US$5,000 in patent filing incentives, accelerating AI product development and deployment.

Applications are open now through April 30, 2026 and will be evaluated based on innovation, technical feasibility, potential societal impact, and commercial relevance. The program will be implemented in two phases. The application phase is open to eligible startups incorporated and registered in Japan, Singapore, or South Korea. Shortlisted startups will enter the mentorship phase, receiving one-on-one guidance, online training, technical support, and access to Qualcomm-powered hardware platforms and development kits for product development. They will also receive a shortlist grant of up to US$10,000 and may be eligible for a patent filing incentive of up to US$5,000. At the conclusion of the program, shortlisted startups may be invited to showcase their innovations at a signature Demo Day in late 2026, engaging with industry leaders, investors, and potential collaborators across the APAC innovation ecosystem.

Comment and Analysis:

Qualcomm is a strong believer in Edge AI—the practice of running AI models directly on devices (smartphones, cars, IoT, PCs) rather than in the cloud—because they view it as the next major technological paradigm shift, overcoming limitations inherent in cloud computing. Despite the challenges of power consumption and processing limitations, Qualcomm’s strategy hinges on specialized, heterogenous computing rather than relying solely on RISC-based CPU cores.

Key Issues for Qualcomm’s Edge AI solutions:

1.  The “Heterogeneous” Solution to Processing Limits
While it is true that standard CPU cores (even RISC-based ones) are inefficient for AI, Qualcomm does not use them for AI workloads. Instead, they use a heterogeneous architecture:
  • Qualcomm® AI Engine: This combines specialized hardware, including the Hexagon NPU (Neural Processing Unit), Adreno GPU, and CPU. The NPU is specifically designed to handle high-performance, complex AI workloads (like Generative AI) far more efficiently than a generic CPU.
  • Custom Oryon CPU: The latest Snapdragon X Elite platform features customized cores that provide high performance while outperforming traditional x86 solutions in power efficiency for everyday tasks.
2. Overcoming Power Consumption (Performance/Watt)
Qualcomm focus on “Performance per Watt” rather than raw power.
  • Specialization Saves Power: By using specialized AI engines (NPUs) rather than general-purpose CPU/GPU cores, Qualcomm can run inference tasks at a fraction of the power cost.
  • Lower Overall Energy: Doing AI at the edge can save total energy by avoiding the need to send data to a power-hungry data center, which requires network infrastructure, and then sending it back.
  • Intelligent Efficiency: The Snapdragon 8 Elite, for example, saw a 27% reduction in power consumption while increasing AI performance significantly.
3. Critical Advantages of Edge over Cloud
Qualcomm believes edge is essential because cloud AI cannot solve certain critical problems:
  • Instant Responsiveness (Low Latency): For autonomous vehicles or industrial robotics, a few milliseconds of latency to the cloud can be catastrophic. Edge AI provides real-time, instantaneous analysis.
  • Privacy and Security: Data never leaves the device. This is crucial for privacy-conscious users (biometrics) and compliance (GDPR), which is a major advantage over cloud-based AI.
  • Offline Capability: Edge devices, such as agricultural sensors or smart home devices in remote areas, continue to function without internet connectivity.
4. Market Expansion and Economic Drivers
  • Diversification: With the smartphone market maturing, Qualcomm sees the “Connected Intelligent Edge” as a huge growth opportunity, extending their reach into automotive, IoT, and PCs.
  • “Ecosystem of You”: Qualcomm aims to connect billions of devices, making AI personal and context-aware, rather than generic.
5. Bridging the Gap: Software & Model Optimization
Qualcomm is not just providing hardware; they are simplifying the deployment of AI:
  • Qualcomm AI Hub: This makes it easier for developers to deploy optimized models on Snapdragon devices.
  • Model Optimization: They specialize in making AI models smaller and more efficient (using quantization and specialized AI inference) to run on devices without requiring massive, cloud-sized computing power.
In summary, Qualcomm believes in Edge AI because they are building highly specialized hardware designed to excel within tight power and thermal constraints.
……………………………………………………………………………………………………………………………………………………………………………

References:

https://www.prnewswire.com/apac/news-releases/qualcomm-ai-program-for-innovators-2026–apac-officially-kicks-off—empowering-startups-across-japan-singapore-and-south-korea-to-lead-the-ai-innovation-302676025.html

Qualcomm CEO: AI will become pervasive, at the edge, and run on Snapdragon SoC devices

Huawei, Qualcomm, Samsung, and Ericsson Leading Patent Race in $15 Billion 5G Licensing Market

Private 5G networks move to include automation, autonomous systems, edge computing & AI operations

Nvidia’s networking solutions give it an edge over competitive AI chip makers

Nvidia AI-RAN survey results; AI inferencing as a reinvention of edge computing?

CES 2025: Intel announces edge compute processors with AI inferencing capabilities

Qualcomm CEO: expect “pre-commercial” 6G devices by 2028

 

Analysis: SpaceX FCC filing to launch up to 1M LEO satellites for solar powered AI data centers in space

SpaceX has applied to the Federal Communications Commission (FCC) for permission to launch up to 1 million LEO satellites for a new solar-powered AI data center system in space.  The private company, 40% owned by Elon Musk, envisions an orbital data center system with “unprecedented computing capacity” needed to run large-scale AI inference and applications for billions of users, according to SpaceX’s filing entered late on Friday.

Data centers are the physical backbone of artificial intelligence, requiring massive amounts of power. “By directly harnessing near-constant solar power with little operating or maintenance costs, these satellites will achieve transformative cost and energy efficiency while significantly reducing the environmental impact associated with terrestrial data centers,” the FCC filing said. Musk would need the telecom regulator’s approval to move forward.

Credit: Blueee/Alamy Stock Photo

The proposed new satellites would operate in “narrow orbital shells” of up to 50 kilometers each. The satellites would operate at altitudes of between 500 kilometers and 2,000 kilometers, and 30 degrees, and “sun-synchronous orbit inclinations” to capture power from the sun. The system is designed to be interconnected via optical links with existing Starlink broadband satellites, which would transmit data traffic back to ground Earth stations.

SpaceX’s request bets heavily on reduced costs of Starship, the company’s next-generation reusable rocket under development.  Starship has test-launched 11 times since 2023. Musk expects the rocket, which is crucial for expanding Starlink with more powerful satellites, to put its first payloads into orbit this year.
“Fortunately, the development of fully reusable launch vehicles like Starship that can deploy millions of tons of mass per year to orbit when launching at rate, means on-orbit processing capacity can reach unprecedented scale and speed compared to terrestrial buildouts, with significantly reduced environmental impact,” SpaceX said.
SpaceX is positioning orbital AI compute as the definitive solution to the terrestrial capacity crunch, arguing that space-based infrastructure represents the most efficient path for scaling next-generation workloads. As ground-based data centers face increasing grid density constraints and power delivery limitations, SpaceX intends to leverage high-availability solar irradiation to bypass Earth’s energy bottlenecks.The company’s technical rationale hinges on several key architectural advantages:
  • Energy Density & Sustainability: By tapping into “near-constant solar power,” SpaceX aims to utilize a fraction of the Sun’s output—noting that even a millionth of its energy exceeds current civilizational demand by four orders of magnitude.
  • Thermal Management: To address the cooling requirements of high-density AI clusters, these satellites will utilize radiative heat dissipation, eliminating the water-intensive cooling loops required by terrestrial facilities.
  • Opex & Scalability: The financial viability of this orbital layer is tethered to the Starship launch platform. SpaceX anticipates that the radical reduction in $/kg launch costs provided by a fully reusable heavy-lift vehicle will enable rapid scaling and ensure that, within years, the lowest LCOA (Levelized Cost of AI) will be achieved in orbit.
The transition to orbital AI compute introduces a fundamental shift in network topology, moving processing from terrestrial hubs to a decentralized, space-based edge layer. The latency implications are characterized by three primary architectural factors:
  • Vacuum-Speed Data Transmission: In a vacuum, light propagates roughly 50% faster than through terrestrial fiber optic cables. By utilizing Starlink’s optical inter-satellite links (OISLs)—a “petabit” laser mesh—data can bypass terrestrial bottlenecks and subsea cables. This potentially reduces intercontinental latency for AI inference to under 50ms, surpassing many long-haul terrestrial routes.
  • Edge-Native Processing & Data Gravity: Current workflows require downlinking massive raw datasets (e.g., Synthetic Aperture Radar imagery) for terrestrial processing, a process that can take hours. Shifting to orbital edge computing allows for “in-situ” AI inference, processing data onboard to deliver actionable insights in minutes rather than hours. This “Space Cloud” architecture eliminates the need to route raw data back to the Earth’s internet backbone, reducing data transmission volumes by up to 90%.
  • LEO Proximity vs. Terrestrial Hops: While terrestrial fiber remains the “gold standard” for short-range latency (typically 1–10ms), it is often hindered by inefficient routing and multiple hops. SpaceX’s LEO constellation, operating at altitudes between 340km and 614km, currently delivers median peak-hour latencies of ~26ms in the US. Future orbital configurations may feature clusters at varying 50km intervals to optimize for specific workload and latency tiers.

………………………………………………………………………………………………………………………………………………………………………………………

The SpaceX FCC filing on Friday follows an exclusive report by Reuters that Elon Musk is considering merging SpaceX with his xAI (Grok chatbot) company ahead of an IPO later this year. Under the proposed merger, shares of xAI would be exchanged for shares in SpaceX. Two entities have been set up in Nevada to facilitate the transaction, Reuters said.  Musk also runs electric automaker Tesla, tunnel company The Boring Co. and neurotechnology company Neuralink.

………………………………………………………………………………………………………………………………………………………………………………………

References:

https://www.reuters.com/business/aerospace-defense/spacex-seeks-fcc-nod-solar-powered-satellite-data-centers-ai-2026-01-31/

https://www.lightreading.com/satellite/spacex-seeks-fcc-approval-for-mega-ai-data-center-constellation

https://www.reuters.com/world/musks-spacex-merger-talks-with-xai-ahead-planned-ipo-source-says-2026-01-29/

Google’s Project Suncatcher: a moonshot project to power ML/AI compute from space

Blue Origin announces TeraWave – satellite internet rival for Starlink and Amazon Leo

China ITU filing to put ~200K satellites in low earth orbit while FCC authorizes 7.5K additional Starlink LEO satellites

Amazon Leo (formerly Project Kuiper) unveils satellite broadband for enterprises; Competitive analysis with Starlink

Telecoms.com’s survey: 5G NTNs to highlight service reliability and network redundancy

 

Huge significance of EchoStar’s AWS-4 spectrum sale to SpaceX

U.S. BEAD overhaul to benefit Starlink/SpaceX at the expense of fiber broadband providers

Telstra selects SpaceX’s Starlink to bring Satellite-to-Mobile text messaging to its customers in Australia

SpaceX launches first set of Starlink satellites with direct-to-cell capabilities

AST SpaceMobile to deliver U.S. nationwide LEO satellite services in 2026

GEO satellite internet from HughesNet and Viasat can’t compete with LEO Starlink in speed or latency

How will fiber and equipment vendors meet the increased demand for fiber optics in 2026 due to AI data center buildouts?

Subsea cable systems: the new high-capacity, high-resilience backbone of the AI-driven global network

China’s open source AI models to capture a larger share of 2026 global AI market

Overview of AI Models – China vs U.S. :

Chinese AI language models (LMs) have advanced rapidly and are now contesting with the U.S. for global market leadership.  Alibaba’s Qwen-Image-2512 is emerging as a top-performing, free, open-source model capable of high-fidelity human, landscape, and text rendering. Other key, competitive models include Zhipu AI’s GLM-Image (trained on domestic chips), ByteDance’s Seedream 4.0, and UNIMO-G.

Today, Alibaba-backed Moonshot AI released an upgrade of its flagship AI model, heating up a domestic arms race ahead of an expected rollout by Chinese AI hotshot DeepSeek. The latest iteration of Moonshot’s Kimi can process text, images, and videos simultaneously from a single prompt, the company said in a statement, aligning with a trend toward so-called omni models pioneered by industry leaders like OpenAI and Alphabet Inc.’s Google.

Moonshot AI Kimi website. Photographer: Raul Ariano/Bloomberg

………………………………………………………………………………………………………………………………………………………………………………………………………………………………………………..

Chinese AI models are rapidly narrowing the gap with Western counterparts in quality and accessibility.  That shift is forcing U.S. AI leaders like Alphabet’s Google, Microsoft’s Copilot, OpenAI, and Anthropic to fight harder to  maintain their technological lead in AI.  That’s despite their humongous spending on AI data centers, related AI models and infrastructure.

In early 2025, investors seized on DeepSeek’s purportedly lean $5.6 million LM training bill as a sign that Nvidia’s high-end GPUs were already a relic and that U.S. hyperscalers had overspent on AI infrastructure. Instead, the opposite dynamic played out: as models became more capable and more efficient, usage exploded, proving out a classic Jevons’ Paradox and validating the massive data-center build-outs by Microsoft, Amazon, and Google.

The real competitive threat from DeepSeek and its peers is now coming from a different direction. Many Chinese foundation models are released as “open source” or “open weight” AI models which makes them effectively free to download, easy to modify, and cheap to run at scale. By contrast, most leading U.S. players keep tight control over their systems, restricting access to paid APIs and higher-priced subscriptions that protect margins but limit diffusion.

That strategic divergence is visible in how these systems are actually used. U.S. models such as Google’s Gemini, Anthropic’s Claude, and OpenAI’s GPT series still dominate frontier benchmarks [1′] and high‑stakes reasoning tasks. According to a recently published report by OpenRouter, a third-party AI model aggregator, and venture capital firm Andreessen Horowitz. Chinese open-source models have  captured roughly 30% of the “working” AI market. They are especially strong in coding support and roleplay-style assistants—where developers and enterprises optimize for cost efficiency, local customization, and deployment freedom rather than raw leaderboard scores.

Note 1. A frontier benchmark for AI models is a high-difficulty evaluation designed to test the absolute limits of artificial intelligence in complex,, often unsolved, reasoning tasks. FrontierMath, for example, is a prominent benchmark focusing on expert-level mathematics, requiring AI to solve hundreds of unpublished problems that challenge, rather than merely measure, current capabilities.

China’s open playbook:

China’s more permissive stance on model weights is not just a pricing strategy — it’s an acceleration strategy. Opening weights turns the broader developer community into an extension of the R&D pipeline, allowing users to inspect internals, pressure‑test safety, and push incremental improvements upstream.

As Kyle Miller at Georgetown’s Center for Security and Emerging Technology argues, China is effectively trading away some proprietary control to gain speed and breadth: by letting capability diffuse across the ecosystem, it can partially offset the difficulty of going head‑to‑head with tightly controlled U.S. champions like OpenAI and Anthropic. That diffusion logic is particularly potent in a system where state planners, big tech platforms, and startups are all incentivized to show visible progress in AI.

Even so, the performance gap has not vanished. Estimates compiled by Epoch AI suggest that Chinese models, on average, trail leading U.S. releases by about seven months. The window briefly narrowed during DeepSeek’s R1 launch in early 2025, when it looked like Chinese labs might have structurally compressed the lag; since then, the gap has widened again as U.S. firms have pushed ahead at the frontier.

Capital, chips, and the power problem:

The reason the U.S. lead has held is massive AI infrastructure spending. Consensus forecasts put capital expenditure by largely American hyperscalers at roughly $400 billion in 2025 and more than $520 billion in 2026, according to Goldman Sachs Research. By comparison, UBS analysts estimate that China’s major internet platforms collectively spent only about $57 billion last year—a fraction of U.S. outlays, even if headline Chinese policy rhetoric around AI is more aggressive.

But sustaining that level of investment runs into a physical constraint that can’t be hand‑waved away: electricity. The newest data-center designs draw more than a gigawatt of power each—about the output of a nuclear reactor—turning grid capacity into a strategic bottleneck. China now generates more than twice as much power as the U.S., and its centralized planning system can more readily steer incremental capacity toward AI clusters than America’s fragmented, heavily regulated electricity market.

That asymmetry is already shaping how some on Wall Street frame the race. Christopher Woods, global head of equity strategy at Jefferies, recently reiterated that China’s combination of open‑source models and abundant cheap power makes it a structurally formidable AI competitor. In his view, the “DeepSeek moment” of early last year remains a warning that markets have largely chosen to ignore as they rotate back into U.S. AI mega‑caps.

A fragile U.S. AI advantage:

For now, U.S. companies still control the most important chokepoint in the stack: advanced AI accelerators. Access to Nvidia’s cutting‑edge GPUs remains a decisive advantage.  Yesterday, Microsoft announced the Maia 200 chip – their first silicon and system platform optimized specifically for AI inference.  The chip was  was designed for efficiency, both in terms of its ability to deliver tokens per dollar and performance per watt of power used.

“Maia 200 can deliver 30% better performance per dollar than the latest generation hardware in our fleet today,” Microsoft EVP for Cloud and AI Scott Guthrie wrote in a blog post.

Image Credit: Microsoft

……………………………………………………………………………………………………………………………………………………………………………………………………………………………….

Leading Chinese AI research labs have struggled to match training results using only domestic designed silicon. DeepSeek, which is developing the successor to its flagship model and is widely expected to release it around Lunar New Year, reportedly experimented with chips from Huawei and other local vendors before concluding that performance was inadequate and turning to Nvidia GPUs for at least part of the training run.

That reliance underscores the limits of China’s current self‑reliance push—but it also shouldn’t be comforting to U.S. strategists. Chinese firms are actively working around the hardware gap, not waiting for it to close. DeepSeek’s latest research focuses on training larger models with fewer chips through more efficient memory design, an incremental but important reminder that architectural innovation can partially offset disadvantages in raw compute.

From a technology‑editorial perspective, the underlying story is not simply “China versus the U.S.” at the model frontier. It is a clash between two AI industrial strategies: an American approach that concentrates capital, compute, and control in a handful of tightly integrated platforms, and a Chinese approach that leans on open weights, diffusion, and state‑backed infrastructure to pull the broader ecosystem forward.

The question for 2026 is whether U.S. AI firms’ lead in capability and chips can keep outrunning China’s advantages in openness and power—or whether the market will again wait for a shock like DeepSeek to re‑price that risk.

Deepseek and Other Chinese AI Models:

DeepSeek published research this month outlining a method of training larger models using fewer chips through a more efficient memory design. “We view DeepSeek’s architecture as a new, promising engineering solution that could enable continued model scaling without a proportional increase in GPU capacity,” wrote UBS analyst Timothy Arcuri.

Export controls haven’t prevented Chinese companies from training advanced models, but challenges emerge when the models are deployed at scale. Zhipu AI, which released its open-weight GLM 4.7 model in December, said this month it was rationing sales of its coding product to 20% of previous capacity after demand from users overwhelmed its servers.

Moonshot, Zhipu AI and MiniMax Group Inc are among a handful of AI LM front-runners in a hotly contested battle among Chinese large language model makers, which at one point was dubbed the “War of One Hundred Models.”

“I don’t see compute constraints limiting [Chinese companies’] ability to make models that are better and compete near the U.S. frontier,” Georgetown’s Miller says. “I would say compute constraints hit on the wider ecosystem level when it comes to deployment.”

Gaining access to Nvidia AI chips:

U.S. President Donald Trump’s plan to allow Nvidia to sell its H200 chips to China could be pivotal. Alibaba Group and ByteDance, TikTok’s parent company, have privately indicated interest in ordering more than 200,000 units each, Bloomberg reported.  The H200 outperforms any Chinese-produced AI chip, with a roughly 32% processing-power advantage over Huawei’s Ascend 910C.

With access to Nvidia AI chips, Chinese labs could build AI-training supercomputers as capable as American ones at 50% extra cost compared with U.S.-made ones, according to the Institute for Progress. Subsidies by the Chinese government could cover that differential, leveling the playing field, the institute says.

Conclusions:

A combination of open-source innovation and loosened chip controls could create a cheaper, more capable Chinese AI ecosystem. The possibility is emerging just as OpenAI and Anthropic consider public stock listings (IPOs) and U.S. hyperscalers such as Microsoft and Meta Platforms face pressure to justify heavy spending.

The risk for U.S. AI leaders is no longer theoretical; China’s open‑weight, low‑cost model ecosystem is already eroding the moat that Google, OpenAI, and Anthropic thought they were building. By prioritizing diffusion over tight control, Chinese firms are seeding a broad developer base, compressing iteration cycles, and normalizing expectations that powerful models should be cheap—or effectively free—to adapt and deploy.

U.S. AI leaders could face pressure on pricing and profit margins from China AI competitors while having to deal with AI infrastructure costs and power constraints. Their AI advantage remains real, but fragile—highly exposed to regulatory shifts, export controls, and any breakthrough in China’s workarounds on hardware and training efficiency. The uncomfortable prospect for U.S. AI incumbents is that they could win the race for the best models and still lose ground in the market if China’s diffusion‑driven strategy defines how AI is actually consumed at scale.

…………………………………………………………………………………………………………………………………………………………………………………………………………………..

References:

https://www.barrons.com/articles/deepseek-ai-gemini-chatgpt-stocks-ccde892c

https://blogs.microsoft.com/blog/2026/01/26/maia-200-the-ai-accelerator-built-for-inference/

https://www.bloomberg.com/news/articles/2026-01-27/china-s-moonshot-unveils-new-ai-model-ahead-of-deepseek-release

https://www.scmp.com/tech/tech-trends/article/3335602/chinas-open-source-models-make-30-global-ai-usage-led-qwen-and-deepseek

China gaining on U.S. in AI technology arms race- silicon, models and research

U.S. export controls on Nvidia H20 AI chips enables Huawei’s 910C GPU to be favored by AI tech giants in China

Goldman Sachs: Big 3 China telecom operators are the biggest beneficiaries of China’s AI boom via DeepSeek models; China Mobile’s ‘AI+NETWORK’ strategy

Bloomberg: China Lures Billionaires Into Race to Catch U.S. in AI

Comparing AI Native mode in 6G (IMT 2030) vs AI Overlay/Add-On status in 5G (IMT 2020)

Telecom operators investing in Agentic AI while Self Organizing Network AI market set for rapid growth

Telecom companies are planning to use Agentic AI [1.] for customer experience and network automation. A recent RADCOM survey shows 71% of network operators plan to deploy agentic AI in 2026, while 14% have already begun, prioritizing areas that directly influence trust and customer satisfaction: security and fraud prevention (57%) and customer service and support (56%).  The top use cases are automated customer complaint resolution and autonomous fault resolution.

Operators are betting on agentic AI to remove friction before customers feel it, with the highest-value use cases reflecting this shift, including:

  • 57% – automated customer complaint resolution
  • 54% – autonomous fault resolution before it impacts service
  • 52% – predicting experience to prevent churn

This technology is shifting networks from simply detecting issues to preventing them before customers notice. In contact centers, 2026 is expected to see a rise in human and AI agent collaboration to improve efficiency and customer service.

Note 1.  Agentic AI refers to autonomous artificial intelligence systems that can perceive, reason, plan, and act independently to achieve complex goals with minimal human intervention, going beyond simple command-response to manage multi-step tasks, use various tools, and adapt to new information for proactive automation in dynamic environments. These intelligent agents function like digital coworkers, coordinating internally and with other systems to execute sophisticated workflows.

……………………………………………………………………………………………………………………………………………………………………………………………

ResearchAndMarkets.com has just published a “Self-Organizing Network Artificial Intelligence (AI) Global Market Report 2025.” The market research firm says that the self-organizing network AI [2.] is forecast to expand from $5.19 billion in 2024 to $6.18 billion in 2025, at a CAGR of 19.2%. This surge is driven by the integration of machine learning and AI in telecom networks, smart network management investment, and the growing demand for features like self-healing and self-optimization, as well as predictive maintenance technologies.driven by the expansion of 5G, increasing automation demands, and AI integration for network optimization. Opportunities include AI-driven RRM and predictive maintenance. Asia-Pacific emerges as the fast-growing region, boosting telecom innovations amid global trade shifts.

Note 2.  Self-organizing network AI leverages software, hardware, and services to dynamically optimize and manage telecom networks, applicable across various network types and deployment modes. The market encompasses a broad range of solutions, from network optimization software to AI-driven planning products, underscoring its expansive potential.

Looking further ahead, the market is expected to reach $12.32 billion by 2029, with a CAGR of 18.8%. Key drivers during this period include heightened demand for automation, increased 5G deployments, and growing network densification, accompanied by rising data traffic and subscriber numbers. Trends such as AI-driven network automation advancements, machine learning integration for real-time optimization, and the rise of generative AI for analytics are reshaping the landscape.

The expansion of 5G networks plays a pivotal role in propelling this growth. These networks, characterized by high-speed data and ultra-low latency, significantly enhance the capabilities of self-organizing network AI. The integration facilitates real-time data processing, supporting automation, optimization, and predictive maintenance, thereby improving service quality and user experience. A notable development in 2023 saw UK outdoor 5G coverage rise to 85-93%, reflecting growing demand and technological advancement.

Huawei Technologies and other major tech companies, are pioneering innovative solutions like AI-driven radio resource management (RRM), which optimizes network performance and enhances user experience. These solutions rely on AI and machine learning for dynamic spectrum and network resource management. For instance, Huawei’s AI Core Network, introduced at MWC 2025, marks a substantial leap in intelligent telecommunications, integrating AI into core systems for seamless connectivity and real-time decision-making.

Strategic acquisitions are also shaping the market, exemplified by Amdocs Limited acquiring TEOCO Corporation in 2023 to bolster its network optimization and analytics capabilities. This acquisition aims to enhance end-to-end network intelligence and operational efficiency.

Leading players in the market include Huawei, Cisco Systems Inc., Qualcomm Incorporated, and many others, driving innovation and competition. Europe held the largest market share in 2024, with Asia-Pacific poised to be the fastest-growing region through the forecast period.

References:

Operator Priorities for 2026 and Beyond: Data, Automation, Customer Experience

https://uk.finance.yahoo.com/news/self-organizing-network-artificial-intelligence-105400706.html

Ericsson integrates agentic AI into its NetCloud platform for self healing and autonomous 5G private network

Agentic AI and the Future of Communications for Autonomous Vehicles (V2X)

IDC Report: Telecom Operators Turn to AI to Boost EBITDA Margins

Omdia: How telcos will evolve in the AI era

Palo Alto Networks and Google Cloud expand partnership with advanced AI infrastructure and cloud security

Hyperscaler capex > $600 bn in 2026 a 36% increase over 2025 while global spending on cloud infrastructure services skyrockets

Hyperscaler capex for the “big five” (Amazon, Alphabet/Google, Microsoft, Meta/Facebook, Oracle) is now widely forecast to exceed $600 bn in 2026, a 36% increase over 2025. Roughly 75%, or $450 bn, of that spend is directly tied to AI infrastructure (i.e., servers, GPUs, datacenters, equipment), rather than traditional cloud.  Hyperscalers are increasingly leaning on debt markets to bridge the gap between rapidly rising AI capex budgets and internal free cash flow, transforming historically cash-funded business models into ones utilizing leverage, albeit with still very strong balance sheets. Aggregate capex for “the big five”, after buybacks and dividends are included, are now above projected cash flows, thereby necessitating external funding needs.

……………………………………………………………………………………………………………………………………………………………………………………………………………………..

According to market research from Omdia (owned by Informa) global spending on cloud infrastructure services reached $102.6 billion in Q3 2025 — a 25% year-on-year increase. It was the fifth consecutive quarter in which cloud spending growth remained above 20%.  Omdia says it “reflects a significant shift in the technology landscape as enterprise demand for AI moves beyond early experimentation toward scaled production deployment.” AWS, Microsoft Azure, and Google Cloud – maintained their market rankings from the previous quarter, and collectively accounted for 66% of global cloud infrastructure spending. Together, the three firms had 29% year-on-year growth in their cloud spending.

Hyperscaler AI strategies are shifting from a focus on incremental model performance to platform-driven, production-ready approaches. Enterprises are now evaluating AI platforms based not solely on model capabilities, but also on their support for multi-model strategies and agent-based applications. This evolution is accelerating hyperscalers’ move toward platform-level AI capabilities. According to the report, Amazon Web Services (AWS), Microsoft Azure, and Google Cloud are integrating proprietary foundation models with a growing range of third-party and open-weight models to meet these new demands.

“Collaboration across the ecosystem remains critical,” said Rachel Brindley, Senior Director at Omdia. “Multi-model support is increasingly viewed as a production requirement rather than a feature, as enterprises seek resilience, cost control, and deployment flexibility across generative AI workloads.”

Facing challenges with practical application, major cloud providers are boosting resources for AI agent lifecycle management, including creation and operationalization, as enterprise-level deployment proves more intricate than anticipated.

Yi Zhang, Senior Analyst at Omdia, said, “Many enterprises still lack standardized building blocks that can support business continuity, customer experience, and compliance at the same time, which is slowing the real-world deployment of AI agents. This is where hyperscalers are increasingly stepping in, using platform-led approaches to make it easier for enterprises to build and run agents in production environments.”

This past October, Omdia released a report forecasting that growth of cloud adoption among communications service providers (CSPs) will double this year. It also forecasted a compound annual growth rate (CAGR) of 7.3% to 2030, resulting in the telco cloud market being worth $24.8 billion.

………………………………………………………………………………………………………………………………………………………………………………………………………………………..

Editor’s Note:  Does anyone remember the stupendous increase in fiber optic spending from 1998-2001 till that bubble burst?  Caveat Emptor!

………………………………………………………………………………………………………………………………………………………………………………………………………………………..

References:

https://www.mufgamericas.com/sites/default/files/document/2025-12/AI_Chart_Weekly_12_19_Financing_the_AI_Supercycle.pdf

https://www.telecoms.com/public-cloud/global-cloud-infrastructure-spend-up-25-in-q3

https://www.telecoms.com/public-cloud/telco-investment-in-cloud-infrastructure-is-accelerating-omdia

AI infrastructure spending boom: a path towards AGI or speculative bubble?

Expose: AI is more than a bubble; it’s a data center debt bomb

Can the debt fueling the new wave of AI infrastructure buildouts ever be repaid?

AI spending boom accelerates: Big tech to invest an aggregate of $400 billion in 2025; much more in 2026!

Gartner: AI spending >$2 trillion in 2026 driven by hyperscalers data center investments

AI spending is surging; companies accelerate AI adoption, but job cuts loom large

Will billions of dollars big tech is spending on Gen AI data centers produce a decent ROI?

Big tech spending on AI data centers and infrastructure vs the fiber optic buildout during the dot-com boom (& bust)

Canalys & Gartner: AI investments drive growth in cloud infrastructure spending

Sovereign AI infrastructure for telecom companies: implementation and challenges

AI Echo Chamber: “Upstream AI” companies huge spending fuels profit growth for “Downstream AI” firms

Custom AI Chips: Powering the next wave of Intelligent Computing

 

Sovereign AI infrastructure for telecom companies: implementation and challenges

Sovereign AI infrastructure refers to the domestic capability of a nation or an organization to own and control the entire technology stack for artificial intelligence (AI) systems within its own borders, subject to local laws and governance. This includes the physical data centers, specialized hardware (like GPUs), software, data, and skilled workforce.  Sovereign AI infrastructure involves a full “stack” designed to ensure national control and reduce reliance on foreign providers. A few key features:

  • Policies and technical controls (e.g., data localization, encryption) to ensure that sensitive data used for training and inference remains within the jurisdiction.
  • Development and hosting of proprietary or locally tailored AI models and software frameworks that align with national values, languages, and ethical standards.
  • Workforce Development: Investing in domestic talent, including data scientists, engineers, and legal experts, to build and maintain the local AI ecosystem.
  • Regulatory Framework: A comprehensive legal and ethical framework for AI development and deployment that ensures compliance with national laws and standards.

Why It’s Important – The pursuit of sovereign AI infrastructure is driven by several strategic considerations for both governments and private enterprises:

  • National Security: To ensure that critical systems in defense, intelligence, and public infrastructure are not dependent on potentially adversarial foreign technologies or subject to extraterritorial access laws (like the U.S. CLOUD Act).
  • Economic Competitiveness: To foster a domestic tech industry, create high-skilled jobs, protect intellectual property, and capture the significant economic benefits of AI-driven growth.
  • Data Privacy and Compliance: To comply with stringent local data protection regulations (e.g., GDPR in the EU) and build public trust by ensuring citizen data is handled securely and according to local laws. Cultural Preservation: To train AI models on local datasets and languages, preserving cultural nuances and avoiding bias found in generalized, globally trained models.

Image Credit: Nvidia

………………………………………………………………………………………………………………………………………………………………………………………………………..

Governments around the world are starting to build sovereign AI infrastructure, and according to a new report from Morningstar DBRS, which opines that major telecommunications companies are uniquely positioned to benefit from that shift.  Here are a few take-aways from the report:

  • Sovereign AI funding opens a new growth path for telcos – Governments investing in domestic AI infrastructure are increasingly turning to operators, whose network and regulatory strengths position them to capture a large share of this emerging market.
  • Telcos’ capabilities align with sovereignty needs – Their expertise in large-scale networks, local presence, and established government relationships give them an edge over hyperscalers for sensitive, sovereignty-focused AI projects.
  • Early adopters gain advantage – Operators in Canada and Europe are already moving into sovereign AI, positioning themselves to secure higher-margin enterprise and government workloads as national AI buildouts accelerate.
Infrastructure advantages provide a strategic head start for telecommunications companies. Telcos currently manage extensive data centers, fiber optic networks, and computing infrastructure nationwide. Leveraging these established physical assets can significantly reduce the barriers to implementing sovereign AI solutions, contrasting favorably with the greenfield development required by other entities. 
The sophisticated data governance expertise within telcos is well-suited for the stringent requirements of sovereign AI. Their decades of experience managing and processing massive datasets have resulted in mature data handling practices directly applicable to the data infrastructure demands of secure, sovereign AI systems.
Furthermore, existing edge computing capabilities offer a distinct competitive advantage. Telecom networks facilitate localized AI processing near data sources while adhering to data residency requirements—a crucial combination for sovereign AI deployments.  This translates to “embedding AI within their network fabric for both optimization and distributed inference,” enabling AI consumption that offers lower latency, reduced cost, and applicability for high-sensitivity use cases in sectors like government and national security.
The opportunity to integrate AI workloads with emerging 5G and 6G infrastructures creates additional strategic value. Sovereign AI represents a pivotal opportunity for telecom operators to position themselves as central players in national AI strategies, evolving their role beyond primary connectivity provisioning.
……………………………………………………………………………………………………………………………………………………………………………….
Implementing sovereign AI presents substantial challenges despite its strategic potential. Key bottlenecks and technical complexities include:
  • Infrastructure Demands: Building robust domestic AI ecosystems requires specialized expertise spanning hardware, software, data governance, and policy.
  • Resource Constraints: Dr. Matt Hasan, CEO at aiRESULTS and a former AT&T executive, highlights specific bottlenecks:
    • Compute Density at Scale.
    • Spectrum Allocation amidst political pressures.
    • Energy Demand exceeding existing grid capacity.
  • Intensified Reliability Requirements: Sovereign AI implementation places heightened demands on telecom providers for system uptime, reliability, quality, and data privacy. This necessitates a focus on efficient power consumption, resilient routing and backups, robust encryption, and comprehensive cybersecurity measures.
  • Supply Chain Vulnerabilities: Geopolitical tensions introduce risks to the supply of critical components such as GPUs and specialized chips, underscoring the interconnected nature of global hardware supply chains.
  • The rapid evolution of AI technology mandates continuous investment and technical agility to ensure sovereign deployments remain current.
Competitive landscape dynamics:
  • The interplay between global hyperscalers and regional telecom operators is expected to shift.
  • Hasan predicts a collaborative model, with regional telcos leveraging their position as sovereign partners through joint ventures, rather than an outright displacement of hyperscalers.
Ultimately, the objective of sovereign AI is strategic resilience, not complete digital isolation. Nations must judiciously balance sovereignty goals with the advantages of global technological collaboration. For telecom operators, adeptly managing these complexities and investment demands will define sovereign AI’s realization as a viable growth opportunity.
…………………………………………………………………………………………………………………………………………………………………………….

References:

Telcos Across Five Continents Are Building NVIDIA-Powered Sovereign AI Infrastructure

https://dbrs.morningstar.com/research/468155/telecoms-are-well-placed-to-benefit-from-sovereign-ai-infrastructure-plans

How “sovereign AI” could shape telecom

https://www.rcrwireless.com/20251202/ai/sovereign-ai-telcos

Subsea cable systems: the new high-capacity, high-resilience backbone of the AI-driven global network

Analysis: OpenAI and Deutsche Telekom launch multi-year AI collaboration

AI infrastructure spending boom: a path towards AGI or speculative bubble?

Market research firms Omdia and Dell’Oro: impact of 6G and AI investments on telcos

Omdia: How telcos will evolve in the AI era

OpenAI announces new open weight, open source GPT models which Orange will deploy

Expose: AI is more than a bubble; it’s a data center debt bomb

Can the debt fueling the new wave of AI infrastructure buildouts ever be repaid?

Custom AI Chips: Powering the next wave of Intelligent Computing

AI spending boom accelerates: Big tech to invest an aggregate of $400 billion in 2025; much more in 2026!

IBM and Groq Partner to Accelerate Enterprise AI Inference Capabilities

Dell’Oro: Analysis of the Nokia-NVIDIA-partnership on AI RAN

 

Page 1 of 14
1 2 3 14