IDC: Public Cloud software at 2/3 of all enterprise applications revenue in 2026; SaaS is essential!
IDC forecasts that worldwide revenue for enterprise applications will grow from $279.6 billion in 2022 to $385.2 billion in 2026 with a five-year compound annual growth rate (CAGR) of 8.0%. Nearly all this growth will come from investments in public cloud software, which is expected to represent nearly two thirds of all enterprise applications revenue in 2026.
While the process of migrating from on-premises applications to the cloud can take years, enterprise software vendors and their customers will continue the transition to the cloud as this is an essential part of business operations in the digital world. Companies that do not pursue this technology will sustain losses due to profound opportunity costs as their competitors adopt cloud technologies and the use of application programming interfaces (APIs), moving beyond the reach of technological holdouts with on-premises or homemade solutions.
“It’s no longer enough for businesses to sit back and rely on their technological debt of software and hardware assets to keep the company running. In the digital world, enterprise software needs to constantly innovate to keep up with demand for speed, scale, and a resilient business,” said Heather Hershey, research director, Worldwide Digital Commerce at IDC. “Organizations must invest in new tools to keep their application portfolio up to date as they move into the digital era, automating all processes while also leveraging innovation and a wealth of data to become a more creative and resilient company in the digital realm.”
In addition to the ongoing cloud migration, IDC has identified a number of other significant market developments that are driving growth in the enterprise applications market.
- SaaS and cloud-based, modular, and intelligent applications are no longer “nice to have” but are instead essential for business. Organizations that want to stay in business need AI-driven software that is cloud enabled, modular, and intelligent.
- Application programmable interface technology will continue to be the backbone of the enterprise applications market. APIs will always resonate as a sound investment to companies that understand the pivotal role they play in connecting all the disparate code bases that make up the modern world.
- Phasic migration to cloud with TaskApps augmentation will continue, particularly in B2B enterprises. TaskApps and low-code/no-code development tools are being used to close gaps, extend processes, or change up the business at a faster pace throughout the transition to digital first.
- New global regulations around data privacy and ethics have changed the way organizations collect and use data, pushing governance to the forefront of the conversation. Compliance has become a differentiating factor for enterprises that prioritize trustworthiness.
“The digital world is completely altering the way software is utilized and incorporated into the organization from modularity to APIs to low code/no code to business process automation to TaskApps and even with innovation,” said Mickey North Rizza, group vice president, Enterprise Software at IDC. “Organizations are stretching their visions from filling technology gaps to optimizing processes globally to going the last mile with complete differentiators for their clients. The business world is finally starting to leverage the opportunity technology brings to it.”
Photo Credit: Unsplash
The enterprise applications market is a competitive market that includes software specific to certain industries as well as software that can handle requirements for multiple industries. Enterprise applications can be delivered as a pre-integrated suite of applications (featuring common data and process models across functional areas) or as standalone applications that automate specific functional business processes, such as accounting, human capital management, or supply chain execution. The enterprise applications market consists of the following secondary markets: enterprise resource management, customer relationship management, engineering applications, supply chain management applications, and production applications.
The IDC report, Worldwide Enterprise Applications Software Forecast, 2022–2026: Digital Era Software on the Rise (Doc #US48563522), presents a five-year forecast for worldwide enterprise applications revenues, including spending by geographic region and deployment type (public cloud and on premises). The report also provides insight into the market’s evolution through 2026, including deployment models, trends, and significant market developments.
…………………………………………………………………………………………………………………………………………………………
In a separate report titled Worldwide Quarterly Enterprise Infrastructure Tracker: Buyer and Cloud Deployment, IDC sas that spending on compute and storage infrastructure products for cloud deployments, including dedicated and shared IT environments, increased 24.7% year over year in the third quarter of 2022 (3Q22) to $23.9 billion. Spending on cloud infrastructure continues to outgrow the non-cloud segment although the latter had strong growth in 3Q22 as well, increasing at 16.5% year over year to $16.8 billion. The market continues to benefit from high demand and large backlogs, coupled with an improving infrastructure supply chain.
Spending on shared cloud infrastructure reached $16.8 billion in the quarter, increasing 24.4% compared to a year ago. IDC expects to see continuous strong demand for shared cloud infrastructure with spending expected to surpass non-cloud infrastructure spending in 2023. The dedicated cloud infrastructure segment grew 25.3% year over year in 3Q22 to $7.1 billion. Of the total dedicated cloud infrastructure, 45.2% was deployed on customer premises.
For the full year 2022, IDC is forecasting cloud infrastructure spending to grow 19.6% year over year to $88.1 billion – a noticeable increase from 8.6% annual growth in 2021. Non-cloud infrastructure is expected to grow 10.7% to $64.7 billion. Shared cloud infrastructure is expected to grow 19.0% year over year to $60.9 billion for the full year while spending on dedicated cloud infrastructure is expected to grow 21.2% to $27.3 billion for the full year.
About IDC:
International Data Corporation (IDC) is the premier global provider of market intelligence, advisory services, and events for the information technology, telecommunications, and consumer technology markets. With more than 1,300 analysts worldwide, IDC offers global, regional, and local expertise on technology, IT benchmarking and sourcing, and industry opportunities and trends in over 110 countries. IDC’s analysis and insight helps IT professionals, business executives, and the investment community to make fact-based technology decisions and to achieve their key business objectives. Founded in 1964, IDC is a wholly owned subsidiary of International Data Group (IDG), the world’s leading tech media, data, and marketing services company. To learn more about IDC, please visit www.idc.com. Follow IDC on Twitter at @IDC and LinkedIn. Subscribe to the IDC Blog for industry news and insights.
References:
2 thoughts on “IDC: Public Cloud software at 2/3 of all enterprise applications revenue in 2026; SaaS is essential!”
Comments are closed.
FT- 17 January 2023: Microsoft’s $10bn bet on ChatGPT developer marks new era of AI
Tech giants race to stake out their place in new field of generative artificial intelligence
by Richard Waters and Tabby Kinder in San Francisco
The $10bn investment that Microsoft is considering in San Francisco-based research firm OpenAI looks set to become the defining deal for a new era of artificial intelligence.
If the US software giant is right about the far-reaching implications of the technology, it could also trigger a realignment in the AI world as other tech groups race to stake out their place in the new field of generative AI.
OpenAI grabbed global headlines last month with the launch of ChatGPT, an AI system that can answer queries and produce text in natural-sounding language.
But Microsoft executives believe the technology behind the service will soon have a deeper impact throughout the tech world.
“These [AI] models are going to change the way that people interact with computers,” said Eric Boyd, head of AI platforms at Microsoft. Talking to a computer as naturally as a person will revolutionise the everyday experience of using technology, he added.
“They understand your intent in a way that hasn’t been possible before and can translate that to computer actions,” Boyd said in an interview with the Financial Times before news of the possible deal.
Microsoft’s potential investment, first reported by the newsletter Semafor last week and confirmed by two people familiar with the situation, would see it take a significant minority stake that would value OpenAI, after the investment, at $29bn. Microsoft declined to comment.
The potential landmark investment comes as venture capitalists are rushing to back the latest AI trend at a time when previous investment fads like blockchain and cryptocurrencies have faded.
Microsoft made its first $1bn investment in OpenAI in 2019, sealing a role as the tech platform for the firm’s highly demanding AI models and giving it first rights to commercialise its technology.
The software giant has already used OpenAI’s technology in a number of its own products, though its executives say this only scratches the surface of what will come next.
Its cloud customers have been able to pay for access to GPT-3, a text-generating AI model, since 2021. Dall-E 2 — part of a wave of image generating systems that took the AI world by storm last year — is the foundation of a recent Microsoft graphic design product called Designer, and has also been made available through the Bing search engine.
Meanwhile Codex, a system which prompts software developers with suggestions of which lines of code to write next, has been turned into a product by GitHub, a Microsoft service for developers.
The speed at which AI tools like this are passing from advanced research to everyday product may be unparalleled in tech history, according to AI experts. Codex was introduced in an OpenAI research paper only in the middle of 2021, but within a year Microsoft had turned it into a commercial subscription service.
According to GitHub chief executive Thomas Dohmke, 40 per cent of the code created by developers who use the service, called Copilot, is automatically generated by the AI system, halving the time it takes to create new code — a huge leap in efficiency after a decade of largely ineffective efforts to boost developer productivity.
“It’s a mind-blowing productivity statistic,” said Dohmke.
Much of OpenAI’s tech stems from the creation of so-called large language models, which are trained on huge amounts of text. Unlike earlier forms of machine learning — which has dominated AI for the last decade — the technique has led to systems that can be used in a wider variety of circumstances, boosting their commercial value.
“The real power of these models is they have the capability to do so many different tasks at the same time,” said Boyd at Microsoft. He added that makes it possible to do so-called “zero-shot” learning — using the AI for new tasks without needing to train them.
Google and other tech giants, as well as a number of start-ups, have also ploughed resources into creating giant AI models like this. But since GPT3 stunned the AI world in 2020 with its ability to produce large blocks of text on demand, OpenAI has set the pace with a succession of eye-catching public demonstrations.
Microsoft executives are looking to use the technology in a wide range of products. Speaking at a company event late last year, chief executive Satya Nadella predicted that generative AI would lead to “a world where everyone, no matter their profession” would be able to get support from the technology “for everything they do”.
Generative AI is set to become a central part of “productivity” applications like Microsoft’s Office, said Oren Etzioni, an adviser and board member at A12, the AI research institute set up by Microsoft co-founder Paul Allen.
All workers will eventually use productivity software that presents relevant information to them, checks their work and offers to generate content automatically, he said.
The potential upheaval this could cause in the software world has not been lost on Microsoft’s rivals, who see the technology as a rare opportunity to break into markets dominated by Big Tech.
Emad Mostaque, head of London-based Stability AI, which made a splash last year with the launch of its open-source image generating system, claimed his organisation was building a “PowerPoint killer” — an AI tool that is designed to make it much easier to create presentations than the widely-used Microsoft application.
That makes the move both defensive and offensive for Microsoft, as it tries to protect established products like Office while also mounting a stronger challenge in markets like internet search.
With its potential OpenAI investment, meanwhile, Microsoft is also trying to use its technology and financial muscle to position itself as the main platform on which the next era of AI will be built.
“The amount of cloud computing power [OpenAI] needs is beyond the capability of a start-up” or venture capital investor to support, one of the company’s investors said. That meant OpenAI had little choice but to seek financial backing from one of the handful of tech giants, this person added.
Microsoft has sought to use its first investment in OpenAI to get a head start, building what it describes as a supercomputer to train the research firm’s giant AI models. The same technology platform is also now used by Facebook parent Meta for its AI work.
Nadella claimed recently that the head start it got from working with OpenAI meant that calculations carried out by its AI supercomputer cost only around half as much as its biggest rivals. Any cost advantage could be key: analysts at Morgan Stanley estimate that the higher cost of natural language processing means that answering a query using ChatGPT costs around seven times as much as a typical internet search.
Microsoft’s biggest cloud computing rivals have also been seeking to align themselves with some of the most promising generative AI companies, though none of the other start-ups in the field has produced AI models with the scale or range of OpenAI.
Amazon’s cloud division has a three-year deal to act as the computing platform for Stability AI. Another AI start-up, CohereAI, which was founded by three researchers from Google, reached a deal in 2021 to use the search company’s computing platform to train their own AI.
If a handful of tech giants become the central platforms for — and investors in — the start-ups building the next generation of AI technology, it could stir concern among regulators.
One person familiar with Microsoft’s investment plans conceded that its alliance with OpenAI was likely to come under close scrutiny, but added that the minority investment should not provoke any regulatory intervention.
But as owners of the cloud computing platforms needed to support the coming age of generative AI, it seems inevitable that Big Tech will have a significant say in what comes next.
Public cloud providers offer telecom operators a radical business transformation opportunity but many remain wary of relinquishing their assets and forsaking control. Today, public cloud providers have accelerated time-to-market for many enterprises. With 5G core networks considered inherently cloud-native, many 5G standalone networks are being deployed with a cloud-based architecture. Meanwhile, concerns remain over the suitability of public cloud for operator network infrastructure and workloads.
This article sets out to explore what communication service providers (CSPs) and other industry experts consider some of the biggest challenges and drivers for the adoption of public cloud.
CSP Drivers:
With telcos’ growing need to expand revenues beyond connectivity, research from analyst firm STL Partners suggests there is an opportunity for operators to capitalise on the cloud and to partner with public cloud providers. As such, the analysts suggest the migration of IT environments and telco workloads to the public cloud are inevitable for operators.
Further, a recent study by Capgemini, an IT services and consulting firm, shows that on average CSPs are expected to invest around one billion US dollars into their network cloud transformation over the next 3-5 years. Per operator, this sum reportedly equals to 38-44% of their total network transformation budgets.
With such level of investment needed to develop a cloud platform, applications, and services, combined with the opportunity to capitalize from public cloud, some in the industry see hyperscalers and other large public cloud vendors as ideal partners for CSPs. The existing investments into building and operating large cloud infrastructures and offerings can significantly reduce their total cost of ownership (TCO) for CSPs, making TCO the primary driver for telcos to embrace public cloud.
Operational efficiencies gained from running IT systems on the public cloud are another area often discussed in relation with drivers for moving workloads. In terms of scale, public cloud is believed to enable CSPs to expand their addressable market, improve their B2B offerings, and support B2B2x solutions.
Public cloud also promises to offer a pay-as-you-go model based on microservices and Kubernetes (also known as k8s) applications. Such microservices enable telcos with rapid deployment of their go-to-market strategies and facilitate real-time data streaming. Though it is worth noting here that k8s can also be deployed in an on-premise environment and as such are not unique to public cloud.
Public cloud also offers deployment benefits such as flexible and agile customer and revenue management and scalable offerings to increase monetisation of services. Therefore, hosting business support systems (BSS) and the operating support systems (OSS) in the public cloud are seen as particularly valuable use cases. However, there are also security considerations when hosting transactional data and systems (i.e. BSS workloads) on the public cloud. This is due to its multitenancy nature (more to this further below).
In terms of leveraging additional public cloud capabilities, research from Omdia suggests that there are additional advantages for CSPs. These include capabilities offered by AI and ML to help automate processes and monetise insights gained through data. Here Google’s Cloud Platform is cited as the frontrunner. Automation is valuable to CSP tasks such as testing and repetitive tasks, though the research is also quick to flag that maintaining control is equally important for operators.
Some may argue public cloud is an attractive proposition especially for new market entrants. In the US, greenfield operator Dish presents such an example. In early 2021 it announced a collaboration with Amazon Web Services (AWS) for constructing its full 5G network. The operator connected all its 5G hardware and network management systems through the cloud provider.
There are also public cloud partnership examples between hyperscalers and brownfield operators. These include AT&T and Microsoft Azure in the US, Bell Canada and Google, and more recently Tele2 and AWS in Norway. But despite some uptake in such partnerships, many operators remain somewhat sceptical and unclear about the best approaches towards public cloud. Here are some of the key concerns we have gathered.
CSP concerns over public cloud:
The telecom industry is highly regulated and often considered as part of critical infrastructure. As such, it is vital that CSPs maintain control, know where their networks and customer data reside, and can ensure those locations are safe.
In a recent survey by Telecoms.com more than four in five industry respondents feared security concerns over running telco applications in the public cloud, including 37% who find it hard to make the business case for public cloud as private cloud remains vital in addressing security issues. This also means that any efficiency gains are offset by the IT environment and the network running over two cloud types.
Meanwhile, many in the industry also fear vendor lock-in and lack of orchestration from public cloud providers. Around a third of industry experts from the same survey find it a compelling reason not to embrace and move workloads to the public cloud unless applications can run on all versions of public cloud and are portable among cloud vendors.
This brings us to the issue of interoperability and interconnectedness. Proponents of public cloud may argue that through a multi-cloud approach operators can use offerings from various vendors and thus introduce a diverse set of public cloud providers in their networks. However, as it stands, the services of different public cloud vendors are indeed not interconnected nor interoperable for the same types of workloads. This concern is one of the drivers to avoid public cloud, according to some operators.
Alongside vendor lock-in and interoperability are also considerations over TCO savings versus long-term costs. In that, CSPs are unclear over the long-term financial benefits and worry over potential risks of increased costs once their network is locked-in with a public cloud provider.
When it comes to network or service availability, there is still a lack of reliability on hyperscalers for five nines (a telecom network uptime goal of 99.999% of the time), according to some voices in the industry. As critical infrastructure, service availability (which contractually is typically defined under a service level agreement) is a hugely significant characteristic for network operators. Network availability can also have a significant impact on operators both commercially as well as from a customer experience perspective.
As an increasing number of mission critical application and services are added on telecom networks, the ability to support five nine KPIs and delivering on service level agreements becomes even more critical. Collaborations among vendors and public cloud providers can help optimise platforms and support network availability.
These concerns are also evident by the recent operator-driven initiative Sylva. It’s been argued that the project highlights the operator community’s keenness on taking telco cloud into own hands. The initiative also highlights, alongside security, data sovereignty as another key area of concern for CSPs. This further emphasizes the need for alternative solutions to public cloud that are built with open, interoperable, and interconnected solutions in mind has moved up CSPs’ priority list.
Considering these challenges, it is not surprising to see that public cloud still makes up a small portion of total telecoms networks and workloads. The below graph from a research paper by Capgemini illustrates which type of cloud architecture is currently being used or is planned in the telecoms core network, the Radio Access Network (RAN), OSS and orchestration, and for the use of multi-access edge computing.
Source: Capgemini
Hybrid cloud could ease the transition
With 5G core leading the migration of network workloads to the cloud, it is more important than ever to find the optimum road to cloud adoption. Many in the industry are considering a balanced approach in their journey towards becoming cloud-native. In that, hybrid cloud enables CSPs to identify the best use cases for each cloud strategy and thus to transition telco applications and workloads more gradually and systematically. It also allows CSPs to consume public cloud applications and benefits in a private manner.
For instance, in a bid to modernise its BSS to match the capabilities of its e-commerce and application offerings, European CSP Vodafone entered into a partnership deal with cloud solutions provider Oracle last year. Here, due to the risks associated with transactional data hosted on the public cloud, the operator opted for a hybrid model. As such, Oracle has built the operator a complete set of public cloud offering on the CSP’s own data centres, i.e. keeping the workloads on-premise.
The operator also has several hybrid cloud partnership agreements with the other US-based hyperscalers for its operating support systems, data processing and analytics, and edge computing services. Other European Opcos taking such hybrid approaches include Deutsche Telekom (DT) and Telefonica.
Conclusions
The use of hybrid cloud and the Vodafone example paint a picture of a cloud journey that cherry picks the best of both worlds. However, there are still few considerations to keep in mind. Hybrid cloud introduces some difficulties and complexities around operating multiple interfaces. This approach may eventually require internal resources to acquire new skills to operate several new cloud technologies. Additionally, hybrid cloud built by hyperscalers will intrinsically consist of some level of proprietary technology. Some in the industry view this as a de facto lock-in.
Since security and service availability are also key concerns preventing many CSPs from the move to public cloud – especially when it comes to core networks – adopting industry-standard approaches and security strategies become key considerations. These can include incorporating the most modern security approaches (e.g. DevSecOps) early in the adoption process.
While it may be more difficult for some operators (potentially smaller CSPs) to disregard the initial cost savings offered by public cloud, many incumbent operators are still reluctant to relinquish their network workloads fully. Among those most vocal about it is US-based MNO Verizon, which suggested it would never put its core networks in public cloud, and European multi-nationals such as DT which also has no plans to move its network functions into the public cloud domain.
This is despite both CSPs having adopted various telco cloud approaches, including hybrid cloud solutions, in telco IT, edge computing and data processing. For wider context it is also worth noting that DT is among the European operator groups spearheading project Sylva, alongside its peers Orange, Telefonica, TIM and Vodafone together with Nordic kit vendors Nokia and Ericsson.
To sum up, despite the TCO savings and operational efficiencies, the majority of CSPs have not initiated a public cloud deployment strategy. By all accounts, when it comes to telco procurement, experienced CSPs may well aim to avoid risking all their resources in one single venture, relinquishing control, and getting cut off from their own assets. Meanwhile, most CSPs introducing public cloud to their networks, it appears, are taking a considered and steady transformational approach. But regardless of public, hybrid, or private, the future of telco networks is eventually headed to the clouds.
https://telecoms.com/520055/telcos-and-the-public-cloud-drivers-and-challenges/