Want to read Slashdot from your mobile device? Point it at m.slashdot.org and keep reading!

 



Forgot your password?
typodupeerror
AI Cloud Businesses The Almighty Buck

Enterprise AI Adoption Stalls As Inferencing Costs Confound Cloud Customers 16

According to market analyst firm Canalys, enterprise adoption of AI is slowing due to unpredictable and often high costs associated with model inferencing in the cloud. Despite strong growth in cloud infrastructure spending, businesses are increasingly scrutinizing cost-efficiency, with some opting for alternatives to public cloud providers as they grapple with volatile usage-based pricing models. The Register reports: [Canalys] published stats that show businesses spent $90.9 billion globally on infrastructure and platform-as-a-service with the likes of Microsoft, AWS and Google in calendar Q1, up 21 percent year-on-year, as the march of cloud adoption continues. Canalys says that growth came from enterprise users migrating more workloads to the cloud and exploring the use of generative AI, which relies heavily on cloud infrastructure.

Yet even as organizations move beyond development and trials to deployment of AI models, a lack of clarity over the ongoing recurring costs of inferencing services is becoming a concern. "Unlike training, which is a one-time investment, inference represents a recurring operational cost, making it a critical constraint on the path to AI commercialization," said Canalys senior director Rachel Brindley. "As AI transitions from research to large-scale deployment, enterprises are increasingly focused on the cost-efficiency of inference, comparing models, cloud platforms, and hardware architectures such as GPUs versus custom accelerators," she added.

Canalys researcher Yi Zhang said many AI services follow usage-based pricing models that charge on a per token or API call basis. This makes cost forecasting hard as the use of the services scale up. "When inference costs are volatile or excessively high, enterprises are forced to restrict usage, reduce model complexity, or limit deployment to high-value scenarios," Zhang said. "As a result, the broader potential of AI remains underutilized." [...] According to Canalys, cloud providers are aiming to improve inferencing efficiency via a modernized infrastructure built for AI, and reduce the cost of AI services.
The report notes that AWS, Azure, and Google Cloud "continue to dominate the IaaS and PaaS market, accounting for 65 percent of customer spending worldwide."

"However, Microsoft and Google are slowly gaining ground on AWS, as its growth rate has slowed to 'only' 17 percent, down from 19 percent in the final quarter of 2024, while the two rivals have maintained growth rates of more than 30 percent."

Enterprise AI Adoption Stalls As Inferencing Costs Confound Cloud Customers

Comments Filter:
  • Nice click bait headline. 20% growth is hardly stalling.
    • by gweihir ( 88907 ) on Friday June 13, 2025 @10:28PM (#65448571)

      For a get-rich-quick scheme based on lies? Yes, that is stalling.

      • When your head is firmly buried in the sand, it may look like that.

        The volatile pricing of inference is correctly seen as hard to plan for, which simply means that inference providers will start offering more rate limited flat-fee options and other pricing models that are attractive to customers who need the predictability. This is nothing special (cloud computing has had similar issues and solutions for years) and definitely not evidence for "OMG AI IS HYPE YOU IDIOTS" with which you so consistently pollut

  • Run away
    AI needs local compute

    • by Tony Isaac ( 1301187 ) on Friday June 13, 2025 @11:22PM (#65448635) Homepage

      It depends.

      It's kind of like people who insist that renting an apartment is worse financially than owning a home. It depends.

      If you move every year or two, renting is a much better deal than buying. If you move to a new city and don't yet know what part of town you want to live in, renting is better. If you don't have the ability or desire to maintain your own place / grounds / HVAC, renting is better.

      On the other hand, if you are stable, stay put, are raising a family, and are able and willing to maintain your own place, owning is definitely better.

      Computer hardware has many of the same tradeoffs between owning (on prem) and renting (cloud).

    • The problem is local compute is absurdly front loaded cost wise. The big Nvidia cards that can run the larger models (And no a 5080 isn't even close) the costs can run in the tens of thousands for just one. If your needs are modest a Mac Studio MIGHT cover inference (256gb unified memory goes a long way, and most of the key libraries now support metal) but for serious work you gotta go Nvidia, and that aint cheap.

      • Had a long conversation [perplexity.ai] about Cerebras or Graphcore and the power saving plus software developments in power savings make for some impressive numbers.

        Bottom Line
        The combined effect of next-gen hardware (Cerebras, Graphcore, etc.) and modern software optimizations can realistically reduce AI’s energy use per task by 70–98% compared to just a few years ago.

        This is transformative for both cost and sustainability—especially as AI adoption continues to surge.
        The biggest wins will come from deploying both hardware and software advances together.

  • AI isn't cheap (Score:4, Interesting)

    by Tony Isaac ( 1301187 ) on Friday June 13, 2025 @11:28PM (#65448643) Homepage

    For all the hype about AI being cheaper than human labor, the truth isn't so clear. Human call center workers, for example, make maybe $10 per hour, and in that hour, they can take a dozen or more calls, depending on the complexity. AI equivalents, by contrast, often license their tools per use, per call, per command, etc. It's not at all clear that the cost/benefit ratio will swing in favor of AI in the near future.

    Those "free" AI tools you get from the big providers? Those are loss leaders, getting people to think they are cheap and easy to use. As soon as you want to do real work with AI, the prices escalate.

    • Re:AI isn't cheap (Score:4, Interesting)

      by DamnOregonian ( 963763 ) on Saturday June 14, 2025 @04:00AM (#65448811)
      The big providers charge dollars per millions of tokens.

      The real cost is in the middle men.
      A crazy market has popped up selling the tools that turn inference providers into stuff that's useful- and they charge 1000% markups to businesses.
      I have an... acquaintance that is in that like of "work".
      Company has need A, he solves need A with inference provider B, paying the cost-per-Mtok, while company pays him a flat fee of $5,000-$10,000 a month (for $hundreds of dollars of inference)

      It's pretty fucking insane.
      • by Njovich ( 553857 )

        The big providers charge dollars per millions of tokens.

        You aren't wrong about people charging large markups, but the big providers are up there as well. The price for chatgpt audio tokens is around $80 per million tokens, which adds up to around $10 per hour for a call. That is just purely for the chatgpt audio model, but you need a lot more stuff to actually make a useful product. Then in many cases the AI can't actually solve the issues and you still need a human (more cost), and then the customer is frustrated (potential loss of business is more cost).

      • While it may be true that the big providers charge a few dollars per million tokens, a raw "inference" isn't generally a useful product for specific use cases. Businesses need There is a significant amount of refinement required to make a useful AI product, from "inferences." That's why SalesForce, for example, can change $500 for "packs of 100,000". https://coastalcloud.us/resour... [coastalcloud.us] And even at that rate, you can't just sign up for the SalesForce AI, and tell it to start selling your product. SalesForce i

  • by dsgrntlxmply ( 610492 ) on Saturday June 14, 2025 @02:08AM (#65448771)
    This is not inference. It is at best autoconjecture. Polly wanna syllogism.
    • You do raise an interesting point in the nomenclature.

      The result of a statistical model is generally called inference... And realistically, the result of an LLM is no difference.
      But problematically, we assign more meaning to the produced tokens than the model promises.
  • MarcusLion.com has a solution for that

"If you own a machine, you are in turn owned by it, and spend your time serving it..." -- Marion Zimmer Bradley, _The Forbidden Tower_

Working...