Ptechhub
  • News
  • Industries
    • Enterprise IT
    • AI & ML
    • Cybersecurity
    • Finance
    • Telco
  • Brand Hub
    • Lifesight
  • Blogs
No Result
View All Result
  • News
  • Industries
    • Enterprise IT
    • AI & ML
    • Cybersecurity
    • Finance
    • Telco
  • Brand Hub
    • Lifesight
  • Blogs
No Result
View All Result
PtechHub
No Result
View All Result

Qualcomm gears up for AI inference revolution | Computer Weekly

By Computer Weekly by By Computer Weekly
October 28, 2025
Home Uncategorized
Share on FacebookShare on Twitter


Qualcomm’s answer to Nvidia’s dominance in the artificial acceleration market is a pair of new chips for server racks, the A1200 and A1250, based on its existing neural processing unit (NPU) technology.

Significantly, Qualcomm has developed a novel memory architecture for the A1250 based on near-memory computing, which it claims provides “a generational leap in efficiency and performance for AI inference workloads”. It does so, according to Qualcomm, by delivering greater than 10x higher effective memory bandwidth and much lower power consumption.

The A1200 is being positioned by Qualcomm as being purpose-built for running AI inference using a cluster of server racks. The company claimed it has been designed to deliver low total cost of ownership (TCO). Qualcomm said the A1200 has been optimised for large language model (LLM) and multimodal model (MMM) inference and other AI workloads.

To accompany the A1200 and A1250, Qualcomm is providing a software stack, which it said offers “seamless compatibility with leading AI frameworks” and enables enterprises and developers to deploy secure, scalable generative AI across datacentres.

As analyst Forrester points out, these chips appear to be targeting Nvidia and AMD with GPU and rack-scale products. According to Forrester senior analyst Alvin Nguyen, the Qualcomm offerings make sense given that the market for rack-scale AI inference is highly profitable and the current providers of rack-based inference hardware are unable to fully satisfy demand.

“The core of their AI looks to be based on existing NPU designs, so this lowers their barrier to entry. It also seems that they are creating GPUs with larger memory capacity than Nvidia or AMD (768 GB) which could give it an advantage with certain AI workloads,” he added.

In a LinkedIn article posted in March, Jack Gold, strategic adviser and technology analyst at J Gold Associates, predicted that within two to three years, 85% of enterprise AI workloads will be inference-based, rather than the current predominance of training workloads. Training generally requires the high-performance AI-optimised server infrastructure that resides in hyperscale datacentres provided by the likes of AWS, Azure and the Google Cloud Platform.

Like many industry watchers, Gold believes that most AI workloads run by enterprises in hyperscale infrastructure are pilot projects. Evidence from numerous surveys of business and IT leaders shows that these projects often fail to mature into production. But as Gold points out, once an AI model is trained, which requires high performance graphics processor unit (GPU) AI acceleration hardware, it can then be used on more modest hardware.

“Most enterprise AI workloads running today are still experimental and/or small scale. As AI moves to production level inference-based solutions, the need for high-end GPUs is less important and standard server SoCs [systems on a chip] are more appropriate,” Gold said.

This is the market opportunity Qualcomm is hoping to address with the A1200 and A1250 hardware. Durga Malladi, senior vice-president and general manager of technology planning, edge solutions and datacentre of Qualcomm Technologies, said the two products offer a way for organisations to run AI inference AI models more easily.

“With seamless compatibility for leading AI frameworks and one-click model deployment, Qualcomm AI200 and AI250 are designed for frictionless adoption and rapid innovation,” Malladi added.



Source link

By Computer Weekly

By Computer Weekly

Next Post
IoT Market to Reach ,372.46 Billion by 2034 Globally, at 14.1% CAGR: Allied Market Research

IoT Market to Reach $5,372.46 Billion by 2034 Globally, at 14.1% CAGR: Allied Market Research

Recommended.

AEWIN Showcases Tailored Server & Network Appliance Solutions at SC25

AEWIN Showcases Tailored Server & Network Appliance Solutions at SC25

October 31, 2025
IBM To Boost SAP Regulated Industries Skills With Cognitus Buy

IBM To Boost SAP Regulated Industries Skills With Cognitus Buy

October 16, 2025

Trending.

Chai AI Announces Upcoming Rollout of Apple and Google Age Verification APIs to Enhance Platform Safety

Chai AI Announces Upcoming Rollout of Apple and Google Age Verification APIs to Enhance Platform Safety

March 10, 2026
Huawei lanceert Next Generation FAN-oplossing

Huawei lanceert Next Generation FAN-oplossing

March 7, 2026
Baidu Announces Fourth Quarter and Fiscal Year 2025 Results

Baidu Announces Fourth Quarter and Fiscal Year 2025 Results

February 26, 2026
Half of Google’s software development now AI-generated | Computer Weekly

Half of Google’s software development now AI-generated | Computer Weekly

February 5, 2026
Ghost Campaign Uses 7 npm Packages to Steal Crypto Wallets and Credentials

Ghost Campaign Uses 7 npm Packages to Steal Crypto Wallets and Credentials

March 24, 2026

PTechHub

A tech news platform delivering fresh perspectives, critical insights, and in-depth reporting — beyond the buzz. We cover innovation, policy, and digital culture with clarity, independence, and a sharp editorial edge.

Follow Us

Industries

  • AI & ML
  • Cybersecurity
  • Enterprise IT
  • Finance
  • Telco

Navigation

  • About
  • Advertise
  • Privacy & Policy
  • Contact

Subscribe to Our Newsletter

  • About
  • Advertise
  • Privacy & Policy
  • Contact

Copyright © 2025 | Powered By Porpholio

No Result
View All Result
  • News
  • Industries
    • Enterprise IT
    • AI & ML
    • Cybersecurity
    • Finance
    • Telco
  • Brand Hub
    • Lifesight
  • Blogs

Copyright © 2025 | Powered By Porpholio