Electronics Era

  • About Us
  • Advertise with Us
  • Contact Us
  • e-Mag
  • Webinars
Header logo on website
Advertisement
Advertisement
Menu
  • News
    • Industry News
    • Product News
  • TECH ROOM
    • Sensor
    • VR / AR
    • Embedded
    • Medical Electronics
    • Industry 4.0
    • Robotic
    • Automation
    • Smart Machine
    • Component
    • Manufacturing
    • Aerospace & Defence
    • Security
    • Policy
  • Semiconductor
    • AUTOMOTIVE ELECTRONICS
      • EVs
      • HEVs
      • ADAS
      • Connected Cars
    • IoT-Internet of Things
      • Development Kit
      • IoT Design
    • Power Electronics
      • AC-DC/DC-DC Converters
      • Mosfets
      • IGBTs
      • LEDs
  • T & M
    • 5G testing
    • Oscilloscopes
    • SDN & NFV
    • RF & Wireless
  • AI/ML
  • Telecom
    • 5G/6G
  • RENEWABLES
    • Sustainability
  • Future Tech
    • Data Center
    • Cloud Computing
    • Big Data Analytics
  • Webinars
  • Editor’s Pick
    • Tech Article
    • Tech Blog
    • White Papers
    • EE-Tech Talk
    • Market Research
  • EE Awards
    • EE Awards 2025
    • EE Awards 2024
  • MORE
    • E-Mag
    • Events
    • Subscription
    • Contact Us
Home AI/ML

Cerebras Selects Qualcomm to Deliver Unprecedented Performance in Artificial Intelligence Inference

Best-in-class solution developed with Qualcomm® AI Cloud 100 Ultra offers up to 10x number of tokens per dollar, radically lowering operating costs of AI deployment

Editorial by Editorial
March 15, 2024
in AI/ML, Semiconductor
Reading Time: 3 mins read
Cerebras QUALCOMM
Share on FacebookShare on TwitterShare on LinkedIn

Bangalore, India. – Cerebras Systems has announced the company’s plans to deliver groundbreaking performance and value for production artificial intelligence (AI). By using Cerebras’ industry-leading CS-3 AI accelerators for training with the AI 100 Ultra, a product of Qualcomm Technologies, Inc., for inference, production grade deployments can realize up to a 10x price-performance improvement.

“These joint efforts are aimed at ushering in a new era of high-performance low-cost inference and the timing couldn’t be better. Our customers are focused on training the highest quality state-of-the-art models that won’t break the bank at time of inference,” said Andrew Feldman, CEO and co-founder of Cerebras. “Utilizing the AI 100 Ultra from Qualcomm Technologies, we can radically reduce the cost of inference – without sacrificing model quality — leading to the most efficient deployments available today.”

Leveraging the latest cutting-edge ML techniques and world-class AI expertise, Cerebras will work with Qualcomm Technologies’ AI 100 Ultra to speed up AI inference. Some of the advanced techniques to be used are as follows:

  • Unstructured Sparsity: Cerebras and Qualcomm Technologies solutions can perform training and inference using unstructured, dynamic sparsity – a hardware accelerated AI technique that dramatically improves performance efficiency. For example, a Llama 13B model trained on Cerebras hardware with 85% sparsity trains 3-4x faster and using AI 100 Ultra inference generates tokens with a 2-3x higher throughput.
  • Speculative Decoding: this advanced AI technique marries the high throughput of a small LLM with the accuracy of a large LLM. The Cerebras Software Platform can automatically train and generate both models, which are seamlessly ingested via the Qualcomm® AI Stack, a product of Qualcomm Technologies. The resulting model can output tokens at up to 2x the throughput with uncompromised accuracy.
  •  
  • Efficient MX6 inference: The AI 100 Ultra supports MX6, an industry standard micro-exponent format that performs high accuracy inference using half the memory footprint and twice the throughput of FP16.
  • NAS service from Cerebras: Using Network Architecture Search for targeted use cases the Cerebras platform can deliver models that are optimized for the Qualcomm AI architecture leading to up to 2x higher inference performance.

A combination of these and other advanced techniques are designed to allow the Cerebras and Qualcomm Technologies solutions to deliver an order of magnitude performance improvement while enabling it at model release, resulting in inference-ready models that can be deployed on Qualcomm cloud instances anywhere.

“The combination of Cerebras’ AI training solution with the AI 100 Ultra helps deliver industry leading perf/TCO$ for AI Inference, as well as optimized and deployment-ready AI models to customers helping reduce time to deployment and time to RoI,” said Rashid Attar, Vice President, Cloud Computing, Qualcomm Technologies, Inc.

By training on Cerebras, customers can now unlock massive performance and cost advantages with inference-aware training. Models trained on Cerebras are optimized to run inference on the AI 100 Ultra leading to friction-free deployments. 

“AI has become a key part of pharmaceutical research and development, and the cost of operating models is a critical consideration in the research budget,” said Kim Branson, Sr. Vice President and Global Head of AI/ML at GlaxoSmithKline. “Techniques like sparsity and speculative decoding that make inference faster while lowering operating costs are critical: this allows everyone to integrate and experiment with AI.”

Tags: AIAI 100 UltraCerebras SystemsQualcomm
Editorial

Editorial

Join Our Newsletter

* indicates required
Electronics Era

Electronics Era, India's no.1 growing B2B news forum on Electronics and Cutting Edge Technology is exploring the editorial opportunity for organizations working in the Electronics Manufacturing Services(EMS) Industry.

Follow Us

Browse by Category

  • 5G testing
  • 5G/6G
  • AC-DC/DC-DC Converters
  • ADAS
  • Aerospace & Defence
  • AI/ML
  • Automation
  • AUTOMOTIVE ELECTRONICS
  • Big Data Analytics
  • Blockchain
  • Cloud Computing
  • Component
  • Connected Cars
  • Data Center
  • Editor's Desk
  • EE-Tech Talk
  • Electronics Components
  • Embedded
  • EVs
  • Future Tech
  • HEVs
  • Industry 4.0
  • Industry News
  • IoT-Internet of Things
  • LED & Lighting
  • LEDs
  • Manufacturing
  • Market Research
  • Medical Electronics
  • Mosfets
  • News
  • Oscilloscopes
  • Policy
  • Power Electronics
  • Product News
  • RENEWABLES
  • RF & Wireless
  • Robotic
  • SDN & NFV
  • Security
  • Semiconductor
  • Sensor
  • Smart Machine
  • SMT/PCB/EMS
  • Sustainability
  • T & M
  • Tech Article
  • Tech Blog
  • TECH ROOM
  • Telecom
  • Uncategorized
  • VR / AR
  • White Papers

Recent News

Telecom

World Telecom Day Being Commemorated on 17th May with this year’s theme ‘Gender Equality in Digital Transformation’.

May 17, 2025
Advances On Manufacturing And Next Generation Devices

Advances On Manufacturing And Next Generation Devices

May 16, 2025
  • About Us
  • Advertise with Us
  • Contact Us

© 2022-23 TechZone Print Media | All Rights Reserved

No Result
View All Result
  • News
    • Industry News
    • Product News
  • TECH ROOM
    • Sensor
    • VR / AR
    • Embedded
    • Medical Electronics
    • Industry 4.0
    • Robotic
    • Automation
    • Smart Machine
    • Component
    • Manufacturing
    • Aerospace & Defence
    • Security
    • Policy
  • Semiconductor
    • AUTOMOTIVE ELECTRONICS
      • EVs
      • HEVs
      • ADAS
      • Connected Cars
    • IoT-Internet of Things
      • Development Kit
      • IoT Design
    • Power Electronics
      • AC-DC/DC-DC Converters
      • Mosfets
      • IGBTs
      • LEDs
  • T & M
    • 5G testing
    • Oscilloscopes
    • SDN & NFV
    • RF & Wireless
  • AI/ML
  • Telecom
    • 5G/6G
  • RENEWABLES
    • Sustainability
  • Future Tech
    • Data Center
    • Cloud Computing
    • Big Data Analytics
  • Webinars
  • Editor’s Pick
    • Tech Article
    • Tech Blog
    • White Papers
    • EE-Tech Talk
    • Market Research
  • EE Awards
    • EE Awards 2025
    • EE Awards 2024
  • MORE
    • E-Mag
    • Events
    • Subscription
    • Contact Us

© 2022-23 TechZone Print Media | All Rights Reserved

Advertisement
Advertisement