• Menu
  • Skip to right header navigation
  • Skip to main content
  • Skip to primary sidebar

DigiBanker

Bringing you cutting-edge new technologies and disruptive financial innovations.

  • Home
  • Pricing
  • Features
    • Overview Of Features
    • Search
    • Favorites
  • Share!
  • Log In
  • Home
  • Pricing
  • Features
    • Overview Of Features
    • Search
    • Favorites
  • Share!
  • Log In

Study finds running gen AI models on the phones instead of in the cloud consumed anywhere from 75% to 95% less power, with associated sharp decreases in water consumption and overall carbon footprint

July 1, 2025 //  by Finnovate

One of the easiest ways to minimize AI’s environmental impact may be to move where the processing is done, per new academic research conducted in partnership with Qualcomm. Running AI on devices instead of in the cloud slashes power consumption of queries by about 90%, the study finds. The industry has long touted the benefits of running models locally on devices instead of in the cloud — not just in energy terms, but also potentially making them cheaper and more private.  Researchers at the University of California, Riverside ran a series of experiments comparing the performance of various generative AI models, both in the cloud and on phones powered with Qualcomm chips. Running any of six different models on the phones consumed anywhere from 75% to 95% less power, with associated sharp decreases in water consumption and overall carbon footprint. Qualcomm is also developing an AI simulator and calculator that illustrates, for any given query and user location, what the responses would look like on-device versus the cloud, and how much less power and water they would use. One example — running a coding skills question on the Llama-2-7B model in California — was 94% more power efficient and 96% more water efficient on-device. For all six models in the study, the inference time on the phones, measured in seconds, was higher than in the cloud. Narrowing or eliminating that gap, particularly on the most powerful and popular models, will be crucial to accelerating on-device adoption. For many AI users, the data center in your pocket might be all you need.

Read Article

Category: AI & Machine Economy, Innovation Topics

Previous Post: « Volcano Exchange’s financial RWA digital asset leverages blockchain tech to transform high-threshold private banking services into divisible and tradable digital assets lowering the entry barrier for retail investors
Next Post: OpenAI’s API platform allows developers to express intent, not just configure model flows through built-in capabilities for knowledge retrieval, web search, and function calling for supporting real-world agent workflows »

Copyright © 2025 Finnovate Research · All Rights Reserved · Privacy Policy
Finnovate Research · Knyvett House · Watermans Business Park · The Causeway Staines · TW18 3BA · United Kingdom · About · Contact Us · Tel: +44-20-3070-0188

We use cookies to provide the best website experience for you. If you continue to use this site we will assume that you are happy with it.OkayPrivacy policy