Google has finally revealed how much electricity and water consumes its AI. Estimates could not be more wrong

We knew that generative artificial intelligence was a monster that was forcing companies to make large investments in energybut Google’s first detailed analysis has put the figures for the first time on the table.

We go to the point. According to him Google Technical Reportbased on data from May 2025, an average text consultation to Gemini consumes 0.24 Electricity watts. To put it in context, it is something like watching nine seconds of television with a conventional TV of 100 W.

Water consumption, which is still necessary to refrigerate serversis 0.26 milliliters per consultation; The equivalent of five drops of water. The carbon footprint of the entire inference process, according to the report, is 0.03 grams of equivalent.

Wrong estimates. Just a year ago, third party analysis They estimated that a single consultation of AI in the Google search engine, such as those of AI overViews, could consume about 3 Wh, ten times more than a traditional search. This led to calculations as striking as the deployment of AI in the search engine would consume enough energy to load seven electric cars per second.

With Google’s official data in hand, we see that this estimate was wrong by a 12.5 factor. The new software techniques (such as speculative decoding) and the most efficient models architectures (such as the Mixture-OF-Experts paradigm) have completely changed the panorama.

Inference, no training. These figures, the most concrete published to date by the company, only take into account Gemini’s consumption by inferring user response. The expensive process of training the great language models that feed these tools remains a mystery, but Google is justified by saying that the massive adoption of generative AI, integrated even in its search engine, has put the focus on inference.

In this direct relationship with the user it is also where greater efficiency jumps are getting large technological companies. Google says that, in the last 12 months, energy consumption has divided by 33 and by 44 the carbon footprint of each consultation to Gemini. Much of this jump has to do, not only with more efficient models, but with the improvement of AI accelerators (Tpus and Gpus), a hardware that Google develops internally.

Graph of the consumption of the different models of AI
Graph of the consumption of the different models of AI

The amount of “prompts” per kWh that process the different models of AI

Less than Netflix. Google is not alone in this new era of transparency. Sam Altman, CEO of Openai, also shed some light on the consumption of chatgpt. In one June 2025 publicationAltman said that an average consultation to ChatgPT consumes approximately 0.34 or energy Wh and about 0.3 ml of water.

The energy figure is slightly higher than Gemini’s, although it is a difficult comparison. Altman did not give details of his methodology, so we do not know if his calculation includes all the factors that Google has considered (such as electrical consumption in refrigeration and in “idle” machines; that is, inactive, but ready for rapid consumption peaks).

Both companies have been compared to television: “An hour of Netflix consumes 100 times more electricity than Chatgpt,” says an official OpenAI slide. The same that says that the total impact of AI on US carbon emissions would be around 0.5%.

Images | Google

In Xataka | The consumption of AI is overestimated and we must worry more about the air conditioning, according to the IAE

Leave your vote

Leave a Comment

GIPHY App Key not set. Please check settings

Log In

Forgot password?

Forgot password?

Enter your account data and we will send you a link to reset your password.

Your password reset link appears to be invalid or expired.

Log in

Privacy Policy

Add to Collection

No Collections

Here you'll find all collections you've created before.