Published Date : 19/06/2025
Whether it be writing an email or planning a vacation, about a quarter of Americans say they interact with artificial intelligence several times a day, while another 28% say their use is about once a day. But many people might be unaware of the environmental impact of their searches. A request made using ChatGPT, for example, consumes 10 times the electricity of a Google search, according to the International Energy Agency. In addition, data centers, which are essential for powering AI models, represented 4.4% of all the electricity consumed in the U.S. in 2023—and by 2028 they’re expected to consume approximately 6.7 to 12% of the country’s electricity.
A new study, published in Frontiers, aims to draw more attention to the issue. Researchers analyzed the number of “tokens”—the smallest units of data that a language model uses to process and generate text—required to produce responses, and found that certain prompts can release up to 50 times more CO2 emissions than others.
Different AI models use a different number of parameters; those with more parameters often perform better. The study examined 14 large language models (LLMs) ranging from seven to 72 billion parameters, asking them the same 1,000 benchmark questions across a range of subjects. Parameters are the internal variables that a model learns during training, and then uses to produce results.
Reasoning-enabled models, which are able to perform more complex tasks, on average created 543.5 “thinking” tokens per question (these are additional units of data that reasoning LLMs generate before producing an answer). That’s compared to more concise models which required just 37.7 tokens per question. The more tokens were used, the higher the emissions—regardless of whether or not the answer was correct.
The subject matter of the topics impacted the amount of emissions produced. Questions on straightforward topics, like high school history, produced up to six times fewer emissions than subjects like abstract algebra or philosophy, which required lengthy reasoning processes.
Currently, many models have an inherent “accuracy-sustainability trade-off,” researchers say. The model which researchers deemed the most accurate, the reasoning-enabled Cogito model, produced three times more CO2 emissions than similar sized models that generated more concise answers. The inherent challenge then, in the current landscape of AI models, is to be able to optimize both energy efficiency and accuracy. “None of the models that kept emissions below 500 grams of CO₂ equivalent achieved higher than 80% accuracy on answering the 1,000 questions correctly,” first author Maximilian Dauner, a researcher at Hochschule München University of Applied Sciences, said in a press release.
It’s not just the types of questions asked or the degree of the answer’s accuracy, but the models themselves that can lead to the difference in emissions. Researchers found that some language models produce more emissions than others. For DeepSeek R1 (70 billion parameters) to answer 600,000 questions would create CO2 emissions equal to a round-trip flight from London to New York, while Qwen 2.5 (72 billion parameters) can answer over three times as many questions—about 1.9 million—with similar accuracy rates and the same number of emissions.
The researchers hope that users might be more mindful of the environmental impact of their AI use. “If users know the exact CO₂ cost of their AI-generated outputs, such as casually turning themselves into an action figure, they might be more selective and thoughtful about when and how they use these technologies,” said Dauner.
Q: What is the environmental impact of using ChatGPT compared to Google?
A: A request made using ChatGPT consumes 10 times the electricity of a Google search, according to the International Energy Agency.
Q: How do different AI models affect carbon emissions?
A: Different AI models use varying numbers of parameters, and those with more parameters often produce more emissions. The study found that reasoning-enabled models, which perform more complex tasks, can create up to 50 times more CO2 emissions than more concise models.
Q: What subjects produce the most emissions in AI models?
A: Subjects like abstract algebra or philosophy, which require lengthy reasoning processes, produce up to six times more emissions than straightforward topics like high school history.
Q: What is the 'accuracy-sustainability trade-off' in AI models?
A: Many AI models have an inherent trade-off between accuracy and sustainability. The most accurate models often produce more emissions, while more energy-efficient models may not achieve the same level of accuracy.
Q: How can users be more mindful of the environmental impact of AI?
A: Users can be more selective and thoughtful about when and how they use AI technologies by being aware of the CO2 cost of their AI-generated outputs. This can help reduce the environmental impact of AI use.