By: Johannes Fiegenbaum on 7/29/25 7:23 PM
AI token prices have dropped dramatically – from €36 to as low as €0.07 per million tokens. This opens up entirely new opportunities for you to use AI technologies cost-effectively. Especially for small and medium-sized enterprises, AI projects are now becoming affordable. At the same time, the price drop brings challenges, such as increased competition and potential environmental impacts due to higher energy consumption. According to industry analysis, this price reduction represents one of the most significant cost shifts in the AI sector to date, reshaping the competitive landscape and accessibility of advanced AI tools.
Now is the time to adapt your business models and benefit from these developments—without losing sight of long-term risks and sustainability considerations.
The falling prices for AI tokens can be traced back to three main factors: technological advances, increasing competition, and economies of scale. Each of these drivers is accelerating the democratization of AI, making advanced capabilities more accessible than ever before.
Technological progress plays a central role in reducing costs. Smaller language models, in particular, have shown that fewer parameters don’t necessarily mean worse results. An impressive example: While Google’s PaLM required 540 billion parameters in 2022 to achieve over 60 percent on the MMLU benchmark, Microsoft’s Phi-3-mini achieved this in 2024 with just 3.8 billion parameters—a 142-fold reduction in model size. This leap in efficiency demonstrates how targeted model design can drastically cut computational requirements without sacrificing performance.
This efficiency directly impacts costs, as smaller models require less computing power and energy. At the same time, hardware innovations such as NPUs, TPUs, and specialized AI accelerators enable faster processing with lower latency and reduced energy consumption. Estimates suggest that hardware costs for AI computations are dropping by about 30 percent annually, while energy efficiency is increasing by around 40 percent per year (McKinsey, 2023). A concrete example: NVIDIA managed to reduce the cost per token by a factor of 20 and increase inference speed by 25 times in February 2025, highlighting the rapid pace of hardware-driven cost optimization.
In addition to technological advances, intense competition between established providers and new market entrants is driving prices down. Open-source AI models play a key role here, as they spur innovation and reduce costs. Open-source AI is transforming the landscape by enabling more cost-effective implementations and lowering the barriers to entry for businesses of all sizes.
Two-thirds of surveyed organizations believe that open-source AI is more cost-effective to implement than proprietary models (Linux Foundation, 2024). Nearly half of companies cite cost savings as the main reason for choosing open-source solutions. An analysis by Permutable.ai shows that using OpenAI technology costs about $1 million per year—20 times more than using in-house models. This shift is also reflected in the growing adoption of open-source AI in sectors such as finance, healthcare, and logistics, where cost efficiency and customization are critical.
"The findings in this report make it clear: open source AI is a catalyst for economic growth and opportunity. As adoption scales across sectors, we're seeing measurable cost savings, increased productivity and rising demand for AI-related skills that can boost wages and career prospects. Open source AI is not only transforming how businesses operate – it's reshaping how people work."
- Hilary Carter, SVP of Research at The Linux Foundation
Technological advances and increased competition also contribute to economies of scale, further reducing the cost of AI tokens. Today, companies can bundle and process millions of requests in parallel, lowering the cost per token. These efficiency gains result from optimized infrastructure and smarter resource management, as seen in hyperscale cloud providers who maximize server utilization and energy efficiency.
Interestingly, increased demand has led to lower prices, as providers can better utilize their infrastructure. According to researchers, companies would have to spend 3.5 times more if there were no open-source software (Linux Foundation, 2024). Smaller companies, in particular, use open-source AI more frequently than larger ones, leveraging community-driven innovation to stay competitive.
"AI is the essential infrastructure of our time, just as electricity and the internet once were."
- Jensen Huang, NVIDIA Founder and CEO
The falling token prices are fundamentally changing the cost structure of AI applications and directly impacting business models and sustainability strategies. This development opens up new opportunities for digital transformation and ESG compliance, but also brings significant environmental risks. Below, we take a closer look at the opportunities and challenges of this digital transformation, drawing on recent research and industry case studies.
With up to 90% cost reduction, AI adoption is now economically attractive for small and medium-sized enterprises and startups for the first time. In particular, German SMEs, which have so far been hesitant to adopt AI due to high investment costs, benefit from this development. The previously high entry barriers for AI technologies are significantly lowered, creating space for innovation and growth. According to a McKinsey survey, 55% of SMEs plan to increase their AI investments in the next two years, citing affordability as a primary driver.
The reduced token costs offer companies a variety of ways to advance their digital transformation and achieve ESG goals. For example, AI-powered solutions can be used in sustainability reporting or climate risk analysis. According to forecasts, AI could help reduce global greenhouse gas emissions by 1.5 to 4% by 2030 (Capgemini, 2023).
Real-world examples show how modern AI analytics can evaluate massive amounts of data in real time. This enables progress in areas such as fleet optimization, risk management, and climate-related decision-making. Billions of data points are analyzed daily to gain insights into decarbonizing transportation, minimize supply chain risks, and assess climate-related hazards at millions of locations. For instance, Maersk uses AI to optimize shipping routes, reducing fuel consumption and emissions by up to 10% (Maersk, 2023).
Given the weather-related damages in Germany, estimated at around €145 billion between 2000 and 2021, AI-powered analytics can play a crucial role in identifying risks early and developing targeted prevention measures (BMUV, 2023).
Despite the lower token prices, the associated environmental costs must not be overlooked. Data centers could account for up to 21% of global energy demand by 2030—a dramatic increase compared to the current 1–2% (IEA, 2024). The energy consumption of a single ChatGPT request is almost ten times higher than that of a Google search. Forecasts predict that the energy demand of data centers due to AI applications could rise by 160% by 2030 (Nature, 2023).
Vijay Gadepally, lead scientist at the MIT Lincoln Laboratory, warns:
"As we move from text to video to images, these AI models are getting larger, and so are their energy impacts. This will result in a fairly significant energy consumption and a growing contribution to global emissions."
By 2030, the increased energy demand from data centers could reach about 8.4 TWh—equivalent to 3.25 gigatons of CO₂, or the same as 5 billion transcontinental flights in the US. In addition, global water demand from AI applications is expected to reach 4.2 to 6.6 billion cubic meters by 2027 (Nature, 2023).
For German companies, this means that the seemingly low token prices may bring hidden long-term environmental costs. These could jeopardize the achievement of sustainability goals and pose a challenge for ESG strategies. Reducing inference costs by 40% could not only make AI more affordable but also significantly reduce its environmental impact.
Falling token prices are prompting companies to adapt their business models. Now is the right time to take advantage of cost benefits while also introducing responsible practices for the long term. The following approaches will help you integrate AI effectively while acting in an environmentally conscious way. These measures directly address the challenges and opportunities created by lower token prices.
First, you should thoroughly reconsider your existing AI investment plans. With lower token costs, projects that were previously uneconomical are suddenly feasible. Take this opportunity to adjust your budget and implement additional AI applications.
Measuring AI emissions plays an important role here. This involves analyzing your computing power consumption and the energy sources used. Did you know that each ChatGPT request generates about 4.32g CO₂e? With millions of queries, these values add up to significant environmental impacts. Carbon calculators can help you better understand and specifically reduce your actual environmental costs. The Green AI Initiative provides tools and benchmarks for organizations to track and reduce their AI-related emissions.
The use of modern hardware and optimized algorithms is also becoming increasingly important. Large companies like Microsoft, Google, and Amazon are already relying on carbon-negative operations and renewable energy. German companies can follow this example by also switching their AI processes to green energy sources, as recommended by the International Energy Agency.
The new investment opportunities also open up chances to embed AI into your ESG strategies (Environmental, Social, and Governance). According to a study, 89% of investors consider ESG factors crucial for their decisions (PwC, 2023).
A first step is to define relevant ESG metrics, such as carbon emissions, energy consumption, social impact, or governance practices. Then you can select AI tools that integrate seamlessly into your existing systems.
Through automation and centralization of data collection, AI can gather ESG data from various sources and consolidate it on one platform. This enables real-time insights and interactive dashboards. At the same time, AI-powered validation tools improve data quality by comparing information with benchmarks and regulatory requirements. For example, Unilever uses AI-driven platforms to monitor supply chain sustainability and ensure compliance with global standards (Unilever, 2023).
Predictive analytics can also be used to forecast future ESG risks and opportunities based on historical data. With natural language processing, supplier communications can also be monitored to detect potential social conflicts early.
The economic benefits of AI should always be combined with a sustainable approach. Data center electricity consumption rose by 72% between 2019 and 2023—a clear signal for the need for sustainable strategies (IEA, 2024).
Your goal should be to develop resource-efficient AI models. This also means avoiding unnecessary AI applications to reduce the load on data centers. At the same time, responsible management of electronic waste is becoming increasingly important. By 2030, generative AI could cause between 1.2 and 5 million tons of e-waste annually (Nature, 2023).
A key approach is the transition to renewable energy. Data centers consume about 1–2% of global electricity, and training a single deep learning model can generate as much CO₂ as five cars over their entire lifetime. Choosing locations with optimal energy supply and more efficient cooling systems can help minimize negative environmental impacts. According to a NASA study, renewable energy could meet global electricity demand by 2050 if adopted at scale.
Additionally, you should foster a culture of sustainability within your company. Encourage your employees to adopt AI-powered sustainable practices and offer training to deepen their understanding of AI and its benefits. These measures form the foundation for a long-term and successful transformation of your business model.
The drastic reduction in token prices—from €36.00 per million tokens in March 2023 to just €0.07 with some providers—clearly shows how important it is for companies to remain flexible. Business models must be continuously adapted to secure economic advantages and act responsibly in the long term.
Dynamic price developments require regular review of operational processes. A leading streaming company demonstrates what effective AI monitoring looks like in practice: It continuously monitors input data, model outputs, and user engagement metrics. If a recommendation model suddenly suggests content that users skip more often, an internal review is automatically triggered. You too should regularly analyze your AI usage and choose service models that fit current pricing structures.
The numbers underline the trend: Already 35% of companies will rely on AI and machine learning for cost reduction in 2025, up from 22% in 2024 (Gartner, 2024). An example from the financial sector shows how proactive monitoring strategies in credit risk assessment and fraud detection ensure that machine learning models remain accurate even as market conditions change.
As the analyses show, AI monitoring is a crucial tool. Experts emphasize that continuous monitoring of the performance, behavior, and reliability of AI systems is essential to detect issues such as model drift, latency spikes, or data quality problems early. An automated alert system that makes adjustments in case of performance drops or data issues can be decisive here (McKinsey, 2023).
At the same time, it is important to strike a balance between cost optimization and sustainable development as part of ESG strategies. As token prices continue to fall, companies should ensure that increased AI use does not lead to unnecessary environmental burdens. Make sure to closely monitor performance and costs, scale computing resources as needed, and deploy smaller, more efficient model variants during peak times.
Companies that invest in continuous monitoring and flexible adaptation processes today can benefit from nearly free AI tokens tomorrow. At the same time, they create new business models that combine cost reduction with the necessary strategies for risk management and a sustainable digital future.
Small and medium-sized enterprises (SMEs) in Germany can benefit significantly from the reduced costs of AI tokens. This development makes it possible to implement data-intensive applications economically. As a result, numerous opportunities arise, such as:
In addition, AI-powered tools provide effective support for achieving sustainability goals. Examples include monitoring emissions, reducing energy consumption, or creating accurate sustainability reports. Thanks to lower token prices, access to advanced AI technology is easier. This not only helps SMEs remain competitive but also reduce their costs—a win-win situation for both business and the environment. According to McKinsey, SMEs adopting AI are 1.5 times more likely to report revenue growth than those that do not.
Companies can leverage AI to achieve their ESG goals (Environmental, Social, and Governance) more efficiently while simultaneously reducing environmental impact. One example is the use of AI to reduce energy consumption and promote resource-efficient operations. Automated analysis of ESG data enables informed decision-making and more transparent reporting. For instance, Siemens uses AI-driven analytics to track and optimize energy usage across its global facilities, resulting in a 15% reduction in emissions over three years (Siemens, 2023).
AI-powered technologies also offer the ability to make environmental monitoring more precise. For instance, emissions can be analyzed or supply chains monitored. This allows companies to identify potential environmental impacts early and take targeted action. It is important to also keep an eye on the energy efficiency of the AI systems used to further minimize your own ecological footprint.
Using open-source AI models comes with some security risks, including potential cyberattacks and data misuse. Since these models often come from publicly accessible sources, they can be vulnerable to manipulation. Additional challenges include data protection issues, ethical conflicts, unclear liability questions, and high costs that may arise from potential misuse. The NIST AI Risk Management Framework provides guidelines for mitigating such risks.
To minimize these risks, it is important to implement comprehensive security measures. This includes using firewalls, encryption technologies, and regular updates and security checks. Clear ethical guidelines should also be defined and responsibilities established within the company. Thoroughly reviewing the models and their origins helps identify and address vulnerabilities early on. Engaging with the open-source community and contributing to shared security standards can further enhance protection and trust.
A solo consultant supporting companies to shape the future and achieve long-term growth.
More aboutThe technology sector faces a major challenge: energy consumption and CO₂ emissions from AI and...
The EU AI Act is the world’s first legal framework specifically for artificial intelligence. It...