Smiling man running in mid-air

Next-gen AI – the answer to cost and energy concerns?

In the last issue of our three-part series exploring user sentiment toward AI, we test thesis 3: Next-gen AI – low-energy, low-cost.


In brief:

  • Efficient models like DeepSeek deliver strong results while using fewer resources, challenging the need for massive, energy-hungry AI systems. 
  • Open-source AI democratizes access, enabling more developers to use and adapt advanced models without relying on big tech infrastructure.
  • AI’s future lies in balancing efficiency for practical tasks with powerful models that drive innovation and solve complex, novel problems.

The rise of efficient models like DeepSeek – trained in China at a fraction of ChatGPT’s cost – has sparked a compelling debate in the AI world. These models deliver strong performance based on far fewer training resources and can rival leaders like ChatGPT or Claude in many tasks. Their success raises an important question: Will cheaper, more efficient models eventually reduce the need for massive, energy-intensive AI systems? The answer may be more complex than it appears.

Thesis 3: Smart, low-cost AI models can contain energy consumption as AI use spreads

A lightweight AI model developed in China, DeepSeek delivers competitive performance while using significantly fewer training resources than peers like ChatGPT. It thus demonstrates that high-quality results don’t always require massive datasets or energy-intensive infrastructure. As the race to develop AI drives ever-larger models with soaring energy demands, DeepSeek underscores that smarter design and efficient training can offer a more sustainable path forward – potentially making the high costs and energy requirements of today’s leading models less critical.

The challenge: The race to develop AI has only just begun

What requires more brainpower – Einstein developing the theory of relativity, or a physics student learning it in class? Einstein spent years grappling with fundamental questions about the universe, challenging established paradigms and constructing entirely new frameworks of understanding. In contrast, while it is certainly an effort for a student to master these concepts, they're following a well-mapped path of established knowledge.

 

Similarly, today’s large language models build on existing knowledge. They rely on human feedback and carefully crafted algorithms to determine what information to prioritize or withhold when answering questions. It is now becoming apparent that smaller models like DeepSeek can be much more efficient than previously assumed, yet still provide high-quality responses based on what they’ve learned.

 

When these models encounter novel problems and are given time to work through them step-by-step like humans do, they can reach new conclusions from their fundamental knowledge. However, solving complex problems this way requires immense computational time – potentially hours, days or even months. In the current race to advance AI and expand data center capabilities, the focus isn’t just on processing user queries faster. Rather, the long-term aim is to develop AI systems capable of restructuring their understanding of the world based on fundamental principles – mirroring, in a sense, the human capacity for deep reasoning.

 

Models designed for specialized domains require significant computing power to research new approaches and integrate them into their functionality. This extends to the goal of enabling AI to learn independently, much like humans do.

The future energy demands of developing AI systems that surpass current human understanding remain difficult to predict. However, it is expected that algorithmic advances will primarily accelerate progress toward artificial superintelligence (ASI), rather than drastically cut energy usage. ASI refers to a hypothetical AI system whose intellectual abilities surpass those of any human –featuring advanced cognitive functions and reasoning capabilities far beyond our own.

The counterargument: An opportunity to democratize AI access

The rise of efficient open-weight models such as DeepSeek, Mixtral and Llama brings significant opportunities. These systems make sophisticated language models available to a broader range of companies and developers, thereby democratizing access to AI technology. Open-weight models are not only more affordable, but they can also be deployed without having to transfer data to a third-party provider. Additionally, they can be further fine-tuned using proprietary data.

There is growing evidence that highly intelligent models like GPT-4+ can be used to train smaller, more efficient models that come close to their trainers in terms of intelligence and capability. Much like applying relativity theory in practice doesn’t require Einstein himself, but rather a capable student who can work with the distilled insights, smaller models can effectively build on complex knowledge.

Not every application requires the most advanced AI capabilities. Many practical use cases can be served efficiently by lightweight models, addressing a current challenge: while powerful AI systems integrated into businesses remain expensive, smaller models often lack the intelligence needed for broader application.

The success of open-source models like DeepSeek challenges the assumption that “bigger is always better”, and encourages innovation in model architecture and training methods. It’s becoming increasingly clear that cutting-edge infrastructure isn’t always a prerequisite for innovation. By focusing on the right strategies, developers can rival proprietary advancements from industry giants.

Competitive pressure of this kind is healthy for the AI landscape. It drives improvements in both efficiency and performance, fostering a more dynamic and inclusive environment for technological progress.

The outlook: Dual-track development

The future of AI development is likely to unfold along two parallel paths. On the one hand, we can expect the rise of efficient models designed to make AI more accessible for everyday applications and increase its practical utility. Such models excel at understanding and applying existing knowledge – for example, in content creation, data analysis and solving routine problems.

On the other hand, advanced models will continue to push the boundaries of what artificial intelligence can achieve. Although these systems demand significant computing power, they focus on tasks that require genuine innovation and novel problem-solving approaches. In this way, they serve a role similar to research labs in the AI ecosystem – driving breakthroughs that may eventually lead to more efficient implementations.

While lower-cost models will make AI more widely available, the pursuit of new insights and capabilities will continue to rely on substantial computational resources. The future of AI won’t be about choosing between efficiency and performance, but about understanding when each approach is best suited to the task at hand.

Summary

Efficient AI models like DeepSeek are reshaping the landscape by offering high-quality performance while using fewer resources, challenging the dominance of massive, energy-intensive systems. These lightweight, often open-source models democratize AI access and enable practical applications without sacrificing quality. However, breakthroughs in advanced AI still require significant computing power, mirroring research labs in their pursuit of innovation. As AI evolves, development will follow two parallel paths: one focused on efficiency and accessibility, the other on pushing technological boundaries. The future lies not in choosing between them, but in knowing which approach best suits the task.

Explore how EY can help you with AI

Explore our insights that look at how you can build confidence in AI, drive exponential value throughout your organization and deliver positive human impact.

Brazil town from sky

Related articles

Has AI enthusiasm peaked?

Are people growing tired of AI-generated content? Explore alternative theses on user sentiment toward AI.

Is AGI the inevitable next step for businesses?

Is human-level AI feasible in a business context? Find out how it measures up in a corporate environment in terms of its potential, cost and risks.

    About this article

    Request for proposal (RFP) - exclusively for Switzerland

    |

    Submit your request now!