9 Mar 2024
Watt's in our Query? Decoding the Energy of AI Interactions
Resources
As we greet the New Year with aspirations and resolutions, let's add a critical one to our list: sustainability in our digital lives. With every leap in technology, like GPT-4, we marvel at the new horizons of human-like text generation and problem-solving capabilities. However, as we stand at the dawn of 2024, it's time to shine a light on a less-discussed aspect of these advancements: their environmental impact.
The mechanism that allows LLMs to converse, create, and compute is underpinned by a complex web of power consumption that stretches far beyond the data center. As the community becomes increasingly aware of our ecological responsibilities, we're introducing a game-changer: the AI Carbon Tracker Chrome Extension: Carbon ScaleDown. It's not just a tool; it's a movement towards mindful AI usage that aligns with our planet's health.
In this blog, we'll unfold the narrative of LLM’s power dynamics, from the hefty energy demands of models like GPT-4 to the nuanced differences between text and image processing. We'll explore the role of 'inference' - the AI's day job - in the broader picture of sustainability. Most importantly, we'll showcase how our Chrome Extension, built on React and Tailwind CSS and hosted on AWS, isn't just tracking but actively helping reduce the carbon footprint of your AI interactions.
As we venture into the details of AI's environmental footprint, it's crucial to understand the genesis of our concerns. Foundational models like BERT and GPT-2, while setting benchmarks in machine learning, also highlighted the intensive energy requirements of such technologies. The training of these models is not just a marvel of computation but also a hefty draw on power resources.
Here, the "Energy and Policy Considerations for Deep Learning in NLP" [1] paper comes into sharp focus, examining the carbon ledger of AI and questioning the sustainability of our digital advancements. The paper talks about BERT’s training, for instance, is not a mere computational task; it's a power-intensive process. To put this into context, the BERT base model with 110 million parameters required 96 hours of training on 16 TPU chips. This is akin to leaving a modern LED light bulb turned on for over a decade. GPT-2, even larger with 1.5 billion parameters, demanded a week of training on 32 TPUv3 chips—a testament to its colossal energy footprint.
This power consumption isn't just about electricity bills but the carbon footprint. Training these models is equivalent to carbon emissions from an average American's life for a year. When we talk about LLM training, we speak the language of kilowatt-hours and carbon emissions, which translate to real-world environmental impact.
What's less discussed in this paper is the ongoing environmental cost as these models are put to work daily by millions. It's this continuous use—that far outstrips the duration of their training.
The Power Dynamic of AI Tasks
Moving on from the training phase, let's discuss inference when a machine learning model is used to power applications. Here's where it gets interesting: not all machine learning tasks are equal, especially regarding energy consumption.
Energy consumption varies drastically, with more complex tasks consuming more power. GPT-4, with its billions of parameters, is akin to a digital polymath, capable of composing poetry, coding, and even creating art from textual descriptions. However, this versatility comes with an energy demand that's not just a step but a leap from its predecessors.
Imagine AI tasks as different appliances in your home. Some, like your LED bulb, sip electricity gently. Others, like your air conditioner on a hot day, gulp it down. The paper "Power Hungry Processing" [2] explains this dynamic. Text classification, for instance, is like your LED bulb - it's relatively energy-efficient. But when you have a model that generates text or images, the energy consumption jumps significantly – think of it as moving from the bulb to the air conditioner.