Acknowledgments: This article was made possible by contributions and collaboration from the entire Machine Learning team at eyeo. Thank you for running the benchmarks, crunching the numbers and reviewing the content.
Sustainable AI is defined by two branches– “AI for sustainability” and “sustainability of AI”. (1) AI for sustainability is using AI to help promote environmental sustainability, e.g. using AI to decarbonize media by optimizing the supply chain of ads (supply-path-optimization) in the advertising industry. (2) The sustainability of AI is developing AI with minimal negative costs to the environment.
While much of the ML community is moving towards bigger datasets and multi-billion parameter models, we are moving in the opposite direction— making our models as lean and as energy efficient as practically possible. This blog will illustrate our approach to ensuring that our entire machine-learning infrastructure is just as effective as any other model but also aligns with our purpose of transforming the web into a safe, sustainable and accessible place.
If we’re talking about using AI ethically, the environmental impact of the technology cannot be ignored. The energy consumption of popular ML models is staggering and growing exponentially. Here are some figures relating to the training and inference of Machine Language Models according to a 2019 study by the University of Massachusetts Amherst:
To protect user privacy, it is imperative for our models to run in the browser, which has limited computing power and memory thus necessitating the need for leaner models. To focus on keeping our carbon footprint close to zero without compromising effectiveness, we have consciously adopted the following best practices to optimize our ML models and support infrastructure.
In theory, all those best practices sound helpful but are they really effective? We plugged our dataset into the Machine Learning Emissions Calculator to estimate our carbon footprint. Here are the findings:
Training emissions
Training our general-purpose ad-filtering model experiments were conducted:
Total emissions are estimated to be 8.68 kg CO2 of which 100 percent were directly offset by the cloud provider (GCP in our case).
Training energy consumption
Our model saves more than 1000x the energy than the standard Large Language Models for training:
Inference energy consumption
On average our energy consumption is ~0.1 kWh per user session. Much less than the energy consumed by a fluorescent light bulb and an incandescent lamp:
Inference emissions
The EPA estimate that 1 kilowatt-hour of energy consumption generates 0.43 kg of carbon emissions. Our model, which only consumes 0.1 kilowatt-hour of energy per user session, emits 0.043 kg of carbon dioxide.
If we are going to create something to promote sustainability, we need to consider the ethics and synchronicity of doing it in a sustainable way.
eyeo Director of Engineering, Dr. Humera Noor Minhas states,
“Through our commitment to sustainability and ethical AI, we harness the power of machine learning responsibly, ensuring our technological advancements align with a greener future.” It is a clear sign of our imperatives to sustainability that all signposts and stations of the process are approached with an eye toward optimization.
We encourage others to look at their own models to see where they can reduce their carbon emissions. In this challenge, you may just find an opportunity.
Come back for the next blog in the series “eyeo and Ethical AI” where we discuss using AI to protect user choice.
––
Machine Learning and AI will be one of the focus topics of the upcoming 2023 Ad-Filtering Dev Summit in Amsterdam and online from 4-5 October 2023. Registration is now open, save your spot here.