PhD Defense
The defense is scheduled for Tuesday, October 15th, 2024, at 3:00 pm. The event will take place at ENSIMAG (Amphithéâtre H, Building H), 161 Rue des Mathématiques 89, 38400 Saint-Martin-d’Hères (maps link).
It will also be broadcast live: https://youtube.com/live/5V0VVWN5Qm4. Save the link!
The reviewed manuscript of my thesis can be found here: Link to Manuscript. It is entitled: “A Versatile Methodology for Assessing the Electricity Consumption and Environmental Footprint of Machine Learning Training: from Supercomputers to Edge Devices”. The abstract can be found at the bottom of this page. The presentation will be in English.
Jury
Reviewers
- Aurélie BUGEAU, Professeure des Universités, Université de Bordeaux
- Anne-Laure LIGOZAT, Professeure des Universités, ENSIIE
Examinators
- Emma STRUBELL, Assistant Professor, Carnegie Mellon University
- Sylvain BOUVERET, Maître de conférences, Grenoble INP - Université de Grenoble Alpes
- Claude LEPAPE, Ingénieur de Recherche, Schneider Electric
- Claudia RONCANCIO, Professeure des Universités, Grenoble INP - Université Grenoble Alpes
Guest
- Bruno MONNET, Ingénieur, Hewlett Packard Enterprise
PhD Supervisors
- Denis TRYSTRAM, Professeur des Universités, Grenoble INP - Université de Grenoble Alpes
- Laurent LEFÈVRE, Chargé de Recherche HDR, Inria
Abstract
The number of Artificial Intelligence applications being developed and deployed is continually increasing. The effects of these activities on the biosphere, particularly on climate change, have attracted attention since 2019, but assessment methodologies still require improvement. More advanced evaluation methods and a deeper understanding of these impacts are necessary to minimize the environmental impacts of artificial intelligence.
With an emphasis on the training phase, this thesis investigates how machine learning (ML) affects the environment.
First, we question how the electricity consumption of IT infrastructures is measured by comparing power meters currently in use with different benchmarks and infrastructures, focusing on Graphic Processing Units (GPUs). These findings are used to analyze the electricity required to train models selected from the MLPerf benchmark on various ML infrastructures, ranging from an edge device to a supercomputer. Finally, the thesis shifts toward examining the more general environmental impacts of ML, based on an estimation of the embodied impacts of ML infrastructures. These impacts are allocated to each model training, enabling a comparison with the impacts of electricity usage. While numerous ML environmental impact indicators exist, this study focuses on primary energy consumption, global warming potential, and abiotic depletion potential for minerals and metals.
In conclusion, this thesis proposes a methodology that enables a reproducible multi-criteria evaluation of the impact of machine learning training on the environment and can be applied to different ML infrastructures, thus enabling fair comparison and enlightened choices.