Energy and Carbon Considerations of Fine-Tuning BERT

3Citations
Citations of this article
21Readers
Mendeley users who have this article in their library.

Abstract

Despite the popularity of the pre-train then fine-tune paradigm in the NLP community, existing work quantifying energy costs and associated carbon emissions has largely focused on language model pre-training. Although a single pre-training run draws substantially more energy than fine-tuning, fine-tuning is performed more frequently by many more individual actors, and thus must be accounted for when considering the energy and carbon footprint of NLP. In order to better characterize the role of finetuning in the landscape of energy and carbon emissions in NLP, we perform a careful empirical study of the computational costs of finetuning across tasks, datasets, hardware infrastructure and measurement modalities. Our experimental results allow us to place fine-tuning energy and carbon costs into perspective with respect to pre-training and inference, and outline recommendations to NLP researchers and practitioners who wish to improve their finetuning energy efficiency.

References Powered by Scopus

Aligning books and movies: Towards story-like visual explanations by watching movies and reading books

1676Citations
N/AReaders
Get full text

Green AI

737Citations
N/AReaders
Get full text

Aligning artificial intelligence with climate change mitigation

175Citations
N/AReaders
Get full text

Cited by Powered by Scopus

UFEL: a By-Design Understandable and Frugal Entity Linking System for French Microposts

0Citations
N/AReaders
Get full text

The Sunk Carbon Fallacy: Rethinking Carbon Footprint Metrics for Effective Carbon-Aware Scheduling

0Citations
N/AReaders
Get full text

Inference-Optimized Metaheuristic Approach for a Prospective AI Training/Inference Inversion Paradigm in Optimized Energy-Aware Computing

0Citations
N/AReaders
Get full text

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Cite

CITATION STYLE

APA

Wang, X., Na, C., Strubell, E., Friedler, S. A., & Luccioni, S. (2023). Energy and Carbon Considerations of Fine-Tuning BERT. In Findings of the Association for Computational Linguistics: EMNLP 2023 (pp. 9058–9069). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.findings-emnlp.607

Readers over time

‘23‘24‘25036912

Readers' Seniority

Tooltip

PhD / Post grad / Masters / Doc 8

73%

Researcher 2

18%

Lecturer / Post doc 1

9%

Readers' Discipline

Tooltip

Computer Science 8

73%

Medicine and Dentistry 1

9%

Physics and Astronomy 1

9%

Linguistics 1

9%

Save time finding and organizing research with Mendeley

Sign up for free
0