Reducing the carbon footprint of natural language processing
The recent dramatic advances in natural language processing (NLP) technology, such as neural machine translation (NMT) and large language models (LLM), are changing the way people work and interact with technology. These new NLP technologies have the potential to increase productivity and levels of automation in a wide variety of fields.
The downside of the new NLP technology is its enormous energy consumption. At a time when energy efficiency has become essential due to the climate crisis, the advances in NLP are vastly increasing the energy usage of the IT sector. The GreenNLP project addresses this issue by developing more environmentally sustainable ways of building and using NLP applications.
An important part of the GreenNLP project is to disseminate guides and best-practices on efficient training of large language models in supercomputers. The first version of our “Working with large language models on supercomputers”-guide has been published on CSC’s documentation site.
13 September 2024
LUMI User Support Team (LUST) together with CSC and DeiC arranges a two day workshop on how to use LUMI for training AI models. Participants will get to try out fine-tuning a language model on LUMI and scaling it up to multiple GPUs and multiple nodes.
20 May 2024
CSC, a partner in the GreenNLP project, has evaluated the scalability of large language model (LLM) training on the LUMI supercomputer. The results indicate that there are no fundamental scaling bottlenecks even when training with thousands of GPUs.
11 January 2024