A Short Overview of Rebound Effects
in Methods of Artificial Intelligence
Martina Willenbacher1,2*, Torsten Hornauer2 and Volker Wohlgemuth2
1Leuphana University of Lüneburg, Institute for Environmental Communication, Germany
2University of Applied Sciences HTW Berlin, School of Engineering - Technology and Life, Berlin
Submission: July 27, 2021; Published: August 12, 2021
*Corresponding author: Martina Willenbacher, Leuphana University of Lüneburg, Institute for Environmental Communication, Universitätsallee 1, 21335 Lüneburg, Germany & University of Applied Sciences HTW Berlin, School of Engineering - Technology and Life, Treskowallee 8, 10318 Berlin
How to cite this article: Martina W, Torsten H, Volker W. A Short Overview of Rebound Effects in Methods of Artificial Intelligence. Int J Environ Sci Nat Res.
2021; 28(5): 556246. DOI: 10.19080/IJESNR.2021.28.556246
Artificial intelligence (AI) is one of the pioneering technologies of the digital revolution in terms of the already existing and potential application areas in companies. On the technical side, this paper deals with the energy requirements for the use of artificial intelligence methods and the identification of efficiency approaches. Productivity increases often lead to increased energy demand, which contradicts sustainability in terms of reducing CO2 emissions. Therefore, it is investigated to what extent rebound effects can reduce the potential for energy savings in relation to artificial intelligence methods and what are the main factors of the resulting CO2 emissions.
Methods of artificial intelligence have gained enormous popularity in recent years, also due to increasingly cheap storage media and graphical computing units. In particular, artificial multi-layered neural networks have received a lot of media attention in the last ten years and also  have achieved a sharp increase in popularity in the scientific field. Especially in the field of artificial neural networks, enormous amounts of data are processed based on nonlinear functions e.g., to enable the system to autonomously find relationships within the data set .
The more extensive the training of the AI, the more complex the calculations and the associated hardware requirements, especially regarding the tensor processing unit (TPU), but also the GPU. For the calculation of the algebraic basics, these processors are better suited than a conventional CPU . When using artificial intelligence methods, most of the energy used is not required by the processing of the matrix multiplications themselves, but by accessing the storage areas and the associated rewriting of the data to be processed. Using the example of a CPU-based 8-core server CPU on a 45-nanocenter infrastructure, it was shown that both addition and multiplication processes require only a fraction of the energy of data access. Even the comparatively complex multiplication of floating-point numbers with a size of 32 bits requires little more than a third of the energy of memory access to 8 KB in the processor's cache. If the data is in the main memory, the energy requirement increases by a factor of approx. 1000  (Figure 1).
With increasing size of the training data, the power consumption increases exorbitantly simply by the necessary rewriting of the data.
If the increase in efficiency in the energy consumption of one sector leads to the fact that the savings achieved by the lower energy demand are reinvested elsewhere to consume more energy in another way, this is called the rebound effect. The strength of the energy consumption associated with AI processes and the greenhouse gas potential essentially depends on the question of how strongly the purchased electricity is composed of fossil primary energy sources and how efficiently the hardware works for the respective calculation . In addition to transport and extraction, the manufacturing processes themselves have the potential to exacerbate regional conflicts or  release toxic chemicals into the environment (Figure 2).
Strubel et al.  from the University of Massachusetts estimated, among other things, the energy requirements, the Co2 equivalent and the approximate computing costs for cloud providers to train several popular  natural language programming models. Particularly noteworthy were the values for a Transformer model. The basis of this model was the TensorFlow library "Tensor2Tensor" [9,10]. A model with 213 million parameters was used for the Weighting. Neural architecture search (NAS) was used in the existing transformers architecture as an additional feature. For the calculations, a TP integrated Uv2 (tensor processor) chip was used to train the AI. The TPU needed 10 hours . Strubel et al.  calculates that when using 8 GPUs (Nvidia Tesla P100), this would correspond to a training time of just over 274 hours. An exponential trend in machine learning in terms of the demand for computing capacity through relatively popular sets has been observed since about 2012.
Before 2012, it was not common for GPU to be used for computations. The further back in time you look, the more relevant the price differences for RAM and persistent memory may have been. In addition, the development of machine learning may have been accelerated by more powerful algorithms and by the popularity of economic applications.
In addition to the enormous environmental impacts described above, the resource requirements of current AI models also represent another critical trend: In orders of magnitude beyond the Evolved Transformer, the costs for electricity and hardware (or cloud service rental) can reach seven or eight figures. The Green AI concept pursues the approach of bringing not only the improvement of a performative best, but also the efficiency with which a certain result was achieved, into the scientific discourse on AI. The weighting could thus be composed of the number of examples used (E), the size of the training data set (D) and the number of AI trainings by hyperparameter experiments (H) and thus scale the result (R) or the accuracy of the AI .
The resulting weighted costs place a strong focus on efficient AI, as the parameters E, D and H are directly reflected in the computational effort. All in all, Green AI is trying to establish a more complex approach. The focus should not only be on the accuracy of an AI, but also on additional factors such as money, energy consumption and, as a result, the environmental impact.
Just as Green AI seeks to convey a more holistic picture of the AI community's goal, from a sustainability perspective, it is necessary to expand efficiency to levels such as consistency and sufficiency. The proportion of recyclable components should be as high as possible. Ideally, the same processed resources are recycled and reused repeatedly for new products . Since intangible goods are created in software development, this logic can only be transferred to the product indirectly through concepts such as frameworks. Rather, software without protection mechanisms is only bound to the hardware requirements in terms of its duplicability. The use of artificial intelligence also raises the question of whether techniques are used for their own sake or whether innovations should be seriously promoted. Rebound effects can also be used as a synonym for growth ambitions. Of course, this is also the case if economic growth is to be stimulated with the help of AI. If a society and its economic system are dependent on growth, there will also be rebound effects. In combination with an efficient circular economy and conscious consumer behavior, the effects of rebound effects can potentially be reduced.
Innovative technologies such as AI can make meaningful contributions – at best if they are used in a targeted manner. More digitization does not necessarily lead to more efficiency.