Home >Technology peripherals >AI >Behind the large AI model, there are staggering carbon emissions
Since large-scale language models like ChatGPT became a global sensation, few people have noticed that training and running large-scale language models is generating staggering carbon emissions.
Although neither OpenAI nor Google have said how much their respective products’ computing costs are, according to third-party researchers According to the analysis, the ChatGPT part of the training consumed 1,287 megawatt hours and resulted in more than 550 tons of carbon dioxide emissions, which is equivalent to a person traveling back and forth between New York and San Francisco 550 times.
In fact, this is only the emission during training. More carbon dioxide will be emitted when the AI large model is running.
Martin Bouchard, co-founder of QScale, a Canadian data center company, believes that in order to meet the growing needs of search engine users, Microsoft and Google have added generative AI such as ChatGPT to searches, which will lead to Each search increases the computational effort by at least 4 to 5 times.
If you have to retrain the model frequently and add more parameters, the scale of calculation is completely different.
According to the International Energy Agency (International Energy Agency), greenhouse gas emissions from data centers already account for about 1% of global greenhouse gas emissions.
This number is expected to rise as large AI models and demand for cloud computing grow.
AI large models are becoming an important source of carbon emissions.
The training and operation process of AI models consumes a lot of energy, but the key issue is how to know and measure a single machine How many greenhouse gas emissions are learning experiments generating, and how much can be reduced?
Currently, data scientists still cannot easily and reliably obtain measurement results in this field, which also hinders the further development of feasible response strategies.
In response to this problem, Google published a study detailing the energy costs of state-of-the-art language models, including early and larger versions of LaMDA.
Research results show that combining efficient models, processors and data centers with clean energy can reduce the carbon footprint of machine learning systems by up to 1,000 times.
The team proposed four basic methods to significantly reduce the carbon (and energy) footprint of machine learning workloads, which are currently used at Google and anyone using Google Cloud services Can be used by anyone.
Google energy and carbon footprint reduction best practices (4Ms) are as follows:
The cloud-based data center is a new, custom-designed warehouse with energy-efficient features that can accommodate 50,000 servers. They provide exceptionally efficient power utilization (PUE).
Therefore, computing in the cloud instead of locally can save 1.4-2 times energy and reduce pollution.
Google data shows that machine learning training and inference accounted for only 10% to 15% of Google’s overall energy use over the past three years, with 35% used annually. For inference, 25% is used for training.
To find improved machine learning models, Google uses Neural Architecture Search (NAS).
NAS is typically performed only once per problem domain/search space combination, and the resulting model can then be reused for hundreds of applications. The one-time cost of NAS is typically used on an ongoing basis. offset by emission reductions.
Researchers conducted a study to train the Transformer model.
To do this, they used Nvidia P100 GPUs in a typical data center with an energy mix similar to the global average, while using next-generation ML hardware such as TPUv4, performance improved over the P100 14 times.
At the same time, efficient cloud data centers save 1.4 times more energy than ordinary data centers, thus reducing total energy consumption by 83 times.
Additionally, data centers powered by low-carbon energy can reduce carbon emissions by a further nine times, for a total reduction of 747 times over four years.
The Google team believes that in the field of information technology, the life cycle costs of manufacturing computing devices of all types and sizes are much higher than the operating costs of machine learning.
Emissions Estimated manufacturing costs include the embedded carbon emitted by manufacturing all relevant components, from chips to data center buildings.
Of course, in addition to using the 4Ms approach, service providers and users can also take simple steps to improve their carbon footprint performance, such as:
Customers should analyze and reduce their energy use and carbon footprint by having data center providers report on data center efficiency and cleanliness of energy supply for each location.
Engineers should train models on the fastest processors in the greenest data centers, which are increasingly in the cloud.
Machine learning researchers should focus on designing more efficient models, such as taking advantage of sparsity or including retrieval to reduce the model.
Additionally, they should report their energy consumption and carbon impact. Not only will this encourage competition beyond model quality, but it will also ensure that their work is properly accounted for.
Although the large AI model is a major contributor to carbon emissions, cutting-edge technologies represented by AI are also making contributions to reducing carbon emissions. Make a contribution.
A study jointly conducted by Baidu and the consulting agency IDC (International Data Corporation) shows that the contribution of AI-related technologies to carbon reduction will increase year by year and will reach at least 70% by 2060. , the total amount of carbon reduction is expected to exceed 35 billion tons.
Taking the transportation industry as an example, the carbon emissions of China's transportation industry in 2020 are estimated to be 1.04 billion tons, accounting for 9% of the country's overall emissions.
In the process of driving the transportation industry to reduce carbon emissions, the use of congestion-relieving intelligent transportation technology based on intelligent information control can effectively improve the traffic efficiency of major urban road intersections. Therefore, cities with a population of tens of millions can reduce at least 41,600 tons of carbon emissions every year - which is equivalent to the carbon emissions of 14,000 private cars driving for one year.
Judging from current practice, the key to understanding and achieving emission reduction is to predict and monitor the effect of emission reduction, and AI has the function of predicting emissions and monitoring emissions in energy conservation and emission reduction. , three key applications to reduce emissions.
According to the "White Paper on Carbon Neutral Industry Development", in terms of predicting emissions, AI can predict future carbon emissions based on current emissions reduction efforts and needs, and at the same time determine carbon emissions. Lower emissions guidelines.
In terms of monitoring emissions, AI can track carbon footprint data in real time and collect data from all aspects of procurement, production, sales, operation and maintenance, logistics, etc., to improve monitoring accuracy.
In terms of reducing emissions, after AI collects data from each link, it can optimize and adjust the workflow of each link from a global perspective.
In fact, in terms of AI assisting carbon emission reduction, it has been applied in many domestic fields.
In the field of new energy, the prominent problem lies in its volatility, randomness and intermittent characteristics.
Use AI technology combined with simulation calculations to predict the instability of wind and photovoltaic power, such as: combining wind speed, wind direction, light intensity and other natural meteorological characteristics to reasonably predict future power generation forecast, output more accurate power generation plans to the power grid, and shield the uncertainty and instability of new energy sources under the technical layer.
For another example, the jurisdiction of the Water Group includes raw water, water production, water supply, drainage, sewage, water conservation, etc.
Take the water supply for residents as an example. If the water pressure is too high, it will require a lot of energy and the leakage rate of the pipe network will be high, which may cause pipe explosions; if the water pressure is too low, it may cause pipe explosions. It will cause inconvenience for residents to use water.
In order to solve this problem, the water group deployed hardware sensors underground to monitor water pressure and built a water brain. On the premise of ensuring safe and stable water supply, AI technology can be used to achieve intelligence. Voltage regulation control and energy consumption optimization.
Not only that, AI carbon reduction technology is also used in business scenarios with high energy consumption such as power plants, parks, and data centers to accurately predict and control their production electricity needs. Optimize power-consuming equipment and carbon footprint.
The advancement of AI technology has brought many conveniences to mankind, but we must also pay attention to environmental issues during development.
How AI can achieve sustainable development in the future and how AI can better support changes in the dual-carbon field are still issues that need to be solved by all industries.
The above is the detailed content of Behind the large AI model, there are staggering carbon emissions. For more information, please follow other related articles on the PHP Chinese website!