Techno Blender
Digitally Yours.

ChatGPT’s Electricity Consumption | by Kasper Groes Albin Ludvigsen | Mar, 2023

0 34


Opinion

ChatGPT may have consumed as much electricity as 175,000 people in January 2023.

Image: Midjourney

I recently wrote an article in which I guesstimated ChatGPT’s daily carbon footprint to the around 24 kgCO2e. With little information available at the time about ChatGPT’s user base, my estimate was based on the assumption that the ChatGPT service was running on 16 A100 GPUs. In lieu of recent reports that estimate that ChatGPT had 590 million visits in January [1], it’s likely that ChatGPT requires way more GPUs to servce its users.

From this it also follows naturally that ChatGPT is probably deployed in multiple geographic locations. This makes it very difficult to estimate ChatGPT’s total daily carbon footprint, because we would need to know exactly how many GPUs are running in which regions in order to incorporate the carbon intensity of electricity in each region into the carbon footprint estimate.

Estimating ChatGPT’s electricity consumption, on the other hand, is in principle simpler, because we do not need to know in which geographic regions ChatGPT is running. Below I explain how one can go about estimating ChatGPT’s energy consumption and I specifically produce an estimate of ChatGPT’s electricity use in January 2023. The scope is limited to January 2023 because we have some ChatGPT traffic estimates for this month.

Here’s how ChatGPT’s electricity consumption can be estimated:

  1. Estimate ChatGPT’s electricity consumption per query
  2. Estimate ChatGPT’s total number of queries for a given period in time
  3. Multiply those two

To calculate a range within which ChatGPT’s electricity consumption may lie, I’ll define 3 different values for each of points 1 and 2. Let’s first take a look at point 1.

Step 1: ChatGPT’s electricity consumption per query

BLOOM is a language model similar in size to ChatGPT’s underlying language model, GPT-3.

BLOOM once consumed 914 KWh of electricity during an 18 day period where it was deployed on 16 Nvidia A100 40 GB GPUs and handled an average of 558 requests per hour for a total of 230,768 queries. Queries were not batched. The 914 KWh account for CPU, RAM and GPU usage [2].

So BLOOM’s electricity consumption in that period was 0,00396 KWh per query.

Let’s consider how good a proxy BLOOM’s electricity usage might be for ChatGPT’s electricity usage.

First things first: BLOOM has 176b parameters, and GPT-3 (ChatGPT’s underlying model) has 175b parameters, so everything else held equal, their energy consumption should be quite similar.

Now, a few things suggest that ChatGPT’s electricity consumption per query could be lower than BLOOM’s. For instance, ChatGPT receives way more requests per hour than BLOOM did, which could lead to a lower energy consumption per query. Figure 1 below plots BLOOM’s electricity use against the number of requests and we can see that even at 0 requests, BLOOM had an idle consumption of 0.3 KWh, and at 1,000 requests the electricity consumption was less than 5 times higher than at 200 requests.

Figure 1: “The quantity of energy used by the GCP instance (on the y axis) versus the number of requests received by the instance in a 10 minute interval (on the x axis). It can be seen that even when zero requests are received by the instance in this time span (bottom left of the graph), the energy consumption remains at approximately 0.28 kWh.” (Figure and text by [2] p. 6)

Although the authors of the paper that measured BLOOM’s electricity consumption state that “since inference requests are unpredictable, optimization of GPU memory using techniques such as batching and padding […] is not possible” ([2] p. 5), we might expect that ChatGPT experiences a constant rate of requests high enough to make batching feasible. I would guess that batching of queries would lead to a lower energy consumption per query.

One thing that could increase ChatGPT’s energy consumption per query is if ChatGPT on average generates more words per query, but we don’t have access to any information that let’s us approximate this.

To err on the side of caution, I’ll assume in the following that ChatGPT’s electricity consumption is at max the same as BLOOM’s, so I’ll make an estimate of ChatGPT’s electricity consumption for each of the following assumptions:

  1. ChatGPT’s electricity consumption per query is the same as BLOOM’s, i.e. 0.00396 KWh
  2. ChatGPT’s electricity consumption per query is 75 % of BLOOM’s, i.e. 0.00297 KWh
  3. ChatGPT’s electricity consumption per query is 50 % of BLOOM’s, i.e. 0.00198 KWh

Step 2: Number of queries sent to ChatGPT

Now that we have some estimates of ChatGPT’s electricity consumption per query, we need to estimate how many queries ChatGPT received in January.

It’s been estimated that ChatGPT had 590m visits in January 2023 [1]. How many queries did each visit result in? We don’t know, but let’s make some assumptions. I would expect that each visit results in at least 1 query, so that’s the minimum assumption. Then, I’ll make a scenario with 5 and 10 queries per visit. Annecdotally, I’ve heard several people report that they use ChatGPT extensively in their daily tasks, so I believe these numbers are not unreasonable.

  1. Queries per visit = 1, so total number of queries = 590m
  2. Queries per visit = 5, so total number of queries = 2.95b
  3. Queries per visit = 10, so total number of queries = 5.9b

# 3 : Combining electricity consumption estimates with number of queries

Above, I outlined three different assumptions regarding ChatGPT’s electricity use per query and three different assumptions regarding the total number of queries.

This gives us 9 unique set of assumptions that I will refer to as “scenarios”.

Table 1 below shows the estimated electricity consumption for each of the scenarios.

The table shows that ChatGPT’s electricity consumption in January 2023 is estimated to be between 1,168,200 KWh and 23,364,000 KWh.

Here’s an online spreadsheet if you want to check my calculations.

Table 1: ChatGPT’s estimated electricity consumption in January 2023 in different energy efficiency and total queries scenarios. Table by Kasper Groes Albin Ludvigsen, based on Luccioni et al, 2022 [2]

Let’s put the numbers from Table 1 into perspective.

The annual electricity consumption of the average Dane is 1,600 KWh [3].

If the annual consumption is spread out evenly over the year, the average Dane used about 133.33 KWh of electricity in January.

So when we divide ChatGPT’s estimated electricity consumption by 133.3, we can see ChatGPT’s estimated consumption correspond to the monthly electricity consumption of between 8,762 and 175,234 Danes.

To put it into another perspective, you can run a standard 7w light bulb for around 19,050 years with 1,168,200 KWh of electricity (1,168,200 KWh / (7 / 1000) / 24 / 365).

Training ChatGPT’s underlying language model, GPT-3, consumed an estimated 1,287 MWh [4].

So at the low end of the estimated range for ChatGPT’s electricity consumption in January 2023, training GPT-3 and running ChatGPT for one month took roughly the same amount of energy. At the high end of the range, running ChatGPT took 18 times the energy it took to train.

Although handling a single query requires miniscule resources, the total amount of resources (e.g. electricity) add up to noticeable amounts when a product like ChatGPT scales. The search engine Bing now has incorporated ChatGPT [5] which might significantly increase the number of queries ChatGPT should handle, thereby also increasing the electricity consumption. In my opinion, this requires a debate about whether the incorporation of large machine learning models into existing products elevate the user experience to a level that justifies the large energy consumption.

In addition, unless the estimates put forward here are way too high, this article shows that running a machine learning based product in production can easily be way more energy consuming than training the model — especially if the product has a long life time and/or a large number of users. This should be taken into consideration when debating the environmental impact of AI and when thinking of how to reduce the life cycle carbon footprint of machine learning models.

It is important for me to emphasize that the estimates put forward in this article come with great uncertainties as very little information about how ChatGPT is run is publicly available. In the best of all worlds, OpenAI and Microsoft — and other tech companies for that matter — would disclose the energy consumption and carbon footprint of their products. But while we wait for that to happen, I hope my estimates will spur a debate and inspire others to come up with better estimates.


Opinion

ChatGPT may have consumed as much electricity as 175,000 people in January 2023.

Image: Midjourney

I recently wrote an article in which I guesstimated ChatGPT’s daily carbon footprint to the around 24 kgCO2e. With little information available at the time about ChatGPT’s user base, my estimate was based on the assumption that the ChatGPT service was running on 16 A100 GPUs. In lieu of recent reports that estimate that ChatGPT had 590 million visits in January [1], it’s likely that ChatGPT requires way more GPUs to servce its users.

From this it also follows naturally that ChatGPT is probably deployed in multiple geographic locations. This makes it very difficult to estimate ChatGPT’s total daily carbon footprint, because we would need to know exactly how many GPUs are running in which regions in order to incorporate the carbon intensity of electricity in each region into the carbon footprint estimate.

Estimating ChatGPT’s electricity consumption, on the other hand, is in principle simpler, because we do not need to know in which geographic regions ChatGPT is running. Below I explain how one can go about estimating ChatGPT’s energy consumption and I specifically produce an estimate of ChatGPT’s electricity use in January 2023. The scope is limited to January 2023 because we have some ChatGPT traffic estimates for this month.

Here’s how ChatGPT’s electricity consumption can be estimated:

  1. Estimate ChatGPT’s electricity consumption per query
  2. Estimate ChatGPT’s total number of queries for a given period in time
  3. Multiply those two

To calculate a range within which ChatGPT’s electricity consumption may lie, I’ll define 3 different values for each of points 1 and 2. Let’s first take a look at point 1.

Step 1: ChatGPT’s electricity consumption per query

BLOOM is a language model similar in size to ChatGPT’s underlying language model, GPT-3.

BLOOM once consumed 914 KWh of electricity during an 18 day period where it was deployed on 16 Nvidia A100 40 GB GPUs and handled an average of 558 requests per hour for a total of 230,768 queries. Queries were not batched. The 914 KWh account for CPU, RAM and GPU usage [2].

So BLOOM’s electricity consumption in that period was 0,00396 KWh per query.

Let’s consider how good a proxy BLOOM’s electricity usage might be for ChatGPT’s electricity usage.

First things first: BLOOM has 176b parameters, and GPT-3 (ChatGPT’s underlying model) has 175b parameters, so everything else held equal, their energy consumption should be quite similar.

Now, a few things suggest that ChatGPT’s electricity consumption per query could be lower than BLOOM’s. For instance, ChatGPT receives way more requests per hour than BLOOM did, which could lead to a lower energy consumption per query. Figure 1 below plots BLOOM’s electricity use against the number of requests and we can see that even at 0 requests, BLOOM had an idle consumption of 0.3 KWh, and at 1,000 requests the electricity consumption was less than 5 times higher than at 200 requests.

Figure 1: “The quantity of energy used by the GCP instance (on the y axis) versus the number of requests received by the instance in a 10 minute interval (on the x axis). It can be seen that even when zero requests are received by the instance in this time span (bottom left of the graph), the energy consumption remains at approximately 0.28 kWh.” (Figure and text by [2] p. 6)

Although the authors of the paper that measured BLOOM’s electricity consumption state that “since inference requests are unpredictable, optimization of GPU memory using techniques such as batching and padding […] is not possible” ([2] p. 5), we might expect that ChatGPT experiences a constant rate of requests high enough to make batching feasible. I would guess that batching of queries would lead to a lower energy consumption per query.

One thing that could increase ChatGPT’s energy consumption per query is if ChatGPT on average generates more words per query, but we don’t have access to any information that let’s us approximate this.

To err on the side of caution, I’ll assume in the following that ChatGPT’s electricity consumption is at max the same as BLOOM’s, so I’ll make an estimate of ChatGPT’s electricity consumption for each of the following assumptions:

  1. ChatGPT’s electricity consumption per query is the same as BLOOM’s, i.e. 0.00396 KWh
  2. ChatGPT’s electricity consumption per query is 75 % of BLOOM’s, i.e. 0.00297 KWh
  3. ChatGPT’s electricity consumption per query is 50 % of BLOOM’s, i.e. 0.00198 KWh

Step 2: Number of queries sent to ChatGPT

Now that we have some estimates of ChatGPT’s electricity consumption per query, we need to estimate how many queries ChatGPT received in January.

It’s been estimated that ChatGPT had 590m visits in January 2023 [1]. How many queries did each visit result in? We don’t know, but let’s make some assumptions. I would expect that each visit results in at least 1 query, so that’s the minimum assumption. Then, I’ll make a scenario with 5 and 10 queries per visit. Annecdotally, I’ve heard several people report that they use ChatGPT extensively in their daily tasks, so I believe these numbers are not unreasonable.

  1. Queries per visit = 1, so total number of queries = 590m
  2. Queries per visit = 5, so total number of queries = 2.95b
  3. Queries per visit = 10, so total number of queries = 5.9b

# 3 : Combining electricity consumption estimates with number of queries

Above, I outlined three different assumptions regarding ChatGPT’s electricity use per query and three different assumptions regarding the total number of queries.

This gives us 9 unique set of assumptions that I will refer to as “scenarios”.

Table 1 below shows the estimated electricity consumption for each of the scenarios.

The table shows that ChatGPT’s electricity consumption in January 2023 is estimated to be between 1,168,200 KWh and 23,364,000 KWh.

Here’s an online spreadsheet if you want to check my calculations.

Table 1: ChatGPT’s estimated electricity consumption in January 2023 in different energy efficiency and total queries scenarios. Table by Kasper Groes Albin Ludvigsen, based on Luccioni et al, 2022 [2]

Let’s put the numbers from Table 1 into perspective.

The annual electricity consumption of the average Dane is 1,600 KWh [3].

If the annual consumption is spread out evenly over the year, the average Dane used about 133.33 KWh of electricity in January.

So when we divide ChatGPT’s estimated electricity consumption by 133.3, we can see ChatGPT’s estimated consumption correspond to the monthly electricity consumption of between 8,762 and 175,234 Danes.

To put it into another perspective, you can run a standard 7w light bulb for around 19,050 years with 1,168,200 KWh of electricity (1,168,200 KWh / (7 / 1000) / 24 / 365).

Training ChatGPT’s underlying language model, GPT-3, consumed an estimated 1,287 MWh [4].

So at the low end of the estimated range for ChatGPT’s electricity consumption in January 2023, training GPT-3 and running ChatGPT for one month took roughly the same amount of energy. At the high end of the range, running ChatGPT took 18 times the energy it took to train.

Although handling a single query requires miniscule resources, the total amount of resources (e.g. electricity) add up to noticeable amounts when a product like ChatGPT scales. The search engine Bing now has incorporated ChatGPT [5] which might significantly increase the number of queries ChatGPT should handle, thereby also increasing the electricity consumption. In my opinion, this requires a debate about whether the incorporation of large machine learning models into existing products elevate the user experience to a level that justifies the large energy consumption.

In addition, unless the estimates put forward here are way too high, this article shows that running a machine learning based product in production can easily be way more energy consuming than training the model — especially if the product has a long life time and/or a large number of users. This should be taken into consideration when debating the environmental impact of AI and when thinking of how to reduce the life cycle carbon footprint of machine learning models.

It is important for me to emphasize that the estimates put forward in this article come with great uncertainties as very little information about how ChatGPT is run is publicly available. In the best of all worlds, OpenAI and Microsoft — and other tech companies for that matter — would disclose the energy consumption and carbon footprint of their products. But while we wait for that to happen, I hope my estimates will spur a debate and inspire others to come up with better estimates.

FOLLOW US ON GOOGLE NEWS

Read original article here

Denial of responsibility! Techno Blender is an automatic aggregator of the all world’s media. In each content, the hyperlink to the primary source is specified. All trademarks belong to their rightful owners, all materials to their authors. If you are the owner of the content and do not want us to publish your materials, please contact us by email – [email protected]. The content will be deleted within 24 hours.

Leave a comment