-
-
Scout Monitoring
Free Django app performance insights with Scout Monitoring. Get Scout setup in minutes, and let us sweat the small stuff. A couple lines in settings.py is all you need to start monitoring your apps. Sign up for our free tier today.
I was curious so I tried to answer this question
---
Training Llama 3 models emitted 2290 tons CO2e (https://github.com/meta-llama/llama3/blob/main/MODEL_CARD.md), and took 7.7 million GPU hours. Those GPU hours are for H100s, which consume 700W. So the conversion is approximately 2290 / (7.7e6 * 3600 * 700 / 1e9) ~= 0.12 tons CO2e per GPU-gigajoule.
A100s (what Huggingface offers) consume 400W (https://www.nvidia.com/content/dam/en-zz/Solutions/Data-Cent...) and cost $2.21/hour (in e.g. CoreWeave https://www.coreweave.com/gpu-cloud-pricing). So $10 million in H100s buys you ($10e6 / $2.21/h * 3600s/h) * 400W ~= 6515 Gigajoules in GPU-hours.
So Huggingface's offering will emit ~781 tons CO2e. Less if they've inflated the value of the compute they provide, which they have an incentive to do, but let's round to 800 tons.
---
According to https://www.carbonindependent.org/22.html, one Boeing-737-400 flying for 926km emits (3.61 tons fuel/flight * 3.15(g CO2e / g fuel)) = 11.37 tons CO2e .
So $10million in compute is like ~72 Boeing-737-400 international flights.