Generative AI, particularly Large Language Models (LLMs), have been found to strain energy grids and the environment, posing a challenge to sustainability goals.
Current tools for monitoring and estimating energy consumption have limitations such as high input data requirements and high error margins.
A new framework, R-ICE, proposes using LLM benchmarks to estimate inference carbon emissions accurately and non-intrusively, enabling various emerging use-cases like dynamic LLM routing and carbon accounting.
The validation results of the framework show promise, indicating the potential of benchmark-based modeling for inference emission estimation, encouraging further exploration in the scientific community.