AI has successfully shifted me from "plan, design, build" to "I can ship code" mindset.
I can deploy a state-of-the-art language model in minutes.
I can fine-tune BERT variants before my coffee gets cold.
I can build and ship RAG systems faster than ever before..*
But then I read something that stopped me cold.
Training a single AI model can emit more CO2 than five cars will produce in their entire lifetimes. And I've deployed dozens of models this year alone.
This led me back to my IKIGAI: What should I build with AI?
β‘ Be part of building solutions to all or some of humanity's problems
If my AI solutions contribute to climate change, am I really solving problems?
This repository tracks the environmental and qualitative responsibility of AI models. We believe AI should:
- β Help the economy, not just optimize production
- β Be selective β can't be the solution for everything & the value of nothing
- β Reduce cost of living (e.g., cost-saving drugs, affordable groceries)
- β Solve humanity's problems β be part of building sustainable AI solutions
This repository provides frameworks and tools to evaluate AI models across two dimensions:
- CO2 emissions from model training
- Hardware efficiency metrics
- Energy consumption analysis
- Visualization dashboards
- Accuracy metrics
- Precision & Recall
- F1-Score analysis
- Task-specific benchmarks
- Combined environmental + performance metrics
- Model selection guide for responsible AI
- Trade-off analysis (efficiency vs. accuracy)
python >= 3.8
pip install -r requirements.txtcd notebooks
jupyter notebook co2_emissions_tracker.ipynbGenerated files are saved to outputs/:
co2_emissions_latest.csvβ Clean dataset with CO2 metricsco2_emissions_bar_chart.pngβ Top models by emissionsco2_emissions_histogram.pngβ Distribution analysisco2_emissions_categories.pngβ Emission categories
responsible-ai/
βββ README.md # This file
βββ LICENSE # MIT License
βββ CONTRIBUTING.md # How to contribute
βββ requirements.txt # Python dependencies
β
βββ notebooks/
β βββ README_CO2_EMISSIONS.md # Detailed notebook guide
β βββ co2_emissions_tracker.ipynb # Main analysis notebook
β
βββ src/
β βββ __init__.py
β βββ data_collection.py # HF Hub data fetching
β βββ co2_estimator.py # CO2 calculation engine
β βββ visualization.py # Chart generation utilities
β
βββ docs/
β βββ METHODOLOGY.md # CO2 estimation formulas & references
β βββ ROADMAP.md # Phase-by-phase development plan
β βββ FUTURE_ENHANCEMENTS.md # GitHub Actions automation, etc.
β
βββ data/
β βββ sample/ # Sample datasets
β βββ .gitkeep
β
βββ outputs/
β βββ [Generated by notebooks]
β
βββ tests/
βββ [Test files]
| Document | Purpose |
|---|---|
| notebooks/README_CO2_EMISSIONS.md | Step-by-step guide for CO2 analysis notebook |
| docs/METHODOLOGY.md | CO2 estimation formulas, hardware assumptions, references |
| docs/ROADMAP.md | Three-phase development plan |
| CONTRIBUTING.md | How to extend with qualitative metrics |
Based on Hugging Face Hub analysis (10K+ models scanned):
- Model Coverage: ~2-5% of models report CO2 emissions
- Emission Range: From <1 kg (small models) to >100K kg (large LLMs)
- Most Emitting Models: Large language models (70B+)
- Green Alternatives: Distilled models show 40-60% lower emissions
- Data Collection: Hugging Face Hub API
- Analysis: pandas, numpy
- Visualization: matplotlib, seaborn
- Future: scikit-learn (metrics), GitHub Actions (automation)
- Read docs/METHODOLOGY.md to understand CO2 calculation
- Run
notebooks/co2_emissions_tracker.ipynbto collect fresh data - Generate visualizations for your papers/presentations
- Check outputs/ for latest analysis
- Use
data/sample/to find green models - Compare models using the responsibility framework (Phase 2+)
- Read CONTRIBUTING.md
- Start with Phase 2: Adding qualitative metrics
- Follow the modular structure in
src/
If you use this repository in research, please cite:
@software{responsible_ai_2025,
title={Responsible AI: Environmental & Qualitative Model Tracking},
author={Your Name},
year={2025},
url={https://github.com/yourusername/responsible-ai}
}Key Papers on AI CO2 Emissions:
- Strubell et al. (2019): Energy and Policy Considerations for Deep Learning in NLP
- Patterson et al. (2021): Carbon Emissions and Large Neural Network Training
- https://mlco2.github.io/impact/
https://huggingface.co/blog/carbon-emissions-on-the-hub
Data Sources:
- Hugging Face Hub β 10K+ models with metadata
MIT License β See LICENSE file
We welcome contributions! See CONTRIBUTING.md for guidelines.
- π§ Open an Issue
- π‘ Start a Discussion
- π Submit a Pull Request
Last Updated: January 2025
Status: Active Development (Phase 1 β
, Phase 2 π§, Phase 3 π)