EnviroLLM gives you the tools to track and optimize resource usage when running models on your own hardware.
Run one command (no installation needed):
npx envirollm startThen visit the dashboard to see your metrics in real-time!
Requirements: Node.js and Python 3.7+
The CLI automatically finds most popular LLM setups:
Absolutely! Everything's available on GitHub.
Via Web Interface:
npx envirollm startVia CLI:
npx envirollm benchmark --models llama3:8b,phi3:miniTask presets are pre-written prompts designed to test different workload types. We provide 7 presets:
These help you test how models perform across different use cases and enable reproducible comparisons.
All benchmark results are stored locally on your machine at:
~/.envirollm/benchmarks.dbYour data never leaves your machine. You can:
npx envirollm cleanEnviroLLM uses a simplified power estimation model:
This provides relative comparisons between models rather than absolute values. The measurements are consistent enough to identify which models are more efficient and track trends over time.
For research-grade accuracy, consider using specialized hardware power meters.
LLM-as-a-Judge uses another LLM (by default, gemma3:1b running locally) to evaluate response quality on a 0-100 scale.
This helps you assess quality-efficiency tradeoffs: a faster, more energy-efficient model might produce lower-quality responses.
When it's used:
Yes! This is one of the most useful features. Compare Q4, Q8, and FP16 quantizations of the same model:
ollama pull llama3:8bollama pull llama3:8b-q8ollama pull llama3:8b-fp16Research shows quantization can reduce energy by up to 45% while maintaining acceptable quality for many tasks.
LLMs are a fascinating technology to me, but running them locally can be a black box. I wanted to create a tool that gives users visibility and control over the environmental impact of their AI experiments. Since I'm not able to impact cloud-based inference, I thought this would be a good way to contribute to more sustainable AI practices.