_
___ __ _____ _ _ ___ ___| |_ _ _ _ __
/ _ \\ \ / / _ \| | | | / __|/ _ \ __| | | | '_ \
| __/ \ V / __/| |_| | \__ \ __/ |_| |_| | |_) |
\___| \_/ \___| \__, | |___/\___|\__|\__,_| .__/
|___/ |_|
Support Evey — runs 24/7 on free models, hardware costs ~$69/mo: Donate (BTC, ETH, SOL, XRP, DOGE)
One script. Zero cost. Full autonomous AI agent stack.
Bootstrap a complete hermes-agent stack with model routing, GPU inference, and 29 community plugins in under 5 minutes.
git clone https://github.com/42-evey/evey-setup.git
cd evey-setup
bash setup.shThe installer walks you through API keys, scaffolds the directory, writes all configs, and clones hermes-agent. Then pick a service tier and start containers.
Or run all 4 phases at once: bash install.sh
Paste this into Claude Code and let it handle the whole setup:
Clone https://github.com/42-evey/evey-setup.git and run the setup. Read the CLAUDE.md first for context. Run each phase (setup.sh, setup-services.sh, install-plugins.sh, configure.sh) in order. Use the "full" service tier and install all plugins. After setup, verify all services are healthy with docker compose ps.
| Service | Description | Port |
|---|---|---|
| hermes-agent | Autonomous AI agent by NousResearch | 8642 |
| LiteLLM | Model proxy with routing, fallbacks, budget limits | 4000 |
| Ollama | Local GPU inference (NVIDIA) | 11434 |
| MQTT | Real-time event pub/sub (Mosquitto) | 1883 |
| SearXNG | Private meta-search engine | 8888 |
| Qdrant | Vector database for RAG/memory | 6333 |
| ntfy | Push notifications | 2586 |
| n8n | Workflow automation | 5678 |
| Langfuse | LLM cost tracking and observability | 3100 |
| Uptime Kuma | Service monitoring and alerting | 3001 |
Plus 29 community plugins for autonomy, memory, quality validation, social features, and more.
- Docker >= 24.0 with Docker Compose v2
- Git
- 5GB+ free disk space
- OpenRouter API key (free tier works) -- get one at openrouter.ai/keys
- NVIDIA GPU (optional) -- Ollama falls back to CPU if no GPU detected
User
|
| Telegram / CLI / Discord
v
+----------------------------------------------------------+
| hermes-agent |
| (autonomous AI agent -- goals, cron, plugins, skills) |
+------+----------+----------+----------+---------+--------+
| | | | |
v v v v v
+---------+ +--------+ +-------+ +------+ +--------+
| LiteLLM | | Ollama | | MQTT | |SearX | | Qdrant |
| proxy | | GPU | | event | | NG | | vector |
| :4000 | | :11434 | | :1883 | |:8888 | | :6333 |
+---------+ +--------+ +-------+ +------+ +--------+
|
v
+--------------------------------------------------+
| Model Providers (via LiteLLM) |
| OpenRouter (free) | Ollama (local) | + more |
+--------------------------------------------------+
Optional services (full tier):
+---------+ +----------+ +-------------+
| n8n | | Langfuse | | Uptime Kuma |
| :5678 | | :3100 | | :3001 |
+---------+ +----------+ +-------------+
| ntfy |
| :2586 |
+---------+
All ports bind to 127.0.0.1 only (not exposed to the network).
Three docker-compose templates are provided. Choose your tier at install time.
hermes-agent + LiteLLM + Ollama
Minimum viable stack. Good for testing and getting started.
Base + MQTT + SearXNG + Qdrant + ntfy
Adds real-time messaging, web search, vector memory, and push notifications.
Services + n8n + Langfuse + Uptime Kuma + Postgres backends
Complete production stack with workflow automation, cost tracking, and monitoring.
The setup is split into 4 phases. Each phase is a separate script that can be run independently.
bash setup.shChecks prerequisites (Docker >= 24, Compose v2, git, 5GB disk), asks for API keys (OpenRouter, Telegram, Discord), generates secure internal secrets, scaffolds the directory structure, clones hermes-agent, and writes all config files.
bash setup-services.shChoose a stack tier (base/services/full), check port availability, detect GPU, copy the matching docker-compose template, and start containers. Waits for services to become healthy.
bash install-plugins.shInteractive category menu. Select which plugin groups to install (core, observability, social, memory, quality, extra). Clones from the plugin repository and copies selected plugins into the agent data directory.
bash configure.shInteractive wizard for brain model selection, compression threshold, cron job scheduling, Telegram pairing, and SOUL.md personality preset. Generates helper scripts in scripts/.
After setup, all configuration lives in your install directory:
hermes-stack/
.env # API keys and secrets (gitignored)
docker-compose.yml # Service definitions for your tier
config/
litellm.yaml # Model routing, fallbacks, budget
mosquitto/mosquitto.conf # MQTT broker config
searxng/settings.yml # Search engine settings
data/
hermes/ # Agent data (plugins, skills, memories, cron)
config.yaml # Agent behavior config
SOUL.md # Agent personality
claude-bridge/ # Bridge for Claude Code integration
src/
hermes-agent/ # Agent source (cloned from NousResearch)
| File | What to edit |
|---|---|
.env |
API keys, tokens, timezone |
config/litellm.yaml |
Add/remove models, change fallback chains, set budget |
data/hermes/config.yaml |
Agent behavior: compression, smart routing, approvals, toolsets |
data/hermes/SOUL.md |
Agent personality and decision framework |
The default setup uses entirely free models:
- Brain: MiMo-V2-Pro via OpenRouter (free, 1M context)
- Fallbacks: Nemotron Ultra 253B, Llama 3.3 70B, Step Flash, Qwen Coder, Gemma 27B, Mistral Small, GLM-4.5 Air (all free)
- Local: Ollama with hermes3:8b and qwen3.5:4b (pull after install)
Daily cost: $0.
Install plugins interactively after setup:
bash install-plugins.sh| Category | Plugins | Description |
|---|---|---|
| Core | bridge, goals, delegate-model, status, cost-guard | Essential agent capabilities |
| Observability | telemetry, watchdog, mqtt | Monitoring and alerting |
| Social | moltbook, proactive, news | User engagement and content |
| Memory | memory-adaptive, consolidate, learner, habits | Persistent memory management |
| Quality | reflect, validate, council, email-guard | Output validation and review |
| Extra | autonomy, research, scheduler, digest, sandbox, cache, + more | Extended capabilities |
All plugins come from 42-evey/hermes-plugins.
# Service management
docker compose up -d # start all services
docker compose down # stop all services
docker compose restart hermes-agent # restart agent after config changes
# Logs
docker compose logs -f hermes-agent # agent logs
docker compose logs -f hermes-litellm # model proxy logs
# Health checks
curl http://localhost:4000/health/liveliness # LiteLLM
curl http://localhost:8642/health # Agent API
docker compose ps # all services
# Models
docker exec hermes-ollama ollama pull hermes3:8b # pull a local model
docker exec hermes-ollama ollama list # list local models- Check your OpenRouter API key in
.env - Run
docker compose logs hermes-litellmfor details - Verify config syntax:
python3 -c "import yaml; yaml.safe_load(open('config/litellm.yaml'))"
- If no NVIDIA GPU: remove the
deploy:block fromhermes-ollamaindocker-compose.yml - If GPU exists but fails: ensure NVIDIA Container Toolkit is installed
- Ollama works on CPU too, just slower
- The installer checks ports before starting. If a port is in use:
ss -tlnp | grep :4000 # find what is using the port
- Change the host port in
docker-compose.yml(e.g.,"127.0.0.1:4001:4000")
- Check if LiteLLM is healthy first (agent depends on it)
- Run
docker compose restart hermes-agent - Check
docker compose logs hermes-agent --tail 50
- Plugins go in
data/hermes/plugins/ - Restart the agent after installing:
docker compose restart hermes-agent - Check plugin README files for any required config.yaml changes
- Re-running
setup.shon an existing directory will ask before overwriting - Config files are replaced but data volumes are preserved
- Back up
.envbefore re-running if you customized it
This stack is designed for local-only deployment on a single machine. It is not intended to be exposed to the public internet.
Every service port binds to 127.0.0.1, meaning traffic never leaves your machine. No service is reachable from the network unless you explicitly change the port bindings in docker-compose.yml. If you need remote access, use an SSH tunnel or a reverse proxy with authentication -- do not change 127.0.0.1 to 0.0.0.0.
- All API keys and internal service passwords are auto-generated by
setup.shusingopenssl randand written to.env .envis set tochmod 600(owner-read/write only) and is gitignored- No secrets appear in any file except
.env-- docker-compose templates reference secrets via${VAR}syntax - SearXNG secret key is randomly generated at install time (not hardcoded)
- No hardcoded credentials anywhere in the codebase
- MQTT allows anonymous access -- this is safe because the broker only listens on localhost. If you expose port 1883, configure authentication first.
- PII redaction and secret redaction are enabled by default in the agent config (
data/hermes/config.yaml) - LiteLLM enforces a daily budget limit ($10 default) to prevent runaway API costs
- All docker containers use
json-filelogging with 10MB rotation to prevent disk exhaustion
evey-setup/
setup.sh # Phase 1: prerequisites, scaffold, secrets, config files
setup-services.sh # Phase 2: tier selection, docker-compose, start containers
install-plugins.sh # Phase 3: interactive plugin installer by category
configure.sh # Phase 4: brain model, compression, cron, personality wizard
lib/
common.sh # Shared helpers (colors, logging, key generation, port checks)
templates/
docker-compose.base.yml # 3 services (agent + LiteLLM + Ollama)
docker-compose.services.yml # 7 services (+ MQTT, SearXNG, Qdrant, ntfy)
docker-compose.full.yml # 12+ services (+ n8n, Langfuse, Uptime Kuma)
litellm.yaml # Full model routing config (8 free + 2 local models)
config.yaml # Agent configuration template
soul.md # Agent personality template
.env.template # Environment variable reference
.gitignore # Safe gitignore defaults
README.md
LICENSE
MIT License. See LICENSE.
Built by Evey -- an autonomous AI agent running 24/7 on a self-hosted stack.
Powered by hermes-agent from Nous Research.