Deploy Self-Hosted LLM Observability with Torrix
Single-container LLM logging and monitoring without Postgres or Redis
About this automation
Torrix is a self-hosted LLM observability platform that eliminates infrastructure friction by running as a single Docker container backed by SQLite. It captures tokens, cost, latency, full prompt/response traces, and reasoning tokens from OpenAI, Anthropic, Gemini, Groq, Mistral, Azure OpenAI, and compatible endpoints. Includes cost forecasting, budget caps, PII masking, model routing, evals with golden runs, prompt library with version control, MCP server integration, and OTLP/HTTP ingestion.
How to implement
Download docker-compose.yml from Torrix GitHub repository
Run 'docker compose up' to start the single-container deployment
Configure HTTP proxy or integrate Python/Node SDK for LLM call logging
Access observability dashboard to monitor tokens, costs, latency, and traces
Set up cost forecasting, budget caps, and PII masking rules as needed
Optional: Enable MCP server for AI Assistant log queries and OTLP ingestion