ai ml2024

AI Infrastructure Template

Self-hosted MLOps platform for rapid AI product development

Tech Stack

PythonFastAPIDockerPostgreSQLpgvectorRedisMinIOPrometheusGrafanaLokiAWS BedrockReactTypeScriptLangChainPrompt EngineeringSystematic EvalsAI Tracing

A production-ready, self-hosted MLOps platform designed to build any type of AI product with built-in best practices. Implements the three-phase AI Engineering maturity model: Development (prompts + evals), Production (traces + scorecards), and Evolution (backtests + leaderboards). Features a compositional prompt architecture with auto-versioning, systematic evaluation framework, response caching to optimize costs, and complete observability with Prometheus, Grafana, and Loki. Deployed on VPS with Coolify, supporting multiple LLM providers (AWS Bedrock, OpenAI, Anthropic). Demonstrates production-grade infrastructure patterns and DevOps maturity.

Screenshots

AI Infrastructure Template screenshot 1
1 / 8

Key Highlights

Prompt management system with auto-versioning

Systematic evaluation framework with multiple matcher types

Multi-provider LLM support (Bedrock, OpenAI, Anthropic)

Complete observability stack (Prometheus, Grafana, Loki)

Response caching and cost optimization

Production-grade infrastructure patterns

Self-hosted on VPS with Coolify

Metrics

10

Microservices

Production

Status