I build production-grade AI systems — from RAG pipelines and multi-agent architectures to event-driven microservices at scale. Comfortable taking a model from prototype to a containerised, observable service running in Kubernetes.
A multi-tenant-ecommerce-RAG system which is production-ready RAG (Retrieval-Augmented Generation) system serving three e-commerce platforms (Amazon, Flipkart, Myntra) from a single deployment. Implemented tenant-aware vector search using Qdrant with automatic context switching based on API keys, achieving 40% reduction in support response time. The system features real-time document ingestion via Kafka streaming, PostgreSQL for metadata management, and Redis caching, all containerized with Docker for seamless scalability.
Production-grade multi-agent code review system built on LangGraph. Submitted code is analyzed in parallel by specialized agents — bug detection, quality, and security — then synthesized into a structured report with scoring. SHA256 caching eliminates duplicate LLM calls. Fully containerised with FastAPI backend and PostgreSQL persistence.
Production-ready FastAPI service template with structured logging, Prometheus metrics, health/readiness probes, and a multi-stage Dockerfile. Demonstrates how to run a Python ASGI app correctly in Kubernetes — resource limits, liveness probes, config from environment, secrets from K8s Secrets.
Open to backend AI engineering roles, contract work, and interesting problems in production AI systems.