Back to All Projects
DevOps AI Assistant - Local LLM with RAG
A production-ready AI assistant powered by local LLMs (Ollama) with Retrieval-Augmented Generation (RAG) for DevOps documentation. Query Kubernetes, Terraform, Docker, Ansible, and other DevOps tools using natural language.
Architecture
High-level overview of the RAG architecture.
Project Overview
This project provides a production-ready AI assistant that uses local LLMs (Ollama) and Retrieval-Augmented Generation (RAG) for DevOps documentation. It allows users to query various DevOps tools like Kubernetes, Terraform, Docker, and Ansible using natural language.
Key Features
- **Local LLM Inference**: Utilizes Ollama and supports models such as Llama 3.1, Mistral, and CodeLlama.
- **RAG Pipeline**: Employs Qdrant for vector search to provide accurate and context-aware responses.
- **DevOps Documentation**: Pre-configured to ingest documentation for Kubernetes, Terraform, Docker, Ansible, AWS, and more.
- **Web UI**: Features a clean and responsive chat interface.
- **REST API**: A FastAPI backend enables integration with other tools.
- **GPU Acceleration**: Optimized for NVIDIA GPUs.
Technology Stack
Python
FastAPI
Ollama
Qdrant
Docker
React