A plug-and-play Retrieval-Augmented Generation (RAG) framework built in Go, enabling seamless switching between vector databases and LLMs, with support for text and PDF-based context ingestion. Designed for developers and researchers building intelligent systems without handling backend orchestration manually.
- Language: Go (Golang)
- Vector Databases: Qdrant (REST API), Weaviate (GraphQL)
- LLMs: OpenAI GPT-4, Mistral 7B (API-based)
- Embeddings: External Hugging Face model (via
ghcr.io) - Tools: Docker (containerization), CLI-based interface
- π Modular CLI system with plug-and-play components
- π Ingest context via raw text or PDF files
- π§ Choose between Qdrant and Weaviate for vector storage
- π£ Use OpenAI or Mistral as the LLM engine
- π Secure API key integration for LLM access
- π¦ Embeddings generated via external Hugging Face model pulled from
ghcr.io
β Enter plain text or upload a PDF
β Select between Qdrant or Weaviate
β Select between OpenAI or Mistral (provide API key)
β System retrieves relevant context and generates a response using the selected LLM
- π§ Transition from CLI tool to a full-fledged package
- π Integrate with Gofr framework (open-source collaboration) for routing, logging & config management
- π§ͺ Add unit tests, CI/CD, and packaging for public release