Overview
Features
Local MCP server with no APIs
Runs locally with no external APIs exposed; designed for offline-style operation and easy local integration.
DuckDuckGo web search integration
Queries DuckDuckGo to fetch 10 search results for each query to drive retrieval.
Embedding-based ranking
Fetches embeddings from Google's MediaPipe Text Embedder and ranks results by similarity to the query.
Context extraction from URLs
Extracts contextual content from the top results and formats it as Markdown for the LLM.
MCP client compatibility
Designed to work with MCP clients that support tool calling; tested with Claude Desktop, Cursor, Goose.
Primitive RAG-like workflow
Implements a retrieval-augmented generation flow to provide fresh web context to the LLM.
Configurable via MCP settings
Configured through MCP server settings JSON using uvx or Docker-based commands.
Dockerized deployment
Provides a Docker image ghcr.io/nkapila6/mcp-local-rag:latest for straightforward deployment.
Who Is This For?
- LLM developers:Integrate a local, API-free MCP server to fetch recent web information for chat prompts.
- Researchers:Prototype RAG-like retrieval workflows on local infrastructure using MCP clients that support tool calls.




