Documentation
The MCP Documentation Server lets you manage and search your documents using AI.
It runs entirely on your local machine and creates a private, searchable knowledge base from your files.
You can add text, markdown, or PDF files and then ask questions about them through an MCP client like Claude Desktop and Cursor.
The server uses AI embeddings to understand the meaning of your text, which results in much smarter search results than simple keyword matching.
Features
- π Document Management – Add, list, retrieve, and delete documents with comprehensive metadata support
- π Semantic Search – AI-powered search using advanced embeddings for contextual results
- π File Upload Support – Direct processing of .txt, .md, and .pdf files through uploads folder
- π§© Smart Text Chunking – Automatic content splitting for improved search accuracy and relevance
- ποΈ Complete Document Deletion – Clean removal of documents and all associated chunks
- π Multilingual Support – High-quality embeddings that work across multiple languages
- πΎ Local Data Storage – All information stored locally in your home directory
- β‘ Zero Configuration Setup – Works immediately without database configuration
Use Cases
- API Documentation Management: Keep multiple API references searchable in one place. Instead of switching between browser tabs or PDF files, search across all your API docs to find specific endpoints, parameters, or implementation examples.
- Code Knowledge Base: Store coding standards, architectural decisions, and team guidelines. When onboarding new developers or making technical decisions, quickly find relevant precedents and established patterns.
- Research Paper Organization: Academic researchers and engineers can upload technical papers and search for specific methodologies, algorithms, or experimental results without manually reading through entire documents.
- Technical Specification Lookup: Maintain searchable access to hardware specifications, protocol definitions, and technical standards. Find specific requirements or compatibility information across multiple technical documents instantly.
How to Use It
1. Installation and Setup
npx @andrea9293/mcp-documentation-serverFor permanent installation:
npm install -g @andrea9293/mcp-documentation-server
mcp-documentation-server2. Add the server to your MCP client configuration file (such as Claude Desktop’s config):
{
"mcpServers": {
"documentation": {
"command": "npx",
"args": [
"-y",
"@andrea9293/mcp-documentation-server"
],
"env": {
"MCP_EMBEDDING_MODEL": "Xenova/all-MiniLM-L6-v2"
}
}
}
}3. Available Tools and Methods
Document Management Tools:
add_document– Creates new documents with title, content, and metadata- Parameters: title (string), content (string), metadata (object)
- Returns: Document ID and confirmation
list_documents– Retrieves all documents with their metadata- Returns: Array of document objects with IDs, titles, and metadata
get_document– Fetches complete document content by ID- Parameters: id (string)
- Returns: Full document with all chunks
delete_document– Removes document and all associated chunks- Parameters: id (string)
- Returns: Deletion confirmation
Search Tools:
search_documents– Performs semantic search within specific documents- Parameters: document_id (string), query (string), limit (number)
- Returns: Ranked search results with relevance scores
File Processing Tools:
get_uploads_path– Returns the uploads directory path- Returns: Full path to uploads folder (~/.mcp-documentation-server/uploads/)
list_uploads_files– Shows all files in uploads directory- Returns: Array of file names and types
process_uploads– Converts uploaded files into searchable documents- Processes: .txt, .md, and .pdf files
- Returns: Processing results and new document IDs
4. Configuration Options
Embedding Models: Set via MCP_EMBEDDING_MODEL environment variable:
Xenova/all-MiniLM-L6-v2(default) – Fast processing, good quality resultsXenova/paraphrase-multilingual-mpnet-base-v2– Best quality, supports multiple languages
Data Storage: All data is stored locally in your home directory:
~/.mcp-documentation-server/
βββ data/ # Document storage in JSON format
βββ uploads/ # Files awaiting processing5. Step-by-Step Usage Workflow
Add Documents Directly:
{
"tool": "add_document",
"arguments": {
"title": "React Hooks Guide",
"content": "React Hooks allow you to use state and other React features...",
"metadata": {
"category": "frontend",
"framework": "react",
"tags": ["hooks", "state-management"]
}
}
}Upload Files for Processing:
- Get uploads path:
get_uploads_path - Copy your .txt, .md, or .pdf files to the uploads directory
- Run
process_uploadsto convert files into searchable documents
Search Your Documents:
{
"tool": "search_documents",
"arguments": {
"document_id": "doc-abc123",
"query": "useState hook implementation",
"limit": 10
}
}FAQs
Q: Why does the server seem to hang the first time I run it?
A: The first time you start the server, it needs to download the AI embedding model. The recommended model is over 400MB, so it can take a few minutes depending on your internet connection. Just let it run in the background; it will become responsive once the download is complete.
Q: What happens if I change the embedding model after adding documents?
A: The embeddings created by different models are not compatible. If you switch models, your existing documents will no longer be found in searches. You must delete all documents and re-add them to be indexed with the new model.
Q: Does this server extract text from scanned PDFs or images?
A: No, it does not perform Optical Character Recognition (OCR). It can only extract text that is already digitally embedded in the PDF file.
Q: Is my data sent to any external services?
A: No. The entire server, including the documents and the AI models, runs on your local machine. Your data is stored in the ~/.mcp-documentation-server/ directory and is never sent over the internet.
Latest MCP Servers
Obsidian Web
Claude Peers
Memo
Featured MCP Servers
Claude Peers
Excalidraw
Claude Context Mode
FAQs
Q: What exactly is the Model Context Protocol (MCP)?
A: MCP is an open standard, like a common language, that lets AI applications (clients) and external data sources or tools (servers) talk to each other. It helps AI models get the context (data, instructions, tools) they need from outside systems to give more accurate and relevant responses. Think of it as a universal adapter for AI connections.
Q: How is MCP different from OpenAI's function calling or plugins?
A: While OpenAI's tools allow models to use specific external functions, MCP is a broader, open standard. It covers not just tool use, but also providing structured data (Resources) and instruction templates (Prompts) as context. Being an open standard means it's not tied to one company's models or platform. OpenAI has even started adopting MCP in its Agents SDK.
Q: Can I use MCP with frameworks like LangChain?
A: Yes, MCP is designed to complement frameworks like LangChain or LlamaIndex. Instead of relying solely on custom connectors within these frameworks, you can use MCP as a standardized bridge to connect to various tools and data sources. There's potential for interoperability, like converting MCP tools into LangChain tools.
Q: Why was MCP created? What problem does it solve?
A: It was created because large language models often lack real-time information and connecting them to external data/tools required custom, complex integrations for each pair. MCP solves this by providing a standard way to connect, reducing development time, complexity, and cost, and enabling better interoperability between different AI models and tools.
Q: Is MCP secure? What are the main risks?
A: Security is a major consideration. While MCP includes principles like user consent and control, risks exist. These include potential server compromises leading to token theft, indirect prompt injection attacks, excessive permissions, context data leakage, session hijacking, and vulnerabilities in server implementations. Implementing robust security measures like OAuth 2.1, TLS, strict permissions, and monitoring is crucial.
Q: Who is behind MCP?
A: MCP was initially developed and open-sourced by Anthropic. However, it's an open standard with active contributions from the community, including companies like Microsoft and VMware Tanzu who maintain official SDKs.



