Give your AI the ability to understand meaning, not just match keywords.
Context Lens transforms any content into a searchable knowledge base for your AI assistant. This self-contained Model Context Protocol (MCP) server with built-in serverless vector storage (LanceDB) brings semantic search to your conversations. Point it at any content - codebases, documentation, contracts, or text files - and your AI can instantly understand and answer questions about the content.
Traditional keyword search finds files containing specific words. Miss the exact term? Miss the content.
Context Lens understands meaning. Ask about "authentication" and find code about login, credentials, tokens, OAuth, and access control - even if those files never use the word "authentication."
Want to understand how Context-Lens works? Here's the fun part: you can use Context-Lens to learn about Context-Lens.
Demo: Using Claude Desktop with Context-Lens to index and query this repository itself. No git clone, no scrolling through code - just questions and answers.
Context Lens uses LanceDB - a modern, serverless vector database:
- π Completely Free & Local - No cloud services, API keys, or subscriptions
- β‘ Zero Infrastructure - Embedded database, just a file on disk
- π Fast & Efficient - Built on Apache Arrow, optimized for vector search
- πΎ Simple Storage - Single file database, easy to backup or move
Think of it as "SQLite for AI embeddings" - all the power of vector search without the complexity.
- π Semantic Search - Understand meaning, not just keywords
- π Zero Setup - No installation, no configuration, no API keys
- πΎ Serverless Storage - Built-in LanceDB, no external database
- π 100% Local & Private - All data stays on your machine
- π Local & GitHub - Index local files or public GitHub repositories
- π― Smart Parsing - Language-aware chunking for better results
When you add content to Context Lens, it doesn't just dump text into a database. Here's what actually happens:
Smart Reading: Context Lens detects your file type and uses specialized parsers. Python files are analyzed with AST parsing, JSON is parsed structurally, Markdown is split by headers. This preserves the natural structure of your content.
Meaningful Chunks: Instead of arbitrary character limits, content is chunked intelligently - complete functions, logical paragraphs, full sections. Your code never gets split mid-function.
Semantic Vectors: Each chunk is converted to a 384-dimensional vector using a local embedding model. These vectors capture meaning, not just words. "authentication" and "login system" become similar vectors even though they share no words.
Local Storage: Everything goes into LanceDB - a serverless vector database that's just a file on your disk. No cloud services, no API calls, completely private.
Conceptual Search: When you ask a question, it becomes a vector too. Context Lens finds chunks with similar vectors (similar meaning) and ranks them by relevance. You get answers based on concepts, not keyword matching.
| Component | Details |
|---|---|
| Embedding Model | sentence-transformers/all-MiniLM-L6-v2 |
| Vector Dimensions | 384 dimensions |
| Model Size | ~90MB (downloads on first use) |
| Chunk Size | 1000 characters (default, configurable) |
| Chunk Overlap | 200 characters (default, configurable) |
| Vector Database | LanceDB (serverless, file-based) |
| Storage Format | Apache Arrow columnar format |
| Search Method | Cosine similarity |
| Processing | 100% local, no external API calls |
π Want to customize? See SETUP.md for configuration options and TECHNICAL.md for performance benchmarks.
Add to .kiro/settings/mcp.json:
{
"mcpServers": {
"context-lens": {
"command": "uvx",
"args": ["context-lens"],
"autoApprove": ["list_documents", "search_documents"]
}
}
}Reload: Command Palette β "MCP: Reload Servers"
Add to .cursor/mcp.json:
{
"mcpServers": {
"context-lens": {
"command": "uvx",
"args": ["context-lens"]
}
}
}For Claude Desktop, Continue.dev, or any MCP-compatible client:
{
"mcpServers": {
"context-lens": {
"command": "uvx",
"args": ["context-lens"]
}
}
}π Need detailed setup instructions? See SETUP.md for all clients, programmatic usage, and configuration options.
Context Lens is published to the official Model Context Protocol Registry as io.github.cornelcroi/context-lens.
π Registry details and verification: See REGISTRY.md for installation verification and registry information.
Use Context Lens directly in your Python applications:
#!/usr/bin/env python3
import os
from dotenv import load_dotenv
from mcp import StdioServerParameters, stdio_client
from strands import Agent
from strands.models.openai import OpenAIModel
from strands.tools.mcp import MCPClient
def main():
# Load environment variables from .env file
load_dotenv()
# Create MCP client for context-lens server
mcp_client = MCPClient(
lambda: stdio_client(
StdioServerParameters(command="uvx", args=["context-lens"])
)
)
# Create an agent with OpenAI model and MCP tools
model = OpenAIModel(model_id="gpt-4o-mini")
agent = Agent(model=model, tools=[mcp_client])
print("Chatbot started! Type 'quit' to exit.")
while True:
user_input = input("\nYou: ").strip()
if user_input.lower() in ['quit', 'exit', 'bye']:
print("Goodbye!")
break
if not user_input:
continue
try:
response = agent(user_input)
print(f"Bot: {response}")
except Exception as e:
print(f"Error: {e}")
if __name__ == "__main__":
main()Example conversation:
You: Add https://github.com/fastapi/fastapi to the knowledge base
Bot: β Added 247 files from repository with 1,543 chunks
You: How does FastAPI handle dependency injection?
Bot: FastAPI uses a sophisticated dependency injection system...
π Full examples: See SETUP.md for complete code and more frameworks.
You: Add all Python files from ./src to the knowledge base
LLM: β Added 23 Python files with 156 chunks
You: How does the authentication system work?
LLM: The authentication system uses JWT tokens. The login flow starts in
auth/login.py where credentials are validated, then a token is generated
with the user's ID and expiration time...
You: Show me all the database models
LLM: Based on the codebase, here are the main database models:
1. User model (models/user.py) - handles user accounts and authentication
2. Post model (models/post.py) - manages blog posts with relationships...
You: Add https://github.com/fastapi/fastapi to the knowledge base
LLM: β Added 247 files from repository with 1,543 chunks
You: How does FastAPI handle dependency injection?
LLM: FastAPI uses a sophisticated dependency injection system. The Depends()
function in fastapi/dependencies.py analyzes function signatures and
automatically resolves dependencies. Here's how it works...
You: Compare FastAPI's routing to Flask
LLM: FastAPI uses decorator-based routing similar to Flask, but with key
differences: 1) Type hints for automatic validation, 2) Async support
built-in, 3) Automatic OpenAPI documentation generation...
You: Add all contracts from ./legal/contracts
LLM: β Added 12 contract documents with 487 chunks
You: Find all termination clauses
LLM: I found termination clauses in 8 contracts. Here's a summary:
- Service Agreement: 30-day notice required for either party
- Vendor Contract: Immediate termination for breach of confidentiality...
You: What are the payment terms across all contracts?
LLM: Payment terms vary by contract type:
- Service contracts: Net 30 days from invoice date
- Vendor agreements: 50% upfront, 50% on delivery...
You: Add https://raw.githubusercontent.com/fastapi/fastapi/master/README.md
LLM: β Document 'README.md' added successfully with 42 chunks
You: What is FastAPI and what are its key features?
LLM: FastAPI is a modern, fast web framework for building APIs with Python 3.7+.
Key features include:
- Fast performance (on par with NodeJS and Go)
- Automatic API documentation
- Type hints for validation
- Async support built-in...
π More examples and query patterns: USAGE_GUIDE.md
Context Lens doesn't just split text blindly - it understands code structure and creates intelligent chunks that respect language boundaries.
The difference: Generic chunking splits code arbitrarily by character count, often breaking functions mid-way. Smart parsing understands your code's structure and creates complete, meaningful chunks.
- π Python (
.py,.pyw) - Functions, classes, imports - β‘ JavaScript/TypeScript (
.js,.jsx,.ts,.tsx,.mjs,.cjs) - Functions, classes, imports - π¦ JSON (
.json,.jsonc) - Top-level keys, nested objects - π YAML (
.yaml,.yml) - Top-level keys, lists, mappings - π Markdown (
.md,.markdown,.mdx) - Header hierarchy, code blocks - π¦ Rust (
.rs) - Structs, traits, impl blocks, functions - π Other Files (
.txt,.log,.cpp,.java, etc.) - Intelligent paragraph/sentence splitting
β
Complete Code Units - Never splits functions or classes mid-way
β
Preserved Context - Docstrings, comments, and structure stay intact
β
Better Search - Find complete, understandable code snippets
β
Automatic - No configuration needed, works based on file extension
π Want to see how it works? Check out PARSING_EXAMPLES.md for detailed examples.
Context Lens works with text-based files from multiple sources:
- π Local files & folders - Your projects, documentation, any text files
- π GitHub repositories - Public repos, specific branches, directories, or files
- π Direct file URLs - Any HTTP/HTTPS accessible file
- π Documents - Contracts, policies, research papers, technical docs
Supported file types: .py, .js, .ts, .java, .cpp, .go, .rs, .rb, .php, .json, .yaml, .md, .txt, .sh, and more (25+ extensions)
Maximum file size: 10 MB (configurable via MAX_FILE_SIZE_MB environment variable)
Examples:
./src/- Local directory/path/to/file.py- Single local filehttps://github.com/fastapi/fastapi- Entire repositoryhttps://github.com/django/django/tree/main/django/contrib/auth- Specific directoryhttps://example.com/config.yaml- Direct file URL/path/to/contracts/- Legal documents
π See more examples: USAGE_GUIDE.md
- π₯ add_document - Add files, folders, or GitHub URLs
- π search_documents - Semantic search across all content
- π list_documents - Browse indexed documents
- βΉοΈ get_document_info - Get metadata about a document
- ποΈ remove_document - Remove specific documents
- π§Ή clear_knowledge_base - Remove all documents
π See detailed examples: USAGE_GUIDE.md
How does this compare to GitHub's MCP server?
They serve different purposes and complement each other:
Context Lens is better for:
- π§ Semantic understanding - "Find authentication code" returns login, credentials, tokens, OAuth - even without exact keywords
- π Learning codebases - Ask "How does X work?" and get conceptually relevant results across the entire project
- π Pattern discovery - Find similar code patterns, error handling approaches, or architectural decisions
- πΎ Offline development - Once indexed, works without internet connection
- π Privacy - All processing happens locally, no data sent to external services
GitHub's MCP server is better for:
- π§ Repository management - Create issues, manage PRs, handle CI/CD operations
- π Real-time state - Always fetches the latest version from GitHub
- π GitHub-specific features - Integrates with GitHub's ecosystem (Actions, Projects, etc.)
Key difference: Context Lens clones once and indexes everything for fast semantic search (offline). GitHub MCP makes API calls per query for real-time access (online). Use Context Lens to understand code, GitHub MCP to manage repositories.
Why is the first run slow?
The embedding model (~100MB) downloads on first use. This only happens once.
Do I need an API key?
No! Context Lens runs completely locally. No API keys, no cloud services.
Where is my data stored?
Context-Lens stores data in platform-specific directories:
- macOS:
~/Library/Application Support/context-lens/ - Linux:
~/.local/share/context-lens/ - Windows:
%LOCALAPPDATA%\context-lens\
You can change the base directory by setting CONTEXT_LENS_HOME environment variable:
{
"mcpServers": {
"context-lens": {
"command": "uvx",
"args": ["context-lens"],
"env": {
"CONTEXT_LENS_HOME": "/path/to/your/data"
}
}
}
}Or override individual paths with LANCE_DB_PATH (database) and EMBEDDING_CACHE_DIR (models).
Can I use this with private code?
Yes! All processing happens locally. Nothing is sent to external services.
How much disk space does it use?
~100MB for the model + ~1KB per text chunk. A 10MB codebase uses ~5-10MB of database space.
π More questions: TROUBLESHOOTING.md
- π Setup Guide - Detailed setup for all clients, configuration options
- π Usage Guide - Examples, queries, and best practices
- π¨ Parsing Examples - How smart parsing works
- π§ Troubleshooting - Common issues and solutions
- βοΈ Technical Details - Architecture, stack, and performance
- π Registry Information - MCP Registry verification and installation
- π€ Contributing - How to contribute, roadmap
- π¦ Publishing Guide - MCP Registry publishing process (for maintainers)
Contributions are welcome! Please:
- Open an issue first to discuss your idea
- Get approval before starting work
- Submit a PR referencing the issue
See CONTRIBUTING.md for details.
MIT License - see LICENSE for details.
Star the repo if you find it useful! β

