home / mcp / berryrag mcp server
Local vector DB with Playwright MCP integration to scrape, embed, and query Claude-facing content.
Configuration
View docs{
"mcpServers": {
"berrydev-ai-berry-rag": {
"url": "https://mcp.berrydev.ai/berry-rag"
}
}
}You set up BerryRAG as an MCP-enabled local knowledge base that combines a vector database with Playwright-based web scraping. This lets Claude access up-to-date, scraped content processed into embeddings, so you can search, retrieve context, and get precise answers from your own data.
Configure Claude Desktop to connect to the two MCP servers used by BerryRAG. This enables you to drive Playwright scraping through Claude and to operate the local vector database directly from Claudeβs workflow.
"mcpServers": {
"playwright": {
"command": "npx",
"args": ["@playwright/mcp@latest"]
},
"berry-rag": {
"command": "node",
"args": ["mcp_servers/vector_db_server.js"],
"cwd": "/Users/eberry/BerryDev/berry-rag"
}
}1) Scrape target content via Playwright MCP through Claude using your configured servers. 2) Process scraped content into the local vector database. 3) Query and search your knowledge base to retrieve context and answers.
npm run process-scraped
npm run search "React hooks"Adds content directly to the local vector DB so it can be searched and contextualized.
Find similar content in the vector DB to support answering questions.
Return Claude-formatted context for a given query, suitable for feeding into a chat or assistant.
List all stored documents in the vector DB to inspect your knowledge base.
Provide vector database statistics for monitoring health and size.
Process files scraped by Playwright MCP and embed them into the vector DB.
Save scraped material for later processing or auditing.
(BerryExa Server) Extract and crawl web content with subpage support.
(BerryExa Server) Retrieve internal links to enable subpage discovery.
(BerryExa Server) Preview content without full processing to speed up workflows.