Content
# Docfork MCP: 🌿 @latest Docs to 9000+ Code Libraries in a Single Tool Call
[](https://cursor.com/install-mcp?name=docfork&config=eyJjb21tYW5kIjoibnB4IC15IGRvY2ZvcmtAbGF0ZXN0In0%3D) [](https://docfork.com) [](https://smithery.ai/server/@docfork/mcp) [<img alt="Install in VS Code (npx)" src="https://img.shields.io/badge/VS_Code-VS_Code?style=flat-square&label=Install%20Docfork%20MCP&color=0098FF">](https://insiders.vscode.dev/redirect?url=vscode%3Amcp%2Finstall%3F%7B%22name%22%3A%22docfork%22%2C%22command%22%3A%22npx%22%2C%22args%22%3A%5B%22-y%22%2C%22docfork%40latest%22%5D%7D)
## ❌ The Problem: Expired Knowledge
- Out of date code examples & stale data from year-old model training
- Hallucinated syntax & APIs
- Old or mismatched versions
## ✅ The Solution: @latest docs at warp speed
- Always in sync with the latest version of docs
- Accurate descriptions and code examples
- Sub-second retrieval results (500ms @ p95) in your AI code editor
Docfork MCP pulls @latest documentation and code examples straight from the source - and adds them right into your context.
Just tell Cursor to **`use docfork`**:
```txt
Create a basic Next.js app with the App Router. use docfork
```
## 🛠️ Installation
### 📋 Requirements
- Node.js ≥ v18
- Cursor/Windsurf/Claude Desktop (any MCP client)
<details>
<summary><b>Installing via Smithery</b></summary>
### Installing via Smithery
To install Docfork MCP Server for any client automatically via [Smithery](https://smithery.ai/server/@docfork/mcp):
```bash
npx -y @smithery/cli@latest install @docfork/mcp --client <CLIENT_NAME> --key <YOUR_SMITHERY_KEY>
```
You can find your Smithery key in the [Smithery.ai webpage](https://smithery.ai/server/@docfork/mcp).
</details>
<details>
<summary><b>Install in Cursor</b></summary>
Go to: `Settings` -> `Cursor Settings` -> `Tools & Integrations` -> `Add a custom MCP server`
Pasting the following config into your Cursor `~/.cursor/mcp.json` file is the recommended approach. You can also install in a specific project by creating `.cursor/mcp.json` in your project folder. See [Cursor MCP docs](https://docs.cursor.com/context/model-context-protocol) for more info.
#### Cursor Remote Server Connection
[](https://cursor.com/en/install-mcp?name=docfork&config=JTdCJTIydXJsJTIyJTNBJTIyaHR0cHMlM0ElMkYlMkZtY3AuZG9jZm9yay5jb20lMkZtY3AlMjIlN0Q%3D)
```json
{
"mcpServers": {
"docfork": {
"url": "https://mcp.docfork.com/mcp"
}
}
}
```
#### Cursor Local Server Connection
[](https://cursor.com/install-mcp?name=docfork&config=eyJjb21tYW5kIjoibnB4IC15IGRvY2ZvcmtAbGF0ZXN0In0%3D)
```json
{
"mcpServers": {
"docfork": {
"command": "npx",
"args": ["-y", "docfork"]
}
}
}
```
</details>
<details>
<summary>Alternative: Use Bun</summary>
```json
{
"mcpServers": {
"docfork": {
"command": "bunx",
"args": ["-y", "docfork"]
}
}
}
```
</details>
<details>
<summary>Alternative: Use Deno</summary>
```json
{
"mcpServers": {
"docfork": {
"command": "deno",
"args": ["run", "--allow-env", "--allow-net", "npm:docfork"]
}
}
}
```
</details>
<details>
<summary><b>Install in Claude Code</b></summary>
### Install in Claude Code
Run this command. See [Claude Code MCP docs](https://docs.anthropic.com/en/docs/agents-and-tools/claude-code/tutorials#set-up-model-context-protocol-mcp) for more info.
#### Claude Code Remote Server Connection
```sh
claude mcp add --transport sse docfork https://mcp.docfork.com/sse
```
#### Claude Code Local Server Connection
```sh
claude mcp add docfork -- npx -y docfork
```
</details>
<details>
<summary><b>Install in Claude Desktop</b></summary>
### Install in Claude Desktop
Add this to your Claude Desktop `claude_desktop_config.json` file. See [Claude Desktop MCP docs](https://modelcontextprotocol.io/quickstart/user) for more info.
```json
{
"mcpServers": {
"docfork": {
"command": "npx",
"args": ["-y", "docfork"]
}
}
}
```
</details>
<details>
<summary><b>Install in Windsurf</b></summary>
### Install in Windsurf
Add this to your Windsurf MCP config. See [Windsurf MCP docs](https://docs.windsurf.com/windsurf/mcp) for more info.
#### Windsurf Remote Server Connection
```json
{
"mcpServers": {
"docfork": {
"serverUrl": "https://mcp.docfork.com/sse"
}
}
}
```
#### Windsurf Local Server Connection
```json
{
"mcpServers": {
"docfork": {
"command": "npx",
"args": ["-y", "docfork"]
}
}
}
```
</details>
<details>
<summary><b>Install in VS Code</b></summary>
### Install in VS Code
Add this to your VS Code MCP config. See [VS Code MCP docs](https://code.visualstudio.com/docs/copilot/chat/mcp-servers) for more info.
#### VS Code Remote Server Connection
```json
{
"mcpServers": {
"docfork": {
"type": "http",
"url": "https://mcp.docfork.com/mcp"
}
}
}
```
#### VS Code Local Server Connection
```json
{
"servers": {
"docfork": {
"type": "stdio",
"command": "npx",
"args": ["-y", "docfork"]
}
}
}
```
</details>
<details>
<summary><b>Install in Zed</b></summary>
### Install in Zed
One-click install:
→ Get the [Docfork Extension](https://zed.dev/extensions?query=Docfork&filter=context-servers)
Or Manual config (for power users):
```json
{
"context_servers": {
"docfork": {
"command": {
"path": "npx",
"args": ["-y", "docfork"]
},
"settings": {}
}
}
}
```
</details>
<details>
<summary><b>Install in BoltAI</b></summary>
### Install in BoltAI
Open the "Settings" page of the app, navigate to "Plugins," and enter the following JSON:
```json
{
"mcpServers": {
"docfork": {
"command": "npx",
"args": ["-y", "docfork"]
}
}
}
```
More info is available on [BoltAI's Documentation site](https://docs.boltai.com/docs/plugins/mcp-servers). For BoltAI on iOS, [see this guide](https://docs.boltai.com/docs/boltai-mobile/mcp-servers).
</details>
<details>
<summary><b>Using Docker</b></summary>
### Using Docker
If you prefer to run the MCP server in a Docker container:
1. **Build the Docker Image:**
First, create a `Dockerfile` in the project root (or anywhere you prefer):
<details>
<summary>Click to see Dockerfile content</summary>
```Dockerfile
FROM node:18-alpine
WORKDIR /app
# Install the latest version globally
RUN npm install -g docfork
# Expose default port if needed (optional, depends on MCP client interaction)
# EXPOSE 3000
# Default command to run the server
CMD ["docfork"]
```
</details>
Then, build the image using a tag (e.g., `docfork-mcp`). **Make sure Docker Desktop (or the Docker daemon) is running.** Run the following command in the same directory where you saved the `Dockerfile`:
```bash
docker build -t docfork .
```
2. **Configure Your MCP Client:**
Update your MCP client's configuration to use the Docker command.
_Example for a cline_mcp_settings.json:_
```json
{
"mcpServers": {
"docfork": {
"autoApprove": [],
"disabled": false,
"timeout": 60,
"command": "docker",
"args": ["run", "-i", "--rm", "docfork-mcp"],
"transportType": "stdio"
}
}
}
```
_Note: This is an example configuration. Please refer to the specific examples for your MCP client (like Cursor, VS Code, etc.) earlier in this README to adapt the structure (e.g., `mcpServers` vs `servers`). Also, ensure the image name in `args` matches the tag used during the `docker build` command._
</details>
<details>
<summary><b>Install in Windows</b></summary>
### Install in Windows
The configuration on Windows is slightly different compared to Linux or macOS (_`Cline` is used in the example_). The same principle applies to other editors; refer to the configuration of `command` and `args`.
```json
{
"mcpServers": {
"github.com/docfork/mcp": {
"command": "cmd",
"args": ["/c", "npx", "-y", "docfork@latest"],
"disabled": false,
"autoApprove": []
}
}
}
```
</details>
<details>
<summary><b>Install in Augment Code</b></summary>
### Install in Augment Code
To configure Docfork MCP in Augment Code, follow these steps:
1. Press Cmd/Ctrl Shift P or go to the hamburger menu in the Augment panel
2. Select Edit Settings
3. Under Advanced, click Edit in settings.json
4. Add the server configuration to the `mcpServers` array in the `augment.advanced` object
```json
"augment.advanced": {
"mcpServers": [
{
"name": "docfork",
"command": "npx",
"args": ["-y", "docfork"]
}
]
}
```
Once the MCP server is added, restart your editor. If you receive any errors, check the syntax to make sure closing brackets or commas are not missing.
</details>
<details>
<summary><b>Install in Roo Code</b></summary>
### Install in Roo Code
Add this to your Roo Code MCP configuration file. See [Roo Code MCP docs](https://docs.roocode.com/features/mcp/using-mcp-in-roo) for more info.
#### Roo Code Remote Server Connection
```json
{
"mcpServers": {
"docfork": {
"type": "streamable-http",
"url": "https://mcp.docfork.com/mcp"
}
}
}
```
#### Roo Code Local Server Connection
```json
{
"mcpServers": {
"docfork": {
"command": "npx",
"args": ["-y", "docfork"]
}
}
}
```
</details>
<details>
<summary><b>Install in Trae</b></summary>
Use the Add manually feature and fill in the JSON configuration information for that MCP server.
For more details, visit the [Trae documentation](https://docs.trae.ai/ide/model-context-protocol?_lang=en).
#### Trae Remote Server Connection
```json
{
"mcpServers": {
"docfork": {
"url": "https://mcp.docfork.com/mcp"
}
}
}
```
#### Trae Local Server Connection
```json
{
"mcpServers": {
"docfork": {
"command": "npx",
"args": ["-y", "docfork"]
}
}
}
```
</details>
<details>
<summary><b>Install in Visual Studio 2022</b></summary>
You can configure Docfork MCP in Visual Studio 2022 by following the [Visual Studio MCP Servers documentation](https://learn.microsoft.com/visualstudio/ide/mcp-servers?view=vs-2022).
Add this to your Visual Studio MCP config file (see the [Visual Studio docs](https://learn.microsoft.com/visualstudio/ide/mcp-servers?view=vs-2022) for details):
```json
{
"mcp": {
"servers": {
"docfork": {
"type": "http",
"url": "https://mcp.docfork.com/mcp"
}
}
}
}
```
Or, for a local server:
```json
{
"mcp": {
"servers": {
"docfork": {
"type": "stdio",
"command": "npx",
"args": ["-y", "docfork"]
}
}
}
}
```
For more information and troubleshooting, refer to the [Visual Studio MCP Servers documentation](https://learn.microsoft.com/visualstudio/ide/mcp-servers?view=vs-2022).
</details>
<details>
<summary><b>Install in Gemini CLI</b></summary>
See [Gemini CLI Configuration](https://github.com/google-gemini/gemini-cli/blob/main/docs/cli/configuration.md) for details.
1. Open the Gemini CLI settings file. The location is `~/.gemini/settings.json` (where `~` is your home directory).
2. Add the following to the `mcpServers` object in your `settings.json` file:
```json
{
"mcpServers": {
"docfork": {
"httpUrl": "https://mcp.docfork.com/mcp"
}
}
}
```
Or, for a local server:
```json
{
"mcpServers": {
"docfork": {
"command": "npx",
"args": ["-y", "docfork"]
}
}
}
```
If the `mcpServers` object does not exist, create it.
</details>
<details>
<summary><b>Install in Crush</b></summary>
Add this to your Crush configuration file. See [Crush MCP docs](https://github.com/charmbracelet/crush#mcps) for more info.
#### Crush Remote Server Connection (HTTP)
```json
{
"$schema": "https://charm.land/crush.json",
"mcp": {
"docfork": {
"type": "http",
"url": "https://mcp.docfork.com/mcp"
}
}
}
```
#### Crush Remote Server Connection (SSE)
```json
{
"$schema": "https://charm.land/crush.json",
"mcp": {
"docfork": {
"type": "sse",
"url": "https://mcp.docfork.com/sse"
}
}
}
```
#### Crush Local Server Connection
```json
{
"$schema": "https://charm.land/crush.json",
"mcp": {
"docfork": {
"type": "stdio",
"command": "npx",
"args": ["-y", "docfork"]
}
}
}
```
</details>
<details>
<summary>
<b>Install in Cline</b>
</summary>
You can easily install Docfork through the [Cline MCP Server Marketplace](https://cline.bot/mcp-marketplace) by following these instructions:
1. Open **Cline**.
2. Click the hamburger menu icon (☰) to enter the **MCP Servers** section.
3. Use the search bar within the **Marketplace** tab to find _Docfork_.
4. Click the **Install** button.
</details>
<details>
<summary><b>Install in Zencoder</b></summary>
To configure Docfork MCP in Zencoder, follow these steps:
1. Go to the Zencoder menu (...)
2. From the dropdown menu, select Agent tools
3. Click on the Add custom MCP
4. Add the name and server configuration from below, and make sure to hit the Install button
```json
{
"command": "npx",
"args": ["-y", "docfork@latest"]
}
```
Once the MCP server is added, you can easily continue using it.
</details>
<details>
<summary><b>Install in Amazon Q Developer CLI</b></summary>
Add this to your Amazon Q Developer CLI configuration file. See [Amazon Q Developer CLI docs](https://docs.aws.amazon.com/amazonq/latest/qdeveloper-ug/command-line-mcp-configuration.html) for more details.
```json
{
"mcpServers": {
"docfork": {
"command": "npx",
"args": ["-y", "docfork@latest"]
}
}
}
```
</details>
<details>
<summary><b>Install in Qodo Gen</b></summary>
See [Qodo Gen docs](https://docs.qodo.ai/qodo-documentation/qodo-gen/qodo-gen-chat/agentic-mode/agentic-tools-mcps) for more details.
1. Open Qodo Gen chat panel in VSCode or IntelliJ.
2. Click Connect more tools.
3. Click + Add new MCP.
4. Add the following configuration:
```json
{
"mcpServers": {
"docfork": {
"url": "https://mcp.docfork.com/mcp"
}
}
}
```
</details>
<details>
<summary><b>Install in JetBrains AI Assistant</b></summary>
See [JetBrains AI Assistant Documentation](https://www.jetbrains.com/help/ai-assistant/configure-an-mcp-server.html) for more details.
1. In JetBrains IDEs go to `Settings` -> `Tools` -> `AI Assistant` -> `Model Context Protocol (MCP)`
2. Click `+ Add`.
3. Click on `Command` in the top-left corner of the dialog and select the As JSON option from the list
4. Add this configuration and click `OK`
```json
{
"mcpServers": {
"docfork": {
"command": "npx",
"args": ["-y", "docfork"]
}
}
}
```
5. Click `Apply` to save changes.
6. The same way docfork could be added for JetBrains Junie in `Settings` -> `Tools` -> `Junie` -> `MCP Settings`
</details>
<details>
<summary><b>Install in Warp</b></summary>
See [Warp Model Context Protocol Documentation](https://docs.warp.dev/knowledge-and-collaboration/mcp#adding-an-mcp-server) for details.
1. Navigate `Settings` > `AI` > `Manage MCP servers`.
2. Add a new MCP server by clicking the `+ Add` button.
3. Paste the configuration given below:
```json
{
"Docfork": {
"command": "npx",
"args": ["-y", "docfork"],
"env": {},
"working_directory": null,
"start_on_launch": true
}
}
```
4. Click `Save` to apply the changes.
</details>
<details>
<summary><b>Install in Opencode</b></summary>
Add this to your Opencode configuration file. See [Opencode MCP docs](https://opencode.ai/docs/mcp-servers) docs for more info.
#### Opencode Remote Server Connection
```json
"mcp": {
"docfork": {
"type": "remote",
"url": "https://mcp.docfork.com/mcp",
"enabled": true
}
}
```
#### Opencode Local Server Connection
```json
{
"mcp": {
"docfork": {
"type": "local",
"command": ["npx", "-y", "docfork"],
"enabled": true
}
}
}
```
</details>
<details>
<summary><b>Install in Copilot Coding Agent</b></summary>
## Using Docfork with Copilot Coding Agent
Add the following configuration to the `mcp` section of your Copilot Coding Agent configuration file Repository->Settings->Copilot->Coding agent->MCP configuration:
```json
{
"mcpServers": {
"docfork": {
"type": "http",
"url": "https://mcp.docfork.com/mcp",
"tools": ["get-library-docs"]
}
}
}
```
For more information, see the [official GitHub documentation](https://docs.github.com/en/enterprise-cloud@latest/copilot/how-tos/agents/copilot-coding-agent/extending-copilot-coding-agent-with-mcp).
</details>
<details>
<summary><b>Install in Kiro</b></summary>
See [Kiro Model Context Protocol Documentation](https://kiro.dev/docs/mcp/configuration/) for details.
1. Navigate `Kiro` > `MCP Servers`
2. Add a new MCP server by clicking the `+ Add` button.
3. Paste the configuration given below:
```json
{
"mcpServers": {
"Docfork": {
"command": "npx",
"args": ["-y", "docfork"],
"env": {},
"disabled": false,
"autoApprove": []
}
}
}
```
4. Click `Save` to apply the changes.
</details>
<details>
<summary><b>Install in OpenAI Codex</b></summary>
See [OpenAI Codex](https://github.com/openai/codex) for more information.
Add the following configuration to your OpenAI Codex MCP server settings:
```toml
[mcp_servers.docfork]
args = ["-y", "docfork"]
command = "npx"
```
</details>
<details>
<summary><b>Install in LM Studio</b></summary>
See [LM Studio MCP Support](https://lmstudio.ai/blog/lmstudio-v0.3.17) for more information.
#### One-click install:
[](https://lmstudio.ai/install-mcp?name=docfork&config=eyJjb21tYW5kIjoibnB4IiwiYXJncyI6WyIteSIsImRvY2ZvcmsiXX0%3D)
#### Manual set-up:
1. Navigate to `Program` (right side) > `Install` > `Edit mcp.json`.
2. Paste the configuration given below:
```json
{
"mcpServers": {
"Docfork": {
"command": "npx",
"args": ["-y", "docfork"]
}
}
}
```
3. Click `Save` to apply the changes.
4. Toggle the MCP server on/off from the right hand side, under `Program`, or by clicking the plug icon at the bottom of the chat box.
</details>
<details>
<summary><b>Install in Perplexity Desktop</b></summary>
See [Local and Remote MCPs for Perplexity](https://www.perplexity.ai/help-center/en/articles/11502712-local-and-remote-mcps-for-perplexity) for more information.
1. Navigate `Perplexity` > `Settings`
2. Select `Connectors`.
3. Click `Add Connector`.
4. Select `Advanced`.
5. Enter Server Name: `Docfork`
6. Paste the following JSON in the text area:
```json
{
"args": ["-y", "docfork"],
"command": "npx",
"env": {}
}
```
7. Click `Save`.
</details>
## 🔨 Available Tools
Docfork MCP provides the following tool that LLMs can use:
- `get-library-docs`: Searches the library and returns its documentation.
- `libraryName` (required): The name of the library to search for
- `topic` (required): Focus the docs on a specific topic (e.g., "routing", "hooks")
- `tokens` (optional, default 10000, max 50000): Max number of tokens to return. Values less than the configured `DEFAULT_MINIMUM_TOKENS` value or the default value of 10000 are automatically increased to that value.
## 💡 Tips
### Add a Rule
If you don't want to add `use docfork` to every prompt, you can define a simple rule from your `Cursor Settings > Rules` section in Cursor (or the equivalent in your MCP client) to auto-invoke Docfork on any code question:
```markdown
---
alwaysApply: true
---
when the user requests code examples, setup or configuration steps, or library/API documentation
use docfork.
```
From then on you'll get Docfork's docs in any related conversation without typing anything extra. You can add your use cases to the match part.
### Use Specific Library Names
When you know exactly which library you want to use, be specific in your prompts. This helps Docfork find the right documentation faster and more accurately:
```txt
implement basic authentication with supabase. use docfork
```
```txt
create a Next.js middleware for rate limiting. use docfork
```
```txt
configure Tailwind CSS with custom typography. use docfork
```
The more specific you are about the library and what you want to accomplish, the better documentation you'll receive.
## Development
Clone the project and install dependencies:
```bash
npm i
```
Build:
```bash
npm run build
```
<details>
<summary><b>Environment Variables</b></summary>
The Docfork MCP server supports the following environment variables:
- `DEFAULT_MINIMUM_TOKENS`: Set the minimum token count for documentation retrieval (default: 10000)
### For HTTP/SSE Transport Only
The following environment variables are only relevant when running the server as an HTTP/SSE service (not for standard `npx` usage):
- `MCP_TRANSPORT`: Set the transport type for MCP communication (default: `stdio`, options: `streamable-http`, `stdio`, `sse`)
- `PORT`: Set the port number for HTTP/SSE transport (default: `3000`, only used when MCP_TRANSPORT is `streamable-http` or `sse`)
</details>
<details>
<summary><b>Example Configurations</b></summary>
**Standard node server configuration (most common):**
```json
{
"mcpServers": {
"docfork": {
"command": "npx",
"args": ["-y", "docfork@latest"],
"env": {
"DEFAULT_MINIMUM_TOKENS": "10000"
}
}
}
}
```
**HTTP/SSE server configuration (for custom deployments):**
These environment variables are used when you're running your own instance of the Docfork server, not when connecting to remote servers. For remote server connections, use the URL-based configurations shown earlier in this README (e.g., `"url": "https://mcp.docfork.com/mcp"`).
If you're self-hosting and want to run the server with HTTP/SSE transport:
```bash
# Set environment variables and run
MCP_TRANSPORT=streamable-http PORT=3000 npx -y docfork@latest
```
</details>
<details>
<summary><b>Local Configuration Example</b></summary>
```json
{
"mcpServers": {
"docfork": {
"command": "npx",
"args": ["tsx", "/path/to/folder/docfork/src/index.ts"]
}
}
}
```
</details>
<details>
<summary><b>Testing with MCP Inspector</b></summary>
```bash
npx -y @modelcontextprotocol/inspector npx docfork
```
</details>
## 🚨 Troubleshooting
<details>
<summary><b>Module Not Found Errors</b></summary>
If you encounter `ERR_MODULE_NOT_FOUND`, try using `bunx` instead of `npx`:
```json
{
"mcpServers": {
"docfork": {
"command": "bunx",
"args": ["-y", "docfork"]
}
}
}
```
This often resolves module resolution issues in environments where `npx` doesn't properly install or resolve packages.
</details>
<details>
<summary><b>ESM Resolution Issues</b></summary>
For errors like `Error: Cannot find module 'uriTemplate.js'`, try the `--experimental-vm-modules` flag:
```json
{
"mcpServers": {
"docfork": {
"command": "npx",
"args": ["-y", "--node-options=--experimental-vm-modules", "docfork"]
}
}
}
```
</details>
<details>
<summary><b>TLS/Certificate Issues</b></summary>
Use the `--experimental-fetch` flag to bypass TLS-related problems:
```json
{
"mcpServers": {
"docfork": {
"command": "npx",
"args": ["-y", "--node-options=--experimental-fetch", "docfork"]
}
}
}
```
</details>
<details>
<summary><b>General MCP Client Errors</b></summary>
1. Try adding `@latest` to the package name
2. Use `bunx` as an alternative to `npx`
3. Consider using `deno` as another alternative
4. Ensure you're using Node.js v18 or higher for native fetch support
</details>
## ⚠️ Disclaimer
Docfork is an open, community-driven catalogue. Although we review submissions, we make no warranties—express or implied—about the accuracy, completeness, or security of any linked documentation or code. Projects listed here are created and maintained by their respective authors, not by Docfork.
If you spot content that is suspicious, inappropriate, or potentially harmful, please contact us.
By using Docfork, you agree to do so at your own discretion and risk.
## 🌟 Let's Connect!
Stay in the loop and meet the community:
- 🐦 Follow us on [X](https://x.com/docfork_ai) for product news and updates
- 🌐 Visit our [Website](https://docfork.com)
## Star History
[](https://www.star-history.com/#docfork/mcp&Date)
## License
MIT
You Might Also Like
Ollama
Ollama enables easy access to large language models on various platforms.

n8n
n8n is a secure workflow automation platform for technical teams with 400+...
OpenWebUI
Open WebUI is an extensible web interface for customizable applications.
mssql_mcp_server
MSSQL MCP Server for secure SQL Server access via Claude Desktop.
mcp-reddit
MCP Reddit Server for fetching and analyzing Reddit content.
ros-mcp-server
The ROS MCP Server connects LLMs with robot control via natural language commands.