drio
Open app

LLM_MCP

Source

Building MCP client and server for LLM

Catalog onlyCatalog onlySTDIO

Overview

LLM_MCP is a project focused on building a client and server for a Large Language Model (LLM) that facilitates communication and processing.

To use LLM_MCP, clone the repository from GitHub, set up the server and client as per the instructions provided in the documentation, and start interacting with the LLM.

  • Client-server architecture for efficient communication with LLMs. - Support for various LLM functionalities. - Open-source project under Apache-2.0 license.
  1. Developing applications that require natural language understanding.
  2. Creating chatbots that leverage LLM capabilities.
  3. Integrating LLMs into existing software solutions for enhanced functionality.

Add to your AI client

Use these steps to connect LLM_MCP in Cursor, Claude, VS Code, and other MCP-compatible apps. The same JSON appears in the Use with menu above for one-click copy.

Cursor

Add this to your .cursor/mcp.json file in your project root, then restart Cursor.

.cursor/mcp.json

{
  "mcpServers": {
    "llm-mcp-nkarnaud": {
      "command": "npx",
      "args": [
        "-y",
        "@modelcontextprotocol/server-llm-mcp-nkarnaud"
      ]
    }
  }
}

Claude Desktop

Add this server entry to the mcpServers object in your Claude Desktop config, then restart the app.

~/Library/Application Support/Claude/claude_desktop_config.json (macOS) or %APPDATA%\Claude\claude_desktop_config.json (Windows)

{
  "mcpServers": {
    "llm-mcp-nkarnaud": {
      "command": "npx",
      "args": [
        "-y",
        "@modelcontextprotocol/server-llm-mcp-nkarnaud"
      ]
    }
  }
}

Claude Code

Add this to your project's .mcp.json file. Claude Code will detect it automatically.

.mcp.json (project root)

{
  "mcpServers": {
    "llm-mcp-nkarnaud": {
      "command": "npx",
      "args": [
        "-y",
        "@modelcontextprotocol/server-llm-mcp-nkarnaud"
      ]
    }
  }
}

VS Code (Copilot)

Add this to your .vscode/mcp.json file. Requires the GitHub Copilot extension with MCP support enabled.

.vscode/mcp.json

{
  "servers": {
    "llm-mcp-nkarnaud": {
      "type": "stdio",
      "command": "npx",
      "args": [
        "-y",
        "@modelcontextprotocol/server-llm-mcp-nkarnaud"
      ]
    }
  }
}

Windsurf

Add this to your Windsurf MCP config file, then restart Windsurf.

~/.codeium/windsurf/mcp_config.json

{
  "mcpServers": {
    "llm-mcp-nkarnaud": {
      "command": "npx",
      "args": [
        "-y",
        "@modelcontextprotocol/server-llm-mcp-nkarnaud"
      ]
    }
  }
}

Cline

Open Cline settings, navigate to MCP Servers, and add this server configuration.

Cline MCP Settings (via UI)

{
  "mcpServers": {
    "llm-mcp-nkarnaud": {
      "command": "npx",
      "args": [
        "-y",
        "@modelcontextprotocol/server-llm-mcp-nkarnaud"
      ]
    }
  }
}

FAQ

What is the purpose of LLM_MCP?

LLM_MCP aims to provide a robust framework for building applications that utilize large language models.

Is LLM_MCP free to use?

Yes! LLM_MCP is open-source and free to use under the Apache-2.0 license.

How can I contribute to LLM_MCP?

Contributions are welcome! You can fork the repository, make your changes, and submit a pull request.