drio
Open app

LibraryAI

Source

[C#/AI] (LibraryAI/ConsoleAI) AI 知识库与数据处理程序

Catalog onlyCatalog onlySTDIO

Overview

LibraryAI is an AI knowledge base and data processing program designed to facilitate text chunk processing, vectorization, and search functionalities using a vector database.

To use LibraryAI, developers can integrate its core library into their applications, utilize the ConsoleAI for text processing, and connect to vector databases through the MCP framework.

  • Text chunk processing and vectorization for efficient data handling. - MCP protocol support for seamless AI integration with vector databases. - Debugging tools for testing core library functionalities.
  1. Implementing AI-driven search functionalities in applications.
  2. Enhancing data processing capabilities in machine learning projects.
  3. Developing applications that require efficient text handling and vector database integration.

Add to your AI client

Use these steps to connect LibraryAI in Cursor, Claude, VS Code, and other MCP-compatible apps. The same JSON appears in the Use with menu above for one-click copy.

Cursor

Add this to your .cursor/mcp.json file in your project root, then restart Cursor.

.cursor/mcp.json

{
  "mcpServers": {
    "libraryai-yelandaokong": {
      "command": "npx",
      "args": [
        "-y",
        "@modelcontextprotocol/server-libraryai-yelandaokong"
      ]
    }
  }
}

Claude Desktop

Add this server entry to the mcpServers object in your Claude Desktop config, then restart the app.

~/Library/Application Support/Claude/claude_desktop_config.json (macOS) or %APPDATA%\Claude\claude_desktop_config.json (Windows)

{
  "mcpServers": {
    "libraryai-yelandaokong": {
      "command": "npx",
      "args": [
        "-y",
        "@modelcontextprotocol/server-libraryai-yelandaokong"
      ]
    }
  }
}

Claude Code

Add this to your project's .mcp.json file. Claude Code will detect it automatically.

.mcp.json (project root)

{
  "mcpServers": {
    "libraryai-yelandaokong": {
      "command": "npx",
      "args": [
        "-y",
        "@modelcontextprotocol/server-libraryai-yelandaokong"
      ]
    }
  }
}

VS Code (Copilot)

Add this to your .vscode/mcp.json file. Requires the GitHub Copilot extension with MCP support enabled.

.vscode/mcp.json

{
  "servers": {
    "libraryai-yelandaokong": {
      "type": "stdio",
      "command": "npx",
      "args": [
        "-y",
        "@modelcontextprotocol/server-libraryai-yelandaokong"
      ]
    }
  }
}

Windsurf

Add this to your Windsurf MCP config file, then restart Windsurf.

~/.codeium/windsurf/mcp_config.json

{
  "mcpServers": {
    "libraryai-yelandaokong": {
      "command": "npx",
      "args": [
        "-y",
        "@modelcontextprotocol/server-libraryai-yelandaokong"
      ]
    }
  }
}

Cline

Open Cline settings, navigate to MCP Servers, and add this server configuration.

Cline MCP Settings (via UI)

{
  "mcpServers": {
    "libraryai-yelandaokong": {
      "command": "npx",
      "args": [
        "-y",
        "@modelcontextprotocol/server-libraryai-yelandaokong"
      ]
    }
  }
}

FAQ

What programming language is LibraryAI built with?

LibraryAI is developed using C#.

Is LibraryAI suitable for production use?

LibraryAI is in the early stages of development and may require further testing before production use.

How can I contribute to LibraryAI?

Contributions can be made through the GitHub repository.