bigsweetpotatostudio_hyperchat

bigsweetpotatostudio_hyperchat

by bigsweetpotatostudio
HyperChat is an open chat client that integrates APIs from various LLMs and implements productivity tools through the MCP protocol.

HyperChat MCP Server: A Multi-LLM Chat Client with Productivity Tools

Overview

HyperChat is an open-source chat client designed to provide the best chat experience by leveraging APIs from various Large Language Models (LLMs). It fully supports the Model Context Protocol (MCP) and includes productivity tools built on native Agents. HyperChat is compatible with multiple platforms, including Windows, macOS, and Linux, and offers a user-friendly interface for managing and interacting with LLMs.

Key Features

  • Multi-Platform Support: Runs on Windows, macOS, and Linux.
  • Multi-LLM Integration: Supports OpenAI, Claude (via OpenRouter), Qwen, Deepseek, GLM, and Ollama.
  • MCP Plugin Marketplace: Built-in marketplace for MCP plugins with one-click installation.
  • Third-Party MCP Support: Manual installation of third-party MCPs by specifying command, args, and env.
  • Productivity Tools: Includes tools for managing resources, prompts, and scheduled tasks.
  • Dark Mode: Supports dark mode for better user experience.
  • Multi-Language Support: Available in both English and Chinese.
  • Rendering Capabilities: Supports rendering of Artifacts, SVG, and HTML.
  • WebDAV Syncing: Enables synchronization of data via WebDAV.
  • RAG Integration: Adds Retrieval-Augmented Generation (RAG) based on MCP knowledge bases.
  • ChatSpace Concept: Supports multiple simultaneous conversations.

Demo

You can try the HyperChat demo on Docker: HyperChat Demo

Installation

Command Line Execution

npx -y @dadigua/hyper-chat

Default port: 16100, password: 123456. Access via http://localhost:16100/123456/

Docker

docker pull dadigua/hyperchat-mini:latest

Prerequisites

  • uv & nodejs: Ensure you have uv and nodejs installed on your system.

Installing uv

# MacOS
brew install uv
# Windows
winget install --id=astral-sh.uv -e

Installing nodejs

# MacOS
brew install node
# Windows
winget install OpenJS.NodeJS.LTS

Usage

  1. Configure APIKEY: Ensure your LLM service is compatible with OpenAI-style APIs.
  2. Run HyperChat: Use the command line or Docker to start the server.
  3. Access the Web Interface: Open your browser and navigate to the provided URL.

Development

To contribute to HyperChat, follow these steps:

cd electron && npm install
cd web && pnpm install
npm install
npm run dev

Examples

Calling Shell MCP

Shell MCP

Scheduled Task List

Scheduled Tasks

MCP Marketplace

MCP Marketplace

HTML Rendering

HTML Rendering

Disclaimer

This project is intended for learning and communication purposes only. Any actions taken using this project, such as web scraping, are the responsibility of the user and not the developers.

About

HyperChat is developed by BigSweetPotatoStudio and is available under an open-source license. For more information, visit the GitHub repository.

Topics

License

View the license for more details.

Contributors

Languages

  • JavaScript 72.2%
  • TypeScript 25.9%
  • CSS 1.6%
  • Other 0.3%

Features & Capabilities

Categories
mcp_server model_context_protocol javascript typescript docker api_integration llm claude openai productivity_tools chat_application

Implementation Details

Stats

0 Views
397 GitHub Stars

Repository Info

bigsweetpotatostudio Organization

Similar MCP Servers

continuedev_continue by continuedev
25049
21423
9300