Ryan-yang125 / ChatLLM-Web

๐Ÿ—ฃ๏ธ Chat with LLM like Vicuna totally in your browser with WebGPU, safely, privately, and with no server. Powered by web llm.

Home Page:https://chat-llm-web.vercel.app

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

icon

ChatLLM Web

English / ็ฎ€ไฝ“ไธญๆ–‡ / ๆ—ฅๆœฌ่ชž

๐Ÿ—ฃ๏ธ Chat with LLM like Vicuna totally in your browser with WebGPU, safely, privately, and with no server. Powered By web-llm.

Try it now

cover

Features

  • ๐Ÿค– Everything runs inside the browser with no server support and is accelerated with WebGPU.

  • โš™๏ธ Model runs in a web worker, ensuring that it doesn't block the user interface and providing a seamless experience.

  • ๐Ÿš€ Easy to deploy for free with one-click on Vercel in under 1 minute, then you get your own ChatLLM Web.

  • ๐Ÿ’พ Model caching is supported, so you only need to download the model once.

  • ๐Ÿ’ฌ Multi-conversation chat, with all data stored locally in the browser for privacy.

  • ๐Ÿ“ Markdown and streaming response support: math, code highlighting, etc.

  • ๐ŸŽจ responsive and well-designed UI, including dark mode.

  • ๐Ÿ’ป PWA supported, download it and run totally offline.

Instructions

  • ๐ŸŒ To use this app, you need a browser that supports WebGPU, such as Chrome 113 or Chrome Canary. Chrome versions โ‰ค 112 are not supported.

  • ๐Ÿ’ป You will need a GPU with about 6.4GB of memory. If your GPU has less memory, the app will still run, but the response time will be slower.

  • ๐Ÿ“ฅ The first time you use the app, you will need to download the model. For the Vicuna-7b model that we are currently using, the download size is about 4GB. After the initial download, the model will be loaded from the browser cache for faster usage.

  • โ„น๏ธ For more details, please visit mlc.ai/web-llm

Roadmap

  • [โœ…] LLM: using web worker to create an LLM instance and generate answers.

  • [โœ…] Conversations: Multi-conversation support is available.

  • [โœ…] PWA

  • [] Settings:

    • ui: dark/light theme
    • device:
      • gpu device choose
      • cache usage and manage
    • model:
      • support multi models: vicuna-7bโœ… RedPajama-INCITE-Chat-3B []
      • params config: temperature, max-length, etc.
      • export & import model

Deploy to Vercel

  1. Click Deploy with Vercel, follow the instructions, and finish in just 1 minute.
  2. Enjoy it ๐Ÿ˜Š

Development

git clone https://github.com/Ryan-yang125/ChatLLM-Web.git
cd ChatLLM-Web
npm i
npm run dev

Screenshots

Home

More

LICENSE

MIT

About

๐Ÿ—ฃ๏ธ Chat with LLM like Vicuna totally in your browser with WebGPU, safely, privately, and with no server. Powered by web llm.

https://chat-llm-web.vercel.app

License:MIT License


Languages

Language:JavaScript 70.4%Language:TypeScript 29.6%Language:CSS 0.1%