There are lots of ever more powerful open LLMs (HF hub), wonderful frameworks to build (train or merge) your own models (trl, axolotl), as well as reliable and efficient solutions to serve your models (vllm, tgi). But, or so I find, there are relatively few simple, local, open-source chat clients that work well with custom (self-hosted) LLMs and allow you to use your models in a straightforward way. (However, check chat clients below.)
simple-chat
is a minimalistic chat client that runs locally (in your browser) and connects to a local or remote LLM. It works out of the box, but can also be used as a boilerplate to build more sophisticated agents.
-
Clone the repository.
git clone https://github.com/debatelab/simple-chat.git cd simple-chat
-
Set base URL. Create a text file named
.env
(e.g. with text editor) that contains the following line:BASE_URL="<insert-your-inference-server-url-here>"
-
Install poetry (python package manager) and its dotenv plugin.
-
Install the dependencies.
poetry install
-
Run the app.
poetry run chainlit run src/simple_chat/app.py
- Sanctum AI: privacy-focused local chat client
- jan.ai: open-source chat client to interact with local and remote LLMs