CPU

Machine Learning in Linux: Serge – web interface for chatting

Artificial intelligence icon Our Machine Learning in Linux series focuses on apps that make it easy to experiment with machine learning. All the apps covered in the series can be self-hosted.

Large Languages Models trained on massive amount of text can perform new tasks from textual instructions. They can generate creative text, solve maths problems, answer reading comprehension questions, and much more.

Serge is a chat interface crafted with LLaMA for running GGUF models.

Installation

We evaluated Serge using Manjaro, an Arch-based distro, as well as the ubiquitous Ubuntu 24.10 distro.

With both distros, installation is a breeze. We used Docker to install and run Serge with Docker. Issue the command:

docker run -d \
    --name serge \
    -v weights:/usr/src/app/weights \
    -v datadb:/data/db/ \
    -p 8008:8008 \
    ghcr.io/serge-chat/serge:latest

This is what you’ll see.

Installing Serge

Next point your web browser to http://localhost:8008 so that you can download one or more models. There’s a whole raft of models available.

Download models
Click image for full size

In the above image, we’re downloading a couple of models. Many of the models are big downloads. Some of the models may not work if your system doesn’t have sufficient RAM. But storing the models locally lets you run the software self-hosted.

The software is cross-platform. Besides Linux, it runs under macOS and Windows.

Next page: Page 2 – In Operation and Summary

Pages in this article:
Page 1 – Introduction and Installation
Page 2 – In Operation and Summary

Subscribe
Notify of
guest

This site uses Akismet to reduce spam. Read our Comment FAQ.

0 Comments
Oldest
Newest Most Voted
Inline Feedbacks
View all comments