Skip to content

Latest commit

 

History

History
10 lines (7 loc) · 515 Bytes

README.md

File metadata and controls

10 lines (7 loc) · 515 Bytes

How to run a local LLM with a chat UI in four easy steps

This is the repo for the article Set up a local LLM on CPU with chat UI in 15 minutes.

The process consists of these simple steps:

  1. Select a model on Huggingface, e.g. "RJuro/munin-neuralbeagle-7b"
  2. Quantize the model by running quantize.py
  3. Wrap model in Ollama image
  4. Build and run a Docker container that wraps the GUI, e.g. Chatbot Ollama