Install Local LLM (Quick Guide)

This guide will walk you through the process of installing Ollama on a fresh Ubuntu setup, enabling you to run the model on your own hardware without relying on cloud-based services.
1. Install curl
sudo apt update
sudo apt install curl -y
2. Install Ollama
curl -fsSL https://ollama.com/install.sh | sh
Verify:
ollama --version

3. Run a Model
Example (recommended for CPU):
ollama run llama3:8b
Or faster 7B option:
ollama run mistral:7b
First run will download the model

4. Monitor CPU / RAM (Optional)
sudo apt install htop -y
htop

5. Start interactive chat
