$ -weight: 500;">curl -fsSL https://ollama.com/-weight: 500;">install.sh | sh
-weight: 500;">curl -fsSL https://ollama.com/-weight: 500;">install.sh | sh
-weight: 500;">curl -fsSL https://ollama.com/-weight: 500;">install.sh | sh
# Pull a model
ollama pull llama3.2 # Run it in the terminal
ollama run llama3.2 # Or use a specific model
ollama run qwen2.5:14b
# Pull a model
ollama pull llama3.2 # Run it in the terminal
ollama run llama3.2 # Or use a specific model
ollama run qwen2.5:14b
# Pull a model
ollama pull llama3.2 # Run it in the terminal
ollama run llama3.2 # Or use a specific model
ollama run qwen2.5:14b
# Simple -weight: 500;">curl call
-weight: 500;">curl http://localhost:11434/api/generate \ -d '{ "model": "llama3.2", "prompt": "Explain Docker in one paragraph", "stream": false }'
# Simple -weight: 500;">curl call
-weight: 500;">curl http://localhost:11434/api/generate \ -d '{ "model": "llama3.2", "prompt": "Explain Docker in one paragraph", "stream": false }'
# Simple -weight: 500;">curl call
-weight: 500;">curl http://localhost:11434/api/generate \ -d '{ "model": "llama3.2", "prompt": "Explain Docker in one paragraph", "stream": false }'
# Python
import requests response = requests.post("http://localhost:11434/api/generate", json={ "model": "llama3.2", "prompt": "Write a Python function that reads a CSV and returns the top 5 rows", "stream": False
}) print(response.json()["response"])
# Python
import requests response = requests.post("http://localhost:11434/api/generate", json={ "model": "llama3.2", "prompt": "Write a Python function that reads a CSV and returns the top 5 rows", "stream": False
}) print(response.json()["response"])
# Python
import requests response = requests.post("http://localhost:11434/api/generate", json={ "model": "llama3.2", "prompt": "Write a Python function that reads a CSV and returns the top 5 rows", "stream": False
}) print(response.json()["response"])
# Same endpoint format as OpenAI
-weight: 500;">curl http://localhost:11434/v1/chat/completions \ -H "Content-Type: application/json" \ -d '{ "model": "llama3.2", "messages": [ {"role": "user", "content": "What is a homelab?"} ] }'
# Same endpoint format as OpenAI
-weight: 500;">curl http://localhost:11434/v1/chat/completions \ -H "Content-Type: application/json" \ -d '{ "model": "llama3.2", "messages": [ {"role": "user", "content": "What is a homelab?"} ] }'
# Same endpoint format as OpenAI
-weight: 500;">curl http://localhost:11434/v1/chat/completions \ -H "Content-Type: application/json" \ -d '{ "model": "llama3.2", "messages": [ {"role": "user", "content": "What is a homelab?"} ] }'
-weight: 500;">docker run -d \ -p 3000:8080 \ --add-host=host.-weight: 500;">docker.internal:host-gateway \ -v open-webui:/app/backend/data \ --name open-webui \ ---weight: 500;">restart always \ ghcr.io/open-webui/open-webui:main
-weight: 500;">docker run -d \ -p 3000:8080 \ --add-host=host.-weight: 500;">docker.internal:host-gateway \ -v open-webui:/app/backend/data \ --name open-webui \ ---weight: 500;">restart always \ ghcr.io/open-webui/open-webui:main
-weight: 500;">docker run -d \ -p 3000:8080 \ --add-host=host.-weight: 500;">docker.internal:host-gateway \ -v open-webui:/app/backend/data \ --name open-webui \ ---weight: 500;">restart always \ ghcr.io/open-webui/open-webui:main - Any modern CPU (Intel 10th gen+, Ryzen 3000+)
- 8GB RAM (16GB better)
- No GPU required — CPU inference works, just slower
- ~5-10GB disk space per model - NVIDIA GPU with 6GB+ VRAM for 7B models
- 8-12GB VRAM for 13-14B models
- 16GB VRAM for comfortable 27B models
- AMD GPU works too (ROCm support, somewhat newer) - Model switching mid-conversation
- Document upload and chat (RAG)
- Conversation history
- System prompt customization - Specific, clear instructions
- Examples of the output format you want
- System prompts that set context and constraints - Install Ollama: -weight: 500;">curl -fsSL https://ollama.com/-weight: 500;">install.sh | sh
- Pull a model: ollama pull llama3.2
- Run it: ollama run llama3.2
- If you have a decent GPU: try qwen2.5:14b
- Explore Open WebUI for a proper chat interface