@claude-code-commands/code-tools-ollama-setup-ollama
Set up Ollama on the machine for local LLM inference
prpm install @claude-code-commands/code-tools-ollama-setup-ollama2 total downloads
📄 Full Prompt Content
---
description: Set up Ollama on the machine for local LLM inference
tags: [ai, ml, ollama, llm, setup, project, gitignored]
---
You are helping the user set up Ollama for local LLM inference.
## Process
1. **Check if Ollama is already installed**
- Run: `ollama --version`
- Check if service is running: `systemctl status ollama` or `sudo systemctl status ollama`
2. **Install Ollama if needed**
- Download and install: `curl -fsSL https://ollama.com/install.sh | sh`
- Or manual install from https://ollama.com/download
- Verify installation: `ollama --version`
3. **Start Ollama service**
- Start service: `systemctl start ollama` or `sudo systemctl start ollama`
- Enable on boot: `systemctl enable ollama` or `sudo systemctl enable ollama`
- Check status: `systemctl status ollama`
4. **Verify GPU support (for AMD on Daniel's system)**
- Check if ROCm is detected: `rocm-smi` or `rocminfo`
- Ollama should auto-detect AMD GPU
- Check Ollama logs for GPU recognition: `journalctl -u ollama -n 50`
5. **Configure Ollama**
- Check default model storage: `~/.ollama/models`
- Environment variables (if needed):
- `OLLAMA_HOST` - change port/binding
- `OLLAMA_MODELS` - custom model directory
- `OLLAMA_NUM_PARALLEL` - parallel requests
- Edit systemd service if needed: `/etc/systemd/system/ollama.service`
6. **Test Ollama**
- Pull a test model: `ollama pull llama2` (or smaller: `ollama pull tinyllama`)
- Run a test: `ollama run tinyllama "Hello, how are you?"`
- Verify GPU usage during inference
7. **Suggest initial models**
- Based on Daniel's hardware (AMD GPU), suggest:
- General: llama3.2, qwen2.5
- Code: codellama, deepseek-coder
- Fast: tinyllama, phi
- Vision: llava, bakllava
## Output
Provide a summary showing:
- Ollama installation status and version
- Service status
- GPU detection status
- Default configuration
- Recommended models to pull
- Next steps for usage
💡 Suggested Test Inputs
Loading suggested inputs...
🎯 Community Test Results
Loading results...
📦 Package Info
- Format
- claude
- Type
- slash-command
- Category
- general