AI Setup
Configure local AI for your PC2 personal cloud. This guide covers Ollama installation, model selection, GPU acceleration, and connecting external providers.
With Ollama (local AI), your conversations stay 100% private - they never leave your machine.
Installing Ollama
curl -fsSL https://ollama.com/install.sh | shVerify installation:
ollama --versionRecommended Models
| Model | Size | RAM Needed | Best For |
|---|---|---|---|
deepseek-r1:1.5b | 1GB | 4GB | Fast responses, basic tasks |
llama3.2:3b | 2GB | 6GB | Good balance |
phi3:mini | 2GB | 6GB | Microsoft’s efficient model |
mistral:7b | 4GB | 8GB | Strong general purpose |
llama3.2:8b | 5GB | 12GB | Complex reasoning |
codellama:7b | 4GB | 8GB | Code generation |
Install a Model
ollama pull deepseek-r1:1.5bOr via PC2: Settings → AI Setup → Click “Install” on any model.
List Installed Models
ollama listGPU Acceleration
GPU dramatically improves AI speed.
NVIDIA GPUs (CUDA)
Ollama automatically uses NVIDIA GPUs if CUDA is available.
# Check GPU is being used
nvidia-smiApple Silicon (M1/M2/M3)
Ollama automatically uses Metal acceleration. No configuration needed.
No GPU?
CPU-only works for smaller models (1.5b-3b). Larger models will be slow but functional.
Connecting to Remote Ollama
If running Ollama on a different machine (like a powerful server):
On the Ollama Server
OLLAMA_HOST=0.0.0.0 ollama serveIn PC2 Settings
- Go to Settings → AI Setup
- Set Ollama URL to
http://server-ip:11434 - Save
External AI Providers
PC2 also supports cloud providers for when you need more power:
| Provider | Models | Get API Key |
|---|---|---|
| OpenAI | GPT-4, GPT-3.5 | platform.openai.com |
| Anthropic | Claude 3 | console.anthropic.com |
| Gemini | aistudio.google.com | |
| xAI | Grok | xAI dashboard |
Cloud providers send data to their servers. Use Ollama for maximum privacy.
Troubleshooting
”Ollama not available”
# Check if running
curl http://localhost:11434/api/tags
# Start Ollama
ollama serveSlow Responses
- Use smaller model (
deepseek-r1:1.5b) - Enable GPU (see above)
- Check system resources:
htop
Out of Memory
- Use smaller/quantized model
- Close other applications
- Add more RAM or swap
Model Recommendations by Use Case
| Use Case | Recommended Model |
|---|---|
| General chat | llama3.2:3b, mistral:7b |
| Coding | codellama:7b, deepseek-coder:6.7b |
| Writing | mistral:7b, llama3.2:8b |
| Fast responses | deepseek-r1:1.5b, phi3:mini |