Ollama Setup Guide
Enable local AI features in Fiavaion products by installing Ollama. Your data stays on your device—no cloud processing, no API fees, complete privacy.
What You'll Need
Operating System
Windows 10/11, macOS 11+, or Linux
Memory (RAM)
8GB minimum, 16GB+ recommended
Disk Space
5GB minimum for basic models
GPU (Optional)
NVIDIA GPU with 6GB+ VRAM for faster processing
Installation
Select your operating system:
Download Ollama
Visit the official Ollama website and download the Windows installer.
Download for WindowsRun the Installer
Double-click the downloaded OllamaSetup.exe file and follow the installation wizard. Accept the default settings.
Verify Installation
Open Command Prompt or PowerShell and run:
ollama --version You should see a version number like ollama version 0.5.x
Download Your First Model
Download the recommended model for general AI tasks:
ollama pull llama3.2:3b This downloads a 2GB model optimized for most tasks. The download may take a few minutes depending on your internet speed.
Test the Model
Verify everything works by running a quick test:
ollama run llama3.2:3b "Hello, are you working?" If you see a response, Ollama is ready! Press Ctrl+C to exit.
Ollama Runs in Background
After installation, Ollama runs automatically as a background service. Look for the llama icon in your system tray (bottom-right of taskbar). Our apps connect to it at localhost:11434.
Download Ollama
Visit the official Ollama website and download the macOS app.
Download for macOSInstall the App
- Open the downloaded
Ollama-darwin.zipfile - Drag Ollama.app to your Applications folder
- Double-click Ollama in Applications to launch it
- If prompted, click "Open" to allow the app from an identified developer
Allow Background Access
When Ollama first runs, it may ask for permission to run in the background. Click Allow to let it serve AI requests.
You'll see a llama icon appear in your menu bar (top-right).
Open Terminal and Download a Model
Open Terminal (search in Spotlight with Cmd+Space) and run:
ollama pull llama3.2:3b This downloads the recommended 2GB model. Wait for the download to complete.
Test the Model
Verify everything works:
ollama run llama3.2:3b "Hello, are you working?" If you see a response, you're all set! Press Ctrl+C to exit.
Mac Performance Tips
Excellent performance. The GPU is used automatically. 16GB+ unified memory recommended for larger models.
Good performance on CPU. Stick to smaller models like llama3.2:3b or phi3:mini for best results.
Alternative: Install via Homebrew
If you prefer Homebrew:
brew install ollama ollama serve & ollama pull llama3.2:3b Install with One Command
Open your terminal and run the official install script:
curl -fsSL https://ollama.com/install.sh | sh This automatically detects your system and installs Ollama.
Start the Ollama Service
The installer usually starts Ollama automatically. If not, run:
ollama serve Or to run as a systemd service:
sudo systemctl enable ollama sudo systemctl start ollama Download a Model
ollama pull llama3.2:3b Test the Model
ollama run llama3.2:3b "Hello, are you working?" Recommended Models
Select your GPU's VRAM to see which models work best for your hardware. Apple Silicon users: match your unified memory to these tiers.
llama3.2:3b Required Summarization, simplification, study questions
2GB downloadphi3:mini Optional Fast responses, good for quick tasks
2.3GB downloadInstall Commands
ollama pull llama3.2:3bollama pull phi3:mini llama3.2:3b Required Summarization, simplification, study questions
2GB downloadllava Recommended Image descriptions and diagram understanding
4.7GB downloadmistral:7b Optional Better writing analysis and text quality
4.1GB downloadInstall Commands
ollama pull llama3.2:3bollama pull llavaollama pull mistral:7b llama3.1:8b Recommended Noticeably better quality for all tasks
4.7GB downloadllava Recommended Image descriptions and diagram understanding
4.7GB downloadmistral:7b Optional Creative writing and detailed analysis
4.1GB downloadInstall Commands
ollama pull llama3.1:8bollama pull llavaollama pull mistral:7b llama3.1:8b Recommended High quality default for all tasks
4.7GB downloadllava:13b Recommended Superior image understanding and descriptions
8GB downloadgemma2:9b Optional Strong reasoning and comprehension
5.4GB downloadInstall Commands
ollama pull llama3.1:8bollama pull llava:13bollama pull gemma2:9b llama3.1:13b Recommended Excellent quality for all tasks
7.4GB downloadllava:13b Recommended Superior image understanding
8GB downloaddeepseek-r1:14b Optional Advanced reasoning and problem solving
8.5GB downloadInstall Commands
ollama pull llama3.1:13bollama pull llava:13bollama pull deepseek-r1:14b mixtral:8x7b Recommended State-of-the-art mixture-of-experts model
26GB downloadllama3.1:13b Recommended High quality, fast responses
7.4GB downloadllava:34b Optional Best-in-class image understanding
20GB downloadInstall Commands
ollama pull mixtral:8x7bollama pull llama3.1:13bollama pull llava:34b Windows: Open Task Manager → Performance → GPU to see dedicated GPU memory. Mac: Apple menu → About This Mac shows your total unified memory—subtract 4-8GB for macOS and apps to estimate what's available for models.
Verify Connection
Fiavaion apps automatically detect Ollama. Here's how to verify it's running:
Check in Browser
Open localhost:11434 in your browser. You should see "Ollama is running".
Check in Terminal
Run ollama list to see your installed models.
Check in Our Apps
Open any Fiavaion app with AI features. It will show a green indicator if Ollama is connected.
Troubleshooting
"Ollama not detected" error
- Make sure Ollama is running (check for the llama icon in system tray/menu bar)
- Try restarting Ollama
- Verify localhost:11434 is accessible
- Check if a firewall is blocking the connection
Slow responses
- Try a smaller model:
ollama pull phi3:mini - Close other memory-intensive applications
- If using CPU only, responses will be slower than with a GPU
- Check available RAM with Task Manager (Windows) or Activity Monitor (Mac)
Model download fails
- Check your internet connection
- Ensure you have enough disk space (models are 2-8GB each)
- Try a smaller model first:
ollama pull phi3:mini - If behind a proxy, configure it in your terminal environment
Out of memory errors
- Use a smaller model (3b or mini variants)
- Close other applications to free up RAM
- Restart Ollama to clear any stuck processes
- Consider upgrading RAM if you frequently hit limits
macOS: "Cannot be opened" error
- Right-click (or Ctrl+click) the Ollama app
- Select "Open" from the context menu
- Click "Open" in the dialog that appears
- This only needs to be done once
Windows: Ollama not starting
- Check Windows Services (services.msc) for "Ollama"
- Right-click and select "Start" if it's stopped
- Try reinstalling Ollama
- Run as Administrator if needed
Your Data Stays Local
With Ollama, all AI processing happens on your computer. Your documents, text, and images are never sent to any server. This makes it safe for:
- Student work (FERPA compliant)
- Healthcare data (HIPAA safe)
- Personal documents (GDPR friendly)
- Sensitive business information