Follow these simple steps to have SageAI running on your machine. No complex setup, no cloud accounts, just pure local AI power.
Ollama runs your AI models locally. Choose your operating system:
Run this single command in your terminal:
curl -fsSL https://ollama.com/install.sh | sh
Note: On most Linux distributions, Ollama will automatically start as a service.
Option 1: Using Homebrew (recommended)
brew install ollama
Option 2: Download the installer
Download for macOS
Tip: After installation, start Ollama from
Applications or run
ollama serve
Download and run the Windows installer:
Download for WindowsNote: Ollama will run in the system tray after installation. It starts automatically on boot.
Download the Llama 3 model (recommended for best results):
ollama pull llama3:latest
llama3:latest
- Full model
~4.7GB
llama3.2:1b -
Smaller, faster
~1.3GB
llama3.2:3b -
Balanced
~2GB
💡 For slower machines, try the smaller models. For best quality, use llama3:latest.
This will take a few minutes depending on your internet speed. Go grab a coffee! ☕
Choose the configuration that matches your setup:
For local development with hot-reload and debugging enabled.(Note: the source code needed to be cloned to run dev compose files)
💡 Recommendation: Start with the Dev version for your OS. It's easier to debug!
For Dev mode, you need to clone the repository to build the services locally. For Prod mode, you can skip to Step 5 and just download the compose file.
# Clone the repository git clone --branch submission/ridowan https://github.com/WhisperNet/research-paper-rag-assessment.git # Navigate to the project directory cd research-paper-rag-assessment # For Linux Dev(Run this from the project root) docker compose -f infra/docker-compose.linux.dev.yml up --build # For Mac/Windows Dev(Run this from the project root) docker compose -f infra/docker-compose.mac-win.dev.yml up --build
💡 Note: This step is only required if you're using Dev compose files. Production files use pre-built images and don't need the source code.
Navigate to the
resources
folder and run the compose file:
# Navigate to resources folder cd resources # For Linux Dev (Run this from the project root) docker compose -f infra/docker-compose.linux.dev.yml up --build # Or for Mac/Windows Dev (Run this from the project root) docker compose -f infra/docker-compose.mac-win.dev.yml up --build # Or for Production (any OS) docker compose -f docker-compose.linux.prod.yml up -d docker compose -f docker-compose.mac-win.prod.yml up -d
⏱️ First build takes 3-5 minutes. Subsequent starts are under 30 seconds!
Open your browser and navigate to:
Make sure Ollama is running:
ollama serve
Check if it's accessible:
curl http://localhost:11434
Stop conflicting services or modify the ports in the compose
file.
Check what's using a port:
lsof -i :8080
(Mac/Linux)
Ensure you have enough disk space (at least 10GB free).
Try:
docker system prune
to clean up old images.