Install & Configure Ollama
Run powerful AI models locally on your broadcasting machine — no cloud subscriptions, no data leaving your network. Generate artist bios, smart playlists, and show content with a single REST API on localhost:11434.
What is Ollama?
A lightweight, open-source runtime that lets you download and run large language models entirely on your own hardware. No API keys, no monthly fees, no data sent to third parties.
100% Local
Every model runs on your CPU or GPU. Your prompts, your data, and your responses never leave the machine. Perfect for stations handling sensitive playlist data or unreleased content metadata.
REST API on :11434
Ollama exposes a clean JSON API at http://localhost:11434. Any application on your machine — including Mcaster1Studio and Mcaster1AMP — can send prompts and receive completions over plain HTTP.
Model Library
Pull from hundreds of pre-quantized models with a single command. Llama 3.1, Mistral, Gemma, Phi, CodeLlama, and more — each optimized for different hardware profiles and use cases.
Installation
Ollama supports macOS, Linux, and Windows. Installation takes under a minute on most systems.
macOS
Linux
Windows
Verify Installation
Start the Server
Hardware Tiers
Not every broadcast machine is a powerhouse. Pick the right model size for your available RAM, CPU cores, and storage to keep your station running smoothly.
Entry-Level Broadcast PC
- Phi-3 Mini (3.8B) — compact & fast
- Gemma 2B — efficient text generation
- TinyLlama 1.1B — minimal footprint
- Best for: short bios, tag cleanup, basic prompts
Mid-Range Workstation
- Llama 3.1 8B — general purpose workhorse
- Mistral 7B — fast inference, great quality
- Gemma 7B — strong reasoning ability
- Best for: artist bios, playlist logic, show notes
Dedicated AI Server
- Llama 3.1 70B — near-cloud quality
- Mixtral 8x7B — mixture-of-experts speed
- CodeLlama 34B — code & scripting tasks
- Best for: long-form content, multi-step workflows, analytics
Pull Your First Model
Each ollama pull command downloads a quantized model
optimized for CPU inference. Models are cached locally so subsequent runs start instantly.
Test Interactively
Integration with Mcaster1
Once Ollama is running, Mcaster1 products detect it automatically and unlock AI-powered features — no configuration files to edit, no API keys to manage.
Mcaster1Studio
The broadcast automation suite connects to Ollama's REST API at localhost:11434 to generate artist bios on the fly, create show notes from your playlist history, and draft social media posts for upcoming segments.
Auto-detectedMcaster1AMP
The intelligent media player uses local AI models to analyze your music library, suggest playlist transitions based on tempo and genre, and generate descriptive metadata for tracks missing artist information.
Auto-detectedHow It Works
Both products send standard HTTP POST requests to Ollama's
/api/generate and /api/chat
endpoints. You choose which model to use in each product's AI settings panel.
Responses stream back in real time.
Ollama API Quick Reference
These are the core endpoints you will use most often. All requests go to http://localhost:11434 and accept/return JSON.
Generate Text
List Installed Models
Chat Format (Multi-Turn)
Model Information
Ready to Broadcast Smarter?
With Ollama running locally, you have a private AI assistant purpose-built for your station. Explore our model recommendations to find the perfect fit for your hardware and workflow.