ollama-setup

Configure auto-configure Ollama when user needs local LLM deployment, free AI alternatives, or wants to eliminate hosted API costs. Trigger phrases: "install ollama", "local AI", "free LLM", "self-hosted AI", "replace OpenAI", "no API costs". Use when appropriate context detected. Trigger with relevant phrases based on skill purpose. allowed-tools: Read, Write, Bash(cmd:*) version: 1.0.0 author: Jeremy Longshore <jeremy@intentsolutions.io> license: MIT

Allowed Tools

No tools specified

Provided by Plugin

ollama-local-ai

Run AI models locally with Ollama - free alternative to OpenAI, Anthropic, and other paid LLM APIs. Zero-cost, privacy-first AI infrastructure.

ai ml v1.0.0
View Plugin

Installation

This skill is included in the ollama-local-ai plugin:

/plugin install ollama-local-ai@claude-code-plugins-plus

Click to copy

Instructions

# Ollama Setup ## Overview This skill provides automated assistance for the described functionality. ## Prerequisites - Appropriate file access permissions - Required dependencies installed ## Instructions 1. Invoke this skill when the trigger conditions are met 2. Provide necessary context and parameters 3. Review the generated output 4. Apply modifications as needed ## Output The skill produces structured output relevant to the task. ## Error Handling See `{baseDir}/references/errors.md` for comprehensive error handling. ## Examples See `{baseDir}/references/examples.md` for detailed examples. ## Resources - Project documentation - Related skills and commands

Skill file: plugins/ai-ml/ollama-local-ai/skills/ollama-setup/SKILL.md