Getting Started¶
New to Loom? Start with Concepts to understand the mental model.
Quick Start — Try a Worker (No Data Files Needed)¶
The fastest way to see Loom work. Paste any text and get structured output.
# 1. Install
pip install loom-ai[workshop]
# 2. Configure (interactive wizard — detects Ollama, prompts for API keys)
loom setup
# 3. Open the Workshop web UI
loom workshop
Open your browser at http://localhost:8080. You'll see a list of workers.
Click any worker → Test → paste text into the input box → click Run.
Try these first:
- summarizer — paste an article, get a structured summary with key points
- classifier — paste text and a list of categories, get a classification with confidence score
- extractor — paste a contract or report, define fields to extract (names, dates, amounts)
- qa — paste a passage and ask a question, get an answer with source citations
All four work with Ollama (free, local). If you don't have Ollama, loom setup
will prompt for an Anthropic API key instead.
See Workers Reference for full I/O schemas and the Workshop Tour for what each screen does.
RAG Quickstart¶
If you have Telegram JSON exports, Loom can ingest, chunk, embed, and search them:
# 1. Install with RAG support
pip install loom-ai[rag]
# 2. Configure
loom setup
# 3. Ingest Telegram exports
loom rag ingest /path/to/telegram/exports/*.json
# 4. Search
loom rag search "earthquake damage reports"
# 5. Open the dashboard
loom rag serve
The loom setup wizard detects Ollama, prompts for API keys, and writes
~/.loom/config.yaml. All settings can be overridden via environment
variables or CLI flags. See Configuration for details.
Build Your First Worker¶
Once you've tested the shipped workers, build your own — or chain existing ones into a pipeline:
| Worker | What it does | Tier |
|---|---|---|
summarizer |
Compress text into structured summary with key points | local |
classifier |
Assign text to categories with confidence | local |
extractor |
Pull structured fields from unstructured text | standard |
translator |
Translate between languages with auto-detection | local |
qa |
Answer questions from provided context with citations | local |
reviewer |
Review content quality against configurable criteria | standard |
# Create your own worker interactively (generates YAML for you)
loom new worker
# Chain workers into a pipeline
loom new pipeline
# Validate configs (no infrastructure needed)
loom validate configs/workers/*.yaml
# Test in the Workshop
loom workshop
No NATS or infrastructure needed. The Workshop calls LLM backends directly.
That's it for basic usage. Everything below is for when you need the full distributed infrastructure (multi-user, scaling, continuous processing).
Prerequisites (Full Infrastructure)¶
- Python 3.11+
- uv package manager (recommended for development)
- At least one LLM backend (Ollama recommended to start)
- NATS and Valkey for full infrastructure (not needed for Quick Start)
1. Install Python Dependencies¶
Loom has optional extras for integrations:
uv sync --extra rag # RAG pipeline (DuckDB + Ollama embeddings)
uv sync --extra lancedb # LanceDB vector store (ANN search, alternative to DuckDB)
uv sync --extra telegram # Live Telegram channel capture via Telethon
uv sync --extra duckdb # DuckDB tools and query backends
uv sync --extra redis # Redis/Valkey-backed checkpoint store
uv sync --extra local # Ollama client for local models
uv sync --extra workshop # Worker Workshop web UI (FastAPI, Jinja2, DuckDB)
uv sync --extra mcp # MCP gateway (Model Context Protocol)
uv sync --extra otel # OpenTelemetry distributed tracing
uv sync --extra tui # Terminal dashboard (Textual)
uv sync --extra mdns # mDNS/Bonjour service discovery on LAN
uv sync --extra scheduler # Cron expression parsing (croniter)
uv sync --extra eval # DeepEval LLM output quality evaluation
uv sync --extra docs # MkDocs-Material API documentation generation
2. Configure LLM Backends¶
The easiest path — run the setup wizard:
This auto-detects Ollama, prompts for API keys, and writes ~/.loom/config.yaml.
Or configure manually via environment variables:
# Option A: Ollama (free, local, recommended to start)
brew install ollama
ollama serve &
ollama pull llama3.2:3b
export OLLAMA_URL=http://localhost:11434
# Option B: Anthropic API
export ANTHROPIC_API_KEY=sk-ant-...
# Option C: Any OpenAI-compatible API (vLLM, LiteLLM, llama.cpp server)
# See OpenAICompatibleBackend in src/loom/worker/backends.py
Settings resolution priority: CLI flags > environment variables > ~/.loom/config.yaml > built-in defaults. See Configuration for the full reference.
3. Run the Unit Tests (No Infrastructure Needed)¶
This runs all unit tests without needing NATS or Valkey.
4. Set Up Infrastructure (NATS + Valkey)¶
Skip this if you only need the RAG pipeline or Workshop. The steps below are for the full distributed actor mesh.
# Install via Homebrew (Mac) or use Docker
brew install nats-server valkey
# Start them
nats-server &
valkey-server &
Or with Docker:
docker run -d --name nats -p 4222:4222 nats:2.10-alpine
docker run -d --name valkey -p 6379:6379 valkey/valkey:8-alpine
5. Start the Router, Orchestrator, and a Worker¶
# Terminal 1: Start the router
uv run loom router --nats-url nats://localhost:4222
# Terminal 2: Start the orchestrator
uv run loom orchestrator --config configs/orchestrators/default.yaml --nats-url nats://localhost:4222
# Terminal 3: Start a summarizer worker
uv run loom worker --config configs/workers/summarizer.yaml --tier local --nats-url nats://localhost:4222
6. Submit a Test Task¶
# Terminal 4: Send a task through the system
uv run loom submit "Summarize the main points of the UN Charter preamble" --nats-url nats://localhost:4222
Monitor what's happening:
# Option 1: TUI dashboard (recommended)
uv sync --extra tui
uv run loom ui --nats-url nats://localhost:4222
# Option 2: NATS CLI (raw messages)
brew tap nats-io/nats-tools && brew install nats-io/nats-tools/nats
nats sub "loom.>" --server=nats://localhost:4222
7. Create Your Own Worker¶
Edit the file — define a system prompt, input/output schemas, and default tier. Then start it:
Or test it without NATS using the Workshop:
What's Next¶
| Goal | Guide |
|---|---|
| Understand the mental model | Concepts |
| How Loom compares to other tools | Why Loom? |
| Tour the Workshop UI | Workshop Tour |
| Build workers, pipelines, tools | Building Workflows |
| Set up RAG analysis pipeline | RAG How-To |
| Configure settings and API keys | Configuration |
| Full CLI command reference | CLI Reference |
| Test and evaluate workers | Workshop |
| Deploy to production | Local Deployment / Kubernetes |
| Debug issues | Troubleshooting |
| Understand design decisions | Design Invariants |
For architecture details, see Architecture. For the full CLI reference, see CLI Reference.