Every AI subscription you're paying for—ChatGPT, Copilot, Claude, Gemini—is sending your data to someone else's server. With LM Studio, you run powerful models on your own machine: your documents, your conversations, your data. Never leaves. Zero cost per query. Works offline on a plane. This guide shows 11 ways to connect local AI to every tool you already use.
Step 0: LM Studio Setup (5 Minutes)
LM Studio is a desktop app that lets you download, manage, and run open-source LLMs locally. It exposes a local server at http://localhost:1234 using the same OpenAI-compatible API—so anything that works with OpenAI works with LM Studio.
Installation
- Download from lmstudio.ai (Mac, Windows, Linux)
- Install and launch
Recommended Models (2026)
Tip: Use Q4_K_M quantization for the best quality-to-speed trade-off. Avoid Q2 (too degraded) and Q8 (unnecessary for most tasks).
Start the Local Server
- Click Developer tab (left sidebar,
<->icon) - Select your downloaded model in the top dropdown
- Enable CORS toggle (required for browser extensions)
- Click Start Server
Your local AI is now running at http://localhost:1234/v1. This endpoint is fully OpenAI API-compatible — any tool that supports "custom OpenAI endpoint" will work.
1. Obsidian Copilot — Chat with Your Notes
Setup: 10 min | Best for: Knowledge workers, students, researchers
Turn your Obsidian vault into an intelligent knowledge base. Ask questions about your notes, generate ideas, and get summaries — all processed locally, never uploaded.
Setup:
- Obsidian → Settings → Community Plugins → Browse → Search "Copilot" → Install (by Logan Yang)
- Settings → Copilot → Model tab → Add Custom Model:
- Model Name: Your LM Studio model (e.g.,
qwen2.5-7b-instruct) - Provider: LM Studio
- Base URL:
http://localhost:1234/v1 - API Key:
lm-studio(any text works)
- Model Name: Your LM Studio model (e.g.,
- Click Verify → Add Model
Commands that work well:
"Summarize this note and extract 3 key insights""@vault What do I know about productivity systems?""Generate an outline for a post about [topic from my notes]"- Highlight text → right-click → "Ask Copilot"
Pro tip: Enable "Chat Memory" so Copilot remembers context across your session. Combine with the Templater plugin to auto-generate daily note summaries.
2. AnythingLLM — RAG for Any Document Type
Setup: 10 min | Best for: PDF research, contract analysis, studying
Drop in PDFs, Word docs, CSVs, URLs, or entire folders. Ask questions and get answers with source citations. Zero data uploaded — documents stay on your machine.
Setup:
- Download AnythingLLM Desktop from anythingllm.com/desktop
- On first launch → LLM Preference → LM Studio → URL:
http://localhost:1234/v1 - Create a Workspace (think of it as a project folder)
- Drag in your files → "Save and Embed"
Works with: PDF, DOCX, TXT, Markdown, CSV, YouTube URLs (transcribes), GitHub repos (clones and indexes)
3. HARPA AI — Browser Intelligence with Local LLM
Setup: 5 min | Best for: Web research, article summaries, content extraction
Browser extension that puts local AI on every webpage. Summarize articles, extract key points, fill forms, and automate repetitive research tasks — without sending page content to any external service.
Setup:
- Install HARPA AI from harpa.ai (Chrome, Edge, Brave, Arc)
- Click HARPA icon → Settings → AI Backend → Custom Model / OpenAI-Compatible:
- API Endpoint:
http://localhost:1234/v1 - API Key:
lm-studio - Model: your LM Studio model name
- API Endpoint:
Built-in slash commands:
Press Alt+A on any page to open HARPA — it reads the page content automatically.
4. Fabric — Extract Wisdom from Any Content
Setup: 15 min | Best for: YouTube, podcasts, articles, newsletters
Fabric is a command-line tool with 200+ AI "patterns" for processing content. Run a YouTube video through the extract_wisdom pattern and get the speaker's key ideas, quotes, and action items in 30 seconds.
# Install (Go-based, cross-platform)
go install github.com/danielmiessler/fabric@latest
# Or via pip
pip install fabric-ai
# Configure for LM Studio
fabric --setup
# → Choose "LM Studio" or enter custom URL: http://localhost:1234/v1
# Summarize a YouTube video (uses yt-dlp for transcript)
fabric -y "https://youtube.com/watch?v=VIDEO_ID" | fabric --pattern summarize
# Extract all wisdom (quotes, ideas, frameworks, action items)
fabric -y "https://youtube.com/watch?v=VIDEO_ID" | fabric --pattern extract_wisdom
# Process a URL
fabric -u "https://article-url.com" | fabric --pattern extract_main_idea
# Pipe any text
echo "your text here" | fabric --pattern create_study_notes
Most useful patterns:
5. Elephas (Mac) — AI in Every App with One Shortcut
Setup: 10 min | Best for: Mac users who want system-wide AI
Press Cmd + / in any Mac application — Mail, Slack, Notes, Safari, VS Code — and get an AI overlay that can read, write, and rewrite your content without switching apps.
Setup:
- Download from elephas.app (or via Setapp)
- Grant accessibility permissions when prompted
- Preferences → AI Models → LM Studio:
- URL:
http://localhost:1234/v1 - Test connection
- URL:
- Optional: Super Brain → Add your document folders (local RAG)
Use cases:
- In Mail: Select an email →
Cmd + /→ "Write a polite decline" - In Safari: On any page →
Cmd + /→ "Summarize what I'm reading" - In Slack: In a message box →
Cmd + /→ "Make this more concise and professional" - Anywhere:
Cmd + /→ "Translate this to Arabic" or "Fix the grammar"
Windows/Linux alternative: Jan.ai — cross-platform desktop app with similar system-level AI capabilities and built-in OpenAI-compatible server.
6. Voice Notes to Action Items — Hands-Free Capture
Setup: 10 min | Best for: Ideas while walking, commuting, or driving
Record a voice memo → transcribe locally with Whisper → process with LM Studio to extract action items, create structured notes, or summarize meetings. Every step runs on your device.
MacWhisper (Mac, easiest):
- Get MacWhisper from goodsnooze.gumroad.com
- Record or drop in any audio file (MP3, M4A, WAV)
- Select model (Tiny = fast, Base = balanced, Large = most accurate)
- Copy transcript → paste into any local AI chat
Command-line (all platforms):
# Install Whisper
pip install openai-whisper
# Transcribe audio
whisper meeting.mp3 --model small --output_format txt
# Process with LM Studio
curl http://localhost:1234/v1/chat/completions \
-H "Content-Type: application/json" \
-d '{
"messages": [{
"role": "user",
"content": "Extract action items and assign owners from this meeting transcript:\n\nTRANSCRIPT"
}],
"max_tokens": 500
}'
Workflow:
- Voice memo: "Call dentist Friday, review Q2 proposal by Monday, remind team about standup time change"
- Whisper transcription: plain text
- LM Studio output:
- ☐ Call dentist — by Friday
- ☐ Review Q2 proposal — by Monday
- ☐ Send team notice about standup time change
7. LibreChat / Open WebUI — Full ChatGPT-Style Interface
Setup: 10 min | Best for: Anyone who wants a proper chat UI with history
Miss ChatGPT's interface? LibreChat (or the simpler Open WebUI) gives you conversation history, model switching, and markdown rendering — connected to your local LLMs.
Open WebUI (Simpler — recommended to start):
docker run -d -p 3000:8080 \
--add-host=host.docker.internal:host-gateway \
-v open-webui:/app/backend/data \
--name open-webui \
ghcr.io/open-webui/open-webui:main
Open http://localhost:3000 → Settings → Connections → Add LM Studio:
- URL:
http://host.docker.internal:1234/v1 - API Key:
lm-studio
LibreChat (More features — multi-user, plugins, presets):
git clone https://github.com/danny-avila/LibreChat.git
cd LibreChat
cp .env.example .env
docker compose up -d
# Open http://localhost:3080
Both give you:
- Persistent conversation history (saved locally in Docker volume)
- Model switching mid-conversation
- Markdown, code highlighting, image rendering
- System prompt templates
- Conversation search
8. SillyTavern — Creative Writing and Character AI
Setup: 15 min | Best for: Writers, worldbuilders, interactive fiction
A powerful UI for character-driven AI conversations. Create characters with detailed personalities, build story worlds, write interactive fiction — and because it runs locally, there are no content filters that would stop your creative work.
git clone https://github.com/SillyTavern/SillyTavern.git
cd SillyTavern
# Mac/Linux
./start.sh
# Windows
start.bat
Open http://localhost:8000 → API (top menu) → Chat Completion → Custom (OpenAI-compatible):
- Endpoint:
http://localhost:1234/v1 - API Key:
lm-studio - Click Connect (look for green indicator)
Writing use cases:
- Create an "Editor" character with specific writing style preferences for manuscript feedback
- Build a "World Historian" character to keep your fictional world's lore consistent
- Use branching story mode to explore different narrative paths before committing
- Interview your characters to discover their voice before writing dialogue
9. n8n — Automate Workflows with Local AI
Setup: 20 min | Best for: Developers and power users who want AI automation
Connect LM Studio to 400+ apps via n8n's visual workflow builder. Build automations that would normally require cloud AI — but your data never leaves your network.
# Start n8n with Docker
docker run -it --rm \
--name n8n \
-p 5678:5678 \
-v n8n_data:/home/node/.n8n \
docker.n8n.io/n8nio/n8n
# Or npm
npm install -g n8n && n8n start
Open http://localhost:5678 → New Workflow → Add HTTP Request node:
- Method: POST
- URL:
http://host.docker.internal:1234/v1/chat/completions - Body (JSON):
{
"messages": [{ "role": "user", "content": "Summarize: {{ $json.text }}" }],
"max_tokens": 500
}
Example automations:
10. Raycast AI (Mac) — Instant AI from Anywhere
Setup: 5 min | Best for: Mac power users who live in the keyboard
Raycast replaces Spotlight. With an AI extension configured for LM Studio, pressing Cmd + Space gives you instant AI access without opening any app—translate text, fix grammar, get quick answers, run custom AI commands.
Setup:
- Install Raycast from raycast.com
- Open Raycast → Extensions Store → Install "AI Commands" or "OpenAI" extension
- Extension Preferences → Set:
- API Base:
http://localhost:1234/v1 - API Key:
lm-studio - Model: your LM Studio model name
- API Base:
Example custom commands you can create:
Windows/Linux alternative: Flow Launcher with OpenAI plugin → point to LM Studio.
11. Automated Reading List — Never Fall Behind
Setup: 15 min | Best for: Heavy readers, newsletter subscribers, researchers
Automatically process your saved articles, RSS feeds, and newsletters with local AI. Get daily digests that summarize and extract the key insights from everything you saved — ready when you wake up.
n8n-based reading digest workflow:
[Schedule: Daily 7:30am]
↓
[RSS Feed nodes × 3-5 sources]
↓
[Filter: Articles from last 24 hours]
↓
[HTTP Request: LM Studio]
Body: "Summarize in 3 bullet points and rate relevance 1-10: {{article_text}}"
↓
[Filter: Relevance >= 7]
↓
[Email: Daily digest with summaries + links]
Manual workflow (no setup required):
- Save articles to Pocket or Instapaper
- Open AnythingLLM → paste article text
- Ask: "Summarize in 3 bullets and extract 1 actionable insight"
Quick Reference
Total time: ~2 hours for all 11 | Cost: $0/month
Troubleshooting
Start Here: Which Integration First?
Key Takeaways
http://localhost:1234/v1 and use any string as the API key.localhost refers to the container, not your host machine. Use host.docker.internal:1234 when configuring n8n, LibreChat, Open WebUI, or any Docker-based tool.