# Install
pip install llmswap
# or Homebrew
brew tap llmswap/tap && brew install llmswap
# Create your first workspacecd~/my-project
llmswap workspace init
# Chat with AI that remembers everything
llmswap chat "Help me with Flask routing"# AI has full project context + all past learnings!
🆕 Use Any Model from Any Provider! New model just launched? Use it immediately. llmswap's pass-through architecture means GPT-5, Claude Opus 4, Gemini 2.5 Pro work the day they release. Currently supports 10 providers (OpenAI, Anthropic, Gemini, Cohere, Perplexity, IBM watsonx, Groq, Ollama, xAI Grok, Sarvam AI).
✅ Battle-Tested with LMArena Top Models: All 10 providers tested and validated with top-rated models from LMArena leaderboard. From Grok-4 (xAI's flagship) to Claude Sonnet 4.5 (best coding model) to Gemini 2.0 Flash Exp - every model in our defaults is production-validated and arena-tested for real-world use.
The First AI Tool with Project Memory & Learning Journals - v5.1.0 introduces revolutionary workspace system that remembers your learning journey across projects. Build apps without vendor lock-in (SDK) or use from terminal (CLI). Works with your existing subscriptions: Claude, OpenAI, Gemini, Cohere, Perplexity, IBM watsonx, Groq, Ollama, xAI Grok, Sarvam AI (10 providers). Use any model from your provider - even ones released tomorrow. Pass-through architecture means GPT-5, Gemini 2.5 Pro, Claude Opus 4? They work the day they launch.
🎯 Solve These Common Problems:
❌ "I need multiple second brains for different aspects of my life" 🆕
❌ "AI strays over time, I need to re-steer it constantly" 🆕
❌ "I keep explaining the same context to AI over and over"
❌ "AI forgets what I learned yesterday"
❌ "I lose track of architecture decisions across projects"
❌ "Context switching between projects is exhausting"
❌ "I want AI to understand my specific codebase, not generic answers"
✅ llmswap v5.1.0 Solves All These:
✅ Multiple independent "second brains" per project/life aspect 🆕
✅ Persistent context prevents AI from straying 🆕
✅ Per-project workspaces that persist context across sessions
✅ Auto-tracked learning journals - never forget what you learned
✅ Architecture decision logs - all your technical decisions documented
✅ Zero context switching - AI loads the right project automatically
✅ Project-aware AI - mentor understands YOUR specific tech stack
⚡ CLI tool - Terminal interface that works with any subscription (bonus!)
Why llmswap?
🔓 No vendor lock-in - Switch providers with 1 line of code (SDK) or 1 command (CLI)
🎓 Teaching-first AI - Eklavya mentorship system (guru, coach, socrates personas)
💰 Cost optimizer - Automatic caching saves 50-90% on API calls
🔧 For apps AND terminal - One tool, two ways to use it
v5.1.0: Revolutionary AI mentorship with project memory, workspace-aware context, auto-tracked learning journals, and persistent mentor relationships. The first AI tool that truly remembers your learning journey across projects.
NEW in v5.1.0:
🧠 Workspace Memory - Per-project context that persists across sessions
📚 Auto-Learning Journal - Automatically tracks what you learn in each project
🎯 Context-Aware Mentorship - AI mentor understands your project and past learnings
🔄 Cross-Project Intelligence - Learn patterns from one project, apply to another
💡 Proactive Learning - AI suggests next topics based on your progress
🗂️ Project Knowledge Base - Custom prompt library per workspace
🧠 Finally: An Elegant Solution for Multiple Second Brains
The Problem Industry Leaders Can't Solve:
"I still haven't found an elegant solution to the fact that I need several second brains for the various aspects of my life, each with different styles and contexts." - Industry feedback
The llmswap Solution: Workspace System
Each aspect of your life gets its own "brain" with independent memory:
💼 Work Projects - ~/work/api-platform - Enterprise patterns, team conventions
📚 Learning - ~/learning/rust - Your learning journey, struggles, progress
🚀 Side Projects - ~/personal/automation - Personal preferences, experiments
🌐 Open Source - ~/oss/django - Community patterns, contribution history
What Makes It "Elegant":
✅ Zero configuration - just cd to project directory
✅ Auto-switching - AI loads the right "brain" automatically
✅ No context bleed - work knowledge stays separate from personal
✅ Persistent memory - each brain remembers across sessions
✅ Independent personas - different teaching style per project if you want
Stop Re-Explaining Context. Start Building.
🎯 Transform AI Into Your Personal Mentor with Project Memory
Inspired by Eklavya - the legendary self-taught archer who learned from dedication and the right guidance - llmswap transforms any AI provider into a personalized mentor that adapts to your learning style and remembers your journey.
The Challenge: Developers struggle to learn effectively from AI because:
🔴 Responses are generic, lack personality, and don't adapt to individual needs
🔴 AI loses context between sessions - you repeat the same explanations
🔴 No learning history - AI doesn't know what you already learned
🔴 Project context is lost - AI doesn't understand your codebase
Our Solution v5.1.0: Choose your mentorship style, initialize a workspace, and ANY AI provider becomes your personalized guide that remembers everything:
# 🆕 v5.1.0: Initialize workspace for your projectcd~/my-flask-app
llmswap workspace init
# Creates .llmswap/ with context.md, learnings.md, decisions.md# Now your AI mentor KNOWS your project
llmswap chat --mentor guru --alias "Guruji"# Mentor has full context: your tech stack, past learnings, decisions made# 🆕 Auto-tracked learning journal# Every conversation automatically saves key learnings
llmswap workspace journal
# View everything you've learned in this project# 🆕 Architecture decision log
llmswap workspace decisions
# See all technical decisions documented automatically# View all your workspaces
llmswap workspace list
# Get wisdom and deep insights from a patient teacher
llmswap chat --mentor guru --alias "Guruji"# High-energy motivation when you're stuck
llmswap ask "How do I debug this?" --mentor coach
# Collaborative peer learning for exploring ideas
llmswap chat --mentor friend --alias "CodeBuddy"# Question-based learning for critical thinking
llmswap ask "Explain REST APIs" --mentor socrates
# 🆕 Use Claude Sonnet 4.5 - Best coding model
llmswap chat --provider anthropic --model claude-sonnet-4-5
# Or set as default in config for all queries
🔄 Rotate Personas to Expose Blind Spots
Industry Insight: "Rotate personas: mentor, skeptic, investor, end-user. Each lens exposes blind spots differently."
Use Case: Reviewing API Design
# Round 1: Long-term wisdom
llmswap chat --mentor guru "Design API for multi-tenant SaaS"# Catches: scalability, technical debt, maintenance# Round 2: Critical questions
llmswap chat --mentor socrates "Review this API design"# Catches: assumptions, alternatives, edge cases# Round 3: Practical execution
llmswap chat --mentor coach "What's the fastest path to v1?"# Catches: over-engineering, paralysis by analysis
Same project context. Different perspectives. Complete understanding.
What Makes v5.1.0 Revolutionary:
🧠 Works with ANY provider - Transform Claude, GPT-4, or Gemini into your mentor
🎭 6 Teaching Personas - Guru, Coach, Friend, Socrates, Professor, Tutor
📊 Project Memory - Per-project context that persists across sessions ⭐ NEW
📚 Auto-Learning Journal - Automatically tracks what you learn ⭐ NEW
📖 Decision Tracking - Documents architecture decisions ⭐ NEW
🎓 Age-Appropriate - Explanations tailored to your level (--age 10, --age 25, etc.)
💰 Cost Optimized - Use cheaper providers for learning, premium for complex problems
🔄 Workspace Detection - Automatically loads project context ⭐ NEW
Traditional AI tools give you answers. llmswap v5.1.0 gives you a personalized learning journey that REMEMBERS.
🏆 Production-Validated with LMArena Top Models
Every model in llmswap's defaults comes from LMArena's top performers:
All 10 providers ship with carefully selected default models based on LMArena rankings and real-world production testing. We track arena performance and update defaults to ensure you're always using validated, battle-tested models.
Provider
Default Model
Arena Status
Why We Chose It
Anthropic
claude-sonnet-4-5
#1 Coding
Best coding model in the world (Sept 2025)
xAI
grok-4-0709
Top 5 Overall
Advanced reasoning, real-time data access
Gemini
gemini-2.0-flash-exp
Top 10
Lightning-fast, multimodal, cutting-edge
OpenAI
gpt-4o-mini
Cost Leader
Best price/performance ratio
Cohere
command-r-08-2024
Top RAG
Enterprise-grade retrieval-augmented generation
Perplexity
sonar
Web Search
Real-time web-connected AI with citations
Groq
llama-3.1-8b-instant
Speed King
840+ tokens/second ultra-fast inference
Sarvam
sarvam-m
Multilingual
24B params, best for 10 Indian languages
Watsonx
granite-3-8b-instruct
Enterprise
IBM's production-grade AI for business
Ollama
granite-code:8b
Local AI
Privacy-first, runs on your hardware
✅ Battle-tested with real API calls - Every provider validated in production, not simulated tests.
✅ Weekly model updates - We monitor LMArena rankings and deprecation notices to keep defaults current.
✅ Zero lock-in - Don't like our defaults? Override with any model: LLMClient(model="gpt-5") or llmswap config set provider.models.openai gpt-5
🔓 Use Any Model Your Provider Supports (Zero-Wait Model Support)
Here's something cool: llmswap doesn't restrict which models you can use. When GPT-5 or Gemini 2.5 Pro drops tomorrow, you can start using it immediately. No waiting for us to update anything.
How? We use pass-through architecture. Whatever model name you pass goes directly to your provider's API. We don't gatekeep.
# Use any OpenAI model (even ones that don't exist yet)
llmswap chat --provider openai --model gpt-5
llmswap chat --provider openai --model o3-mini
# Use any Anthropic model
llmswap chat --provider anthropic --model claude-opus-4
llmswap chat --provider anthropic --model claude-sonnet-4-5
# Use any Gemini model
llmswap chat --provider gemini --model gemini-2-5-pro
llmswap chat --provider gemini --model gemini-ultra-2
# Set as default so you don't have to type it every time
llmswap config set provider.models.openai gpt-5
llmswap config set provider.models.anthropic claude-opus-4
fromllmswapimportLLMClient# Use whatever model your provider offersclient=LLMClient(provider="openai", model="gpt-5")
client=LLMClient(provider="anthropic", model="claude-opus-4")
client=LLMClient(provider="gemini", model="gemini-2-5-pro")
# Model just released? Use it right nowclient=LLMClient(provider="openai", model="gpt-6") # works!
The point: You're not limited to what we've documented. If your provider supports it, llmswap supports it.
🆚 llmswap vs Single-Provider Tools
For Python Developers Building Apps:
Your Need
Single-Provider SDKs
llmswap SDK
Build chatbot/app
Import openai library (locked in)
Import llmswap (works with any provider)
Switch providers
Rewrite all API calls
Change 1 line: provider="anthropic"
Try different models
Sign up, new SDK, refactor code
Just change config, same code
Use new models
Wait for SDK update
Works immediately (pass-through)
Cost optimization
Manual implementation
Built-in caching (50-90% savings)
Use multiple providers
Maintain separate codebases
One codebase, switch dynamically
For Developers Using Terminal:
Your Need
Vendor CLIs
llmswap CLI
Have Claude subscription
Install Claude Code (Claude only)
Use llmswap (works with Claude)
Have OpenAI subscription
Build your own scripts
Use llmswap (works with OpenAI)
Have multiple subscriptions
Install 3+ different CLIs
One CLI for all subscriptions
New model launches
Wait for CLI update
Use it same day (pass-through)
Want AI to teach you
Not available
Built-in Eklavya mentorship
Switch providers mid-chat
Can't - locked in
/switch anthropic command
The Bottom Line:
Building an app? Use llmswap SDK - no vendor lock-in
Using terminal? Use llmswap CLI - works with your existing subscriptions
Both? Perfect - it's the same tool!
# 🆕 NEW v5.1.0: Workspace System - Project Memory That Persists
llmswap workspace init
# Creates .llmswap/ directory with:# - workspace.json (project metadata)# - context.md (editable project description)# - learnings.md (auto-tracked learning journal)# - decisions.md (architecture decision log)
llmswap workspace list # View all your workspaces
llmswap workspace info # Show current workspace statistics
llmswap workspace journal # View learning journal
llmswap workspace decisions # View decision log
llmswap workspace context # Edit project context# 🆕 NEW v5.1.0: Context-Aware Mentorship# AI mentor automatically loads project context, past learnings, and decisions
llmswap chat
# Mentor knows: your tech stack, what you've learned, decisions made# 🆕 NEW v5.0: Age-Appropriate AI Explanations
llmswap ask "What is Docker?" --age 10
# Output: "Docker is like a magic lunch box! 🥪 When your mom packs..."
llmswap ask "What is blockchain?" --audience "business owner"# Output: "Think of blockchain like your business ledger system..."# 🆕 NEW v5.0: Teaching Personas & Personalization
llmswap ask "Explain Python classes" --teach --mentor developer --alias "Sarah"# Output: "[Sarah - Senior Developer]: Here's how we handle classes in production..."# 🆕 NEW v5.0: Conversational Chat with Provider Switching
llmswap chat --age 25 --mentor tutor
# In chat: /switch anthropic # Switch mid-conversation# In chat: /provider # See current provider# Commands: /help, /switch, /clear, /stats, /quit# 🆕 NEW v5.0: Provider Management & Configuration
llmswap providers # View all providers and their status
llmswap config set provider.models.cohere command-r-plus-08-2024
llmswap config set provider.default anthropic
llmswap config show
# Code Generation (GitHub Copilot CLI Alternative)
llmswap generate "sort files by size in reverse order"# Output: du -sh * | sort -hr
llmswap generate "Python function to read JSON with error handling" --language python
# Output: Complete Python function with try/catch blocks# Advanced Log Analysis with AI
llmswap logs --analyze /var/log/app.log --since "2h ago"
llmswap logs --request-id REQ-12345 --correlate
# Code Review & Debugging
llmswap review app.py --focus security
llmswap debug --error "IndexError: list index out of range"
# ❌ Problem: Vendor Lock-inimportopenai# Locked to OpenAI foreverclient=openai.Client(api_key="sk-...")
response=client.chat.completions.create(
model="gpt-4",
messages=[{"role": "user", "content": "Hello"}]
)
# To switch to Claude? Rewrite everything.# ✅ Solution: llmswap SDK - Universal InterfacefromllmswapimportLLMClient# Works with any provider you're subscribed toclient=LLMClient() # Auto-detects from env varsresponse=client.query("Hello")
# Want Claude instead? Just change provider:client=LLMClient(provider="anthropic") # That's it!# Want to try Gemini? client=LLMClient(provider="gemini") # Same code, different provider# Built-in cost optimization:# - Automatic response caching (50-90% savings)# - Provider cost comparison# - Smart provider selection based on query type
🆕 v5.1.0: Workspace System - Real-World Scenarios
🎯 Scenario 1: New Developer Learning Flask
Problem: Junior developer learning Flask keeps asking AI the same questions because AI forgets previous conversations.
Solution with llmswap v5.1.0:
cd~/my-first-flask-app
llmswap workspace init --name "Learning Flask"# Day 1: Learn about routing
llmswap chat --mentor professor
"How do Flask routes work?"# AI explains. Learning auto-saved to learnings.md# Day 2: Same workspace, AI remembers!
llmswap chat
"Can I use decorators for authentication?"# AI response: "Building on what you learned about routes yesterday..."# No need to re-explain basics!# View your learning journey
llmswap workspace journal
# See: Day 1 - Routes, Day 2 - Authentication, etc.
Result: 60% faster learning because AI builds on previous knowledge instead of repeating basics.
🏢 Scenario 2: Team Onboarding on Legacy Project
Problem: New team member joins 2-year-old codebase. Spends weeks understanding architecture decisions.
Solution with llmswap v5.1.0:
cd~/legacy-ecommerce-app
llmswap workspace init
# Edit context.md with project overview
llmswap workspace context
# Add: Tech stack, key components, known issues# Ask questions - AI has full context
llmswap ask "Why did we choose MongoDB over PostgreSQL?" --mentor guru
# AI suggests checking decisions.md# If documented: "According to your decision log from 2023-05..."# If not: AI helps document it now
llmswap workspace decisions
# See all past architectural decisions in one place
Result: Onboarding time reduced from 3 weeks to 1 week.
Problem: Freelancer switches between 5 client projects daily. Context switching is exhausting.
Solution with llmswap v5.1.0:
# Morning: Client A's React projectcd~/client-a-dashboard
llmswap chat
# AI loads: React patterns you learned, components built, state management decisions# Afternoon: Client B's Python APIcd~/client-b-api
llmswap chat
# AI switches context: Python best practices, API design decisions, database schema# List all projects
llmswap workspace list
# See: 5 workspaces, each with independent context and learnings# Each workspace has separate:# - Learning journal (React patterns vs Python patterns)# - Decision log (frontend vs backend decisions)# - Project context (different tech stacks)
Result: Zero mental overhead for context switching. AI handles it automatically.
🎓 Scenario 4: Learning Journey Across Technologies
Problem: Developer learning full-stack wants to track progress across frontend, backend, DevOps.
Solution with llmswap v5.1.0:
# Frontend projectcd~/react-app
llmswap workspace init --name "React Learning"
llmswap chat --mentor tutor
# Learn: Hooks, State, Components# All auto-tracked in learnings.md# Backend projectcd~/python-api
llmswap workspace init --name "Python API"
llmswap chat --mentor tutor
# Learn: FastAPI, SQLAlchemy, Testing# Separate learning journal# View all learning across projects
llmswap workspace list
# See progress in each area# Each workspace shows:# - Total queries# - Learnings count# - Last accessed
Result: Complete visibility into learning journey across all technologies.
🚀 Scenario 5: Open Source Contributor
Problem: Contributing to 3 different OSS projects. Each has different conventions, patterns, testing approaches.
Solution with llmswap v5.1.0:
# Project 1: Djangocd~/django-oss
llmswap workspace init
# Document in context.md: Coding style, PR process, testing patterns# Project 2: FastAPIcd~/fastapi-oss
llmswap workspace init
# Different conventions, different patterns# Ask project-specific questionscd~/django-oss
llmswap ask "How should I write tests here?"# AI knows: This project uses pytest with Django TestCasecd~/fastapi-oss
llmswap ask "How should I write tests here?"# AI knows: This project uses pytest with async fixtures# Each workspace maintains separate:# - Testing patterns learned# - Code review feedback# - Architecture understanding
Result: Contribute confidently to multiple projects without mixing up conventions.
💡 Real-World Use Cases (v5.0 Features)
📚 Learning & Skill Development
# Junior developer learning system design
llmswap chat --mentor professor --alias "Dr. Smith""Explain microservices architecture"# Response adapts to teaching style with examples and deep explanations# Career switcher exploring new domains
llmswap ask "What is Kubernetes?" --age 25 --audience "beginner"# Age-appropriate explanation without assuming prior DevOps knowledge
🚀 Professional Development
# Get different perspectives on technical decisions
llmswap ask "Should I use MongoDB or PostgreSQL?" --mentor guru
# Wisdom-based response considering trade-offs and bigger picture
llmswap ask "Should I use MongoDB or PostgreSQL?" --mentor developer
# Practical engineering perspective with production experience
# Explain technical concepts to non-technical stakeholders
llmswap ask "Explain our API architecture" --audience "business owner"# Business-friendly explanation without jargon# Onboard new team members
llmswap chat --mentor friend --alias "TeamBuddy"# Collaborative, approachable learning environment
🎭 Same Question, Different Perspectives
Watch how the SAME question gets tailored responses based on your role:
# As a Developer: Implementation focus
llmswap ask "Explain Model Context Protocol (MCP)" --audience developer
Response: "MCP is a standardized protocol for connecting AI models to external data sources and tools. Unlike traditional chatbots that bundle everything into one monolithic system, MCP uses a client-server architecture:- Server: Exposes tools, resources, and prompts via standard interface- Client: Your AI application that consumes these capabilities- Transport: JSON-RPC over stdio/HTTPKey difference from older chatbots:OLD: Hardcoded integrations, tight coupling, monolithic deploymentNEW: Pluggable architecture, loose coupling, independent tool serversCode example:from mcp import Serverserver = Server("my-tools")@server.tool()def get_weather(city: str): ..."
---
# As an Architect: System design focus
llmswap ask "Explain Model Context Protocol (MCP)" --audience architect
Response: "MCP introduces a service-oriented architecture for AI systems, addressing key architectural concerns:**System Architecture:**┌─────────────┐ ┌──────────────┐ ┌─────────────┐│ AI Client │◄────►│ MCP Protocol │◄────►│ Tool Server ││ (Claude App)│ │ (JSON-RPC) │ │ (Resources) │└─────────────┘ └──────────────┘ └─────────────┘**Design Patterns:**- Service Discovery: Dynamic capability negotiation- Loose Coupling: Clients/servers evolve independently - Extensibility: Add new tools without modifying core- Standardization: Uniform interface across tools**vs Traditional Chatbots:**- Monolithic → Microservices for AI tools- Static → Dynamic capability loading- Vendor lock-in → Provider-agnostic standard**Scalability:** Deploy tool servers independently, horizontal scaling per service"
---
# As a Business Leader: ROI and strategy focus
llmswap ask "Explain Model Context Protocol (MCP)" --audience "business leader"
Response: "Model Context Protocol (MCP) is like moving from custom-built integrations to standardized APIs - but for AI systems.**Business Impact:**1. **Reduced Integration Costs** - OLD: $50K-$100K per custom AI integration - NEW: Plug-and-play tools, 80% cost reduction - Time to market: Weeks instead of months2. **Vendor Independence** - Not locked to one AI provider - Switch between Claude, GPT-4, Gemini seamlessly - Negotiating power with vendors3. **Reusable Components** - Build once, use across multiple AI applications - Internal tools become valuable assets - Ecosystem of third-party MCP tools4. **Competitive Advantage** - Deploy AI features 10x faster than competitors - Respond quickly to market changes - Lower total cost of ownership**vs Traditional Chatbots:** Traditional bots were closed systems requiring custom development for each feature. MCP creates an open ecosystem where your team (or partners) can extend capabilities without vendor dependencies.**ROI Example:** A company saved $200K in integration costs and reduced deployment time from 6 months to 3 weeks using MCP architecture."
The Power: Same command, same AI provider, completely different responses tailored to what each role needs to know.
⚠️ Privacy & Usage Disclaimers
llmswap is an interface tool only - all AI interactions and billing are handled directly by your chosen providers.
🔒 Context Privacy & Provider Switching
No context sharing: When switching providers mid-conversation, NO conversation history is transferred between providers
Fresh start: Each provider switch starts a completely new conversation thread
Legal compliance: This protects your privacy and complies with provider Terms of Service
Your control: You decide what each provider sees in their separate conversations
Cost estimates (~$X.XX estimated) are approximate based on public pricing and common tokenizers. Actual costs may differ. You are responsible for all provider costs and billing.
Legal Notice: llmswap provides estimates and interface functionality for convenience only. We are not responsible for billing differences, provider charges, pricing changes, or data handling by individual providers. Always verify costs with your provider's billing dashboard.
⚡ Get Started in 30 Seconds
🍺 Homebrew (Recommended - macOS/Linux)
# Add our tap and install
brew tap llmswap/tap
brew install llmswap
# Ready to use immediately!
llmswap --help
Why Homebrew? No virtualenv needed, global access, automatic dependency management, and easier updates.
🚀 Quick Start with Workspaces (v5.1.0)
Complete Beginner's Guide - 3 Steps:
Step 1: Install llmswap
pip install llmswap
# or
brew install llmswap
Step 2: Set up API key (one provider is enough)
export ANTHROPIC_API_KEY="your-key-here"# For Claude# orexport OPENAI_API_KEY="your-key-here"# For GPT-4# or any other provider
Step 3: Initialize workspace in your project
cd~/my-project
llmswap workspace init
# Start chatting - AI has full project context!
llmswap chat --mentor guru
# Ask anything about your project# Learnings automatically tracked# Decisions automatically documented# View your learning journey
llmswap workspace journal
That's it! Your AI mentor now remembers everything about your project. 🎉
Without Workspace (Classic Mode)
# Works exactly like v5.0 - no workspace needed
llmswap ask "How do I deploy a Flask app?"
llmswap chat --mentor tutor
llmswap generate "Python function to read CSV"
📋 Quick Reference - v5.1.0 Commands
🆕 Workspace Commands (NEW!)
Command
Description
Example
llmswap workspace init
Initialize workspace in current directory
Creates .llmswap/ with context, learnings, decisions
llmswap workspace init --name
Initialize with custom project name
llmswap workspace init --name "My API"
llmswap workspace info
Show current workspace statistics
Displays queries, learnings, decisions count
llmswap workspace list
List all registered workspaces
Shows all projects with llmswap workspaces
llmswap workspace journal
View learning journal
See everything you've learned
llmswap workspace decisions
View architecture decision log
See all technical decisions
llmswap workspace context
Edit project context
Opens context.md in default editor
Provider & Config Commands (v5.0)
Command
Description
Example
llmswap providers
View all providers and their status
Shows configured/missing API keys
llmswap config set provider.models.<provider> <model>
Update default model for any provider
llmswap config set provider.models.cohere command-r-plus-08-2024
llmswap config list
View current configuration
Shows all settings and models
/switch (in chat)
Switch providers mid-conversation
Privacy-compliant provider switching
/provider (in chat)
Show current provider and available options
Quick status check
🔧 First-Time Setup (v5.0.4 NEW!)
# First run automatically creates ~/.llmswap/config.yaml with defaults
llmswap ask "Hello world"# Output: 🔧 Creating config file at ~/.llmswap/config.yaml# ✅ Default configuration created# View all providers and their configuration status
llmswap providers
# Set up your API keys and start usingexport ANTHROPIC_API_KEY="your-key-here"
llmswap ask "Explain Docker in simple terms"
💡 Smart Defaults: llmswap comes pre-configured with sensible model defaults for all 8 providers. No configuration needed to get started!
fromllmswapimportLLMClient# Works with any provider you have configuredclient=LLMClient() # Auto-detects from environment/configresponse=client.query("Explain quantum computing in 50 words")
print(response.content)
🎯 Why llmswap v5.1.0 for AI Development?
Feature
llmswap v5.1.0
Claude Code
Cursor AI
Aider
LangChain
Direct APIs
Project Memory
✅ Workspace system
❌ No memory
❌ No memory
❌ No memory
❌ Manual
❌ None
Learning Journal
✅ Auto-tracked
❌ Not available
❌ Not available
❌ Not available
❌ Manual
❌ None
Context Awareness
✅ Project-specific
❌ Generic
❌ Generic
❌ Generic
❌ Manual
❌ None
AI Providers
✅ 8+ providers, instant switch
❌ Claude only
❌ Few providers
❌ OpenAI only
⚠️ 50+ complex setup
❌ 1 per codebase
Conversational Mode
✅ Provider-native, all
✅ Claude only
✅ Limited
❌ Not available
⚠️ Manual setup
❌ Not available
Memory Usage
✅ 99% reduction
⚠️ Local storage
⚠️ Local storage
⚠️ Local storage
❌ Heavy framework
❌ Manual
Configuration
✅ Git-like, shareable
⚠️ Basic settings
⚠️ Basic settings
⚠️ Basic settings
❌ Complex files
❌ None
Cost Analytics
✅ Real-time tracking
❌ No cost info
❌ No cost info
❌ No cost info
❌ External tools
❌ Manual
Provider Switching
✅ Mid-conversation
❌ Locked to Claude
⚠️ Limited
❌ Locked to OpenAI
❌ Restart required
❌ New session
Workspace System
✅ Per-project context
❌ Not available
❌ Not available
❌ Not available
❌ Not available
❌ None
CLI Commands
✅ 15+ powerful tools
⚠️ Limited
❌ IDE only
⚠️ Limited
❌ Separate packages
❌ None
SDK + CLI
✅ Both included
❌ CLI only
❌ IDE only
❌ CLI only
✅ SDK only
⚠️ SDK only
Teaching Personas
✅ 6 mentors
❌ Not available
❌ Not available
❌ Not available
❌ Not available
❌ None
Open Source
✅ 100% MIT licensed
❌ Proprietary
❌ Proprietary
✅ Open source
✅ Open source
⚠️ Varies
Key Differentiators for v5.1.0:
🧠 Only AI tool with persistent project memory - Never repeat context again
📚 Automatic learning journals - Track your progress without manual work
🎯 Workspace-aware mentorship - AI understands your specific project
🔄 Zero context switching overhead - Change projects, AI adapts automatically
💡 Learning extraction - AI summarizes key takeaways from conversations
🚀 Three Ways to Use llmswap:
📚 1. Python Library/SDK
fromllmswapimportLLMClientclient=LLMClient() # Import into any codebaseresponse=client.query("Analyze this data")
⚡ 2. CLI Tools
llmswap generate "sort files by size"# GitHub Copilot alternative
llmswap generate "Python function to read JSON"# Multi-language code generation
llmswap ask "Debug this error"# Terminal AI assistant
llmswap costs # Cost optimization insights
📊 3. Enterprise Analytics
stats=client.get_usage_stats() # Track AI spendcomparison=client.get_provider_comparison() # Compare costs
🆕 Revolutionary Workspace & Memory Features
🧠 Workspace System: Per-project memory with .llmswap/ directories (inspired by .git/)
📚 Auto-Learning Journal: AI automatically tracks what you learn in learnings.md
📖 Architecture Decision Log: Document technical decisions in decisions.md
🎯 Context-Aware Mentorship: AI mentor loads project context, past learnings, and decisions
🔍 Workspace Detection: Automatically finds .llmswap/ in current or parent directories
🗂️ Project Knowledge Base: Editable context.md for project-specific information
📊 Workspace Statistics: Track queries, learnings, and decisions per project
🌐 Global Workspace Registry: Manage all workspaces from ~/.llmswap/workspaces/registry.json
💡 Learning Extraction: Uses fast AI (Groq) to extract key learnings from conversations
🔄 Workspace Switching: Change directories, AI automatically loads different context
Teaching & Conversational Features (v5.0)
🎓 Age-Appropriate AI: First CLI with age-targeted explanations (--age 10, --audience "teacher")
🧑🏫 Teaching Personas: 6 AI mentors (teacher, developer, tutor, professor, mentor, buddy)
👤 Personalized Aliases: Custom AI names (--alias "Sarah" for your personal tutor)
💬 Multi-Provider Chat: Provider-native conversational mode with mid-chat switching
🧠 Zero Local Storage: 99% memory reduction, all context at provider level
📊 Session Analytics: Real-time cost and token tracking
Provider & Model Flexibility
🔓 Pass-Through Architecture: Use ANY model from your provider - GPT-5, Claude Opus 4, Gemini 2.5 Pro work immediately
⚡ Zero-Wait Updates: New model released? Use it the same day, no llmswap update needed
🌐 10 Providers Currently: OpenAI, Anthropic, Gemini, Cohere, Perplexity, IBM watsonx, Groq, Ollama, xAI (Grok), Sarvam AI
🆕 v5.1.4: Added xAI Grok and Sarvam AI support for cutting-edge and Indian language AI
1️⃣ Python SDK - Multi-Provider Intelligence
fromllmswapimportLLMClient# Auto-detects available providersclient=LLMClient()
# Or specify your preferenceclient=LLMClient(provider="anthropic") # Claude 3 Opus/Sonnet/Haikuclient=LLMClient(provider="openai") # GPT-4, GPT-3.5client=LLMClient(provider="gemini") # Google Gemini Pro/Flashclient=LLMClient(provider="watsonx") # IBM watsonx.ai Graniteclient=LLMClient(provider="ollama") # Llama, Mistral, Phi, 100+ localclient=LLMClient(provider="groq") # Groq ultra-fast inferenceclient=LLMClient(provider="cohere") # Cohere Command models for RAGclient=LLMClient(provider="perplexity") # Perplexity web-connected AIclient=LLMClient(provider="xai") # xAI Grok models 🆕client=LLMClient(provider="sarvam") # Sarvam AI (Indian languages) 🆕# Automatic failoverclient=LLMClient(fallback=True)
response=client.query("Hello") # Tries multiple providers# Save 50-90% with intelligent cachingclient=LLMClient(cache_enabled=True)
response1=client.query("Expensive question") # $$$ API callresponse2=client.query("Expensive question") # FREE from cache# 🆕 v5.1.0: Workspace-Aware SDK (Auto-detects .llmswap/)fromllmswapimportLLMClient# SDK automatically detects workspace in current directoryclient=LLMClient() # Loads workspace context if .llmswap/ exists# Query with full project contextresponse=client.query("How should I structure my API?")
# AI has access to: project context, past learnings, architecture decisions# Check if workspace is loadedifclient.workspace_manager:
workspace_data=client.workspace_manager.load_workspace()
print(f"Working in: {workspace_data['project_name']}")
print(f"Learnings tracked: {workspace_data['statistics']['learnings_count']}")
# Learnings are automatically saved after each query# No manual tracking needed!# 🆕 v5.1.0: Eklavya Mentor Integration with WorkspacefromllmswapimportLLMClientfromllmswap.eklavya.mentorimportEklavyaMentor# Initialize client and mentorclient=LLMClient(provider="anthropic")
mentor=EklavyaMentor(persona="guru", alias="Guruji")
# Generate teaching system prompt with workspace contextteaching_prompt=mentor.generate_system_prompt()
# Use mentor for teaching-focused responsesresponse=client.query(
"Explain Python decorators",
system_prompt=teaching_prompt
)
print(response.content) # Guru-style teaching response# Different personas for different learning stylescoach=EklavyaMentor(persona="coach", alias="Coach Sarah") # Motivationalfriend=EklavyaMentor(persona="friend", alias="CodeBuddy") # Collaborativesocrates=EklavyaMentor(persona="socrates") # Question-based learning# 🆕 v5.0: Conversational Sessions (Provider-Native)client.start_chat_session()
response=client.chat("Tell me about Python") # Context maintainedresponse=client.chat("What are its best features?") # Remembers previousclient.end_chat_session() # Clean provider-level cleanup# 🆕 v5.0: Async Support for High PerformanceimportasynciofromllmswapimportAsyncLLMClientasyncdefmain():
async_client=AsyncLLMClient()
response=awaitasync_client.query_async("Process this data")
asyncio.run(main())
2️⃣ CLI Suite - 15+ Powerful Terminal Commands
🆕 v5.1.0: Workspace Commands
# Initialize workspace in current project
llmswap workspace init
llmswap workspace init --name "My Flask App"# View workspace information
llmswap workspace info # Current workspace stats
llmswap workspace list # All workspaces
llmswap workspace journal # View learning journal
llmswap workspace decisions # View decision log
llmswap workspace context # Edit project context# Workspace automatically detected when you run:
llmswap chat # Loads workspace context
llmswap ask "How do I test this?"# Uses project-specific context
CLI Commands (All Features)
# 🆕 v5.0: Conversational Chat with Provider-Native Context
llmswap chat # Interactive AI assistant with memory# 🆕 v5.0: Configuration Management (Git-like)
llmswap config set provider.default anthropic
llmswap config export --file team-config.yaml
# Generate code from natural language (GitHub Copilot alternative)
llmswap generate "sort files by size in reverse order"
llmswap generate "Python function to read JSON file" --language python
llmswap generate "find large files over 100MB" --execute
# Ask one-line questions
llmswap ask "How to optimize PostgreSQL queries?"# Interactive AI chat
llmswap chat
# AI code review
llmswap review app.py --focus security
# Debug errors instantly
llmswap debug --error "ConnectionTimeout at line 42"# Analyze logs with AI
llmswap logs --analyze app.log --since "2h ago"
3️⃣ Provider Management & Model Configuration (v5.0.4 NEW!)
🎯 View All Providers and Models:
# Beautiful table showing all providers, their status, and default models
llmswap providers
# Update any provider's default model
llmswap config set provider.models.openai gpt-4o-mini
llmswap config set provider.models.cohere command-r-plus-08-2024
llmswap config set provider.models.anthropic claude-3-5-haiku-20241022
# Set default provider
llmswap config set provider.default anthropic
# View current configuration
llmswap config list
# Export/import team configurations
llmswap config export team-config.yaml
llmswap config import team-config.yaml --merge
🚀 Handle Model Deprecations:
When providers deprecate models (like Cohere's command-r-plus → command-r-plus-08-2024):
# Simply update your config - no code changes needed!
llmswap config set provider.models.cohere command-r-plus-08-2024
llmswap providers # Verify the change
⚙️ Configuration File Location:
User config: ~/.llmswap/config.yaml (created automatically on first run)
Custom location: Set LLMSWAP_CONFIG_HOME environment variable
Team sharing: Export/import YAML configs for team standardization
💬 Interactive Chat Commands:
llmswap chat # Start interactive conversation# Available commands in chat:
/help # Show all commands
/provider # Show current provider and available providers
/switch # Switch to different provider (privacy-compliant)
/clear # Clear conversation history
/stats # Show session statistics
/quit # Exit chat# Example session:
[0] > Hello, I'm working on a Python project[anthropic] Hi! I'd be happy to help with your Python project...
[1] > /switch
📋 Available providers: anthropic, gemini, perplexity, watsonx, groq
Enter provider name: gemini
🔒 PRIVACY NOTICE: Switching to gemini
✅ NO conversation history will be shared with the new provider
✅ This protects your privacy and complies with provider Terms of Service
Continue? (y/n): y
✅ Switched to gemini
💬 Starting fresh conversation with gemini
4️⃣ Analytics & Cost Optimization (v4.0 NEW!)
# Compare provider costs before choosing
llmswap compare --input-tokens 1000 --output-tokens 500
# Output: Gemini $0.0005 | OpenAI $0.014 | Claude $0.011# Track your actual usage and spending
llmswap usage --days 30 --format table
# Shows: queries, tokens, costs by provider, response times# Get AI spend optimization recommendations
llmswap costs
# Suggests: Switch to Gemini, enable caching, use Ollama for dev
# Python SDK - Full analytics suiteclient=LLMClient(analytics_enabled=True)
# Automatic conversation memoryresponse=client.chat("What is Python?")
response=client.chat("How is it different from Java?") # Remembers context# Real-time cost trackingstats=client.get_usage_stats()
print(f"Total queries: {stats['totals']['queries']}")
print(f"Total cost: ${stats['totals']['cost']:.4f}")
print(f"Avg response time: {stats['avg_response_time_ms']}ms")
# Cost optimization insightsanalysis=client.get_cost_breakdown()
print(f"Potential savings: ${analysis['optimization_opportunities']['potential_provider_savings']:.2f}")
print(f"Recommended provider: {analysis['recommendations'][0]}")
# Compare providers for your specific use casecomparison=client.get_provider_comparison(input_tokens=1500, output_tokens=500)
print(f"Cheapest: {comparison['cheapest']} (${comparison['cheapest_cost']:.6f})")
print(f"Savings vs current: {comparison['max_savings_percentage']:.1f}%")
# Context-aware caching for multi-tenant appsresponse=client.query(
"Get user data",
cache_context={"user_id": "user123"} # Isolated cache
)
Provider Comparison
# Compare responses from different modelscomparison=client.compare_providers(
"Solve this problem",
providers=["anthropic", "openai", "gemini"]
)
📊 Real-World Use Cases & Examples
🏢 Enterprise: Content Generation at Scale
Netflix-style recommendation descriptions for millions of items:
fromllmswapimportLLMClient# Start with OpenAI, switch to Gemini for 96% cost savingsclient=LLMClient(provider="gemini", cache_enabled=True)
defgenerate_descriptions(items):
foriteminitems:
# Cached responses save 90% on similar contentdescription=client.query(
f"Create engaging description for {item['title']}",
cache_context={"category": item['category']}
)
yielddescription.content# Cost: $0.0005 per description vs $0.015 with OpenAI
👨💻 Developers: AI-Powered Code Review
GitHub Copilot alternative for your team:
# CLI for instant code review
$ llmswapreviewapi_handler.py--focussecurity# Python SDK for CI/CD integrationfromllmswapimportLLMClientclient=LLMClient(analytics_enabled=True)
review=client.query(f"Review this PR for bugs: {pr_diff}")
# Track costs across your teamstats=client.get_usage_stats()
print(f"This month's AI costs: ${stats['totals']['cost']:.2f}")
🎓 Education: AI Tutoring Platform
Khan Academy-style personalized learning:
client=LLMClient(provider="ollama") # Free for schools!defai_tutor(student_question, subject):
# Use watsonx for STEM, Ollama for general subjectsifsubjectin ["math", "science"]:
client.set_provider("watsonx")
response=client.query(
f"Explain {student_question} for a {subject} student",
cache_context={"grade_level": student.grade}
)
returnresponse.content# Zero cost with Ollama, enterprise-grade with watsonx
🚀 Startups: Multi-Modal Customer Support
Shopify-scale merchant assistance:
fromllmswapimportLLMClient# Start with Anthropic, fallback to others if rate-limitedclient=LLMClient(fallback=True, cache_enabled=True)
asyncdefhandle_support_ticket(ticket):
# 90% of questions are similar - cache saves thousandsresponse=awaitclient.aquery(
f"Help with: {ticket.issue}",
cache_context={"type": ticket.category}
)
# Auto-escalate complex issuesifresponse.confidence<0.8:
client.set_provider("anthropic") # Use best modelresponse=awaitclient.aquery(ticket.issue)
returnresponse.content
📱 Content Creators: Writing Assistant
Medium/Substack article generation:
# Quick blog post ideas
llmswap ask "10 trending topics in AI for developers"# Full article draft
llmswap chat
> Write a 1000-word article on prompt engineering
> Make it more technical
> Add code examples
🔧 DevOps Engineers: Infrastructure as Code
Kubernetes and Docker automation:
# Generate Kubernetes deployment
llmswap generate "Kubernetes deployment for React app with 3 replicas" --save k8s-deploy.yaml
# Docker multi-stage build
llmswap generate "Docker multi-stage build for Node.js app with Alpine" --language dockerfile
# Terraform AWS infrastructure
llmswap generate "Terraform script for AWS VPC with public/private subnets" --save main.tf
🎯 Data Scientists: Analysis Workflows
Pandas, visualization, and ML pipeline generation:
# Data analysis scripts
llmswap generate "Pandas script to clean CSV and handle missing values" --language python
# Visualization code
llmswap generate "Matplotlib script for correlation heatmap" --save plot.py
# ML pipeline
llmswap generate "scikit-learn pipeline for text classification with TF-IDF" --language python
💬 App Developers: Full Applications
Complete app generation with modern frameworks:
# Streamlit chatbot
llmswap generate "Streamlit chatbot app with session state and file upload" --save chatbot.py
# FastAPI REST API
llmswap generate "FastAPI app with CRUD operations for user management" --save api.py
# React component
llmswap generate "React component for data table with sorting and filtering" --language javascript --save DataTable.jsx
🤖 AI/ML Engineers: Model Deployment
Production-ready ML workflows and deployments:
# LangChain RAG pipeline
llmswap generate "LangChain RAG system with ChromaDB and OpenAI embeddings" --language python --save rag_pipeline.py
# Hugging Face model fine-tuning
llmswap generate "Script to fine-tune BERT for sentiment analysis with Hugging Face" --save finetune.py
# Gradio ML demo app
llmswap generate "Gradio app for image classification with drag and drop" --save demo.py
# Vector database setup
llmswap generate "Pinecone vector database setup for semantic search" --language python
🔒 Security Engineers: Vulnerability Scanning
Security automation and compliance scripts:
# Security audit script
llmswap generate "Python script to scan for exposed API keys in codebase" --save security_scan.py
# OAuth2 implementation
llmswap generate "FastAPI OAuth2 with JWT tokens implementation" --language python
# Rate limiting middleware
llmswap generate "Redis-based rate limiting for Express.js" --language javascript
🛠️ AI Agent Development: Tool Creation
Build tools and functions for AI agents (inspired by Anthropic's writing tools):
# Create tool functions for agents
llmswap generate "Python function for web scraping with BeautifulSoup error handling" --save tools/scraper.py
# Database interaction tools
llmswap generate "SQLAlchemy functions for CRUD operations with type hints" --save tools/database.py
# File manipulation utilities
llmswap generate "Python class for safe file operations with context managers" --save tools/file_ops.py
# API integration tools
llmswap generate "Async Python functions for parallel API calls with rate limiting" --save tools/api_client.py
# Agent orchestration
llmswap generate "LangChain agent with custom tools for research tasks" --language python
🏆 Hackathon Power Kit: Win Your Next Hackathon
Build complete MVPs in minutes, not hours:
# RAG Chatbot for Document Q&A (Most requested hackathon project)
llmswap generate "Complete RAG chatbot with OpenAI embeddings, Pinecone vector store, and Streamlit UI for PDF document Q&A" --save rag_chatbot.py
# Full-Stack SaaS Starter (0 to production in 5 minutes)
llmswap generate "Next.js 14 app with Clerk auth, Stripe payments, Prisma ORM, and PostgreSQL schema for SaaS platform" --save saas_mvp.js
# Install package
pip install llmswap
# Set any API key (one is enough to get started)export ANTHROPIC_API_KEY="sk-..."# For Claudeexport OPENAI_API_KEY="sk-..."# For GPT-4export GEMINI_API_KEY="..."# For Google Geminiexport WATSONX_API_KEY="..."# For IBM watsonxexport WATSONX_PROJECT_ID="..."# watsonx projectexport GROQ_API_KEY="gsk_..."# For Groq ultra-fast inferenceexport COHERE_API_KEY="co_..."# For Cohere Command modelsexport PERPLEXITY_API_KEY="pplx-..."# For Perplexity web search# Or run Ollama locally for 100% free usage
Task Type
Input Tokens
Output Tokens
Estimated Cost
Active Development: Regular updates and feature releases
Built with ❤️ for developers who value simplicity, efficiency, and learning. Star us on GitHub if llmswap saves you time, money, or helps you learn faster!
v5.1.0 Release: The first AI tool that truly remembers your learning journey. 🚀