Documentation Index
Fetch the complete documentation index at: https://docs.cascadeflow.ai/llms.txt
Use this file to discover all available pages before exploring further.
Python
Minimal install
Core dependencies: pydantic>=2.0.0, httpx>=0.25.0, tiktoken>=0.5.0, rich>=13.0.0.
With providers
pip install "cascadeflow[providers]" # OpenAI + Anthropic + Groq
Individual providers:
pip install "cascadeflow[openai]" # OpenAI
pip install "cascadeflow[anthropic]" # Anthropic
pip install "cascadeflow[groq]" # Groq
pip install "cascadeflow[huggingface]" # Hugging Face
pip install "cascadeflow[together]" # Together AI
With framework integrations
pip install "cascadeflow[langchain]" # LangChain/LangGraph
pip install "cascadeflow[openai-agents]" # OpenAI Agents SDK
pip install "cascadeflow[crewai]" # CrewAI (Python 3.10+)
pip install "cascadeflow[google-adk]" # Google ADK (Python 3.10+)
Local inference
pip install "cascadeflow[vllm]" # vLLM (Python 3.10-3.13)
Ollama does not need a Python package — cascadeflow communicates with Ollama via HTTP at localhost:11434. Install Ollama separately from ollama.ai.
Everything
pip install "cascadeflow[all]" # All providers + semantic routing
Development
git clone https://github.com/lemony-ai/cascadeflow.git
cd cascadeflow
pip install -e ".[dev]"
TypeScript
Core
npm install @cascadeflow/core
Framework packages
npm install @cascadeflow/langchain # LangChain integration
npm install @cascadeflow/vercel-ai # Vercel AI SDK middleware
npm install @cascadeflow/n8n-nodes-cascadeflow # n8n community node
Provider Setup
Set API keys as environment variables:
export OPENAI_API_KEY="sk-..."
export ANTHROPIC_API_KEY="sk-ant-..."
export GROQ_API_KEY="gsk_..."
cascadeflow auto-detects available providers based on which API keys are set.
Verify Installation
python -c "import cascadeflow; print(cascadeflow.__version__)"
python -c "from cascadeflow import init, run, HarnessConfig, HarnessRunContext; print('OK')"
Next Step
Start observing your LLM calls with zero code changes. Quickstart: Observe Mode →