Talk to your PostgreSQL databases through AI. No dashboards, no BI tools, just questions and answers.
dbecho is an MCP server that gives AI agents (Claude Code, Cursor, Windsurf, or any MCP client) direct read-only access to your PostgreSQL databases. Point it at your databases, ask questions in plain language, get instant analytics.
You: What are my most popular blog posts and why?
Claude: [runs schema → query → analyze → trend across 29 tables]
Here's what the data shows...
11 tools that cover the full analytics workflow:
| Tool | Purpose |
|---|---|
list_databases |
Show all connected databases |
health |
Check connectivity, PostgreSQL version, database size |
schema |
Full schema: tables, columns, types, PKs, row counts, sizes |
query |
Run read-only SQL (SELECT, WITH, EXPLAIN, SHOW) |
analyze |
Profile a table: nulls, cardinality, distributions, top values |
compare |
Same query across multiple databases, side by side |
summary |
Overview: table counts, total rows, largest tables |
trend |
Time-series: counts/averages grouped by day/week/month/year |
anomalies |
Data quality: high nulls, outliers, duplicates, future dates |
sample |
Preview rows from any table |
erd |
Entity-relationship diagram: PKs and foreign keys |
Plus 3 MCP Resources (schema/summary per database) and 3 MCP Prompts (guided exploration, cross-database comparison, data quality audit).
The problem: You have PostgreSQL databases across projects. Getting answers means context-switching to psql, pgAdmin, or a BI tool, writing SQL, formatting results, then bringing insights back to your conversation.
The fix: dbecho stays inside your AI agent's workflow. The agent explores schema, writes SQL, cross-references tables, and builds analysis without you leaving the conversation. One config file, zero context switches.
What makes it different from just giving an agent a connection string:
- Multi-database. Connect 1 or 20 databases. Compare across them with one tool call.
- Safe by default. Read-only connections, query timeouts, row limits, SQL injection prevention. You can't accidentally
DROP TABLE. - Agent-optimized output. Schema, stats, and query results are formatted so LLMs parse them efficiently, not as raw psql dumps.
- Zero infrastructure. No containers, no web UI, no background processes. A Python package that speaks MCP over stdio.
You: Show me what's in the database
Claude: [calls list_databases → schema → summary]
Database: ecommerce (12 tables, 847 MB)
Largest: orders (2.1M rows), products (45K rows), users (128K rows)
You: How are sales trending this quarter?
Claude: [calls trend on orders table, grouped by week]
Week | orders | total_revenue
2026-01-06 | 3,241 | $482,100
2026-01-13 | 3,892 | $571,340
2026-01-20 | 2,107 | $298,500 ← 46% drop
...
Significant drop in week of Jan 20. Let me check what happened.
[calls query to investigate, then anomalies on orders table]
Found: 89% of orders that week have status='cancelled'.
The cancellation spike correlates with a payment gateway outage
(payments table shows 0 successful transactions Jan 19-21).
You: Any data quality issues I should worry about?
Claude: [calls anomalies on each table]
Found 4 issues:
- users.email: 342 possible duplicates
- orders.shipped_at: 127 dates in the future
- products.price: 23 outliers (IQR: $5..$89, found items at $0 and $9,999)
- events.user_id: 94.2% NULL (missing user attribution)
One conversation, zero context switches. The agent picks the right tools automatically.
pip install dbechoOr from source:
git clone https://github.com/ginkida/dbecho.git
cd dbecho
pip install .Requires Python 3.10+.
Create dbecho.toml in your project directory:
[databases.myapp]
url = "postgres://user:pass@localhost:5432/myapp"
description = "Main application"
[databases.analytics]
url = "postgres://user:pass@localhost:5432/analytics"
description = "Analytics warehouse"
[settings]
row_limit = 500 # max rows returned per query (default: 500)
query_timeout = 30 # seconds before query is killed (default: 30)Environment variables work with ${VAR} syntax:
[databases.production]
url = "${DATABASE_URL}"
description = "Production (read replica)"Claude Code (project-level, recommended):
Create .mcp.json in your project root:
{
"mcpServers": {
"dbecho": {
"command": "dbecho",
"args": ["--config", "/path/to/dbecho.toml"]
}
}
}Claude Code (global):
Add to ~/.claude.json:
{
"mcpServers": {
"dbecho": {
"command": "dbecho"
}
}
}When no --config is passed, dbecho searches for config in:
./dbecho.toml(current directory)~/.config/dbecho/config.toml~/.dbecho.toml
Other MCP clients (Cursor, Windsurf, etc.): use the same command/args in your client's MCP server configuration.
Show me a summary of all my databases
How many users signed up each month this year?
Compare order counts between staging and production
Find data quality issues in the events table
What's the relationship between users, orders, and products?
Which columns have the most nulls?
Show me the trend of daily revenue for the last 90 days
The agent picks the right tools automatically. You don't need to know the tool names.
dbecho is designed to be safe to point at any database, including production:
- Read-only connections. Every connection sets
default_transaction_read_only=onat the PostgreSQL level. Even if someone crafts malicious SQL, the database rejects writes. - Query whitelist. Only
SELECT,WITH,EXPLAIN, andSHOWstatements are allowed. Checked before execution. - SQL injection prevention. All table/column identifiers use
psycopg.sql.Identifier()parameterization. User input is validated against^[a-zA-Z_][a-zA-Z0-9_]*$. - Query timeout. Default 30 seconds. Prevents runaway queries from locking your database.
- Row limit. Default 500 rows per query. Prevents the agent from pulling entire tables into context.
- Local only. No network calls, no telemetry, no cloud. Data stays on your machine.
src/dbecho/
config.py TOML config loading, env var expansion, validation
db.py DatabaseManager: connections, schema cache, queries, stats, trends, anomalies
server.py FastMCP server: 11 tools, 3 resources, 3 prompts
~1000 lines of Python total. No framework beyond mcp and psycopg.
git clone https://github.com/ginkida/dbecho.git
cd dbecho
pip install -e ".[dev]"
pytest -vTests are fully mocked, no PostgreSQL instance needed. CI runs on Python 3.10-3.13.
MIT
