// Analyzes and optimizes SQL/NoSQL queries for performance. Use when reviewing query performance, optimizing slow queries, analyzing EXPLAIN output, suggesting indexes, identifying N+1 problems, recommending query rewrites, or improving database access patterns. Supports PostgreSQL, MySQL, SQLite, MongoDB, Redis, DynamoDB, and Elasticsearch.
| name | optimizing-queries |
| description | Analyzes and optimizes SQL/NoSQL queries for performance. Use when reviewing query performance, optimizing slow queries, analyzing EXPLAIN output, suggesting indexes, identifying N+1 problems, recommending query rewrites, or improving database access patterns. Supports PostgreSQL, MySQL, SQLite, MongoDB, Redis, DynamoDB, and Elasticsearch. |
Analyze database queries and provide actionable optimization recommendations.
Copy and track progress:
Query Optimization:
- [ ] Identify database type
- [ ] Analyze query structure
- [ ] Check for red flags
- [ ] Provide recommendations
- [ ] Suggest verification
For complex queries (3+ JOINs or nested subqueries):
For each query, evaluate:
SELECT * on large tablesWHERE function(column) prevents index useLIKE '%pattern' (leading wildcard)OR conditions spanning different columnsOFFSET valuesNOT IN with nullable subqueryLIMIT on unbounded queries$where, unbounded arraysKEYS *, missing TTL, wrong data structureExample 1: Missing Index
Input:
SELECT * FROM orders WHERE customer_email = 'user@example.com';
Output:
## Issue [HIGH]
Full table scan on `orders` - no index on `customer_email`
## Impact
Query time grows linearly with table size; causes lock contention
## Solution
CREATE INDEX idx_orders_customer_email ON orders(customer_email);
## Verification
EXPLAIN ANALYZE SELECT * FROM orders WHERE customer_email = 'user@example.com';
-- Should show "Index Scan" instead of "Seq Scan"
Example 2: N+1 Query Pattern
Input:
users = db.query("SELECT * FROM users")
for user in users:
orders = db.query(f"SELECT * FROM orders WHERE user_id = {user.id}")
Output:
## Issue [HIGH]
N+1 query pattern - executes 1 + N queries instead of 1
## Impact
100 users = 101 queries; network latency multiplied
## Solution
SELECT u.*, o.* FROM users u
LEFT JOIN orders o ON u.id = o.user_id;
## Verification
Check query count before/after; should reduce from N+1 to 1
Example 3: When NOT to Optimize
Input:
SELECT * FROM config WHERE key = 'app_version';
Output:
## No Issues Detected
This query on a small lookup table (typically <100 rows) is acceptable.
Adding an index would add write overhead without meaningful read benefit.
## When to Reconsider
- If table grows beyond 10,000 rows
- If query is called >100 times/second
Example 4: Incomplete Information
Input:
SELECT * FROM large_table WHERE status = ?;
Output:
## Issue [MEDIUM]
Preliminary analysis without schema information
## Potential Issues
1. `SELECT *` - may fetch unnecessary columns
2. `status` column - if low cardinality (e.g., 5 values), index may not help
## Information Needed
- Table row count
- Cardinality of `status` column
- Existing indexes: `SHOW INDEX FROM large_table;`
Example 5: MongoDB COLLSCAN
Input:
db.orders.find({ customer_email: "user@example.com" })
Output:
## Issue [HIGH]
COLLSCAN on `orders` - no index on `customer_email`
## Impact
Query scans entire collection; O(n) performance
## Solution
db.orders.createIndex({ customer_email: 1 })
## Verification
db.orders.find({ customer_email: "..." }).explain("executionStats")
// Should show "stage": "IXSCAN" instead of "COLLSCAN"
Structure recommendations as:
## Issue [HIGH/MEDIUM/LOW]
[Specific problem identified]
## Impact
[Performance/resource cost]
## Solution
[Concrete fix with code]
## Verification
[EXPLAIN command or metric to confirm]
When uncertainty exists, indicate confidence:
<table_name>| Issue Type | Impact | Priority |
|---|---|---|
| Missing JOIN index | Query blocks | P0 |
| Full table scan (>100k rows) | Slow response | P0 |
| N+1 pattern | Latency × N | P1 |
| SELECT * | Memory/bandwidth | P1 |
| Missing LIMIT | Resource exhaustion | P1 |
| Large OFFSET | Slow pagination | P2 |
| Suboptimal ORDER BY | Sort overhead | P2 |
Consult these for detailed patterns (use grep to find specific sections):
references/sql-patterns.md - SQL optimization
## Index, ## Query Rewrite, ## Join, ## Aggregationreferences/nosql-patterns.md - MongoDB, Redis, DynamoDB, Elasticsearch
## MongoDB, ## Redis, ## DynamoDB, ## Elasticsearchreferences/explain-analysis.md - Execution plans
## PostgreSQL, ## MySQL, ## MongoDBreferences/common-scenarios.md - Common patterns
## Slow pagination, ## N+1, ## Dashboard, ## Bulkreferences/orm-patterns.md - ORM fixes
### Django, ### SQLAlchemy, ### ActiveRecord, ### PrismaAll scripts use Python standard library only (no pip install required).
Run these for automated analysis:
analyze_query.py - Detect anti-patterns and score complexity:
python scripts/analyze_query.py "SELECT * FROM orders WHERE ..."
python scripts/analyze_query.py --file query.sql --json
suggest_index.py - Generate index recommendations:
python scripts/suggest_index.py "SELECT ... FROM orders WHERE status = 'pending'"
parse_explain.py - Analyze EXPLAIN output:
python scripts/parse_explain.py --pg explain_output.txt
python scripts/parse_explain.py --mysql explain_output.txt
diff_explain.py - Compare before/after EXPLAIN outputs:
python scripts/diff_explain.py --pg before.txt after.txt
python scripts/diff_explain.py --mysql before.txt after.txt --json
parse_slow_log.py - Analyze slow query logs:
python scripts/parse_slow_log.py --pg /var/log/postgresql.log --top 20
python scripts/parse_slow_log.py --mysql slow-query.log --json
analyze_schema.py - Find schema optimization opportunities:
python scripts/analyze_schema.py schema.sql
detect_orm_issues.py - Find ORM anti-patterns in code:
python scripts/detect_orm_issues.py app.py
python scripts/detect_orm_issues.py --dir ./models
check_migration.py - Check migration safety:
python scripts/check_migration.py --pg migration.sql
python scripts/check_migration.py --mysql migration.sql
For query optimization:
analyze_query.py → detect issuessuggest_index.py → generate indexesdiff_explain.py → verify improvementFor verifying optimization impact:
# Save before state
psql -c "EXPLAIN (ANALYZE, BUFFERS) SELECT ..." > before.txt
# Apply optimization (add index, rewrite query, etc.)
# Save after state
psql -c "EXPLAIN (ANALYZE, BUFFERS) SELECT ..." > after.txt
# Compare
python scripts/diff_explain.py --pg before.txt after.txt
For codebase audit:
detect_orm_issues.py --dir ./ → find N+1 patternsanalyze_schema.py → check schemaFor migration review:
check_migration.py → safety checkProvide appropriate commands:
-- PostgreSQL
EXPLAIN (ANALYZE, BUFFERS) SELECT ...;
-- MySQL
EXPLAIN ANALYZE SELECT ...;
-- SQLite
EXPLAIN QUERY PLAN SELECT ...;
// MongoDB
db.collection.find({...}).explain("executionStats")
When suggesting indexes:
-- PostgreSQL/MySQL
CREATE INDEX [CONCURRENTLY] idx_table_columns
ON table(col1, col2, col3);
-- Rationale: [why this column order, covering considerations]
// MongoDB (ESR: Equality, Sort, Range)
db.collection.createIndex({ equality_field: 1, sort_field: -1, range_field: 1 })
When user provides incomplete context:
No schema provided:
"To give precise recommendations, I need the table schema. Could you share:
- Column data types
- Existing indexes
- Approximate row counts
Meanwhile, here's what I can infer from the query..."
No EXPLAIN output:
"Run this command and share the output:
EXPLAIN (ANALYZE, BUFFERS) <their query>;This will show actual vs estimated rows and scan types."
Unknown database type:
"Which database are you using? (PostgreSQL, MySQL, SQLite, MongoDB, etc.) The optimization strategies differ significantly between them."
Analyzing a query:
analyze_query.py firstInterpreting EXPLAIN output:
parse_explain.py with appropriate flagFull optimization request:
Complex optimization (3+ issues or architectural decisions): Use extended reasoning to:
If scripts fail:
Common errors:
FileNotFoundError → Check file path, use absolute pathsjson.JSONDecodeError → Input not valid JSON, check --pg/--mysql flagsWhen scripts are unavailable: Perform manual analysis using Quick Reference red flags and reference files.
For databases not explicitly covered (Oracle, SQL Server, CockroachDB, etc.):
references/sql-patterns.md