Scaling Overview
Scaling strategies differ based on your cluster plan:- Basic: Automatic scaling based on usage
- Standard: Manual capacity provisioning
- Advanced: Independent node and storage scaling
Basic Cluster Scaling
Basic clusters automatically scale based on your workload: Automatic Scaling Features:- Compute scales with Request Units consumed
- Storage automatically provisioned as data grows
- No manual intervention required
- Pay only for resources used
Basic clusters are billed based on Request Units (RUs). Set spend limits to control costs while allowing automatic scaling.
Standard Cluster Scaling
Standard clusters use provisioned compute capacity with usage-based storage.Scale Compute Capacity
Access Capacity Settings
- Navigate to cluster’s Overview page
- Click Update capacity in the Capacity section
Adjust vCPUs
Use the slider or input field to set provisioned vCPUs:Guidelines:
- Minimum: 4 vCPUs for production
- Recommended: 40% headroom above peak
- Maximum: Contact support for >128 vCPUs
- Small: 4-8 vCPUs (dev/test)
- Medium: 16-32 vCPUs (production)
- Large: 64+ vCPUs (high-traffic applications)
Monitor Resource Usage
Track resource utilization to determine when to scale: Key Metrics to Monitor:- CPU Usage: Scale up if consistently >70%
- RU Consumption: Track against provisioned capacity
- SQL Statement Latency: Increase with resource constraints
- Connection Count: High connection counts may need more capacity
Review Performance
Check the following dashboards:
- Overview: CPU, RUs, and SQL statements
- Request Units: Detailed RU breakdown
- SQL: Query performance and latency
Storage Scaling
Storage automatically scales with data growth:- Automatic provisioning: Added as needed
- Usage-based billing: Pay for storage used
- Per-region pricing: Rates vary by cloud provider and region
- Three replicas included: No extra charge for standard replication
Advanced Cluster Scaling
Advanced clusters offer independent scaling of nodes and storage.Add or Remove Nodes
Modify Nodes
Click Add nodes or select nodes to removeAdding Nodes:
- Select region(s)
- Specify number of nodes per region
- Review cost impact
- Click Add nodes
- Select nodes to remove
- Ensure sufficient nodes remain (minimum 3)
- Confirm removal
| Cluster Size | Use Case |
|---|---|
| 3 nodes | Development, testing |
| 5-9 nodes | Production, single region |
| 9+ nodes | Production, multi-region |
| 15+ nodes | High-traffic, global applications |
Increase Storage per Node
Change Node Size
You cannot directly change node size (vCPUs/RAM) for existing clusters. To resize nodes:Scaling Best Practices
Plan for Growth
Capacity Planning
- Monitor current usage trends
- Project growth over 6-12 months
- Allow 40% headroom for traffic spikes
- Test at expected peak load
Optimize Before Scaling
Before adding capacity, optimize your workload: Query Optimization:- Review slow queries in SQL Activity
- Add appropriate indexes
- Use
EXPLAINto analyze query plans - Implement caching where appropriate
- Review table designs
- Partition large tables
- Use appropriate data types
- Implement TTL for old data
- Use connection pooling
- Limit maximum connections
- Close idle connections
- Monitor connection counts
Multi-Region Scaling
For multi-region clusters: Balanced Topology:- Keep node counts equal across regions
- Distribute primary replicas evenly
- Use regional tables for locality
- Place clusters close to users
- Use follower reads for read-heavy workloads
- Configure table localities appropriately
Cost Optimization
Right-size Your Cluster:- Don’t over-provision capacity
- Monitor actual usage vs. provisioned
- Scale down during off-peak periods (where allowed)
- Use Basic tier for dev/test environments
- Implement data retention policies
- Use TTL to expire old data
- Archive historical data
- Compress large text/binary columns
Scaling for Specific Workloads
OLTP Workloads
Characteristics:- High transaction rates
- Low latency requirements
- Many concurrent connections
- Provision sufficient CPU capacity
- Use connection pooling
- Add nodes for increased throughput
- Optimize frequently-run queries
Analytical Workloads
Characteristics:- Complex queries
- Large data scans
- Batch processing
- Increase storage capacity
- Add nodes for parallel processing
- Use changefeeds to offload analytics
- Consider separate read replicas
Mixed Workloads
Characteristics:- Both transactional and analytical queries
- Variable traffic patterns
- Multiple applications
- Balanced node and storage scaling
- Separate clusters for different workloads
- Use workload isolation techniques
- Monitor and adjust based on usage patterns
Monitoring Scaling Operations
Track scaling operations and their impact:During Scaling
- Cluster Status: Check Overview page for operation progress
- Performance: Monitor query latency during scaling
- Rebalancing: Watch data distribution in Metrics
- Alerts: Review any alerts triggered
After Scaling
Check Performance
Review metrics to ensure improved performance:
- CPU utilization decreased
- Query latency improved
- Throughput increased
Troubleshooting Scaling Issues
Scaling Operation Stuck
Symptoms: Scaling operation doesn’t complete Solutions:- Check cluster status for error messages
- Verify no ongoing maintenance operations
- Contact Support if operation is stuck >30 minutes
Performance Not Improved
Symptoms: Scaling up didn’t improve performance Possible Causes:- Bottleneck is elsewhere (queries, indexes)
- Data not rebalanced yet
- Application connection pool not increased
- Review query performance in SQL Activity
- Wait for rebalancing to complete
- Check application connection pool settings
- Analyze with Insights for recommendations
Scaling Down Blocked
Symptoms: Cannot decrease capacity Causes:- Already scaled down 3 times in 7 days (Standard)
- Capacity decrease would cause issues
- Wait for 7-day period to reset
- Contact Support for assistance
- Consider other optimization approaches
Scaling Automation
Use the Cloud API to automate scaling:- Schedule capacity changes
- Respond to CloudWatch/Datadog alerts
- Implement auto-scaling policies
- Integrate with CI/CD pipelines
Next Steps
Performance Tuning
Optimize query performance
Monitoring
Set up comprehensive monitoring
Resource Usage
Understand resource consumption
Pricing
Understand cost implications