Overview
This guide covers proven strategies for monitoring website health, optimising performance, and maintaining high availability with Adapt.Scheduling Strategy
Choose the Right Interval
Adapt supports recurring crawls at 6, 12, 24, or 48-hour intervals:6-Hour Interval
6-Hour Interval
Best for:
- High-traffic production sites
- E-commerce platforms
- Sites with frequent content updates
- Uses more daily page quota
- Provides rapid issue detection
- Keeps cache consistently warm
12-Hour Interval
12-Hour Interval
Best for:
- Business websites
- Marketing sites with regular updates
- SaaS application landing pages
- Balanced quota usage
- Twice-daily health checks
- Good cache coverage
24-Hour Interval
24-Hour Interval
Best for:
- Corporate websites
- Portfolio sites
- Documentation sites
- Blogs
- Efficient quota usage
- Daily health monitoring
- Standard recommendation
48-Hour Interval
48-Hour Interval
Best for:
- Low-traffic sites
- Archive sites
- Development environments
- Minimal quota usage
- Less frequent updates
- Lower cache coverage
Cache Warming Strategy
When to Warm Cache
- After Publishing: Run a crawl immediately after deploying new content
- Before Traffic Spikes: Warm cache before expected high-traffic events
- After Cache Purges: Re-warm after manual cache clearing
- Regular Maintenance: Schedule recurring crawls to keep cache fresh
Priority-Based Warming
Connect Google Analytics to enable priority-based cache warming:Without Analytics, Adapt warms the homepage first, then processes pages in discovery order.
Crawl Configuration
Sitemap vs Link Crawling
Use both methods for comprehensive coverage:| Method | Pros | Cons |
|---|---|---|
| Sitemap | Fast, comprehensive, respects your structure | Requires sitemap.xml |
| Link Crawling | Finds unlisted pages, validates internal links | Slower, may miss isolated pages |
Setting Max Pages
Production Sites
Production Sites
Set
max_pages: 0 (unlimited) for complete site coverage:Testing & Development
Testing & Development
Use a limit for initial testing:
Large Sites (1000+ Pages)
Large Sites (1000+ Pages)
Consider multiple focused crawls:
Concurrency Settings
Adjust concurrency based on your hosting:| Hosting Type | Recommended Concurrency |
|---|---|
| Shared hosting | 5-10 |
| VPS / Cloud | 20-30 |
| Dedicated server | 30-50 |
| CDN (Cloudflare, etc.) | 50-100 |
Higher concurrency speeds up crawls but increases server load. Start conservative and increase if your server handles it well.
Monitoring & Alerts
Set Up Slack Notifications
Monitor Usage Limits
Check usage regularly to avoid hitting limits:Multi-Organisation Workflows
Organise by Client or Project
Agency Use Case
Agency Use Case
Create one organisation per client:
- Client A → Organisation “Client A”
- Client B → Organisation “Client B”
- Internal → Organisation “Agency Internal”
- Isolated data and limits
- Easy client handoff
- Clear billing separation
Multi-Site Business
Multi-Site Business
Create organisations by environment or brand:
- Production sites → Organisation “Production”
- Staging sites → Organisation “Staging”
- Partner sites → Organisation “Partners”
- Environment isolation
- Separate quota pools
- Different team access
Team Management
Assign Multiple Admins
Assign Multiple Admins
Always have at least 2 admins per organisation:
Use Member Role for Limited Access
Use Member Role for Limited Access
Grant “member” role to team members who need to view results but not manage settings.
Performance Optimisation
Identify Bottlenecks
Review Slow Pages
Export slow pages report and analyse common patterns:
- Large images
- Slow database queries
- Third-party scripts
- Cache misses
Cache Optimisation
Maximise cache hit ratio:- Enable Cache Warming: Run crawls after publishing
- Monitor Hit Ratio: Aim for 80%+ cache hits
- Fix Cache Misses: Investigate pages with consistent MISS status
- Warm High-Traffic Pages: Prioritise pages with most visitors
Broken Link Management
Weekly Review Process
Categorise Issues
Group broken links by:
- Internal vs external
- High-traffic vs low-traffic
- Critical vs non-critical
Proactive Prevention
Pre-Delete Checks
Pre-Delete Checks
Before deleting pages, search your content for internal links to that page and update or remove them.
Link Validation
Link Validation
Run a crawl in staging before deploying to production.
Regular Audits
Regular Audits
Schedule quarterly reviews of external links, as third-party sites change frequently.
API Integration Patterns
CI/CD Integration
Trigger crawls automatically after deployments:Automated Reporting
Build custom reports using the API:Monitoring Scripts
Track performance trends:Security Best Practices
Use API Keys for Integrations
Use API Keys for Integrations
Create scoped API keys instead of using JWT tokens in automation:
Rotate Keys Regularly
Rotate Keys Regularly
Rotate API keys quarterly and immediately after team member departures.
Restrict Member Access
Restrict Member Access
Use the “member” role for users who only need to view results, reserving “admin” for users who manage settings and billing.
Troubleshooting Common Issues
Job Takes Too Long
Job Takes Too Long
Solutions:
- Reduce concurrency to respect server rate limits
- Set a
max_pageslimit for testing - Check if robots.txt specifies a high crawl-delay
- Verify your hosting can handle the load
High Cache Miss Rate
High Cache Miss Rate
Solutions:
- Run crawls more frequently to keep cache warm
- Check CDN settings for cache TTL
- Verify pages are cacheable (not authenticated)
- Review cache-control headers
Many 404 Errors
Many 404 Errors
Solutions:
- Check sitemap.xml for outdated URLs
- Review recent content deletions
- Validate internal link updates
- Check for broken external links
Hitting Usage Limits
Hitting Usage Limits
Solutions:
- Reduce crawl frequency
- Set max_pages limits
- Upgrade to a higher plan
- Stagger crawls across multiple days