Quick Start
This guide will walk you through scraping your first rental property listings from Jiji Ghana or Meqasa using the interactive CLI.Make sure you’ve completed the installation before proceeding.
Two-Phase Workflow
ScrapeAccraProperties uses a two-phase approach:- Collect listing URLs from search/result pages
- Scrape listing details by visiting each URL
Your First Scrape
Collect listing URLs
Select option For this quickstart, let’s choose This will scrape the first 2 pages of Jiji rental listings.The spider will start crawling and display progress:URLs are saved to:
1 to collect listing URLs.Choose your source:1 (Jiji only).Configure pagination:outputs/urls/jiji_urls.csvScrape listing details
Run the CLI again and select option Select Specify the URL CSV path:Press Enter to use the default path. The spider will:
2 to scrape listing details:2. Scrape listing detailsChoose your source:- Read URLs from
outputs/urls/jiji_urls.csv - Visit each listing page
- Extract structured data (title, location, price, bedrooms, amenities, etc.)
- Write incrementally to
outputs/data/jiji_data.csv
The cleaning step (producing
raw.csv) is currently Jiji-only.Explore your data
Check the View the data:You’ll see rental listings with fields like:
outputs/ directory for your scraped data:url- Listing URLtitle- Property titlelocation- Area/neighborhoodhouse_type- Apartment, house, etc.bedrooms- Number of bedroomsbathrooms- Number of bathroomsprice- Rental priceamenities- List of amenitiesdescription- Full listing descriptionfetch_date- When the data was scraped
Resume Scraping
If your scrape is interrupted, you can resume without re-scraping existing listings:Review the queue summary
The CLI will show which URLs are already scraped vs. pending:Only the 13 pending URLs will be scraped.
Running Spiders Directly
You can bypass the interactive CLI and run spiders directly withscrapy:
Next Steps
Understand the Workflow
Learn about the two-phase workflow, resume mode, and data cleaning pipeline
Configure Settings
Customize spider behavior, pagination, concurrency, and Playwright options
Output Schema
Understand the structure of URL CSVs and listing data CSVs
Troubleshooting
Fix common issues like browser failures, empty fields, and resume errors