Scrapy Books Example

Scrapy Books Example

Example of Python Scrapy project. It scrapes book data from https://books.toscrape.com/.

DEVELOPER_TOOLSOPEN_SOURCEApify

Actor Scrapy Books Example

This project serves as an example of Python Scrapy project. It scrapes book data from books.toscrape.com.

Getting Started

Install Apify CLI

To use this scraper, you need to install the Apify CLI. Follow the instructions here.

Install Python and Virtualenv

Make sure you have Python installed. If not, download it here. Any version supported by Apify SDK and Scrapy should be fine.

Additionally, install Virtualenv using the following command:

pip install virtualenv

Run the Actor locally

Prepare Python environment

Create a Python virtual environment by running:

python3.12 -m virtualenv .venv

Activate the virtual environment:

source .venv/bin/activate

Install Python dependencies:

pip install -r requirements.txt -r requirements-dev.txt

Run the scraper as Scrapy project

The project is still runnable as a Scrapy project. Execute the following command:

scrapy crawl book_spider -o books.json

Run the scraper as Apify Actor

Run the scraper as an Apify Actor using:

apify run --purge

Deploy on Apify

Log in to Apify

You will need to provide your Apify API Token to complete this action.

apify login

Deploy your Actor

This command will deploy and build the Actor on the Apify Platform. You can find your newly created Actor under Actors -> My Actors.

apify push

Documentation reference

To learn more about Apify and Actors, take a look at the following resources:

Frequently Asked Questions

Is it legal to scrape job listings or public data?

Yes, if you're scraping publicly available data for personal or internal use. Always review Websute's Terms of Service before large-scale use or redistribution.

Do I need to code to use this scraper?

No. This is a no-code tool — just enter a job title, location, and run the scraper directly from your dashboard or Apify actor page.

What data does it extract?

It extracts job titles, companies, salaries (if available), descriptions, locations, and post dates. You can export all of it to Excel or JSON.

Can I scrape multiple pages or filter by location?

Yes, you can scrape multiple pages and refine by job title, location, keyword, or more depending on the input settings you use.

How do I get started?

You can use the Try Now button on this page to go to the scraper. You’ll be guided to input a search term and get structured results. No setup needed!