Github Marketplace Scraper

Github Marketplace Scraper

Scrape GitHub Marketplace listings. Extracts relevant metadata, developer details, pricing information, and even hidden fields like emails.

LEAD_GENERATIONAUTOMATIONAIApify

This Apify actor scrapes listing details from GitHub Marketplace using a search query-based approach. It extracts relevant metadata, developer details, pricing information, and more.

Features

  • Crawls GitHub Marketplace search results for application listings
  • Extracts application metadata such as name, description, and logo
  • Retrieves developer information including name and support email
  • Captures pricing plans for each listing
  • Collects number of installs and category information

Use Cases

  • Competitive analysis of GitHub Marketplace applications
  • Market research on GitHub ecosystem apps
  • Aggregating and analyzing developer and pricing data
  • Monitoring changes in listings over time

How It Works

  1. Search Query Crawling: The scraper loads search results from GitHub Marketplace based on a predefined query.
  2. Data Extraction: For each listing, it retrieves relevant metadata using Cheerio.
  3. Data Storage: Extracted data is pushed to the Apify dataset for further processing.

Input Schema

This scraper does not require user input as it automatically crawls GitHub Marketplace using a predefined search query.

Example Output

1{
2  "app_name": "Sample App",
3  "developer_name": "Sample Developer",
4  "description": "This is a sample GitHub Marketplace listing.",
5  "description_long": "Detailed description of the app.",
6  "logo_url": "https://example.com/logo.png",
7  "number_of_installs": 5000,
8  "privacy_policy": "https://example.com/privacy",
9  "support_email": "support@example.com",
10  "categories": [
11    {"name": "DevOps"},
12    {"name": "Security"}
13  ],
14  "plans": [
15    {"name": "Basic", "price": "$10/month"},
16    {"name": "Pro", "price": "$50/month"}
17  ],
18  "app_url": "https://github.com/marketplace/sample-app"
19}

Need More Features?

If you'd like to add new data fields to this scraper or need a custom scraper for another purpose, feel free to file an issue or get in touch! We are open to customizing the scraper to suit your needs.

Why Choose This Scraper?

  • Efficient: Handles large-scale scraping with robust session and concurrency management.
  • Customizable: Easily tailor input options for specific scraping requirements.
  • Reliable: Implements error-handling mechanisms to minimize interruptions.

Proxies and Anti-blocking

This scraper uses the Apify Proxy or your custom proxy settings to reduce the risk of being blocked. It supports automatic session management for uninterrupted scraping.

Resources

Get Started

  • Clone this scraper or use it directly on the Apify platform.
  • Customize the input settings as needed.
  • Run the scraper and export the data in JSON, CSV, or Excel format.

Unlock valuable insights from the Github Marketplace with this powerful scraper. Get started today!

Frequently Asked Questions

Is it legal to scrape job listings or public data?

Yes, if you're scraping publicly available data for personal or internal use. Always review Websute's Terms of Service before large-scale use or redistribution.

Do I need to code to use this scraper?

No. This is a no-code tool — just enter a job title, location, and run the scraper directly from your dashboard or Apify actor page.

What data does it extract?

It extracts job titles, companies, salaries (if available), descriptions, locations, and post dates. You can export all of it to Excel or JSON.

Can I scrape multiple pages or filter by location?

Yes, you can scrape multiple pages and refine by job title, location, keyword, or more depending on the input settings you use.

How do I get started?

You can use the Try Now button on this page to go to the scraper. You’ll be guided to input a search term and get structured results. No setup needed!