Pageripper API: a commercial data-extraction service

Pageripper API a commercial data-extraction service


Pageripper is a commercial API that extracts data from webpages, even if they're rendered with Javascript. Pageripper extracts:

  • links
  • email addresses
  • Twitter handles

from any public website.

Who is it for?

Pageripper is a tool for data analysts, digital marketers, web developers and more who need to reliably retrieve links from websites.

Pageripper extracts data from websites such as links, email addresses and Twitter handles

Pageripper is often ideal for folks who could write their own data extraction API themselves, but don't want the hassle of hosting or maintaining it.

A simple example

Let's look at an example. Suppose we want to extract all the links from the New York times homepage:

curl -X POST \
  -H 'Content-type: application/json' \
  -d '{"options": {"includeUrls": true, "waitUntil":"domcontentloaded"}, "url": ""}' \

We'd then get back:

  "emails": [],
  "twitterHandles": [],
  "urls": [url1, url2, url3]

You can optionally request email addresses and Twitter handles, too. And you can specify which network event Puppeteer should wait for when fetching a site to extract data from.

What problem is Pageripper solving?

As Javascript-heavy websites have become increasingly popular over the years, we've seen a sharp rise in Single Page Applications (SPAs). These sites behave differently from traditional dynamic websites that might be rendered by the backend. In the name of increased performance and optimizing for the end user experience, SPAs send a single minimal HTML file that links to a Javascript bundle.

Pageripper handles even Javascript-rendered sites

This Javascript bundle contains the entirety of the frontend application's logic for rendering the UI, including all links. While this approach vastly improves loading times and reducing the overall time until a given site is interactive for the end user, it complicates data-extraction tasks such as getting all the links out of a given page.

For more detail and a demonstration of what scrapers see under the hood when fetching these two different site architectures, see Evolving web scraping: How Pageripper handles Javascript-heavy sites.

The Pageripper API addresses this challenge by using Chrome's Pupeteer project under the hood. This means that the API renders sites in exactly the same way a fully fledged web browser, such as Chrome, would.

This enables extraction of data such as links, email addresses and Twitter handles, even from sites that are rendered entirely in Javascript.

Pageripper lets you tweak Pupeteer's behavior on a per-request basis in order to specify which network events to wait for. This additional control allows users to get the best data-extraction experience from a much wider array of sites on the public internet.

Scalability and performance

Pageripper is a Node.js project written in TypeScript that is Dockerized into a container. The container includes Puppeteer and a headless Chrome browser as a dependency, which it uses to render websites.

The API is defined as an ECS service for scalability and deployed on AWS behind a load balancer.

Pageripper is defined via Infrastructure as Code (IaC) with Pulumi. Additional capacity can be added quickly in order to meet demand.

Key features and benefits

Pageripper allows you to fetch and extract data from any public website. On a per-request basis, you can even pass options to tweak and tune Puppeteer's own behavior, such as which network event to wait for when fetching a site to extract data from.

Use cases

Pageripper is ideal for getting usable data out of large pages. For example, a conference site might publish the contact information and company websites for speakers at an upcoming event that you're attending. Pageripper is ideal for fetching this data from the page and making it available to you easily so that you can process it further according to your needs.

Integration and usage

Pageripper is currently available via RapidAPI. The API's OpenAPI spec is available in the public GitHub repository.

You can test out the API from RapidAPI if you're curious to kick the tires. In the future, there will likely be client SDKs generated for various languages to ease integration.

Pricing and Commercial uses

Pageripper is available for commercial use. There are currently three pricing plans to support different needs and scaling requirements.

Getting started

Please give Pageripper a shot and let me know what you think!