Integrations

MrScraper can integrate smoothly into any workflow.

At this point, you need to decide how you want to work with the data.

Some people choose to download the results and manually process them. However, in scenarios involving a large amount of data or real-time requirements, integration is necessary. You can add the results to a database, Google Sheet, or even send them by email.

Available integration options

There is an infinite number of integrations you can make with your scrapers. To provide flexibility and cover various scenarios, we currently have three available methods of integration:

Webhooks

Webhooks are great because they send information in real-time as soon as a scraper has finished.

Most people use webhooks to integrate the results of scraping with their own applications using their favorite programming language.

There are also no-code apps such as Zapier or Make.com that accept incoming webhooks and enable you to build powerful integrations.

API

In contrast to webhooks, the API works the other way around. Your application or no-code app can request information from MrScraper at any time by performing an API call.

This is mostly used when real-time is not needed or when we have to check information that is not sent by the webhooks.

No code apps

No-code apps are a combination of the previous two options, but made easy for non-technical people.

Apps such as Zapier or Make.com allow you to receive webhooks or perform API requests without having any programming knowledge.

Furthermore, MrScraper has an official Zapier app to make things even easier without having to write a single line of code.

Let's integrate our scraper

We will use the scraper we just created to add MrScraper blog posts to a Google Sheet.

Create a new Zap

Create or log in to your Zapier account and create a new Zap.

Add the MrScraper app

First of all, we'll need to create the step that triggers the new Zap. Select the MrScraper app and choose the "New Result" event.

Once MrScraper has a new scraping result ready, it will trigger this zap.

MrScraper create zap step 1

Convert the results to line items

The next step is to transform the scraping results into a format that Zapier can understand, referred to as "line items."

We will utilize the Zapier code tool to generate a JSON object from the scraping results.

MrScraper create zap step 2

Add to the database

Lastly, we will save the posts to a database. In this example, we will utilize Zapier's new databases, which make it incredibly easy to store information.

Create a database, add the columns for title and link, and attach it to this final step.

MrScraper create zap step 3

And we are done!

MrScraper create zap step 4

Integrating your scraper is as simple as that. There are lots of possibilities the limit is your imagination!

Blur logo

Community & Support

Head over to our community where you can engage with us and our community directly.

Questions? Ask our team via live chat 24/5 or just poke us on our official Twitter or our founder. We’re always happy to help.

Help center →
avatar

John Madrak

Founder, Waddling Technology

We're able to quickly and painlessly create automated
scrapers across a variety of sites without worrying about
getting blocked (loading JS, rotating proxies, etc.),
scheduling, or scaling up when we want more data
- all we need to do is open the site that we want to
scrape in devtools, find the elements that we want to
extract, and MrScraper takes care of the rest! Plus, since
MrScraper's pricing is based on the size of the data that
we're extracting it's quite cheap in comparison to most
other services. I definitely recommend checking out
MrScraper if you want to take the complexity
out of scraping.

avatar

Kim Moser

Computer consultant

Now that I've finally set-up and tested my first scraper,
I'm really impressed. It was much easier to set up than I
would have guessed, and specifying a selector made it
dead simple. Results worked out of the box, on a site
that is super touch about being scraped.

avatar

John

MrScraper User

I actually never expected us to be making this many
requests per month but MrScraper is so easy that we've
been increasing the amount of data we're collecting -
I have a few more scrapers that I need to add soon.
You're truly building a great product.

avatar

Ben

Russel

If you're needing a webscaper, for your latest project,
you can't go far wrong with MrScraper. Really clean,
intuitive UI. Easy to create queries. Great support.
Free option, for small jobs. Subscriptions for
larger volumes.

avatar

John Madrak

Founder, Waddling Technology

We're able to quickly and painlessly create automated
scrapers across a variety of sites without worrying about
getting blocked (loading JS, rotating proxies, etc.),
scheduling, or scaling up when we want more data
- all we need to do is open the site that we want to
scrape in devtools, find the elements that we want to
extract, and MrScraper takes care of the rest! Plus, since
MrScraper's pricing is based on the size of the data that
we're extracting it's quite cheap in comparison to most
other services. I definitely recommend checking out
MrScraper if you want to take the complexity
out of scraping.

avatar

Kim Moser

Computer consultant

Now that I've finally set-up and tested my first scraper,
I'm really impressed. It was much easier to set up than I
would have guessed, and specifying a selector made it
dead simple. Results worked out of the box, on a site
that is super touch about being scraped.

avatar

John

MrScraper User

I actually never expected us to be making this many
requests per month but MrScraper is so easy that we've
been increasing the amount of data we're collecting -
I have a few more scrapers that I need to add soon.
You're truly building a great product.

avatar

Ben

Russel

If you're needing a webscaper, for your latest project,
you can't go far wrong with MrScraper. Really clean,
intuitive UI. Easy to create queries. Great support.
Free option, for small jobs. Subscriptions for
larger volumes.