An Introduction to Scraping Websites with Axios and Cheerio

An Introduction to Scraping Websites with Axios and Cheerio
Last edit: Oct 06, 2022

We have a lot of articles about scraping using Python, but it isn't the only language suitable for data extraction. Therefore, today we will talk about NodeJS scraping.

Firstly, let's look at general information about what it is and how to prepare your computer to work with NodeJS and build the scraper using Axios and Cheerio.

Introduction to NodeJS

NodeJS is not a different programming language but some kind of environment that allows you to work with JavaScript outside the browser. It has the same base engine as Chrome - V8. 

So it will be easy to start for those who have already dealt with JavaScript in the frontend.

To put it simply, Node.js is a C++ application that takes JavaScript code as input and executes it.

The primary advantage of NodeJS is that it allows you to implement an asynchronous architecture in a single thread. In addition, thanks to an active community, NodeJS has ready-made solutions for all occasions, but we will discuss this later.

So let's prepare for scraping:

  1. Install NodeJS.
  2. Let's update NPM.

To install NodeJS, visit the official site and download the installation file. Two versions are currently available:

Available Node.js versions
Available Node.js versions

Version 16.17.1 LTS is stable, and 18.9.1 is the latest. The choice of version is up to the user and will not affect the example in this article.

After downloading and installing the package, check that everything works well. You can do this on the command line using the commands:

node -v
npm -v

To update NPM, you can run the following command:

npm install -g npm

After that, you can start working with NPM packages.

NPM packages or Solutions for All Situations

NPM (node package manager) is a standard package manager used to download packages from the npm cloud server or upload packages to these servers by any user. It means that every user who has written a JavaScript library or tool to perform their tasks can upload it to the server, after which anyone can upload it.

With this approach, there are many ready-made NPM packages today, and if you wish, you can find a ready-made solution for almost any problem or task.

To write a simple scraper, we need two of these tools:

  1. A tool that allows you to execute a request and get the page code.
  2. A tool that processes the source code and gets specific data.

Let’s try to decide which libraries will be better for us.

Work with Requests

There are two popular libraries for getting the page code and sending POST and GET requests in NodeJS: Axios and Fetch. However, even though Axios has a lot of flaws, and Fetch is a relatively new library that doesn't have such errors, Axios is still more popular. And all because:

  1. Code using Axios is more concise.
  2. The Axios library reports network errors, while the Fetch API doesn't.
  3. Axios can monitor the progress of unloading data, but Fetch isn't.

In addition to these two libraries, also worth mentioning are Ky (a tiny and well-thought-out HTTP client based on window.fetch) and Superagent (a small HTTP client library based on XMLHttpRequest). 

When the project is small, and it is possible to implement the last two libraries, it is better to use them.

Parse Library

To date, Cheerio is the most popular parser that creates the DOM tree of the page and allows you to work with it conveniently. Cheerio parses the markup and provides functions to process the received data.

As an analog of Cheerio, you can use Osmosis, which is similar in functionality but has fewer dependencies. 

Creating a project and adding Axios

So, we will use Axios and Cheerio. Let’s specify a directory and initiate NPM in it. Go to the folder where your scraper will be and run the command line in this folder. Then enter the command:

npm init -y

This will create a package.json file in the folder describing our project and specifying its dependencies.

Let's get down to the central part of creating a scraper. We will consider it on the example page Axios allows you to make requests and receive data. For this project, we only need HTTP GET requests.

Write the dependency on the command line:

npm install axios

Let's now try to get the page code. To do this, in the same folder, create an index.js file. In it, we will connect Axios, execute the request and display it on the screen:

const axios = require('axios'); 
const url = '';
    .then(response => {

Also, in the end, we will add an error handler so that the program, if an error occurs, is not interrupted and display it on the screen:

 .catch(error => {

To run the project, go to the command line and run the command:

node index.js page source code will be displayed on the screen:

<!doctype html>
    <title>Example Domain</title>
    <meta charset="utf-8" />
    <meta http-equiv="Content-type" content="text/html; charset=utf-8" />
    <meta name="viewport" content="width=device-width, initial-scale=1" />
    <style type="text/css">
    body {
        background-color: #f0f0f2;
        margin: 0;
        padding: 0;
        font-family: -apple-system, system-ui, BlinkMacSystemFont, "Segoe UI", "Open Sans", "Helvetica Neue", Helvetica, Arial, sans-serif;
    div {
        width: 600px;
        margin: 5em auto;
        padding: 2em;
        background-color: #fdfdff;
        border-radius: 0.5em;
        box-shadow: 2px 3px 7px 2px rgba(0,0,0,0.02);
    a:link, a:visited {
        color: #38488f;
        text-decoration: none;
    @media (max-width: 700px) {
        div {
            margin: 0 auto;
            width: auto;
    <h1>Example Domain</h1>
    <p>This domain is for use in illustrative examples in documents. You may use this
    domain in literature without prior coordination or asking for permission.</p>
    <p><a href="">More information...</a></p>

Now let's try to get only the title in the <h1> tag.

Processing a request with Cheerio

Cheerio allows you to parse materials that are HTML and XML markup. So, we can use it to get specific data from the source code we just received.

Let's go to the command line and add dependencies:

npm install cheerio

And also include it in the index file index.js:

const cheerio = require('cheerio');

Let's assume that we can have more than one header, and we want to get them all:

  .then(({ data }) => {  
    const $ = cheerio.load(data); 
    const title = $('h1') 
      .map((_, titles) => { 
        const $titles = $(titles); 
        return $titles.text() 

You can set up more complex queries or select specific data using CSS selectors.

Read more about CSS selectors: "The Ultimate CSS Selectors Cheat Sheet for Web Scraping."

Example vs. Reality

Our code worked on a simple HTML page, but for most sites, the structure isn't static – it's dynamic. Moreover, if we constantly scrape the site, we will be banned or offered to solve a captcha. Such a simple scraper, unfortunately, will not cope with this.

However, to solve the first problem, it is enough to use headless browsers, to solve the second problem - proxies, and to solve the third one - services for solving captchas. But there is an even simpler solution - to use the web scraping API, which will solve all these difficulties. 

Let's try to explore the home page and get the article titles from it using the web scraping API and NodeJS. To use the API, register on the Scrape-It.Cloud and copy the API key from your account. We will use it later.

To understand which selectors contain the information we need, go to the site and the developer tools or DevTools (F12). 

If you haven't worked with DevTools before, we recommend reading our article: «5 Chrome DevTools Tips & Tricks for Scraping».

After getting on DevTools, press Shift+Ctrl+C to be able to select an element on the page and view its code, then choose the title of any article. home page home page

The code highlights this element on the right side of the window. Great, now let's go to the site and go to the Web Scraping API tab. In the request builder, select NodeJS and get ready-made code for page scraping.

However, this is not enough. Let's set up our request so that it returns precisely the titles of the articles that are stored in the h2 tag. To do this, we use Extraction Rules.

Request Builder
Request Builder

We can run the resulting script from the site or save it to a file. Let's save it to a file and change the result output line (line 24) to the following code:


This is necessary to display only the titles of the articles. If we do it, we will get the list:

"Saturn’s Rings are Made of a Broken-Up Moon",
"What does Earth look like from across the Universe?",
"Have Rocket, Will Travel: The Coming Age of Intercontinental Spaceflight",
"Why Should You Care About the Bar Exam?",
"Are Iranian women defining the point of no return?",
"Corn is in Nearly Everything We Eat. Does It Matter?",
"Health Care Roulette",
"America’s Brief Experiment With Universal Healthcare Is Ending",
"There is More than Meets the Eye in Patagonia’s Bold Activism",
"It’s What You See",
"How Civil Rights Groups are Pushing Back Against Amazon’s New Show",
"Why “Microhistories” Rock",
"How to Be a Tourist",
"You Don’t Feel My Pain",
"The Servitude Of Our Times",
"10 Mental Models for Learning Anything",
"The Joy of Repairing Laptops",
"Why Adnan Syed Is Free"

Such scraping is safe and easy and helps avoid blocks and scrape dynamic data.

Conclusion and Takeaways

NodeJS is a very flexible and user-friendly data scraping tool that allows you to scrape data from various sites quickly and efficiently.

Full example code without API:

const axios = require('axios'); 
const cheerio = require('cheerio');
const url = '';
  .then(({ data }) => {  
    const $ = cheerio.load(data); 
    const title = $('h1') 
      .map((_, titles) => { 
        const $titles = $(titles); 
        return $titles.text() 

Full example code with API:

var axios = require('axios');
var data = JSON.stringify({
  "extract_rules": {
    "Title": "h2"
  "wait": 0,
  "screenshot": true,
  "block_resources": false,
  "url": ""
var config = {
  method: 'post',
  url: '',
  headers: {
    'x-api-key': 'YOUR-API-KEY',
    'Content-Type': 'application/json'
  data : data
.then(function (response) {
.catch(function (error) {

If the topic interests you, you can find out more in the articles “Web Scraping with node.js: How to Leverage the power of JavaScript” and “How to scrape Yelp using NodeJS.”

Tired of getting blocked while scraping the web?

Try out Web Scraping API with proxy rotation, CAPTCHA bypass, and Javascript rendering.

  • 1,000 Free API Credits
  • No Credit Card Required
  • 30-Day Trial
Try now for free

Collect structured data without any coding!

Our no-code scrapers make it easy to extract data from popular websites with just a few clicks.

  • CSV, XLSX, and JSON Formats
  • No Coding or Software Required
  • Save Time and Effort
Scrape with No Code
Valentina Skakun

I'm a technical writer who believes that data parsing can help in getting and analyzing data. I'll tell about what parsing is and how to use it.