Easy Web Scraping With Node.js

Posted by
on under

Web scraping is a technique used to extract data from websites using a computer program that acts as a web browser. The program requests pages from web servers in the same way a web browser does, and it may even simulate a user logging in to obtain access. It downloads the pages containing the desired data and extracts the data out of the HTML code. Once the data is extracted it can be reformatted and presented in a more useful way.

In this article I'm going to show you how to write web scraping scripts in Javascript using Node.js.

Why use web scraping?

Here are a few of examples where web scraping can be useful:

  • You have several bank accounts with different institutions and you want to generate a combined report that includes all your accounts.
  • You want to see data presented by a website in a different format. For example, the website shows a table, but you want to see a chart.
  • A web site presents related data across multiple pages. You want to see all or part of this data combined in a single report.
  • You are an app developer with apps in iTunes and several Android app stores, and you want to have a report of monthly sales across all app stores.

Web scraping can also be used in ways that are dishonest and sometimes even illegal. Harvesting of email addresses for spam purposes, or sniping Ebay auctions are examples of such uses. As a matter of principle I only use web scraping to collect and organize information that is either available to everyone (stock prices, movie showtimes, etc.) or only available to me (personal bank accounts, etc.). I avoid using this technique for profit, I just do it to simplify the task of obtaining information.

In this article I'm going to show you a practical example that implements this technique. Ready? Then let's get started!

Tools for web scraping

In its most basic form, a web scraping script just needs to have a way to download web pages and then search for data in them. All modern languages provide functions to download web pages, or at least someone wrote a library or extension that can do it, so this is not a problem. Locating and isolating data in HTML pages, however, is difficult. An HTML page has content, layout and style elements all intermixed, so a non trivial effort is required to parse and identify the interesting parts of the page.

For example, consider the following HTML page:

<html>
    <head>...</head>
    <body>
        <div id="content">
            <div id="sidebar">
            ...
            </div>
            <div id="main">
                <div class="breadcrumbs">
                ...
                </div>
                <table id="data">
                    <tr><th>Name</th><th>Address</th></tr>
                    <tr><td class="name">John</td><td class="address">Address of John</td></tr>
                    <tr><td class="name">Susan</td><td class="address">Address of Susan</td></tr>
                </table>
            </div>
        </div>
    </body>
</html>

Let's say we want to extract the names of the people that appear in the table with id="data" that is in the page. How do we get to those?

Typically the web page will be downloaded into a string, so it would be simple to just search this string for all the occurrences of <td class="name"> and extract what comes after that and until the following </td>.

But this could easily make us find incorrect data. The page could have other tables, either before or after the one we want that use the same CSS classes for some of its cells. Or worst, maybe this simple search algoritm works fine for a while, but one day the layout of the page changes so that the old <td class="name"> becomes <td align="left" class="name"> making our search find nothing.

While there is always a risk that a change to the target web page can break a scraping script, it is a good idea to be smart about how items are located in the HTML so that the script does not need to be revised every time the web site changes.

If you have ever written client-side Javascript for the browser using a library like jQuery then you know how the tricky task of locating DOM elements becomes much easier using CSS selectors.

For example, in the browser we could easily extract the names from the above web page as follows:

$('#data .name').each(function() {
    alert($(this).text());
});

The CSS selector is what goes inside jQuery's $ function, #data .name in this example, This is saying that we want to locate all the elements that are children of an element with the id data and have a CSS class name. Note that we are not saying anything about the data being in a table in this case. CSS selectors have great flexibility in how you specify search terms for elements, and you can be as specific or vague as you want.

The each function will just call the function given as an argument for all the elements that match the selector, with the this context set to the matching element. If we were to run this in the browser we would see an alert box with the name John, and then another one with the name "Susan".

Wouldn't it be nice if we could do something similar outside of the context of a web browser? Well, this is exactly what we are about to do.

Introducing Node.js

Javascript was born as a language to be embedded in web browsers, but thanks to the open source Node.js project we can now write stand-alone scripts in Javascript that can run on a desktop computer or even on a web server.

Manipulating the DOM inside a web browser is something that Javascript and libraries like jQuery do really well so to me it makes a lot of sense to write web scraping scripts in Node.js, since we can use many techniques that we know from DOM manipulation in the client-side code for the web browser.

If you would like to try the examples I will present in the rest of this article then this is the time to download and install Node.js. Installers for Windows, Linux and OS X are available at http://nodejs.org.

Node.js has a large library of packages that simplify different tasks. For web scraping we will use two packages called request and cheerio. The request package is used to download web pages, while cheerio generates a DOM tree and provides a subset of the jQuery function set to manipulate it. To install Node.js packages we use a package manager called npm that is installed with Node.js. This is equivalent to Ruby's gem or Python's easy_install and pip, it simplifies the download and installation of packages.

So let's start by creating a new directory where we will put our web scraping scripts and install these two modules in it:

$ mkdir scraping
$ cd scraping
$ npm install request cheerio

Node.js modules will be installed in the scraping/node_modules subdirectory and will only be accessible to scripts that are in the scraping directory. It is also possible to install Node.js packages globally, but I prefer to keep things organized by installing modules locally.

Now that we have all the tools installed let's see how we can implement the above scraping example using cheerio. Let's call this script example.js:

var cheerio = require('cheerio');
$ = cheerio.load('<html><head></head><body><div id="content"><div id="sidebar"></div><div id="main"><div id="breadcrumbs"></div><table id="data"><tr><th>Name</th><th>Address</th></tr><tr><td class="name">John</td><td class="address">Address of John</td></tr><tr><td class="name">Susan</td><td class="address">Address of Susan</td></tr></table></div></div></body></html>');

$('#data .name').each(function() {
    console.log($(this).text());
});

The first line imports the cheerio package into the script. The require statement is similar to #include in C/C++, require in Ruby or import in Python.

In the second line we instantiate a DOM for our example HTML, by sending the HTML string to cheerio.load(). The return value is the constructed DOM, which we store in a variable called $ to match how the DOM is accessed in the browser when using jQuery.

Once we have a DOM created we just go about business as if we were using jQuery on the client side. So we use the proper selector and the each iterator to find all the occurrences of the data we want to extract. In the callback function we use the console.log function to write the extracted data. In Node.js console.log writes to the console, so it is handy to dump data to the screen.

Here is how to run the script and what output it produces:

$ node example.js
John
Susan

Easy, right? In the following section we'll write a more complex scraping script.

Real world scraping

Let's use web scraping to solve a real problem.

The Tualatin Hills Park and Recreation District (THPRD) is a Beaverton, Oregon organization that offers area residents a number of recreational options, among them swimming. There are eight swimming pools, all in the area, each offering swimming instruction, lap swimming, open swim and a variety of other programs. The problem is that THPRD does not publish a combined schedule for the pools, it only publishes individual schedules for each pool. But the pools are all located close to each other, so many times the choice of pool is less important than what programs are offered at a given time. If I wanted to find the time slots a given program is offered at any of the pools I would need to access eight different web pages and search eight schedules.

For this example we will say that we want to obtain the list of times during the current week when there is an open swim program offered in any of the pools in the district. This requires obtaining the schedule pages for all the pools, locating the open swim entries and listing them.

Before we start, click here to open one of the pool schedules in another browser tab. Feel free to inspect the HTML for the page to familiarize yourself with the structure of the schedule.

The schedule pages for the eight pools have a URL with the following structure:

http://www.thprd.org/schedules/schedule.cfm?cs_id=<id>

The id is what selects which pool to show a schedule for. I took the effort to open all the schedules manually to take note of the names of each pool and its corresponding id, since we will need those in the script. We will also use an array with the names of the days of the week. We can scrape these names from the web pages, but since this is information that will never change we can simplify the script by incorporating the data as constants.

Web scraping skeleton

With the above information we can sketch out the structure of our scraping script. Let's call the script thprd.js:

var request = require('request');
var cheerio = require('cheerio');

pools = {
    'Aloha': 3,
    'Beaverton': 15,
    'Conestoga': 12,
    'Harman': 11,
    'Raleigh': 6,
    'Somerset': 22,
    'Sunset': 5,
    'Tualatin Hills': 2
};
days = ['Monday', 'Tuesday', 'Wednesday', 'Thursday', 'Friday', 'Saturday', 'Sunday'];

for (pool in pools) {
    var url = 'http://www.thprd.org/schedules/schedule.cfm?cs_id=' + pools[pool];
    request(url, function(err, resp, body) {
        if (err)
            throw err;
        $ = cheerio.load(body);
        console.log(pool);
        // TODO: scraping goes here!
    });
}

We begin the script importing the two packages that we are going to use and defining the constants for the eight pools and the days of the week.

Then we download the schedule web pages of each of the pools in a loop. For this we construct the URL of each pool schedule and send it to the request function. This is an asynchronous function that takes a callback as its second argument. If you are not very familiar with Javascript this may seem odd, but in this language asynchronous functions are very common. The request() function returns immediately, so it is likely that the eight request() calls will be issued almost simultaneously and will be processed concurrently by background threads.

When a request completes its callback function will be invoked with three arguments, an error code, a response object and the body of the response. Inside the callback we make sure there is no error and then we just send the body of the response into cheerio to create a DOM from it. When we reach this point we are ready to start scraping.

We will look at how to scrape this content later, for now we just print the name of the pool as a placeholder. If you run this first version of our script you'll get a surprise:

$ node thprd.js
Tualatin Hills
Tualatin Hills
Tualatin Hills
Tualatin Hills
Tualatin Hills
Tualatin Hills
Tualatin Hills
Tualatin Hills

What? Why do we get the same pool name eight times? Shouldn't we see all the pool names here?

Javascript scoping

Remember I said above that the request() function is asynchronous? The for loop will do its eight iterations, spawning a background job in each. The loop then ends, leaving the loop variable set to the pool name that was used in the last iteration. When the callback functions are invoked a few seconds later they will all see this value and print it.

I made this mistake on purpose to demonstrate one of the main sources of confusion among developers that are used to traditional languages and are new to Javascript's asynchronous model.

How can we get the correct pool name to be sent to each callback function then?

The solution is to bind the name of the pool to the callback function at the time the callback is created and sent to the request() function, because that is when the pool variable has the correct value.

As we've seen before the callback function will execute some time in the future, after the loop in the main script completed. But the callback function can still access the loop variable even though the callback runs outside of the context of the main script. This is because the scope of Javascript functions is defined at the time the function is created. When we created the callback function the loop variable was in scope, so the variable is accessible to the callback. The url variable is also in the scope, so the callback can also make use of it if necessary, though the same problem will exist with it, its last value will be seen by all callbacks.

So what I'm basically saying is that the scope of a function is determined at the time the function is created, but the values of the variables in the scope are only retrieved at the time the function is called.

We can take advantage of these seemingly odd scoping rules of Javascript to insert any variable into the scope of a callback function. Let's do this with a simple function:

function main()
{
    var a = 1;
    var f = function() { console.log(a); }
    a = 2;
    f();
}
main();

Can you guess what the output of this script will be? The output will be 2, because that's the value of variable a at the time the function stored in variable f is invoked.

To freeze the value of a at the time f is created we need to insert the current value of a into the scope:

function main()
{
    var a = 1;
    var f = ( function(a) { return function() { console.log(a); } } )(a);
    a = 2;
    f();
}
main();

Let's analyze this alternative way to create f one step at a time:

var f = (...)(a);

We clearly see that the expression enclosed in parenthesis supposedly returns a function, and we invoke that function and pass the current value of a as an argument. This is not a callback function that will execute later, this is executing right away, so the current value of a that is passed into the function is 1.

var f = ( function(a) { return ... } )(a);

Here we see a bit more of what's inside the parenthesis. The expression is, in fact, a function that expects one argument. We called that argument a, but we could have used a different name.

In Javascript a construct like the above is called a self-executing function. You could consider this the reverse of a callback function. While a callback function is a function that is created now but runs later, a self-executing function is created and immediately executed. Whatever this function returns will be the result of the whole expression, and will get assigned to f in our example.

Why would you want to use a self-executing function when you can make any code execute directly without enclosing it inside a function? The difference is subtle. By putting code inside a function we are creating a new scope level, and that gives us the chance to insert variables into that scope simply by passing them as arguments to the self-executing function.

We know f should be a function, since later in the script we want to invoke it. So the return value of our self-executing function must be the function that will get assigned to f:

var f = ( function(a) { return function() { console.log(a); } } )(a);

Does it make a bit more sense now? The function that is assigned to f now has a parent function that received a as an argument. That a is a level closer than the original a in the scope of f, so that is the a that the scope of f sees. When you run the modified script you will get a 1 as output.

Here is how the self-executing trick can be applied to our web scraping script:

for (pool in pools) {
    var url = 'http://www.thprd.org/schedules/schedule.cfm?cs_id=' + pools[pool];
    request(url, ( function(pool) {
        return function(err, resp, body) {
            if (err)
                throw err;
            $ = cheerio.load(body);
            console.log(pool);
            // TODO: scraping goes here!
        }
    } )(pool));
}

This is pretty much identical to the simpler example above using the pool variable instead of a. Running this script again gives us the expected result:

$ node thprd.js
Aloha
Raleigh
Conestoga
Beaverton
Harman
Somerset
Tualatin Hills
Sunset

Scraping the swimming pool schedules

To be able to scrape the contents of the schedule tables we need to discover how these schedules are structured. In rough terms the schedule table is located inside a page that looks like this:

<html>
<head> ... </head>
<body>
    <div id="container">
        <div id="mainContent_lv13">
            <div id="level3body">
                <div>
                    <table id="calendar">
                        <tr class="header"> ... </tr>
                        <tr class="days">
                            <td><!-- schedule for Monday --></td>
                            <td><!-- schedule for Tuesday --></td>
                            <td><!-- schedule for Wednesday --></td>
                            <td><!-- schedule for Thursday --></td>
                            <td><!-- schedule for Friday --></td>
                            <td><!-- schedule for Saturday --></td>
                            <td><!-- schedule for Sunday --></td>
                        </tr>
                        <tr class="footer"> ... </tr>
                    </table>
                </div>
            </div>
        </div>
    </div>
</body>
</html>

Inside each of these <td> elements that hold the daily schedules there is a <div> wrapper around each scheduled event. Here is a simplified structure for a day:

<td>
    <a href="...">link</a>
    <div>
        <a href="...">
            <strong>FROM-TO</strong>
            <br>
            EVENT NAME
            <br>
        </a>
    </div>
    <div>
        <a href="...">
            <strong>FROM-TO</strong>
            <br>
            EVENT NAME
            <br>
        </a>
    </div>
    ...
</td>

Each <td> element contains a link at the top that we are not interested in, then a sequence of <div> elements, each containing the information for an event.

One way we can get to these event <div> elements is with the following selector:

$('#calendar .days td div').each(...);

The problem with the above selector, though, is that we will get all the events of all the days in sequence, so we will not know what events happen on which day.

Instead, we can separate the search in two parts. First we locate the <td> element that defines a day, then we search for <div> elements within it:

$('#calendar .days td').each(function(day) {
    $(this).find('div').each(function() {
        console.log(pool + ',' + days[day] + ',' + $(this).text());
    });
});

The function that we pass to the each() iterator receives the index number of the found element as a first argument. This is handy because for our outer search this is telling us which day we are in. We do not need an index number in the inner search, so there we do not need to use an argument in our function.

Running the script now shows the pool name, then the day of the week and then the text inside the event <div>, which has the information that we want. The text() function applied to any element of the DOM returns the constant text filtering out any HTML elements, so this gets rid of the <strong> and <br> elements that exist there and just returns the filtered text.

We are now very close. The only remaining problem is that the text we extracted from the <div> element has a lot of whitespace in it. There is whitespace at the start and end of the text and also in between the event time and event description. We can eliminate the leading and trailing whitespace with trim():

console.log(pool + ',' + days[day] + ',' + $(this).text().trim());

This leaves us with a few lines of whitespace in between the event time and the description. To remove that we can use replace():

console.log(pool + ',' + days[day] + ',' + $(this).text().trim().replace(/\s\s+/g, ','));

Note the regular expression that we use to remove the spaces requires at least two whitespace characters. This is because the event description can contain spaces as well, if we search for two or more spaces we will just find the large whitespace block in the middle and not affect the description.

When we run the script now this is what we get:

$ node thprd.js
Raleigh,Monday,6:00a-10:00p,Pool Closed
Raleigh,Tuesday,6:00a-10:00p,Pool Closed
Raleigh,Wednesday,6:00a-10:00p,Pool Closed
Raleigh,Thursday,6:00a-10:00p,Pool Closed
Raleigh,Friday,6:00a-10:00p,Pool Closed
Raleigh,Saturday,6:00a-10:00p,Pool Closed
Raleigh,Sunday,6:00a-10:00p,Pool Closed
Beaverton,Monday,7:00a-8:00a,Deep Water Aerobics (7-8)
Beaverton,Monday,7:00a-8:50a,Aquajog lane
Beaverton,Monday,7:00a-8:50a,All Age Lap
Beaverton,Monday,8:55a-10:30a,Aquajog lane^
...

And this is just a CSV version of all the pool schedules combined!

We said that for this exercise we were only interested in obtaining the open swim events, so we need to add one more filtering layer to just print the targeted events:

event = $(this).text().trim().replace(/\s\s+/g, ',').split(',');
if (event.length >= 2 && event[1].match(/open swim/i))
    console.log(pool + ',' + days[day] + ',' + event[0] + ',' + event[1]);

And now we have completed our task. Here is the final version of our web scraping script:

var request = require('request');
var cheerio = require('cheerio');

days = ['Monday', 'Tuesday', 'Wednesday', 'Thursday', 'Friday', 'Saturday', 'Sunday'];
pools = {
    'Aloha': 3,
    'Beaverton': 15,
    'Conestoga': 12,
    'Harman': 11,
    'Raleigh': 6,
    'Somerset': 22,
    'Sunset': 5,
    'Tualatin Hills': 2
};
for (pool in pools) {
    var url = 'http://www.thprd.org/schedules/schedule.cfm?cs_id=' + pools[pool];

    request(url, (function(pool) { return function(err, resp, body) {
        $ = cheerio.load(body);
        $('#calendar .days td').each(function(day) {
            $(this).find('div').each(function() {
                event = $(this).text().trim().replace(/\s\s+/g, ',').split(',');
                if (event.length >= 2 && (event[1].match(/open swim/i) || event[1].match(/family swim/i)))
                    console.log(pool + ',' + days[day] + ',' + event[0] + ',' + event[1]);
            });
        });
    }})(pool));
}

Running the script gives us this output:

$ node thprd.js
Conestoga,Monday,4:15p-5:15p,Open Swim - M/L
Conestoga,Monday,7:45p-9:00p,Open Swim - M/L
Conestoga,Tuesday,7:30p-9:00p,Open Swim - M/L
Conestoga,Wednesday,4:15p-5:15p,Open Swim - M/L
Conestoga,Wednesday,7:45p-9:00p,Open Swim - M/L
Conestoga,Thursday,7:30p-9:00p,Open Swim - M/L
Conestoga,Friday,6:30p-8:30p,Open Swim - M/L
Conestoga,Saturday,1:00p-4:15p,Open Swim - M/L
Conestoga,Sunday,2:00p-4:15p,Open Swim - M/L
Aloha,Monday,1:05p-2:20p,Open Swim
Aloha,Monday,7:50p-8:25p,Open Swim
Aloha,Tuesday,1:05p-2:20p,Open Swim
Aloha,Tuesday,8:45p-9:30p,Open Swim
Aloha,Wednesday,1:05p-2:20p,Open Swim
Aloha,Wednesday,7:50p-8:25p,Open Swim
Aloha,Thursday,1:05p-2:20p,Open Swim
Aloha,Thursday,8:45p-9:30p,Open Swim
Aloha,Friday,1:05p-2:20p,Open Swim
Aloha,Friday,7:50p-8:25p,Open Swim
Aloha,Saturday,2:00p-3:30p,Open Swim
Aloha,Saturday,4:30p-6:00p,Open Swim
Aloha,Sunday,2:00p-3:30p,Open Swim
Aloha,Sunday,4:30p-6:00p,Open Swim
Harman,Monday,4:25p-5:30p,Open Swim*
Harman,Monday,7:30p-8:55p,Open Swim
Harman,Tuesday,4:25p-5:10p,Open Swim*
Harman,Wednesday,4:25p-5:30p,Open Swim*
Harman,Wednesday,7:30p-8:55p,Open Swim
Harman,Thursday,4:25p-5:10p,Open Swim*
Harman,Friday,2:00p-4:55p,Open Swim*
Harman,Saturday,1:30p-2:25p,Open Swim
Harman,Sunday,2:00p-2:55p,Open Swim
Beaverton,Tuesday,10:45a-12:55p,Open Swim (No Diving Well)
Beaverton,Tuesday,8:35p-9:30p,Open Swim No Diving Well
Beaverton,Thursday,10:45a-12:55p,Open Swim (No Diving Well)
Beaverton,Thursday,8:35p-9:30p,Open Swim No Diving Well
Beaverton,Saturday,2:30p-4:00p,Open Swim
Beaverton,Sunday,4:15p-6:00p,Open Swim
Sunset,Tuesday,1:00p-2:30p,Open Swim/One Lap Lane
Sunset,Thursday,1:00p-2:30p,Open Swim/One Lap Lane
Sunset,Sunday,1:30p-3:00p,Open Swim/One Lap Lane
Tualatin Hills,Monday,7:35p-9:00p,Open Swim-Diving area opens at 8pm
Tualatin Hills,Wednesday,7:35p-9:00p,Open Swim-Diving area opens at 8pm
Tualatin Hills,Sunday,1:30p-3:30p,Open Swim
Tualatin Hills,Sunday,4:00p-6:00p,Open Swim

From this point on it is easy to continue to massage this data to get it into a format that is useful. My next step would be to sort the list by day and time instead of by pool, but I'll leave that as an exercise to interested readers.

Final words

I hope this introduction to web scraping was useful to you and the example script serves you as a starting point for your own projects.

If you have any questions feel free to leave them below in the comments section.

Thanks!

Miguel

Become a Patron!

Hello, and thank you for visiting my blog! If you enjoyed this article, please consider supporting my work on this blog on Patreon!

68 comments
  • #1 Nano said

    Thanks, very well explained, is very useful me.

  • #2 fallanic said

    This is a great tutorial, very detailed.

    For those interested I also wrote a simple mode package to easily scrape a web page and save the result in a CSV file.

    https://npmjs.org/package/scrape2csv

  • #3 roshan agarwal said

    nice tutorial Miguel, we want to be get in touch with you and want to know you also work for private project.

  • #4 Kishore said

    Great article for beginners. How to handle login required sites with cookies.

  • #5 Miguel Grinberg said

    @Kishore: it gets a bit more complicated to handle logins. If the page requires HTTP Authentication then you just need to add a header to the request with the credentials (which could be given as command line arguments to the script, I guess). You can login to a site by posting the credentials as a form and then capture the cookie in the response. You will then send the cookie back in any follow up requests. I might write a second part to this article showing more advanced examples. Thanks!

  • #6 Matt said

    Great tutorial, thanks!

  • #7 Victor said

    I've tried using the request module and it doesn't seem to handle the case of dynamic websites where javascript is run after the page has finished loading. Do you have any ideas on how to handle these types of sites? The only one I can find at the moment is PhantomJS.

  • #8 Evis said

    Thanks for this tutorial, Miguel!
    How to submit a form for searching in a website?
    And if the form is more complicated containing radiobuttons or dropdownlists?

    Thanks!
    Evis

  • #9 Miguel Grinberg said

    @Victor: I don't know of any other solution besides PhantomJS.

  • #10 Miguel Grinberg said

    @Evis: the request module can send POST requests with form data.

  • #11 Marko said

    Thank you very much! Great tutorial and explanation.

    I was reading one book about Node.js (Sams edition) and after I've read the part about HTTP clients in node, I immediately thought of some kind of scraping tool with power of jQuery.

    Few seconds later, I'm here, reading and doing your example. Cool stuff... I've already started doing something for my own use after this.

  • #12 Jinjo Johnson said

    Nice one , Thank u ...

  • #13 Alexandru Cobuz said

    Impresive tutorial, thanks for showing how to use cheerio. I usually like to save the data into a database separating server from client.

  • #14 David Konsumer said

    Connection pooling is needed for anything larger than a trivial example, or the async nature of node will quickly spiral out of control with all the requests.

    See https://github.com/coopernurse/node-pool for an easy-to-use lib to help with this.

    See this for a complete example: https://github.com/sylvinus/node-crawler

    It uses jsdom, which I don't like as much as cheerio, but it's a good illustration of avoiding the connection limit.

  • #15 David Konsumer said

    also, jsdom even though a little heavy for jquery-like parsing tasks that cheerio excels at, is a full DOM implementation, so you can actually execute dynamic javascript pages, and inject your own code to parse it, in a lighter manner than phantomjs.

  • #16 Max said

    This was a fantastic introduction to scraping for a novice developer. Really, your explanations were very thorough. I could understand more experienced devs skipping certain parts but the fact that you put in all the extra explanation about async functions and self executing functions was great. Also the description of what was going on with the regexp. Thank you soo much!

  • #17 ponk said

    Hi Miguel--

    how would i utilize this data into my site? let's say i have a domain called example.com how would i get this data on that domain?

  • #18 Miguel Grinberg said

    @ponk: this tutorial is not for a web based scraper, this is just the scraping engine. You will call this scraper as a function from your website, maybe written in express.

  • #19 Carlos said

    Great work Miguel! Your explanation about JS scoping is very well written. Thank you!

  • #20 Trevor said

    Thanks for the great post. Looks like a great next step in my JS exploration. Also, I love the nod to THPRD. That was my first job and here I am again- full circle!

  • #21 dhar said

    hi Miguel-

    what if i'd like to scrap multiple urls? do you have any suggestion ?

  • #22 Miguel Grinberg said

    @dhar: this example does scrap multiple URLs. What are you thinking that is different than what I presented above?

  • #23 Mark Thien said

    Hi Mig,

    I tried to scrap

    http://www.finviz.com/screener.ashx?v=111&f=cap_small,fa_peg_u1,geo_usa,ta_beta_o1.5&ft=4

    and the body return undefined.

    below are the code:

    var cheerio = require('/usr/local/lib/node_modules/cheerio');
    var request = require('/usr/local/lib/node_modules/request');

    var url = 'http://www.finviz.com/screener.ashx?v=111&f=cap_small,fa_peg_u1,geo_usa,ta_beta_o1.5&ft=4';

    request(url, function(err, resp, body) {
    if (err) {

        throw err;
    
    }
    

    console.log(body);
    $ = cheerio.load(body);

    $('.tab-link').each(function() {
        console.log($(this).text());
    });
    

    });

    Would you kindly assist? thanks !

    Mark

  • #24 Miguel Grinberg said

    @Mark: Some web sites want to see a "User-Agent" header that they recognize. The following call makes it look like it is in iPad making the request, and this works fine:

    request({ url: url, headers: {'User-Agent': 'Mozilla/5.0 (iPad; U; CPU OS 3_2_1 like Mac OS X; en-us) AppleWebKit/531.21.10 (KHTML, like Gecko) Mobile/7B405'} }, function(err, resp, body) { ... });

  • #25 sotiris said

    Hello Miguel and thank you for your wonderful post. Although everything is working fine in most cases, i am facing the following problem. The content i would like to access is included in a jquery accordion (F.A.Q list). This means that part of it is is collapsed and the css display property of the surrounding divs is set to none witch in turn removes the content from the DOM. So when i donwload it there is not content. Do you have any idea how i can overcome this problem?

Leave a Comment