Site icon R-bloggers

How to Scrape Data from a JavaScript Website with R

[This article was first published on Curiosity Killed the Cat, and kindly contributed to R-bloggers]. (You can report issue about the content on this page here)
Want to share your content on R-bloggers? click here if you have a blog, or here if you don't.

In September 2017, I found myself working on a project that required odds data for football. At the time I didn’t know about resources such as Football-Data or the odds-api, so I decided to build a scraper to collect data directly from the bookmakers. However, most of them used JavaScript to display their odds, so I couldn’t collect the data with R and rvest alone. In this article, I’ll demonstrate how PhantomJS can be used with R to scrape JS-rendered content from the web.

< !--more-->

Requirements

In addition to R’s base packages, I’ll need the following for this example:

Collecting Data

For this example I’ll stick to the original intent of the project and collect odds from an online bookmaker. Paddy Power (no affiliation) uses JavaScript and also doesn’t have an API to my knowledge, so I’ll collect odds from their website.

Without PhantomJS

To collect odds data with rvest, we need to:

  1. Read the HTML file from a URL.
  2. Extract the HTML elements containing the odds. (The <div> elements that contain the data I am interested in have the attribute class="avb-item".)
  3. Extract the contents of the elements.
  4. Clean the data if necessary.
> withoutJS <- xml2::read_html("https://www.paddypower.com/football?tab=today") %>%
+   rvest::html_nodes(".avb-item") %>%
+   rvest::html_text()

> withoutJS
character(0)

We ended up with an empty character vector because the HTML document contains only JS snippets in the source code instead of the content we see in the browser. That means the CSS selector we can see in the browser does not exist in the document R is trying to extract data from.

With PhantomJS

The first step is to save the following script in a separate file:

// scraper_PaddyPower.js

// Create a webpage object
var page = require('webpage').create();

// Include the File System module for writing to files
var fs = require('fs');

// Specify source and path to output file
var url  = 'https://www.paddypower.com/football?tab=today'
var path = 'odds_PaddyPower.html'

page.open(url, function (status) {
  var content = page.content;
  fs.write(path,content,'w')
  phantom.exit();
});

The purpose of this script is to retrieve the HTML file from the specified URL and store it into a local HTML file, so that R can read contents from that file instead of reading the contents directly from the URL.

We can then use the system() function to invoke the script from R and repeat the steps outlined in the attempt without PhantomJS to extract the data:

> system("/usr/local/phantomjs-2.1.1/bin/phantomjs scraper_PaddyPower.js")

> withJS <- xml2::read_html("odds_PaddyPower.html") %>%
+   rvest::html_nodes(".avb-item") %>%
+   rvest::html_text()

Note for Windows users: The invocation of the script has the same format as the command for Linux, but you may have to add .exe to phantomjs for it to work.

The content is currently a mess and the variable is one large chunk of text, i.e. a character vector of length 1. If I remove the white space and split it into individual elements, here’s what I get:

> pp_odds_data <- withJS %>% 
+   gsub("[\n\t]", "", .) %>%
+   stringr::str_trim() %>%
+   gsub("\\s+", " ", .) %>%
+   strsplit(" ") %>% 
+   unlist()

> pp_odds_data[1:40]
 [1] "Bahrain"        "Tajikistan"     "4/7"            "13/5"           "5/1"            "17:00"          "More"
 [8] "Bets"           ">"              "Hatayspor"      "Genclerbirligi" "1/14"           "13/2"           "19/1"
[15] "More"           "Bets"           ">"              "Darica"         "Genclerbirligi" "Antalyaspor"    "4/1"
[22] "23/10"          "4/7"            "More"           "Bets"           ">"              "Akhisar"        "Belediye"
[29] "Fatih"          "Karagumrukspor" "1/3"            "10/3"           "6/1"            "16:00"          "More"
[36] "Bets"           ">"              "Giresunspor"    "Fenerbahce"     "9/2"

There’s still some useless data mixed in, and there is no guarantee that I can turn this vector into a data frame to store and analyze the data. In this particular case, team names can have more than two words, so I should extract information about matches and odds separately:

> matches <- xml2::read_html("odds_PaddyPower.html") %>%
+   rvest::html_nodes(".avb-item .ui-scoreboard-coupon") %>%
+   rvest::html_text() %>%
+   stringr::str_trim() %>%
+   gsub("\\s\\s+"," vs ", .)

> odds <- xml2::read_html("odds_PaddyPower.html") %>%
+   rvest::html_nodes(".avb-item .btn-odds") %>%
+   rvest::html_text() %>%
+   stringr::str_trim()

> odds_selector <- seq(1, length(odds), 3)
> pp_odds_df <- data.frame(Match = matches, 
                           Home = odds[odds_selector], 
                           Draw = odds[(odds_selector + 1)], 
                           Away = odds[(odds_selector + 2)])
                           
> head(pp_odds_df)
                                     Match  Home  Draw  Away
1                    Bahrain vs Tajikistan   4/7  13/5   5/1
2              Hatayspor vs Genclerbirligi  1/14  13/2  19/1
3     Darica Genclerbirligi vs Antalyaspor   4/1 23/10   4/7
4 Akhisar Belediye vs Fatih Karagumrukspor   1/3  10/3   6/1
5                Giresunspor vs Fenerbahce   9/2  11/4   1/2
6             Botosani vs Dunarea Calarasi  8/13   5/2   9/2

Now I can further transform the data before storing and analyzing it, e.g. convert fractional odds into decimal, split the Match column into “Home Team” and “Away Team” columns, and so on.

Final Comments

Although the script I used still works today, I should note that the development of PhantomJS is currently suspended and its git repository is archived. However, there are many ways to collect data from a JavaScript-rendered website, so you can also try using rvest and V8 from R or Python and Selenium to collect the data you need. Good luck!

To leave a comment for the author, please follow the link and comment on their blog: Curiosity Killed the Cat.

R-bloggers.com offers daily e-mail updates about R news and tutorials about learning R and many other topics. Click here if you're looking to post or find an R/data-science job.
Want to share your content on R-bloggers? click here if you have a blog, or here if you don't.