Web Scraping Google Play Books with Nodejs

What will be scraped

Using Google Play Books Store API from SerpApi

This section is to show the comparison between the DIY solution and our solution.

The biggest difference is that you don't need to create the parser from scratch and maintain it.

There's also a chance that the request might be blocked at some point from Google, we handle it on our backend so there's no need to figure out how to do it yourself or figure out which CAPTCHA, proxy provider to use.

First, we need to install google-search-results-nodejs:

npm i google-search-results-nodejs

Here's the full code example, if you don't need an explanation:

const SerpApi = require("google-search-results-nodejs");
const search = new SerpApi.GoogleSearch(process.env.API_KEY); //your API key from serpapi.com

const params = {
  engine: "google_play", // search engine
  gl: "us", // parameter defines the country to use for the Google search
  hl: "en", // parameter defines the language to use for the Google search
  store: "books", // parameter defines the type of Google Play store
  store_device: "phone", // parameter defines the search device. Options: phone, tablet, tv, chromebook, watch, car
  // if you need to find books from one of the categories you need to uncomment the "books_category" parameter
  // books_category: "audiobooks", // you can see the full list of supported categories on https://serpapi.com/google-play-books-categories
};

const getJson = () => {
  return new Promise((resolve) => {
    search.json(params, resolve);
  });
};

const getResults = async () => {
  const json = await getJson();
  const booksResults = json.organic_results.reduce((result, category) => {
    const { title: categoryTitle, items } = category;
    const books = items.map((book) => {
      const { title, link, rating = "No rating", original_price, price, thumbnail, product_id } = book;
      const returnedBook = {
        title,
        link,
        rating,
        price,
        thumbnail,
        movieId: product_id,
      };
      if (original_price) returnedBook.originalPrice = original_price;
      return returnedBook;
    });
    return {
      ...result,
      [categoryTitle]: books,
    };
  }, {});
  return booksResults;
};

getResults().then((result) => console.dir(result, { depth: null }));

Code explanation

First, we need to declare SerpApi from google-search-results-nodejs library and define new search instance with your API key from SerpApi:

const SerpApi = require("google-search-results-nodejs");
const search = new SerpApi.GoogleSearch(API_KEY);

Next, we write the necessary parameters for making a request (if you want to set search category you need to uncomment the books_category parameter):

const params = {
  engine: "google_play", // search engine
  gl: "us", // parameter defines the country to use for the Google search
  hl: "en", // parameter defines the language to use for the Google search
  store: "books", // parameter defines the type of Google Play store
  store_device: "phone", // parameter defines the search device. Options: phone, tablet, tv, chromebook, watch, car
  // if you need to find books from one of the categories you need to uncomment the "books_category" parameter
  // books_category: "audiobooks", // you can see the full list of supported categories on https://serpapi.com/google-play-books-categories
};

The full list of supported categories looks like:

{
  "coll_1665": {
    "section": "Books",
    "html_structure": "rows",
    "category": "Arts & entertainment"
  },
  "subj_Art___Humor.AH_Art": {
    "section": "Books",
    "html_structure": "list",
    "category": "Arts & entertainment",
    "subcategory": "Art"
  },
  "subj_Art___Humor.AH_Drama": {
    "section": "Books",
    "html_structure": "list",
    "category": "Arts & entertainment",
    "subcategory": "Drama"
  },
  "subj_Art___Humor.AH_Humor": {
    "section": "Books",
    "html_structure": "list",
    "category": "Arts & entertainment",
    "subcategory": "Humor"
  },
  "subj_Art___Humor.AH_Music": {
    "section": "Books",
    "html_structure": "list",
    "category": "Arts & entertainment",
    "subcategory": "Music"
  },
  ...and other categories
}

Next, we wrap the search method from the SerpApi library in a promise to further work with the search results:

const getJson = () => {
  return new Promise((resolve) => {
    search.json(params, resolve);
  });
};

And finally, we declare the function getResult that gets data from the page and return it:

const getResults = async () => {
  ...
};

In this function first, we get json with results, then we need to iterate organic_results array in the received json. To do this we use reduce() method (it's allow to make the object with results). On each itaration step we return previous step result (using spread syntax) and add the new category with name from categoryTitle constant:

  const json = await getJson();
  const booksResults = json.organic_results.reduce((result, category) => {
    ...
    return {
      ...result,
      [categoryTitle]: books,
    };
  }, {});
  return booksResults;

Next, we destructure category element, redefine title to categoryTitle constant, and itarate the items array to get all books from this category. To do this we need to destructure the book element, set default value "No rating" for rating and return this constants:

const { title: categoryTitle, items } = category;
const books = items.map((book) => {
  const { title, link, rating = "No rating", original_price, price, thumbnail, product_id } = book;
  const returnedBook = {
    title,
    link,
    rating,
    price,
    thumbnail,
    movieId: product_id,
  };
  if (original_price) returnedBook.originalPrice = original_price;
  return returnedBook;
});

After, we run the getResults function and print all the received information in the console with the console.dir method, which allows you to use an object with the necessary parameters to change default output options:

getResults().then((result) => console.dir(result, { depth: null }));

Output

{
   "New releases":[
      {
         "title":"The Golden Enclaves: A Novel",
         "link":"https://play.google.com/store/books/details/Naomi_Novik_The_Golden_Enclaves?id=7qBSEAAAQBAJ",
         "rating":4.7,
         "price":"$13.99",
         "thumbnail":"https://books.google.com/books/publisher/content/images/frontcover/7qBSEAAAQBAJ?fife=w256-h256",
         "movieId":"7qBSEAAAQBAJ"
      },
        ... and other results
   ],
   "Advice for a better life":[
      {
         "title":"How to Stop Feeling Like Sh*t: 14 Habits that Are Holding You Back from Happiness",
         "link":"https://play.google.com/store/books/details/Andrea_Owen_How_to_Stop_Feeling_Like_Sh_t?id=ekfiDQAAQBAJ",
         "rating":4.4,
         "price":"$11.99",
         "thumbnail":"https://books.google.com/books/publisher/content/images/frontcover/ekfiDQAAQBAJ?fife=w256-h256",
         "movieId":"ekfiDQAAQBAJ"
      },
        ... and other results
   ],
   ... and other categories
}

DIY Code

If you don't need an explanation, have a look at the full code example in the online IDE

const puppeteer = require("puppeteer-extra");
const StealthPlugin = require("puppeteer-extra-plugin-stealth");

puppeteer.use(StealthPlugin());

const searchParams = {
  hl: "en", // Parameter defines the language to use for the Google search
  gl: "us", // parameter defines the country to use for the Google search
  device: "phone", // parameter defines the search device. Options: phone, tablet, tv, chromebook
  category: null, // you can see the full list of supported categories on https://serpapi.com/google-play-books-categories
};

const URL = searchParams.category
  ? `https://play.google.com/store/books/category/${searchParams.category}?hl=${searchParams.hl}&gl=${searchParams.gl}&device=${searchParams.device}`
  : `https://play.google.com/store/books?hl=${searchParams.hl}&gl=${searchParams.gl}&device=${searchParams.device}`;

async function scrollPage(page, scrollContainer) {
  let lastHeight = await page.evaluate(`document.querySelector("${scrollContainer}").scrollHeight`);
  while (true) {
    await page.evaluate(`window.scrollTo(0, document.querySelector("${scrollContainer}").scrollHeight)`);
    await page.waitForTimeout(4000);
    let newHeight = await page.evaluate(`document.querySelector("${scrollContainer}").scrollHeight`);
    if (newHeight === lastHeight) {
      break;
    }
    lastHeight = newHeight;
  }
}

async function getBooksFromPage(page) {
  const books = await page.evaluate(() => {
    const mainPageInfo = Array.from(document.querySelectorAll("section .oVnAB")).reduce((result, block) => {
      const categoryTitle = block.querySelector(".kcen6d").textContent.trim();
      const books = Array.from(block.parentElement.querySelectorAll(".ULeU3b")).map((book) => {
        const link = `https://play.google.com${book.querySelector(".Si6A0c")?.getAttribute("href")}`;
        const bookId = link.slice(link.indexOf("?id=") + 4);
        return {
          title: book.querySelector(".hP61id div:first-child")?.getAttribute("title"),
          link,
          rating: parseFloat(book.querySelector(".LrNMN[aria-label]")?.getAttribute("aria-label").slice(6, 9)) || "No rating",
          originalPrice: book.querySelector(".LrNMN .SUZt4c")?.textContent.trim(),
          price: book.querySelector(".LrNMN .VfPpfd")?.textContent.trim(),
          thumbnail: book.querySelector(".TjRVLb img")?.getAttribute("srcset").slice(0, -3),
          bookId,
        };
      });
      return {
        ...result,
        [categoryTitle]: books,
      };
    }, {});

    return mainPageInfo;
  });
  return books;
}

async function getMainPageInfo() {
  const browser = await puppeteer.launch({
    headless: true, // if you want to see what the browser is doing, you need to change this option to "false"
    args: ["--no-sandbox", "--disable-setuid-sandbox"],
  });

  const page = await browser.newPage();

  await page.setDefaultNavigationTimeout(60000);
  await page.goto(URL);

  await page.waitForSelector(".oVnAB");

  await scrollPage(page, ".T4LgNb");

  const books = await getBooksFromPage(page);

  await browser.close();

  return books;
}

getMainPageInfo().then((result) => console.dir(result, { depth: null }));

Preparation

First, we need to create a Node.js* project and add npm packages puppeteer, puppeteer-extra and puppeteer-extra-plugin-stealth to control Chromium (or Chrome, or Firefox, but now we work only with Chromium which is used by default) over the DevTools Protocol in headless or non-headless mode.

To do this, in the directory with our project, open the command line and enter npm init -y, and then npm i puppeteer puppeteer-extra puppeteer-extra-plugin-stealth.

*If you don't have Node.js installed, you can download it from nodejs.org and follow the installation documentation.

πŸ“ŒNote: also, you can use puppeteer without any extensions, but I strongly recommend using it with puppeteer-extra with puppeteer-extra-plugin-stealth to prevent website detection that you are using headless Chromium or that you are using web driver. You can check it on Chrome headless tests website. The screenshot below shows you a difference.

Process

First of all, we need to scroll through all books listings until there are no more listings loading which is the difficult part described below.

The next step is to extract data from HTML elements after scrolling is finished. The process of getting the right CSS selectors is fairly easy via SelectorGadget Chrome extension which enables us to grab CSS selectors by clicking on the desired element in the browser. However, it is not always working perfectly, especially when the website is heavily used by JavaScript.

We have a dedicated Web Scraping with CSS Selectors blog post at SerpApi if you want to know a little bit more about them.

The Gif below illustrates the approach of selecting different parts of the results using SelectorGadget.

Code explanation

Declare puppeteer to control Chromium browser from puppeteer-extra library and StealthPlugin to prevent website detection that you are using web driver from puppeteer-extra-plugin-stealth library:

const puppeteer = require("puppeteer-extra");
const StealthPlugin = require("puppeteer-extra-plugin-stealth");

Next, we "say" to puppeteer use StealthPlugin, write the necessary request parameters and search URL using ternary operator (the URL may differ depending on whether the category is specified):

const searchParams = {
  hl: "en", // Parameter defines the language to use for the Google search
  gl: "us", // parameter defines the country to use for the Google search
  device: "phone", // parameter defines the search device. Options: phone, tablet, tv, chromebook
  category: null, // you can see the full list of supported categories on https://serpapi.com/google-play-books-categories
};

const URL = searchParams.category
  ? `https://play.google.com/store/books/category/${searchParams.category}?hl=${searchParams.hl}&gl=${searchParams.gl}&device=${searchParams.device}`
  : `https://play.google.com/store/books?hl=${searchParams.hl}&gl=${searchParams.gl}&device=${searchParams.device}`;

If the category parameter is set to null, that means we use the default category (Ebooks) and the URL will look like this:

"https://play.google.com/store/books?hl=en&gl=US";

Otherwise, the URL will look like this:

"https://play.google.com/store/books/category/coll_1665?hl=en&gl=US";

The URL only changes when a category needs to be applied:

The full list of supported categories looks like:

{
  "coll_1665": {
    "section": "Books",
    "html_structure": "rows",
    "category": "Arts & entertainment"
  },
  "subj_Art___Humor.AH_Art": {
    "section": "Books",
    "html_structure": "list",
    "category": "Arts & entertainment",
    "subcategory": "Art"
  },
  "subj_Art___Humor.AH_Drama": {
    "section": "Books",
    "html_structure": "list",
    "category": "Arts & entertainment",
    "subcategory": "Drama"
  },
  "subj_Art___Humor.AH_Humor": {
    "section": "Books",
    "html_structure": "list",
    "category": "Arts & entertainment",
    "subcategory": "Humor"
  },
  "subj_Art___Humor.AH_Music": {
    "section": "Books",
    "html_structure": "list",
    "category": "Arts & entertainment",
    "subcategory": "Music"
  },
  ...and other categories
}

Next, we write a function to scroll the page to load all the articles:

async function scrollPage(page, scrollContainer) {
  ...
}

In this function, first, we need to get scrollContainer height (using evaluate() method). Then we use while loop in which we scroll down scrollContainer, wait 2 seconds (using waitForTimeout method), and get a new scrollContainer height.

Next, we check if newHeight is equal to lastHeight we stop the loop. Otherwise, we define newHeight value to lastHeight variable and repeat again until the page was not scrolled down to the end:

let lastHeight = await page.evaluate(`document.querySelector("${scrollContainer}").scrollHeight`);
while (true) {
  await page.evaluate(`window.scrollTo(0, document.querySelector("${scrollContainer}").scrollHeight)`);
  await page.waitForTimeout(4000);
  let newHeight = await page.evaluate(`document.querySelector("${scrollContainer}").scrollHeight`);
  if (newHeight === lastHeight) {
    break;
  }
  lastHeight = newHeight;
}

Next, we write a function to get books data from the page:

async function getBooksFromPage(page) {
  ...
}

In this function, we get information from the page context and save it in the returned object. Next, we need to get all HTML elements with "section .oVnAB" selector (querySelectorAll() method). Then we use reduce() method (it's allowed to make the object with results) to iterate an array built with Array.from() method:

const books = await page.evaluate(() => {
  const mainPageInfo = Array.from(document.querySelectorAll("section .oVnAB")).reduce((result, block) => {
      ...
    }, {});

    return mainPageInfo;
});
return books;

And finally, we need to get categoryTitle, and title, link, rating, originalPrice, price, thumbnail and bookId(we can cut it fromlink using slice() and indexOf() methods) of each app from the selected category (querySelectorAll(), querySelector(), getAttribute(), textContent and trim() methods.

On each itaration step we return previous step result (using spread syntax) and add the new category with name from categoryTitle constant:

const categoryTitle = block.querySelector(".kcen6d").textContent.trim();
const books = Array.from(block.parentElement.querySelectorAll(".ULeU3b")).map((book) => {
  const link = `https://play.google.com${book.querySelector(".Si6A0c")?.getAttribute("href")}`;
  const bookId = link.slice(link.indexOf("?id=") + 4);
  return {
    title: book.querySelector(".hP61id div:first-child")?.getAttribute("title"),
    link,
    rating: parseFloat(book.querySelector(".LrNMN[aria-label]")?.getAttribute("aria-label").slice(6, 9)) || "No rating",
    originalPrice: book.querySelector(".LrNMN .SUZt4c")?.textContent.trim(),
    price: book.querySelector(".LrNMN .VfPpfd")?.textContent.trim(),
    thumbnail: book.querySelector(".TjRVLb img")?.getAttribute("srcset").slice(0, -3),
    bookId,
  };
});
return {
  ...result,
  [categoryTitle]: books,
};

Next, write a function to control the browser, and get information:

async function getMainPageInfo() {
  ...
}

In this function first we need to define browser using puppeteer.launch({options}) method with current options, such as headless: true and args: ["--no-sandbox", "--disable-setuid-sandbox"].

These options mean that we use headless mode and array with arguments which we use to allow the launch of the browser process in the online IDE. And then we open a new page:

const browser = await puppeteer.launch({
  headless: true, // if you want to see what the browser is doing, you need to change this option to "false"
  args: ["--no-sandbox", "--disable-setuid-sandbox"],
});

const page = await browser.newPage();

Next, we change default (30 sec) time for waiting for selectors to 60000 ms (1 min) for slow internet connection with .setDefaultNavigationTimeout() method, go to URL with .goto() method and use .waitForSelector() method to wait until the selector has loaded:

await page.setDefaultNavigationTimeout(60000);
await page.goto(URL);
await page.waitForSelector(".oVnAB");

And finally, we wait until the page was scrolled, save books data from the page in the books constant, close the browser, and return the received data:

await scrollPage(page, ".T4LgNb");

const books = await getBooksFromPage(page);

await browser.close();

return books;

Now we can launch our parser:

$ node YOUR_FILE_NAME # YOUR_FILE_NAME is the name of your .js file

Output

{
   "Start a new series":[
      {
         "title":"Magic Lessons: The Prequel to Practical Magic",
         "link":"https://play.google.com/store/books/details/Alice_Hoffman_Magic_Lessons?id=sejNDwAAQBAJ",
         "rating":4.7,
         "price":"$12.99",
         "thumbnail":"https://books.google.com/books/publisher/content/images/frontcover/sejNDwAAQBAJ?fife=w512-h512",
         "bookId":"sejNDwAAQBAJ"
      },
      ... and other results
   ],
   "New to rent":[
      {
         "title":"The Forever War",
         "link":"https://play.google.com/store/books/details/Joe_Haldeman_The_Forever_War?id=SUFOBQAAQBAJ",
         "rating":4.5,
         "originalPrice":"$2.99",
         "price":"$2.69",
         "thumbnail":"https://books.google.com/books/publisher/content/images/frontcover/SUFOBQAAQBAJ?fife=w512-h512",
         "bookId":"SUFOBQAAQBAJ"
      },
      ... and other results
   ],
   ... and other categories
}

If you want to see some projects made with SerpApi, write me a message.


Join us on Twitter | YouTube

Add a Feature RequestπŸ’« or a Bug🐞