Skip to main content

How to grab the page source from any dynamically generated webpage and then process it

· 3 min read
Andrew Golightly

I had an issue where I needed to scrape some data from a dynamically generated webpage. Open this page in a browser and you’ll see what I mean.

I tried a lot of options, and the solution that ended up working for me was to use Headless Chrome. This basically allows you to launch Chrome and use it as a tool from within your code. It’s recommended to install Chrome Canary and you can get it from here.

Essentially the pseudo code for my script goes as follows..

  1. launch Chrome in headless mode (no visible window)
  2. load a page and wait until the page is loaded
  3. then wait a little longer until all the JS on the page has completed
  4. then click on something using a querySelector to change the language to English
  5. then grab all the source code from the page
  6. then load that source code into cheerio and perform queries as needed

The full source code for this is below. You’ll of course need to install the 3 packages too.

> yarn add chrome-remote-interface chrome-launcher cheerio

const CDP = require('chrome-remote-interface');
const chromeLauncher = require('chrome-launcher');
const cheerio = require('cheerio');

(async function() {
const launchChrome = () =>
chromeLauncher.launch({ chromeFlags: ['--disable-gpu', '--headless'] });

const chrome = await launchChrome();
const protocol = await CDP({ port: chrome.port });

const timeout = ms => new Promise(resolve => setTimeout(resolve, ms));

// See API docs:
const { Page, Runtime, DOM } = protocol;
await Promise.all([Page.enable(), Runtime.enable(), DOM.enable()]);

Page.navigate({ url: '' });

// wait until the page says it's loaded...
Page.loadEventFired(async () => {
try {
console.log('Page loaded! Now waiting a few seconds for all the JS to load...');
await timeout(3000); // give the JS some time to load
console.log('Selecting English..');

// first set the language to English
const result = await Runtime.evaluate({
"document.querySelector('.lang-switcher li:nth-of-type(2) a').click()"

// get the page source
const rootNode = await DOM.getDocument({ depth: -1 });
const pageSource = await DOM.getOuterHTML({
nodeId: rootNode.root.nodeId

// load the page source into cheerio
console.log('Processing page source...');
const $ = cheerio.load(pageSource.outerHTML);

// perform queries
console.log('Getting movie times for', $('.showtime-cinema-name').text());
$('.showtime-box').each((i, movieElement) => {
console.log($(movieElement).find('.movie-detail .name').text());
} catch (err) {

View this gist on GitHub

I hope that helps some people! It took me ages to finally get to this place where I could scrape a dynamically generated webpage from Node.js.