Puppeteer vs Selenium: A Web Scraper's Experience-Driven Comparison

Jan 9, 2024 · 5 min read

As someone who has spent countless late nights battling challenging sites reluctant to surrender their data to automated extraction scripts, I've formed plenty of hard-earned opinions on the nuanced tradeoffs between Puppeteer and Selenium.

While textbook feature checklists paint a sterile picture, I want to share gritty truths learned scraping production systems relying on both WebDriver powered frameworks over the years.

In this post, we'll shun theoretical analysis in favor of street-tested anecdotes highlighting where each tool respectively flounders or flourishes when scraping those finicky sites denying data access politely.

Let's dive in to where precisely Puppeteer and Selenium differ through a developer's lens with plenty of battle scars!

Task Suitability: Testing vs. Data Retrieval

First, let's clarify the origins of both tools, as those genesis use cases significantly impact their applicability for assorted tasks:

Selenium arose as an open source web application test automation framework allowing QA teams to programmatically validate functionality and assertions across real browsers like Chrome, Firefox and Safari.

Puppeteer conversely exists solely to provide a high-level Node.js API for controlling headless Chrome and Chromium enabling scraping and screenshot generation.

So Selenium targets web app testing but Puppeteer focuses squarely on web data extraction and harvesting.

Where This Caused Me Grief

Early on, I would routinely attempt to utilize Puppeteer just like Selenium to drive test automation scripts with occasionally painful outcomes.

While Puppeteer can technically trigger application flows and simulate users, the lack of built-in synchronization primitives led to hopeless races between UI state stalled mid-update versus my script barreling onwards errantly assuming pages had fully loaded.

After days wasted forcing Puppeteer into acting like a blunt Selenium stand-in for testing reactive single page apps, I finally accepted its data harvesting strengths and pivoted to a dual framework approach:

  • Selenium for qualified test case automation
  • Puppeteer for data scraping and static site harvesting
  • Lesson learned — emphasize the core competencies of both tools, don't force square pegs into round holes!

    Example Scraping Task: Retail Inventory Auditing

    To better illustrate subtle differences in utilizing Puppeteer and Selenium for scraping data, let me walk through a representative use case:

    Automatically audit inventory counts changes daily across assigned retail products to identify possible database sync issues when figures diverge significantly without explanations like upcoming sales.

    This requires:

    1. Login via form credentials
    2. Navigate to inventory dashboard
    3. Extract current product counts
    4. Compare vs. historical baseline

    For the purposes of this post, I'll focus specifically on steps 1-3 with brief code snippets highlighting nuanced implementations between both tools:

    Step 1 - Login Form Submission


    const usernameSelector = '#username';
    await page.type(usernameSelector, 'puppeteer_maestro');
    const passwordSelector = '#password';
    await page.type(passwordSelector, 'test_password');
    await Promise.all([


    WebElement username = driver.findElement(By.id("username"));
    WebElement password = driver.findElement(By.id("password"));
    WebDriverWait wait = new WebDriverWait(driver, 10);


  • Puppeteer leverages simple page.$ shorthand for element lookup forcing explicit waits with page.waitForSelector. Mixing implicit and explicit waits risks stale element errors
  • Selenium bakes in configurable implicit waits allowing decoupled actions without redundant expected condition checks
  • Winner: Selenium for flexibility abstracting away waits

    Step 2 - Navigate to Dashboard


    // Wait explicitly for inventory link selector
    await page.waitForSelector('.inventory-link');
    // Click inventory link when available
    await page.click('.inventory-link');
    // Locate product row container explicitly again
    const products = await page.$('#product-rows');


    // Click directly with configurable implicit waits
    // Timeout exceptions provide implicit waiting built-in
    List<WebElement> rows = driver.findElements(By.id("product-rows"));


  • Puppeteer forces coding discipline awaiting selectors available before interacting. Brittle but explicit
  • Selenium leaning on timeouts liberates actions at cost of intermittent issues
  • Winner: Toss Up based on preferences

    Step 3 - Extract Product Inventory Counts


    // Retrieve row cells using convenient page.$$eval shorthand
    const counts = await page.$$eval('#inventory tr td:nth-child(3)', cells => {
      // Map and extract needed data
      return cells.map(cell => parseInt(cell.innerText));


    // Fallback to Java 8 streams approach
    List<WebElement> cells = driver.findElements(By.cssSelector("#inventory tr td:nth-child(3)"));
    List<Integer> counts = cells.stream().map(e -> Integer.parseInt(e.getText())).collect(Collectors.toList());


  • Puppeteer eliminates DOM traversal with concise $$eval and lambda parsing
  • Selenium succinct for data flows with Streams minus element heartache
  • Winner: Situational depending on DOM complexity

    Key Takeways from Getting Burned

    Through ample time stuck executing inventive maneuvers to wrestle data from rigid sites across a spectrum of use cases with both Puppeteer and Selenium, a few principles emerged:

  • Prefer Selenium for test automation flows, Puppeteer for data scraping
  • Leverage Selenium's configurable implicit waits to remove timing fragility
  • Master Puppeteer's element lookup and data extraction shorthand
  • Integrate both tools for optimal test coverage and harvesting
  • Beware default timeouts and polling race conditions causing brittleness
  • I hope by relaying a sampling of painful lessons etched through lost hours and inverted eyelids, your own journey taming web apps for test and harvesting proves smoother!

    No developers should endure the involuntary JavaScript puzzle abuse I overcame getting either framework to submit to my data extraction will.

    Let my suffering spare you similar misery!

    Browse by tags:

    Browse by language:

    The easiest way to do Web Scraping

    Get HTML from any page with a simple API call. We handle proxy rotation, browser identities, automatic retries, CAPTCHAs, JavaScript rendering, etc automatically for you

    Try ProxiesAPI for free

    curl "http://api.proxiesapi.com/?key=API_KEY&url=https://example.com"

    <!doctype html>
        <title>Example Domain</title>
        <meta charset="utf-8" />
        <meta http-equiv="Content-type" content="text/html; charset=utf-8" />
        <meta name="viewport" content="width=device-width, initial-scale=1" />


    Don't leave just yet!

    Enter your email below to claim your free API key: