Skip to main content
Transform your testing process with: Real Device Features, Company-wide Licences, & Test Observability
No Result Found

Web performance testing using Lighthouse

Learn to measure and enhance the user experience of your web pages by seamlessly integrating Lighthouse into your Puppeteer tests.

Google Lighthouse runs a series of audits on your web pages and provides the results across following categories: performance, accessibility, best practices, SEO, and progressive web app (PWA). The generated report helps you measure the Web Vitals and provides insights on optimizing the quality of user experience of your web pages.

  • To get started with web performance testing using BrowserStack Automate, refer to our getting started page.
  • This feature is exclusive to the Automate Desktop and Mobile Pro plan. Visit our pricing page for more details.

You have the following two ways of integrating Lighthouse with Puppeteer tests:

To enable js executor, set the following capability in your configuration file:

“performance: assert” 

Assert Lighthouse configuration

You can assert on the Lighthouse categories & metrics using JavascriptExecutor. Post this, you can generate a Lighthouse report.

Let us look at some scenarios on asserting the Lighthouse configuration.

  • The Lighthouse integration supports generating reports for up to the first 10 web pages (URLs) opened during a test session.
  • Currently, Lighthouse report generation is supported on browsers Chrome and Edge 90 and above.
  • Supported operating system versions include Windows 10 and later and macOS Mojave and later.


Generate Lighthouse report of URL

To generate a Lighthouse report of a URL in the current test context through JavascriptExecutor, set the action parameter to lighthouseAudit. To pass a custom URL for the Lighthouse audit, mention the URL in the arguments object. Use the sample JavascriptExecutor request below:

In the below example, the current URL in the browser context is picked up.

await page.evaluate(_ => {}, `browserstack_executor: ${JSON.stringify({
    "action": "lighthouseAudit"
})}`);

In the below example, the specified URL is picked up:

await page.evaluate(_ => {}, `browserstack_executor: ${JSON.stringify({
    "action": "lighthouseAudit", 
    "arguments":{
      "url": "https://www.bstackdemo.com"
    }
})}`);

Assert Lighthouse results

To assert the Lighthouse audit scores against the specified value, mention your assertion parameters in the assertResult object. In the below example, if the Lighthouse performance score is below 40, the lighthouse assertion fails. Similarly, for the accessibility parameter, 50 is taken as the default threshold. In metrics, in the first-contentful-paint, the score (metricUnit) is more than 50. Use the sample JavascriptExecutor request below:

await page.evaluate(_ => {}, `browserstack_executor: ${JSON.stringify({
    "action": "lighthouseAudit",
    "arguments":{
      "url": "https://bstackdemo.com/",
      "assertResult": { 
        "categories": {
          "performance": 40,
          "best-practices": 50
        },
        "metrics": {
          "first-contentful-paint": {
              "moreThan": 50,
              "metricUnit": "score"             
          },
          "largest-contentful-paint": {
              "lessThan": 4000,
              "metricUnit": "numeric"              
          },
          "total-blocking-time": {
              "lessThan": 600,
              "metricUnit": "numeric"              
          },
          "cumulative-layout-shift": {
              "moreThan": 50,
              "metricUnit": "score"              
          } 
        } 
      }
    }
})}`);

Mark test session status as `failed`

Use the code below to mark test sessions as failed if any of the assertions in the assertResult object across any URLs fail:

  await page.evaluate(_ => {}, `browserstack_executor: ${JSON.stringify({
    "action": "lighthouseAudit",
    "arguments":{
      "url": "https://bstackdemo.com/",
      "assertResult": { 
        "sessionFail": true,
        "categories": {
          "performance": 40
        },
        "metrics": {
          "largest-contentful-paint": {
              "lessThan": 4000,
              "metricUnit": "numeric"              
          }
        } 
      }
    }
})}`);

Access lighthouse JSON reports in the test script

await page.evaluate(_ => {}, `browserstack_executor: ${JSON.stringify({
  "action": "lighthouseAudit", 
  "arguments":{
    "url": "https://www.bstackdemo.com",
    "executorOutput": "json"   
    }
})}`);

If any of the above parameters are not passed, the default lighthouse configuration will be effective. No validation is to be applied to your inputs.

The request returns the below assertion result in the response:


  "url": "https://www.browserstack.com",
  "assertResult": {
    "categories": {
        "performance": "passed",
        "best-practices": "failed"
    }
    "metrics": {
        "first-contentful-paint": "passed",
        "largest-contentful-paint": "failed",
        "total-blocking-time": "passed",
        "cumulative-layout-shift": "failed"
    }

Description of the parameters in the JavascriptExecutor request

Object Mandatory/Optional Description
categories Optional Allows you to add assertion on lighthouse categories’ score, available in the categories object in the result JSON - Lighthouse.
Valid value(s) for the categories: performance, accessibility, best-practices, seo, pwa. Valid values for the categories’ score: null, integer from 0 to 100, 50 (default).
metrics Optional Allows you to add assertions on lighthouse metrics result present in the audits objects in the result JSON - Lighthouse. Valid value(s) for the metrics are the parameters in the audits object in the lighthouse result.
metrics Child:
a. Either moreThan or lessThan - allows you to provide the threshold value for assertion. Both parameters are of float type.
b. metricUnit - allows user to specify the threshold value type.
Valid values are: score, numeric. score compares with the score parameter and numeric compares with numericValue in the audits child object in the Lighthouse result JSON.

Set Lighthouse-specific configurations

You can emulate the device for lighthouse audit by passing various configurations that are specific to Lighthouse.

To configure Lighthouse, pass your configurations in the lhConfig object.

You can assert on the Lighthouse settings along with setting the configurations in the lhconfig object.

Here are some sample scenarios for using the Lighthouse-specific configurations:

Emulate the device for Lighthouse audit

The lhConfig object allows you to pass Lighthouse-specific configurations. By default the view is mobile view.

Lighthouse runs on Slow 4G network and MOTOG4 mobile screens.

Use the code below to emulate the desktop devices for the Lighthouse audit:

await page.evaluate(_ => {}, `browserstack_executor: ${JSON.stringify({
    "action": "lighthouseAudit",
    "arguments":{
      "url": "https://bstackdemo.com/",
      "lhConfig": {
        "extends": "lighthouse:default",
        "settings": {
          "formFactor": "desktop",
          "screenEmulation": {
            "mobile": false,
            "width": 1350,
            "height": 940,
            "deviceScaleFactor": 1,
            "disabled": false
          },
        },
      },
      "assertResult": { 
        "categories": {
          "performance": 40
        },
        "metrics": {
          "largest-contentful-paint": {
              "lessThan": 4000,
              "metricUnit": "numeric"              
          }
        } 
      }
    }
})}`);

Use the code below to emulate the mobile devices for the Lighthouse audit:

await page.evaluate(_ => {}, `browserstack_executor: ${JSON.stringify({
    "action": "lighthouseAudit",
    "arguments":{
      "url": "https://bstackdemo.com/",
      "lhConfig": {
        "extends": "lighthouse:default",
        "settings": {
          "formFactor": "mobile",
          "screenEmulation":{
            "mobile": true,
            "width": 412,
            "height": 843,
            "deviceScaleFactor": 1.75,
            "disabled": false
          }
        }
      },
      "assertResult": { 
        "categories": {
          "performance": 40
        },
        "metrics": {
          "largest-contentful-paint": {
              "lessThan": 4000,
              "metricUnit": "numeric"              
          }
        } 
      }
    }
})}`);

Simulate the network connection for Lighthouse audit

Use the code below to simulate the network connection for predefined settings, using the mobileRegular3G, desktopDense4G, and mobileSlow4G settings.

await page.evaluate(_ => {}, `browserstack_executor: ${JSON.stringify({
        "action": "lighthouseAudit",
        "arguments":{
          "url": "https://bstackdemo.com/",
          "throttling": "mobileRegular3G",
          "assertResult": { 
            "categories": {
              "performance": 40
            },
            "metrics": {
              "largest-contentful-paint": {
                  "lessThan": 4000,
                  "metricUnit": "numeric"              
              }
            } 
          }
        }
})}`);

Use the code below to simulate the network connection by adding customised settings for the Lighthouse audit or disable the default network simulation by Lighthouse:

await page.evaluate(_ => {}, `browserstack_executor: ${JSON.stringify({
    "action": "lighthouseAudit",
    "arguments":{
      "url": "https://bstackdemo.com/",
      "lhConfig": {
        "extends": "lighthouse:default",
        "settings": {    
          "throttling": {         
            "rttMs": 150,
            "throughputKbps": 1638.4,
            "cpuSlowdownMultiplier": 4,
            "requestLatencyMs": 562.5,
            "downloadThroughputKbps": 1474.56,
            "uploadThroughputKbps": 675
          }
        }
      },
      "assertResult": { 
        "categories": {
          "performance": 40
        },
        "metrics": {
          "largest-contentful-paint": {
              "lessThan": 4000,
              "metricUnit": "numeric"              
          }
        } 
      }
    }
})}`);

Click here for the description of the Lighthouse parameters.

View reports on the Automate dashboard

On your Automate dashboard, select the Web Performance tab to view the report generated by Lighthouse.

You can view the assertion results and the config parameters passed under the Text Logs tab. You can view the report by clicking the View Report link.

Text Logs

Lighthouse Performance results

To view all the reports of the web performance in one place, select the Web Performance tab.

Lighthouse report

Generate Lighthouse reports

To generate a lighthouse report for the open URLs in a session, set the following capabilities in your test script:

  • For a Lighthouse report to generate, a test session must be at least 30-60 seconds long. If your test is likely to take less time, consider introducing the required amount of artificial delay.
  • To successfully generate lighthouse report for more than one URL, ensure 30 to 60 seconds processing time after each opened URL.
browserstack.performance: report;
  • The Lighthouse integration supports generating reports for up to the first 10 web pages (URLs) opened during a test session.
  • Currently, Lighthouse report generation is supported on browsers Chrome 85 and above and Edge 85 and above. However, it’s important to note that the performance score is unavailable for Chrome and Edge versions ranging from 85 to 90.
  • Supported operating system versions include Windows 10 and later and macOS Mojave and later.

Important points to remember:

  • The setting is false by default. You can see the Web Performance tab on the Automate dashboard only when you set the capability to true.
  • If a pop-up appears on the web page of your website under test when Lighthouse launches, the report does not get generated.
  • If you incorporate two consecutive open URLs in your code, the system generates a report solely for the first open URL session. To acquire a report for subsequent open URLs, ensure you utilize the wait command for 30 seconds after each open URL.
  • The web performance feature is currently available for desktop browsers only.

View reports on the Automate dashboard

On your Automate dashboard, select the Web Performance tab to view the report generated by Lighthouse.

Lighthouse report

We're sorry to hear that. Please share your feedback so we can do better

Contact our Support team for immediate help while we work on improving our docs.

We're continuously improving our docs. We'd love to know what you liked





Thank you for your valuable feedback

Is this page helping you?

Yes
No

We're sorry to hear that. Please share your feedback so we can do better

Contact our Support team for immediate help while we work on improving our docs.

We're continuously improving our docs. We'd love to know what you liked





Thank you for your valuable feedback!

Talk to an Expert
Download Copy Check Circle