What is the best time to scrape data from Trustpilot?

When deciding the best time to scrape data from Trustpilot, or any other website, there are several factors to consider:

  1. Website Traffic: Higher website traffic can sometimes mean slower response times, which might affect the efficiency of your scraping operation. It might be beneficial to scrape during off-peak hours when the website traffic is lower.

  2. Terms of Service: Always review the Terms of Service (ToS) of Trustpilot before scraping. The ToS will specify what is allowed and what isn't. Violating the ToS can lead to legal issues or being blocked from the site.

  3. Rate Limiting: Trustpilot may have rate limiting in place to prevent excessive requests to their servers. This means that if you send too many requests in a short period, your IP address could be temporarily banned.

  4. Server Maintenance: Sometimes, websites have scheduled maintenance windows. It's wise to avoid scraping during these times as you might encounter downtime or errors.

  5. Data Update Frequency: If you're scraping Trustpilot for the most current reviews or ratings, you'll want to time your scrapes for when this information is updated on the website.

There isn't a universally "best" time to scrape data since it can vary based on the factors above. However, here are some general guidelines:

  • Late Night or Early Morning: Websites generally experience lower traffic during these times, which could be advantageous for scraping.
  • After Data Updates: If Trustpilot updates its data at a specific time, it's best to scrape afterward to ensure you are getting the most current information.
  • Avoid Peak Hours: Try to avoid scraping during business hours or times when you know the website will be busiest.
  • Compliance: Make sure your scraping is compliant with Trustpilot’s ToS and respects their robots.txt file, which may specify the scrapable paths and crawl-delay.

Technical Considerations for Scraping

When you are ready to scrape, make sure you follow these technical best practices:

  • Rate Limiting: Implement a delay between your requests to avoid hitting rate limits or being perceived as a Denial-of-Service (DoS) attack.
  • Headers: Include headers in your requests that mimic a real browser, such as User-Agent, to avoid being blocked.
  • Proxy Rotation: Use a rotation of different IP addresses to avoid getting your IP banned.
  • Respect robots.txt: Check Trustpilot’s robots.txt file for rules about scraping.

Example of a Python Scraper with Time Considerations

import requests
import time
from bs4 import BeautifulSoup

# Function to scrape Trustpilot
def scrape_trustpilot(url):
    headers = {
        'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/58.0.3029.110 Safari/537.3'}
    response = requests.get(url, headers=headers)
    # Check if we got a successful response from the server
    if response.status_code == 200:
        soup = BeautifulSoup(response.text, 'html.parser')
        # Process the page content with BeautifulSoup or any other parsing method
        # ...
        print(f"Failed to retrieve data: {response.status_code}")

# Main scraping logic with delay to avoid rate limiting
def main():
    urls = ['https://www.trustpilot.com/review/example1.com',
    for url in urls:
        time.sleep(10)  # Sleep for 10 seconds between requests to avoid rate limiting

if __name__ == "__main__":

JavaScript (Node.js) Example with Time Considerations

const axios = require('axios');
const cheerio = require('cheerio');

// Function to scrape Trustpilot
async function scrapeTrustpilot(url) {
    try {
        const response = await axios.get(url, {
            headers: {
                'User-Agent': 'Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/58.0.3029.110 Safari/537.3'
        const $ = cheerio.load(response.data);
        // Process the page content with Cheerio or any other parsing method
        // ...
    } catch (error) {
        console.error(`Failed to retrieve data: ${error}`);

// Main scraping logic with delay to avoid rate limiting
async function main() {
    const urls = ['https://www.trustpilot.com/review/example1.com',
    for (const url of urls) {
        await scrapeTrustpilot(url);
        await new Promise(resolve => setTimeout(resolve, 10000)); // Sleep for 10 seconds between requests


Always remember to use web scraping responsibly and ethically, respecting the terms and limitations imposed by the target website.

Related Questions

Get Started Now

WebScraping.AI provides rotating proxies, Chromium rendering and built-in HTML parser for web scraping