How do I handle redirects while scraping with JavaScript?

When web scraping with JavaScript, handling redirects is an important aspect of dealing with dynamic webpage behavior. Redirects happen when a server responds with a status code that instructs the client (in this case, your scraping script) to fetch a different URL. Common redirect status codes are 301 (Moved Permanently), 302 (Found or Temporary Redirect), and 307 (Temporary Redirect).

If you are using Node.js for web scraping, you can handle redirects using the popular axios or request libraries. Below are examples of how to handle redirects in JavaScript using Node.js with these libraries.

Using Axios

Axios follows redirects by default. However, if you want to capture the redirect URLs or limit the number of redirects, you can do so by configuring Axios:

const axios = require('axios');

axios.get('http://example.com/page', {
  maxRedirects: 5, // Set the maximum number of redirects to follow
  validateStatus: function (status) {
    return status >= 200 && status < 300; // default
  }
})
.then(response => {
  console.log('Final URL:', response.request.res.responseUrl); // Final URL after redirects
  console.log(response.data);
})
.catch(error => {
  console.error('Error:', error);
});

If you want to disable automatic following of redirects, you can intercept the response and handle it manually:

const axios = require('axios');

axios.get('http://example.com/page', {
  maxRedirects: 0, // Disable following redirects automatically
})
.then(response => {
  // This block will not be executed for redirect responses
})
.catch(error => {
  if (error.response && error.response.status >= 300 && error.response.status < 400) {
    console.log('Redirect to:', error.response.headers.location);
    // Here you can handle the redirect manually by making a new request to the location header
  } else {
    console.error('Error:', error);
  }
});

Using Request (Deprecated)

The request library has been deprecated, but if you are maintaining legacy code that uses it, here's how to handle redirects:

const request = require('request');

request({
  url: 'http://example.com/page',
  followRedirect: true, // This is true by default
  maxRedirects: 10 // Maximum number of redirects to follow
}, (error, response, body) => {
  if (!error && response.statusCode == 200) {
    console.log('Final URL:', response.request.href); // Final URL after redirects
    console.log(body); // Print the HTML of the web page.
  } else {
    console.error('Error:', error);
  }
});

To disable following redirects and handle them manually:

const request = require('request');

request({
  url: 'http://example.com/page',
  followRedirect: false // Do not automatically follow redirects
}, (error, response, body) => {
  if (response && response.headers.location) {
    console.log('Redirect to:', response.headers.location);
    // Handle the redirect manually here
  } else {
    console.error('Error:', error);
  }
});

Using Fetch API (Client-Side JavaScript)

In a browser environment, the Fetch API also follows redirects automatically. If you want to handle redirects manually, you can check the redirected property of the response:

fetch('http://example.com/page')
  .then(response => {
    if (response.redirected) {
      console.log('Redirected to:', response.url);
    }
    return response.text();
  })
  .then(html => {
    // Process the HTML here
  })
  .catch(error => {
    console.error('Error:', error);
  });

To disable following redirects, you can set the redirect option to 'manual':

fetch('http://example.com/page', {
  redirect: 'manual' // Do not follow redirects automatically
})
.then(response => {
  if (response.type === 'opaqueredirect') {
    console.log('Response was a redirect');
    // The actual redirect location is not exposed to the script
  }
})
.catch(error => {
  console.error('Error:', error);
});

In the case of the fetch API, note that due to security reasons (CORS policy), the redirect location may not always be available to the script. It depends on the headers sent by the server.

Always remember to handle redirects with respect to the website's terms of service and privacy policy. Additionally, consider the legal and ethical implications when writing web scraping scripts.

Get Started Now

WebScraping.AI provides rotating proxies, Chromium rendering and built-in HTML parser for web scraping
Icon