Rottentomatoes.com is a popular website for movie ratings and reviews. The platform offers valuable information on movies, TV shows and even audience opinions. Rotten Tomatoes has data for movie lovers, researchers, and developers.

This blog will show you how to scrape Rotten Tomatoes to get movie ratings using Python. Since Rotten Tomatoes uses JavaScript rendering, we will use the Crawlbase Crawling API to handle the dynamic content loading. By the end of this blog, you will know how to extract key movie data like ratings, release dates, and reviews and store it in a structured format like JSON.

Here’s a detailed tutorial on how to scrape Rotten Tomatoes:

Table of Contents

  1. Why Scrape Rotten Tomatoes for Movie Ratings?
  2. Key Data Points to Extract from Rotten Tomatoes
  3. Crawlbase Crawling API for Rotten Tomatoes Scraping
  4. Setting Up Your Python Environment
  • Installing Python and Required Libraries
  • Setting Up a Virtual Environment
  • Choosing an IDE
  1. Scraping Rotten Tomatoes Movie Listings
  • Inspecting the HTML Structure
  • Writing the Rotten Tomatoes Movie Listings Scraper
  • Handling Pagination
  • Storing Data in a JSON File
  • Complete Code Example
  1. Scraping Rotten Tomatoes Movie Details
  • Inspecting the HTML Structure
  • Writing the Rotten Tomatoes Movie Details Scraper
  • Storing Data in a JSON File
  • Complete Code Example
  1. Final Thoughts
  2. Frequently Asked Questions

Why Scrape Rotten Tomatoes for Movie Ratings?

Rotten Tomatoes is a reliable source for movie ratings and reviews so it’s a good website to scrape for movie data. Whether you’re a movie lover, a data analyst or a developer, you can scrape Rotten Tomatoes to gain insights into movie trends, audience preferences and critic ratings.

Here are a few reasons why you should scrape Rotten Tomatoes for movie ratings.

  1. Access to Critic Reviews: Rotten Tomatoes provides critic reviews, allowing you to see how individual professionals perceive the movie.
  2. Audience Scores: Get audience ratings which shows how the general public feels about a movie.
  3. Film Details: Rotten Tomatoes offers data such as titles, genres, release dates, and more.
  4. Popularity Tracking: By scraping ratings over time, you can monitor trends in genres, directors, or actors’ popularity.
  5. Create a Personal Movie Database: Gather ratings and reviews to build a custom database for research, recommendations, or personal projects.

Key Data Points to Extract from Rotten Tomatoes

When scraping movie ratings from Rotten Tomatoes, you should focus on obtaining the most important data points. These data points shall provide you with essential info about movie’s reception, popularity and performance. Here are the data points you should extract:

Rotten Tomatoes Key Data Points
  1. Movie Title: The title of the movie is the first piece of information that they have. This will assist you in categorizing your scraped data according to certain particular films.
  2. Critic Score (Tomatometer): Rotten Tomatoes has a critics score which is referred to as the ‘Tomatometer’. This score is computed with reviews from approved critics and provides an overall picture of a movie’s critical consensus.
  3. Audience Score: The audience score reflects how the general public prefers the particular movie. It is therefore useful when looking at the comparisons between the professional and public opinions on various movies.
  4. Number of Reviews: Critic as well as audience ratings are calculated according to the number of reviews available. To establish the reliability of the scores, you will need to extract this data as well.One of the ways to use this knowledge is to compare movies of different years and analyze changes over time knowing the release date of a given movie.
  5. Release Date: Release Date can be used to compare movies of different years and analyze changes over time.
  6. Genre: The movies in the database of Rotten Tomatoes are classified according to genre like drama, action, or comedy. Hence, genres are relevant to categorize the movie ratings according to the viewers’ interests.
  7. Movie Synopsis: Extracting the brief description/synopsis of the movie will help you to get a background information about the movie’s storyline and themes.
  8. Cast and Crew: Rotten Tomatoes has the cast and the crew list of the movies. This data is useful to track movies of a particular director, actor or writer.

Crawlbase Crawling API for Rotten Tomatoes Scraping

We will use the Crawlbase Crawling API to get movie ratings and other data from Rotten Tomatoes. Scrape Rotten Tomatoes with simple approaches is hard because the website loads its content dynamically using JavaScript. Crawlbase Crawling API is designed to handle dynamic, JavaScript-heavy websites, so it’s the best way to scrape Rotten Tomatoes fast and easy.

Why Scrape Rotten Tomatoes with Crawlbase?

JavaScript is used to load dynamic content from Rotten Tomatoes pages, including audience scores, reviews, and ratings. Such websites are tough for standard web scraping libraries like requests to handle since they cannot manage JavaScript rendering. The Crawlbase Crawling API solves this issue by making sure you get fully-loaded HTML, complete with all necessary data, through server-side JavaScript rendering.

Here’s why Crawlbase is a solid choice for scraping Rotten Tomatoes:

  1. JavaScript Rendering: It automatically deals with pages that depend on JavaScript to load content, like ratings or reviews.
  2. Built-in Proxies: Crawlbase includes rotating proxies to avoid IP blocks and captchas, keeping your scraping smooth.
  3. Customizable Parameters: You can adjust API parameters like ajax_wait and page_wait to make sure every piece of content is fully loaded before you start scraping.
  4. Reliable and Fast: Designed for efficiency, Crawlbase lets you scrape large datasets from Rotten Tomatoes quickly, with minimal interruptions.

Crawlbase Python Library

Crawlbase provide its own Python library for simplicity. To use it, you’ll need an access token from Crawlbase, which you can obtain by registering an account.

Here’s a sample function to send requests with the Crawlbase Crawling API:

1
2
3
4
5
6
7
8
9
10
11
12
13
from crawlbase import CrawlingAPI

# Initialize Crawlbase API with your token
crawling_api = CrawlingAPI({'token': 'YOUR_CRAWLBASE_TOKEN'})

def make_crawlbase_request(url):
response = crawling_api.get(url)

if response['headers']['pc_status'] == '200':
return response['body'].decode('utf-8')
else:
print(f"Failed to fetch the page. Crawlbase status code: {response['headers']['pc_status']}")
return None

Note: For static sites, use the Normal Token. For dynamic sites like Rotten Tomatoes, use the JS Token. Crawlbase offers 1,000 free requests to get started, with no credit card needed. For more details, refer to the Crawlbase Crawling API documentation.

In the next sections we will cover how to set up your Python environment and the code to scrape movie data using Crawlbase Crawling API.

Setting Up Your Python Environment

To scrape data from Rotten Tomatoes, the first thing you’ll need to do is set up your Python environment. This includes installing Python, creating a virtual environment, and ensuring all the libraries you need are in place.

Installing Python

The first step is making sure Python is installed on your system. Head over to the official Python website to download the latest version. Remember to choose the version that matches your operating system (Windows, macOS or Linux).

Setting Up a Virtual Environment

Using a virtual environment is a smart way to manage your project dependencies. It helps you keep things clean and prevents conflicts with other projects. Here’s how you can do it:

  1. Open your terminal or command prompt.

  2. Go to your project folder.

  3. Run the following command to create a virtual environment:

    1
    python -m venv myenv
  4. Activate the virtual environment:

  • On Windows:

    1
    myenv\Scripts\activate
  • On macOS/Linux:

    1
    source myenv/bin/activate

Extracting Data Using BeautifulSoup and Crawlbase

Next you’ll need to install the required libraries, including Crawlbase and BeautifulSoup for data handling. Run the following command in your terminal:

1
pip install crawlbase beautifulsoup4
  • Crawlbase: Used to interact with the Crawlbase products including Crawling API.
  • BeautifulSoup: For parsing HTML and extracting the required data.

Choosing an IDE

To write and run your Python scripts smoothly, use a good IDE (Integrated Development Environment). Here are a few:

Now that you have your environment set up, let’s start scraping Rotten Tomatoes for movie ratings with Python. In the next section we’ll get into the code that will extract movie ratings and other info from Rotten Tomatoes.

Scraping Rotten Tomatoes Movie Listings

Here we will scrape movie listings from Rotten Tomatoes. We will look at the HTML, write the scraper, handle pagination and organize the data. We will use the Crawlbase Crawling API for JavaScript and dynamic content.

Inspecting the HTML Structure

Before we write the scraper we need to inspect the Rotten Tomatoes page to see what the structure looks like. This will help us know what to target.

  1. Open the Rotten Tomatoes page: Go to the page you want to scrape. For example we are scraping Top Box Office movies List.
  2. Open Developer Tools: Right-click on the page and select “Inspect” or press Ctrl+Shift+I (Windows) or Cmd+Option+I (Mac).
Rotten Tomatoes Listing Page
  1. Identify the Movie Container: Movies on Rotten Tomatoes are usually inside

    elements with class flex-container which are inside a parent
    with attribute data-qa=”discovery-media-list”.

  2. Locate Key Data:

    • Title: Usually inside a <span> with attribute like data-qa="discovery-media-list-item-title". This is the movie title.
    • Critics Score: Inside an rt-text element with slot="criticsScore". This is the critics score for the movie.
    • Audience Score: Also inside an rt-text element but with slot="audienceScore". This is the audience score for the movie.
    • Link: The movie link is usually inside an <a> tag with data-qa attribute that starts with discovery-media-list-item. You can extract the href attribute from this element to get the link to the movie’s page.

Writing the Rotten Tomatoes Movie Listings Scraper

Now we know the structure, let’s write the scraper. We will use Crawlbase Crawling API to fetch the HTML and BeautifulSoup to parse the page and extract titles, ratings and links.

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
from crawlbase import CrawlingAPI
from bs4 import BeautifulSoup

# Initialize Crawlbase API with your token
crawling_api = CrawlingAPI({'token': 'CRAWLBASE_JS_TOKEN'})

def fetch_html(url):
options = {
'ajax_wait': 'true',
'page_wait': '5000'
}
response = crawling_api.get(url, options)
if response['headers']['pc_status'] == '200':
return response['body'].decode('utf-8')
else:
print(f"Failed to fetch data. Status code: {response['headers']['pc_status']}")
return None

def parse_movies(html):
soup = BeautifulSoup(html, 'html.parser')
movies = soup.select('div[data-qa="discovery-media-list"] > div.flex-container')

movie_data = []
for movie in movies:
title = movie.select_one('span[data-qa="discovery-media-list-item-title"]').text.strip() if movie.select_one('span[data-qa="discovery-media-list-item-title"]') else ''
criticsScore = movie.select_one('rt-text[slot="criticsScore"]').text.strip() if movie.select_one('rt-text[slot="criticsScore"]') else ''
audienceScore = movie.select_one('rt-text[slot="audienceScore"]').text.strip() if movie.select_one('rt-text[slot="audienceScore"]') else ''
link = movie.select_one('a[data-qa^="discovery-media-list-item"]')['href'] if movie.select_one('a[data-qa^="discovery-media-list-item"]') else ''

movie_data.append({
'title': title,
'critics_score': criticsScore,
'audience_score': audienceScore,
'link': 'https://www.rottentomatoes.com' + link
})

return movie_data

Handling Pagination

Rotten Tomatoes uses button-based pagination for their movie listings. We need to handle pagination by clicking the “Load More” button. Crawlbase Crawling API allows us to handle pagination with the css_click_selector parameter.

1
2
3
4
5
6
7
8
9
10
11
12
def fetch_html_with_pagination(url):
options = {
'ajax_wait': 'true',
'page_wait': '5000',
'css_click_selector': 'button[data-qa="dlp-load-more-button"]' # CSS Selector for "Load More" button
}
response = crawling_api.get(url, options)
if response['headers']['pc_status'] == '200':
return response['body'].decode('utf-8')
else:
print(f"Failed to fetch data. Status code: {response['headers']['pc_status']}")
return None

This code will click the “Load More” button to load more movie listings before scraping the data.

Storing Data in a JSON File

After scraping the movie data, we can save it to a file in a structured format like JSON.

1
2
3
4
5
6
import json

def save_to_json(data, filename='movies.json'):
with open(filename, 'w') as file:
json.dump(data, file, indent=4)
print(f"Data saved to {filename}")

Complete Code Example

Here is the full code that brings it all together, fetching the HTML, parsing the movie data, handling pagination and saving the results to a JSON file.

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
from crawlbase import CrawlingAPI
from bs4 import BeautifulSoup
import json

# Initialize Crawlbase API with your token
crawling_api = CrawlingAPI({'token': 'CRAWLBASE_JS_TOKEN'})

def fetch_html_with_pagination(url):
options = {
'ajax_wait': 'true',
'page_wait': '5000',
'css_click_selector': 'button[data-qa="dlp-load-more-button"]' # CSS Selector for "Load More" button
}
response = crawling_api.get(url, options)
if response['headers']['pc_status'] == '200':
return response['body'].decode('utf-8')
else:
print(f"Failed to fetch data. Status code: {response['headers']['pc_status']}")
return None

def parse_movies(html):
soup = BeautifulSoup(html, 'html.parser')
movies = soup.select('div[data-qa="discovery-media-list"] > div.flex-container')

movie_data = []
for movie in movies:
title = movie.select_one('span[data-qa="discovery-media-list-item-title"]').text.strip() if movie.select_one('span[data-qa="discovery-media-list-item-title"]') else ''
criticsScore = movie.select_one('rt-text[slot="criticsScore"]').text.strip() if movie.select_one('rt-text[slot="criticsScore"]') else ''
audienceScore = movie.select_one('rt-text[slot="audienceScore"]').text.strip() if movie.select_one('rt-text[slot="audienceScore"]') else ''
link = movie.select_one('a[data-qa^="discovery-media-list-item"]')['href'] if movie.select_one('a[data-qa^="discovery-media-list-item"]') else ''

movie_data.append({
'title': title,
'critics_score': criticsScore,
'audience_score': audienceScore,
'link': 'https://www.rottentomatoes.com' + link
})

return movie_data

def save_to_json(data, filename='movies.json'):
with open(filename, 'w') as file:
json.dump(data, file, indent=4)
print(f"Data saved to {filename}")

if __name__ == "__main__":
url = 'https://www.rottentomatoes.com/browse/movies_in_theaters/sort:top_box_office'
html_content = fetch_html_with_pagination(url)
if html_content:
movies_data = parse_movies(html_content)
save_to_json(movies_data)

Example Output:

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
[
{
"title": "Beetlejuice Beetlejuice",
"critics_score": "77%",
"audience_score": "81%",
"link": "https://www.rottentomatoes.com/m/beetlejuice_beetlejuice"
},
{
"title": "Deadpool & Wolverine",
"critics_score": "79%",
"audience_score": "95%",
"link": "https://www.rottentomatoes.com/m/deadpool_and_wolverine"
},
{
"title": "Alien: Romulus",
"critics_score": "80%",
"audience_score": "85%",
"link": "https://www.rottentomatoes.com/m/alien_romulus"
},
{
"title": "It Ends With Us",
"critics_score": "57%",
"audience_score": "91%",
"link": "https://www.rottentomatoes.com/m/it_ends_with_us"
},
{
"title": "The Forge",
"critics_score": "73%",
"audience_score": "99%",
"link": "https://www.rottentomatoes.com/m/the_forge"
},
{
"title": "Twisters",
"critics_score": "75%",
"audience_score": "91%",
"link": "https://www.rottentomatoes.com/m/twisters"
},
{
"title": "Blink Twice",
"critics_score": "74%",
"audience_score": "69%",
"link": "https://www.rottentomatoes.com/m/blink_twice"
},
... more
]

In the next section, we will discuss scraping individual movie details.

Scraping Rotten Tomatoes Movie Details

Now, we’ll move on to scraping individual movie details from Rotten Tomatoes. Once you have the movie listings, you need to extract detailed information for each movie, like release date, director, genre etc. This part will walk you through inspecting the HTML of a movie page, writing the scraper and saving the data to a JSON file.

Inspecting the HTML Structure

Before we start writing the scraper, we need to inspect the HTML structure of a specific movie’s page.

  1. Open Movie Page: Go to a movie’s page from the list you scraped earlier.

  2. Open Developer Tools: Right-click on the page and select “Inspect” or press Ctrl+Shift+I (Windows) or Cmd+Option+I (Mac).

  3. Locate Key Data:

  • Title: Usually in an

    with a slot="titleIntro" attribute.
  • Synopsis: In a
    with a class of synopsis-wrap and an rt-text element without the .key class.
  • Movie Details: These are in a list format with <dt> elements (keys) and <dd> elements (values). The data is usually in rt-link and rt-text tags.

Writing the Rotten Tomatoes Movie Details Scraper

Now that we have the HTML structure, we can write the scraper to get the details from each movie page. Lets write the scraper code.

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
from crawlbase import CrawlingAPI
from bs4 import BeautifulSoup

# Initialize Crawlbase API with your token
crawling_api = CrawlingAPI({'token': 'CRAWLBASE_JS_TOKEN'})

def fetch_html(url):
options = {
'ajax_wait': 'true',
'page_wait': '5000'
}
response = crawling_api.get(url, options)
if response['headers']['pc_status'] == '200':
return response['body'].decode('utf-8')
else:
print(f"Failed to fetch data. Status code: {response['headers']['pc_status']}")
return None

def fetch_movie_details(movie_url):
html = fetch_html(movie_url)
if html:
soup = BeautifulSoup(html, 'html.parser')

# Extract movie title
title = soup.select_one('h1[slot="titleIntro"]').text.strip()

# Extract synopsis
synopsis = soup.select_one('div.synopsis-wrap rt-text:not(.key)').text.strip() if soup.select_one('div.synopsis-wrap rt-text:not(.key)') else 'N/A'

# Get all movie details
movie_details = {dt.text.strip(): ', '.join([item.text.strip() for item in dd.find_all(['rt-link', 'rt-text']) if item.name != 'rt-text' or 'delimiter' not in item.get('class', [])]) for dt, dd in zip(soup.select('dt.key rt-text'), soup.select('dd'))}

# Return the collected details as a dictionary
return {
'title': title,
'synopsis': synopsis,
'movie_details': movie_details
}
else:
print("Failed to fetch movie details.")
return None

fetch_movie_details function fetches the movie title, release date, director and genres from the movie URL. It uses BeautifulSoup for HTML parsing and structures the data into a dictionary.

Storing Movie Details in a JSON File

After scraping the movie details, you will want to save the data in a structured format like JSON. Here is the code to save the movie details to a JSON file.

1
2
3
4
5
6
import json

def save_movie_details_to_json(movie_data, filename='movie_details.json'):
with open(filename, 'w') as file:
json.dump(movie_data, file, indent=4)
print(f"Movie details saved to {filename}")

Complete Code Example

Here is the complete code to scrape movie details from Rotten Tomatoes and save to a JSON file.

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
from crawlbase import CrawlingAPI
from bs4 import BeautifulSoup
import json

# Initialize Crawlbase API with your token
crawling_api = CrawlingAPI({'token': 'CRAWLBASE_JS_TOKEN'})

def fetch_html(url):
options = {
'ajax_wait': 'true',
'page_wait': '5000'
}
response = crawling_api.get(url, options)
if response['headers']['pc_status'] == '200':
return response['body'].decode('utf-8')
else:
print(f"Failed to fetch data. Status code: {response['headers']['pc_status']}")
return None

def fetch_movie_details(movie_url):
html = fetch_html(movie_url)
if html:
soup = BeautifulSoup(html, 'html.parser')

# Extract movie title
title = soup.select_one('h1[slot="titleIntro"]').text.strip()

# Extract synopsis
synopsis = soup.select_one('div.synopsis-wrap rt-text:not(.key)').text.strip() if soup.select_one('div.synopsis-wrap rt-text:not(.key)') else 'N/A'

# Get all movie details
movie_details = {dt.text.strip(): ', '.join([item.text.strip() for item in dd.find_all(['rt-link', 'rt-text']) if item.name != 'rt-text' or 'delimiter' not in item.get('class', [])]) for dt, dd in zip(soup.select('dt.key rt-text'), soup.select('dd'))}

# Return the collected details as a dictionary
return {
'title': title,
'synopsis': synopsis,
'movie_details': movie_details
}
else:
print("Failed to fetch movie details.")
return None

def save_movie_details_to_json(movie_data, filename='movie_details.json'):
with open(filename, 'w') as file:
json.dump(movie_data, file, indent=4)
print(f"Movie details saved to {filename}")

if __name__ == "__main__":
movie_url = 'https://www.rottentomatoes.com/m/beetlejuice_beetlejuice'
movie_details = fetch_movie_details(movie_url)
if movie_details:
save_movie_details_to_json(movie_details)

Example Output:

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
{
"title": "Beetlejuice Beetlejuice",
"synopsis": "Beetlejuice is back! After an unexpected family tragedy, three generations of the Deetz family return home to Winter River. Still haunted by Beetlejuice, Lydia's life is turned upside down when her rebellious teenage daughter, Astrid, discovers the mysterious model of the town in the attic and the portal to the Afterlife is accidentally opened. With trouble brewing in both realms, it's only a matter of time until someone says Beetlejuice's name three times and the mischievous demon returns to unleash his very own brand of mayhem.",
"movie_details": {
"Director": "Tim Burton",
"Producer": "Marc Toberoff, Dede Gardner, Jeremy Kleiner, Tommy Harper, Tim Burton",
"Screenwriter": "Alfred Gough, Miles Millar",
"Distributor": "Warner Bros. Pictures",
"Production Co": "Tommy Harper, Plan B Entertainment, Marc Toberoff, Tim Burton Productions",
"Rating": "PG-13 (Macabre and Bloody Images|Brief Drug Use|Some Suggestive Material|Strong Language|Violent Content)",
"Genre": "Comedy, Fantasy",
"Original Language": "English",
"Release Date (Theaters)": "Sep 6, 2024, Wide",
"Box Office (Gross USA)": "$111.0M",
"Runtime": "1h 44m",
"Aspect Ratio": "Flat (1.85:1)"
}
}

Scrape Movie Ratings on Rotten Tomatoes with Crawlbase

Scraping movie ratings from Rotten Tomatoes has many uses. You can collect data to analyze, research or just have fun. It helps you see what’s popular. The Crawlbase Crawling API can deal with dynamic content so you can get the data you need. When you scrape Rotten Tomatoes, you can get public opinions, box office stats or movie details to use in any project you like.

This blog showed you how to scrape movie listings, ratings, and get details like release dates, directors, and genres. We used Python, Crawlbase BeautifulSoup, and JSON to gather and sort the data so you can use and study it. This blog showed you how to scrape movie listings and ratings and get details like release dates, directors, and genres.

If you want to do more web scraping, check out our guides on scraping other key websites.

📜 How to Scrape Monster.com
📜 How to Scrape Groupon
📜 How to Scrape TechCrunch
📜 How to Scrape Clutch.co

If you have any questions or feedback, our support team can help you with web scraping. Happy scraping!

Frequently Asked Questions

Q. How do I scrape Rotten Tomatoes if the site changes its layout?

If Rotten Tomatoes alters its layout or HTML, your scraper will stop working. To address this:

  • Keep an eye on the site for any changes.
  • Examine the new HTML and revise your CSS selectors.
  • Make changes to your scraper’s code.

Q. What should I keep in mind when scraping Rotten Tomatoes?

When you scrape Rotten Tomatoes or similar websites:

  • Look at the Robots.txt: Make sure the site allows scraping by checking its robots.txt file.
  • Put throttling to use: Add some time between requests to avoid overwhelming the server and reduce your chances of getting blocked.
  • Deal with errors: Include ways to handle request failures or changes in how the site is built.

Q. How do I manage pagination with Crawlbase Crawling API while scraping Rotten Tomatoes?

Rotten Tomatoes may use different ways to show more content, like “Load More” buttons or endless scrolling.

  • For Buttons: Use the css_click_selector parameter in the Crawlbase Crawling API to click the “Load More” button.
  • For Infinite Scrolling: Use page_wait or ajax_wait parameter in the Crawlbase Crawling API to wait for all content to load before you capture.