In this blog, we'll guide you through the process of using Vite and Crawlee to build a website that recommends Netflix shows based on their categories and genres. To do that, we will first scrape the shows and categories from Netflix using Crawlee, and then visualize the scraped data in a React app built with Vite. By the end of this guide, you'll have a functional web show recommender that can provide Netflix show suggestions.
Note: One of our community members wrote this blog as a contribution to Crawlee Blog. If you want to contribute blogs like these to Crawlee Blog, please reach out to us on our discord channel.
Crawlee—A web scraping and browser automation library for Node.js to build reliable crawlers. In JavaScript and TypeScript. Extract data for AI, LLMs, RAG, or GPTs. Download HTML, PDF, JPG, PNG, and other files from websites. Works with Puppeteer, Playwright, Cheerio, JSDOM, and raw HTTP. Both headful and headless mode. With proxy rotation.
A web scraping and browser automation library
Crawlee covers your crawling and scraping end-to-end and helps you build reliable scrapers. Fast.
Your crawlers will appear human-like and fly under the radar of modern bot protections even with the default configuration. Crawlee gives you the tools to crawl the web for links, scrape data, and store it to disk or cloud while staying configurable to suit your project's needs.
First, we will create a React app (for the front end) using Vite. Run this command in the terminal to create it:
npx create-vite@latest
You can check out the Vite Docs for more details on how to create a React app.
Once the React app is created, open it in VS Code.
This will be the structure of your React app.
Run npm run dev command in the terminal to run the app.
This will be the output displayed.
Adding Scraper code
As per our project requirements, we will scrape the genres and the titles of the shows available on Netflix.
Let’s start building the scraper code.
Installation
Run this command to install Crawlee:
npm install crawlee
Crawlee utilizes Cheerio for HTML parsing and scraping of static websites. While faster and less resource-intensive, it can only scrape websites that do not require JavaScript rendering, making it unsuitable for SPAs (single page applications).
In this tutorial, we can extract data from the HTML structure, so we will go with Cheerio, but for extracting data from SPAs or JavaScript-rendered websites, Crawlee also supports headless browser libraries like Playwright and Puppeteer
After installing the libraries, it’s time to create the scraper code.
Create a file in src directory and name it scraper.js. The entire scraper code will be created in this file.
Scraping genres and shows
To scrape the genres and shows, we will utilize the browser DevTools to identify the tags and CSS selectors targeting the genre elements on the Netflix website.
We can capture the HTML structure and call $(element) to query the element's subtree.
Here, we can observe that the name of the genre is captured by a span tag with nm-collections-row-name class. So we can use the span.nm-collections-row-name selector to capture this and similar elements.
Similarly, we can observe that the title of the show is captured by the span tag having nm-collections-title-name class. So we can use the span.nm-collections-title-name selector to capture this and similar elements.
// Use parseWithCheerio for efficient HTML parsingconst$=awaitparseWithCheerio();// Extract genre and shows directly from the HTML structureconstdata=$('[data-uia="collections-row"]').map((_,el)=>{constgenre=$(el).find('[data-uia="collections-row-title"]').text().trim();constitems=$(el).find('[data-uia="collections-title"]').map((_,itemEl)=>$(itemEl).text().trim()).get();return{genre,items};}).get();constgenres=data.map((d)=>d.genre);constshows=data.map((d)=>d.items);
In the code snippet given above, we are using parseWithCheerio to parse the HTML content of the current page and extract genres and shows information from the HTML structure using Cheerio.
This will give the genres and shows array having list of genres and shows stored in it respectively.
Storing data
Now we have all the data that we want for our project and it’s time to store or save the scraped data. To store the data, Crawlee comes with a pushData() method.
The pushData() method creates a storage folder in the project directory and stores the scraped data in JSON format.
awaitpushData({genres:genres,shows:shows,});
This will save the genres and shows arrays as values in the genres and shows keys.
Here’s the full code that we will use in our project:
import{CheerioCrawler,log,Dataset}from"crawlee";constcrawler=newCheerioCrawler({requestHandler:async ({request,parseWithCheerio,pushData})=>{log.info(`Processing: ${request.url}`);// Use parseWithCheerio for efficient HTML parsingconst$=awaitparseWithCheerio();// Extract genre and shows directly from the HTML structureconstdata=$('[data-uia="collections-row"]').map((_,el)=>{constgenre=$(el).find('[data-uia="collections-row-title"]').text().trim();constitems=$(el).find('[data-uia="collections-title"]').map((_,itemEl)=>$(itemEl).text().trim()).get();return{genre,items};}).get();// Prepare data for pushingconstgenres=data.map((d)=>d.genre);constshows=data.map((d)=>d.items);awaitpushData({genres:genres,shows:shows,});},// Limit crawls for efficiencymaxRequestsPerCrawl:20,});awaitcrawler.run(["https://www.netflix.com/in/browse/genre/1191605"]);awaitDataset.exportToJSON("results");
Now, we will run Crawlee to scrape the website. Before running Crawlee, we need to tweak the package.json file. We will add the start script targeting the scraper.js file to run Crawlee.
Add the following code in 'scripts' object:
"start": "node src/scraper.js"
and save it. Now run this command to run Crawlee to scrape the data:
npm start
After running this command, you will see a storage folder with the key_value_stores/default/results.json file. The scraped data will be stored in JSON format in this file.
Now we can use this JSON data and display it in the App.jsx component to create the project.
In the App.jsx component, we will import jsonData from the results.json file:
import{useState}from"react";import"./App.css";importjsonDatafrom"../storage/key_value_stores/default/results.json";functionHeaderAndSelector({handleChange}){return (<><h1className="header">NetflixWebShowRecommender</h1>
<divclassName="genre-selector"><selectonChange={handleChange}className="select-genre"><optionvalue="">Selectyourgenre</option>
{jsonData[0].genres.map((genres,key)=>{return (<optionkey={key}value={key}>{genres}</option>
);})}</select>
</div>
</>
);}functionApp(){const[count,setCount]=useState(null);consthandleChange=(event)=>{constvalue=event.target.value;if (value)setCount(parseInt(value));};// Validate count to ensure it is within the bounds of the jsonData.shows arrayconstisValidCount=count!==null&&count<=jsonData[0].shows.length;return (<divclassName="app-container"><HeaderAndSelectorhandleChange={handleChange}/>
<divclassName="shows-container">{isValidCount&&(<><divclassName="shows-list"><ul>{jsonData[0].shows[count].slice(0,20).map((show,index)=>(<likey={index}className="show-item">{show}</li>
))}</ul>
</div>
<divclassName="shows-list"><ul>{jsonData[0].shows[count].slice(20).map((show,index)=>(<likey={index}className="show-item">{show}</li>
))}</ul>
</div>
</>
)}</div>
</div>
);}exportdefaultApp;
In this code snippet, the genre array is used to display the list of genres. User can select their desired genre and based upon that a list of web shows available on Netflix will be displayed using the shows array.
In this project, we used Crawlee to scrape Netflix; similarly, Crawlee can be used to scrape single application pages (SPAs) and JavaScript-rendered websites. The best part is all of this can be done while coding in JavaScript/TypeScript and using a single library.