Python web scraping with BeautifulSoup

Pierre - Jan 22 '20 - - Dev Community

In this article, we will see how to extract structured information from web page leveraging BeautifulSoup and CSS selectors.

WebScraping with BeautifulSoup

Pulling the HTML out

BeautifulSoup is not a web scraping library per se. It is a library that allows you to efficiently and easily pull out information from HTML, in the real world, it is very often used for web scraping project.

So to begin, we'll need HTML. We will begin by pulling out HackerNews landing page HTML using requests python package.

import requests
response = requests.get("https://news.ycombinator.com/")
if response.status_code != 200:
    print("Error fetching page")
    exit()
else:
    content = response.content
print(content)

> b'<html op="news"><head><meta name="referrer" content="origin"><meta name="viewport" 
> content="width=device-width, initial-scale=1.0"><link rel="stylesheet" type="text/css" href="news.css? ...
Enter fullscreen mode Exit fullscreen mode

Parsing the HTML with BeautifulSoup

Now that the HTML is accessible we will use BeautifulSoup to parse it. If you haven't done already you need to install the package by doing a simple pip install beautifullsoup4. In the rest of this article, we will refer to BeautifulSoup4 as BS4.

We now need to parse the HTML and load it into a BS4 structure.

from bs4 import BeautifulSoup
soup = BeautifulSoup(response.content, 'html.parser')
Enter fullscreen mode Exit fullscreen mode

This soup object is very handy and allows us to easily access many useful pieces of information such as:

# The HTML title of the page
print(soup.title)
> <title>Hacker News</title>

# The test title of the page
print(soup.title.string)
> Hacker News

# All links in the page
nb_links = len(soup.find_all('a'))
print(f"There are {nb_links} links in this page")
> There are 231 links in this page

# Text from the page
print(soup.get_text())
> Hacker News
> Hacker News
> new | past | comments | ask | show | jobs | submit
> login
> ...
Enter fullscreen mode Exit fullscreen mode

Targeting DOM elements

You might begin to see a pattern in how to use this library. This library allows you to quickly and elegantly target the DOM elements you need.

If you need to select DOM elements from its tag (<p>, <a>, <span>, ....) you can simply do soup.<tag> to select it. The caveat is that it will only select the first HTML element with that tag.

For example if I want the first link I just have to do

 first_link = soup.a
 print(first_link)
 ><a href="https://news.ycombinator.com"><img height="18" src="y18.gif" style="border:1px white solid;" width="18"/></a>
Enter fullscreen mode Exit fullscreen mode

And this element will also have many useful method to quickly extract informations:

# The text of the link
print(first_link.text)
# Empty because first link is an <img>
>""

# The href of the link
print(first_link.get('href')
> https://news.ycombinator.com
Enter fullscreen mode Exit fullscreen mode

This is a simple example, if you want to select the first element based on its id or class it is not much more difficult:

pagespace = soup.find(id="pagespace")
print(pagespace)
> <tr id="pagespace" style="height:10px" title=""></tr>

# class is a reserved keyword in Python, hence the '_'
athing = soup.find(class_="athing")
print(athing)
> <tr class="athing" id="22115671">
> ...
Enter fullscreen mode Exit fullscreen mode

That is as simple.

And if you don't want the first matching element but all matching elements, just replace find by find_all.

This simple and elegant interface allows you to quickly write short and powerful Python snippets. For example, let's say that I want to extract all links in this page, and find the top 3 links that appears the most in the page, all I have to do is this:

from collections import Counter
all_hrefs = [a.get('href') for a in soup.find_all('a')]
top_3_links = Counter(all_hrefs).most_common(3)
print(top_3_links)
> [('from?site=github.com', 3), ('item?id=22115671', 2), ('item?id=22113827', 2)]
Enter fullscreen mode Exit fullscreen mode

Advanced usage

BeautifulSoup is a great example of a library that is both easy to use and powerful. There is much more you can do to select elements, we won't cover those cases in this article but here are few examples of advanced things you can do with the relevant documentation links:

We also only covered how to target elements but there is also a whole section about updating and writing HTML but again we won't cover it in this article.

Let's now talk about CSS selectors

CSS selectors

But why learn about CSS selectors if BeautifulSoup can select all elements with its pre-made method ?

Well, you'll soon understand.

Hard dom

Sometimes the HTML document you'll work with won't have useful class and id. So to select elements with BS4 without relying on that information it can be quite verbose.

For example, let's say that you want to extract score of post in HN homepage and that you can't use class name or id in your code, here is how you could do it:

results = []
all_tr = soup.find_all('tr')
for tr in all_tr:
    if len(tr.contents) == 2:
        print(len(tr.contents[1]))
        if len(tr.contents[0].contents) == 0 and len(tr.contents[1].contents) == 13:
            points = tr.contents[1].text.split(' ')[0].strip()
            results.append(points)
print(results)
>['168', '80', '95', '344', '7', '84', '76', '2827', '247', '185', '91', '2025', '482', '68', '47', '37', '6', '89', '32', '17', '47', '1449', '25', '73', '35', '463', '44', '329', '738', '17']
Enter fullscreen mode Exit fullscreen mode

Not that great right?

If you rely on CSS selectors, it becomes easier.

all_results = soup.select('td:nth-child(2) > span:nth-child(1)')
results = [r.text.split(' ')[0].strip() for r in all_results]
print(results)

>['168', '80', '95', '344', '7', '84', '76', '2827', '247', '185', '91', '2025', '482', '68', '47', '37', '6', '89', '32', '17', '47', '1449', '25', '73', '35', '463', '44', '329', '738', '17']
Enter fullscreen mode Exit fullscreen mode

Much clearer and simpler right? Of course this example artificially hightlights the usefullness of the css selector but quickly, you will see that the dom structure of a page is more reliable than than the class name.

Easily debuggable

Another thing that makes CSS expression great for web scraping is that there are easily debuggable, I'll show you how. Open Chrome, your developers' tools, (left-click -> "Inspect"), click on the document panel and do "Ctrl-F or CMD-F" to be in search mode.

In the search bar, you'll be able to write any CSS expression you want and Chrome will find instantly all elements matching it.

Iterate over results by pressing Enter to check that you are correctly getting everything you need.

But what is great with Chrome is that it works the other way around too, you can also left click on an element, click "Copy -> Copy Selector" and your selector will be pasted in your clipboard.

Powerful

CSS selectors, and particularly pseudo-classes allow you to select any elements you want with only one simple string.

Child and descendants

You can select direct child and descendant with:

# all <p> directly inside and <a>
a > p

# all <p> descendant of an <a>
a p
Enter fullscreen mode Exit fullscreen mode

And you can mix them together

a > p > .test .example > span
Enter fullscreen mode Exit fullscreen mode

This will totally works

Siblings

This one is one of my favorites because it allows you to select elements based on the elements on the same level in the DOM hierarchy, hence the sibling expression.

#html example
<p>...</p>
<section>
    <p>...</p>
    <h2>...</h2>
    <p>This paragraph will be selected</p> (match h2 ~ p / h2 + p)
    <div>
        <p>...</p>
    </div>
    <p>This paragraph will be selected</p> (match h2 ~ p)
</section>
Enter fullscreen mode Exit fullscreen mode

To select all p coming after an h2 you can use the h2 ~ p selector (it will match two p). You can also use h2 + p if you only want to select p coming directly after an h2 (it will match only one p)

Attribute selectors

Attribute selectors allow you to select element with particular attributes values, p[data-test="foo"] will match

<p data-test="foo"></p>
Enter fullscreen mode Exit fullscreen mode

Position pseudo classes

If you want to select last p inside a section, you can also do it in "pure" CSS by leveraging position pseudo-classes. For this particular example, you just need this selector: section p:last-child(). If you want to learn more about this I suggest you take a look at this article

Maintainable code

I also personally think that CSS expressions are easier to maintain. For example at ScrapingBee, when we do custom web scraping tasks all our scripts begins like this:

    TITLE_SELECTOR = "title"
    SCORE_SELECTOR = "td:nth-child(2) > span:nth-child(1)"
    ...
Enter fullscreen mode Exit fullscreen mode

It makes it easy and quick to fix scripts when DOM changes appear. The laziest way to do it is to simply copy/paste what Chrome gives you when you left-click on an element. If you do this be careful, Chrome tends to add a lot of useless selectors when you use this trick so do not hesitate to clear them a bit before using them in your script.

Conclusion

In the end, everything you do with pure CSS selector you can do it with BeautifulSoup4 but I think privileging the former is the good way to go.

I hoped you liked this article about web scraping in Python and that it will make your life easier.

If you'd like to read more about web scraping in Python do not hesitate to check out our extensive Python web scraping guide).

Happy Scraping

Pierre de Wulf

. . . . . . . . . . . . . . . . . . . . . . . . . . .