Skip to main content

A simplest HTML parsing library.

Project description

README

A simplest html parsing library.

Key features:

  • no third-party dependencies
  • no need to know CSS, Xpath or complicated rules to find element
  • interaction with native python lambda syntax or function-predicate
  • opportunity to work with damaged html
  • ability to use element relations (find ancestor, descendant, siblings)
  • a standard find first element or find all by current filter

Installation

Via pip:

pip install py_parse

First example

Lets get src attribute (link) of the Google logo on google.com

import requests
from py_parse import Parser

# get content of the google web page
content = requests.get('https://www.google.com/').text
# find first element with img-tag and 'alt' attribute equal to Google (logo)
google_logo = Parser().parse(content).find(lambda e: e.tag == 'img' and e.alt == 'Google')
# prints src attribute of the logo element
print(google_logo.src)

You will see following result

/images/branding/googlelogo/1x/googlelogo_white_background_color_272x92dp.png

If there is no element with current filter, you will get exception with filter text (if lambda was used): For code above lets say we use wrong filter

google_logo = Parser().parse(content).find(lambda e: e.tag == 'img' and e.alt == 'Wrong')

You will see following result

...traceback...
py_parse.exceptions.NoSuchElementError: No elements with current filter (e.tag == 'img' and e.alt == 'Wrong')

HOW IT WORKS

During parsing all html-elements in DOM converts to Node objects, which remains all relations (parent, child, sibling) and get all their attributes from html-element.

For example

<div class="some" type="submit">My text</div>

will return (after parsing):

from py_parse import Parser

element = '<div class="some" type="submit">My text</div>'
html_element = Parser().parse(element)[0]
print(html_element.text)  # My text
print(html_element.tag)  # div
print(html_element.class_)  # some

As you can see, all html attributes became object attributes, so you can use it in your filters

But remember:

  • Attribute tag is required, always present and can't be None
  • Attribute text always present BUT can be None
  • Attribute class became class_ in an object (html_element.class_), and it is not required

As you know, web page is a hierarchy, where html is a ancestor for all elements, and they all are nested in html. Function parse returns Nodes object, that is just container (like list) for Node objects. And for most cases that Nodes will have just one element (html), which contains all other elements inside (nested). So, for using search, you need to use methods like find or find_all of the Nodes.

The main idea is simple parsing without knowledge of CSS or XPATH, user need to know just a basic part of html. If you can understand what is html-element, tag, attribute, parent, child -you can parse all that you need.

Text - is that we looking for

User parse html to get some text of other values from it. Attributes is simple as get something from any python object. For example, when we parse following html

<div class="some" value="text"></div>

we can get value simple by using attribute:

value = Parser().parse('<div class="some" value="text"></div>').find_tag('div').value
print(value)  # text

If text of the element is split by other tags, it will be split by newline character in resulting text. For example div text had some br's inside:

<div>One<br>Two<br>Three</div>

result will be

text_with_n = Parser().parse('<div>One<br>Two<br>Three</div>').find_tag('div').text
print(text_with_n)  # One\nTwo\nThree
texts = text_with_n.split('\n')
print(texts)  # ['One', 'Two', 'Three']

And finally we can get cleared text - the text without newlines characters, tabs or leading/following spaces. so following html

<div>  One<br>  Two<br>         Three   </div>

transforms to

clear_text = Parser().parse('<div>  One<br>\nTwo<br>\tThree   </div>').find_tag('div').cleared_text()
print(clear_text)  # OneTwoThree

Find methods

Method find_all of the Nodes objects returns all found elements. If you not specify a filter, then all elements be in result. With filter, you get only elements, that satisfying the condition in it. If there are no such elements, then empty Nodes container returns.

Method find based on find_all, but returns just first element with that filter. If there are no results, then exception will be raised.

Methods find_tags and find_tag are similar to find_all, find, but it makes search with tag more convinient. For example

find_all(lambda e: e.tag == 'div')

is equivalent to

find_tags('div')

Simple Filtering

For all examples we will use content of the python documentation page https://docs.python.org/3/

So, start of all code is

import requests
from py_parse import Parser

content = requests.get('https://docs.python.org/3/').text

1. Find by tag

Let's find first element with 'strong' tag and get it text

first_strong = Parser().parse(content).find(lambda e: e.tag == 'strong')
print(first_strong.text)  # Parts of the documentation:

or you can use special find_tag method:

first_strong = Parser().parse(content).find_tag('strong')
print(first_strong.text)  # Parts of the documentation:

2. Find by tag and text (always present in any element)

tables = Parser().parse(content).find(lambda e: e.tag == 'strong' and e.text and e.text == 'Indices and tables:')
print(tables.text)  # Indices and tables:

or

tables = Parser().parse(content).find_tag('strong', lambda e: e.text == 'Indices and tables:')
print(tables.text)  # Indices and tables:

3. Find by containing text

copyright_ = Parser().parse(content).find(lambda e: 'pyri' in e.text)  # pyri is a part of Copyright
print(copyright_)  # <a class="biglink" href="copyright.html">Copyright</a>

In this example we print Node object itself, but not its text attribute.

4. Find element which has id

element_with_id = Parser().parse(content).find(lambda e: 'id' in e)  # 'id' in e - checks element has "id" attribute
print(element_with_id)
# <script id="documentation_options" data-url_root="./" src="_static/documentation_options.js"></script>

5. Find element by tag and type, then get it value

Let's find 'Go' button to search on documentation page

# finds element with input tag, which has type and this type equal to submit
go = Parser().parse(content).find(lambda e: e.tag == 'input' and e.type == 'submit')
print(go.value)  # Go

or

go = Parser().parse(content).find_tag('input', lambda e: e.type == 'submit')
print(go.value)  # Go

6. Finds all script elements

scripts = Parser().parse(content).find_all(lambda e: e.tag == 'script')  # Using find_all to finds all elements
for script in scripts:
    print(script)
# <script id="documentation_options" data-url_root="./" src="_static/documentation_options.js"></script>
# <script src="_static/jquery.js"></script>
# <script src="_static/underscore.js"></script>
# <script src="_static/doctools.js"></script>
# <script src="_static/language_data.js"></script>
# <script src="_static/sidebar.js"></script>
# <script type="text/javascript" src="_static/copybutton.js"></script>
# <script type="text/javascript">$('.inline-search').show(0);</script>
# <script type="text/javascript">$('.inline-search').show(0);</script>
# <script type="text/javascript" src="_static/switchers.js"></script>

or with special find_tags method:

scripts = Parser().parse(content).find_tags('script')  # Using find_tags to finds all elements with that tag
for script in scripts:
    print(script)

7. Finds all elements by part of the class name

sphinxes = Parser().parse(content).find_all(
    lambda e: 'sphinx' in e.class_)  # finds all elements containing 'sphinx' in class name
for sphinx in sphinxes:
    print(sphinx)
# <div class="sphinxsidebar" role="navigation" aria-label="main navigation"></div>
# <div class="sphinxsidebarwrapper"></div>

8. Using function-predicate

Sometimes your search filter became too long and heavy to use it in lambda, in that cases you should create a function-predicate with clear name to use in filters!

def images_with_alt_and_style(e) -> bool: # predicate function (takes html element and returns bool)
    return e.tag == 'img' and 'alt' in e and 'middle' in e.style


li_s = Parser().parse(content).find_all(images_with_alt_and_style)
for li in li_s:
    print(li)

# <img src="_static/py.png" alt="" style="vertical-align: middle; margin-top: -1px"></img>
# <img src="_static/py.png" alt="" style="vertical-align: middle; margin-top: -1px"></img>

Relations filtering

Contact me

Lexman2@yandex.ru

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

py_parse-0.2.0.tar.gz (13.3 kB view details)

Uploaded Source

Built Distribution

py_parse-0.2.0-py3-none-any.whl (12.1 kB view details)

Uploaded Python 3

File details

Details for the file py_parse-0.2.0.tar.gz.

File metadata

  • Download URL: py_parse-0.2.0.tar.gz
  • Upload date:
  • Size: 13.3 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/3.2.0 pkginfo/1.5.0.1 requests/2.24.0 setuptools/50.3.1 requests-toolbelt/0.9.1 tqdm/4.50.2 CPython/3.8.6

File hashes

Hashes for py_parse-0.2.0.tar.gz
Algorithm Hash digest
SHA256 7f7ac2a44491e71d9caa0d9e44e3c28847fe57c1c756a4e381ddd599869cd7b8
MD5 9202d73083f966a3a4dd8e3fbbdad2e0
BLAKE2b-256 f85c908a773baf83185ab10398d795aaddf127f3f21957927adadcb8474849c6

See more details on using hashes here.

File details

Details for the file py_parse-0.2.0-py3-none-any.whl.

File metadata

  • Download URL: py_parse-0.2.0-py3-none-any.whl
  • Upload date:
  • Size: 12.1 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/3.2.0 pkginfo/1.5.0.1 requests/2.24.0 setuptools/50.3.1 requests-toolbelt/0.9.1 tqdm/4.50.2 CPython/3.8.6

File hashes

Hashes for py_parse-0.2.0-py3-none-any.whl
Algorithm Hash digest
SHA256 3e71ed8043b5d78865552a6ecb1a7bdbf5324e8d594b44d040c91a14ed810503
MD5 f12f6299c5f70adaf8e45f7675f34b45
BLAKE2b-256 a7284cb7fc7f45baf5f33a57d861e63e23c28db666f8a9f26bba0709a22a6ff0

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page