Skip to main content

A coding challenge webscraper for leetcode, and other websites!

Project description

leetscraper · Downloads · Average time to resolve an issue · Percentage of issues still open

leetscraper is a coding challenge webscraper for leetcode, and other websites!
It was created as a way to gather coding problems to solve without having to sign up to a website or submit code to a problem checker.


Usage

Install package and dependencies

pip install leetscraper tqdm urllib3 beautifulsoup4 selenium webdriver-manager

Examples

Import the module and Instantiate the class. The class has some kwargs options to control the behaviour of the scraper. However, all the default values will start to scrape all problems from leetcode.com to the cwd.

The most basic usage looks like this:

from leetscraper import Leetscraper

if __name__ == "__main__":
    Leetscraper()

The avaliable kwargs to control the behaviour of the scraper are:

"""
website_name: "leetcode.com", "projecteuler.net", "codechef.com" ("leetcode.com" is set if ignored)
scraped_path: "path/to/save/scraped_problems" (Current working directory is set if ignored)
scrape_limit: Integer of how many problems to scrape at a time (-1 is set if ignored, which is no limit)
auto_scrape: "True", "False" (True is set if ignored)
"""

Example of how to automatically scrape the first 50 problems from projecteuler.net to a directory called SOLVE-ME:

from leetscraper import Leetscraper

if __name__ == "__main__":
    Leetscraper(website_name="projecteuler.net", scraped_path="~/SOLVE-ME", scrape_limit=50)

Example of how to scrape all problems from all supported websites:

from leetscraper import Leetscraper

if __name__ == "__main__":
    websites = ["leetcode.com", "projecteuler.net", "codechef.com"]

    for site in websites:
        Leetscraper(website_name=site)

You can pass through different arguments for different websites to control exactly how the scraper behaves. You can also disable scraping problems at time of instantiation by using the kwarg auto_scrape=False. This allows you to call the class functions in different order, or one at a time. This will change how the scraper works, as its designed to look in a directory for already scraped problems to avoid duplicates. I would encourage you to look at the function docstrings if you wish to use this scraper outside of its intended automated use.


Contributing

If you would like to contribute, adding support for a new coding challenge website, or fixing current bugs is always appreciated! I would encourage you to see CONTRIBUTING.md for further details. If you would like to report bugs or suggest websites to support, please add a card to Issues. Thank you to all contributors to this project!


Code of Conduct

Contributing to this project means you are willing to follow the same conduct that others are held to! Please see Code of Conduct for further details.


Licence

This project uses the GPL-2.0 License, As generally speaking, I want you to be able to do whatever you want with this project, But still have the ability to add your changes to this codebase should you make improvements or extend support. For further details on what this licence allows, please see LICENSE.md

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

leetscraper-1.1.1.tar.gz (14.0 kB view hashes)

Uploaded Source

Built Distribution

leetscraper-1.1.1-py3-none-any.whl (13.4 kB view hashes)

Uploaded Python 3

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page