Skip to main content

Scrappe all products and theirs related suppliers existing on Alibaba based on keywords provided by user and save results into a database (Mysql/Sqlite).

Project description

overview image

Alibaba-CLI-Scraper

Is a python package that provides a dedicated CLI interface for scraping data from Alibaba.com. The purpose of this project is to extract products and theirs related suppliers informations from Alibaba.com and store it in a local database (SQLite or MySQL). The project utilizes asynchronous requests for efficient handling of numerous requests and allows users to easily run the scraper and manage the database using a user-friendly command-line interface (CLI).

Installation

It's recommended to use pipx instead of pip for end-user applications written in Python. pipx installs the package, exposes his CLI entrypoints in an isolated environment and makes it available everywhere this guarantees no dependency conflicts and clean uninstall. If you'd like to use pip instead, just replace pipx with pip but obviously as usual you'll need to create a virtual environment and activate it before to use aba-cli-scrapper to avoid any dependency conflicts issues. let's install aba-cli-scrapper using pipx:

   pipx install aba-cli-scrapper

Using the CLI Interface

Available Commands:

Need Help? run any commands followed by --help for detailed informations about its usage and options. For example: aba-run --help will show you all subcommands available and how to use them.

command result 1

Warnings:

  • aba-run is the base command means all other commands that will be introduce bellow are sub-commands and should always be preceded by aba-run.

  • asynchronous requests are used by default, but is not available for moment cause bright-data api has been exhausted. Instead you can use sync api by specifying --sync-api or -sa flag parameter with scraper sub-command is works perfecly fine. So let's jump to the tutorial.

Practice make perfect isn't ? So let's get started with a use case example. Let's assume that you want to scrape data about electric bikes from Alibaba.com.

Scraper Demo

https://user-images.githubusercontent.com/49741340/238535232-459847af-a15c-4d9b-91ac-fab9958bc74f.mp4

  • scraper sub-command: Initiates scraping of Alibaba.com based on the provided keywords. this command takes two required arguments and one optional argument:

      • key_words (required): The search term(s) for finding products on Alibaba. Enclose multiple keywords in quotes.
      • --page-results or -pr (required): Usually keys words will results to many pages macthing them. Then you must to indicate how many of them you want to pull out.If any value is not provided 10 will be used by default.
      • --html-folder or -hf (optional): Specifies the directory to store the raw HTML files. If omitted, a folder with sanitized keywords as name will be automatically created. In this case electric_bikes will be used as a results folder name.

    Example:

    aba-run scraper "electric bikes" -hf "bike_results" -pr 15
    

by default scrapper will use async which is as explained unstable. then if you want to use sync api run:

aba-run scraper "electric bikes" -hf "bike_results" -pr 15  --sync-api
and voila! 

Now bike_results (since you already provided name you wish to have) directory has been created and should contains all html files from alibaba.com matching your keywords.

db-init Demo

https://user-images.githubusercontent.com/49741340/238535232-459847af-a15c-4d9b-91ac-fab9958bc74f.mp4

Then you must initialize a database. Mysql and sqlite are supported.
  • db-init sub-command: Creates a new database mysql/sqlite. this command takes one required arguments and six optional arguments(depends on engine you choose):
      • engine (required): Choose either sqlite or mysql.
      • --sqlite-file or -f(optional, SQLite only): The name for your SQLite database file (without any extension).
      • --host or -h, --port or -p, --user or -u, --password or -pw, --db-nameor -db (required for MySQL): Your MySQL database connection details.
    • --only-with or -ow(optional Mysql): If you just want to update some details of your credentials in db_credentials.json file but not all before to initialize a brand new database.

  • NB: --host and --port are respectively set to localhost and 3306 by default.

MySQL Use case:

aba-run db-init mysql -u "mysql_username" -pw "mysql_password" -db "alibaba_products" 

Assuming that you have already initialized your database,and you want to created a new one with a new database name without to set password and username again , simply run :

aba-run db-init mysql --only-with -db "alibaba_products" 

NB: When you initialize your mysql as engine, the db-init sub-command will save your credentials in db_credentials.json file, so when you will need to update your database, simply run aba-run db-update mysql --kw-results bike_results\ to automatically update your database by using your saved credentials

SQLite Use case :

aba db-init sqlite --sqlite-file alibaba_data

As soons as your database has been initialized, you can update it with the scraped data.

db-update Demo

https://user-images.githubusercontent.com/49741340/238535232-459847af-a15c-4d9b-91ac-fab9958bc74f.mp4

  • db-update sub-command: add scraped data from html files to your database (you can't use this command twice with same database credentals to avoid UNIQUE CONSTRAINT ERROR).

this command takes two required arguments and two optional arguments:

    • --db-engine (required): Select your database engine: sqlite or mysql.
    • --kw-results (required): The path to the folder containing the HTML files generated by the scraper sub command.
    • --filename (required for SQLite): If you're using SQLite, provide the desired filename for your database. whitout any extension.
    • --db-name (optional for MySQL): If you're using MySQL,and want to push the data to a different database, provide the desired database name.

MySQL Use case:

aba-run db-update  mysql --kw-results bike_results\ 

NB:What if you want to change something while you updating the database? Assuming that you have run another scraping command and you want to save this data in another database name whitout update credential file or rewriting all theses parameter just to change your database name then, simply run aba-run db-update mysql --kw-results another_keyword_folder_result\ --db-name "another_database_name".

SQLite Use case:

aba-run db-update  sqlite --kw-results bike_results\ --filename alibaba_data

Features:

  • Asynchronous Scraping: Utilizes asynchronous API of Playwright for efficient handling of numerous pages results.
  • Database Integration: Stores scraped data in a database (SQLite or MySQL) for structured persistence.
  • User-Friendly CLI: Provides easy-to-use commands for running the scraper and managing the database.

Future Enhancements

This project has a lot of potential for growth! Here are some exciting features I'm considering for the future:

  • Data Export: Add functionality to export scraped data to various formats like CSV and Excel spreadsheets for easier analysis and sharing.
  • Retrieval Augmented Generation (RAG): Integrate a RAG system that allows users to ask natural language questions about the scraped data, making it even more powerful for insights.

Contributions Welcome!

I believe in the power of open source! If you'd like to contribute to this project, feel free to fork the repository, make your changes, and submit a pull request. I'm always open to new ideas and improvements.

License

This project is licensed under the Gnu General Public License Version 3.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

aba_cli_scrapper-0.1.10.tar.gz (3.2 MB view details)

Uploaded Source

Built Distribution

aba_cli_scrapper-0.1.10-py3-none-any.whl (3.3 MB view details)

Uploaded Python 3

File details

Details for the file aba_cli_scrapper-0.1.10.tar.gz.

File metadata

  • Download URL: aba_cli_scrapper-0.1.10.tar.gz
  • Upload date:
  • Size: 3.2 MB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/1.8.3 CPython/3.12.2 Windows/10

File hashes

Hashes for aba_cli_scrapper-0.1.10.tar.gz
Algorithm Hash digest
SHA256 6c865324837d646d63453cc09fd0a0c636f30149a7afcab4828160d5fc5ad438
MD5 fd2e7c7bcf87a42c972f9973cdb10a76
BLAKE2b-256 95abe5b6fa7a21f10f225f511409fdbe0cec946f34bfe748ac2d3a0bc3ca0844

See more details on using hashes here.

Provenance

File details

Details for the file aba_cli_scrapper-0.1.10-py3-none-any.whl.

File metadata

File hashes

Hashes for aba_cli_scrapper-0.1.10-py3-none-any.whl
Algorithm Hash digest
SHA256 8f1b06dfb1e02a4dfac72eff8586d35bff40b230d1c4225c72dbd1b1e62a80a5
MD5 2c7216b73369d9a9f044123a3fbe5593
BLAKE2b-256 4c47dd2e4f9d8814dd2c49f81dbd24b0d8d9f9332b4f2348f7b66e4103a6773b

See more details on using hashes here.

Provenance

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page