A simple scraper for Youtube
Project description
Youtube Simple Scraper
This is a simple youtube scraper that uses the youtube API to get the videos metadata and comments of a channel.
You don't need an API key to use this scraper, so there are no usage limits or associated costs.
It should be noted that although there are no limits on use, YouTube can block the IP if it detects abusive use of the API.
Features
Scrape the following information of a channel:
- Channel metadata
- Videos metadata and comments
- Shorts metadata and comments
Installation
pip install youtube_simple_scraper
Usage
from youtube_simple_scraper.entities import GetChannelOptions
from youtube_simple_scraper.list_video_comments import ApiVideoCommentRepository, ApiShortVideoCommentRepository
from youtube_simple_scraper.list_videos import ApiChannelRepository
from youtube_simple_scraper.logger import build_default_logger
from youtube_simple_scraper.network import Requester
from youtube_simple_scraper.stop_conditions import ListCommentMaxPagesStopCondition, \
ListVideoMaxPagesStopCondition
if __name__ == '__main__':
##############################
# To Avoid IP Blocking
# Set the request rate per second to 0.5 seconds
Requester.request_rate_per_second = 0.5
# In every request sleep between 1 and 5 seconds
Requester.min_sleep_time_sec = 1
Requester.max_sleep_time_sec = 5
# Every 100 requests sleep 30 seconds
Requester.long_sleep_time_sec = 30
Requester.long_sleep_after_requests = 100
##############################
logger = build_default_logger()
video_comment_repo = ApiVideoCommentRepository()
short_comment_repo = ApiShortVideoCommentRepository()
repo = ApiChannelRepository(
video_comment_repo=video_comment_repo,
shorts_comment_repo=short_comment_repo,
logger=logger,
)
opts = GetChannelOptions(
list_video_stop_conditions=[
ListVideoMaxPagesStopCondition(2) # Stop after 2 pages of videos
],
list_video_comment_stop_conditions=[
ListCommentMaxPagesStopCondition(3) # Stop after 3 pages of comments
],
list_short_stop_conditions=[
ListVideoMaxPagesStopCondition(1) # Stop after 1 page of shorts
],
list_short_comment_stop_conditions=[
ListCommentMaxPagesStopCondition(4) # Stop after 4 pages of comments
]
)
channel_ = repo.get_channel("BancoFalabellaChile", opts)
print(channel_.model_dump_json(indent=2))
Example of the output channel object parsed to json:
{
"id": "UCaY_-ksFSQtTGk0y1HA_3YQ",
"name": "IbaiLlanos",
"target_id": "668be16f-0000-20de-b6a2-582429cfbdec",
"title": "Ibai",
"description": "contenido premium ▶️\n",
"subscriber_count": 11600000,
"video_count": 1400,
"videos": [
{
"id": "VFXu8gzcpNc",
"title": "EL RESTAURANTE MÁS ÚNICO AL QUE HE IDO NUNCA",
"description": "MI CANAL DE DIRECTOS: https://www.youtube.com/@Ibai_TV\nExtraído de mi canal de TWITCH: https://www.twitch.tv/ibai/\nMI PODCAST: \nhttps://www.youtube.com/channel/UC6jNDNkoOKQfB5djK2IBDoA\nTWITTER:...",
"date": "2024-06-02T19:18:27.647137",
"view_count": 1455817,
"like_count": 0,
"dislike_count": 0,
"comment_count": 0,
"thumbnail_url": "https://i.ytimg.com/vi/VFXu8gzcpNc/hqdefault.jpg?sqp=-oaymwEbCKgBEF5IVfKriqkDDggBFQAAiEIYAXABwAEG&rs=AOn4CLCEmoQtslruHk-droajdw0KJUI_KA",
"comments": [
{
"id": "UgzV8lY8eJ4dyHjl9Bp4AaABAg",
"text": "Todo muy rico pero....Y la cuenta?",
"user": "@eliasabregu2813",
"date": "2024-06-03T19:11:28.109467",
"likes": 0
},
{
"id": "UgwHtPZb8jprbCH-ysp4AaABAg",
"text": "Que humilde Ibai, comiendo todo para generar ingresos a los nuevos negocios",
"user": "@user-ui2sk7sr5i",
"date": "2024-06-03T19:04:28.112228",
"likes": 0
}
]
},
// More videos ...
],
"shorts": [
// the shorts videos and comments
]
}
Stop conditions
Videos stop conditions
- ListVideoMaxPagesStopCondition: Stops the scraping process when the number of pages scraped is greater than the specified value.
- ListVideoNeverStopCondition: The scraping process stop when all the videos of the channel are scraped.
Comments stop conditions
- ListCommentMaxPagesStopCondition: Stops the scraping process when the number of pages scraped is greater than the specified value.
- ListCommentNeverStopCondition: The scraping process stop when all the comments of the video are scraped.
The stop conditions are used to stop the scraping process. The following stop conditions are available:
Project details
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file youtube_simple_scraper-0.0.7.tar.gz
.
File metadata
- Download URL: youtube_simple_scraper-0.0.7.tar.gz
- Upload date:
- Size: 18.2 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/5.1.0 CPython/3.12.2
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | ca39258b1abaf2ac0c4f03b482a3c82125ac161a816140d70e443dcf1d3f898d |
|
MD5 | b4f4a823e04d07da09bd3fd924c4defa |
|
BLAKE2b-256 | 60985a6a3cf8d7e3762ae3b28ce08b4e98e7665d20a8c1cc15f053f67a3d264f |
File details
Details for the file youtube_simple_scraper-0.0.7-py3-none-any.whl
.
File metadata
- Download URL: youtube_simple_scraper-0.0.7-py3-none-any.whl
- Upload date:
- Size: 18.8 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/5.1.0 CPython/3.12.2
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 4fea68aa5d98a1dc2c19cd446747dd708956c4844803fb8a984480de7160e40a |
|
MD5 | c0aeb8d3a0985d5f155da3e3b680e014 |
|
BLAKE2b-256 | 3e4fe3924cefa10803083045e1155bcc672b4b6de2f9bcf6761c0762d71dd967 |