Skip to main content

A Python library for scraping the Google search engine.

Project description

googlesearch

googlesearch is a Python library for searching Google, easily. googlesearch uses requests and BeautifulSoup4 to scrape Google.

Installation

To install, run the following command:

python3 -m pip install googlesearch-python

Usage

To get results for a search term, simply use the search function in googlesearch. For example, to get results for "Google" in Google, just run the following program:

from googlesearch import search
search("Google")

Additional options

googlesearch supports a few additional options. By default, googlesearch returns 10 results. This can be changed. To get a 100 results on Google for example, run the following program.

from googlesearch import search
search("Google", num_results=100)

If you want to have unique links in your search result, you can use the unique option as in the following program.

from googlesearch import search
search("Google", num_results=100, unique=True)

In addition, you can change the language google searches in. For example, to get results in French run the following program:

from googlesearch import search
search("Google", lang="fr")

You can also specify the region (Country Codes) for your search results. For example, to get results specifically from the US run the following program:

from googlesearch import search
search("Google", region="us")

If you want to turn off the safe search function (this function is on by default), you can do this:

from googlesearch import search
search("Google", safe=None)

To extract more information, such as the description or the result URL, use an advanced search:

from googlesearch import search
search("Google", advanced=True)
# Returns a list of SearchResult
# Properties:
# - title
# - url
# - description

If requesting more than 100 results, googlesearch will send multiple requests to go through the pages. To increase the time between these requests, use sleep_interval:

from googlesearch import search
search("Google", sleep_interval=5, num_results=200)
If requesting more than 10 results, but want to manage the batching yourself? 
Use `start_num` to specify the start number of the results you want to get:
```python
from googlesearch import search
search("Google", sleep_interval=5, num_results=200, start_result=10)

If you are using a HTTP Rotating Proxy which requires you to install their CA Certificate, you can simply add ssl_verify=False in the search() method to avoid SSL Verification.

from googlesearch import search

proxy = 'http://API:@proxy.host.com:8080/'

j = search("proxy test", num_results=100, lang="en", proxy=proxy, ssl_verify=False)
for i in j:
    print(i)

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

googlesearch_python-1.3.0.tar.gz (5.2 kB view details)

Uploaded Source

Built Distribution

googlesearch_python-1.3.0-py3-none-any.whl (5.6 kB view details)

Uploaded Python 3

File details

Details for the file googlesearch_python-1.3.0.tar.gz.

File metadata

  • Download URL: googlesearch_python-1.3.0.tar.gz
  • Upload date:
  • Size: 5.2 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.0.0 CPython/3.12.8

File hashes

Hashes for googlesearch_python-1.3.0.tar.gz
Algorithm Hash digest
SHA256 c5729b1247c2a8f5c4b48ed73c4f8e9fd558ac4e09de67865479f0a33f2d97dc
MD5 aa7a1d9a7c7380865507b3fe38f3b31d
BLAKE2b-256 00c83f76213025b77de23f11d3f87349ff9825cf3b0054f62156858af1bd94f3

See more details on using hashes here.

File details

Details for the file googlesearch_python-1.3.0-py3-none-any.whl.

File metadata

File hashes

Hashes for googlesearch_python-1.3.0-py3-none-any.whl
Algorithm Hash digest
SHA256 808c4dd390dc4c6a1cfba2f5151f5ef16dceb0a200d9770b388dcd39162b4e19
MD5 0442f95723d267a4e8111a85175add7c
BLAKE2b-256 60a6c1fe6a46a7ac2d3b08acfe88ce3d2b12cd8351c697ee4b300bfa350b7c9a

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page