Collect a citation graph from Google Scholar
étudier is a small Python program that uses Selenium and requests-html to drive a non-headless browser to collect a citation graph around a particular Google Scholar citation or set of search results. The resulting network is written out as a Gephi file and a D3 visualization using networkx. The D3 visualization could use some work, so if you add style to it please submit a pull request.
If you are wondering why it uses a non-headless browser it's because Google is quite protective of this data and routinely will ask you to solve a captcha (identifying street signs, cars, etc in photos). étudier will allow you to complete these tasks when they occur and then will continue on its way collecting data.
You'll need to install ChromeDriver before doing anything else. If you use Homebrew on OS X this is as easy as:
brew install chromedriver
Then you'll want to install Python 3 and:
pip3 install etudier
To use it you first need to navigate to a page on Google Scholar that you are interested in, for example here is the page of citations that reference Sherry Ortner's Theory in Anthropology since the Sixties. Then you start etudier up pointed at that page.
% etudier 'https://scholar.google.com/scholar?start=0&hl=en&as_sdt=20000005&sciodt=0,21&cites=17950649785549691519&scipsc='
If you are interested in starting with keyword search results in Google Scholar you can do that too. For example here is the url for searching for "cscw memory" if I was interested in papers that talk about the CSCW conference and memory:
% etudier 'https://scholar.google.com/scholar?hl=en&as_sdt=0%2C21&q=cscw+memory&btnG='
Note: it's important to quote the URL so that the shell doesn't interpret the ampersands as an attempt to background the process.
By default étudier will collect the 10 citations on that page and then look at the top 10 citatations that reference each one. So you will end up with no more than 100 citations being collected (10 on each page * 10 citations).
If you would like to get more than one page of results use the
example this would result in no more than 400 (20 * 20) results being collected:
% etudier --pages 2 'https://scholar.google.com/scholar?start=0&hl=en&as_sdt=20000005&sciodt=0,21&cites=17950649785549691519&scipsc='
And finally if you would like to look at the citations of the citations you the --depth parameter.
% etudier --depth 2 'https://scholar.google.com/scholar?start=0&hl=en&as_sdt=20000005&sciodt=0,21&cites=17950649785549691519&scipsc='
This will collect the initial set of 10 citations, the top 10 citations for each, and then the top 10 citations of each, so no more than 1000 citations 1000 citations (10 * 10 * 10). It's no more because there is certain to be some duplication of publications in the citations of each.
By default a file called
output.gexf will be written, but you can change this
--output option. The output file will contain rudimentary metadata
collected from Google Scholar including:
- id - the cluster identifier assigned by Google
- url - the url for the publication
- title - the title of the publication
- authors - a comma separated list of the publication authors
- year - the year of publication
- cited-by - the number of other publications that cite the publication
- cited-by-url - a Google Scholar URL for the list of citing publications