This project uses a home server cronjob to scrape Google Scholar data via the scholarly Python package and the wikipedia Python package to get the journal's impact factor (IF) and the publication's DOI.
The JSON can then be used, for example, by uploading the data to a publicly accessible server via Secure Copy (SCP) or rsync, which serves the JSON data via a Flask application.
- Python 3.6+
- Flask
- scholarly
- Wikipedia
- Clone this repository to your local machine.
- Install the required packages.
- Setup.
git clone https://github.com/Luen/scholarly-api
python -m venv scholar
source scholar/bin/activate
pip install -r requirements.txt
playwright install
- Test run.
python main.py ynWS968AAAAJ
- Set up cronjob.
0 * * * * /path/to/your_bash_script.sh
Install pytest and run it using the command pytest
.
Navigate to the project directory and run the Flask application:
python ./serve.py
URL: / Method: GET Description: Displays a welcome message in plain text. Example: /
URL: /author_id Method: GET Description: Searches for authors by id. Parameter: id Example: /ynWS968AAAAJ
This open-sourced project is released under the Unlicense.
Alternative wikipedia packages: wikipedia-api and pymediawiki