coffeepapers

Just run query.py with whatever titles of papers in quotes or using underscores for spaces (you only need enough of the title to uniquely identify it). That script will fetch those papers from arxiv and build a new papers.js, and both old_papers files. Those files are used to build the website located at jvshields.github.io/coffeepapers/

In order to correctly build the website, the workflow should be:

  1. Pull the current files (making sure you have the up to date old_papers files). This should be as simple as a git pull.
  2. Run query.py with the paper names.
  3. Push your updated files back to github. (git add papers.js old_papers.js old_papers.json, git commit -m 'papers', git push)

Extra info for querying arxiv: There are several methods of running the query script, specifying arxiv ID or other paper information instead of titles. By default the script will assume you are choosing papers for the upcoming coffee date, so under most circumstances a date does not need to be specified. Running the script in test mode with the argument -t will show you the papers query.py grabs without overwriting the current and archival .js files.

For more info run $ ./query.py --help