Forgot to add arxiv2bib + Now using argparse

This commit is contained in:
Phyks 2014-05-03 02:05:46 +02:00
parent 1adb1ce947
commit e53a52df2b
2 changed files with 487 additions and 50 deletions

412
arxiv2bib.py Executable file
View File

@ -0,0 +1,412 @@
#! /usr/bin/env python
#
# Copyright (c) 2012, Nathan Grigg
# All rights reserved.
#
# Redistribution and use in source and binary forms, with or without
# modification, are permitted provided that the following conditions are met:
#
# * Redistributions of source code must retain the above copyright
# notice, this list of conditions and the following disclaimer.
# * Redistributions in binary form must reproduce the above copyright
# notice, this list of conditions and the following disclaimer in the
# documentation and/or other materials provided with the distribution.
# * Neither the name of this package nor the
# names of its contributors may be used to endorse or promote products
# derived from this software without specific prior written permission.
#
# This software is provided by the copyright holders and contributors "as
# is" and any express or implied warranties, including, but not limited
# to, the implied warranties of merchantability and fitness for a
# particular purpose are disclaimed. In no event shall Nathan Grigg be
# liable for any direct, indirect, incidental, special, exemplary, or
# consequential damages (including, but not limited to, procurement of
# substitute goods or services; loss of use, data, or profits; or business
# interruption) however caused and on any theory of liability, whether in
# contract, strict liability, or tort (including negligence or otherwise)
# arising in any way out of the use of this software, even if advised of
# the possibility of such damage.
#
# (also known as the New BSD License)
#
# Indiscriminate automated downloads from arXiv.org are not permitted.
# For more information, see http://arxiv.org/help/robots
#
# This script usually makes only one call to arxiv.org per run.
# No caching of any kind is performed.
from __future__ import print_function
from xml.etree import ElementTree
import sys
import re
import os
if sys.version_info < (2, 6):
raise Exception("Python 2.6 or higher required")
# Python 2 compatibility code
PY2 = sys.version_info[0] == 2
if not PY2:
from urllib.parse import urlencode
from urllib.request import urlopen
from urllib.error import HTTPError
print_bytes = lambda s: sys.stdout.buffer.write(s)
else:
from urllib import urlencode
from urllib2 import HTTPError, urlopen
print_bytes = lambda s: sys.stdout.write(s)
# Namespaces
ATOM = '{http://www.w3.org/2005/Atom}'
ARXIV = '{http://arxiv.org/schemas/atom}'
# regular expressions to check if arxiv id is valid
NEW_STYLE = re.compile(r'^\d{4}\.\d{4}(v\d+)?$')
OLD_STYLE = re.compile(r"""(?x)
^(
math-ph
|hep-ph
|nucl-ex
|nucl-th
|gr-qc
|astro-ph
|hep-lat
|quant-ph
|hep-ex
|hep-th
|stat
(\.(AP|CO|ML|ME|TH))?
|q-bio
(\.(BM|CB|GN|MN|NC|OT|PE|QM|SC|TO))?
|cond-mat
(\.(dis-nn|mes-hall|mtrl-sci|other|soft|stat-mech|str-el|supr-con))?
|cs
(\.(AR|AI|CL|CC|CE|CG|GT|CV|CY|CR|DS|DB|DL|DM|DC|GL|GR|HC|IR|IT|LG|LO|
MS|MA|MM|NI|NE|NA|OS|OH|PF|PL|RO|SE|SD|SC))?
|nlin
(\.(AO|CG|CD|SI|PS))?
|physics
(\.(acc-ph|ao-ph|atom-ph|atm-clus|bio-ph|chem-ph|class-ph|comp-ph|
data-an|flu-dyn|gen-ph|geo-ph|hist-ph|ins-det|med-ph|optics|ed-ph|
soc-ph|plasm-ph|pop-ph|space-ph))?
|math
(\.(AG|AT|AP|CT|CA|CO|AC|CV|DG|DS|FA|GM|GN|GT|GR|HO|IT|KT|LO|MP|MG
|NT|NA|OA|OC|PR|QA|RT|RA|SP|ST|SG))?
)/\d{7}(v\d+)?$""")
def is_valid(arxiv_id):
"""Checks if id resembles a valid arxiv identifier."""
return bool(NEW_STYLE.match(arxiv_id)) or bool(OLD_STYLE.match(arxiv_id))
class FatalError(Exception):
"""Error that prevents us from continuing"""
class NotFoundError(Exception):
"""Reference not found by the arxiv API"""
class Reference(object):
"""Represents a single reference.
Instantiate using Reference(entry_xml). Note entry_xml should be
an ElementTree.Element object.
"""
def __init__(self, entry_xml):
self.xml = entry_xml
self.url = self._field_text('id')
self.id = self._id()
self.authors = self._authors()
self.title = self._field_text('title')
if len(self.id) == 0 or len(self.authors) == 0 or len(self.title) == 0:
raise NotFoundError("No such publication", self.id)
self.summary = self._field_text('summary')
self.category = self._category()
self.year, self.month = self._published()
self.updated = self._field_text('updated')
self.bare_id = self.id[:self.id.rfind('v')]
self.note = self._field_text('journal_ref', namespace=ARXIV)
def _authors(self):
"""Extracts author names from xml."""
xml_list = self.xml.findall(ATOM + 'author/' + ATOM + 'name')
return [field.text for field in xml_list]
def _field_text(self, id, namespace=ATOM):
"""Extracts text from arbitrary xml field"""
try:
return self.xml.find(namespace + id).text.strip()
except:
return ""
def _category(self):
"""Get category"""
try:
return self.xml.find(ARXIV + 'primary_category').attrib['term']
except:
return ""
def _id(self):
"""Get arxiv id"""
try:
id_url = self._field_text('id')
return id_url[id_url.find('/abs/') + 5:]
except:
return ""
def _published(self):
"""Get published date"""
published = self._field_text('published')
if len(published) < 7:
return "", ""
y, m = published[:4], published[5:7]
try:
m = ["Jan", "Feb", "Mar", "Apr", "May", "Jun", "Jul",
"Aug", "Sep", "Nov", "Dec"][int(m) - 1]
except:
pass
return y, m
def bibtex(self):
"""BibTex string of the reference."""
lines = ["@article{" + self.id]
for k, v in [("Author", " and ".join(self.authors)),
("Title", self.title),
("Eprint", self.id),
("ArchivePrefix", "arXiv"),
("PrimaryClass", self.category),
("Abstract", self.summary),
("Year", self.year),
("Month", self.month),
("Note", self.note),
("Url", self.url),
("File", self.id + ".pdf"),
]:
if len(v):
lines.append("%-13s = {%s}" % (k, v))
return ("," + os.linesep).join(lines) + os.linesep + "}"
class ReferenceErrorInfo(object):
"""Contains information about a reference error"""
def __init__(self, message, id):
self.message = message
self.id = id
self.bare_id = id[:id.rfind('v')]
# mark it as really old, so it gets superseded if possible
self.updated = '0'
def bibtex(self):
"""BibTeX comment explaining error"""
return "@comment{%(id)s: %(message)s}" % \
{'id': self.id, 'message': self.message}
def __str__(self):
return "Error: %(message)s (%(id)s)" % \
{'id': self.id, 'message': self.message}
def arxiv2bib(id_list):
"""Returns a list of references, corresponding to elts of id_list"""
d = arxiv2bib_dict(id_list)
l = []
for id in id_list:
try:
l.append(d[id])
except:
l.append(ReferenceErrorInfo("Not found", id))
return l
def arxiv_request(ids):
"""Sends a request to the arxiv API."""
q = urlencode([
("id_list", ",".join(ids)),
("max_results", len(ids))
])
xml = urlopen("http://export.arxiv.org/api/query?" + q)
# xml.read() returns bytes, but ElementTree.fromstring decodes
# to unicode when needed (python2) or string (python3)
return ElementTree.fromstring(xml.read())
def arxiv2bib_dict(id_list):
"""Fetches citations for ids in id_list into a dictionary indexed by id"""
ids = []
d = {}
# validate ids
for id in id_list:
if is_valid(id):
ids.append(id)
else:
d[id] = ReferenceErrorInfo("Invalid arXiv identifier", id)
if len(ids) == 0:
return d
# make the api call
while True:
xml = arxiv_request(ids)
# check for error
entries = xml.findall(ATOM + "entry")
try:
first_title = entries[0].find(ATOM + "title")
except:
raise FatalError("Unable to connect to arXiv.org API.")
if first_title is None or first_title.text.strip() != "Error":
break
try:
id = entries[0].find(ATOM + "summary").text.split()[-1]
del(ids[ids.index(id)])
except:
raise FatalError("Unable to parse an error returned by arXiv.org.")
# Parse each reference and store it in dictionary
for entry in entries:
try:
ref = Reference(entry)
except NotFoundError as error:
message, id = error.args
ref = ReferenceErrorInfo(message, id)
if ref.id:
d[ref.id] = ref
if ref.bare_id:
if not (ref.bare_id in d) or d[ref.bare_id].updated < ref.updated:
d[ref.bare_id] = ref
return d
class Cli(object):
"""Command line interface"""
def __init__(self, args=None):
"""Parse arguments"""
self.args = self.parse_args(args)
if len(self.args.id) == 0:
self.args.id = [line.strip() for line in sys.stdin]
# avoid duplicate error messages unless verbose is set
if self.args.comments and not self.args.verbose:
self.args.quiet = True
self.output = []
self.messages = []
self.error_count = 0
self.code = 0
def run(self):
"""Produce output and error messages"""
try:
bib = arxiv2bib(self.args.id)
except HTTPError as error:
if error.getcode() == 403:
raise FatalError("""\
403 Forbidden error. This usually happens when you make many
rapid fire requests in a row. If you continue to do this, arXiv.org may
interpret your requests as a denial of service attack.
For more information, see http://arxiv.org/help/robots.
""")
else:
raise FatalError(
"HTTP Connection Error: {0}".format(error.getcode()))
self.create_output(bib)
self.code = self.tally_errors(bib)
def create_output(self, bib):
"""Format the output and error messages"""
for b in bib:
if isinstance(b, ReferenceErrorInfo):
self.error_count += 1
if self.args.comments:
self.output.append(b.bibtex())
if not self.args.quiet:
self.messages.append(str(b))
else:
self.output.append(b.bibtex())
def print_output(self):
if not self.output:
return
output_string = os.linesep.join(self.output)
try:
print(output_string)
except UnicodeEncodeError:
print_bytes((output_string + os.linesep).encode('utf-8'))
if self.args.verbose:
self.messages.append(
'Could not use system encoding; using utf-8')
def tally_errors(self, bib):
"""calculate error code"""
if self.error_count == len(self.args.id):
self.messages.append("No successful matches")
return 2
elif self.error_count > 0:
self.messages.append("%s of %s matched succesfully" %
(len(bib) - self.error_count, len(bib)))
return 1
else:
return 0
def print_messages(self):
"""print messages to stderr"""
if self.messages:
self.messages.append("")
sys.stderr.write(os.linesep.join(self.messages))
@staticmethod
def parse_args(args):
try:
import argparse
except:
sys.exit("Cannot load required module 'argparse'")
parser = argparse.ArgumentParser(
description="Get the BibTeX for each arXiv id.",
epilog="""\
Returns 0 on success, 1 on partial failure, 2 on total failure.
Valid BibTeX is written to stdout, error messages to stderr.
If no arguments are given, ids are read from stdin, one per line.""",
formatter_class=argparse.RawDescriptionHelpFormatter)
parser.add_argument('id', metavar='arxiv_id', nargs="*",
help="arxiv identifier, such as 1201.1213")
parser.add_argument('-c', '--comments', action='store_true',
help="Include @comment fields with error details")
parser.add_argument('-q', '--quiet', action='store_true',
help="Display fewer error messages")
parser.add_argument('-v', '--verbose', action="store_true",
help="Display more error messages")
return parser.parse_args(args)
def main(args=None):
"""Run the command line interface"""
cli = Cli(args)
try:
cli.run()
except FatalError as err:
sys.stderr.write(err.args[0] + os.linesep)
return 2
cli.print_output()
cli.print_messages()
return cli.code
if __name__ == "__main__":
sys.exit(main())

113
main.py
View File

@ -1,6 +1,7 @@
#!/usr/bin/env python2
# -*- coding: utf8 -*-
import argparse
import os
import shutil
import subprocess
@ -247,72 +248,96 @@ def resync():
if __name__ == '__main__':
parser = argparse.ArgumentParser(description="A bibliography " +
"management tool.")
subparsers = parser.add_subparsers(help="sub-command help")
parser_download = subparsers.add_parser('download', help="download help")
parser_download.add_argument('-t', '--type', default=None,
choices=['article', 'book'],
help="Type of the file to download")
parser_download.add_argument('url', nargs='+',
help="url of the file to import")
parser_download.set_defaults(func='download')
parser_import = subparsers.add_parser('import', help="import help")
parser_import.add_argument('-t', '--type', default=None,
choices=['article', 'book'],
help="Type of the file to import")
parser_import.add_argument('file', nargs='+',
help="path to the file to import")
parser_import.set_defaults(func='import')
parser_delete = subparsers.add_parser('delete', help="delete help")
parser_delete.add_argument('files', metavar='entry', nargs='+',
help="a filename or an identifier")
parser_delete.set_defaults(func='delete')
parser_list = subparsers.add_parser('list', help="list help")
parser_list.set_defaults(func='list')
# TODO
parser_search = subparsers.add_parser('search', help="search help")
parser_search.set_defaults(func='search')
# TODO
parser_open = subparsers.add_parser('open', help="open help")
parser_open.add_argument('ids', metavar='id', nargs='+',
help="an identifier")
parser_open.set_defaults(func='open')
parser_resync = subparsers.add_parser('resync', help="resync help")
parser_resync.set_defaults(func='resync')
args = parser.parse_args()
try:
if len(sys.argv) < 2:
sys.exit("Usage: TODO")
if sys.argv[1] == 'download':
if len(sys.argv) < 3:
sys.exit("Usage: " + sys.argv[0] +
" download FILE [article|book]")
filetype = None
if len(sys.argv) > 3 and sys.argv[3] in ["article", "book"]:
filetype = sys.argv[3].lower()
new_name = downloadFile(sys.argv[2], filetype)
if args.func == 'download':
for url in args.url:
new_name = downloadFile(url, args.t)
if new_name is not False:
print(sys.argv[2]+" successfully imported as "+new_name)
print(url+" successfully imported as "+new_name)
else:
tools.warning("An error occurred while downloading "+url)
sys.exit()
if sys.argv[1] == 'import':
if len(sys.argv) < 3:
sys.exit("Usage: " + sys.argv[0] +
" import FILE [article|book]")
filetype = None
if len(sys.argv) > 3 and sys.argv[3] in ["article", "book"]:
filetype = sys.argv[3].lower()
new_name = addFile(sys.argv[2], filetype)
if args.func == 'import':
for filename in args.file:
new_name = addFile(filename, args.t)
if new_name is not False:
print(sys.argv[2]+" successfully imported as "+new_name+".")
print(sys.argv[2]+" successfully imported as " +
new_name+".")
else:
tools.warning("An error occurred while importing " +
filename)
sys.exit()
elif sys.argv[1] == 'delete':
if len(sys.argv) < 3:
sys.exit("Usage: " + sys.argv[0] + " delete FILE|ID")
elif args.func == 'delete':
for filename in args.file:
confirm = tools.rawInput("Are you sure you want to delete " +
sys.argv[2]+"? [y/N] ")
filename+"? [y/N] ")
if confirm.lower() == 'y':
if not backend.deleteId(sys.argv[2]):
if not backend.deleteFile(sys.argv[2]):
tools.warning("Unable to delete "+sys.argv[2])
if not backend.deleteId(filename):
if not backend.deleteFile(filename):
tools.warning("Unable to delete "+filename)
sys.exit(1)
print(sys.argv[2]+" successfully deleted.")
print(filename+" successfully deleted.")
sys.exit()
elif sys.argv[1] == 'list':
elif args.func == 'list':
raise Exception('TODO')
elif sys.argv[1] == 'search':
elif args.func == 'search':
raise Exception('TODO')
elif sys.argv[1] == 'open':
if len(sys.argv) < 3:
sys.exit("Usage: " + sys.argv[0] +
" open ID1 ID2 …")
for filename in sys.argv[2:]:
elif args.func == 'open':
for filename in args.ids:
if not openFile(filename):
sys.exit("Unable to open file associated " +
"to ident "+filename)
elif sys.argv[1] == 'resync':
if len(sys.argv) > 2 and sys.argv[2] == 'help':
sys.exit("Usage: " + sys.argv[0] + " resync")
elif args.func == 'resync':
confirm = tools.rawInput("Resync files and bibtex index? [y/N] ")
if confirm.lower() == 'y':
resync()