mirror of
https://github.com/bookwyrm-social/bookwyrm.git
synced 2024-12-30 20:10:36 +00:00
233 lines
7.9 KiB
Python
233 lines
7.9 KiB
Python
''' openlibrary data connector '''
|
|
from django.core.exceptions import ObjectDoesNotExist
|
|
from django.core.files.base import ContentFile
|
|
import re
|
|
import requests
|
|
|
|
from fedireads import models
|
|
from .abstract_connector import AbstractConnector, SearchResult
|
|
from .abstract_connector import update_from_mappings, get_date
|
|
from .openlibrary_languages import languages
|
|
|
|
|
|
class Connector(AbstractConnector):
|
|
''' instantiate a connector for OL '''
|
|
def __init__(self, identifier):
|
|
super().__init__(identifier)
|
|
|
|
|
|
def search(self, query):
|
|
''' query openlibrary search '''
|
|
resp = requests.get(
|
|
'%s%s' % (self.search_url, query),
|
|
headers={
|
|
'Accept': 'application/json; charset=utf-8',
|
|
},
|
|
)
|
|
if not resp.ok:
|
|
resp.raise_for_status()
|
|
data = resp.json()
|
|
results = []
|
|
|
|
for doc in data['docs'][:5]:
|
|
key = doc['key']
|
|
key = key.split('/')[-1]
|
|
author = doc.get('author_name') or ['Unknown']
|
|
results.append(SearchResult(
|
|
doc.get('title'),
|
|
key,
|
|
author[0],
|
|
doc.get('first_publish_year'),
|
|
doc
|
|
))
|
|
return results
|
|
|
|
|
|
def get_or_create_book(self, olkey):
|
|
''' pull up a book record by whatever means possible.
|
|
if you give a work key, it should give you the default edition,
|
|
annotated with work data. '''
|
|
|
|
try:
|
|
book = models.Book.objects.select_subclasses().get(
|
|
openlibrary_key=olkey
|
|
)
|
|
return book
|
|
except ObjectDoesNotExist:
|
|
pass
|
|
# no book was found, so we start creating a new one
|
|
model = models.Edition
|
|
if re.match(r'^OL\d+W$', olkey):
|
|
model = models.Work
|
|
book = model(openlibrary_key=olkey)
|
|
return self.update_book(book)
|
|
|
|
|
|
def update_book(self, book):
|
|
''' query openlibrary for data on a book '''
|
|
olkey = book.openlibrary_key
|
|
# load the book json from openlibrary.org
|
|
response = requests.get('%s/works/%s.json' % (self.url, olkey))
|
|
if not response.ok:
|
|
response.raise_for_status()
|
|
data = response.json()
|
|
if not book.source_url:
|
|
book.source_url = response.url
|
|
return self.update_from_data(book, data)
|
|
|
|
|
|
def update_from_data(self, book, data):
|
|
''' update a book from a json blob '''
|
|
mappings = {
|
|
'publish_date': ('published_date', get_date),
|
|
'first_publish_date': ('first_published_date', get_date),
|
|
'description': ('description', get_description),
|
|
'isbn_13': ('isbn', lambda a: a[0]),
|
|
'oclc_numbers': ('oclc_number', lambda a: a[0]),
|
|
'lccn': ('lccn', lambda a: a[0]),
|
|
'languages': ('languages', get_languages),
|
|
'number_of_pages': ('pages', None),
|
|
'series': ('series', lambda a: a[0]),
|
|
}
|
|
book = update_from_mappings(book, data, mappings)
|
|
|
|
if 'identifiers' in data:
|
|
if 'goodreads' in data['identifiers']:
|
|
book.goodreads_key = data['identifiers']['goodreads'][0]
|
|
if 'series' in data and len(data['series']) > 1:
|
|
book.series_number = data['series'][1]
|
|
|
|
if not book.connector:
|
|
book.connector = self.connector
|
|
book.save()
|
|
|
|
# this book sure as heck better be an edition
|
|
if data.get('works'):
|
|
key = data.get('works')[0]['key']
|
|
key = key.split('/')[-1]
|
|
work = self.get_or_create_book(key)
|
|
book.parent_work = work
|
|
|
|
if isinstance(book, models.Work):
|
|
# load editions of a work
|
|
self.get_editions_of_work(book)
|
|
|
|
# we also need to know the author get the cover
|
|
for author_blob in data.get('authors', []):
|
|
# this id is "/authors/OL1234567A" and we want just "OL1234567A"
|
|
author_blob = author_blob.get('author', author_blob)
|
|
author_id = author_blob['key']
|
|
author_id = author_id.split('/')[-1]
|
|
book.authors.add(self.get_or_create_author(author_id))
|
|
if not data.get('authors'):
|
|
book.authors.set(book.parent_work.authors.all())
|
|
|
|
|
|
if book.sync_cover and data.get('covers') and len(data['covers']):
|
|
book.cover.save(*self.get_cover(data['covers'][0]), save=True)
|
|
|
|
return book
|
|
|
|
|
|
def expand_book_data(self, book):
|
|
work = book
|
|
if isinstance(book, models.Edition):
|
|
work = book.parent_work
|
|
self.get_editions_of_work(work, default_only=False)
|
|
|
|
|
|
def get_editions_of_work(self, work, default_only=True):
|
|
''' get all editions of a work '''
|
|
response = requests.get(
|
|
'%s/works/%s/editions.json' % (self.url, work.openlibrary_key))
|
|
edition_data = response.json()
|
|
|
|
options = edition_data.get('entries', [])
|
|
if default_only and len(options) > 1:
|
|
options = [e for e in options if e.get('cover')] or options
|
|
options = [e for e in options if \
|
|
'/languages/eng' in str(e.get('languages'))] or options
|
|
formats = ['paperback', 'hardcover', 'mass market paperback']
|
|
options = [e for e in options if \
|
|
str(e.get('physical_format')).lower() in formats] or options
|
|
options = [e for e in options if e.get('isbn_13')] or options
|
|
options = [e for e in options if e.get('ocaid')] or options
|
|
|
|
if not options:
|
|
options = edition_data.get('entries', [])
|
|
options = options[:1]
|
|
|
|
for data in options:
|
|
try:
|
|
olkey = data['key'].split('/')[-1]
|
|
except KeyError:
|
|
# bad data I guess?
|
|
return
|
|
|
|
try:
|
|
models.Edition.objects.get(openlibrary_key=olkey)
|
|
except ObjectDoesNotExist:
|
|
book = models.Edition.objects.create(openlibrary_key=olkey)
|
|
self.update_from_data(book, data)
|
|
|
|
|
|
def get_or_create_author(self, olkey):
|
|
''' load that author '''
|
|
if not re.match(r'^OL\d+A$', olkey):
|
|
raise ValueError('Invalid OpenLibrary author ID')
|
|
try:
|
|
return models.Author.objects.get(openlibrary_key=olkey)
|
|
except ObjectDoesNotExist:
|
|
pass
|
|
|
|
response = requests.get('%s/authors/%s.json' % (self.url, olkey))
|
|
if not response.ok:
|
|
response.raise_for_status()
|
|
|
|
data = response.json()
|
|
author = models.Author(openlibrary_key=olkey)
|
|
mappings = {
|
|
'birth_date': ('born', get_date),
|
|
'death_date': ('died', get_date),
|
|
'bio': ('bio', get_description),
|
|
}
|
|
author = update_from_mappings(author, data, mappings)
|
|
# TODO this is making some BOLD assumption
|
|
name = data.get('name')
|
|
if name:
|
|
author.last_name = name.split(' ')[-1]
|
|
author.first_name = ' '.join(name.split(' ')[:-1])
|
|
author.save()
|
|
|
|
return author
|
|
|
|
|
|
def get_cover(self, cover_id):
|
|
''' ask openlibrary for the cover '''
|
|
# TODO: get medium and small versions
|
|
image_name = '%s-M.jpg' % cover_id
|
|
url = '%s/b/id/%s' % (self.covers_url, image_name)
|
|
response = requests.get(url)
|
|
if not response.ok:
|
|
response.raise_for_status()
|
|
image_content = ContentFile(response.content)
|
|
return [image_name, image_content]
|
|
|
|
|
|
def get_description(description_blob):
|
|
''' descriptions can be a string or a dict '''
|
|
if isinstance(description_blob, dict):
|
|
return description_blob.get('value')
|
|
return description_blob
|
|
|
|
|
|
def get_languages(language_blob):
|
|
''' /language/eng -> English '''
|
|
langs = []
|
|
for lang in language_blob:
|
|
langs.append(
|
|
languages.get(lang.get('key', ''), None)
|
|
)
|
|
return langs
|
|
|
|
|