# SPDX-License-Identifier: AGPL-3.0-or-later """CORE (science) """ # pylint: disable=missing-function-docstring from json import loads from datetime import datetime from urllib.parse import urlencode from searx import logger from searx.exceptions import SearxEngineAPIException logger = logger.getChild('CORE engine') about = { "website": 'https://core.ac.uk', "wikidata_id": None, "official_api_documentation": 'https://core.ac.uk/documentation/api/', "use_official_api": True, "require_api_key": True, "results": 'JSON', } categories = ['science'] paging = True nb_per_page = 10 api_key = 'unset' logger = logger.getChild('CORE engine') base_url = 'https://core.ac.uk:443/api-v2/search/' search_string = '{query}?page={page}&pageSize={nb_per_page}&apiKey={apikey}' def request(query, params): if api_key == 'unset': raise SearxEngineAPIException('missing CORE API key') search_path = search_string.format( query = urlencode({'q': query}), nb_per_page = nb_per_page, page = params['pageno'], apikey = api_key, ) params['url'] = base_url + search_path logger.debug("query_url --> %s", params['url']) return params def response(resp): results = [] json_data = loads(resp.text) for result in json_data['data']: source = result['_source'] time = source['publishedDate'] or source['depositedDate'] if time : date = datetime.fromtimestamp(time / 1000) else: date = None metadata = [] if source['publisher'] and len(source['publisher']) > 3: metadata.append(source['publisher']) if source['topics']: metadata.append(source['topics'][0]) if source['doi']: metadata.append(source['doi']) metadata = ' / '.join(metadata) results.append({ 'url': source['urls'][0].replace('http://', 'https://', 1), 'title': source['title'], 'content': source['description'], 'publishedDate': date, 'metadata' : metadata, }) return results