mirror of
https://github.com/searxng/searxng.git
synced 2024-11-30 06:31:04 +00:00
[feat] engine: implementation of Anna's Archive
Anna's Archive [1] is a free non-profit online shadow library metasearch engine providing access to a variety of book resources (also via IPFS), created by a team of anonymous archivists [2]. [1] https://annas-archive.org/ [2] https://annas-software.org/AnnaArchivist/annas-archive
This commit is contained in:
parent
fd26f37073
commit
e5637fe7b9
2 changed files with 70 additions and 0 deletions
63
searx/engines/annas-archive.py
Normal file
63
searx/engines/annas-archive.py
Normal file
|
@ -0,0 +1,63 @@
|
||||||
|
# SPDX-License-Identifier: AGPL-3.0-or-later
|
||||||
|
# lint: pylint
|
||||||
|
"""Anna's Archive
|
||||||
|
|
||||||
|
"""
|
||||||
|
from typing import List, Dict, Any, Optional
|
||||||
|
from urllib.parse import quote
|
||||||
|
from lxml import html
|
||||||
|
|
||||||
|
from searx.utils import extract_text, eval_xpath
|
||||||
|
|
||||||
|
# about
|
||||||
|
about: Dict[str, Any] = {
|
||||||
|
"website": "https://annas-archive.org/",
|
||||||
|
"wikidata_id": "Q115288326",
|
||||||
|
"official_api_documentation": None,
|
||||||
|
"use_official_api": False,
|
||||||
|
"require_api_key": False,
|
||||||
|
"results": "HTML",
|
||||||
|
}
|
||||||
|
|
||||||
|
# engine dependent config
|
||||||
|
categories: List[str] = ["files"]
|
||||||
|
paging: bool = False
|
||||||
|
|
||||||
|
# search-url
|
||||||
|
base_url: str = "https://annas-archive.org"
|
||||||
|
|
||||||
|
# xpath queries
|
||||||
|
xpath_results: str = '//main//a[starts-with(@href,"/md5")]'
|
||||||
|
xpath_url: str = ".//@href"
|
||||||
|
xpath_title: str = ".//h3/text()[1]"
|
||||||
|
xpath_authors: str = './/div[contains(@class, "italic")]'
|
||||||
|
xpath_publisher: str = './/div[contains(@class, "text-sm")]'
|
||||||
|
xpath_file_info: str = './/div[contains(@class, "text-xs")]'
|
||||||
|
|
||||||
|
|
||||||
|
def request(query, params: Dict[str, Any]) -> Dict[str, Any]:
|
||||||
|
search_url: str = base_url + "/search?q={search_query}"
|
||||||
|
params["url"] = search_url.format(search_query=quote(query))
|
||||||
|
return params
|
||||||
|
|
||||||
|
|
||||||
|
def response(resp) -> List[Dict[str, Optional[str]]]:
|
||||||
|
results: List[Dict[str, Optional[str]]] = []
|
||||||
|
dom = html.fromstring(resp.text)
|
||||||
|
|
||||||
|
for item in dom.xpath(xpath_results):
|
||||||
|
result: Dict[str, Optional[str]] = {}
|
||||||
|
|
||||||
|
result["url"] = base_url + item.xpath(xpath_url)[0]
|
||||||
|
|
||||||
|
result["title"] = extract_text(eval_xpath(item, xpath_title))
|
||||||
|
|
||||||
|
result["content"] = "{publisher}. {authors}. {file_info}".format(
|
||||||
|
authors=extract_text(eval_xpath(item, xpath_authors)),
|
||||||
|
publisher=extract_text(eval_xpath(item, xpath_publisher)),
|
||||||
|
file_info=extract_text(eval_xpath(item, xpath_file_info)),
|
||||||
|
)
|
||||||
|
|
||||||
|
results.append(result)
|
||||||
|
|
||||||
|
return results
|
|
@ -297,6 +297,13 @@ engines:
|
||||||
shortcut: 9g
|
shortcut: 9g
|
||||||
disabled: true
|
disabled: true
|
||||||
|
|
||||||
|
- name: anna's archive
|
||||||
|
engine: annas-archive
|
||||||
|
paging: False
|
||||||
|
categories: files
|
||||||
|
disabled: true
|
||||||
|
shortcut: aa
|
||||||
|
|
||||||
- name: apk mirror
|
- name: apk mirror
|
||||||
engine: apkmirror
|
engine: apkmirror
|
||||||
timeout: 4.0
|
timeout: 4.0
|
||||||
|
|
Loading…
Reference in a new issue