1
0
mirror of https://github.com/searxng/searxng.git synced 2024-11-09 14:40:11 +01:00
searxng/searx/engines/etools.py

59 lines
1.3 KiB
Python
Raw Normal View History

# SPDX-License-Identifier: AGPL-3.0-or-later
2020-02-12 23:58:50 +01:00
"""
eTools (Web)
"""
from lxml import html
from urllib.parse import quote
from searx.utils import extract_text, eval_xpath
2020-02-12 23:58:50 +01:00
# about
about = {
"website": 'https://www.etools.ch',
"wikidata_id": None,
"official_api_documentation": None,
"use_official_api": False,
"require_api_key": False,
"results": 'HTML',
}
2020-02-12 23:58:50 +01:00
categories = ['general']
paging = False
safesearch = True
base_url = 'https://www.etools.ch'
search_path = (
# fmt: off
'/searchAdvancedSubmit.do'
'?query={search_term}'
'&pageResults=20'
'&safeSearch={safesearch}'
# fmt: on
)
2020-02-12 23:58:50 +01:00
def request(query, params):
if params['safesearch']:
safesearch = 'true'
else:
safesearch = 'false'
params['url'] = base_url + search_path.format(search_term=quote(query), safesearch=safesearch)
return params
def response(resp):
results = []
dom = html.fromstring(resp.text)
for result in eval_xpath(dom, '//table[@class="result"]//td[@class="record"]'):
url = eval_xpath(result, './a/@href')[0]
title = extract_text(eval_xpath(result, './a//text()'))
content = extract_text(eval_xpath(result, './/div[@class="text"]//text()'))
results.append({'url': url, 'title': title, 'content': content})
2020-02-12 23:58:50 +01:00
return results