2013-10-14 23:09:13 +02:00
|
|
|
|
#!/usr/bin/env python
|
2021-05-26 10:16:34 +02:00
|
|
|
|
# SPDX-License-Identifier: AGPL-3.0-or-later
|
|
|
|
|
# lint: pylint
|
2022-01-24 11:44:18 +01:00
|
|
|
|
# pyright: basic
|
2021-05-26 13:32:04 +02:00
|
|
|
|
"""WebbApp
|
2021-03-18 19:59:01 +01:00
|
|
|
|
|
2021-05-26 13:32:04 +02:00
|
|
|
|
"""
|
2015-01-17 21:54:40 +01:00
|
|
|
|
import hashlib
|
2016-10-16 23:40:56 +02:00
|
|
|
|
import hmac
|
|
|
|
|
import json
|
|
|
|
|
import os
|
2021-05-26 10:16:34 +02:00
|
|
|
|
import sys
|
2015-03-10 19:55:22 +01:00
|
|
|
|
|
2014-03-14 09:55:04 +01:00
|
|
|
|
from datetime import datetime, timedelta
|
2021-04-14 17:23:15 +02:00
|
|
|
|
from timeit import default_timer
|
2020-08-06 17:42:46 +02:00
|
|
|
|
from html import escape
|
|
|
|
|
from io import StringIO
|
2022-01-17 07:54:37 +01:00
|
|
|
|
import typing
|
2022-01-17 11:11:39 +01:00
|
|
|
|
from typing import List, Dict, Iterable
|
2021-05-26 10:16:34 +02:00
|
|
|
|
|
[fix] debug log: UnicodeEncodeError: 'ascii' codec can't encode
The issue exists only in the debug log::
--- Logging error ---
Traceback (most recent call last):
File "/usr/lib/python3.9/logging/__init__.py", line 1086, in emit
stream.write(msg + self.terminator)
UnicodeEncodeError: 'ascii' codec can't encode characters in position 79-89: ordinal not in range(128)
Call stack:
File "/usr/local/searx/searx-pyenv/lib/python3.9/site-packages/flask/app.py", line 2464, in __call__
return self.wsgi_app(environ, start_response)
File "/usr/local/searx/searx-src/searx/webapp.py", line 1316, in __call__
return self.app(environ, start_response)
File "/usr/local/searx/searx-pyenv/lib/python3.9/site-packages/werkzeug/middleware/proxy_fix.py", line 169, in __call__
return self.app(environ, start_response)
File "/usr/local/searx/searx-pyenv/lib/python3.9/site-packages/flask/app.py", line 2447, in wsgi_app
response = self.full_dispatch_request()
File "/usr/local/searx/searx-pyenv/lib/python3.9/site-packages/flask/app.py", line 1950, in full_dispatch_request
rv = self.dispatch_request()
File "/usr/local/searx/searx-pyenv/lib/python3.9/site-packages/flask/app.py", line 1936, in dispatch_request
return self.view_functions[rule.endpoint](**req.view_args)
File "/usr/local/searx/searx-src/searx/webapp.py", line 766, in search
number_of_results=format_decimal(number_of_results),
File "/usr/local/searx/searx-pyenv/lib/python3.9/site-packages/flask_babel/__init__.py", line 458, in format_decimal
locale = get_locale()
File "/usr/local/searx/searx-pyenv/lib/python3.9/site-packages/flask_babel/__init__.py", line 226, in get_locale
rv = babel.locale_selector_func()
File "/usr/local/searx/searx-src/searx/webapp.py", line 249, in get_locale
logger.debug("%s uses locale `%s` from %s", request.url, locale, locale_source)
Unable to print the message and arguments - possible formatting error.
Use the traceback above to help find the error.
Signed-off-by: Markus Heiser <markus.heiser@darmarit.de>
2021-04-26 11:36:35 +02:00
|
|
|
|
import urllib
|
2022-01-17 11:16:11 +01:00
|
|
|
|
import urllib.parse
|
2021-06-08 11:22:46 +02:00
|
|
|
|
from urllib.parse import urlencode
|
2021-05-26 10:16:34 +02:00
|
|
|
|
|
|
|
|
|
import httpx
|
2020-08-06 17:42:46 +02:00
|
|
|
|
|
|
|
|
|
from pygments import highlight
|
|
|
|
|
from pygments.lexers import get_lexer_by_name
|
2020-11-03 11:35:53 +01:00
|
|
|
|
from pygments.formatters import HtmlFormatter # pylint: disable=no-name-in-module
|
2020-08-06 17:42:46 +02:00
|
|
|
|
|
2021-05-26 19:43:27 +02:00
|
|
|
|
import flask
|
|
|
|
|
|
2014-02-05 20:24:31 +01:00
|
|
|
|
from flask import (
|
2021-05-26 10:16:34 +02:00
|
|
|
|
Flask,
|
|
|
|
|
render_template,
|
|
|
|
|
url_for,
|
|
|
|
|
make_response,
|
|
|
|
|
redirect,
|
|
|
|
|
send_from_directory,
|
2014-02-05 20:24:31 +01:00
|
|
|
|
)
|
2022-01-24 11:44:18 +01:00
|
|
|
|
from flask.wrappers import Response
|
2020-04-15 23:24:12 +02:00
|
|
|
|
from flask.ctx import has_request_context
|
2016-06-07 23:08:48 +02:00
|
|
|
|
from flask.json import jsonify
|
2021-05-26 10:16:34 +02:00
|
|
|
|
|
|
|
|
|
from babel.support import Translations
|
|
|
|
|
import flask_babel
|
|
|
|
|
from flask_babel import (
|
|
|
|
|
Babel,
|
|
|
|
|
gettext,
|
|
|
|
|
format_date,
|
|
|
|
|
format_decimal,
|
|
|
|
|
)
|
|
|
|
|
|
|
|
|
|
from searx import (
|
2021-09-02 16:15:22 +02:00
|
|
|
|
logger,
|
|
|
|
|
get_setting,
|
2021-05-26 10:16:34 +02:00
|
|
|
|
settings,
|
|
|
|
|
searx_debug,
|
2022-01-15 06:09:37 +01:00
|
|
|
|
user_help,
|
2021-05-26 10:16:34 +02:00
|
|
|
|
)
|
2021-09-18 10:59:56 +02:00
|
|
|
|
from searx.data import ENGINE_DESCRIPTIONS
|
2022-01-17 11:11:39 +01:00
|
|
|
|
from searx.results import Timing, UnresponsiveEngine
|
2021-05-28 18:45:22 +02:00
|
|
|
|
from searx.settings_defaults import OUTPUT_FORMATS
|
2021-09-02 16:15:22 +02:00
|
|
|
|
from searx.settings_loader import get_default_settings_path
|
2016-11-30 18:43:03 +01:00
|
|
|
|
from searx.exceptions import SearxParameterException
|
2021-05-26 10:16:34 +02:00
|
|
|
|
from searx.engines import (
|
2021-12-28 16:12:54 +01:00
|
|
|
|
OTHER_CATEGORY,
|
2021-05-26 10:16:34 +02:00
|
|
|
|
categories,
|
|
|
|
|
engines,
|
|
|
|
|
engine_shortcuts,
|
|
|
|
|
)
|
2020-09-19 18:25:24 +02:00
|
|
|
|
from searx.webutils import (
|
2021-05-26 10:16:34 +02:00
|
|
|
|
UnicodeWriter,
|
|
|
|
|
highlight_content,
|
|
|
|
|
get_static_files,
|
|
|
|
|
get_result_templates,
|
|
|
|
|
get_themes,
|
|
|
|
|
prettify_url,
|
|
|
|
|
new_hmac,
|
2021-12-28 13:44:28 +01:00
|
|
|
|
is_hmac_of,
|
2021-05-26 10:16:34 +02:00
|
|
|
|
is_flask_run_cmdline,
|
2022-01-03 07:24:20 +01:00
|
|
|
|
DEFAULT_GROUP_NAME,
|
|
|
|
|
group_engines_in_tab,
|
2021-05-26 10:16:34 +02:00
|
|
|
|
)
|
|
|
|
|
from searx.webadapter import (
|
|
|
|
|
get_search_query_from_webapp,
|
|
|
|
|
get_selected_categories,
|
|
|
|
|
)
|
|
|
|
|
from searx.utils import (
|
|
|
|
|
html_to_text,
|
|
|
|
|
gen_useragent,
|
|
|
|
|
dict_subset,
|
|
|
|
|
match_language,
|
2014-04-25 01:46:40 +02:00
|
|
|
|
)
|
2021-08-18 19:51:07 +02:00
|
|
|
|
from searx.version import VERSION_STRING, GIT_URL, GIT_BRANCH
|
2016-11-14 22:15:03 +01:00
|
|
|
|
from searx.query import RawTextQuery
|
2022-01-17 07:54:37 +01:00
|
|
|
|
from searx.plugins import Plugin, plugins, initialize as plugin_initialize
|
2017-11-01 13:58:48 +01:00
|
|
|
|
from searx.plugins.oa_doi_rewrite import get_doi_resolver
|
2021-05-26 10:16:34 +02:00
|
|
|
|
from searx.preferences import (
|
|
|
|
|
Preferences,
|
|
|
|
|
ValidationException,
|
|
|
|
|
)
|
2021-09-02 16:15:22 +02:00
|
|
|
|
from searx.answerers import (
|
|
|
|
|
answerers,
|
|
|
|
|
ask,
|
|
|
|
|
)
|
2021-05-26 10:16:34 +02:00
|
|
|
|
from searx.metrics import (
|
|
|
|
|
get_engines_stats,
|
|
|
|
|
get_engine_errors,
|
|
|
|
|
get_reliabilities,
|
|
|
|
|
histogram,
|
|
|
|
|
counter,
|
|
|
|
|
)
|
2021-06-08 10:08:41 +02:00
|
|
|
|
from searx.flaskfix import patch_application
|
2021-05-26 10:16:34 +02:00
|
|
|
|
|
|
|
|
|
# renaming names from searx imports ...
|
|
|
|
|
|
|
|
|
|
from searx.autocomplete import search_autocomplete, backends as autocomplete_backends
|
|
|
|
|
from searx.languages import language_codes as languages
|
2021-10-06 10:26:40 +02:00
|
|
|
|
from searx.locales import LOCALE_NAMES, RTL_LOCALES
|
2021-05-26 10:16:34 +02:00
|
|
|
|
from searx.search import SearchWithPlugins, initialize as search_initialize
|
2021-08-14 20:12:11 +02:00
|
|
|
|
from searx.network import stream as http_stream, set_context_network_name
|
2021-05-26 10:16:34 +02:00
|
|
|
|
from searx.search.checker import get_result as checker_get_result
|
|
|
|
|
|
|
|
|
|
logger = logger.getChild('webapp')
|
2013-12-01 23:52:49 +01:00
|
|
|
|
|
2020-12-21 10:21:10 +01:00
|
|
|
|
# check secret_key
|
|
|
|
|
if not searx_debug and settings['server']['secret_key'] == 'ultrasecretkey':
|
|
|
|
|
logger.error('server.secret_key is not changed. Please use something else instead of ultrasecretkey.')
|
2021-05-26 13:32:04 +02:00
|
|
|
|
sys.exit(1)
|
2020-12-21 10:21:10 +01:00
|
|
|
|
|
2017-01-06 13:23:30 +01:00
|
|
|
|
# about static
|
2021-05-28 18:45:22 +02:00
|
|
|
|
logger.debug('static directory is %s', settings['ui']['static_path'])
|
|
|
|
|
static_files = get_static_files(settings['ui']['static_path'])
|
2014-10-09 19:26:02 +02:00
|
|
|
|
|
2017-01-06 13:23:30 +01:00
|
|
|
|
# about templates
|
2021-05-28 18:45:22 +02:00
|
|
|
|
logger.debug('templates directory is %s', settings['ui']['templates_path'])
|
2015-08-02 19:38:27 +02:00
|
|
|
|
default_theme = settings['ui']['default_theme']
|
2021-05-28 18:45:22 +02:00
|
|
|
|
templates_path = settings['ui']['templates_path']
|
2017-01-06 13:52:59 +01:00
|
|
|
|
themes = get_themes(templates_path)
|
2017-01-06 13:23:30 +01:00
|
|
|
|
result_templates = get_result_templates(templates_path)
|
|
|
|
|
global_favicons = []
|
|
|
|
|
for indice, theme in enumerate(themes):
|
|
|
|
|
global_favicons.append([])
|
2021-05-28 18:45:22 +02:00
|
|
|
|
theme_img_path = os.path.join(settings['ui']['static_path'], 'themes', theme, 'img', 'icons')
|
2017-01-06 13:23:30 +01:00
|
|
|
|
for (dirpath, dirnames, filenames) in os.walk(theme_img_path):
|
|
|
|
|
global_favicons[indice].extend(filenames)
|
2014-04-25 01:46:40 +02:00
|
|
|
|
|
2021-05-26 13:32:04 +02:00
|
|
|
|
STATS_SORT_PARAMETERS = {
|
|
|
|
|
'name': (False, 'name', ''),
|
|
|
|
|
'score': (True, 'score', 0),
|
|
|
|
|
'result_count': (True, 'result_count', 0),
|
|
|
|
|
'time': (False, 'total', 0),
|
|
|
|
|
'reliability': (False, 'reliability', 100),
|
|
|
|
|
}
|
|
|
|
|
|
2017-01-06 13:23:30 +01:00
|
|
|
|
# Flask app
|
2021-12-27 09:26:22 +01:00
|
|
|
|
app = Flask(__name__, static_folder=settings['ui']['static_path'], template_folder=templates_path)
|
2014-01-19 22:59:01 +01:00
|
|
|
|
|
2015-02-14 01:42:06 +01:00
|
|
|
|
app.jinja_env.trim_blocks = True
|
|
|
|
|
app.jinja_env.lstrip_blocks = True
|
2020-11-03 11:35:53 +01:00
|
|
|
|
app.jinja_env.add_extension('jinja2.ext.loopcontrols') # pylint: disable=no-member
|
2021-12-28 14:51:21 +01:00
|
|
|
|
app.jinja_env.filters['group_engines_in_tab'] = group_engines_in_tab # pylint: disable=no-member
|
2014-01-19 22:59:01 +01:00
|
|
|
|
app.secret_key = settings['server']['secret_key']
|
2014-01-14 18:17:19 +01:00
|
|
|
|
|
2014-01-22 00:15:23 +01:00
|
|
|
|
babel = Babel(app)
|
|
|
|
|
|
2016-04-09 19:47:06 +02:00
|
|
|
|
# used when translating category names
|
2021-06-08 18:19:30 +02:00
|
|
|
|
_category_names = (
|
|
|
|
|
gettext('files'),
|
|
|
|
|
gettext('general'),
|
|
|
|
|
gettext('music'),
|
|
|
|
|
gettext('social media'),
|
|
|
|
|
gettext('images'),
|
|
|
|
|
gettext('videos'),
|
|
|
|
|
gettext('it'),
|
|
|
|
|
gettext('news'),
|
|
|
|
|
gettext('map'),
|
|
|
|
|
gettext('onions'),
|
2021-12-27 09:26:22 +01:00
|
|
|
|
gettext('science'),
|
2021-12-28 14:51:21 +01:00
|
|
|
|
# non-tab categories
|
|
|
|
|
gettext('apps'),
|
|
|
|
|
gettext('dictionaries'),
|
|
|
|
|
gettext('lyrics'),
|
|
|
|
|
gettext('packages'),
|
|
|
|
|
gettext('q&a'),
|
|
|
|
|
gettext('repos'),
|
|
|
|
|
gettext('software wikis'),
|
|
|
|
|
gettext('web'),
|
|
|
|
|
gettext(DEFAULT_GROUP_NAME),
|
2021-12-28 16:12:54 +01:00
|
|
|
|
gettext(OTHER_CATEGORY),
|
2021-06-08 18:19:30 +02:00
|
|
|
|
)
|
2015-02-10 23:14:37 +01:00
|
|
|
|
|
2021-12-27 09:26:22 +01:00
|
|
|
|
_simple_style = (gettext('auto'), gettext('light'), gettext('dark'))
|
2021-11-19 13:49:16 +01:00
|
|
|
|
|
2021-04-14 18:11:35 +02:00
|
|
|
|
#
|
2021-04-26 11:12:02 +02:00
|
|
|
|
timeout_text = gettext('timeout')
|
|
|
|
|
parsing_error_text = gettext('parsing error')
|
|
|
|
|
http_protocol_error_text = gettext('HTTP protocol error')
|
|
|
|
|
network_error_text = gettext('network error')
|
|
|
|
|
exception_classname_to_text = {
|
|
|
|
|
None: gettext('unexpected crash'),
|
|
|
|
|
'timeout': timeout_text,
|
|
|
|
|
'asyncio.TimeoutError': timeout_text,
|
|
|
|
|
'httpx.TimeoutException': timeout_text,
|
|
|
|
|
'httpx.ConnectTimeout': timeout_text,
|
|
|
|
|
'httpx.ReadTimeout': timeout_text,
|
|
|
|
|
'httpx.WriteTimeout': timeout_text,
|
|
|
|
|
'httpx.HTTPStatusError': gettext('HTTP error'),
|
|
|
|
|
'httpx.ConnectError': gettext("HTTP connection error"),
|
|
|
|
|
'httpx.RemoteProtocolError': http_protocol_error_text,
|
|
|
|
|
'httpx.LocalProtocolError': http_protocol_error_text,
|
|
|
|
|
'httpx.ProtocolError': http_protocol_error_text,
|
|
|
|
|
'httpx.ReadError': network_error_text,
|
|
|
|
|
'httpx.WriteError': network_error_text,
|
|
|
|
|
'httpx.ProxyError': gettext("proxy error"),
|
|
|
|
|
'searx.exceptions.SearxEngineCaptchaException': gettext("CAPTCHA"),
|
|
|
|
|
'searx.exceptions.SearxEngineTooManyRequestsException': gettext("too many requests"),
|
|
|
|
|
'searx.exceptions.SearxEngineAccessDeniedException': gettext("access denied"),
|
|
|
|
|
'searx.exceptions.SearxEngineAPIException': gettext("server API error"),
|
|
|
|
|
'searx.exceptions.SearxEngineXPathException': parsing_error_text,
|
|
|
|
|
'KeyError': parsing_error_text,
|
|
|
|
|
'json.decoder.JSONDecodeError': parsing_error_text,
|
|
|
|
|
'lxml.etree.ParserError': parsing_error_text,
|
2021-04-14 18:11:35 +02:00
|
|
|
|
}
|
|
|
|
|
|
2020-02-24 05:46:26 +01:00
|
|
|
|
|
|
|
|
|
# monkey patch for flask_babel.get_translations
|
2021-08-17 08:18:30 +02:00
|
|
|
|
_flask_babel_get_translations = flask_babel.get_translations
|
2021-12-27 09:26:22 +01:00
|
|
|
|
|
|
|
|
|
|
2022-01-17 07:54:37 +01:00
|
|
|
|
class ExtendedRequest(flask.Request):
|
|
|
|
|
"""This class is never initialized and only used for type checking."""
|
|
|
|
|
|
|
|
|
|
preferences: Preferences
|
|
|
|
|
errors: List[str]
|
|
|
|
|
user_plugins: List[Plugin]
|
|
|
|
|
form: Dict[str, str]
|
|
|
|
|
start_time: float
|
|
|
|
|
render_time: float
|
2022-01-17 08:06:31 +01:00
|
|
|
|
timings: List[Timing]
|
2022-01-17 07:54:37 +01:00
|
|
|
|
|
|
|
|
|
|
|
|
|
|
request = typing.cast(ExtendedRequest, flask.request)
|
|
|
|
|
|
|
|
|
|
|
2020-02-24 05:46:26 +01:00
|
|
|
|
def _get_translations():
|
2020-04-15 23:24:12 +02:00
|
|
|
|
if has_request_context() and request.form.get('use-translation') == 'oc':
|
2020-02-24 05:46:26 +01:00
|
|
|
|
babel_ext = flask_babel.current_app.extensions['babel']
|
2020-04-15 23:24:12 +02:00
|
|
|
|
return Translations.load(next(babel_ext.translation_directories), 'oc')
|
|
|
|
|
return _flask_babel_get_translations()
|
2021-12-27 09:26:22 +01:00
|
|
|
|
|
|
|
|
|
|
2021-08-17 08:18:30 +02:00
|
|
|
|
flask_babel.get_translations = _get_translations
|
2020-02-24 05:46:26 +01:00
|
|
|
|
|
|
|
|
|
|
2021-08-17 08:18:30 +02:00
|
|
|
|
@babel.localeselector
|
|
|
|
|
def get_locale():
|
2022-01-24 11:44:18 +01:00
|
|
|
|
locale = 'en'
|
|
|
|
|
|
|
|
|
|
if has_request_context():
|
|
|
|
|
value = request.preferences.get_value('locale')
|
|
|
|
|
if value:
|
|
|
|
|
locale = value
|
|
|
|
|
|
2021-08-17 08:18:30 +02:00
|
|
|
|
if locale == 'oc':
|
|
|
|
|
request.form['use-translation'] = 'oc'
|
|
|
|
|
locale = 'fr_FR'
|
2021-10-06 10:26:40 +02:00
|
|
|
|
if locale == '':
|
|
|
|
|
# if there is an error loading the preferences
|
|
|
|
|
# the locale is going to be ''
|
|
|
|
|
locale = 'en'
|
|
|
|
|
# babel uses underscore instead of hyphen.
|
|
|
|
|
locale = locale.replace('-', '_')
|
2021-08-17 08:18:30 +02:00
|
|
|
|
logger.debug("%s uses locale `%s`", urllib.parse.quote(request.url), locale)
|
|
|
|
|
return locale
|
2020-02-24 05:46:26 +01:00
|
|
|
|
|
2014-01-19 23:04:09 +01:00
|
|
|
|
|
2021-08-17 08:18:30 +02:00
|
|
|
|
def _get_browser_language(req, lang_list):
|
2021-05-26 13:32:04 +02:00
|
|
|
|
for lang in req.headers.get("Accept-Language", "en").split(","):
|
2020-03-16 00:22:15 +01:00
|
|
|
|
if ';' in lang:
|
|
|
|
|
lang = lang.split(';')[0]
|
2021-02-03 09:02:21 +01:00
|
|
|
|
if '-' in lang:
|
|
|
|
|
lang_parts = lang.split('-')
|
|
|
|
|
lang = "{}-{}".format(lang_parts[0], lang_parts[-1].upper())
|
2020-02-23 10:03:42 +01:00
|
|
|
|
locale = match_language(lang, lang_list, fallback=None)
|
|
|
|
|
if locale is not None:
|
|
|
|
|
return locale
|
2021-08-17 08:18:30 +02:00
|
|
|
|
return 'en'
|
2014-01-22 00:15:23 +01:00
|
|
|
|
|
|
|
|
|
|
2021-10-06 10:26:40 +02:00
|
|
|
|
def _get_locale_rfc5646(locale):
|
|
|
|
|
"""Get locale name for <html lang="...">
|
|
|
|
|
Chrom* browsers don't detect the language when there is a subtag (ie a territory).
|
|
|
|
|
For example "zh-TW" is detected but not "zh-Hant-TW".
|
|
|
|
|
This function returns a locale without the subtag.
|
|
|
|
|
"""
|
|
|
|
|
parts = locale.split('-')
|
|
|
|
|
return parts[0].lower() + '-' + parts[-1].upper()
|
|
|
|
|
|
|
|
|
|
|
2014-12-20 23:33:03 +01:00
|
|
|
|
# code-highlighter
|
|
|
|
|
@app.template_filter('code_highlighter')
|
|
|
|
|
def code_highlighter(codelines, language=None):
|
|
|
|
|
if not language:
|
|
|
|
|
language = 'text'
|
|
|
|
|
|
2014-12-22 16:26:45 +01:00
|
|
|
|
try:
|
|
|
|
|
# find lexer by programing language
|
|
|
|
|
lexer = get_lexer_by_name(language, stripall=True)
|
2021-05-26 13:32:04 +02:00
|
|
|
|
|
|
|
|
|
except Exception as e: # pylint: disable=broad-except
|
|
|
|
|
logger.exception(e, exc_info=True)
|
2014-12-22 16:26:45 +01:00
|
|
|
|
# if lexer is not found, using default one
|
|
|
|
|
lexer = get_lexer_by_name('text', stripall=True)
|
|
|
|
|
|
2014-12-20 23:33:03 +01:00
|
|
|
|
html_code = ''
|
|
|
|
|
tmp_code = ''
|
|
|
|
|
last_line = None
|
2022-01-24 11:44:18 +01:00
|
|
|
|
line_code_start = None
|
2014-12-20 23:33:03 +01:00
|
|
|
|
|
|
|
|
|
# parse lines
|
|
|
|
|
for line, code in codelines:
|
|
|
|
|
if not last_line:
|
|
|
|
|
line_code_start = line
|
|
|
|
|
|
|
|
|
|
# new codeblock is detected
|
2021-12-27 09:26:22 +01:00
|
|
|
|
if last_line is not None and last_line + 1 != line:
|
2014-12-20 23:33:03 +01:00
|
|
|
|
|
|
|
|
|
# highlight last codepart
|
2021-12-27 09:26:22 +01:00
|
|
|
|
formatter = HtmlFormatter(linenos='inline', linenostart=line_code_start, cssclass="code-highlight")
|
2014-12-20 23:33:03 +01:00
|
|
|
|
html_code = html_code + highlight(tmp_code, lexer, formatter)
|
2014-12-22 16:26:45 +01:00
|
|
|
|
|
2014-12-20 23:33:03 +01:00
|
|
|
|
# reset conditions for next codepart
|
|
|
|
|
tmp_code = ''
|
|
|
|
|
line_code_start = line
|
|
|
|
|
|
|
|
|
|
# add codepart
|
|
|
|
|
tmp_code += code + '\n'
|
2014-12-22 16:26:45 +01:00
|
|
|
|
|
2014-12-20 23:33:03 +01:00
|
|
|
|
# update line
|
|
|
|
|
last_line = line
|
|
|
|
|
|
|
|
|
|
# highlight last codepart
|
2020-09-16 08:53:05 +02:00
|
|
|
|
formatter = HtmlFormatter(linenos='inline', linenostart=line_code_start, cssclass="code-highlight")
|
2014-12-20 23:33:03 +01:00
|
|
|
|
html_code = html_code + highlight(tmp_code, lexer, formatter)
|
|
|
|
|
|
|
|
|
|
return html_code
|
|
|
|
|
|
|
|
|
|
|
2022-01-17 11:16:11 +01:00
|
|
|
|
def get_current_theme_name(override: str = None) -> str:
|
2014-04-25 01:46:40 +02:00
|
|
|
|
"""Returns theme name.
|
|
|
|
|
|
|
|
|
|
Checks in this order:
|
|
|
|
|
1. override
|
|
|
|
|
2. cookies
|
|
|
|
|
3. settings"""
|
|
|
|
|
|
2017-01-13 22:15:11 +01:00
|
|
|
|
if override and (override in themes or override == '__common__'):
|
2014-04-25 01:46:40 +02:00
|
|
|
|
return override
|
2016-04-08 16:38:05 +02:00
|
|
|
|
theme_name = request.args.get('theme', request.preferences.get_value('theme'))
|
2022-01-24 11:44:18 +01:00
|
|
|
|
|
|
|
|
|
if theme_name and theme_name in themes:
|
|
|
|
|
return theme_name
|
|
|
|
|
|
|
|
|
|
return default_theme
|
2014-04-25 01:46:40 +02:00
|
|
|
|
|
|
|
|
|
|
2022-01-17 11:16:11 +01:00
|
|
|
|
def get_result_template(theme_name: str, template_name: str):
|
2021-05-26 13:32:04 +02:00
|
|
|
|
themed_path = theme_name + '/result_templates/' + template_name
|
2015-01-01 18:59:53 +01:00
|
|
|
|
if themed_path in result_templates:
|
|
|
|
|
return themed_path
|
|
|
|
|
return 'result_templates/' + template_name
|
|
|
|
|
|
|
|
|
|
|
2022-01-17 11:16:11 +01:00
|
|
|
|
def url_for_theme(endpoint: str, override_theme: str = None, **values):
|
2015-01-01 17:48:12 +01:00
|
|
|
|
if endpoint == 'static' and values.get('filename'):
|
2014-04-25 01:46:40 +02:00
|
|
|
|
theme_name = get_current_theme_name(override=override_theme)
|
2015-01-01 17:48:12 +01:00
|
|
|
|
filename_with_theme = "themes/{}/{}".format(theme_name, values['filename'])
|
|
|
|
|
if filename_with_theme in static_files:
|
|
|
|
|
values['filename'] = filename_with_theme
|
2020-10-08 14:19:00 +02:00
|
|
|
|
url = url_for(endpoint, **values)
|
|
|
|
|
return url
|
2014-04-25 01:46:40 +02:00
|
|
|
|
|
|
|
|
|
|
2022-01-17 11:16:11 +01:00
|
|
|
|
def proxify(url: str):
|
2016-10-17 00:22:41 +02:00
|
|
|
|
if url.startswith('//'):
|
|
|
|
|
url = 'https:' + url
|
|
|
|
|
|
|
|
|
|
if not settings.get('result_proxy'):
|
|
|
|
|
return url
|
|
|
|
|
|
2022-01-24 11:44:18 +01:00
|
|
|
|
url_params = dict(mortyurl=url)
|
2016-10-29 23:12:24 +02:00
|
|
|
|
|
|
|
|
|
if settings['result_proxy'].get('key'):
|
2021-12-27 09:26:22 +01:00
|
|
|
|
url_params['mortyhash'] = hmac.new(settings['result_proxy']['key'], url.encode(), hashlib.sha256).hexdigest()
|
|
|
|
|
|
|
|
|
|
return '{0}?{1}'.format(settings['result_proxy']['url'], urlencode(url_params))
|
2016-10-17 00:22:41 +02:00
|
|
|
|
|
|
|
|
|
|
2022-01-17 11:16:11 +01:00
|
|
|
|
def image_proxify(url: str):
|
2015-01-16 16:26:15 +01:00
|
|
|
|
|
|
|
|
|
if url.startswith('//'):
|
|
|
|
|
url = 'https:' + url
|
|
|
|
|
|
2016-04-09 18:32:07 +02:00
|
|
|
|
if not request.preferences.get_value('image_proxy'):
|
2015-01-16 16:26:15 +01:00
|
|
|
|
return url
|
|
|
|
|
|
2020-06-22 13:57:33 +02:00
|
|
|
|
if url.startswith('data:image/'):
|
|
|
|
|
# 50 is an arbitrary number to get only the beginning of the image.
|
2021-12-27 09:26:22 +01:00
|
|
|
|
partial_base64 = url[len('data:image/') : 50].split(';')
|
|
|
|
|
if (
|
|
|
|
|
len(partial_base64) == 2
|
|
|
|
|
and partial_base64[0] in ['gif', 'png', 'jpeg', 'pjpeg', 'webp', 'tiff', 'bmp']
|
|
|
|
|
and partial_base64[1].startswith('base64,')
|
|
|
|
|
):
|
2020-06-22 13:57:33 +02:00
|
|
|
|
return url
|
2021-05-26 13:32:04 +02:00
|
|
|
|
return None
|
2019-01-18 09:04:40 +01:00
|
|
|
|
|
2016-10-30 21:15:18 +01:00
|
|
|
|
if settings.get('result_proxy'):
|
|
|
|
|
return proxify(url)
|
|
|
|
|
|
2020-08-06 17:42:46 +02:00
|
|
|
|
h = new_hmac(settings['server']['secret_key'], url.encode())
|
2015-01-18 09:54:24 +01:00
|
|
|
|
|
2021-12-27 09:26:22 +01:00
|
|
|
|
return '{0}?{1}'.format(url_for('image_proxy'), urlencode(dict(url=url.encode(), h=h)))
|
2015-01-16 16:26:15 +01:00
|
|
|
|
|
|
|
|
|
|
2021-03-16 11:07:04 +01:00
|
|
|
|
def get_translations():
|
|
|
|
|
return {
|
|
|
|
|
# when there is autocompletion
|
2021-09-18 11:01:39 +02:00
|
|
|
|
'no_item_found': gettext('No item found'),
|
|
|
|
|
# /preferences: the source of the engine description (wikipedata, wikidata, website)
|
|
|
|
|
'Source': gettext('Source'),
|
2021-03-16 11:07:04 +01:00
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
2022-01-17 11:16:11 +01:00
|
|
|
|
def _get_enable_categories(all_categories: Iterable[str]):
|
2021-06-08 11:56:30 +02:00
|
|
|
|
disabled_engines = request.preferences.engines.get_disabled()
|
2021-06-08 18:19:30 +02:00
|
|
|
|
enabled_categories = set(
|
2021-07-03 17:51:39 +02:00
|
|
|
|
# pylint: disable=consider-using-dict-items
|
2021-12-27 09:26:22 +01:00
|
|
|
|
category
|
|
|
|
|
for engine_name in engines
|
2021-06-08 18:19:30 +02:00
|
|
|
|
for category in engines[engine_name].categories
|
|
|
|
|
if (engine_name, category) not in disabled_engines
|
|
|
|
|
)
|
|
|
|
|
return [x for x in all_categories if x in enabled_categories]
|
2021-06-08 11:56:30 +02:00
|
|
|
|
|
|
|
|
|
|
2022-01-17 11:16:11 +01:00
|
|
|
|
def get_pretty_url(parsed_url: urllib.parse.ParseResult):
|
2021-11-21 21:38:00 +01:00
|
|
|
|
path = parsed_url.path
|
|
|
|
|
path = path[:-1] if len(path) > 0 and path[-1] == '/' else path
|
|
|
|
|
path = path.replace("/", " › ")
|
2021-12-27 09:26:22 +01:00
|
|
|
|
return [parsed_url.scheme + "://" + parsed_url.netloc, path]
|
2021-11-21 21:38:00 +01:00
|
|
|
|
|
|
|
|
|
|
2022-01-17 11:16:11 +01:00
|
|
|
|
def render(template_name: str, override_theme: str = None, **kwargs):
|
2021-06-08 11:17:55 +02:00
|
|
|
|
# values from the HTTP requests
|
|
|
|
|
kwargs['endpoint'] = 'results' if 'q' in kwargs else request.endpoint
|
|
|
|
|
kwargs['cookies'] = request.cookies
|
|
|
|
|
kwargs['errors'] = request.errors
|
2014-03-29 16:29:19 +01:00
|
|
|
|
|
2021-06-08 11:17:55 +02:00
|
|
|
|
# values from the preferences
|
|
|
|
|
kwargs['preferences'] = request.preferences
|
|
|
|
|
kwargs['method'] = request.preferences.get_value('method')
|
|
|
|
|
kwargs['autocomplete'] = request.preferences.get_value('autocomplete')
|
|
|
|
|
kwargs['results_on_new_tab'] = request.preferences.get_value('results_on_new_tab')
|
2021-06-08 11:56:30 +02:00
|
|
|
|
kwargs['advanced_search'] = request.preferences.get_value('advanced_search')
|
2021-11-06 12:26:48 +01:00
|
|
|
|
kwargs['query_in_title'] = request.preferences.get_value('query_in_title')
|
2021-06-08 11:17:55 +02:00
|
|
|
|
kwargs['safesearch'] = str(request.preferences.get_value('safesearch'))
|
|
|
|
|
kwargs['theme'] = get_current_theme_name(override=override_theme)
|
2022-01-04 18:00:45 +01:00
|
|
|
|
kwargs['categories_as_tabs'] = list(settings['categories_as_tabs'].keys())
|
2021-12-22 15:51:26 +01:00
|
|
|
|
kwargs['categories'] = _get_enable_categories(categories.keys())
|
2021-12-28 16:12:54 +01:00
|
|
|
|
kwargs['OTHER_CATEGORY'] = OTHER_CATEGORY
|
2014-03-29 16:29:19 +01:00
|
|
|
|
|
2021-06-08 11:17:55 +02:00
|
|
|
|
# i18n
|
2021-12-27 09:26:22 +01:00
|
|
|
|
kwargs['language_codes'] = [l for l in languages if l[0] in settings['search']['languages']]
|
2021-06-08 11:17:55 +02:00
|
|
|
|
kwargs['translations'] = json.dumps(get_translations(), separators=(',', ':'))
|
2014-03-29 16:29:19 +01:00
|
|
|
|
|
2019-03-28 18:07:03 +01:00
|
|
|
|
locale = request.preferences.get_value('locale')
|
2021-10-06 10:26:40 +02:00
|
|
|
|
kwargs['locale_rfc5646'] = _get_locale_rfc5646(locale)
|
2021-10-05 15:03:28 +02:00
|
|
|
|
|
2021-08-03 15:13:00 +02:00
|
|
|
|
if locale in RTL_LOCALES and 'rtl' not in kwargs:
|
2015-02-10 15:23:56 +01:00
|
|
|
|
kwargs['rtl'] = True
|
2016-08-06 06:34:56 +02:00
|
|
|
|
if 'current_language' not in kwargs:
|
2021-06-08 18:19:30 +02:00
|
|
|
|
kwargs['current_language'] = match_language(
|
2021-12-27 09:26:22 +01:00
|
|
|
|
request.preferences.get_value('language'), settings['search']['languages']
|
|
|
|
|
)
|
2016-08-06 06:34:56 +02:00
|
|
|
|
|
2021-06-08 11:17:55 +02:00
|
|
|
|
# values from settings
|
2021-12-27 09:26:22 +01:00
|
|
|
|
kwargs['search_formats'] = [x for x in settings['search']['formats'] if x != 'html']
|
2021-07-18 15:38:52 +02:00
|
|
|
|
kwargs['instance_name'] = get_setting('general.instance_name')
|
2021-06-08 11:17:55 +02:00
|
|
|
|
kwargs['searx_version'] = VERSION_STRING
|
2021-07-27 18:37:46 +02:00
|
|
|
|
kwargs['searx_git_url'] = GIT_URL
|
2021-07-18 15:38:52 +02:00
|
|
|
|
kwargs['get_setting'] = get_setting
|
2021-11-21 21:38:00 +01:00
|
|
|
|
kwargs['get_pretty_url'] = get_pretty_url
|
2020-03-25 11:49:33 +01:00
|
|
|
|
|
2021-06-08 11:17:55 +02:00
|
|
|
|
# helpers to create links to other pages
|
|
|
|
|
kwargs['url_for'] = url_for_theme # override url_for function in templates
|
|
|
|
|
kwargs['image_proxify'] = image_proxify
|
|
|
|
|
kwargs['proxify'] = proxify if settings.get('result_proxy', {}).get('url') else None
|
|
|
|
|
kwargs['proxify_results'] = settings.get('result_proxy', {}).get('proxify_results', True)
|
|
|
|
|
kwargs['get_result_template'] = get_result_template
|
2021-06-08 18:19:30 +02:00
|
|
|
|
kwargs['opensearch_url'] = (
|
2021-12-27 09:26:22 +01:00
|
|
|
|
url_for('opensearch') + '?' + urlencode({'method': kwargs['method'], 'autocomplete': kwargs['autocomplete']})
|
2021-06-08 18:19:30 +02:00
|
|
|
|
)
|
2021-03-16 11:07:04 +01:00
|
|
|
|
|
2021-06-08 11:17:55 +02:00
|
|
|
|
# scripts from plugins
|
2015-04-12 19:24:01 +02:00
|
|
|
|
kwargs['scripts'] = set()
|
|
|
|
|
for plugin in request.user_plugins:
|
|
|
|
|
for script in plugin.js_dependencies:
|
|
|
|
|
kwargs['scripts'].add(script)
|
|
|
|
|
|
2021-06-08 11:17:55 +02:00
|
|
|
|
# styles from plugins
|
2015-04-12 19:24:01 +02:00
|
|
|
|
kwargs['styles'] = set()
|
|
|
|
|
for plugin in request.user_plugins:
|
|
|
|
|
for css in plugin.css_dependencies:
|
|
|
|
|
kwargs['styles'].add(css)
|
|
|
|
|
|
2021-05-06 09:39:52 +02:00
|
|
|
|
start_time = default_timer()
|
2021-12-27 09:26:22 +01:00
|
|
|
|
result = render_template('{}/{}'.format(kwargs['theme'], template_name), **kwargs)
|
2021-05-14 11:15:35 +02:00
|
|
|
|
request.render_time += default_timer() - start_time # pylint: disable=assigning-non-slot
|
|
|
|
|
|
2021-05-06 09:39:52 +02:00
|
|
|
|
return result
|
2013-10-15 20:50:12 +02:00
|
|
|
|
|
2014-01-19 22:59:01 +01:00
|
|
|
|
|
2015-03-10 20:44:02 +01:00
|
|
|
|
@app.before_request
|
|
|
|
|
def pre_request():
|
2021-05-14 11:15:35 +02:00
|
|
|
|
request.start_time = default_timer() # pylint: disable=assigning-non-slot
|
|
|
|
|
request.render_time = 0 # pylint: disable=assigning-non-slot
|
|
|
|
|
request.timings = [] # pylint: disable=assigning-non-slot
|
|
|
|
|
request.errors = [] # pylint: disable=assigning-non-slot
|
2016-11-14 22:07:23 +01:00
|
|
|
|
|
2021-05-26 13:32:04 +02:00
|
|
|
|
preferences = Preferences(themes, list(categories.keys()), engines, plugins) # pylint: disable=redefined-outer-name
|
2020-08-18 18:29:58 +02:00
|
|
|
|
user_agent = request.headers.get('User-Agent', '').lower()
|
|
|
|
|
if 'webkit' in user_agent and 'android' in user_agent:
|
|
|
|
|
preferences.key_value_settings['method'].value = 'GET'
|
2021-05-14 11:15:35 +02:00
|
|
|
|
request.preferences = preferences # pylint: disable=assigning-non-slot
|
2021-05-26 13:32:04 +02:00
|
|
|
|
|
2016-09-04 15:56:46 +02:00
|
|
|
|
try:
|
2017-07-10 12:47:25 +02:00
|
|
|
|
preferences.parse_dict(request.cookies)
|
2021-05-26 13:32:04 +02:00
|
|
|
|
|
|
|
|
|
except Exception as e: # pylint: disable=broad-except
|
|
|
|
|
logger.exception(e, exc_info=True)
|
2016-11-14 22:07:23 +01:00
|
|
|
|
request.errors.append(gettext('Invalid settings, please edit your preferences'))
|
2016-04-08 16:38:05 +02:00
|
|
|
|
|
2016-11-14 22:07:23 +01:00
|
|
|
|
# merge GET, POST vars
|
2016-10-22 13:10:31 +02:00
|
|
|
|
# request.form
|
2021-05-14 11:15:35 +02:00
|
|
|
|
request.form = dict(request.form.items()) # pylint: disable=assigning-non-slot
|
2015-03-15 12:13:24 +01:00
|
|
|
|
for k, v in request.args.items():
|
2015-03-10 22:45:59 +01:00
|
|
|
|
if k not in request.form:
|
|
|
|
|
request.form[k] = v
|
2017-10-25 23:56:37 +02:00
|
|
|
|
|
|
|
|
|
if request.form.get('preferences'):
|
|
|
|
|
preferences.parse_encoded_data(request.form['preferences'])
|
|
|
|
|
else:
|
|
|
|
|
try:
|
|
|
|
|
preferences.parse_dict(request.form)
|
2021-05-26 13:32:04 +02:00
|
|
|
|
except Exception as e: # pylint: disable=broad-except
|
|
|
|
|
logger.exception(e, exc_info=True)
|
2017-10-25 23:56:37 +02:00
|
|
|
|
request.errors.append(gettext('Invalid settings'))
|
2015-03-10 20:44:02 +01:00
|
|
|
|
|
2021-08-17 08:18:30 +02:00
|
|
|
|
# language is defined neither in settings nor in preferences
|
|
|
|
|
# use browser headers
|
2019-03-28 18:07:03 +01:00
|
|
|
|
if not preferences.get_value("language"):
|
2021-12-27 09:26:22 +01:00
|
|
|
|
language = _get_browser_language(request, settings['search']['languages'])
|
2021-08-17 08:18:30 +02:00
|
|
|
|
preferences.parse_dict({"language": language})
|
|
|
|
|
|
|
|
|
|
# locale is defined neither in settings nor in preferences
|
|
|
|
|
# use browser headers
|
2019-03-28 18:07:03 +01:00
|
|
|
|
if not preferences.get_value("locale"):
|
2021-10-06 10:26:40 +02:00
|
|
|
|
locale = _get_browser_language(request, LOCALE_NAMES.keys())
|
2021-08-17 08:18:30 +02:00
|
|
|
|
preferences.parse_dict({"locale": locale})
|
2019-03-28 18:07:03 +01:00
|
|
|
|
|
2016-10-22 13:10:31 +02:00
|
|
|
|
# request.user_plugins
|
2021-05-14 11:15:35 +02:00
|
|
|
|
request.user_plugins = [] # pylint: disable=assigning-non-slot
|
2016-04-08 16:38:05 +02:00
|
|
|
|
allowed_plugins = preferences.plugins.get_enabled()
|
|
|
|
|
disabled_plugins = preferences.plugins.get_disabled()
|
2015-03-10 20:44:02 +01:00
|
|
|
|
for plugin in plugins:
|
2021-12-27 09:26:22 +01:00
|
|
|
|
if (plugin.default_on and plugin.id not in disabled_plugins) or plugin.id in allowed_plugins:
|
2015-03-10 20:44:02 +01:00
|
|
|
|
request.user_plugins.append(plugin)
|
|
|
|
|
|
|
|
|
|
|
2020-11-04 17:32:51 +01:00
|
|
|
|
@app.after_request
|
2022-01-17 11:16:11 +01:00
|
|
|
|
def add_default_headers(response: flask.Response):
|
2020-11-04 17:32:51 +01:00
|
|
|
|
# set default http headers
|
2021-05-28 18:45:22 +02:00
|
|
|
|
for header, value in settings['server']['default_http_headers'].items():
|
2020-11-04 17:32:51 +01:00
|
|
|
|
if header in response.headers:
|
|
|
|
|
continue
|
|
|
|
|
response.headers[header] = value
|
|
|
|
|
return response
|
|
|
|
|
|
|
|
|
|
|
2019-07-17 10:38:45 +02:00
|
|
|
|
@app.after_request
|
2022-01-17 11:16:11 +01:00
|
|
|
|
def post_request(response: flask.Response):
|
2021-04-14 17:23:15 +02:00
|
|
|
|
total_time = default_timer() - request.start_time
|
2021-12-27 09:26:22 +01:00
|
|
|
|
timings_all = [
|
|
|
|
|
'total;dur=' + str(round(total_time * 1000, 3)),
|
|
|
|
|
'render;dur=' + str(round(request.render_time * 1000, 3)),
|
|
|
|
|
]
|
2019-07-17 10:38:45 +02:00
|
|
|
|
if len(request.timings) > 0:
|
2022-01-17 08:06:31 +01:00
|
|
|
|
timings = sorted(request.timings, key=lambda t: t.total)
|
2021-06-08 18:19:30 +02:00
|
|
|
|
timings_total = [
|
2022-01-17 08:06:31 +01:00
|
|
|
|
'total_' + str(i) + '_' + t.engine + ';dur=' + str(round(t.total * 1000, 3)) for i, t in enumerate(timings)
|
2021-06-08 18:19:30 +02:00
|
|
|
|
]
|
|
|
|
|
timings_load = [
|
2022-01-17 08:06:31 +01:00
|
|
|
|
'load_' + str(i) + '_' + t.engine + ';dur=' + str(round(t.load * 1000, 3))
|
|
|
|
|
for i, t in enumerate(timings)
|
|
|
|
|
if t.load
|
2021-06-08 18:19:30 +02:00
|
|
|
|
]
|
2019-07-17 10:38:45 +02:00
|
|
|
|
timings_all = timings_all + timings_total + timings_load
|
|
|
|
|
response.headers.add('Server-Timing', ', '.join(timings_all))
|
|
|
|
|
return response
|
|
|
|
|
|
|
|
|
|
|
2022-01-17 11:16:11 +01:00
|
|
|
|
def index_error(output_format: str, error_message: str):
|
2017-01-20 18:52:47 +01:00
|
|
|
|
if output_format == 'json':
|
2021-12-27 09:26:22 +01:00
|
|
|
|
return Response(json.dumps({'error': error_message}), mimetype='application/json')
|
2021-05-26 13:32:04 +02:00
|
|
|
|
if output_format == 'csv':
|
2017-01-20 18:52:47 +01:00
|
|
|
|
response = Response('', mimetype='application/csv')
|
|
|
|
|
cont_disp = 'attachment;Filename=searx.csv'
|
|
|
|
|
response.headers.add('Content-Disposition', cont_disp)
|
|
|
|
|
return response
|
2021-05-26 13:32:04 +02:00
|
|
|
|
|
|
|
|
|
if output_format == 'rss':
|
2017-01-20 18:52:47 +01:00
|
|
|
|
response_rss = render(
|
|
|
|
|
'opensearch_response_rss.xml',
|
|
|
|
|
results=[],
|
|
|
|
|
q=request.form['q'] if 'q' in request.form else '',
|
|
|
|
|
number_of_results=0,
|
2017-02-13 21:36:45 +01:00
|
|
|
|
error_message=error_message,
|
|
|
|
|
override_theme='__common__',
|
2017-01-20 18:52:47 +01:00
|
|
|
|
)
|
|
|
|
|
return Response(response_rss, mimetype='text/xml')
|
2021-05-26 13:32:04 +02:00
|
|
|
|
|
|
|
|
|
# html
|
|
|
|
|
request.errors.append(gettext('search error'))
|
|
|
|
|
return render(
|
2021-12-27 09:16:03 +01:00
|
|
|
|
# fmt: off
|
2021-05-26 13:32:04 +02:00
|
|
|
|
'index.html',
|
|
|
|
|
selected_categories=get_selected_categories(request.preferences, request.form),
|
2021-12-27 09:16:03 +01:00
|
|
|
|
# fmt: on
|
2021-05-26 13:32:04 +02:00
|
|
|
|
)
|
2017-01-20 18:52:47 +01:00
|
|
|
|
|
|
|
|
|
|
2014-01-14 18:19:21 +01:00
|
|
|
|
@app.route('/', methods=['GET', 'POST'])
|
2013-10-14 23:09:13 +02:00
|
|
|
|
def index():
|
2019-07-30 06:25:05 +02:00
|
|
|
|
"""Render index page."""
|
|
|
|
|
|
|
|
|
|
# redirect to search if there's a query in the request
|
|
|
|
|
if request.form.get('q'):
|
2020-11-06 12:11:52 +01:00
|
|
|
|
query = ('?' + request.query_string.decode()) if request.query_string else ''
|
|
|
|
|
return redirect(url_for('search') + query, 308)
|
2019-07-30 06:25:05 +02:00
|
|
|
|
|
|
|
|
|
return render(
|
2021-12-27 09:16:03 +01:00
|
|
|
|
# fmt: off
|
2019-07-30 06:25:05 +02:00
|
|
|
|
'index.html',
|
2020-11-06 12:11:52 +01:00
|
|
|
|
selected_categories=get_selected_categories(request.preferences, request.form),
|
2021-12-27 09:16:03 +01:00
|
|
|
|
# fmt: on
|
2019-07-30 06:25:05 +02:00
|
|
|
|
)
|
|
|
|
|
|
|
|
|
|
|
2021-10-02 17:00:37 +02:00
|
|
|
|
@app.route('/healthz', methods=['GET'])
|
|
|
|
|
def health():
|
|
|
|
|
return Response('OK', mimetype='text/plain')
|
|
|
|
|
|
|
|
|
|
|
2019-07-30 06:25:05 +02:00
|
|
|
|
@app.route('/search', methods=['GET', 'POST'])
|
|
|
|
|
def search():
|
|
|
|
|
"""Search query in q and return results.
|
2014-01-31 07:08:24 +01:00
|
|
|
|
|
|
|
|
|
Supported outputs: html, json, csv, rss.
|
|
|
|
|
"""
|
2021-05-26 13:32:04 +02:00
|
|
|
|
# pylint: disable=too-many-locals, too-many-return-statements, too-many-branches
|
|
|
|
|
# pylint: disable=too-many-statements
|
2013-11-04 00:18:07 +01:00
|
|
|
|
|
2017-01-20 18:52:47 +01:00
|
|
|
|
# output_format
|
|
|
|
|
output_format = request.form.get('format', 'html')
|
2021-05-26 19:43:27 +02:00
|
|
|
|
if output_format not in OUTPUT_FORMATS:
|
2017-01-20 18:52:47 +01:00
|
|
|
|
output_format = 'html'
|
|
|
|
|
|
2021-05-28 18:45:22 +02:00
|
|
|
|
if output_format not in settings['search']['formats']:
|
2021-05-26 19:43:27 +02:00
|
|
|
|
flask.abort(403)
|
|
|
|
|
|
2020-11-06 12:11:52 +01:00
|
|
|
|
# check if there is query (not None and not an empty string)
|
|
|
|
|
if not request.form.get('q'):
|
2017-01-20 18:52:47 +01:00
|
|
|
|
if output_format == 'html':
|
|
|
|
|
return render(
|
2021-12-27 09:16:03 +01:00
|
|
|
|
# fmt: off
|
2017-01-20 18:52:47 +01:00
|
|
|
|
'index.html',
|
2020-09-24 16:26:00 +02:00
|
|
|
|
selected_categories=get_selected_categories(request.preferences, request.form),
|
2021-12-27 09:16:03 +01:00
|
|
|
|
# fmt: on
|
2017-01-20 18:52:47 +01:00
|
|
|
|
)
|
2021-05-26 13:32:04 +02:00
|
|
|
|
return index_error(output_format, 'No query'), 400
|
2014-02-07 03:15:34 +01:00
|
|
|
|
|
2016-10-22 13:10:31 +02:00
|
|
|
|
# search
|
|
|
|
|
search_query = None
|
2019-07-16 16:27:29 +02:00
|
|
|
|
raw_text_query = None
|
2016-10-22 13:10:31 +02:00
|
|
|
|
result_container = None
|
2014-02-07 01:19:07 +01:00
|
|
|
|
try:
|
2021-12-27 09:26:22 +01:00
|
|
|
|
search_query, raw_text_query, _, _ = get_search_query_from_webapp(request.preferences, request.form)
|
2016-10-22 13:10:31 +02:00
|
|
|
|
# search = Search(search_query) # without plugins
|
2021-05-26 13:32:04 +02:00
|
|
|
|
search = SearchWithPlugins(search_query, request.user_plugins, request) # pylint: disable=redefined-outer-name
|
2020-07-03 15:25:04 +02:00
|
|
|
|
|
2016-10-22 13:10:31 +02:00
|
|
|
|
result_container = search.search()
|
2020-07-03 15:25:04 +02:00
|
|
|
|
|
2020-11-03 11:35:53 +01:00
|
|
|
|
except SearxParameterException as e:
|
|
|
|
|
logger.exception('search error: SearxParameterException')
|
|
|
|
|
return index_error(output_format, e.message), 400
|
2021-05-26 13:32:04 +02:00
|
|
|
|
except Exception as e: # pylint: disable=broad-except
|
|
|
|
|
logger.exception(e, exc_info=True)
|
2020-11-03 11:35:53 +01:00
|
|
|
|
return index_error(output_format, gettext('search error')), 500
|
2017-01-20 18:52:47 +01:00
|
|
|
|
|
|
|
|
|
# results
|
2016-10-22 13:10:31 +02:00
|
|
|
|
results = result_container.get_ordered_results()
|
2017-01-20 18:52:47 +01:00
|
|
|
|
number_of_results = result_container.results_number()
|
|
|
|
|
if number_of_results < result_container.results_length():
|
|
|
|
|
number_of_results = 0
|
2014-02-07 01:19:07 +01:00
|
|
|
|
|
2020-07-03 15:25:04 +02:00
|
|
|
|
# checkin for a external bang
|
|
|
|
|
if result_container.redirect_url:
|
|
|
|
|
return redirect(result_container.redirect_url)
|
|
|
|
|
|
2019-07-17 10:38:45 +02:00
|
|
|
|
# Server-Timing header
|
2021-05-14 11:15:35 +02:00
|
|
|
|
request.timings = result_container.get_timings() # pylint: disable=assigning-non-slot
|
2019-07-17 10:38:45 +02:00
|
|
|
|
|
2016-10-22 13:10:31 +02:00
|
|
|
|
# output
|
2016-07-16 21:41:02 +02:00
|
|
|
|
for result in results:
|
2016-10-22 13:10:31 +02:00
|
|
|
|
if output_format == 'html':
|
2016-09-05 22:22:25 +02:00
|
|
|
|
if 'content' in result and result['content']:
|
2016-11-30 18:43:03 +01:00
|
|
|
|
result['content'] = highlight_content(escape(result['content'][:1024]), search_query.query)
|
2019-09-23 17:14:32 +02:00
|
|
|
|
if 'title' in result and result['title']:
|
2020-08-06 17:42:46 +02:00
|
|
|
|
result['title'] = highlight_content(escape(result['title'] or ''), search_query.query)
|
2014-01-10 23:38:08 +01:00
|
|
|
|
else:
|
2015-04-12 17:37:01 +02:00
|
|
|
|
if result.get('content'):
|
2014-01-10 23:38:08 +01:00
|
|
|
|
result['content'] = html_to_text(result['content']).strip()
|
2014-02-04 19:42:32 +01:00
|
|
|
|
# removing html content and whitespace duplications
|
2015-04-12 17:37:01 +02:00
|
|
|
|
result['title'] = ' '.join(html_to_text(result['title']).strip().split())
|
2014-06-24 16:30:04 +02:00
|
|
|
|
|
2019-09-23 17:14:32 +02:00
|
|
|
|
if 'url' in result:
|
|
|
|
|
result['pretty_url'] = prettify_url(result['url'])
|
2013-11-15 19:28:30 +01:00
|
|
|
|
|
2021-05-26 13:32:04 +02:00
|
|
|
|
# TODO, check if timezone is calculated right # pylint: disable=fixme
|
2021-04-04 12:43:55 +02:00
|
|
|
|
if result.get('publishedDate'): # do not try to get a date from an empty string or a None type
|
2016-03-29 11:59:16 +02:00
|
|
|
|
try: # test if publishedDate >= 1900 (datetime module bug)
|
|
|
|
|
result['pubdate'] = result['publishedDate'].strftime('%Y-%m-%d %H:%M:%S%z')
|
|
|
|
|
except ValueError:
|
|
|
|
|
result['publishedDate'] = None
|
2014-03-14 09:55:04 +01:00
|
|
|
|
else:
|
2016-03-29 11:59:16 +02:00
|
|
|
|
if result['publishedDate'].replace(tzinfo=None) >= datetime.now() - timedelta(days=1):
|
|
|
|
|
timedifference = datetime.now() - result['publishedDate'].replace(tzinfo=None)
|
|
|
|
|
minutes = int((timedifference.seconds / 60) % 60)
|
|
|
|
|
hours = int(timedifference.seconds / 60 / 60)
|
|
|
|
|
if hours == 0:
|
2020-08-06 17:42:46 +02:00
|
|
|
|
result['publishedDate'] = gettext('{minutes} minute(s) ago').format(minutes=minutes)
|
2016-03-29 11:59:16 +02:00
|
|
|
|
else:
|
2021-12-27 09:26:22 +01:00
|
|
|
|
result['publishedDate'] = gettext('{hours} hour(s), {minutes} minute(s) ago').format(
|
|
|
|
|
hours=hours, minutes=minutes
|
|
|
|
|
)
|
2016-03-29 11:59:16 +02:00
|
|
|
|
else:
|
|
|
|
|
result['publishedDate'] = format_date(result['publishedDate'])
|
2014-03-14 09:55:04 +01:00
|
|
|
|
|
2016-10-22 13:10:31 +02:00
|
|
|
|
if output_format == 'json':
|
2021-06-08 18:19:30 +02:00
|
|
|
|
x = {
|
|
|
|
|
'query': search_query.query,
|
|
|
|
|
'number_of_results': number_of_results,
|
|
|
|
|
'results': results,
|
|
|
|
|
'answers': list(result_container.answers),
|
|
|
|
|
'corrections': list(result_container.corrections),
|
|
|
|
|
'infoboxes': result_container.infoboxes,
|
|
|
|
|
'suggestions': list(result_container.suggestions),
|
2021-12-27 09:26:22 +01:00
|
|
|
|
'unresponsive_engines': __get_translated_errors(result_container.unresponsive_engines),
|
2021-06-08 18:19:30 +02:00
|
|
|
|
}
|
2021-12-27 09:26:22 +01:00
|
|
|
|
response = json.dumps(x, default=lambda item: list(item) if isinstance(item, set) else item)
|
2021-06-08 18:19:30 +02:00
|
|
|
|
return Response(response, mimetype='application/json')
|
2021-05-26 13:32:04 +02:00
|
|
|
|
|
|
|
|
|
if output_format == 'csv':
|
2016-11-30 18:43:03 +01:00
|
|
|
|
csv = UnicodeWriter(StringIO())
|
2020-03-13 00:50:19 +01:00
|
|
|
|
keys = ('title', 'url', 'content', 'host', 'engine', 'score', 'type')
|
2015-10-03 17:26:07 +02:00
|
|
|
|
csv.writerow(keys)
|
2016-07-16 21:41:02 +02:00
|
|
|
|
for row in results:
|
2015-10-03 17:26:07 +02:00
|
|
|
|
row['host'] = row['parsed_url'].netloc
|
2020-03-13 00:50:19 +01:00
|
|
|
|
row['type'] = 'result'
|
|
|
|
|
csv.writerow([row.get(key, '') for key in keys])
|
|
|
|
|
for a in result_container.answers:
|
|
|
|
|
row = {'title': a, 'type': 'answer'}
|
|
|
|
|
csv.writerow([row.get(key, '') for key in keys])
|
|
|
|
|
for a in result_container.suggestions:
|
|
|
|
|
row = {'title': a, 'type': 'suggestion'}
|
|
|
|
|
csv.writerow([row.get(key, '') for key in keys])
|
|
|
|
|
for a in result_container.corrections:
|
|
|
|
|
row = {'title': a, 'type': 'correction'}
|
2015-10-03 17:26:07 +02:00
|
|
|
|
csv.writerow([row.get(key, '') for key in keys])
|
|
|
|
|
csv.stream.seek(0)
|
2013-11-15 19:28:30 +01:00
|
|
|
|
response = Response(csv.stream.read(), mimetype='application/csv')
|
2020-08-11 16:25:03 +02:00
|
|
|
|
cont_disp = 'attachment;Filename=searx_-_{0}.csv'.format(search_query.query)
|
2014-02-07 02:45:12 +01:00
|
|
|
|
response.headers.add('Content-Disposition', cont_disp)
|
2013-11-15 18:55:18 +01:00
|
|
|
|
return response
|
2020-07-03 15:25:04 +02:00
|
|
|
|
|
2021-05-26 13:32:04 +02:00
|
|
|
|
if output_format == 'rss':
|
2014-01-19 22:59:01 +01:00
|
|
|
|
response_rss = render(
|
|
|
|
|
'opensearch_response_rss.xml',
|
2016-07-16 21:41:02 +02:00
|
|
|
|
results=results,
|
2020-03-13 00:43:05 +01:00
|
|
|
|
answers=result_container.answers,
|
|
|
|
|
corrections=result_container.corrections,
|
|
|
|
|
suggestions=result_container.suggestions,
|
2016-11-02 14:52:22 +01:00
|
|
|
|
q=request.form['q'],
|
2016-07-16 21:37:40 +02:00
|
|
|
|
number_of_results=number_of_results,
|
2017-01-13 22:15:11 +01:00
|
|
|
|
override_theme='__common__',
|
2014-01-19 22:59:01 +01:00
|
|
|
|
)
|
2014-01-14 22:18:21 +01:00
|
|
|
|
return Response(response_rss, mimetype='text/xml')
|
2014-01-14 18:17:19 +01:00
|
|
|
|
|
2019-07-16 16:27:29 +02:00
|
|
|
|
# HTML output format
|
|
|
|
|
|
|
|
|
|
# suggestions: use RawTextQuery to get the suggestion URLs with the same bang
|
2021-06-08 18:19:30 +02:00
|
|
|
|
suggestion_urls = list(
|
|
|
|
|
map(
|
2021-12-27 09:26:22 +01:00
|
|
|
|
lambda suggestion: {'url': raw_text_query.changeQuery(suggestion).getFullQuery(), 'title': suggestion},
|
|
|
|
|
result_container.suggestions,
|
|
|
|
|
)
|
|
|
|
|
)
|
2021-06-08 18:19:30 +02:00
|
|
|
|
|
|
|
|
|
correction_urls = list(
|
|
|
|
|
map(
|
2021-12-27 09:26:22 +01:00
|
|
|
|
lambda correction: {'url': raw_text_query.changeQuery(correction).getFullQuery(), 'title': correction},
|
|
|
|
|
result_container.corrections,
|
|
|
|
|
)
|
|
|
|
|
)
|
2021-06-08 18:19:30 +02:00
|
|
|
|
|
2014-01-19 22:59:01 +01:00
|
|
|
|
return render(
|
2021-12-27 09:16:03 +01:00
|
|
|
|
# fmt: off
|
2014-01-19 22:59:01 +01:00
|
|
|
|
'results.html',
|
2021-06-08 18:19:30 +02:00
|
|
|
|
results = results,
|
2016-11-02 14:52:22 +01:00
|
|
|
|
q=request.form['q'],
|
2021-06-08 18:19:30 +02:00
|
|
|
|
selected_categories = search_query.categories,
|
|
|
|
|
pageno = search_query.pageno,
|
|
|
|
|
time_range = search_query.time_range,
|
|
|
|
|
number_of_results = format_decimal(number_of_results),
|
|
|
|
|
suggestions = suggestion_urls,
|
|
|
|
|
answers = result_container.answers,
|
|
|
|
|
corrections = correction_urls,
|
|
|
|
|
infoboxes = result_container.infoboxes,
|
|
|
|
|
engine_data = result_container.engine_data,
|
|
|
|
|
paging = result_container.paging,
|
|
|
|
|
unresponsive_engines = __get_translated_errors(
|
|
|
|
|
result_container.unresponsive_engines
|
|
|
|
|
),
|
|
|
|
|
current_language = match_language(
|
|
|
|
|
search_query.lang,
|
2021-10-21 10:41:57 +02:00
|
|
|
|
settings['search']['languages'],
|
2021-06-08 18:19:30 +02:00
|
|
|
|
fallback=request.preferences.get_value("language")
|
|
|
|
|
),
|
|
|
|
|
theme = get_current_theme_name(),
|
|
|
|
|
favicons = global_favicons[themes.index(get_current_theme_name())],
|
|
|
|
|
timeout_limit = request.form.get('timeout_limit', None)
|
2021-12-27 09:16:03 +01:00
|
|
|
|
# fmt: on
|
2014-01-19 22:59:01 +01:00
|
|
|
|
)
|
2014-01-01 22:16:53 +01:00
|
|
|
|
|
2013-10-14 23:09:13 +02:00
|
|
|
|
|
2022-01-17 11:11:39 +01:00
|
|
|
|
def __get_translated_errors(unresponsive_engines: Iterable[UnresponsiveEngine]):
|
2021-04-26 11:12:02 +02:00
|
|
|
|
translated_errors = []
|
2021-06-08 18:19:30 +02:00
|
|
|
|
|
|
|
|
|
# make a copy unresponsive_engines to avoid "RuntimeError: Set changed size
|
|
|
|
|
# during iteration" it happens when an engine modifies the ResultContainer
|
|
|
|
|
# after the search_multiple_requests method has stopped waiting
|
|
|
|
|
|
2022-01-17 11:11:39 +01:00
|
|
|
|
for unresponsive_engine in unresponsive_engines:
|
|
|
|
|
error_user_text = exception_classname_to_text.get(unresponsive_engine.error_type)
|
2021-04-26 11:12:02 +02:00
|
|
|
|
if not error_user_text:
|
|
|
|
|
error_user_text = exception_classname_to_text[None]
|
|
|
|
|
error_msg = gettext(error_user_text)
|
2022-01-17 11:11:39 +01:00
|
|
|
|
if unresponsive_engine.suspended:
|
2021-04-13 15:21:53 +02:00
|
|
|
|
error_msg = gettext('Suspended') + ': ' + error_msg
|
2022-01-17 11:11:39 +01:00
|
|
|
|
translated_errors.append((unresponsive_engine.engine, error_msg))
|
2021-06-08 18:19:30 +02:00
|
|
|
|
|
2021-04-26 11:12:02 +02:00
|
|
|
|
return sorted(translated_errors, key=lambda e: e[0])
|
2020-04-17 16:31:02 +02:00
|
|
|
|
|
|
|
|
|
|
2013-10-21 00:28:48 +02:00
|
|
|
|
@app.route('/about', methods=['GET'])
|
|
|
|
|
def about():
|
2022-01-31 11:24:45 +01:00
|
|
|
|
"""Redirect to about page"""
|
|
|
|
|
return redirect(url_for('help_page', pagename='about'))
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
@app.route('/help/en/<pagename>', methods=['GET'])
|
|
|
|
|
def help_page(pagename):
|
|
|
|
|
"""Render help page"""
|
|
|
|
|
page = user_help.PAGES.get(pagename)
|
|
|
|
|
|
|
|
|
|
if page is None:
|
|
|
|
|
flask.abort(404)
|
|
|
|
|
|
2022-01-31 12:42:04 +01:00
|
|
|
|
return render(
|
|
|
|
|
'help.html', page=user_help.PAGES[pagename], all_pages=user_help.PAGES.items(), page_filename=pagename
|
|
|
|
|
)
|
2014-01-17 16:23:23 +01:00
|
|
|
|
|
|
|
|
|
|
2014-03-20 10:28:24 +01:00
|
|
|
|
@app.route('/autocompleter', methods=['GET', 'POST'])
|
|
|
|
|
def autocompleter():
|
|
|
|
|
"""Return autocompleter results"""
|
2014-03-29 16:29:19 +01:00
|
|
|
|
|
2021-02-22 18:13:50 +01:00
|
|
|
|
# run autocompleter
|
|
|
|
|
results = []
|
|
|
|
|
|
2014-10-01 17:18:18 +02:00
|
|
|
|
# set blocked engines
|
2016-04-09 18:26:29 +02:00
|
|
|
|
disabled_engines = request.preferences.engines.get_disabled()
|
2014-10-01 17:18:18 +02:00
|
|
|
|
|
|
|
|
|
# parse query
|
2020-09-21 19:14:24 +02:00
|
|
|
|
raw_text_query = RawTextQuery(request.form.get('q', ''), disabled_engines)
|
[fix] url bar autocomplete (opensearch suggestions)
Since #2593 is merged the OpenSearch-Format is buggy. The loop in [1] will
change raw_text_query object and this will change also the value of
`raw_text_query.query` on every `raw_text_query.changeQuery(result)`.
This patch fixes this issue by storing the initial query value in `sug_prefix`.
[1] https://github.com/searx/searx/blob/ac0fdc3b9670168de8061ed0e656dd66a567d741/searx/webapp.py#L804-L806
OpenSearch-Format::
[ "<query>",
[ "<term 1>", "<term 2>", ... "<term n>" ],
[ "<content 1>", "<content 2>", ..., "<content n>" ],
[ "<url 1>", "<url 2>", ..., "<url n>" ]
]
- https://www.google.com/support/enterprise/static/gsa/docs/admin/current/gsa_doc_set/xml_reference/query_suggestion.html#1080002
- https://developer.mozilla.org/en-US/docs/Archive/Add-ons/Supporting_search_suggestions_in_search_plugins#implementing_search_suggestion_support_on_the_server
Legacy-Format::
[ "<term 1>", "<term 2>", ..., "<term n>" ]
- https://www.google.com/support/enterprise/static/gsa/docs/admin/current/gsa_doc_set/xml_reference/query_suggestion.html#1081079
Signed-off-by: Markus Heiser <markus.heiser@darmarit.de>
2021-04-03 18:18:50 +02:00
|
|
|
|
sug_prefix = raw_text_query.getQuery()
|
2014-03-29 16:29:19 +01:00
|
|
|
|
|
2018-01-19 03:51:27 +01:00
|
|
|
|
# normal autocompletion results only appear if no inner results returned
|
2021-02-22 18:13:50 +01:00
|
|
|
|
# and there is a query part
|
[fix] url bar autocomplete (opensearch suggestions)
Since #2593 is merged the OpenSearch-Format is buggy. The loop in [1] will
change raw_text_query object and this will change also the value of
`raw_text_query.query` on every `raw_text_query.changeQuery(result)`.
This patch fixes this issue by storing the initial query value in `sug_prefix`.
[1] https://github.com/searx/searx/blob/ac0fdc3b9670168de8061ed0e656dd66a567d741/searx/webapp.py#L804-L806
OpenSearch-Format::
[ "<query>",
[ "<term 1>", "<term 2>", ... "<term n>" ],
[ "<content 1>", "<content 2>", ..., "<content n>" ],
[ "<url 1>", "<url 2>", ..., "<url n>" ]
]
- https://www.google.com/support/enterprise/static/gsa/docs/admin/current/gsa_doc_set/xml_reference/query_suggestion.html#1080002
- https://developer.mozilla.org/en-US/docs/Archive/Add-ons/Supporting_search_suggestions_in_search_plugins#implementing_search_suggestion_support_on_the_server
Legacy-Format::
[ "<term 1>", "<term 2>", ..., "<term n>" ]
- https://www.google.com/support/enterprise/static/gsa/docs/admin/current/gsa_doc_set/xml_reference/query_suggestion.html#1081079
Signed-off-by: Markus Heiser <markus.heiser@darmarit.de>
2021-04-03 18:18:50 +02:00
|
|
|
|
if len(raw_text_query.autocomplete_list) == 0 and len(sug_prefix) > 0:
|
|
|
|
|
|
2016-03-30 02:53:31 +02:00
|
|
|
|
# get language from cookie
|
2016-04-09 18:32:07 +02:00
|
|
|
|
language = request.preferences.get_value('language')
|
2019-01-06 15:27:46 +01:00
|
|
|
|
if not language or language == 'all':
|
|
|
|
|
language = 'en'
|
2016-03-30 02:53:31 +02:00
|
|
|
|
else:
|
2016-08-06 06:34:56 +02:00
|
|
|
|
language = language.split('-')[0]
|
[fix] url bar autocomplete (opensearch suggestions)
Since #2593 is merged the OpenSearch-Format is buggy. The loop in [1] will
change raw_text_query object and this will change also the value of
`raw_text_query.query` on every `raw_text_query.changeQuery(result)`.
This patch fixes this issue by storing the initial query value in `sug_prefix`.
[1] https://github.com/searx/searx/blob/ac0fdc3b9670168de8061ed0e656dd66a567d741/searx/webapp.py#L804-L806
OpenSearch-Format::
[ "<query>",
[ "<term 1>", "<term 2>", ... "<term n>" ],
[ "<content 1>", "<content 2>", ..., "<content n>" ],
[ "<url 1>", "<url 2>", ..., "<url n>" ]
]
- https://www.google.com/support/enterprise/static/gsa/docs/admin/current/gsa_doc_set/xml_reference/query_suggestion.html#1080002
- https://developer.mozilla.org/en-US/docs/Archive/Add-ons/Supporting_search_suggestions_in_search_plugins#implementing_search_suggestion_support_on_the_server
Legacy-Format::
[ "<term 1>", "<term 2>", ..., "<term n>" ]
- https://www.google.com/support/enterprise/static/gsa/docs/admin/current/gsa_doc_set/xml_reference/query_suggestion.html#1081079
Signed-off-by: Markus Heiser <markus.heiser@darmarit.de>
2021-04-03 18:18:50 +02:00
|
|
|
|
|
2015-01-10 19:55:21 +01:00
|
|
|
|
# run autocompletion
|
2021-12-27 09:26:22 +01:00
|
|
|
|
raw_results = search_autocomplete(request.preferences.get_value('autocomplete'), sug_prefix, language)
|
2021-02-22 18:13:50 +01:00
|
|
|
|
for result in raw_results:
|
[fix] url bar autocomplete (opensearch suggestions)
Since #2593 is merged the OpenSearch-Format is buggy. The loop in [1] will
change raw_text_query object and this will change also the value of
`raw_text_query.query` on every `raw_text_query.changeQuery(result)`.
This patch fixes this issue by storing the initial query value in `sug_prefix`.
[1] https://github.com/searx/searx/blob/ac0fdc3b9670168de8061ed0e656dd66a567d741/searx/webapp.py#L804-L806
OpenSearch-Format::
[ "<query>",
[ "<term 1>", "<term 2>", ... "<term n>" ],
[ "<content 1>", "<content 2>", ..., "<content n>" ],
[ "<url 1>", "<url 2>", ..., "<url n>" ]
]
- https://www.google.com/support/enterprise/static/gsa/docs/admin/current/gsa_doc_set/xml_reference/query_suggestion.html#1080002
- https://developer.mozilla.org/en-US/docs/Archive/Add-ons/Supporting_search_suggestions_in_search_plugins#implementing_search_suggestion_support_on_the_server
Legacy-Format::
[ "<term 1>", "<term 2>", ..., "<term n>" ]
- https://www.google.com/support/enterprise/static/gsa/docs/admin/current/gsa_doc_set/xml_reference/query_suggestion.html#1081079
Signed-off-by: Markus Heiser <markus.heiser@darmarit.de>
2021-04-03 18:18:50 +02:00
|
|
|
|
# attention: this loop will change raw_text_query object and this is
|
|
|
|
|
# the reason why the sug_prefix was stored before (see above)
|
2021-02-22 18:13:50 +01:00
|
|
|
|
results.append(raw_text_query.changeQuery(result).getFullQuery())
|
|
|
|
|
|
|
|
|
|
if len(raw_text_query.autocomplete_list) > 0:
|
|
|
|
|
for autocomplete_text in raw_text_query.autocomplete_list:
|
|
|
|
|
results.append(raw_text_query.get_autocomplete_full_query(autocomplete_text))
|
|
|
|
|
|
|
|
|
|
for answers in ask(raw_text_query):
|
|
|
|
|
for answer in answers:
|
|
|
|
|
results.append(str(answer['answer']))
|
2014-03-29 16:29:19 +01:00
|
|
|
|
|
2020-08-09 13:59:49 +02:00
|
|
|
|
if request.headers.get('X-Requested-With') == 'XMLHttpRequest':
|
[fix] url bar autocomplete (opensearch suggestions)
Since #2593 is merged the OpenSearch-Format is buggy. The loop in [1] will
change raw_text_query object and this will change also the value of
`raw_text_query.query` on every `raw_text_query.changeQuery(result)`.
This patch fixes this issue by storing the initial query value in `sug_prefix`.
[1] https://github.com/searx/searx/blob/ac0fdc3b9670168de8061ed0e656dd66a567d741/searx/webapp.py#L804-L806
OpenSearch-Format::
[ "<query>",
[ "<term 1>", "<term 2>", ... "<term n>" ],
[ "<content 1>", "<content 2>", ..., "<content n>" ],
[ "<url 1>", "<url 2>", ..., "<url n>" ]
]
- https://www.google.com/support/enterprise/static/gsa/docs/admin/current/gsa_doc_set/xml_reference/query_suggestion.html#1080002
- https://developer.mozilla.org/en-US/docs/Archive/Add-ons/Supporting_search_suggestions_in_search_plugins#implementing_search_suggestion_support_on_the_server
Legacy-Format::
[ "<term 1>", "<term 2>", ..., "<term n>" ]
- https://www.google.com/support/enterprise/static/gsa/docs/admin/current/gsa_doc_set/xml_reference/query_suggestion.html#1081079
Signed-off-by: Markus Heiser <markus.heiser@darmarit.de>
2021-04-03 18:18:50 +02:00
|
|
|
|
# the suggestion request comes from the searx search form
|
|
|
|
|
suggestions = json.dumps(results)
|
|
|
|
|
mimetype = 'application/json'
|
|
|
|
|
else:
|
|
|
|
|
# the suggestion request comes from browser's URL bar
|
|
|
|
|
suggestions = json.dumps([sug_prefix, results])
|
|
|
|
|
mimetype = 'application/x-suggestions+json'
|
2015-01-25 22:52:48 +01:00
|
|
|
|
|
2021-11-28 13:42:35 +01:00
|
|
|
|
if get_current_theme_name() == 'simple':
|
|
|
|
|
suggestions = escape(suggestions, False)
|
[fix] url bar autocomplete (opensearch suggestions)
Since #2593 is merged the OpenSearch-Format is buggy. The loop in [1] will
change raw_text_query object and this will change also the value of
`raw_text_query.query` on every `raw_text_query.changeQuery(result)`.
This patch fixes this issue by storing the initial query value in `sug_prefix`.
[1] https://github.com/searx/searx/blob/ac0fdc3b9670168de8061ed0e656dd66a567d741/searx/webapp.py#L804-L806
OpenSearch-Format::
[ "<query>",
[ "<term 1>", "<term 2>", ... "<term n>" ],
[ "<content 1>", "<content 2>", ..., "<content n>" ],
[ "<url 1>", "<url 2>", ..., "<url n>" ]
]
- https://www.google.com/support/enterprise/static/gsa/docs/admin/current/gsa_doc_set/xml_reference/query_suggestion.html#1080002
- https://developer.mozilla.org/en-US/docs/Archive/Add-ons/Supporting_search_suggestions_in_search_plugins#implementing_search_suggestion_support_on_the_server
Legacy-Format::
[ "<term 1>", "<term 2>", ..., "<term n>" ]
- https://www.google.com/support/enterprise/static/gsa/docs/admin/current/gsa_doc_set/xml_reference/query_suggestion.html#1081079
Signed-off-by: Markus Heiser <markus.heiser@darmarit.de>
2021-04-03 18:18:50 +02:00
|
|
|
|
return Response(suggestions, mimetype=mimetype)
|
2014-03-20 10:28:24 +01:00
|
|
|
|
|
|
|
|
|
|
2014-01-01 22:16:53 +01:00
|
|
|
|
@app.route('/preferences', methods=['GET', 'POST'])
|
|
|
|
|
def preferences():
|
2016-04-08 16:38:05 +02:00
|
|
|
|
"""Render preferences page && save user preferences"""
|
2014-02-07 00:35:15 +01:00
|
|
|
|
|
2021-05-26 13:32:04 +02:00
|
|
|
|
# pylint: disable=too-many-locals, too-many-return-statements, too-many-branches
|
|
|
|
|
# pylint: disable=too-many-statements
|
|
|
|
|
|
2016-04-08 16:38:05 +02:00
|
|
|
|
# save preferences
|
|
|
|
|
if request.method == 'POST':
|
2021-04-12 15:29:08 +02:00
|
|
|
|
resp = make_response(redirect(url_for('index', _external=True)))
|
2016-04-08 16:38:05 +02:00
|
|
|
|
try:
|
|
|
|
|
request.preferences.parse_form(request.form)
|
|
|
|
|
except ValidationException:
|
2016-11-14 22:07:23 +01:00
|
|
|
|
request.errors.append(gettext('Invalid settings, please edit your preferences'))
|
2016-04-08 16:38:05 +02:00
|
|
|
|
return resp
|
|
|
|
|
return request.preferences.save(resp)
|
|
|
|
|
|
|
|
|
|
# render preferences
|
2021-05-26 13:32:04 +02:00
|
|
|
|
image_proxy = request.preferences.get_value('image_proxy') # pylint: disable=redefined-outer-name
|
2016-04-09 18:26:29 +02:00
|
|
|
|
disabled_engines = request.preferences.engines.get_disabled()
|
2016-04-08 16:38:05 +02:00
|
|
|
|
allowed_plugins = request.preferences.plugins.get_enabled()
|
2015-05-30 12:15:23 +02:00
|
|
|
|
|
|
|
|
|
# stats for preferences page
|
2022-01-09 15:53:58 +01:00
|
|
|
|
filtered_engines = dict(filter(lambda kv: request.preferences.validate_token(kv[1]), engines.items()))
|
2015-05-30 12:15:23 +02:00
|
|
|
|
|
2020-02-01 11:01:17 +01:00
|
|
|
|
engines_by_category = {}
|
2021-07-03 17:51:39 +02:00
|
|
|
|
|
2021-12-27 09:26:22 +01:00
|
|
|
|
for c in categories: # pylint: disable=consider-using-dict-items
|
2021-04-14 17:23:15 +02:00
|
|
|
|
engines_by_category[c] = [e for e in categories[c] if e.name in filtered_engines]
|
2021-04-14 18:11:35 +02:00
|
|
|
|
# sort the engines alphabetically since the order in settings.yml is meaningless.
|
|
|
|
|
list.sort(engines_by_category[c], key=lambda e: e.name)
|
2020-02-01 11:01:17 +01:00
|
|
|
|
|
2016-11-05 13:45:20 +01:00
|
|
|
|
# get first element [0], the engine time,
|
|
|
|
|
# and then the second element [1] : the time (the first one is the label)
|
2021-05-26 13:32:04 +02:00
|
|
|
|
stats = {} # pylint: disable=redefined-outer-name
|
2021-04-14 18:11:35 +02:00
|
|
|
|
max_rate95 = 0
|
2021-04-14 17:23:15 +02:00
|
|
|
|
for _, e in filtered_engines.items():
|
|
|
|
|
h = histogram('engine', e.name, 'time', 'total')
|
|
|
|
|
median = round(h.percentage(50), 1) if h.count > 0 else None
|
2021-04-14 18:11:35 +02:00
|
|
|
|
rate80 = round(h.percentage(80), 1) if h.count > 0 else None
|
|
|
|
|
rate95 = round(h.percentage(95), 1) if h.count > 0 else None
|
|
|
|
|
|
|
|
|
|
max_rate95 = max(max_rate95, rate95 or 0)
|
|
|
|
|
|
|
|
|
|
result_count_sum = histogram('engine', e.name, 'result', 'count').sum
|
|
|
|
|
successful_count = counter('engine', e.name, 'search', 'count', 'successful')
|
|
|
|
|
result_count = int(result_count_sum / float(successful_count)) if successful_count else 0
|
2021-04-14 17:23:15 +02:00
|
|
|
|
|
|
|
|
|
stats[e.name] = {
|
2021-05-19 11:24:18 +02:00
|
|
|
|
'time': median,
|
|
|
|
|
'rate80': rate80,
|
|
|
|
|
'rate95': rate95,
|
2021-04-14 17:23:15 +02:00
|
|
|
|
'warn_timeout': e.timeout > settings['outgoing']['request_timeout'],
|
2021-04-14 18:11:35 +02:00
|
|
|
|
'supports_selected_language': _is_selected_language_supported(e, request.preferences),
|
|
|
|
|
'result_count': result_count,
|
2021-04-14 17:23:15 +02:00
|
|
|
|
}
|
2015-05-30 12:15:23 +02:00
|
|
|
|
# end of stats
|
|
|
|
|
|
2021-04-14 18:11:35 +02:00
|
|
|
|
# reliabilities
|
|
|
|
|
reliabilities = {}
|
|
|
|
|
engine_errors = get_engine_errors(filtered_engines)
|
|
|
|
|
checker_results = checker_get_result()
|
2021-12-27 09:26:22 +01:00
|
|
|
|
checker_results = (
|
|
|
|
|
checker_results['engines'] if checker_results['status'] == 'ok' and 'engines' in checker_results else {}
|
|
|
|
|
)
|
2021-04-14 18:11:35 +02:00
|
|
|
|
for _, e in filtered_engines.items():
|
|
|
|
|
checker_result = checker_results.get(e.name, {})
|
|
|
|
|
checker_success = checker_result.get('success', True)
|
|
|
|
|
errors = engine_errors.get(e.name) or []
|
|
|
|
|
if counter('engine', e.name, 'search', 'count', 'sent') == 0:
|
|
|
|
|
# no request
|
|
|
|
|
reliablity = None
|
|
|
|
|
elif checker_success and not errors:
|
|
|
|
|
reliablity = 100
|
|
|
|
|
elif 'simple' in checker_result.get('errors', {}):
|
|
|
|
|
# the basic (simple) test doesn't work: the engine is broken accoding to the checker
|
|
|
|
|
# even if there is no exception
|
|
|
|
|
reliablity = 0
|
|
|
|
|
else:
|
|
|
|
|
reliablity = 100 - sum([error['percentage'] for error in errors if not error.get('secondary')])
|
|
|
|
|
|
|
|
|
|
reliabilities[e.name] = {
|
|
|
|
|
'reliablity': reliablity,
|
|
|
|
|
'errors': [],
|
|
|
|
|
'checker': checker_results.get(e.name, {}).get('errors', {}).keys(),
|
|
|
|
|
}
|
|
|
|
|
# keep the order of the list checker_results[e.name]['errors'] and deduplicate.
|
|
|
|
|
# the first element has the highest percentage rate.
|
|
|
|
|
reliabilities_errors = []
|
|
|
|
|
for error in errors:
|
2021-04-26 11:12:02 +02:00
|
|
|
|
error_user_text = None
|
2021-04-14 18:11:35 +02:00
|
|
|
|
if error.get('secondary') or 'exception_classname' not in error:
|
|
|
|
|
continue
|
2021-04-26 11:12:02 +02:00
|
|
|
|
error_user_text = exception_classname_to_text.get(error.get('exception_classname'))
|
2021-04-14 18:11:35 +02:00
|
|
|
|
if not error:
|
2021-04-26 11:12:02 +02:00
|
|
|
|
error_user_text = exception_classname_to_text[None]
|
|
|
|
|
if error_user_text not in reliabilities_errors:
|
|
|
|
|
reliabilities_errors.append(error_user_text)
|
2021-04-14 18:11:35 +02:00
|
|
|
|
reliabilities[e.name]['errors'] = reliabilities_errors
|
|
|
|
|
|
|
|
|
|
# supports
|
|
|
|
|
supports = {}
|
|
|
|
|
for _, e in filtered_engines.items():
|
|
|
|
|
supports_selected_language = _is_selected_language_supported(e, request.preferences)
|
|
|
|
|
safesearch = e.safesearch
|
|
|
|
|
time_range_support = e.time_range_support
|
|
|
|
|
for checker_test_name in checker_results.get(e.name, {}).get('errors', {}):
|
|
|
|
|
if supports_selected_language and checker_test_name.startswith('lang_'):
|
|
|
|
|
supports_selected_language = '?'
|
|
|
|
|
elif safesearch and checker_test_name == 'safesearch':
|
|
|
|
|
safesearch = '?'
|
|
|
|
|
elif time_range_support and checker_test_name == 'time_range':
|
|
|
|
|
time_range_support = '?'
|
|
|
|
|
supports[e.name] = {
|
|
|
|
|
'supports_selected_language': supports_selected_language,
|
|
|
|
|
'safesearch': safesearch,
|
|
|
|
|
'time_range_support': time_range_support,
|
|
|
|
|
}
|
|
|
|
|
|
2021-06-08 18:19:30 +02:00
|
|
|
|
return render(
|
2021-12-27 09:16:03 +01:00
|
|
|
|
# fmt: off
|
2021-06-08 18:19:30 +02:00
|
|
|
|
'preferences.html',
|
|
|
|
|
selected_categories = get_selected_categories(request.preferences, request.form),
|
2021-08-03 15:13:00 +02:00
|
|
|
|
locales = LOCALE_NAMES,
|
2021-06-08 18:19:30 +02:00
|
|
|
|
current_locale = request.preferences.get_value("locale"),
|
|
|
|
|
image_proxy = image_proxy,
|
|
|
|
|
engines_by_category = engines_by_category,
|
|
|
|
|
stats = stats,
|
|
|
|
|
max_rate95 = max_rate95,
|
|
|
|
|
reliabilities = reliabilities,
|
|
|
|
|
supports = supports,
|
|
|
|
|
answerers = [
|
|
|
|
|
{'info': a.self_info(), 'keywords': a.keywords}
|
|
|
|
|
for a in answerers
|
|
|
|
|
],
|
|
|
|
|
disabled_engines = disabled_engines,
|
|
|
|
|
autocomplete_backends = autocomplete_backends,
|
|
|
|
|
shortcuts = {y: x for x, y in engine_shortcuts.items()},
|
|
|
|
|
themes = themes,
|
|
|
|
|
plugins = plugins,
|
|
|
|
|
doi_resolvers = settings['doi_resolvers'],
|
2021-09-06 08:47:11 +02:00
|
|
|
|
current_doi_resolver = get_doi_resolver(request.preferences),
|
2021-06-08 18:19:30 +02:00
|
|
|
|
allowed_plugins = allowed_plugins,
|
|
|
|
|
theme = get_current_theme_name(),
|
|
|
|
|
preferences_url_params = request.preferences.get_as_url_params(),
|
|
|
|
|
locked_preferences = settings['preferences']['lock'],
|
|
|
|
|
preferences = True
|
2021-12-27 09:16:03 +01:00
|
|
|
|
# fmt: on
|
2021-06-08 18:19:30 +02:00
|
|
|
|
)
|
2014-01-01 22:16:53 +01:00
|
|
|
|
|
|
|
|
|
|
2022-01-17 11:16:11 +01:00
|
|
|
|
def _is_selected_language_supported(engine, preferences: Preferences): # pylint: disable=redefined-outer-name
|
2019-01-06 15:27:46 +01:00
|
|
|
|
language = preferences.get_value('language')
|
2021-06-08 18:19:30 +02:00
|
|
|
|
if language == 'all':
|
|
|
|
|
return True
|
|
|
|
|
x = match_language(
|
2021-12-27 09:26:22 +01:00
|
|
|
|
language, getattr(engine, 'supported_languages', []), getattr(engine, 'language_aliases', {}), None
|
2021-06-08 18:19:30 +02:00
|
|
|
|
)
|
|
|
|
|
return bool(x)
|
2019-01-06 15:27:46 +01:00
|
|
|
|
|
|
|
|
|
|
2015-01-16 16:02:21 +01:00
|
|
|
|
@app.route('/image_proxy', methods=['GET'])
|
|
|
|
|
def image_proxy():
|
2021-08-14 19:36:30 +02:00
|
|
|
|
# pylint: disable=too-many-return-statements, too-many-branches
|
2021-05-26 13:32:04 +02:00
|
|
|
|
|
2021-03-18 19:59:01 +01:00
|
|
|
|
url = request.args.get('url')
|
2015-01-16 16:02:21 +01:00
|
|
|
|
if not url:
|
|
|
|
|
return '', 400
|
|
|
|
|
|
2021-12-28 13:44:28 +01:00
|
|
|
|
if not is_hmac_of(settings['server']['secret_key'], url.encode(), request.args.get('h', '')):
|
2015-01-17 21:54:40 +01:00
|
|
|
|
return '', 400
|
|
|
|
|
|
2021-03-18 19:59:01 +01:00
|
|
|
|
maximum_size = 5 * 1024 * 1024
|
2021-08-14 19:36:30 +02:00
|
|
|
|
forward_resp = False
|
|
|
|
|
resp = None
|
2021-03-18 19:59:01 +01:00
|
|
|
|
try:
|
2021-08-14 19:36:30 +02:00
|
|
|
|
request_headers = {
|
|
|
|
|
'User-Agent': gen_useragent(),
|
|
|
|
|
'Accept': 'image/webp,*/*',
|
|
|
|
|
'Accept-Encoding': 'gzip, deflate',
|
|
|
|
|
'Sec-GPC': '1',
|
|
|
|
|
'DNT': '1',
|
|
|
|
|
}
|
2021-08-14 20:12:11 +02:00
|
|
|
|
set_context_network_name('image_proxy')
|
2022-02-19 14:26:58 +01:00
|
|
|
|
resp, stream = http_stream(method='GET', url=url, headers=request_headers, allow_redirects=True)
|
2021-03-18 19:59:01 +01:00
|
|
|
|
content_length = resp.headers.get('Content-Length')
|
2021-12-27 09:26:22 +01:00
|
|
|
|
if content_length and content_length.isdigit() and int(content_length) > maximum_size:
|
2021-03-18 19:59:01 +01:00
|
|
|
|
return 'Max size', 400
|
|
|
|
|
|
|
|
|
|
if resp.status_code != 200:
|
2021-08-14 19:36:30 +02:00
|
|
|
|
logger.debug('image-proxy: wrong response code: %i', resp.status_code)
|
2021-03-18 19:59:01 +01:00
|
|
|
|
if resp.status_code >= 400:
|
|
|
|
|
return '', resp.status_code
|
|
|
|
|
return '', 400
|
|
|
|
|
|
2021-08-14 19:36:30 +02:00
|
|
|
|
if not resp.headers.get('Content-Type', '').startswith('image/'):
|
|
|
|
|
logger.debug('image-proxy: wrong content-type: %s', resp.headers.get('Content-Type', ''))
|
2021-03-18 19:59:01 +01:00
|
|
|
|
return '', 400
|
2015-01-16 16:02:21 +01:00
|
|
|
|
|
2021-08-14 19:36:30 +02:00
|
|
|
|
forward_resp = True
|
|
|
|
|
except httpx.HTTPError:
|
|
|
|
|
logger.exception('HTTP error')
|
|
|
|
|
return '', 400
|
|
|
|
|
finally:
|
|
|
|
|
if resp and not forward_resp:
|
|
|
|
|
# the code is about to return an HTTP 400 error to the browser
|
|
|
|
|
# we make sure to close the response between searxng and the HTTP server
|
|
|
|
|
try:
|
|
|
|
|
resp.close()
|
|
|
|
|
except httpx.HTTPError:
|
|
|
|
|
logger.exception('HTTP error on closing')
|
|
|
|
|
|
2021-09-28 17:44:00 +02:00
|
|
|
|
def close_stream():
|
|
|
|
|
nonlocal resp, stream
|
|
|
|
|
try:
|
2022-01-24 11:44:18 +01:00
|
|
|
|
if resp:
|
|
|
|
|
resp.close()
|
2021-09-28 17:44:00 +02:00
|
|
|
|
del resp
|
|
|
|
|
del stream
|
|
|
|
|
except httpx.HTTPError as e:
|
|
|
|
|
logger.debug('Exception while closing response', e)
|
|
|
|
|
|
2021-08-14 19:36:30 +02:00
|
|
|
|
try:
|
2021-12-27 09:26:22 +01:00
|
|
|
|
headers = dict_subset(resp.headers, {'Content-Type', 'Content-Encoding', 'Content-Length', 'Length'})
|
|
|
|
|
response = Response(stream, mimetype=resp.headers['Content-Type'], headers=headers, direct_passthrough=True)
|
2021-09-28 17:44:00 +02:00
|
|
|
|
response.call_on_close(close_stream)
|
|
|
|
|
return response
|
2021-03-18 19:59:01 +01:00
|
|
|
|
except httpx.HTTPError:
|
2021-09-28 17:44:00 +02:00
|
|
|
|
close_stream()
|
2021-03-18 19:59:01 +01:00
|
|
|
|
return '', 400
|
2021-09-18 10:59:56 +02:00
|
|
|
|
|
|
|
|
|
|
|
|
|
|
@app.route('/engine_descriptions.json', methods=['GET'])
|
|
|
|
|
def engine_descriptions():
|
|
|
|
|
locale = get_locale().split('_')[0]
|
|
|
|
|
result = ENGINE_DESCRIPTIONS['en'].copy()
|
|
|
|
|
if locale != 'en':
|
|
|
|
|
for engine, description in ENGINE_DESCRIPTIONS.get(locale, {}).items():
|
|
|
|
|
result[engine] = description
|
|
|
|
|
for engine, description in result.items():
|
2021-12-27 09:26:22 +01:00
|
|
|
|
if len(description) == 2 and description[1] == 'ref':
|
2021-09-18 10:59:56 +02:00
|
|
|
|
ref_engine, ref_lang = description[0].split(':')
|
|
|
|
|
description = ENGINE_DESCRIPTIONS[ref_lang][ref_engine]
|
|
|
|
|
if isinstance(description, str):
|
2021-12-27 09:26:22 +01:00
|
|
|
|
description = [description, 'wikipedia']
|
2021-09-18 10:59:56 +02:00
|
|
|
|
result[engine] = description
|
2022-01-23 20:55:14 +01:00
|
|
|
|
|
|
|
|
|
# overwrite by about:description (from settings)
|
|
|
|
|
for engine_name, engine_mod in engines.items():
|
|
|
|
|
descr = getattr(engine_mod, 'about', {}).get('description', None)
|
|
|
|
|
if descr is not None:
|
|
|
|
|
result[engine_name] = [descr, "SearXNG config"]
|
|
|
|
|
|
2021-09-18 10:59:56 +02:00
|
|
|
|
return jsonify(result)
|
2015-01-16 16:02:21 +01:00
|
|
|
|
|
|
|
|
|
|
2013-10-27 01:03:05 +02:00
|
|
|
|
@app.route('/stats', methods=['GET'])
|
|
|
|
|
def stats():
|
2014-01-31 07:08:24 +01:00
|
|
|
|
"""Render engine statistics page."""
|
2021-04-23 21:08:48 +02:00
|
|
|
|
sort_order = request.args.get('sort', default='name', type=str)
|
|
|
|
|
selected_engine_name = request.args.get('engine', default=None, type=str)
|
|
|
|
|
|
2022-01-09 15:53:58 +01:00
|
|
|
|
filtered_engines = dict(filter(lambda kv: request.preferences.validate_token(kv[1]), engines.items()))
|
2021-04-23 21:08:48 +02:00
|
|
|
|
if selected_engine_name:
|
|
|
|
|
if selected_engine_name not in filtered_engines:
|
|
|
|
|
selected_engine_name = None
|
|
|
|
|
else:
|
|
|
|
|
filtered_engines = [selected_engine_name]
|
|
|
|
|
|
2021-04-22 17:47:53 +02:00
|
|
|
|
checker_results = checker_get_result()
|
2021-06-08 18:19:30 +02:00
|
|
|
|
checker_results = (
|
2021-12-27 09:26:22 +01:00
|
|
|
|
checker_results['engines'] if checker_results['status'] == 'ok' and 'engines' in checker_results else {}
|
2021-06-08 18:19:30 +02:00
|
|
|
|
)
|
2021-04-22 17:47:53 +02:00
|
|
|
|
|
2021-04-14 17:23:15 +02:00
|
|
|
|
engine_stats = get_engines_stats(filtered_engines)
|
2021-04-22 17:47:53 +02:00
|
|
|
|
engine_reliabilities = get_reliabilities(filtered_engines, checker_results)
|
|
|
|
|
|
2021-05-26 13:32:04 +02:00
|
|
|
|
if sort_order not in STATS_SORT_PARAMETERS:
|
2021-04-22 17:47:53 +02:00
|
|
|
|
sort_order = 'name'
|
|
|
|
|
|
2021-05-26 13:32:04 +02:00
|
|
|
|
reverse, key_name, default_value = STATS_SORT_PARAMETERS[sort_order]
|
2021-04-22 17:47:53 +02:00
|
|
|
|
|
|
|
|
|
def get_key(engine_stat):
|
2022-01-24 11:44:18 +01:00
|
|
|
|
reliability = engine_reliabilities.get(engine_stat['name'], {}).get('reliablity', 0)
|
2021-04-22 17:47:53 +02:00
|
|
|
|
reliability_order = 0 if reliability else 1
|
|
|
|
|
if key_name == 'reliability':
|
|
|
|
|
key = reliability
|
|
|
|
|
reliability_order = 0
|
|
|
|
|
else:
|
|
|
|
|
key = engine_stat.get(key_name) or default_value
|
|
|
|
|
if reverse:
|
|
|
|
|
reliability_order = 1 - reliability_order
|
|
|
|
|
return (reliability_order, key, engine_stat['name'])
|
|
|
|
|
|
|
|
|
|
engine_stats['time'] = sorted(engine_stats['time'], reverse=reverse, key=get_key)
|
2014-03-21 12:19:48 +01:00
|
|
|
|
return render(
|
2021-12-27 09:16:03 +01:00
|
|
|
|
# fmt: off
|
2014-03-21 12:19:48 +01:00
|
|
|
|
'stats.html',
|
2021-06-08 18:19:30 +02:00
|
|
|
|
sort_order = sort_order,
|
|
|
|
|
engine_stats = engine_stats,
|
|
|
|
|
engine_reliabilities = engine_reliabilities,
|
|
|
|
|
selected_engine_name = selected_engine_name,
|
2021-08-18 10:36:53 +02:00
|
|
|
|
searx_git_branch = GIT_BRANCH,
|
2021-12-27 09:16:03 +01:00
|
|
|
|
# fmt: on
|
2014-03-21 12:19:48 +01:00
|
|
|
|
)
|
2013-10-27 01:03:05 +02:00
|
|
|
|
|
2014-01-01 22:16:53 +01:00
|
|
|
|
|
2020-11-26 15:12:11 +01:00
|
|
|
|
@app.route('/stats/errors', methods=['GET'])
|
|
|
|
|
def stats_errors():
|
2022-01-09 15:53:58 +01:00
|
|
|
|
filtered_engines = dict(filter(lambda kv: request.preferences.validate_token(kv[1]), engines.items()))
|
2021-04-14 17:23:15 +02:00
|
|
|
|
result = get_engine_errors(filtered_engines)
|
2020-11-26 15:12:11 +01:00
|
|
|
|
return jsonify(result)
|
|
|
|
|
|
|
|
|
|
|
2021-01-05 11:24:39 +01:00
|
|
|
|
@app.route('/stats/checker', methods=['GET'])
|
|
|
|
|
def stats_checker():
|
|
|
|
|
result = checker_get_result()
|
|
|
|
|
return jsonify(result)
|
|
|
|
|
|
|
|
|
|
|
2013-12-01 16:10:38 +01:00
|
|
|
|
@app.route('/robots.txt', methods=['GET'])
|
|
|
|
|
def robots():
|
2021-12-27 09:26:22 +01:00
|
|
|
|
return Response(
|
|
|
|
|
"""User-agent: *
|
2013-12-01 16:10:38 +01:00
|
|
|
|
Allow: /
|
|
|
|
|
Allow: /about
|
|
|
|
|
Disallow: /stats
|
2014-02-07 18:43:05 +01:00
|
|
|
|
Disallow: /preferences
|
2016-10-01 20:22:52 +02:00
|
|
|
|
Disallow: /*?*q=*
|
2021-12-27 09:26:22 +01:00
|
|
|
|
""",
|
|
|
|
|
mimetype='text/plain',
|
|
|
|
|
)
|
2013-12-01 16:10:38 +01:00
|
|
|
|
|
2014-01-01 22:16:53 +01:00
|
|
|
|
|
2013-10-16 00:01:08 +02:00
|
|
|
|
@app.route('/opensearch.xml', methods=['GET'])
|
|
|
|
|
def opensearch():
|
2013-10-20 22:37:55 +02:00
|
|
|
|
method = 'post'
|
2015-03-15 20:07:50 +01:00
|
|
|
|
|
2016-04-08 16:38:05 +02:00
|
|
|
|
if request.preferences.get_value('method') == 'GET':
|
2015-03-15 20:07:50 +01:00
|
|
|
|
method = 'get'
|
|
|
|
|
|
2013-10-21 00:28:48 +02:00
|
|
|
|
# chrome/chromium only supports HTTP GET....
|
2013-10-20 22:37:55 +02:00
|
|
|
|
if request.headers.get('User-Agent', '').lower().find('webkit') >= 0:
|
|
|
|
|
method = 'get'
|
2014-03-29 16:29:19 +01:00
|
|
|
|
|
2021-12-27 09:26:22 +01:00
|
|
|
|
ret = render('opensearch.xml', opensearch_method=method, override_theme='__common__')
|
2014-03-29 16:29:19 +01:00
|
|
|
|
|
2021-12-27 09:26:22 +01:00
|
|
|
|
resp = Response(response=ret, status=200, mimetype="application/opensearchdescription+xml")
|
2013-10-16 00:01:08 +02:00
|
|
|
|
return resp
|
|
|
|
|
|
2014-01-19 22:59:01 +01:00
|
|
|
|
|
2013-12-01 23:52:49 +01:00
|
|
|
|
@app.route('/favicon.ico')
|
|
|
|
|
def favicon():
|
2021-05-26 13:32:04 +02:00
|
|
|
|
return send_from_directory(
|
2021-12-27 09:26:22 +01:00
|
|
|
|
os.path.join(app.root_path, settings['ui']['static_path'], 'themes', get_current_theme_name(), 'img'),
|
2021-05-26 13:32:04 +02:00
|
|
|
|
'favicon.png',
|
2021-12-27 09:26:22 +01:00
|
|
|
|
mimetype='image/vnd.microsoft.icon',
|
2021-05-26 13:32:04 +02:00
|
|
|
|
)
|
2013-12-01 23:52:49 +01:00
|
|
|
|
|
2021-12-27 09:26:22 +01:00
|
|
|
|
|
2015-04-07 11:07:48 +02:00
|
|
|
|
@app.route('/clear_cookies')
|
|
|
|
|
def clear_cookies():
|
2021-04-03 13:56:47 +02:00
|
|
|
|
resp = make_response(redirect(url_for('index', _external=True)))
|
2015-04-07 11:07:48 +02:00
|
|
|
|
for cookie_name in request.cookies:
|
|
|
|
|
resp.delete_cookie(cookie_name)
|
|
|
|
|
return resp
|
|
|
|
|
|
|
|
|
|
|
2016-06-07 23:08:48 +02:00
|
|
|
|
@app.route('/config')
|
|
|
|
|
def config():
|
2020-03-25 17:12:02 +01:00
|
|
|
|
"""Return configuration in JSON format."""
|
|
|
|
|
_engines = []
|
|
|
|
|
for name, engine in engines.items():
|
|
|
|
|
if not request.preferences.validate_token(engine):
|
|
|
|
|
continue
|
|
|
|
|
|
|
|
|
|
supported_languages = engine.supported_languages
|
|
|
|
|
if isinstance(engine.supported_languages, dict):
|
|
|
|
|
supported_languages = list(engine.supported_languages.keys())
|
|
|
|
|
|
2021-12-27 09:26:22 +01:00
|
|
|
|
_engines.append(
|
|
|
|
|
{
|
|
|
|
|
'name': name,
|
|
|
|
|
'categories': engine.categories,
|
|
|
|
|
'shortcut': engine.shortcut,
|
|
|
|
|
'enabled': not engine.disabled,
|
|
|
|
|
'paging': engine.paging,
|
|
|
|
|
'language_support': engine.language_support,
|
|
|
|
|
'supported_languages': supported_languages,
|
|
|
|
|
'safesearch': engine.safesearch,
|
|
|
|
|
'time_range_support': engine.time_range_support,
|
|
|
|
|
'timeout': engine.timeout,
|
|
|
|
|
}
|
|
|
|
|
)
|
2020-03-25 17:12:02 +01:00
|
|
|
|
|
|
|
|
|
_plugins = []
|
|
|
|
|
for _ in plugins:
|
|
|
|
|
_plugins.append({'name': _.name, 'enabled': _.default_on})
|
|
|
|
|
|
2021-12-27 09:26:22 +01:00
|
|
|
|
return jsonify(
|
|
|
|
|
{
|
|
|
|
|
'categories': list(categories.keys()),
|
|
|
|
|
'engines': _engines,
|
|
|
|
|
'plugins': _plugins,
|
|
|
|
|
'instance_name': settings['general']['instance_name'],
|
|
|
|
|
'locales': LOCALE_NAMES,
|
|
|
|
|
'default_locale': settings['ui']['default_locale'],
|
|
|
|
|
'autocomplete': settings['search']['autocomplete'],
|
|
|
|
|
'safe_search': settings['search']['safe_search'],
|
|
|
|
|
'default_theme': settings['ui']['default_theme'],
|
|
|
|
|
'version': VERSION_STRING,
|
|
|
|
|
'brand': {
|
|
|
|
|
'CONTACT_URL': get_setting('general.contact_url'),
|
|
|
|
|
'GIT_URL': GIT_URL,
|
|
|
|
|
'GIT_BRANCH': GIT_BRANCH,
|
|
|
|
|
'DOCS_URL': get_setting('brand.docs_url'),
|
|
|
|
|
},
|
|
|
|
|
'doi_resolvers': list(settings['doi_resolvers'].keys()),
|
|
|
|
|
'default_doi_resolver': settings['default_doi_resolver'],
|
|
|
|
|
}
|
|
|
|
|
)
|
2016-06-07 23:08:48 +02:00
|
|
|
|
|
|
|
|
|
|
2016-08-24 19:53:09 +02:00
|
|
|
|
@app.errorhandler(404)
|
2021-05-26 13:32:04 +02:00
|
|
|
|
def page_not_found(_e):
|
2016-09-07 08:32:01 +02:00
|
|
|
|
return render('404.html'), 404
|
2016-08-24 19:53:09 +02:00
|
|
|
|
|
|
|
|
|
|
2021-09-13 19:37:51 +02:00
|
|
|
|
# see https://flask.palletsprojects.com/en/1.1.x/cli/
|
|
|
|
|
# True if "FLASK_APP=searx/webapp.py FLASK_ENV=development flask run"
|
|
|
|
|
flask_run_development = (
|
2021-12-27 09:26:22 +01:00
|
|
|
|
os.environ.get("FLASK_APP") is not None and os.environ.get("FLASK_ENV") == 'development' and is_flask_run_cmdline()
|
2021-09-13 19:37:51 +02:00
|
|
|
|
)
|
|
|
|
|
|
|
|
|
|
# True if reload feature is activated of werkzeug, False otherwise (including uwsgi, etc..)
|
|
|
|
|
# __name__ != "__main__" if searx.webapp is imported (make test, make docs, uwsgi...)
|
|
|
|
|
# see run() at the end of this file : searx_debug activates the reload feature.
|
|
|
|
|
werkzeug_reloader = flask_run_development or (searx_debug and __name__ == "__main__")
|
|
|
|
|
|
|
|
|
|
# initialize the engines except on the first run of the werkzeug server.
|
2021-12-27 09:26:22 +01:00
|
|
|
|
if not werkzeug_reloader or (werkzeug_reloader and os.environ.get("WERKZEUG_RUN_MAIN") == "true"):
|
2021-09-13 19:37:51 +02:00
|
|
|
|
plugin_initialize(app)
|
2021-12-26 22:44:46 +01:00
|
|
|
|
search_initialize(enable_checker=True, check_network=True, enable_metrics=settings['general']['enable_metrics'])
|
2022-01-15 06:09:37 +01:00
|
|
|
|
user_help.render(app)
|
2021-09-13 19:37:51 +02:00
|
|
|
|
|
|
|
|
|
|
2014-01-12 12:40:27 +01:00
|
|
|
|
def run():
|
2021-12-27 09:26:22 +01:00
|
|
|
|
logger.debug('starting webserver on %s:%s', settings['server']['bind_address'], settings['server']['port'])
|
2014-01-19 22:59:01 +01:00
|
|
|
|
app.run(
|
2021-12-27 09:26:22 +01:00
|
|
|
|
debug=searx_debug,
|
|
|
|
|
use_debugger=searx_debug,
|
|
|
|
|
port=settings['server']['port'],
|
|
|
|
|
host=settings['server']['bind_address'],
|
|
|
|
|
threaded=True,
|
|
|
|
|
extra_files=[get_default_settings_path()],
|
2014-01-19 22:59:01 +01:00
|
|
|
|
)
|
2014-01-12 12:40:27 +01:00
|
|
|
|
|
2021-12-27 09:26:22 +01:00
|
|
|
|
|
2015-06-16 19:55:31 +02:00
|
|
|
|
application = app
|
2021-06-08 10:08:41 +02:00
|
|
|
|
patch_application(app)
|
2014-07-03 22:02:53 +02:00
|
|
|
|
|
2014-01-12 12:40:27 +01:00
|
|
|
|
if __name__ == "__main__":
|
|
|
|
|
run()
|