mirror of
https://github.com/searxng/searxng.git
synced 2024-11-17 18:00:12 +01:00
Merge pull request #2189 from dalf/architecture-clean-up
Architecture clean up
This commit is contained in:
commit
8d47142f35
@ -20,9 +20,8 @@ along with searx. If not, see < http://www.gnu.org/licenses/ >.
|
|||||||
import re
|
import re
|
||||||
|
|
||||||
from searx.languages import language_codes
|
from searx.languages import language_codes
|
||||||
from searx.engines import (
|
from searx.engines import categories, engines, engine_shortcuts
|
||||||
categories, engines, engine_shortcuts
|
from searx.search import EngineRef
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
VALID_LANGUAGE_CODE = re.compile(r'^[a-z]{2,3}(-[a-zA-Z]{2})?$')
|
VALID_LANGUAGE_CODE = re.compile(r'^[a-z]{2,3}(-[a-zA-Z]{2})?$')
|
||||||
@ -40,7 +39,7 @@ class RawTextQuery:
|
|||||||
self.disabled_engines = disabled_engines
|
self.disabled_engines = disabled_engines
|
||||||
|
|
||||||
self.query_parts = []
|
self.query_parts = []
|
||||||
self.engines = []
|
self.enginerefs = []
|
||||||
self.languages = []
|
self.languages = []
|
||||||
self.timeout_limit = None
|
self.timeout_limit = None
|
||||||
self.external_bang = None
|
self.external_bang = None
|
||||||
@ -135,24 +134,19 @@ class RawTextQuery:
|
|||||||
parse_next = True
|
parse_next = True
|
||||||
engine_name = engine_shortcuts[prefix]
|
engine_name = engine_shortcuts[prefix]
|
||||||
if engine_name in engines:
|
if engine_name in engines:
|
||||||
self.engines.append({'category': 'none',
|
self.enginerefs.append(EngineRef(engine_name, 'none', True))
|
||||||
'name': engine_name,
|
|
||||||
'from_bang': True})
|
|
||||||
|
|
||||||
# check if prefix is equal with engine name
|
# check if prefix is equal with engine name
|
||||||
elif prefix in engines:
|
elif prefix in engines:
|
||||||
parse_next = True
|
parse_next = True
|
||||||
self.engines.append({'category': 'none',
|
self.enginerefs.append(EngineRef(prefix, 'none', True))
|
||||||
'name': prefix,
|
|
||||||
'from_bang': True})
|
|
||||||
|
|
||||||
# check if prefix is equal with categorie name
|
# check if prefix is equal with categorie name
|
||||||
elif prefix in categories:
|
elif prefix in categories:
|
||||||
# using all engines for that search, which
|
# using all engines for that search, which
|
||||||
# are declared under that categorie name
|
# are declared under that categorie name
|
||||||
parse_next = True
|
parse_next = True
|
||||||
self.engines.extend({'category': prefix,
|
self.enginerefs.extend(EngineRef(engine.name, prefix)
|
||||||
'name': engine.name}
|
|
||||||
for engine in categories[prefix]
|
for engine in categories[prefix]
|
||||||
if (engine.name, prefix) not in self.disabled_engines)
|
if (engine.name, prefix) not in self.disabled_engines)
|
||||||
|
|
||||||
@ -178,23 +172,3 @@ class RawTextQuery:
|
|||||||
def getFullQuery(self):
|
def getFullQuery(self):
|
||||||
# get full querry including whitespaces
|
# get full querry including whitespaces
|
||||||
return ''.join(self.query_parts)
|
return ''.join(self.query_parts)
|
||||||
|
|
||||||
|
|
||||||
class SearchQuery:
|
|
||||||
"""container for all the search parameters (query, language, etc...)"""
|
|
||||||
|
|
||||||
def __init__(self, query, engines, categories, lang, safesearch, pageno, time_range,
|
|
||||||
timeout_limit=None, preferences=None, external_bang=None):
|
|
||||||
self.query = query
|
|
||||||
self.engines = engines
|
|
||||||
self.categories = categories
|
|
||||||
self.lang = lang
|
|
||||||
self.safesearch = safesearch
|
|
||||||
self.pageno = pageno
|
|
||||||
self.time_range = None if time_range in ('', 'None', None) else time_range
|
|
||||||
self.timeout_limit = timeout_limit
|
|
||||||
self.preferences = preferences
|
|
||||||
self.external_bang = external_bang
|
|
||||||
|
|
||||||
def __str__(self):
|
|
||||||
return self.query + ";" + str(self.engines)
|
|
||||||
|
237
searx/search.py
237
searx/search.py
@ -15,27 +15,22 @@ along with searx. If not, see < http://www.gnu.org/licenses/ >.
|
|||||||
(C) 2013- by Adam Tauber, <asciimoo@gmail.com>
|
(C) 2013- by Adam Tauber, <asciimoo@gmail.com>
|
||||||
'''
|
'''
|
||||||
|
|
||||||
|
import typing
|
||||||
import gc
|
import gc
|
||||||
import sys
|
|
||||||
import threading
|
import threading
|
||||||
from time import time
|
from time import time
|
||||||
from uuid import uuid4
|
from uuid import uuid4
|
||||||
from _thread import start_new_thread
|
from _thread import start_new_thread
|
||||||
|
|
||||||
from flask_babel import gettext
|
|
||||||
import requests.exceptions
|
import requests.exceptions
|
||||||
import searx.poolrequests as requests_lib
|
import searx.poolrequests as requests_lib
|
||||||
from searx.engines import (
|
from searx.engines import engines, settings
|
||||||
categories, engines, settings
|
|
||||||
)
|
|
||||||
from searx.answerers import ask
|
from searx.answerers import ask
|
||||||
from searx.external_bang import get_bang_url
|
from searx.external_bang import get_bang_url
|
||||||
from searx.utils import gen_useragent
|
from searx.utils import gen_useragent
|
||||||
from searx.query import RawTextQuery, SearchQuery, VALID_LANGUAGE_CODE
|
|
||||||
from searx.results import ResultContainer
|
from searx.results import ResultContainer
|
||||||
from searx import logger
|
from searx import logger
|
||||||
from searx.plugins import plugins
|
from searx.plugins import plugins
|
||||||
from searx.exceptions import SearxParameterException
|
|
||||||
|
|
||||||
|
|
||||||
logger = logger.getChild('search')
|
logger = logger.getChild('search')
|
||||||
@ -53,6 +48,49 @@ else:
|
|||||||
exit(1)
|
exit(1)
|
||||||
|
|
||||||
|
|
||||||
|
class EngineRef:
|
||||||
|
|
||||||
|
__slots__ = 'name', 'category', 'from_bang'
|
||||||
|
|
||||||
|
def __init__(self, name: str, category: str, from_bang: bool=False):
|
||||||
|
self.name = name
|
||||||
|
self.category = category
|
||||||
|
self.from_bang = from_bang
|
||||||
|
|
||||||
|
def __str__(self):
|
||||||
|
return "(" + self.name + "," + self.category + "," + str(self.from_bang) + ")"
|
||||||
|
|
||||||
|
|
||||||
|
class SearchQuery:
|
||||||
|
"""container for all the search parameters (query, language, etc...)"""
|
||||||
|
|
||||||
|
__slots__ = 'query', 'engineref_list', 'categories', 'lang', 'safesearch', 'pageno', 'time_range',\
|
||||||
|
'timeout_limit', 'external_bang'
|
||||||
|
|
||||||
|
def __init__(self,
|
||||||
|
query: str,
|
||||||
|
engineref_list: typing.List[EngineRef],
|
||||||
|
categories: typing.List[str],
|
||||||
|
lang: str,
|
||||||
|
safesearch: bool,
|
||||||
|
pageno: int,
|
||||||
|
time_range: typing.Optional[str],
|
||||||
|
timeout_limit: typing.Optional[float]=None,
|
||||||
|
external_bang: typing.Optional[str]=False):
|
||||||
|
self.query = query
|
||||||
|
self.engineref_list = engineref_list
|
||||||
|
self.categories = categories
|
||||||
|
self.lang = lang
|
||||||
|
self.safesearch = safesearch
|
||||||
|
self.pageno = pageno
|
||||||
|
self.time_range = time_range
|
||||||
|
self.timeout_limit = timeout_limit
|
||||||
|
self.external_bang = external_bang
|
||||||
|
|
||||||
|
def __str__(self):
|
||||||
|
return self.query + ";" + str(self.engineref_list)
|
||||||
|
|
||||||
|
|
||||||
def send_http_request(engine, request_params):
|
def send_http_request(engine, request_params):
|
||||||
# create dictionary which contain all
|
# create dictionary which contain all
|
||||||
# informations about the request
|
# informations about the request
|
||||||
@ -247,167 +285,11 @@ def default_request_params():
|
|||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
# remove duplicate queries.
|
|
||||||
# FIXME: does not fix "!music !soundcloud", because the categories are 'none' and 'music'
|
|
||||||
def deduplicate_query_engines(query_engines):
|
|
||||||
uniq_query_engines = {q["category"] + '|' + q["name"]: q for q in query_engines}
|
|
||||||
return uniq_query_engines.values()
|
|
||||||
|
|
||||||
|
|
||||||
def get_search_query_from_webapp(preferences, form):
|
|
||||||
# no text for the query ?
|
|
||||||
if not form.get('q'):
|
|
||||||
raise SearxParameterException('q', '')
|
|
||||||
|
|
||||||
# set blocked engines
|
|
||||||
disabled_engines = preferences.engines.get_disabled()
|
|
||||||
|
|
||||||
# parse query, if tags are set, which change
|
|
||||||
# the serch engine or search-language
|
|
||||||
raw_text_query = RawTextQuery(form['q'], disabled_engines)
|
|
||||||
|
|
||||||
# set query
|
|
||||||
query = raw_text_query.getQuery()
|
|
||||||
|
|
||||||
# get and check page number
|
|
||||||
pageno_param = form.get('pageno', '1')
|
|
||||||
if not pageno_param.isdigit() or int(pageno_param) < 1:
|
|
||||||
raise SearxParameterException('pageno', pageno_param)
|
|
||||||
query_pageno = int(pageno_param)
|
|
||||||
|
|
||||||
# get language
|
|
||||||
# set specific language if set on request, query or preferences
|
|
||||||
# TODO support search with multible languages
|
|
||||||
if len(raw_text_query.languages):
|
|
||||||
query_lang = raw_text_query.languages[-1]
|
|
||||||
elif 'language' in form:
|
|
||||||
query_lang = form.get('language')
|
|
||||||
else:
|
|
||||||
query_lang = preferences.get_value('language')
|
|
||||||
|
|
||||||
# check language
|
|
||||||
if not VALID_LANGUAGE_CODE.match(query_lang):
|
|
||||||
raise SearxParameterException('language', query_lang)
|
|
||||||
|
|
||||||
# get safesearch
|
|
||||||
if 'safesearch' in form:
|
|
||||||
query_safesearch = form.get('safesearch')
|
|
||||||
# first check safesearch
|
|
||||||
if not query_safesearch.isdigit():
|
|
||||||
raise SearxParameterException('safesearch', query_safesearch)
|
|
||||||
query_safesearch = int(query_safesearch)
|
|
||||||
else:
|
|
||||||
query_safesearch = preferences.get_value('safesearch')
|
|
||||||
|
|
||||||
# safesearch : second check
|
|
||||||
if query_safesearch < 0 or query_safesearch > 2:
|
|
||||||
raise SearxParameterException('safesearch', query_safesearch)
|
|
||||||
|
|
||||||
# get time_range
|
|
||||||
query_time_range = form.get('time_range')
|
|
||||||
|
|
||||||
# check time_range
|
|
||||||
if query_time_range not in ('None', None, '', 'day', 'week', 'month', 'year'):
|
|
||||||
raise SearxParameterException('time_range', query_time_range)
|
|
||||||
|
|
||||||
# query_engines
|
|
||||||
query_engines = raw_text_query.engines
|
|
||||||
|
|
||||||
# timeout_limit
|
|
||||||
query_timeout = raw_text_query.timeout_limit
|
|
||||||
if query_timeout is None and 'timeout_limit' in form:
|
|
||||||
raw_time_limit = form.get('timeout_limit')
|
|
||||||
if raw_time_limit in ['None', '']:
|
|
||||||
raw_time_limit = None
|
|
||||||
else:
|
|
||||||
try:
|
|
||||||
query_timeout = float(raw_time_limit)
|
|
||||||
except ValueError:
|
|
||||||
raise SearxParameterException('timeout_limit', raw_time_limit)
|
|
||||||
|
|
||||||
# query_categories
|
|
||||||
query_categories = []
|
|
||||||
|
|
||||||
# if engines are calculated from query,
|
|
||||||
# set categories by using that informations
|
|
||||||
if query_engines and raw_text_query.specific:
|
|
||||||
additional_categories = set()
|
|
||||||
for engine in query_engines:
|
|
||||||
if 'from_bang' in engine and engine['from_bang']:
|
|
||||||
additional_categories.add('none')
|
|
||||||
else:
|
|
||||||
additional_categories.add(engine['category'])
|
|
||||||
query_categories = list(additional_categories)
|
|
||||||
|
|
||||||
# otherwise, using defined categories to
|
|
||||||
# calculate which engines should be used
|
|
||||||
else:
|
|
||||||
# set categories/engines
|
|
||||||
load_default_categories = True
|
|
||||||
for pd_name, pd in form.items():
|
|
||||||
if pd_name == 'categories':
|
|
||||||
query_categories.extend(categ for categ in map(str.strip, pd.split(',')) if categ in categories)
|
|
||||||
elif pd_name == 'engines':
|
|
||||||
pd_engines = [{'category': engines[engine].categories[0],
|
|
||||||
'name': engine}
|
|
||||||
for engine in map(str.strip, pd.split(',')) if engine in engines]
|
|
||||||
if pd_engines:
|
|
||||||
query_engines.extend(pd_engines)
|
|
||||||
load_default_categories = False
|
|
||||||
elif pd_name.startswith('category_'):
|
|
||||||
category = pd_name[9:]
|
|
||||||
|
|
||||||
# if category is not found in list, skip
|
|
||||||
if category not in categories:
|
|
||||||
continue
|
|
||||||
|
|
||||||
if pd != 'off':
|
|
||||||
# add category to list
|
|
||||||
query_categories.append(category)
|
|
||||||
elif category in query_categories:
|
|
||||||
# remove category from list if property is set to 'off'
|
|
||||||
query_categories.remove(category)
|
|
||||||
|
|
||||||
if not load_default_categories:
|
|
||||||
if not query_categories:
|
|
||||||
query_categories = list(set(engine['category']
|
|
||||||
for engine in query_engines))
|
|
||||||
else:
|
|
||||||
# if no category is specified for this search,
|
|
||||||
# using user-defined default-configuration which
|
|
||||||
# (is stored in cookie)
|
|
||||||
if not query_categories:
|
|
||||||
cookie_categories = preferences.get_value('categories')
|
|
||||||
for ccateg in cookie_categories:
|
|
||||||
if ccateg in categories:
|
|
||||||
query_categories.append(ccateg)
|
|
||||||
|
|
||||||
# if still no category is specified, using general
|
|
||||||
# as default-category
|
|
||||||
if not query_categories:
|
|
||||||
query_categories = ['general']
|
|
||||||
|
|
||||||
# using all engines for that search, which are
|
|
||||||
# declared under the specific categories
|
|
||||||
for categ in query_categories:
|
|
||||||
query_engines.extend({'category': categ,
|
|
||||||
'name': engine.name}
|
|
||||||
for engine in categories[categ]
|
|
||||||
if (engine.name, categ) not in disabled_engines)
|
|
||||||
|
|
||||||
query_engines = deduplicate_query_engines(query_engines)
|
|
||||||
external_bang = raw_text_query.external_bang
|
|
||||||
|
|
||||||
return (SearchQuery(query, query_engines, query_categories,
|
|
||||||
query_lang, query_safesearch, query_pageno,
|
|
||||||
query_time_range, query_timeout, preferences,
|
|
||||||
external_bang=external_bang),
|
|
||||||
raw_text_query)
|
|
||||||
|
|
||||||
|
|
||||||
class Search:
|
class Search:
|
||||||
"""Search information container"""
|
"""Search information container"""
|
||||||
|
|
||||||
|
__slots__ = "search_query", "result_container", "start_time", "actual_timeout"
|
||||||
|
|
||||||
def __init__(self, search_query):
|
def __init__(self, search_query):
|
||||||
# init vars
|
# init vars
|
||||||
super().__init__()
|
super().__init__()
|
||||||
@ -444,9 +326,6 @@ class Search:
|
|||||||
return False
|
return False
|
||||||
|
|
||||||
def _is_accepted(self, engine_name, engine):
|
def _is_accepted(self, engine_name, engine):
|
||||||
if not self.search_query.preferences.validate_token(engine):
|
|
||||||
return False
|
|
||||||
|
|
||||||
# skip suspended engines
|
# skip suspended engines
|
||||||
if engine.suspend_end_time >= time():
|
if engine.suspend_end_time >= time():
|
||||||
logger.debug('Engine currently suspended: %s', engine_name)
|
logger.debug('Engine currently suspended: %s', engine_name)
|
||||||
@ -462,13 +341,13 @@ class Search:
|
|||||||
|
|
||||||
return True
|
return True
|
||||||
|
|
||||||
def _get_params(self, selected_engine, user_agent):
|
def _get_params(self, engineref, user_agent):
|
||||||
if selected_engine['name'] not in engines:
|
if engineref.name not in engines:
|
||||||
return None, None
|
return None, None
|
||||||
|
|
||||||
engine = engines[selected_engine['name']]
|
engine = engines[engineref.name]
|
||||||
|
|
||||||
if not self._is_accepted(selected_engine['name'], engine):
|
if not self._is_accepted(engineref.name, engine):
|
||||||
return None, None
|
return None, None
|
||||||
|
|
||||||
# set default request parameters
|
# set default request parameters
|
||||||
@ -485,15 +364,13 @@ class Search:
|
|||||||
request_params['safesearch'] = self.search_query.safesearch
|
request_params['safesearch'] = self.search_query.safesearch
|
||||||
request_params['time_range'] = self.search_query.time_range
|
request_params['time_range'] = self.search_query.time_range
|
||||||
|
|
||||||
request_params['category'] = selected_engine['category']
|
request_params['category'] = engineref.category
|
||||||
request_params['pageno'] = self.search_query.pageno
|
request_params['pageno'] = self.search_query.pageno
|
||||||
|
|
||||||
return request_params, engine.timeout
|
return request_params, engine.timeout
|
||||||
|
|
||||||
# do search-request
|
# do search-request
|
||||||
def _get_requests(self):
|
def _get_requests(self):
|
||||||
global number_of_searches
|
|
||||||
|
|
||||||
# init vars
|
# init vars
|
||||||
requests = []
|
requests = []
|
||||||
|
|
||||||
@ -505,14 +382,14 @@ class Search:
|
|||||||
default_timeout = 0
|
default_timeout = 0
|
||||||
|
|
||||||
# start search-reqest for all selected engines
|
# start search-reqest for all selected engines
|
||||||
for selected_engine in self.search_query.engines:
|
for engineref in self.search_query.engineref_list:
|
||||||
# set default request parameters
|
# set default request parameters
|
||||||
request_params, engine_timeout = self._get_params(selected_engine, user_agent)
|
request_params, engine_timeout = self._get_params(engineref, user_agent)
|
||||||
if request_params is None:
|
if request_params is None:
|
||||||
continue
|
continue
|
||||||
|
|
||||||
# append request to list
|
# append request to list
|
||||||
requests.append((selected_engine['name'], self.search_query.query, request_params))
|
requests.append((engineref.name, self.search_query.query, request_params))
|
||||||
|
|
||||||
# update default_timeout
|
# update default_timeout
|
||||||
default_timeout = max(default_timeout, engine_timeout)
|
default_timeout = max(default_timeout, engine_timeout)
|
||||||
@ -535,7 +412,7 @@ class Search:
|
|||||||
actual_timeout = min(query_timeout, max_request_timeout)
|
actual_timeout = min(query_timeout, max_request_timeout)
|
||||||
|
|
||||||
logger.debug("actual_timeout={0} (default_timeout={1}, ?timeout_limit={2}, max_request_timeout={3})"
|
logger.debug("actual_timeout={0} (default_timeout={1}, ?timeout_limit={2}, max_request_timeout={3})"
|
||||||
.format(self.actual_timeout, default_timeout, query_timeout, max_request_timeout))
|
.format(actual_timeout, default_timeout, query_timeout, max_request_timeout))
|
||||||
|
|
||||||
return requests, actual_timeout
|
return requests, actual_timeout
|
||||||
|
|
||||||
@ -567,6 +444,8 @@ class Search:
|
|||||||
class SearchWithPlugins(Search):
|
class SearchWithPlugins(Search):
|
||||||
"""Similar to the Search class but call the plugins."""
|
"""Similar to the Search class but call the plugins."""
|
||||||
|
|
||||||
|
__slots__ = 'ordered_plugin_list', 'request'
|
||||||
|
|
||||||
def __init__(self, search_query, ordered_plugin_list, request):
|
def __init__(self, search_query, ordered_plugin_list, request):
|
||||||
super().__init__(search_query)
|
super().__init__(search_query)
|
||||||
self.ordered_plugin_list = ordered_plugin_list
|
self.ordered_plugin_list = ordered_plugin_list
|
||||||
|
235
searx/webadapter.py
Normal file
235
searx/webadapter.py
Normal file
@ -0,0 +1,235 @@
|
|||||||
|
from typing import Dict, List, Optional, Tuple
|
||||||
|
from searx.exceptions import SearxParameterException
|
||||||
|
from searx.query import RawTextQuery, VALID_LANGUAGE_CODE
|
||||||
|
from searx.engines import categories, engines
|
||||||
|
from searx.search import SearchQuery, EngineRef
|
||||||
|
from searx.preferences import Preferences
|
||||||
|
|
||||||
|
|
||||||
|
# remove duplicate queries.
|
||||||
|
# FIXME: does not fix "!music !soundcloud", because the categories are 'none' and 'music'
|
||||||
|
def deduplicate_engineref_list(engineref_list: List[EngineRef]) -> List[EngineRef]:
|
||||||
|
engineref_dict = {q.category + '|' + q.name: q for q in engineref_list}
|
||||||
|
return engineref_dict.values()
|
||||||
|
|
||||||
|
|
||||||
|
def validate_engineref_list(engineref_list: List[EngineRef], preferences: Preferences)\
|
||||||
|
-> Tuple[List[EngineRef], List[EngineRef], List[EngineRef]]:
|
||||||
|
"""Validate query_engines according to the preferences
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
List[EngineRef]: list of existing engines with a validated token
|
||||||
|
List[EngineRef]: list of unknown engine
|
||||||
|
List[EngineRef]: list of engine with invalid token according to the preferences
|
||||||
|
"""
|
||||||
|
valid = []
|
||||||
|
unknown = []
|
||||||
|
no_token = []
|
||||||
|
for engineref in engineref_list:
|
||||||
|
if engineref.name not in engines:
|
||||||
|
unknown.append(engineref)
|
||||||
|
continue
|
||||||
|
|
||||||
|
engine = engines[engineref.name]
|
||||||
|
if not preferences.validate_token(engine):
|
||||||
|
no_token.append(engineref)
|
||||||
|
continue
|
||||||
|
|
||||||
|
valid.append(engineref)
|
||||||
|
return valid, unknown, no_token
|
||||||
|
|
||||||
|
|
||||||
|
def parse_pageno(form: Dict[str, str]) -> int:
|
||||||
|
pageno_param = form.get('pageno', '1')
|
||||||
|
if not pageno_param.isdigit() or int(pageno_param) < 1:
|
||||||
|
raise SearxParameterException('pageno', pageno_param)
|
||||||
|
return int(pageno_param)
|
||||||
|
|
||||||
|
|
||||||
|
def parse_lang(preferences: Preferences, form: Dict[str, str], raw_text_query: RawTextQuery) -> str:
|
||||||
|
# get language
|
||||||
|
# set specific language if set on request, query or preferences
|
||||||
|
# TODO support search with multible languages
|
||||||
|
if len(raw_text_query.languages):
|
||||||
|
query_lang = raw_text_query.languages[-1]
|
||||||
|
elif 'language' in form:
|
||||||
|
query_lang = form.get('language')
|
||||||
|
else:
|
||||||
|
query_lang = preferences.get_value('language')
|
||||||
|
|
||||||
|
# check language
|
||||||
|
if not VALID_LANGUAGE_CODE.match(query_lang):
|
||||||
|
raise SearxParameterException('language', query_lang)
|
||||||
|
|
||||||
|
return query_lang
|
||||||
|
|
||||||
|
|
||||||
|
def parse_safesearch(preferences: Preferences, form: Dict[str, str]) -> int:
|
||||||
|
if 'safesearch' in form:
|
||||||
|
query_safesearch = form.get('safesearch')
|
||||||
|
# first check safesearch
|
||||||
|
if not query_safesearch.isdigit():
|
||||||
|
raise SearxParameterException('safesearch', query_safesearch)
|
||||||
|
query_safesearch = int(query_safesearch)
|
||||||
|
else:
|
||||||
|
query_safesearch = preferences.get_value('safesearch')
|
||||||
|
|
||||||
|
# safesearch : second check
|
||||||
|
if query_safesearch < 0 or query_safesearch > 2:
|
||||||
|
raise SearxParameterException('safesearch', query_safesearch)
|
||||||
|
|
||||||
|
return query_safesearch
|
||||||
|
|
||||||
|
|
||||||
|
def parse_time_range(form: Dict[str, str]) -> str:
|
||||||
|
query_time_range = form.get('time_range')
|
||||||
|
# check time_range
|
||||||
|
query_time_range = None if query_time_range in ('', 'None') else query_time_range
|
||||||
|
if query_time_range not in (None, 'day', 'week', 'month', 'year'):
|
||||||
|
raise SearxParameterException('time_range', query_time_range)
|
||||||
|
return query_time_range
|
||||||
|
|
||||||
|
|
||||||
|
def parse_timeout(form: Dict[str, str], raw_text_query: RawTextQuery) -> Optional[float]:
|
||||||
|
query_timeout = raw_text_query.timeout_limit
|
||||||
|
if query_timeout is None and 'timeout_limit' in form:
|
||||||
|
raw_time_limit = form.get('timeout_limit')
|
||||||
|
if raw_time_limit in ['None', '']:
|
||||||
|
return None
|
||||||
|
else:
|
||||||
|
try:
|
||||||
|
return float(raw_time_limit)
|
||||||
|
except ValueError:
|
||||||
|
raise SearxParameterException('timeout_limit', raw_time_limit)
|
||||||
|
|
||||||
|
|
||||||
|
def parse_specific(raw_text_query: RawTextQuery) -> Tuple[List[EngineRef], List[str]]:
|
||||||
|
query_engineref_list = raw_text_query.enginerefs
|
||||||
|
additional_categories = set()
|
||||||
|
for engineref in raw_text_query.enginerefs:
|
||||||
|
if engineref.from_bang:
|
||||||
|
additional_categories.add('none')
|
||||||
|
else:
|
||||||
|
additional_categories.add(engineref.category)
|
||||||
|
query_categories = list(additional_categories)
|
||||||
|
return query_engineref_list, query_categories
|
||||||
|
|
||||||
|
|
||||||
|
def parse_category_form(query_categories: List[str], name: str, value: str) -> None:
|
||||||
|
if name == 'categories':
|
||||||
|
query_categories.extend(categ for categ in map(str.strip, value.split(',')) if categ in categories)
|
||||||
|
elif name.startswith('category_'):
|
||||||
|
category = name[9:]
|
||||||
|
|
||||||
|
# if category is not found in list, skip
|
||||||
|
if category not in categories:
|
||||||
|
return
|
||||||
|
|
||||||
|
if value != 'off':
|
||||||
|
# add category to list
|
||||||
|
query_categories.append(category)
|
||||||
|
elif category in query_categories:
|
||||||
|
# remove category from list if property is set to 'off'
|
||||||
|
query_categories.remove(category)
|
||||||
|
|
||||||
|
|
||||||
|
def get_selected_categories(preferences: Preferences, form: Dict[str, str]) -> List[str]:
|
||||||
|
selected_categories = []
|
||||||
|
|
||||||
|
if form is not None:
|
||||||
|
for name, value in form.items():
|
||||||
|
parse_category_form(selected_categories, name, value)
|
||||||
|
|
||||||
|
# if no category is specified for this search,
|
||||||
|
# using user-defined default-configuration which
|
||||||
|
# (is stored in cookie)
|
||||||
|
if not selected_categories:
|
||||||
|
cookie_categories = preferences.get_value('categories')
|
||||||
|
for ccateg in cookie_categories:
|
||||||
|
selected_categories.append(ccateg)
|
||||||
|
|
||||||
|
# if still no category is specified, using general
|
||||||
|
# as default-category
|
||||||
|
if not selected_categories:
|
||||||
|
selected_categories = ['general']
|
||||||
|
|
||||||
|
return selected_categories
|
||||||
|
|
||||||
|
|
||||||
|
def parse_generic(preferences: Preferences, form: Dict[str, str], disabled_engines: List[str])\
|
||||||
|
-> Tuple[List[EngineRef], List[str]]:
|
||||||
|
query_engineref_list = []
|
||||||
|
query_categories = []
|
||||||
|
|
||||||
|
# set categories/engines
|
||||||
|
load_default_categories = True
|
||||||
|
for pd_name, pd in form.items():
|
||||||
|
if pd_name == 'engines':
|
||||||
|
pd_engines = [EngineRef(engine_name, engines[engine_name].categories[0])
|
||||||
|
for engine_name in map(str.strip, pd.split(',')) if engine_name in engines]
|
||||||
|
if pd_engines:
|
||||||
|
query_engineref_list.extend(pd_engines)
|
||||||
|
load_default_categories = False
|
||||||
|
else:
|
||||||
|
parse_category_form(query_categories, pd_name, pd)
|
||||||
|
|
||||||
|
if not load_default_categories:
|
||||||
|
if not query_categories:
|
||||||
|
query_categories = list(set(engine['category']
|
||||||
|
for engine in query_engineref_list))
|
||||||
|
else:
|
||||||
|
if not query_categories:
|
||||||
|
query_categories = get_selected_categories(preferences, None)
|
||||||
|
|
||||||
|
# using all engines for that search, which are
|
||||||
|
# declared under the specific categories
|
||||||
|
for categ in query_categories:
|
||||||
|
query_engineref_list.extend(EngineRef(engine.name, categ)
|
||||||
|
for engine in categories[categ]
|
||||||
|
if (engine.name, categ) not in disabled_engines)
|
||||||
|
|
||||||
|
return query_engineref_list, query_categories
|
||||||
|
|
||||||
|
|
||||||
|
def get_search_query_from_webapp(preferences: Preferences, form: Dict[str, str])\
|
||||||
|
-> Tuple[SearchQuery, RawTextQuery, List[EngineRef], List[EngineRef]]:
|
||||||
|
# no text for the query ?
|
||||||
|
if not form.get('q'):
|
||||||
|
raise SearxParameterException('q', '')
|
||||||
|
|
||||||
|
# set blocked engines
|
||||||
|
disabled_engines = preferences.engines.get_disabled()
|
||||||
|
|
||||||
|
# parse query, if tags are set, which change
|
||||||
|
# the serch engine or search-language
|
||||||
|
raw_text_query = RawTextQuery(form['q'], disabled_engines)
|
||||||
|
|
||||||
|
# set query
|
||||||
|
query = raw_text_query.getQuery()
|
||||||
|
query_pageno = parse_pageno(form)
|
||||||
|
query_lang = parse_lang(preferences, form, raw_text_query)
|
||||||
|
query_safesearch = parse_safesearch(preferences, form)
|
||||||
|
query_time_range = parse_time_range(form)
|
||||||
|
query_timeout = parse_timeout(form, raw_text_query)
|
||||||
|
external_bang = raw_text_query.external_bang
|
||||||
|
|
||||||
|
if raw_text_query.enginerefs and raw_text_query.specific:
|
||||||
|
# if engines are calculated from query,
|
||||||
|
# set categories by using that informations
|
||||||
|
query_engineref_list, query_categories = parse_specific(raw_text_query)
|
||||||
|
else:
|
||||||
|
# otherwise, using defined categories to
|
||||||
|
# calculate which engines should be used
|
||||||
|
query_engineref_list, query_categories = parse_generic(preferences, form, disabled_engines)
|
||||||
|
|
||||||
|
query_engineref_list = deduplicate_engineref_list(query_engineref_list)
|
||||||
|
query_engineref_list, query_engineref_list_unknown, query_engineref_list_notoken =\
|
||||||
|
validate_engineref_list(query_engineref_list, preferences)
|
||||||
|
|
||||||
|
return (SearchQuery(query, query_engineref_list, query_categories,
|
||||||
|
query_lang, query_safesearch, query_pageno,
|
||||||
|
query_time_range, query_timeout,
|
||||||
|
external_bang=external_bang),
|
||||||
|
raw_text_query,
|
||||||
|
query_engineref_list_unknown,
|
||||||
|
query_engineref_list_notoken)
|
@ -67,10 +67,11 @@ from searx.webutils import (
|
|||||||
get_static_files, get_result_templates, get_themes,
|
get_static_files, get_result_templates, get_themes,
|
||||||
prettify_url, new_hmac
|
prettify_url, new_hmac
|
||||||
)
|
)
|
||||||
|
from searx.webadapter import get_search_query_from_webapp, get_selected_categories
|
||||||
from searx.utils import html_to_text, gen_useragent, dict_subset, match_language
|
from searx.utils import html_to_text, gen_useragent, dict_subset, match_language
|
||||||
from searx.version import VERSION_STRING
|
from searx.version import VERSION_STRING
|
||||||
from searx.languages import language_codes as languages
|
from searx.languages import language_codes as languages
|
||||||
from searx.search import SearchWithPlugins, get_search_query_from_webapp
|
from searx.search import SearchWithPlugins
|
||||||
from searx.query import RawTextQuery
|
from searx.query import RawTextQuery
|
||||||
from searx.autocomplete import searx_bang, backends as autocomplete_backends
|
from searx.autocomplete import searx_bang, backends as autocomplete_backends
|
||||||
from searx.plugins import plugins
|
from searx.plugins import plugins
|
||||||
@ -347,25 +348,6 @@ def render(template_name, override_theme=None, **kwargs):
|
|||||||
_get_ordered_categories()
|
_get_ordered_categories()
|
||||||
if x in enabled_categories]
|
if x in enabled_categories]
|
||||||
|
|
||||||
if 'all_categories' not in kwargs:
|
|
||||||
kwargs['all_categories'] = _get_ordered_categories()
|
|
||||||
|
|
||||||
if 'selected_categories' not in kwargs:
|
|
||||||
kwargs['selected_categories'] = []
|
|
||||||
for arg in request.args:
|
|
||||||
if arg.startswith('category_'):
|
|
||||||
c = arg.split('_', 1)[1]
|
|
||||||
if c in categories:
|
|
||||||
kwargs['selected_categories'].append(c)
|
|
||||||
|
|
||||||
if not kwargs['selected_categories']:
|
|
||||||
cookie_categories = request.preferences.get_value('categories')
|
|
||||||
for ccateg in cookie_categories:
|
|
||||||
kwargs['selected_categories'].append(ccateg)
|
|
||||||
|
|
||||||
if not kwargs['selected_categories']:
|
|
||||||
kwargs['selected_categories'] = ['general']
|
|
||||||
|
|
||||||
if 'autocomplete' not in kwargs:
|
if 'autocomplete' not in kwargs:
|
||||||
kwargs['autocomplete'] = request.preferences.get_value('autocomplete')
|
kwargs['autocomplete'] = request.preferences.get_value('autocomplete')
|
||||||
|
|
||||||
@ -524,6 +506,7 @@ def index_error(output_format, error_message):
|
|||||||
request.errors.append(gettext('search error'))
|
request.errors.append(gettext('search error'))
|
||||||
return render(
|
return render(
|
||||||
'index.html',
|
'index.html',
|
||||||
|
selected_categories=get_selected_categories(request.preferences, request.form),
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
@ -545,6 +528,7 @@ def index():
|
|||||||
if output_format == 'html':
|
if output_format == 'html':
|
||||||
return render(
|
return render(
|
||||||
'index.html',
|
'index.html',
|
||||||
|
selected_categories=get_selected_categories(request.preferences, request.form),
|
||||||
)
|
)
|
||||||
else:
|
else:
|
||||||
return index_error(output_format, 'No query'), 400
|
return index_error(output_format, 'No query'), 400
|
||||||
@ -554,7 +538,7 @@ def index():
|
|||||||
raw_text_query = None
|
raw_text_query = None
|
||||||
result_container = None
|
result_container = None
|
||||||
try:
|
try:
|
||||||
search_query, raw_text_query = get_search_query_from_webapp(request.preferences, request.form)
|
search_query, raw_text_query, _, _ = get_search_query_from_webapp(request.preferences, request.form)
|
||||||
# search = Search(search_query) # without plugins
|
# search = Search(search_query) # without plugins
|
||||||
search = SearchWithPlugins(search_query, request.user_plugins, request)
|
search = SearchWithPlugins(search_query, request.user_plugins, request)
|
||||||
|
|
||||||
@ -825,6 +809,8 @@ def preferences():
|
|||||||
# end of stats
|
# end of stats
|
||||||
|
|
||||||
return render('preferences.html',
|
return render('preferences.html',
|
||||||
|
selected_categories=get_selected_categories(request.preferences, request.form),
|
||||||
|
all_categories=_get_ordered_categories(),
|
||||||
locales=settings['locales'],
|
locales=settings['locales'],
|
||||||
current_locale=request.preferences.get_value("locale"),
|
current_locale=request.preferences.get_value("locale"),
|
||||||
image_proxy=image_proxy,
|
image_proxy=image_proxy,
|
||||||
|
@ -1,16 +1,14 @@
|
|||||||
# -*- coding: utf-8 -*-
|
# -*- coding: utf-8 -*-
|
||||||
|
|
||||||
from searx.testing import SearxTestCase
|
from searx.testing import SearxTestCase
|
||||||
from searx.preferences import Preferences
|
from searx.search import SearchQuery, EngineRef
|
||||||
from searx.engines import engines
|
|
||||||
|
|
||||||
import searx.search
|
import searx.search
|
||||||
|
import searx.engines
|
||||||
|
|
||||||
|
|
||||||
SAFESEARCH = 0
|
SAFESEARCH = 0
|
||||||
PAGENO = 1
|
PAGENO = 1
|
||||||
PUBLIC_ENGINE_NAME = 'general dummy'
|
PUBLIC_ENGINE_NAME = 'general dummy'
|
||||||
PRIVATE_ENGINE_NAME = 'general private offline'
|
|
||||||
TEST_ENGINES = [
|
TEST_ENGINES = [
|
||||||
{
|
{
|
||||||
'name': PUBLIC_ENGINE_NAME,
|
'name': PUBLIC_ENGINE_NAME,
|
||||||
@ -20,15 +18,6 @@ TEST_ENGINES = [
|
|||||||
'timeout': 3.0,
|
'timeout': 3.0,
|
||||||
'tokens': [],
|
'tokens': [],
|
||||||
},
|
},
|
||||||
{
|
|
||||||
'name': PRIVATE_ENGINE_NAME,
|
|
||||||
'engine': 'dummy-offline',
|
|
||||||
'categories': 'general',
|
|
||||||
'shortcut': 'do',
|
|
||||||
'timeout': 3.0,
|
|
||||||
'offline': True,
|
|
||||||
'tokens': ['my-token'],
|
|
||||||
},
|
|
||||||
]
|
]
|
||||||
|
|
||||||
|
|
||||||
@ -40,92 +29,57 @@ class SearchTestCase(SearxTestCase):
|
|||||||
|
|
||||||
def test_timeout_simple(self):
|
def test_timeout_simple(self):
|
||||||
searx.search.max_request_timeout = None
|
searx.search.max_request_timeout = None
|
||||||
search_query = searx.query.SearchQuery('test', [{'category': 'general', 'name': PUBLIC_ENGINE_NAME}],
|
search_query = SearchQuery('test', [EngineRef(PUBLIC_ENGINE_NAME, 'general')],
|
||||||
['general'], 'en-US', SAFESEARCH, PAGENO, None, None,
|
['general'], 'en-US', SAFESEARCH, PAGENO, None, None)
|
||||||
preferences=Preferences(['oscar'], ['general'], engines, []))
|
|
||||||
search = searx.search.Search(search_query)
|
search = searx.search.Search(search_query)
|
||||||
search.search()
|
search.search()
|
||||||
self.assertEqual(search.actual_timeout, 3.0)
|
self.assertEqual(search.actual_timeout, 3.0)
|
||||||
|
|
||||||
def test_timeout_query_above_default_nomax(self):
|
def test_timeout_query_above_default_nomax(self):
|
||||||
searx.search.max_request_timeout = None
|
searx.search.max_request_timeout = None
|
||||||
search_query = searx.query.SearchQuery('test', [{'category': 'general', 'name': PUBLIC_ENGINE_NAME}],
|
search_query = SearchQuery('test', [EngineRef(PUBLIC_ENGINE_NAME, 'general')],
|
||||||
['general'], 'en-US', SAFESEARCH, PAGENO, None, 5.0,
|
['general'], 'en-US', SAFESEARCH, PAGENO, None, 5.0)
|
||||||
preferences=Preferences(['oscar'], ['general'], engines, []))
|
|
||||||
search = searx.search.Search(search_query)
|
search = searx.search.Search(search_query)
|
||||||
search.search()
|
search.search()
|
||||||
self.assertEqual(search.actual_timeout, 3.0)
|
self.assertEqual(search.actual_timeout, 3.0)
|
||||||
|
|
||||||
def test_timeout_query_below_default_nomax(self):
|
def test_timeout_query_below_default_nomax(self):
|
||||||
searx.search.max_request_timeout = None
|
searx.search.max_request_timeout = None
|
||||||
search_query = searx.query.SearchQuery('test', [{'category': 'general', 'name': PUBLIC_ENGINE_NAME}],
|
search_query = SearchQuery('test', [EngineRef(PUBLIC_ENGINE_NAME, 'general')],
|
||||||
['general'], 'en-US', SAFESEARCH, PAGENO, None, 1.0,
|
['general'], 'en-US', SAFESEARCH, PAGENO, None, 1.0)
|
||||||
preferences=Preferences(['oscar'], ['general'], engines, []))
|
|
||||||
search = searx.search.Search(search_query)
|
search = searx.search.Search(search_query)
|
||||||
search.search()
|
search.search()
|
||||||
self.assertEqual(search.actual_timeout, 1.0)
|
self.assertEqual(search.actual_timeout, 1.0)
|
||||||
|
|
||||||
def test_timeout_query_below_max(self):
|
def test_timeout_query_below_max(self):
|
||||||
searx.search.max_request_timeout = 10.0
|
searx.search.max_request_timeout = 10.0
|
||||||
search_query = searx.query.SearchQuery('test', [{'category': 'general', 'name': PUBLIC_ENGINE_NAME}],
|
search_query = SearchQuery('test', [EngineRef(PUBLIC_ENGINE_NAME, 'general')],
|
||||||
['general'], 'en-US', SAFESEARCH, PAGENO, None, 5.0,
|
['general'], 'en-US', SAFESEARCH, PAGENO, None, 5.0)
|
||||||
preferences=Preferences(['oscar'], ['general'], engines, []))
|
|
||||||
search = searx.search.Search(search_query)
|
search = searx.search.Search(search_query)
|
||||||
search.search()
|
search.search()
|
||||||
self.assertEqual(search.actual_timeout, 5.0)
|
self.assertEqual(search.actual_timeout, 5.0)
|
||||||
|
|
||||||
def test_timeout_query_above_max(self):
|
def test_timeout_query_above_max(self):
|
||||||
searx.search.max_request_timeout = 10.0
|
searx.search.max_request_timeout = 10.0
|
||||||
search_query = searx.query.SearchQuery('test', [{'category': 'general', 'name': PUBLIC_ENGINE_NAME}],
|
search_query = SearchQuery('test', [EngineRef(PUBLIC_ENGINE_NAME, 'general')],
|
||||||
['general'], 'en-US', SAFESEARCH, PAGENO, None, 15.0,
|
['general'], 'en-US', SAFESEARCH, PAGENO, None, 15.0)
|
||||||
preferences=Preferences(['oscar'], ['general'], engines, []))
|
|
||||||
search = searx.search.Search(search_query)
|
search = searx.search.Search(search_query)
|
||||||
search.search()
|
search.search()
|
||||||
self.assertEqual(search.actual_timeout, 10.0)
|
self.assertEqual(search.actual_timeout, 10.0)
|
||||||
|
|
||||||
def test_query_private_engine_without_token(self):
|
|
||||||
search_query = searx.query.SearchQuery('test', [{'category': 'general', 'name': PRIVATE_ENGINE_NAME}],
|
|
||||||
['general'], 'en-US', SAFESEARCH, PAGENO, None, 2.0,
|
|
||||||
preferences=Preferences(['oscar'], ['general'], engines, []))
|
|
||||||
search = searx.search.Search(search_query)
|
|
||||||
results = search.search()
|
|
||||||
self.assertEqual(results.results_length(), 0)
|
|
||||||
|
|
||||||
def test_query_private_engine_with_incorrect_token(self):
|
|
||||||
preferences_with_tokens = Preferences(['oscar'], ['general'], engines, [])
|
|
||||||
preferences_with_tokens.parse_dict({'tokens': 'bad-token'})
|
|
||||||
search_query = searx.query.SearchQuery('test', [{'category': 'general', 'name': PRIVATE_ENGINE_NAME}],
|
|
||||||
['general'], 'en-US', SAFESEARCH, PAGENO, None, 2.0,
|
|
||||||
preferences=preferences_with_tokens)
|
|
||||||
search = searx.search.Search(search_query)
|
|
||||||
results = search.search()
|
|
||||||
self.assertEqual(results.results_length(), 0)
|
|
||||||
|
|
||||||
def test_query_private_engine_with_correct_token(self):
|
|
||||||
preferences_with_tokens = Preferences(['oscar'], ['general'], engines, [])
|
|
||||||
preferences_with_tokens.parse_dict({'tokens': 'my-token'})
|
|
||||||
search_query = searx.query.SearchQuery('test', [{'category': 'general', 'name': PRIVATE_ENGINE_NAME}],
|
|
||||||
['general'], 'en-US', SAFESEARCH, PAGENO, None, 2.0,
|
|
||||||
preferences=preferences_with_tokens)
|
|
||||||
search = searx.search.Search(search_query)
|
|
||||||
results = search.search()
|
|
||||||
self.assertEqual(results.results_length(), 1)
|
|
||||||
|
|
||||||
def test_external_bang(self):
|
def test_external_bang(self):
|
||||||
search_query = searx.query.SearchQuery('yes yes',
|
search_query = SearchQuery('yes yes',
|
||||||
[{'category': 'general', 'name': PUBLIC_ENGINE_NAME}],
|
[EngineRef(PUBLIC_ENGINE_NAME, 'general')],
|
||||||
['general'], 'en-US', SAFESEARCH, PAGENO, None, None,
|
['general'], 'en-US', SAFESEARCH, PAGENO, None, None,
|
||||||
preferences=Preferences(['oscar'], ['general'], engines, [],),
|
|
||||||
external_bang="yt")
|
external_bang="yt")
|
||||||
search = searx.search.Search(search_query)
|
search = searx.search.Search(search_query)
|
||||||
results = search.search()
|
results = search.search()
|
||||||
# For checking if the user redirected with the youtube external bang
|
# For checking if the user redirected with the youtube external bang
|
||||||
self.assertTrue(results.redirect_url is not None)
|
self.assertTrue(results.redirect_url is not None)
|
||||||
|
|
||||||
search_query = searx.query.SearchQuery('youtube never gonna give you up',
|
search_query = SearchQuery('youtube never gonna give you up',
|
||||||
[{'category': 'general', 'name': PUBLIC_ENGINE_NAME}],
|
[EngineRef(PUBLIC_ENGINE_NAME, 'general')],
|
||||||
['general'], 'en-US', SAFESEARCH, PAGENO, None, None,
|
['general'], 'en-US', SAFESEARCH, PAGENO, None, None)
|
||||||
preferences=Preferences(['oscar'], ['general'], engines, []),)
|
|
||||||
|
|
||||||
search = searx.search.Search(search_query)
|
search = searx.search.Search(search_query)
|
||||||
results = search.search()
|
results = search.search()
|
||||||
|
54
tests/unit/test_webadapter.py
Normal file
54
tests/unit/test_webadapter.py
Normal file
@ -0,0 +1,54 @@
|
|||||||
|
# -*- coding: utf-8 -*-
|
||||||
|
|
||||||
|
from searx.testing import SearxTestCase
|
||||||
|
from searx.preferences import Preferences
|
||||||
|
from searx.engines import engines
|
||||||
|
|
||||||
|
import searx.search
|
||||||
|
from searx.search import EngineRef, SearchQuery
|
||||||
|
from searx.webadapter import validate_engineref_list
|
||||||
|
|
||||||
|
|
||||||
|
PRIVATE_ENGINE_NAME = 'general private offline'
|
||||||
|
TEST_ENGINES = [
|
||||||
|
{
|
||||||
|
'name': PRIVATE_ENGINE_NAME,
|
||||||
|
'engine': 'dummy-offline',
|
||||||
|
'categories': 'general',
|
||||||
|
'shortcut': 'do',
|
||||||
|
'timeout': 3.0,
|
||||||
|
'offline': True,
|
||||||
|
'tokens': ['my-token'],
|
||||||
|
},
|
||||||
|
]
|
||||||
|
SEARCHQUERY = [EngineRef(PRIVATE_ENGINE_NAME, 'general')]
|
||||||
|
|
||||||
|
|
||||||
|
class ValidateQueryCase(SearxTestCase):
|
||||||
|
|
||||||
|
@classmethod
|
||||||
|
def setUpClass(cls):
|
||||||
|
searx.engines.initialize_engines(TEST_ENGINES)
|
||||||
|
|
||||||
|
def test_query_private_engine_without_token(self):
|
||||||
|
preferences = Preferences(['oscar'], ['general'], engines, [])
|
||||||
|
valid, unknown, invalid_token = validate_engineref_list(SEARCHQUERY, preferences)
|
||||||
|
self.assertEqual(len(valid), 0)
|
||||||
|
self.assertEqual(len(unknown), 0)
|
||||||
|
self.assertEqual(len(invalid_token), 1)
|
||||||
|
|
||||||
|
def test_query_private_engine_with_incorrect_token(self):
|
||||||
|
preferences_with_tokens = Preferences(['oscar'], ['general'], engines, [])
|
||||||
|
preferences_with_tokens.parse_dict({'tokens': 'bad-token'})
|
||||||
|
valid, unknown, invalid_token = validate_engineref_list(SEARCHQUERY, preferences_with_tokens)
|
||||||
|
self.assertEqual(len(valid), 0)
|
||||||
|
self.assertEqual(len(unknown), 0)
|
||||||
|
self.assertEqual(len(invalid_token), 1)
|
||||||
|
|
||||||
|
def test_query_private_engine_with_correct_token(self):
|
||||||
|
preferences_with_tokens = Preferences(['oscar'], ['general'], engines, [])
|
||||||
|
preferences_with_tokens.parse_dict({'tokens': 'my-token'})
|
||||||
|
valid, unknown, invalid_token = validate_engineref_list(SEARCHQUERY, preferences_with_tokens)
|
||||||
|
self.assertEqual(len(valid), 1)
|
||||||
|
self.assertEqual(len(unknown), 0)
|
||||||
|
self.assertEqual(len(invalid_token), 0)
|
@ -30,7 +30,9 @@ import codecs
|
|||||||
import searx.query
|
import searx.query
|
||||||
import searx.search
|
import searx.search
|
||||||
import searx.engines
|
import searx.engines
|
||||||
|
import searx.webapdater
|
||||||
import searx.preferences
|
import searx.preferences
|
||||||
|
import searx.webadapter
|
||||||
import argparse
|
import argparse
|
||||||
|
|
||||||
searx.engines.initialize_engines(settings['engines'])
|
searx.engines.initialize_engines(settings['engines'])
|
||||||
@ -64,7 +66,7 @@ form = {
|
|||||||
preferences = searx.preferences.Preferences(['oscar'], searx.engines.categories.keys(), searx.engines.engines, [])
|
preferences = searx.preferences.Preferences(['oscar'], searx.engines.categories.keys(), searx.engines.engines, [])
|
||||||
preferences.key_value_settings['safesearch'].parse(args.safesearch)
|
preferences.key_value_settings['safesearch'].parse(args.safesearch)
|
||||||
|
|
||||||
search_query, raw_text_query = searx.search.get_search_query_from_webapp(preferences, form)
|
search_query, raw_text_query, _, _ = searx.webadapter.get_search_query_from_webapp(preferences, form)
|
||||||
search = searx.search.Search(search_query)
|
search = searx.search.Search(search_query)
|
||||||
result_container = search.search()
|
result_container = search.search()
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user