2013-10-14 23:09:13 +02:00
|
|
|
#!/usr/bin/env python
|
|
|
|
|
2013-10-15 00:33:18 +02:00
|
|
|
'''
|
|
|
|
searx is free software: you can redistribute it and/or modify
|
|
|
|
it under the terms of the GNU Affero General Public License as published by
|
|
|
|
the Free Software Foundation, either version 3 of the License, or
|
|
|
|
(at your option) any later version.
|
|
|
|
|
|
|
|
searx is distributed in the hope that it will be useful,
|
|
|
|
but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
|
|
GNU Affero General Public License for more details.
|
|
|
|
|
|
|
|
You should have received a copy of the GNU Affero General Public License
|
|
|
|
along with searx. If not, see < http://www.gnu.org/licenses/ >.
|
|
|
|
|
|
|
|
(C) 2013- by Adam Tauber, <asciimoo@gmail.com>
|
|
|
|
'''
|
|
|
|
|
2014-02-14 16:16:20 +01:00
|
|
|
if __name__ == '__main__':
|
|
|
|
from sys import path
|
|
|
|
from os.path import realpath, dirname
|
2015-02-10 15:23:56 +01:00
|
|
|
path.append(realpath(dirname(realpath(__file__)) + '/../'))
|
2014-02-14 16:16:20 +01:00
|
|
|
|
2015-01-17 21:54:40 +01:00
|
|
|
import hashlib
|
2016-10-16 23:40:56 +02:00
|
|
|
import hmac
|
|
|
|
import json
|
|
|
|
import os
|
2016-11-30 18:43:03 +01:00
|
|
|
import sys
|
|
|
|
|
2015-04-10 01:10:49 +02:00
|
|
|
import requests
|
2014-01-21 21:28:54 +01:00
|
|
|
|
2015-03-10 19:55:22 +01:00
|
|
|
from searx import logger
|
|
|
|
logger = logger.getChild('webapp')
|
|
|
|
|
|
|
|
try:
|
|
|
|
from pygments import highlight
|
|
|
|
from pygments.lexers import get_lexer_by_name
|
|
|
|
from pygments.formatters import HtmlFormatter
|
|
|
|
except:
|
|
|
|
logger.critical("cannot import dependency: pygments")
|
|
|
|
from sys import exit
|
|
|
|
exit(1)
|
2016-12-09 19:10:33 +01:00
|
|
|
from cgi import escape
|
2014-03-14 09:55:04 +01:00
|
|
|
from datetime import datetime, timedelta
|
2015-02-07 12:28:36 +01:00
|
|
|
from werkzeug.contrib.fixers import ProxyFix
|
2014-02-05 20:24:31 +01:00
|
|
|
from flask import (
|
|
|
|
Flask, request, render_template, url_for, Response, make_response,
|
|
|
|
redirect, send_from_directory
|
|
|
|
)
|
2016-07-04 22:46:43 +02:00
|
|
|
from flask_babel import Babel, gettext, format_date, format_decimal
|
2016-06-07 23:08:48 +02:00
|
|
|
from flask.json import jsonify
|
2016-10-22 19:07:37 +02:00
|
|
|
from searx import settings, searx_dir, searx_debug
|
2016-11-30 18:43:03 +01:00
|
|
|
from searx.exceptions import SearxParameterException
|
2014-02-05 20:24:31 +01:00
|
|
|
from searx.engines import (
|
2016-12-27 17:25:19 +01:00
|
|
|
categories, engines, engine_shortcuts, get_engines_stats, initialize_engines
|
2014-02-05 20:24:31 +01:00
|
|
|
)
|
2014-04-25 01:46:40 +02:00
|
|
|
from searx.utils import (
|
2017-01-06 13:23:30 +01:00
|
|
|
UnicodeWriter, highlight_content, html_to_text, get_resources_directory,
|
|
|
|
get_static_files, get_result_templates, get_themes, gen_useragent,
|
|
|
|
dict_subset, prettify_url
|
2014-04-25 01:46:40 +02:00
|
|
|
)
|
2014-11-18 11:37:42 +01:00
|
|
|
from searx.version import VERSION_STRING
|
2014-01-31 04:35:23 +01:00
|
|
|
from searx.languages import language_codes
|
2016-11-14 22:15:03 +01:00
|
|
|
from searx.search import SearchWithPlugins, get_search_query_from_webapp
|
|
|
|
from searx.query import RawTextQuery
|
2015-01-10 16:42:57 +01:00
|
|
|
from searx.autocomplete import searx_bang, backends as autocomplete_backends
|
2015-03-10 19:55:22 +01:00
|
|
|
from searx.plugins import plugins
|
2016-06-07 21:43:24 +02:00
|
|
|
from searx.preferences import Preferences, ValidationException
|
2016-11-19 20:53:51 +01:00
|
|
|
from searx.answerers import answerers
|
2016-11-30 18:43:03 +01:00
|
|
|
from searx.url_utils import urlencode, urlparse, urljoin
|
2013-12-01 23:52:49 +01:00
|
|
|
|
2017-03-23 15:33:18 +01:00
|
|
|
# check if the pyopenssl package is installed.
|
|
|
|
# It is needed for SSL connection without trouble, see #298
|
2015-04-26 18:58:31 +02:00
|
|
|
try:
|
|
|
|
import OpenSSL.SSL # NOQA
|
|
|
|
except ImportError:
|
2017-03-23 15:33:18 +01:00
|
|
|
logger.critical("The pyopenssl package has to be installed.\n"
|
2016-04-09 14:46:02 +02:00
|
|
|
"Some HTTPS connections will fail")
|
2015-04-26 18:58:31 +02:00
|
|
|
|
2016-11-30 18:43:03 +01:00
|
|
|
try:
|
|
|
|
from cStringIO import StringIO
|
|
|
|
except:
|
|
|
|
from io import StringIO
|
|
|
|
|
|
|
|
|
|
|
|
if sys.version_info[0] == 3:
|
|
|
|
unicode = str
|
|
|
|
|
2016-09-15 13:47:09 +02:00
|
|
|
# serve pages with HTTP/1.1
|
|
|
|
from werkzeug.serving import WSGIRequestHandler
|
2016-12-25 23:31:51 +01:00
|
|
|
WSGIRequestHandler.protocol_version = "HTTP/{}".format(settings['server'].get('http_protocol_version', '1.0'))
|
2013-10-14 23:09:13 +02:00
|
|
|
|
2017-01-06 13:23:30 +01:00
|
|
|
# about static
|
|
|
|
static_path = get_resources_directory(searx_dir, 'static', settings['ui']['static_path'])
|
|
|
|
logger.debug('static directory is %s', static_path)
|
|
|
|
static_files = get_static_files(static_path)
|
2014-10-09 19:26:02 +02:00
|
|
|
|
2017-01-06 13:23:30 +01:00
|
|
|
# about templates
|
2015-08-02 19:38:27 +02:00
|
|
|
default_theme = settings['ui']['default_theme']
|
2017-01-06 13:23:30 +01:00
|
|
|
templates_path = get_resources_directory(searx_dir, 'templates', settings['ui']['templates_path'])
|
|
|
|
logger.debug('templates directory is %s', templates_path)
|
|
|
|
themes = get_themes(static_path)
|
|
|
|
result_templates = get_result_templates(templates_path)
|
|
|
|
global_favicons = []
|
|
|
|
for indice, theme in enumerate(themes):
|
|
|
|
global_favicons.append([])
|
|
|
|
theme_img_path = os.path.join(static_path, 'themes', theme, 'img', 'icons')
|
|
|
|
for (dirpath, dirnames, filenames) in os.walk(theme_img_path):
|
|
|
|
global_favicons[indice].extend(filenames)
|
2014-04-25 01:46:40 +02:00
|
|
|
|
2017-01-06 13:23:30 +01:00
|
|
|
# Flask app
|
2014-01-19 22:59:01 +01:00
|
|
|
app = Flask(
|
|
|
|
__name__,
|
2014-04-25 01:46:40 +02:00
|
|
|
static_folder=static_path,
|
|
|
|
template_folder=templates_path
|
2014-01-19 22:59:01 +01:00
|
|
|
)
|
|
|
|
|
2015-02-14 01:42:06 +01:00
|
|
|
app.jinja_env.trim_blocks = True
|
|
|
|
app.jinja_env.lstrip_blocks = True
|
2014-01-19 22:59:01 +01:00
|
|
|
app.secret_key = settings['server']['secret_key']
|
2014-01-14 18:17:19 +01:00
|
|
|
|
2016-12-28 21:17:51 +01:00
|
|
|
if not searx_debug or os.environ.get("WERKZEUG_RUN_MAIN") == "true":
|
|
|
|
initialize_engines(settings['engines'])
|
|
|
|
|
2014-01-22 00:15:23 +01:00
|
|
|
babel = Babel(app)
|
|
|
|
|
2015-02-10 15:23:56 +01:00
|
|
|
rtl_locales = ['ar', 'arc', 'bcc', 'bqi', 'ckb', 'dv', 'fa', 'glk', 'he',
|
|
|
|
'ku', 'mzn', 'pnb'', ''ps', 'sd', 'ug', 'ur', 'yi']
|
|
|
|
|
2016-04-09 19:47:06 +02:00
|
|
|
# used when translating category names
|
2015-02-10 23:14:37 +01:00
|
|
|
_category_names = (gettext('files'),
|
|
|
|
gettext('general'),
|
|
|
|
gettext('music'),
|
|
|
|
gettext('social media'),
|
|
|
|
gettext('images'),
|
|
|
|
gettext('videos'),
|
|
|
|
gettext('it'),
|
|
|
|
gettext('news'),
|
2016-01-21 10:45:34 +01:00
|
|
|
gettext('map'),
|
2016-01-21 15:57:02 +01:00
|
|
|
gettext('science'))
|
2015-02-10 23:14:37 +01:00
|
|
|
|
2015-08-02 19:38:27 +02:00
|
|
|
outgoing_proxies = settings['outgoing'].get('proxies', None)
|
2014-01-19 23:04:09 +01:00
|
|
|
|
|
|
|
|
2014-01-22 00:15:23 +01:00
|
|
|
@babel.localeselector
|
|
|
|
def get_locale():
|
2014-01-22 00:59:18 +01:00
|
|
|
locale = request.accept_languages.best_match(settings['locales'].keys())
|
|
|
|
|
2016-04-09 18:32:07 +02:00
|
|
|
if request.preferences.get_value('locale') != '':
|
|
|
|
locale = request.preferences.get_value('locale')
|
2014-01-22 00:59:18 +01:00
|
|
|
|
|
|
|
if 'locale' in request.args\
|
|
|
|
and request.args['locale'] in settings['locales']:
|
|
|
|
locale = request.args['locale']
|
|
|
|
|
|
|
|
if 'locale' in request.form\
|
|
|
|
and request.form['locale'] in settings['locales']:
|
|
|
|
locale = request.form['locale']
|
|
|
|
|
|
|
|
return locale
|
2014-01-22 00:15:23 +01:00
|
|
|
|
|
|
|
|
2014-12-20 23:33:03 +01:00
|
|
|
# code-highlighter
|
|
|
|
@app.template_filter('code_highlighter')
|
|
|
|
def code_highlighter(codelines, language=None):
|
|
|
|
if not language:
|
|
|
|
language = 'text'
|
|
|
|
|
2014-12-22 16:26:45 +01:00
|
|
|
try:
|
|
|
|
# find lexer by programing language
|
|
|
|
lexer = get_lexer_by_name(language, stripall=True)
|
|
|
|
except:
|
|
|
|
# if lexer is not found, using default one
|
2015-01-15 18:39:40 +01:00
|
|
|
logger.debug('highlighter cannot find lexer for {0}'.format(language))
|
2014-12-22 16:26:45 +01:00
|
|
|
lexer = get_lexer_by_name('text', stripall=True)
|
|
|
|
|
2014-12-20 23:33:03 +01:00
|
|
|
html_code = ''
|
|
|
|
tmp_code = ''
|
|
|
|
last_line = None
|
|
|
|
|
|
|
|
# parse lines
|
|
|
|
for line, code in codelines:
|
|
|
|
if not last_line:
|
|
|
|
line_code_start = line
|
|
|
|
|
|
|
|
# new codeblock is detected
|
2014-12-22 16:26:45 +01:00
|
|
|
if last_line is not None and\
|
|
|
|
last_line + 1 != line:
|
2014-12-20 23:33:03 +01:00
|
|
|
|
|
|
|
# highlight last codepart
|
2014-12-22 16:26:45 +01:00
|
|
|
formatter = HtmlFormatter(linenos='inline',
|
|
|
|
linenostart=line_code_start)
|
2014-12-20 23:33:03 +01:00
|
|
|
html_code = html_code + highlight(tmp_code, lexer, formatter)
|
2014-12-22 16:26:45 +01:00
|
|
|
|
2014-12-20 23:33:03 +01:00
|
|
|
# reset conditions for next codepart
|
|
|
|
tmp_code = ''
|
|
|
|
line_code_start = line
|
|
|
|
|
|
|
|
# add codepart
|
|
|
|
tmp_code += code + '\n'
|
2014-12-22 16:26:45 +01:00
|
|
|
|
2014-12-20 23:33:03 +01:00
|
|
|
# update line
|
|
|
|
last_line = line
|
|
|
|
|
|
|
|
# highlight last codepart
|
|
|
|
formatter = HtmlFormatter(linenos='inline', linenostart=line_code_start)
|
|
|
|
html_code = html_code + highlight(tmp_code, lexer, formatter)
|
|
|
|
|
|
|
|
return html_code
|
|
|
|
|
|
|
|
|
2015-02-15 19:09:17 +01:00
|
|
|
# Extract domain from url
|
|
|
|
@app.template_filter('extract_domain')
|
|
|
|
def extract_domain(url):
|
|
|
|
return urlparse(url)[1]
|
|
|
|
|
|
|
|
|
2014-01-14 18:17:19 +01:00
|
|
|
def get_base_url():
|
2014-01-19 00:17:02 +01:00
|
|
|
if settings['server']['base_url']:
|
|
|
|
hostname = settings['server']['base_url']
|
2014-01-14 18:17:19 +01:00
|
|
|
else:
|
|
|
|
scheme = 'http'
|
|
|
|
if request.is_secure:
|
|
|
|
scheme = 'https'
|
|
|
|
hostname = url_for('index', _external=True, _scheme=scheme)
|
|
|
|
return hostname
|
|
|
|
|
|
|
|
|
2014-04-25 01:46:40 +02:00
|
|
|
def get_current_theme_name(override=None):
|
|
|
|
"""Returns theme name.
|
|
|
|
|
|
|
|
Checks in this order:
|
|
|
|
1. override
|
|
|
|
2. cookies
|
|
|
|
3. settings"""
|
|
|
|
|
2017-01-13 22:15:11 +01:00
|
|
|
if override and (override in themes or override == '__common__'):
|
2014-04-25 01:46:40 +02:00
|
|
|
return override
|
2016-04-08 16:38:05 +02:00
|
|
|
theme_name = request.args.get('theme', request.preferences.get_value('theme'))
|
2014-04-25 01:46:40 +02:00
|
|
|
if theme_name not in themes:
|
|
|
|
theme_name = default_theme
|
|
|
|
return theme_name
|
|
|
|
|
|
|
|
|
2015-01-01 18:59:53 +01:00
|
|
|
def get_result_template(theme, template_name):
|
|
|
|
themed_path = theme + '/result_templates/' + template_name
|
|
|
|
if themed_path in result_templates:
|
|
|
|
return themed_path
|
|
|
|
return 'result_templates/' + template_name
|
|
|
|
|
|
|
|
|
2014-04-25 01:46:40 +02:00
|
|
|
def url_for_theme(endpoint, override_theme=None, **values):
|
2015-01-01 17:48:12 +01:00
|
|
|
if endpoint == 'static' and values.get('filename'):
|
2014-04-25 01:46:40 +02:00
|
|
|
theme_name = get_current_theme_name(override=override_theme)
|
2015-01-01 17:48:12 +01:00
|
|
|
filename_with_theme = "themes/{}/{}".format(theme_name, values['filename'])
|
|
|
|
if filename_with_theme in static_files:
|
|
|
|
values['filename'] = filename_with_theme
|
2014-04-25 01:46:40 +02:00
|
|
|
return url_for(endpoint, **values)
|
|
|
|
|
|
|
|
|
2016-10-17 00:22:41 +02:00
|
|
|
def proxify(url):
|
|
|
|
if url.startswith('//'):
|
|
|
|
url = 'https:' + url
|
|
|
|
|
|
|
|
if not settings.get('result_proxy'):
|
|
|
|
return url
|
|
|
|
|
2016-10-29 23:12:24 +02:00
|
|
|
url_params = dict(mortyurl=url.encode('utf-8'))
|
|
|
|
|
|
|
|
if settings['result_proxy'].get('key'):
|
|
|
|
url_params['mortyhash'] = hmac.new(settings['result_proxy']['key'],
|
|
|
|
url.encode('utf-8'),
|
|
|
|
hashlib.sha256).hexdigest()
|
2016-10-17 00:22:41 +02:00
|
|
|
|
|
|
|
return '{0}?{1}'.format(settings['result_proxy']['url'],
|
2016-10-29 23:12:24 +02:00
|
|
|
urlencode(url_params))
|
2016-10-17 00:22:41 +02:00
|
|
|
|
|
|
|
|
2015-01-16 16:26:15 +01:00
|
|
|
def image_proxify(url):
|
|
|
|
|
|
|
|
if url.startswith('//'):
|
|
|
|
url = 'https:' + url
|
|
|
|
|
2016-04-09 18:32:07 +02:00
|
|
|
if not request.preferences.get_value('image_proxy'):
|
2015-01-16 16:26:15 +01:00
|
|
|
return url
|
|
|
|
|
2016-10-30 21:15:18 +01:00
|
|
|
if settings.get('result_proxy'):
|
|
|
|
return proxify(url)
|
|
|
|
|
2016-10-17 01:32:31 +02:00
|
|
|
h = hmac.new(settings['server']['secret_key'], url.encode('utf-8'), hashlib.sha256).hexdigest()
|
2015-01-18 09:54:24 +01:00
|
|
|
|
2015-01-16 16:26:15 +01:00
|
|
|
return '{0}?{1}'.format(url_for('image_proxy'),
|
2015-02-09 12:24:54 +01:00
|
|
|
urlencode(dict(url=url.encode('utf-8'), h=h)))
|
2015-01-16 16:26:15 +01:00
|
|
|
|
|
|
|
|
2014-04-25 01:46:40 +02:00
|
|
|
def render(template_name, override_theme=None, **kwargs):
|
2016-04-09 18:26:29 +02:00
|
|
|
disabled_engines = request.preferences.engines.get_disabled()
|
2014-03-29 16:29:19 +01:00
|
|
|
|
2016-04-09 18:26:29 +02:00
|
|
|
enabled_categories = set(category for engine_name in engines
|
|
|
|
for category in engines[engine_name].categories
|
|
|
|
if (engine_name, category) not in disabled_engines)
|
2014-03-29 16:29:19 +01:00
|
|
|
|
2014-10-19 12:41:04 +02:00
|
|
|
if 'categories' not in kwargs:
|
2014-03-08 14:00:20 +01:00
|
|
|
kwargs['categories'] = ['general']
|
|
|
|
kwargs['categories'].extend(x for x in
|
|
|
|
sorted(categories.keys())
|
2014-03-08 14:03:42 +01:00
|
|
|
if x != 'general'
|
2016-04-09 18:26:29 +02:00
|
|
|
and x in enabled_categories)
|
2014-03-29 16:29:19 +01:00
|
|
|
|
2015-06-05 08:48:47 +02:00
|
|
|
if 'all_categories' not in kwargs:
|
|
|
|
kwargs['all_categories'] = ['general']
|
|
|
|
kwargs['all_categories'].extend(x for x in
|
|
|
|
sorted(categories.keys())
|
|
|
|
if x != 'general')
|
|
|
|
|
2014-10-19 12:41:04 +02:00
|
|
|
if 'selected_categories' not in kwargs:
|
2013-10-17 21:46:35 +02:00
|
|
|
kwargs['selected_categories'] = []
|
2014-09-06 15:21:29 +02:00
|
|
|
for arg in request.args:
|
|
|
|
if arg.startswith('category_'):
|
|
|
|
c = arg.split('_', 1)[1]
|
|
|
|
if c in categories:
|
|
|
|
kwargs['selected_categories'].append(c)
|
2015-06-05 08:48:47 +02:00
|
|
|
|
2014-09-06 15:21:29 +02:00
|
|
|
if not kwargs['selected_categories']:
|
2016-04-08 16:38:05 +02:00
|
|
|
cookie_categories = request.preferences.get_value('categories')
|
2013-10-17 21:46:35 +02:00
|
|
|
for ccateg in cookie_categories:
|
2016-04-09 18:32:07 +02:00
|
|
|
kwargs['selected_categories'].append(ccateg)
|
2015-06-05 08:48:47 +02:00
|
|
|
|
2014-09-06 15:21:29 +02:00
|
|
|
if not kwargs['selected_categories']:
|
|
|
|
kwargs['selected_categories'] = ['general']
|
2014-03-29 16:29:19 +01:00
|
|
|
|
2014-10-19 12:41:04 +02:00
|
|
|
if 'autocomplete' not in kwargs:
|
2016-04-09 18:32:07 +02:00
|
|
|
kwargs['autocomplete'] = request.preferences.get_value('autocomplete')
|
2014-03-29 16:29:19 +01:00
|
|
|
|
2015-02-10 15:23:56 +01:00
|
|
|
if get_locale() in rtl_locales and 'rtl' not in kwargs:
|
|
|
|
kwargs['rtl'] = True
|
|
|
|
|
2014-11-18 11:37:42 +01:00
|
|
|
kwargs['searx_version'] = VERSION_STRING
|
|
|
|
|
2016-04-08 16:38:05 +02:00
|
|
|
kwargs['method'] = request.preferences.get_value('method')
|
2014-03-29 16:45:22 +01:00
|
|
|
|
2016-04-08 16:38:05 +02:00
|
|
|
kwargs['safesearch'] = str(request.preferences.get_value('safesearch'))
|
2015-02-08 21:53:37 +01:00
|
|
|
|
2016-08-06 06:34:56 +02:00
|
|
|
kwargs['language_codes'] = language_codes
|
|
|
|
if 'current_language' not in kwargs:
|
|
|
|
kwargs['current_language'] = request.preferences.get_value('language')
|
|
|
|
|
2014-04-25 01:46:40 +02:00
|
|
|
# override url_for function in templates
|
|
|
|
kwargs['url_for'] = url_for_theme
|
|
|
|
|
2015-01-16 16:26:15 +01:00
|
|
|
kwargs['image_proxify'] = image_proxify
|
|
|
|
|
2016-10-17 00:22:41 +02:00
|
|
|
kwargs['proxify'] = proxify if settings.get('result_proxy') else None
|
|
|
|
|
2015-01-01 18:59:53 +01:00
|
|
|
kwargs['get_result_template'] = get_result_template
|
|
|
|
|
2014-04-25 01:46:40 +02:00
|
|
|
kwargs['theme'] = get_current_theme_name(override=override_theme)
|
2014-11-18 19:55:39 +01:00
|
|
|
|
2014-09-22 22:42:29 +02:00
|
|
|
kwargs['template_name'] = template_name
|
2014-04-25 01:46:40 +02:00
|
|
|
|
2015-01-20 16:29:54 +01:00
|
|
|
kwargs['cookies'] = request.cookies
|
|
|
|
|
2016-11-14 22:07:23 +01:00
|
|
|
kwargs['errors'] = request.errors
|
|
|
|
|
2016-02-27 18:16:40 +01:00
|
|
|
kwargs['instance_name'] = settings['general']['instance_name']
|
|
|
|
|
2016-09-04 15:56:46 +02:00
|
|
|
kwargs['results_on_new_tab'] = request.preferences.get_value('results_on_new_tab')
|
|
|
|
|
2016-11-30 18:43:03 +01:00
|
|
|
kwargs['unicode'] = unicode
|
|
|
|
|
2015-04-12 19:24:01 +02:00
|
|
|
kwargs['scripts'] = set()
|
|
|
|
for plugin in request.user_plugins:
|
|
|
|
for script in plugin.js_dependencies:
|
|
|
|
kwargs['scripts'].add(script)
|
|
|
|
|
|
|
|
kwargs['styles'] = set()
|
|
|
|
for plugin in request.user_plugins:
|
|
|
|
for css in plugin.css_dependencies:
|
|
|
|
kwargs['styles'].add(css)
|
|
|
|
|
2014-04-25 01:46:40 +02:00
|
|
|
return render_template(
|
|
|
|
'{}/{}'.format(kwargs['theme'], template_name), **kwargs)
|
2013-10-15 20:50:12 +02:00
|
|
|
|
2014-01-19 22:59:01 +01:00
|
|
|
|
2015-03-10 20:44:02 +01:00
|
|
|
@app.before_request
|
|
|
|
def pre_request():
|
2016-11-14 22:07:23 +01:00
|
|
|
request.errors = []
|
|
|
|
|
2016-11-30 18:43:03 +01:00
|
|
|
preferences = Preferences(themes, list(categories.keys()), engines, plugins)
|
2016-11-15 09:56:18 +01:00
|
|
|
request.preferences = preferences
|
2016-09-04 15:56:46 +02:00
|
|
|
try:
|
|
|
|
preferences.parse_cookies(request.cookies)
|
|
|
|
except:
|
2016-11-14 22:07:23 +01:00
|
|
|
request.errors.append(gettext('Invalid settings, please edit your preferences'))
|
2016-04-08 16:38:05 +02:00
|
|
|
|
2016-11-14 22:07:23 +01:00
|
|
|
# merge GET, POST vars
|
2016-10-22 13:10:31 +02:00
|
|
|
# request.form
|
2015-03-10 22:45:59 +01:00
|
|
|
request.form = dict(request.form.items())
|
2015-03-15 12:13:24 +01:00
|
|
|
for k, v in request.args.items():
|
2015-03-10 22:45:59 +01:00
|
|
|
if k not in request.form:
|
|
|
|
request.form[k] = v
|
2015-03-10 20:44:02 +01:00
|
|
|
|
2016-10-22 13:10:31 +02:00
|
|
|
# request.user_plugins
|
2015-03-10 20:44:02 +01:00
|
|
|
request.user_plugins = []
|
2016-04-08 16:38:05 +02:00
|
|
|
allowed_plugins = preferences.plugins.get_enabled()
|
|
|
|
disabled_plugins = preferences.plugins.get_disabled()
|
2015-03-10 20:44:02 +01:00
|
|
|
for plugin in plugins:
|
2015-03-11 18:57:36 +01:00
|
|
|
if ((plugin.default_on and plugin.id not in disabled_plugins)
|
|
|
|
or plugin.id in allowed_plugins):
|
2015-03-10 20:44:02 +01:00
|
|
|
request.user_plugins.append(plugin)
|
|
|
|
|
|
|
|
|
2017-01-20 18:52:47 +01:00
|
|
|
def index_error(output_format, error_message):
|
|
|
|
if output_format == 'json':
|
|
|
|
return Response(json.dumps({'error': error_message}),
|
|
|
|
mimetype='application/json')
|
|
|
|
elif output_format == 'csv':
|
|
|
|
response = Response('', mimetype='application/csv')
|
|
|
|
cont_disp = 'attachment;Filename=searx.csv'
|
|
|
|
response.headers.add('Content-Disposition', cont_disp)
|
|
|
|
return response
|
|
|
|
elif output_format == 'rss':
|
|
|
|
response_rss = render(
|
|
|
|
'opensearch_response_rss.xml',
|
|
|
|
results=[],
|
|
|
|
q=request.form['q'] if 'q' in request.form else '',
|
|
|
|
number_of_results=0,
|
|
|
|
base_url=get_base_url(),
|
|
|
|
error_message=error_message
|
|
|
|
)
|
|
|
|
return Response(response_rss, mimetype='text/xml')
|
|
|
|
else:
|
|
|
|
# html
|
|
|
|
request.errors.append(gettext('search error'))
|
|
|
|
return render(
|
|
|
|
'index.html',
|
|
|
|
)
|
|
|
|
|
|
|
|
|
2014-05-01 10:14:47 +02:00
|
|
|
@app.route('/search', methods=['GET', 'POST'])
|
2014-01-14 18:19:21 +01:00
|
|
|
@app.route('/', methods=['GET', 'POST'])
|
2013-10-14 23:09:13 +02:00
|
|
|
def index():
|
2014-01-31 07:08:24 +01:00
|
|
|
"""Render index page.
|
|
|
|
|
|
|
|
Supported outputs: html, json, csv, rss.
|
|
|
|
"""
|
2013-11-04 00:18:07 +01:00
|
|
|
|
2017-01-20 18:52:47 +01:00
|
|
|
# output_format
|
|
|
|
output_format = request.form.get('format', 'html')
|
|
|
|
if output_format not in ['html', 'csv', 'json', 'rss']:
|
|
|
|
output_format = 'html'
|
|
|
|
|
|
|
|
# check if there is query
|
2016-11-14 22:07:23 +01:00
|
|
|
if request.form.get('q') is None:
|
2017-01-20 18:52:47 +01:00
|
|
|
if output_format == 'html':
|
|
|
|
return render(
|
|
|
|
'index.html',
|
|
|
|
)
|
|
|
|
else:
|
|
|
|
return index_error(output_format, 'No query'), 400
|
2014-02-07 03:15:34 +01:00
|
|
|
|
2016-10-22 13:10:31 +02:00
|
|
|
# search
|
|
|
|
search_query = None
|
|
|
|
result_container = None
|
2014-02-07 01:19:07 +01:00
|
|
|
try:
|
2016-11-02 14:52:22 +01:00
|
|
|
search_query = get_search_query_from_webapp(request.preferences, request.form)
|
2016-10-22 13:10:31 +02:00
|
|
|
# search = Search(search_query) # without plugins
|
2017-01-02 12:06:04 +01:00
|
|
|
search = SearchWithPlugins(search_query, request.user_plugins, request)
|
2016-10-22 13:10:31 +02:00
|
|
|
result_container = search.search()
|
2017-01-20 18:52:47 +01:00
|
|
|
except Exception as e:
|
|
|
|
# log exception
|
2016-11-14 15:49:06 +01:00
|
|
|
logger.exception('search error')
|
2014-01-29 20:52:04 +01:00
|
|
|
|
2017-01-20 18:52:47 +01:00
|
|
|
# is it an invalid input parameter or something else ?
|
|
|
|
if (issubclass(e.__class__, SearxParameterException)):
|
|
|
|
return index_error(output_format, e.message), 400
|
|
|
|
else:
|
|
|
|
return index_error(output_format, gettext('search error')), 500
|
|
|
|
|
|
|
|
# results
|
2016-10-22 13:10:31 +02:00
|
|
|
results = result_container.get_ordered_results()
|
2017-01-20 18:52:47 +01:00
|
|
|
number_of_results = result_container.results_number()
|
|
|
|
if number_of_results < result_container.results_length():
|
|
|
|
number_of_results = 0
|
2014-02-07 01:19:07 +01:00
|
|
|
|
2016-10-22 13:10:31 +02:00
|
|
|
# UI
|
2016-11-02 14:52:22 +01:00
|
|
|
advanced_search = request.form.get('advanced_search', None)
|
2016-07-16 21:41:02 +02:00
|
|
|
|
2016-10-22 13:10:31 +02:00
|
|
|
# output
|
2016-07-16 21:41:02 +02:00
|
|
|
for result in results:
|
2016-10-22 13:10:31 +02:00
|
|
|
if output_format == 'html':
|
2016-09-05 22:22:25 +02:00
|
|
|
if 'content' in result and result['content']:
|
2016-11-30 18:43:03 +01:00
|
|
|
result['content'] = highlight_content(escape(result['content'][:1024]), search_query.query)
|
|
|
|
result['title'] = highlight_content(escape(result['title'] or u''), search_query.query)
|
2014-01-10 23:38:08 +01:00
|
|
|
else:
|
2015-04-12 17:37:01 +02:00
|
|
|
if result.get('content'):
|
2014-01-10 23:38:08 +01:00
|
|
|
result['content'] = html_to_text(result['content']).strip()
|
2014-02-04 19:42:32 +01:00
|
|
|
# removing html content and whitespace duplications
|
2015-04-12 17:37:01 +02:00
|
|
|
result['title'] = ' '.join(html_to_text(result['title']).strip().split())
|
2014-06-24 16:30:04 +02:00
|
|
|
|
2015-01-29 19:44:52 +01:00
|
|
|
result['pretty_url'] = prettify_url(result['url'])
|
2013-11-15 19:28:30 +01:00
|
|
|
|
2014-03-14 09:55:04 +01:00
|
|
|
# TODO, check if timezone is calculated right
|
|
|
|
if 'publishedDate' in result:
|
2016-03-29 11:59:16 +02:00
|
|
|
try: # test if publishedDate >= 1900 (datetime module bug)
|
|
|
|
result['pubdate'] = result['publishedDate'].strftime('%Y-%m-%d %H:%M:%S%z')
|
|
|
|
except ValueError:
|
|
|
|
result['publishedDate'] = None
|
2014-03-14 09:55:04 +01:00
|
|
|
else:
|
2016-03-29 11:59:16 +02:00
|
|
|
if result['publishedDate'].replace(tzinfo=None) >= datetime.now() - timedelta(days=1):
|
|
|
|
timedifference = datetime.now() - result['publishedDate'].replace(tzinfo=None)
|
|
|
|
minutes = int((timedifference.seconds / 60) % 60)
|
|
|
|
hours = int(timedifference.seconds / 60 / 60)
|
|
|
|
if hours == 0:
|
2016-05-24 12:49:21 +02:00
|
|
|
result['publishedDate'] = gettext(u'{minutes} minute(s) ago').format(minutes=minutes)
|
2016-03-29 11:59:16 +02:00
|
|
|
else:
|
2016-05-24 12:49:21 +02:00
|
|
|
result['publishedDate'] = gettext(u'{hours} hour(s), {minutes} minute(s) ago').format(hours=hours, minutes=minutes) # noqa
|
2016-03-29 11:59:16 +02:00
|
|
|
else:
|
|
|
|
result['publishedDate'] = format_date(result['publishedDate'])
|
2014-03-14 09:55:04 +01:00
|
|
|
|
2016-10-22 13:10:31 +02:00
|
|
|
if output_format == 'json':
|
2016-11-30 18:43:03 +01:00
|
|
|
return Response(json.dumps({'query': search_query.query.decode('utf-8'),
|
2016-07-16 21:37:40 +02:00
|
|
|
'number_of_results': number_of_results,
|
2016-12-10 21:38:34 +01:00
|
|
|
'results': results,
|
|
|
|
'answers': list(result_container.answers),
|
2017-01-14 09:40:37 +01:00
|
|
|
'corrections': list(result_container.corrections),
|
2016-12-10 21:38:34 +01:00
|
|
|
'infoboxes': result_container.infoboxes,
|
|
|
|
'suggestions': list(result_container.suggestions)}),
|
2014-01-20 02:31:20 +01:00
|
|
|
mimetype='application/json')
|
2016-10-22 13:10:31 +02:00
|
|
|
elif output_format == 'csv':
|
2016-11-30 18:43:03 +01:00
|
|
|
csv = UnicodeWriter(StringIO())
|
2013-12-02 21:36:09 +01:00
|
|
|
keys = ('title', 'url', 'content', 'host', 'engine', 'score')
|
2015-10-03 17:26:07 +02:00
|
|
|
csv.writerow(keys)
|
2016-07-16 21:41:02 +02:00
|
|
|
for row in results:
|
2015-10-03 17:26:07 +02:00
|
|
|
row['host'] = row['parsed_url'].netloc
|
|
|
|
csv.writerow([row.get(key, '') for key in keys])
|
|
|
|
csv.stream.seek(0)
|
2013-11-15 19:28:30 +01:00
|
|
|
response = Response(csv.stream.read(), mimetype='application/csv')
|
2016-11-30 18:43:03 +01:00
|
|
|
cont_disp = 'attachment;Filename=searx_-_{0}.csv'.format(search_query.query)
|
2014-02-07 02:45:12 +01:00
|
|
|
response.headers.add('Content-Disposition', cont_disp)
|
2013-11-15 18:55:18 +01:00
|
|
|
return response
|
2016-10-22 13:10:31 +02:00
|
|
|
elif output_format == 'rss':
|
2014-01-19 22:59:01 +01:00
|
|
|
response_rss = render(
|
|
|
|
'opensearch_response_rss.xml',
|
2016-07-16 21:41:02 +02:00
|
|
|
results=results,
|
2016-11-02 14:52:22 +01:00
|
|
|
q=request.form['q'],
|
2016-07-16 21:37:40 +02:00
|
|
|
number_of_results=number_of_results,
|
2017-01-13 22:15:11 +01:00
|
|
|
base_url=get_base_url(),
|
|
|
|
override_theme='__common__',
|
2014-01-19 22:59:01 +01:00
|
|
|
)
|
2014-01-14 22:18:21 +01:00
|
|
|
return Response(response_rss, mimetype='text/xml')
|
2014-01-14 18:17:19 +01:00
|
|
|
|
2014-01-19 22:59:01 +01:00
|
|
|
return render(
|
|
|
|
'results.html',
|
2016-07-16 21:41:02 +02:00
|
|
|
results=results,
|
2016-11-02 14:52:22 +01:00
|
|
|
q=request.form['q'],
|
2016-10-22 13:10:31 +02:00
|
|
|
selected_categories=search_query.categories,
|
|
|
|
pageno=search_query.pageno,
|
|
|
|
time_range=search_query.time_range,
|
2016-07-16 21:37:40 +02:00
|
|
|
number_of_results=format_decimal(number_of_results),
|
2016-10-22 13:10:31 +02:00
|
|
|
advanced_search=advanced_search,
|
|
|
|
suggestions=result_container.suggestions,
|
|
|
|
answers=result_container.answers,
|
2017-01-14 09:40:37 +01:00
|
|
|
corrections=result_container.corrections,
|
2016-10-22 13:10:31 +02:00
|
|
|
infoboxes=result_container.infoboxes,
|
|
|
|
paging=result_container.paging,
|
2016-11-06 03:51:38 +01:00
|
|
|
current_language=search_query.lang,
|
2014-03-04 18:53:56 +01:00
|
|
|
base_url=get_base_url(),
|
2014-12-12 19:09:02 +01:00
|
|
|
theme=get_current_theme_name(),
|
2014-12-13 21:37:28 +01:00
|
|
|
favicons=global_favicons[themes.index(get_current_theme_name())]
|
2014-01-19 22:59:01 +01:00
|
|
|
)
|
2014-01-01 22:16:53 +01:00
|
|
|
|
2013-10-14 23:09:13 +02:00
|
|
|
|
2013-10-21 00:28:48 +02:00
|
|
|
@app.route('/about', methods=['GET'])
|
|
|
|
def about():
|
2014-01-31 07:08:24 +01:00
|
|
|
"""Render about page"""
|
2014-03-21 12:19:48 +01:00
|
|
|
return render(
|
|
|
|
'about.html',
|
|
|
|
)
|
2014-01-17 16:23:23 +01:00
|
|
|
|
|
|
|
|
2014-03-20 10:28:24 +01:00
|
|
|
@app.route('/autocompleter', methods=['GET', 'POST'])
|
|
|
|
def autocompleter():
|
|
|
|
"""Return autocompleter results"""
|
2014-03-29 16:29:19 +01:00
|
|
|
|
2014-10-01 17:18:18 +02:00
|
|
|
# set blocked engines
|
2016-04-09 18:26:29 +02:00
|
|
|
disabled_engines = request.preferences.engines.get_disabled()
|
2014-10-01 17:18:18 +02:00
|
|
|
|
|
|
|
# parse query
|
2016-11-30 18:43:03 +01:00
|
|
|
raw_text_query = RawTextQuery(request.form.get('q', u'').encode('utf-8'), disabled_engines)
|
2016-10-22 13:10:31 +02:00
|
|
|
raw_text_query.parse_query()
|
2014-03-29 16:29:19 +01:00
|
|
|
|
2014-10-01 17:18:18 +02:00
|
|
|
# check if search query is set
|
2016-10-22 13:10:31 +02:00
|
|
|
if not raw_text_query.getSearchQuery():
|
2015-01-10 15:27:42 +01:00
|
|
|
return '', 400
|
2014-03-29 16:29:19 +01:00
|
|
|
|
2016-04-08 16:38:05 +02:00
|
|
|
# run autocompleter
|
|
|
|
completer = autocomplete_backends.get(request.preferences.get_value('autocomplete'))
|
2014-03-29 16:29:19 +01:00
|
|
|
|
2015-01-10 16:42:57 +01:00
|
|
|
# parse searx specific autocompleter results like !bang
|
2016-10-22 13:10:31 +02:00
|
|
|
raw_results = searx_bang(raw_text_query)
|
2015-01-10 16:42:57 +01:00
|
|
|
|
2015-01-25 22:52:48 +01:00
|
|
|
# normal autocompletion results only appear if max 3 inner results returned
|
|
|
|
if len(raw_results) <= 3 and completer:
|
2016-03-30 02:53:31 +02:00
|
|
|
# get language from cookie
|
2016-04-09 18:32:07 +02:00
|
|
|
language = request.preferences.get_value('language')
|
2016-03-30 02:53:31 +02:00
|
|
|
if not language or language == 'all':
|
|
|
|
language = 'en'
|
|
|
|
else:
|
2016-08-06 06:34:56 +02:00
|
|
|
language = language.split('-')[0]
|
2015-01-10 19:55:21 +01:00
|
|
|
# run autocompletion
|
2016-10-22 13:10:31 +02:00
|
|
|
raw_results.extend(completer(raw_text_query.getSearchQuery(), language))
|
2014-10-01 17:18:18 +02:00
|
|
|
|
|
|
|
# parse results (write :language and !engine back to result string)
|
|
|
|
results = []
|
|
|
|
for result in raw_results:
|
2016-10-22 13:10:31 +02:00
|
|
|
raw_text_query.changeSearchQuery(result)
|
2014-10-01 17:18:18 +02:00
|
|
|
|
|
|
|
# add parsed result
|
2016-10-22 13:10:31 +02:00
|
|
|
results.append(raw_text_query.getFullQuery())
|
2014-03-29 16:29:19 +01:00
|
|
|
|
2014-10-01 17:18:18 +02:00
|
|
|
# return autocompleter results
|
2016-11-02 14:52:22 +01:00
|
|
|
if request.form.get('format') == 'x-suggestions':
|
2016-10-22 13:10:31 +02:00
|
|
|
return Response(json.dumps([raw_text_query.query, results]),
|
2014-03-29 16:29:19 +01:00
|
|
|
mimetype='application/json')
|
2015-01-25 22:52:48 +01:00
|
|
|
|
|
|
|
return Response(json.dumps(results),
|
|
|
|
mimetype='application/json')
|
2014-03-20 10:28:24 +01:00
|
|
|
|
|
|
|
|
2014-01-01 22:16:53 +01:00
|
|
|
@app.route('/preferences', methods=['GET', 'POST'])
|
|
|
|
def preferences():
|
2016-04-08 16:38:05 +02:00
|
|
|
"""Render preferences page && save user preferences"""
|
2014-02-07 00:35:15 +01:00
|
|
|
|
2016-04-08 16:38:05 +02:00
|
|
|
# save preferences
|
|
|
|
if request.method == 'POST':
|
|
|
|
resp = make_response(redirect(urljoin(settings['server']['base_url'], url_for('index'))))
|
|
|
|
try:
|
|
|
|
request.preferences.parse_form(request.form)
|
|
|
|
except ValidationException:
|
2016-11-14 22:07:23 +01:00
|
|
|
request.errors.append(gettext('Invalid settings, please edit your preferences'))
|
2016-04-08 16:38:05 +02:00
|
|
|
return resp
|
|
|
|
return request.preferences.save(resp)
|
|
|
|
|
|
|
|
# render preferences
|
|
|
|
image_proxy = request.preferences.get_value('image_proxy')
|
|
|
|
lang = request.preferences.get_value('language')
|
2016-04-09 18:26:29 +02:00
|
|
|
disabled_engines = request.preferences.engines.get_disabled()
|
2016-04-08 16:38:05 +02:00
|
|
|
allowed_plugins = request.preferences.plugins.get_enabled()
|
2015-05-30 12:15:23 +02:00
|
|
|
|
|
|
|
# stats for preferences page
|
|
|
|
stats = {}
|
|
|
|
|
|
|
|
for c in categories:
|
|
|
|
for e in categories[c]:
|
|
|
|
stats[e.name] = {'time': None,
|
|
|
|
'warn_timeout': False,
|
|
|
|
'warn_time': False}
|
2015-08-02 19:38:27 +02:00
|
|
|
if e.timeout > settings['outgoing']['request_timeout']:
|
2015-05-30 12:15:23 +02:00
|
|
|
stats[e.name]['warn_timeout'] = True
|
|
|
|
|
2016-11-05 13:45:20 +01:00
|
|
|
# get first element [0], the engine time,
|
|
|
|
# and then the second element [1] : the time (the first one is the label)
|
2015-05-30 12:15:23 +02:00
|
|
|
for engine_stat in get_engines_stats()[0][1]:
|
|
|
|
stats[engine_stat.get('name')]['time'] = round(engine_stat.get('avg'), 3)
|
2015-08-02 19:38:27 +02:00
|
|
|
if engine_stat.get('avg') > settings['outgoing']['request_timeout']:
|
2015-05-30 12:15:23 +02:00
|
|
|
stats[engine_stat.get('name')]['warn_time'] = True
|
|
|
|
# end of stats
|
|
|
|
|
2014-01-22 01:20:38 +01:00
|
|
|
return render('preferences.html',
|
|
|
|
locales=settings['locales'],
|
2014-01-31 04:35:23 +01:00
|
|
|
current_locale=get_locale(),
|
2015-01-16 17:37:34 +01:00
|
|
|
image_proxy=image_proxy,
|
2015-04-07 18:09:38 +02:00
|
|
|
engines_by_category=categories,
|
2015-05-30 12:15:23 +02:00
|
|
|
stats=stats,
|
2016-11-19 20:53:51 +01:00
|
|
|
answerers=[{'info': a.self_info(), 'keywords': a.keywords} for a in answerers],
|
2016-04-09 18:48:23 +02:00
|
|
|
disabled_engines=disabled_engines,
|
2014-03-29 16:29:19 +01:00
|
|
|
autocomplete_backends=autocomplete_backends,
|
2014-04-25 01:46:40 +02:00
|
|
|
shortcuts={y: x for x, y in engine_shortcuts.items()},
|
|
|
|
themes=themes,
|
2015-03-11 18:57:36 +01:00
|
|
|
plugins=plugins,
|
2016-04-08 16:38:05 +02:00
|
|
|
allowed_plugins=allowed_plugins,
|
2016-08-06 06:34:56 +02:00
|
|
|
theme=get_current_theme_name(),
|
|
|
|
preferences=True)
|
2014-01-01 22:16:53 +01:00
|
|
|
|
|
|
|
|
2015-01-16 16:02:21 +01:00
|
|
|
@app.route('/image_proxy', methods=['GET'])
|
|
|
|
def image_proxy():
|
2015-02-01 10:18:32 +01:00
|
|
|
url = request.args.get('url').encode('utf-8')
|
2015-01-16 16:02:21 +01:00
|
|
|
|
|
|
|
if not url:
|
|
|
|
return '', 400
|
|
|
|
|
2016-10-16 23:40:56 +02:00
|
|
|
h = hmac.new(settings['server']['secret_key'], url, hashlib.sha256).hexdigest()
|
2015-01-17 21:54:40 +01:00
|
|
|
|
|
|
|
if h != request.args.get('h'):
|
|
|
|
return '', 400
|
|
|
|
|
|
|
|
headers = dict_subset(request.headers, {'If-Modified-Since', 'If-None-Match'})
|
|
|
|
headers['User-Agent'] = gen_useragent()
|
|
|
|
|
2015-04-10 01:10:49 +02:00
|
|
|
resp = requests.get(url,
|
|
|
|
stream=True,
|
2015-08-02 19:38:27 +02:00
|
|
|
timeout=settings['outgoing']['request_timeout'],
|
2015-04-25 11:44:53 +02:00
|
|
|
headers=headers,
|
|
|
|
proxies=outgoing_proxies)
|
2015-01-17 21:54:40 +01:00
|
|
|
|
|
|
|
if resp.status_code == 304:
|
|
|
|
return '', resp.status_code
|
2015-01-16 16:02:21 +01:00
|
|
|
|
|
|
|
if resp.status_code != 200:
|
|
|
|
logger.debug('image-proxy: wrong response code: {0}'.format(resp.status_code))
|
|
|
|
if resp.status_code >= 400:
|
|
|
|
return '', resp.status_code
|
|
|
|
return '', 400
|
|
|
|
|
|
|
|
if not resp.headers.get('content-type', '').startswith('image/'):
|
2015-06-21 16:50:42 +02:00
|
|
|
logger.debug('image-proxy: wrong content-type: {0}'.format(resp.headers.get('content-type')))
|
2015-01-16 16:02:21 +01:00
|
|
|
return '', 400
|
|
|
|
|
|
|
|
img = ''
|
|
|
|
chunk_counter = 0
|
|
|
|
|
2015-02-10 15:23:56 +01:00
|
|
|
for chunk in resp.iter_content(1024 * 1024):
|
2015-01-16 16:02:21 +01:00
|
|
|
chunk_counter += 1
|
|
|
|
if chunk_counter > 5:
|
|
|
|
return '', 502 # Bad gateway - file is too big (>5M)
|
|
|
|
img += chunk
|
|
|
|
|
2015-01-17 21:54:40 +01:00
|
|
|
headers = dict_subset(resp.headers, {'Content-Length', 'Length', 'Date', 'Last-Modified', 'Expires', 'Etag'})
|
|
|
|
|
|
|
|
return Response(img, mimetype=resp.headers['content-type'], headers=headers)
|
2015-01-16 16:02:21 +01:00
|
|
|
|
|
|
|
|
2013-10-27 01:03:05 +02:00
|
|
|
@app.route('/stats', methods=['GET'])
|
|
|
|
def stats():
|
2014-01-31 07:08:24 +01:00
|
|
|
"""Render engine statistics page."""
|
2013-10-27 01:03:05 +02:00
|
|
|
stats = get_engines_stats()
|
2014-03-21 12:19:48 +01:00
|
|
|
return render(
|
|
|
|
'stats.html',
|
|
|
|
stats=stats,
|
|
|
|
)
|
2013-10-27 01:03:05 +02:00
|
|
|
|
2014-01-01 22:16:53 +01:00
|
|
|
|
2013-12-01 16:10:38 +01:00
|
|
|
@app.route('/robots.txt', methods=['GET'])
|
|
|
|
def robots():
|
|
|
|
return Response("""User-agent: *
|
|
|
|
Allow: /
|
|
|
|
Allow: /about
|
|
|
|
Disallow: /stats
|
2014-02-07 18:43:05 +01:00
|
|
|
Disallow: /preferences
|
2016-10-01 20:22:52 +02:00
|
|
|
Disallow: /*?*q=*
|
2013-12-01 16:10:38 +01:00
|
|
|
""", mimetype='text/plain')
|
|
|
|
|
2014-01-01 22:16:53 +01:00
|
|
|
|
2013-10-16 00:01:08 +02:00
|
|
|
@app.route('/opensearch.xml', methods=['GET'])
|
|
|
|
def opensearch():
|
2013-10-20 22:37:55 +02:00
|
|
|
method = 'post'
|
2015-03-15 20:07:50 +01:00
|
|
|
|
2016-04-08 16:38:05 +02:00
|
|
|
if request.preferences.get_value('method') == 'GET':
|
2015-03-15 20:07:50 +01:00
|
|
|
method = 'get'
|
|
|
|
|
2013-10-21 00:28:48 +02:00
|
|
|
# chrome/chromium only supports HTTP GET....
|
2013-10-20 22:37:55 +02:00
|
|
|
if request.headers.get('User-Agent', '').lower().find('webkit') >= 0:
|
|
|
|
method = 'get'
|
2014-03-29 16:29:19 +01:00
|
|
|
|
|
|
|
ret = render('opensearch.xml',
|
2014-05-01 10:14:47 +02:00
|
|
|
opensearch_method=method,
|
2016-03-02 11:51:11 +01:00
|
|
|
host=get_base_url(),
|
2017-01-13 22:15:11 +01:00
|
|
|
urljoin=urljoin,
|
|
|
|
override_theme='__common__')
|
2014-03-29 16:29:19 +01:00
|
|
|
|
2013-10-16 00:01:08 +02:00
|
|
|
resp = Response(response=ret,
|
2014-01-20 02:31:20 +01:00
|
|
|
status=200,
|
2015-01-18 14:26:52 +01:00
|
|
|
mimetype="text/xml")
|
2013-10-16 00:01:08 +02:00
|
|
|
return resp
|
|
|
|
|
2014-01-19 22:59:01 +01:00
|
|
|
|
2013-12-01 23:52:49 +01:00
|
|
|
@app.route('/favicon.ico')
|
|
|
|
def favicon():
|
2014-04-25 01:46:40 +02:00
|
|
|
return send_from_directory(os.path.join(app.root_path,
|
2015-01-01 19:24:47 +01:00
|
|
|
'static/themes',
|
2014-04-25 01:46:40 +02:00
|
|
|
get_current_theme_name(),
|
|
|
|
'img'),
|
2014-01-20 02:31:20 +01:00
|
|
|
'favicon.png',
|
|
|
|
mimetype='image/vnd.microsoft.icon')
|
2013-12-01 23:52:49 +01:00
|
|
|
|
|
|
|
|
2015-04-07 11:07:48 +02:00
|
|
|
@app.route('/clear_cookies')
|
|
|
|
def clear_cookies():
|
2015-10-23 23:07:36 +02:00
|
|
|
resp = make_response(redirect(urljoin(settings['server']['base_url'], url_for('index'))))
|
2015-04-07 11:07:48 +02:00
|
|
|
for cookie_name in request.cookies:
|
|
|
|
resp.delete_cookie(cookie_name)
|
|
|
|
return resp
|
|
|
|
|
|
|
|
|
2016-06-07 23:08:48 +02:00
|
|
|
@app.route('/config')
|
|
|
|
def config():
|
|
|
|
return jsonify({'categories': categories.keys(),
|
|
|
|
'engines': [{'name': engine_name,
|
|
|
|
'categories': engine.categories,
|
2016-07-07 09:27:34 +02:00
|
|
|
'shortcut': engine.shortcut,
|
2017-01-07 14:42:55 +01:00
|
|
|
'enabled': not engine.disabled,
|
|
|
|
'paging': engine.paging,
|
|
|
|
'language_support': engine.language_support,
|
|
|
|
'supported_languages':
|
|
|
|
engine.supported_languages.keys()
|
|
|
|
if isinstance(engine.supported_languages, dict)
|
|
|
|
else engine.supported_languages,
|
|
|
|
'safesearch': engine.safesearch,
|
|
|
|
'time_range_support': engine.time_range_support,
|
|
|
|
'timeout': engine.timeout}
|
2016-06-07 23:08:48 +02:00
|
|
|
for engine_name, engine in engines.items()],
|
|
|
|
'plugins': [{'name': plugin.name,
|
|
|
|
'enabled': plugin.default_on}
|
|
|
|
for plugin in plugins],
|
|
|
|
'instance_name': settings['general']['instance_name'],
|
|
|
|
'locales': settings['locales'],
|
|
|
|
'default_locale': settings['ui']['default_locale'],
|
|
|
|
'autocomplete': settings['search']['autocomplete'],
|
|
|
|
'safe_search': settings['search']['safe_search'],
|
2016-10-30 15:47:49 +01:00
|
|
|
'default_theme': settings['ui']['default_theme'],
|
|
|
|
'version': VERSION_STRING})
|
2016-06-07 23:08:48 +02:00
|
|
|
|
|
|
|
|
2016-08-24 19:53:09 +02:00
|
|
|
@app.errorhandler(404)
|
|
|
|
def page_not_found(e):
|
2016-09-07 08:32:01 +02:00
|
|
|
return render('404.html'), 404
|
2016-08-24 19:53:09 +02:00
|
|
|
|
|
|
|
|
2014-01-12 12:40:27 +01:00
|
|
|
def run():
|
2016-11-30 18:43:03 +01:00
|
|
|
logger.debug('starting webserver on %s:%s', settings['server']['port'], settings['server']['bind_address'])
|
2014-01-19 22:59:01 +01:00
|
|
|
app.run(
|
2016-10-22 19:07:37 +02:00
|
|
|
debug=searx_debug,
|
|
|
|
use_debugger=searx_debug,
|
2015-08-02 19:03:55 +02:00
|
|
|
port=settings['server']['port'],
|
2016-09-05 22:44:48 +02:00
|
|
|
host=settings['server']['bind_address'],
|
|
|
|
threaded=True
|
2014-01-19 22:59:01 +01:00
|
|
|
)
|
2014-01-12 12:40:27 +01:00
|
|
|
|
|
|
|
|
2015-06-16 19:55:31 +02:00
|
|
|
class ReverseProxyPathFix(object):
|
|
|
|
'''Wrap the application in this middleware and configure the
|
|
|
|
front-end server to add these headers, to let you quietly bind
|
|
|
|
this to a URL other than / and to an HTTP scheme that is
|
|
|
|
different than what is used locally.
|
|
|
|
|
|
|
|
http://flask.pocoo.org/snippets/35/
|
|
|
|
|
|
|
|
In nginx:
|
|
|
|
location /myprefix {
|
|
|
|
proxy_pass http://127.0.0.1:8000;
|
|
|
|
proxy_set_header Host $host;
|
|
|
|
proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for;
|
|
|
|
proxy_set_header X-Scheme $scheme;
|
|
|
|
proxy_set_header X-Script-Name /myprefix;
|
|
|
|
}
|
|
|
|
|
|
|
|
:param app: the WSGI application
|
|
|
|
'''
|
2016-07-10 16:44:27 +02:00
|
|
|
|
2015-06-16 19:55:31 +02:00
|
|
|
def __init__(self, app):
|
|
|
|
self.app = app
|
|
|
|
|
|
|
|
def __call__(self, environ, start_response):
|
|
|
|
script_name = environ.get('HTTP_X_SCRIPT_NAME', '')
|
|
|
|
if script_name:
|
|
|
|
environ['SCRIPT_NAME'] = script_name
|
|
|
|
path_info = environ['PATH_INFO']
|
|
|
|
if path_info.startswith(script_name):
|
|
|
|
environ['PATH_INFO'] = path_info[len(script_name):]
|
|
|
|
|
|
|
|
scheme = environ.get('HTTP_X_SCHEME', '')
|
|
|
|
if scheme:
|
|
|
|
environ['wsgi.url_scheme'] = scheme
|
|
|
|
return self.app(environ, start_response)
|
2014-07-03 22:02:53 +02:00
|
|
|
|
2015-02-07 12:28:36 +01:00
|
|
|
|
2015-06-16 19:55:31 +02:00
|
|
|
application = app
|
|
|
|
# patch app to handle non root url-s behind proxy & wsgi
|
|
|
|
app.wsgi_app = ReverseProxyPathFix(ProxyFix(application.wsgi_app))
|
2014-07-03 22:02:53 +02:00
|
|
|
|
2014-01-12 12:40:27 +01:00
|
|
|
if __name__ == "__main__":
|
|
|
|
run()
|