mirror of
https://github.com/searxng/searxng.git
synced 2024-11-05 21:00:13 +01:00
2499899554
Partial reverse engineering of the Google engines including a improved language and region handling based on the engine.traits_v1 data. When ever possible the implementations of the Google engines try to make use of the async REST APIs. The get_lang_info() has been generalized to a get_google_info() function / especially the region handling has been improved by adding the cr parameter. searx/data/engine_traits.json Add data type "traits_v1" generated by the fetch_traits() functions from: - Google (WEB), - Google images, - Google news, - Google scholar and - Google videos and remove data from obsolete data type "supported_languages". A traits.custom type that maps region codes to *supported_domains* is fetched from https://www.google.com/supported_domains searx/autocomplete.py: Reversed engineered autocomplete from Google WEB. Supports Google's languages and subdomains. The old API suggestqueries.google.com/complete has been replaced by the async REST API: https://{subdomain}/complete/search?{args} searx/engines/google.py Reverse engineering and extensive testing .. - fetch_traits(): Fetch languages & regions from Google properties. - always use the async REST API (formally known as 'use_mobile_ui') - use *supported_domains* from traits - improved the result list by fetching './/div[@data-content-feature]' and parsing the type of the various *content features* --> thumbnails are added searx/engines/google_images.py Reverse engineering and extensive testing .. - fetch_traits(): Fetch languages & regions from Google properties. - use *supported_domains* from traits - if exists, freshness_date is added to the result - issue 1864: result list has been improved a lot (due to the new cr parameter) searx/engines/google_news.py Reverse engineering and extensive testing .. - fetch_traits(): Fetch languages & regions from Google properties. *supported_domains* is not needed but a ceid list has been added. - different region handling compared to Google WEB - fixed for various languages & regions (due to the new ceid parameter) / avoid CONSENT page - Google News do no longer support time range - result list has been fixed: XPath of pub_date and pub_origin searx/engines/google_videos.py - fetch_traits(): Fetch languages & regions from Google properties. - use *supported_domains* from traits - add paging support - implement a async request ('asearch': 'arc' & 'async': 'use_ac:true,_fmt:html') - simplified code (thanks to '_fmt:html' request) - issue 1359: fixed xpath of video length data searx/engines/google_scholar.py - fetch_traits(): Fetch languages & regions from Google properties. - use *supported_domains* from traits - request(): include patents & citations - response(): fixed CAPTCHA detection (Scholar has its own CATCHA manager) - hardening XPath to iterate over results - fixed XPath of pub_type (has been change from gs_ct1 to gs_cgt2 class) - issue 1769 fixed: new request implementation is no longer incompatible Signed-off-by: Markus Heiser <markus.heiser@darmarit.de>
130 lines
3.6 KiB
Python
130 lines
3.6 KiB
Python
# SPDX-License-Identifier: AGPL-3.0-or-later
|
|
# lint: pylint
|
|
"""This is the implementation of the Google Images engine using the internal
|
|
Google API used by the Google Go Android app.
|
|
|
|
This internal API offer results in
|
|
|
|
- JSON (``_fmt:json``)
|
|
- Protobuf_ (``_fmt:pb``)
|
|
- Protobuf_ compressed? (``_fmt:pc``)
|
|
- HTML (``_fmt:html``)
|
|
- Protobuf_ encoded in JSON (``_fmt:jspb``).
|
|
|
|
.. _Protobuf: https://en.wikipedia.org/wiki/Protocol_Buffers
|
|
"""
|
|
|
|
from typing import TYPE_CHECKING
|
|
|
|
from urllib.parse import urlencode
|
|
from json import loads
|
|
|
|
from searx.engines.google import fetch_traits # pylint: disable=unused-import
|
|
from searx.engines.google import (
|
|
get_google_info,
|
|
time_range_dict,
|
|
detect_google_sorry,
|
|
)
|
|
|
|
if TYPE_CHECKING:
|
|
import logging
|
|
from searx.enginelib.traits import EngineTraits
|
|
|
|
logger: logging.Logger
|
|
traits: EngineTraits
|
|
|
|
|
|
# about
|
|
about = {
|
|
"website": 'https://images.google.com',
|
|
"wikidata_id": 'Q521550',
|
|
"official_api_documentation": 'https://developers.google.com/custom-search',
|
|
"use_official_api": False,
|
|
"require_api_key": False,
|
|
"results": 'JSON',
|
|
}
|
|
|
|
# engine dependent config
|
|
categories = ['images', 'web']
|
|
paging = True
|
|
time_range_support = True
|
|
safesearch = True
|
|
send_accept_language_header = True
|
|
|
|
filter_mapping = {0: 'images', 1: 'active', 2: 'active'}
|
|
|
|
|
|
def request(query, params):
|
|
"""Google-Image search request"""
|
|
|
|
google_info = get_google_info(params, traits)
|
|
|
|
query_url = (
|
|
'https://'
|
|
+ google_info['subdomain']
|
|
+ '/search'
|
|
+ "?"
|
|
+ urlencode(
|
|
{
|
|
'q': query,
|
|
'tbm': "isch",
|
|
**google_info['params'],
|
|
'asearch': 'isch',
|
|
'async': '_fmt:json,p:1,ijn:' + str(params['pageno']),
|
|
}
|
|
)
|
|
)
|
|
|
|
if params['time_range'] in time_range_dict:
|
|
query_url += '&' + urlencode({'tbs': 'qdr:' + time_range_dict[params['time_range']]})
|
|
if params['safesearch']:
|
|
query_url += '&' + urlencode({'safe': filter_mapping[params['safesearch']]})
|
|
params['url'] = query_url
|
|
|
|
params['cookies'] = google_info['cookies']
|
|
params['headers'].update(google_info['headers'])
|
|
return params
|
|
|
|
|
|
def response(resp):
|
|
"""Get response from google's search request"""
|
|
results = []
|
|
|
|
detect_google_sorry(resp)
|
|
|
|
json_start = resp.text.find('{"ischj":')
|
|
json_data = loads(resp.text[json_start:])
|
|
|
|
for item in json_data["ischj"]["metadata"]:
|
|
|
|
result_item = {
|
|
'url': item["result"]["referrer_url"],
|
|
'title': item["result"]["page_title"],
|
|
'content': item["text_in_grid"]["snippet"],
|
|
'source': item["result"]["site_title"],
|
|
'img_format': f'{item["original_image"]["width"]} x {item["original_image"]["height"]}',
|
|
'img_src': item["original_image"]["url"],
|
|
'thumbnail_src': item["thumbnail"]["url"],
|
|
'template': 'images.html',
|
|
}
|
|
|
|
author = item["result"].get('iptc', {}).get('creator')
|
|
if author:
|
|
result_item['author'] = ', '.join(author)
|
|
|
|
copyright_notice = item["result"].get('iptc', {}).get('copyright_notice')
|
|
if copyright_notice:
|
|
result_item['source'] += ' | ' + copyright_notice
|
|
|
|
freshness_date = item["result"].get("freshness_date")
|
|
if freshness_date:
|
|
result_item['source'] += ' | ' + freshness_date
|
|
|
|
file_size = item.get('gsa', {}).get('file_size')
|
|
if file_size:
|
|
result_item['source'] += ' (%s)' % file_size
|
|
|
|
results.append(result_item)
|
|
|
|
return results
|