forked from Ponysearch/Ponysearch
Merge branch 'master' into engines/unsplash
This commit is contained in:
commit
4a127b19de
5 changed files with 27242 additions and 135 deletions
|
@ -32,6 +32,7 @@ RUN echo "@commuedge http://nl.alpinelinux.org/alpine/edge/community" >> /etc/ap
|
|||
openssl-dev \
|
||||
ca-certificates \
|
||||
tini@commuedge \
|
||||
&& pip install --upgrade pip \
|
||||
&& pip install --no-cache -r requirements.txt \
|
||||
&& apk del \
|
||||
build-base \
|
||||
|
|
File diff suppressed because one or more lines are too long
|
@ -1,115 +0,0 @@
|
|||
"""
|
||||
FindX (General, Images, Videos)
|
||||
|
||||
@website https://www.findx.com
|
||||
@provide-api no
|
||||
@using-api no
|
||||
@results HTML
|
||||
@stable no
|
||||
@parse url, title, content, embedded, img_src, thumbnail_src
|
||||
"""
|
||||
|
||||
from dateutil import parser
|
||||
from json import loads
|
||||
import re
|
||||
|
||||
from lxml import html
|
||||
|
||||
from searx import logger
|
||||
from searx.engines.xpath import extract_text
|
||||
from searx.engines.youtube_noapi import base_youtube_url, embedded_url
|
||||
from searx.url_utils import urlencode
|
||||
|
||||
|
||||
paging = True
|
||||
results_xpath = '//script[@id="initial-state"]'
|
||||
search_url = 'https://www.findx.com/{category}?{q}'
|
||||
type_map = {
|
||||
'none': 'web',
|
||||
'general': 'web',
|
||||
'images': 'images',
|
||||
'videos': 'videos',
|
||||
}
|
||||
|
||||
|
||||
def request(query, params):
|
||||
params['url'] = search_url.format(
|
||||
category=type_map[params['category']],
|
||||
q=urlencode({
|
||||
'q': query,
|
||||
'page': params['pageno']
|
||||
})
|
||||
)
|
||||
return params
|
||||
|
||||
|
||||
def response(resp):
|
||||
dom = html.fromstring(resp.text)
|
||||
results_raw_json = dom.xpath(results_xpath)
|
||||
results_json = loads(extract_text(results_raw_json))
|
||||
|
||||
if len(results_json['web']['results']) > 0:
|
||||
return _general_results(results_json['web']['results']['webSearch']['results'])
|
||||
|
||||
if len(results_json['images']['results']) > 0:
|
||||
return _images_results(results_json['images']['results'])
|
||||
|
||||
if len(results_json['video']['results']) > 0:
|
||||
return _videos_results(results_json['video']['results'])
|
||||
|
||||
return []
|
||||
|
||||
|
||||
def _general_results(general_results):
|
||||
results = []
|
||||
for result in general_results:
|
||||
results.append({
|
||||
'url': result['url'],
|
||||
'title': result['title'],
|
||||
'content': result['sum'],
|
||||
})
|
||||
return results
|
||||
|
||||
|
||||
def _images_results(image_results):
|
||||
results = []
|
||||
for result in image_results:
|
||||
results.append({
|
||||
'url': result['sourceURL'],
|
||||
'title': result['title'],
|
||||
'content': result['source'],
|
||||
'thumbnail_src': _extract_url(result['assets']['thumb']['url']),
|
||||
'img_src': _extract_url(result['assets']['file']['url']),
|
||||
'template': 'images.html',
|
||||
})
|
||||
return results
|
||||
|
||||
|
||||
def _videos_results(video_results):
|
||||
results = []
|
||||
for result in video_results:
|
||||
if not result['kind'].startswith('youtube'):
|
||||
logger.warn('Unknown video kind in findx: {}'.format(result['kind']))
|
||||
continue
|
||||
|
||||
description = result['snippet']['description']
|
||||
if len(description) > 300:
|
||||
description = description[:300] + '...'
|
||||
|
||||
results.append({
|
||||
'url': base_youtube_url + result['id'],
|
||||
'title': result['snippet']['title'],
|
||||
'content': description,
|
||||
'thumbnail': _extract_url(result['snippet']['thumbnails']['default']['url']),
|
||||
'publishedDate': parser.parse(result['snippet']['publishedAt']),
|
||||
'embedded': embedded_url.format(videoid=result['id']),
|
||||
'template': 'videos.html',
|
||||
})
|
||||
return results
|
||||
|
||||
|
||||
def _extract_url(url):
|
||||
matching = re.search('(/https?://[^)]+)', url)
|
||||
if matching:
|
||||
return matching.group(0)[1:]
|
||||
return ''
|
|
@ -218,24 +218,6 @@ engines:
|
|||
shortcut : fd
|
||||
disabled : True
|
||||
|
||||
- name : findx
|
||||
engine : findx
|
||||
shortcut : fx
|
||||
categories : general
|
||||
disabled : True
|
||||
|
||||
- name : findx images
|
||||
engine : findx
|
||||
shortcut : fxi
|
||||
categories : images
|
||||
disabled : True
|
||||
|
||||
- name : findx videos
|
||||
engine : findx
|
||||
shortcut : fxv
|
||||
categories : videos
|
||||
disabled : True
|
||||
|
||||
- name : flickr
|
||||
categories : images
|
||||
shortcut : fl
|
||||
|
@ -728,6 +710,19 @@ engines:
|
|||
shortcut : du
|
||||
disabled : True
|
||||
|
||||
- name : seznam
|
||||
shortcut: szn
|
||||
engine: xpath
|
||||
paging : True
|
||||
search_url : https://search.seznam.cz/?q={query}&count=10&from={pageno}
|
||||
results_xpath: //div[@class="Page-content"]//div[@class="Result "]
|
||||
url_xpath : ./h3/a/@href
|
||||
title_xpath : ./h3
|
||||
content_xpath : .//p[@class="Result-description"]
|
||||
first_page_num : 0
|
||||
page_size : 10
|
||||
disabled : True
|
||||
|
||||
# - name : yacy
|
||||
# engine : yacy
|
||||
# shortcut : ya
|
||||
|
|
|
@ -27,12 +27,14 @@ def fetch_supported_languages():
|
|||
if hasattr(engines[engine_name], 'fetch_supported_languages'):
|
||||
try:
|
||||
engines_languages[engine_name] = engines[engine_name].fetch_supported_languages()
|
||||
if type(engines_languages[engine_name]) == list:
|
||||
engines_languages[engine_name] = sorted(engines_languages[engine_name])
|
||||
except Exception as e:
|
||||
print(e)
|
||||
|
||||
# write json file
|
||||
with io.open(engines_languages_file, "w", encoding="utf-8") as f:
|
||||
dump(engines_languages, f, ensure_ascii=False)
|
||||
dump(engines_languages, f, ensure_ascii=False, indent=4, separators=(',', ': '))
|
||||
|
||||
return engines_languages
|
||||
|
||||
|
|
Loading…
Reference in a new issue