forked from Ponysearch/Ponysearch
[mod] change settings file structure according to #314
This commit is contained in:
parent
d0830d4edf
commit
1fcf066a81
7 changed files with 46 additions and 34 deletions
|
@ -29,7 +29,7 @@ from searx.poolrequests import get as http_get
|
|||
|
||||
def get(*args, **kwargs):
|
||||
if 'timeout' not in kwargs:
|
||||
kwargs['timeout'] = settings['server']['request_timeout']
|
||||
kwargs['timeout'] = settings['outgoing']['request_timeout']
|
||||
|
||||
return http_get(*args, **kwargs)
|
||||
|
||||
|
|
|
@ -75,7 +75,7 @@ def load_engine(engine_data):
|
|||
engine.safesearch = False
|
||||
|
||||
if not hasattr(engine, 'timeout'):
|
||||
engine.timeout = settings['server']['request_timeout']
|
||||
engine.timeout = settings['outgoing']['request_timeout']
|
||||
|
||||
if not hasattr(engine, 'shortcut'):
|
||||
engine.shortcut = ''
|
||||
|
|
|
@ -39,11 +39,11 @@ class HTTPAdapterWithConnParams(requests.adapters.HTTPAdapter):
|
|||
block=self._pool_block, **self._conn_params)
|
||||
|
||||
|
||||
if settings.get('source_ips'):
|
||||
if settings['outgoing'].get('source_ips'):
|
||||
http_adapters = cycle(HTTPAdapterWithConnParams(pool_connections=100, source_address=(source_ip, 0))
|
||||
for source_ip in settings['source_ips'])
|
||||
for source_ip in settings['outgoing']['source_ips'])
|
||||
https_adapters = cycle(HTTPAdapterWithConnParams(pool_connections=100, source_address=(source_ip, 0))
|
||||
for source_ip in settings['source_ips'])
|
||||
for source_ip in settings['outgoing']['source_ips'])
|
||||
else:
|
||||
http_adapters = cycle((HTTPAdapterWithConnParams(pool_connections=100), ))
|
||||
https_adapters = cycle((HTTPAdapterWithConnParams(pool_connections=100), ))
|
||||
|
@ -69,7 +69,7 @@ def request(method, url, **kwargs):
|
|||
"""same as requests/requests/api.py request(...) except it use SessionSinglePool and force proxies"""
|
||||
global settings
|
||||
session = SessionSinglePool()
|
||||
kwargs['proxies'] = settings.get('outgoing_proxies', None)
|
||||
kwargs['proxies'] = settings['outgoing'].get('proxies', None)
|
||||
response = session.request(method=method, url=url, **kwargs)
|
||||
session.close()
|
||||
return response
|
||||
|
|
|
@ -1,26 +1,30 @@
|
|||
general:
|
||||
debug : False # Debug mode, only for development
|
||||
|
||||
server:
|
||||
port : 8888
|
||||
bind_address : "127.0.0.1" # address to listen on
|
||||
secret_key : "ultrasecretkey" # change this!
|
||||
debug : False # Debug mode, only for development
|
||||
request_timeout : 2.0 # seconds
|
||||
base_url : False # Set custom base_url. Possible values: False or "https://your.custom.host/location/"
|
||||
image_proxy : False # Proxying image results through searx
|
||||
|
||||
ui:
|
||||
themes_path : "" # Custom ui themes path - leave it blank if you didn't change
|
||||
default_theme : oscar # ui theme
|
||||
useragent_suffix : "" # suffix of searx_useragent, could contain informations like an email address to the administrator
|
||||
image_proxy : False # Proxying image results through searx
|
||||
default_locale : "" # Default interface locale - leave blank to detect from browser information or use codes from the 'locales' config section
|
||||
|
||||
outgoing: # communication with search engines
|
||||
request_timeout : 2.0 # seconds
|
||||
useragent_suffix : "" # suffix of searx_useragent, could contain informations like an email address to the administrator
|
||||
# uncomment below section if you want to use a proxy
|
||||
# see http://docs.python-requests.org/en/latest/user/advanced/#proxies
|
||||
# SOCKS proxies are not supported : see https://github.com/kennethreitz/requests/pull/478
|
||||
#outgoing_proxies :
|
||||
# proxies :
|
||||
# http : http://127.0.0.1:8080
|
||||
# https: http://127.0.0.1:8080
|
||||
|
||||
# uncomment below section only if you have more than one network interface
|
||||
# which can be the source of outgoing search requests
|
||||
#source_ips:
|
||||
# source_ips:
|
||||
# - 1.1.1.1
|
||||
# - 1.1.1.2
|
||||
|
||||
|
|
|
@ -1,14 +1,21 @@
|
|||
general:
|
||||
debug : False
|
||||
|
||||
server:
|
||||
port : 11111
|
||||
bind_address : 127.0.0.1
|
||||
secret_key : "ultrasecretkey" # change this!
|
||||
debug : False
|
||||
request_timeout : 3.0 # seconds
|
||||
base_url: False
|
||||
base_url : False
|
||||
image_proxy : False
|
||||
|
||||
ui:
|
||||
themes_path : ""
|
||||
default_theme : default
|
||||
https_rewrite : True
|
||||
image_proxy : False
|
||||
default_locale : ""
|
||||
|
||||
outgoing:
|
||||
request_timeout : 1.0 # seconds
|
||||
useragent_suffix : ""
|
||||
|
||||
engines:
|
||||
- name : general_dummy
|
||||
|
|
|
@ -26,6 +26,7 @@ ua_versions = ('33.0',
|
|||
ua_os = ('Windows NT 6.3; WOW64',
|
||||
'X11; Linux x86_64',
|
||||
'X11; Linux x86')
|
||||
|
||||
ua = "Mozilla/5.0 ({os}; rv:{version}) Gecko/20100101 Firefox/{version}"
|
||||
|
||||
blocked_tags = ('script',
|
||||
|
@ -40,7 +41,7 @@ def gen_useragent():
|
|||
def searx_useragent():
|
||||
return 'searx/{searx_version} {suffix}'.format(
|
||||
searx_version=VERSION_STRING,
|
||||
suffix=settings['server'].get('useragent_suffix', ''))
|
||||
suffix=settings['outgoing'].get('useragent_suffix', ''))
|
||||
|
||||
|
||||
def highlight_content(content, query):
|
||||
|
|
|
@ -77,11 +77,11 @@ except ImportError:
|
|||
|
||||
|
||||
static_path, templates_path, themes =\
|
||||
get_themes(settings['themes_path']
|
||||
if settings.get('themes_path')
|
||||
get_themes(settings['ui']['themes_path']
|
||||
if settings['ui']['themes_path']
|
||||
else searx_dir)
|
||||
|
||||
default_theme = settings['server'].get('default_theme', 'default')
|
||||
default_theme = settings['ui']['default_theme']
|
||||
|
||||
static_files = get_static_files(searx_dir)
|
||||
|
||||
|
@ -121,15 +121,15 @@ _category_names = (gettext('files'),
|
|||
gettext('news'),
|
||||
gettext('map'))
|
||||
|
||||
outgoing_proxies = settings.get('outgoing_proxies', None)
|
||||
outgoing_proxies = settings['outgoing'].get('proxies', None)
|
||||
|
||||
|
||||
@babel.localeselector
|
||||
def get_locale():
|
||||
locale = request.accept_languages.best_match(settings['locales'].keys())
|
||||
|
||||
if settings['server'].get('default_locale'):
|
||||
locale = settings['server']['default_locale']
|
||||
if settings['ui'].get('default_locale'):
|
||||
locale = settings['ui']['default_locale']
|
||||
|
||||
if request.cookies.get('locale', '') in settings['locales']:
|
||||
locale = request.cookies.get('locale', '')
|
||||
|
@ -640,12 +640,12 @@ def preferences():
|
|||
stats[e.name] = {'time': None,
|
||||
'warn_timeout': False,
|
||||
'warn_time': False}
|
||||
if e.timeout > settings['server']['request_timeout']:
|
||||
if e.timeout > settings['outgoing']['request_timeout']:
|
||||
stats[e.name]['warn_timeout'] = True
|
||||
|
||||
for engine_stat in get_engines_stats()[0][1]:
|
||||
stats[engine_stat.get('name')]['time'] = round(engine_stat.get('avg'), 3)
|
||||
if engine_stat.get('avg') > settings['server']['request_timeout']:
|
||||
if engine_stat.get('avg') > settings['outgoing']['request_timeout']:
|
||||
stats[engine_stat.get('name')]['warn_time'] = True
|
||||
# end of stats
|
||||
|
||||
|
@ -683,7 +683,7 @@ def image_proxy():
|
|||
|
||||
resp = requests.get(url,
|
||||
stream=True,
|
||||
timeout=settings['server'].get('request_timeout', 2),
|
||||
timeout=settings['outgoing']['request_timeout'],
|
||||
headers=headers,
|
||||
proxies=outgoing_proxies)
|
||||
|
||||
|
@ -775,8 +775,8 @@ def clear_cookies():
|
|||
|
||||
def run():
|
||||
app.run(
|
||||
debug=settings['server']['debug'],
|
||||
use_debugger=settings['server']['debug'],
|
||||
debug=settings['general']['debug'],
|
||||
use_debugger=settings['general']['debug'],
|
||||
port=settings['server']['port'],
|
||||
host=settings['server']['bind_address']
|
||||
)
|
||||
|
|
Loading…
Reference in a new issue