forked from Ponysearch/Ponysearch
Merge pull request #8 from return42/manage-script
Replace Makefile boilerplate by shell scripts
This commit is contained in:
commit
a7b9eca98a
21 changed files with 1066 additions and 1205 deletions
3
.github/workflows/data-update.yml
vendored
3
.github/workflows/data-update.yml
vendored
|
@ -40,8 +40,7 @@ jobs:
|
||||||
env:
|
env:
|
||||||
FETCH_SCRIPT: ./searx_extra/update/${{ matrix.fetch }}
|
FETCH_SCRIPT: ./searx_extra/update/${{ matrix.fetch }}
|
||||||
run: |
|
run: |
|
||||||
source local/py3/bin/activate
|
V=1 manage pyenv.cmd python "$FETCH_SCRIPT"
|
||||||
$FETCH_SCRIPT
|
|
||||||
|
|
||||||
- name: Create Pull Request
|
- name: Create Pull Request
|
||||||
id: cpr
|
id: cpr
|
||||||
|
|
7
.github/workflows/integration.yml
vendored
7
.github/workflows/integration.yml
vendored
|
@ -59,7 +59,7 @@ jobs:
|
||||||
- name: Install node dependencies
|
- name: Install node dependencies
|
||||||
run: make V=1 node.env
|
run: make V=1 node.env
|
||||||
- name: Build themes
|
- name: Build themes
|
||||||
run: make V=1 themes
|
run: make V=1 themes.all
|
||||||
|
|
||||||
documentation:
|
documentation:
|
||||||
name: Documentation
|
name: Documentation
|
||||||
|
@ -77,14 +77,15 @@ jobs:
|
||||||
python-version: '3.9'
|
python-version: '3.9'
|
||||||
architecture: 'x64'
|
architecture: 'x64'
|
||||||
- name: Build documentation
|
- name: Build documentation
|
||||||
run: SEARX_DEBUG=1 make V=1 ci-gh-pages
|
run: |
|
||||||
|
make V=1 docs.clean docs.html
|
||||||
- name: Deploy
|
- name: Deploy
|
||||||
if: github.ref == 'refs/heads/master'
|
if: github.ref == 'refs/heads/master'
|
||||||
uses: JamesIves/github-pages-deploy-action@3.7.1
|
uses: JamesIves/github-pages-deploy-action@3.7.1
|
||||||
with:
|
with:
|
||||||
GITHUB_TOKEN: ${{ github.token }}
|
GITHUB_TOKEN: ${{ github.token }}
|
||||||
BRANCH: gh-pages
|
BRANCH: gh-pages
|
||||||
FOLDER: build/gh-pages
|
FOLDER: dist/docs
|
||||||
CLEAN: true # Automatically remove deleted files from the deploy branch
|
CLEAN: true # Automatically remove deleted files from the deploy branch
|
||||||
SINGLE_COMMIT: True
|
SINGLE_COMMIT: True
|
||||||
COMMIT_MESSAGE: build from commit ${{ github.sha }}
|
COMMIT_MESSAGE: build from commit ${{ github.sha }}
|
||||||
|
|
311
Makefile
311
Makefile
|
@ -1,268 +1,107 @@
|
||||||
# -*- coding: utf-8; mode: makefile-gmake -*-
|
# -*- coding: utf-8; mode: makefile-gmake -*-
|
||||||
|
# SPDX-License-Identifier: AGPL-3.0-or-later
|
||||||
|
|
||||||
.DEFAULT_GOAL=help
|
.DEFAULT_GOAL=help
|
||||||
|
export MTOOLS=./manage
|
||||||
|
|
||||||
include utils/makefile.include
|
include utils/makefile.include
|
||||||
|
|
||||||
PYOBJECTS = searx
|
|
||||||
DOC = docs
|
|
||||||
PY_SETUP_EXTRAS ?= \[test\]
|
|
||||||
PYLINT_SEARX_DISABLE_OPTION := I,C,R,W0105,W0212,W0511,W0603,W0613,W0621,W0702,W0703,W1401,E1136
|
|
||||||
PYLINT_ADDITIONAL_BUILTINS_FOR_ENGINES := supported_languages,language_aliases
|
|
||||||
|
|
||||||
include utils/makefile.python
|
|
||||||
include utils/makefile.sphinx
|
|
||||||
|
|
||||||
all: clean install
|
all: clean install
|
||||||
|
|
||||||
PHONY += help-min help-all help
|
PHONY += help
|
||||||
|
|
||||||
help: help-min
|
help:
|
||||||
@echo ''
|
@./manage --help
|
||||||
@echo 'to get more help: make help-all'
|
@echo '----'
|
||||||
|
@echo 'run - run developer instance'
|
||||||
|
@echo 'install - developer install of searx into virtualenv'
|
||||||
|
@echo 'uninstall - uninstall developer installation'
|
||||||
|
@echo 'clean - clean up working tree'
|
||||||
|
@echo 'search.checker - check search engines'
|
||||||
|
@echo 'test - run shell & CI tests'
|
||||||
|
@echo 'test.shell - test shell scripts'
|
||||||
|
@echo 'ci.test - run CI tests'
|
||||||
|
|
||||||
help-min:
|
|
||||||
@echo ' test - run developer tests'
|
|
||||||
@echo ' docs - build documentation'
|
|
||||||
@echo ' docs-live - autobuild HTML documentation while editing'
|
|
||||||
@echo ' run - run developer instance'
|
|
||||||
@echo ' install - developer install (./local)'
|
|
||||||
@echo ' uninstall - uninstall (./local)'
|
|
||||||
@echo ' gh-pages - build docs & deploy on gh-pages branch'
|
|
||||||
@echo ' clean - drop builds and environments'
|
|
||||||
@echo ' project - re-build generic files of the searx project'
|
|
||||||
@echo ' buildenv - re-build environment files (aka brand)'
|
|
||||||
@echo ' themes - re-build build the source of the themes'
|
|
||||||
@echo ' docker - build Docker image'
|
|
||||||
@echo ' node.env - download & install npm dependencies locally'
|
|
||||||
@echo ''
|
|
||||||
@$(MAKE) -e -s make-help
|
|
||||||
|
|
||||||
help-all: help-min
|
|
||||||
@echo ''
|
|
||||||
@$(MAKE) -e -s python-help
|
|
||||||
@echo ''
|
|
||||||
@$(MAKE) -e -s docs-help
|
|
||||||
|
|
||||||
PHONY += install
|
|
||||||
install: buildenv pyenvinstall
|
|
||||||
|
|
||||||
PHONY += uninstall
|
|
||||||
uninstall: pyenvuninstall
|
|
||||||
|
|
||||||
PHONY += clean
|
|
||||||
clean: pyclean docs-clean node.clean test.clean
|
|
||||||
$(call cmd,common_clean)
|
|
||||||
|
|
||||||
PHONY += run
|
PHONY += run
|
||||||
run: buildenv pyenvinstall
|
run: install
|
||||||
$(Q) ( \
|
$(Q) ( \
|
||||||
sleep 2 ; \
|
sleep 2 ; \
|
||||||
xdg-open http://127.0.0.1:8888/ ; \
|
xdg-open http://127.0.0.1:8888/ ; \
|
||||||
) &
|
) &
|
||||||
SEARX_DEBUG=1 $(PY_ENV)/bin/python ./searx/webapp.py
|
SEARX_DEBUG=1 ./manage pyenv.cmd python ./searx/webapp.py
|
||||||
|
|
||||||
# docs
|
PHONY += install uninstall
|
||||||
# ----
|
install uninstall:
|
||||||
|
$(Q)./manage pyenv.$@
|
||||||
|
|
||||||
sphinx-doc-prebuilds:: buildenv pyenvinstall prebuild-includes
|
PHONY += clean
|
||||||
|
clean: py.clean docs.clean node.clean test.clean
|
||||||
|
$(Q)./manage build_msg CLEAN "common files"
|
||||||
|
$(Q)find . -name '*.orig' -exec rm -f {} +
|
||||||
|
$(Q)find . -name '*.rej' -exec rm -f {} +
|
||||||
|
$(Q)find . -name '*~' -exec rm -f {} +
|
||||||
|
$(Q)find . -name '*.bak' -exec rm -f {} +
|
||||||
|
|
||||||
PHONY += docs
|
PHONY += search.checker search.checker.%
|
||||||
docs: sphinx-doc-prebuilds
|
search.checker: install
|
||||||
$(call cmd,sphinx,html,docs,docs)
|
$(Q)./manage pyenv.cmd searx-checker -v
|
||||||
|
|
||||||
PHONY += docs-live
|
search.checker.%: install
|
||||||
docs-live: sphinx-doc-prebuilds
|
$(Q)./manage pyenv.cmd searx-checker -v "$(subst _, ,$(patsubst search.checker.%,%,$@))"
|
||||||
$(call cmd,sphinx_autobuild,html,docs,docs)
|
|
||||||
|
|
||||||
PHONY += prebuild-includes
|
PHONY += test ci.test test.shell
|
||||||
prebuild-includes:
|
ci.test: test.pep8 test.pylint test.unit test.robot
|
||||||
$(Q)mkdir -p $(DOCS_BUILD)/includes
|
test: test.pep8 test.pylint test.unit test.robot test.shell
|
||||||
$(Q)./utils/searx.sh doc | cat > $(DOCS_BUILD)/includes/searx.rst
|
test.shell:
|
||||||
$(Q)./utils/filtron.sh doc | cat > $(DOCS_BUILD)/includes/filtron.rst
|
$(Q)shellcheck -x -s bash \
|
||||||
$(Q)./utils/morty.sh doc | cat > $(DOCS_BUILD)/includes/morty.rst
|
utils/brand.env \
|
||||||
|
./manage \
|
||||||
|
utils/lib.sh \
|
||||||
|
utils/filtron.sh \
|
||||||
|
utils/searx.sh \
|
||||||
|
utils/morty.sh \
|
||||||
|
utils/lxc.sh \
|
||||||
|
utils/lxc-searx.env \
|
||||||
|
.config.sh
|
||||||
|
$(Q)./manage build_msg TEST "$@ OK"
|
||||||
|
|
||||||
|
|
||||||
$(GH_PAGES)::
|
# wrap ./manage script
|
||||||
@echo "doc available at --> $(DOCS_URL)"
|
|
||||||
|
|
||||||
# update project files
|
MANAGE += buildenv
|
||||||
# --------------------
|
MANAGE += babel.compile
|
||||||
|
MANAGE += data.all data.languages data.useragents
|
||||||
|
MANAGE += docs.html docs.live docs.gh-pages docs.prebuild docs.clean
|
||||||
|
MANAGE += docker.build docker.push
|
||||||
|
MANAGE += gecko.driver
|
||||||
|
MANAGE += node.env node.clean
|
||||||
|
MANAGE += py.build py.clean
|
||||||
|
MANAGE += pyenv pyenv.install pyenv.uninstall
|
||||||
|
MANAGE += pypi.upload pypi.upload.test
|
||||||
|
MANAGE += test.pylint test.pep8 test.unit test.coverage test.robot test.clean
|
||||||
|
MANAGE += themes.all themes.oscar themes.simple themes.bootstrap
|
||||||
|
|
||||||
PHONY += project engines.languages useragents.update buildenv
|
PHONY += $(MANAGE)
|
||||||
|
|
||||||
project: buildenv useragents.update engines.languages
|
$(MANAGE):
|
||||||
|
$(Q)$(MTOOLS) $@
|
||||||
|
|
||||||
engines.languages: pyenvinstall
|
# deprecated
|
||||||
$(Q)echo "fetch languages .."
|
|
||||||
$(Q)$(PY_ENV_ACT); python ./searx_extra/update/update_languages.py
|
|
||||||
$(Q)echo "updated searx/data/engines_languages.json"
|
|
||||||
$(Q)echo "updated searx/languages.py"
|
|
||||||
|
|
||||||
useragents.update: pyenvinstall
|
PHONY += docs docs-clean docs-live docker themes
|
||||||
$(Q)echo "fetch useragents .."
|
|
||||||
$(Q)$(PY_ENV_ACT); python ./searx_extra/update/update_firefox_version.py
|
|
||||||
$(Q)echo "updated searx/data/useragents.json with the most recent versions of Firefox."
|
|
||||||
|
|
||||||
buildenv: pyenv
|
docs: docs.html
|
||||||
$(Q)$(PY_ENV_ACT); SEARX_DEBUG=1 python utils/build_env.py
|
$(Q)./manage build_msg WARN $@ is deprecated use docs.html
|
||||||
|
|
||||||
# node / npm
|
docs-clean: docs.clean
|
||||||
# ----------
|
$(Q)./manage build_msg WARN $@ is deprecated use docs.clean
|
||||||
|
|
||||||
node.env: buildenv
|
docs-live: docs.live
|
||||||
$(Q)./manage.sh npm_packages
|
$(Q)./manage build_msg WARN $@ is deprecated use docs.live
|
||||||
|
|
||||||
node.clean:
|
docker: docker.build
|
||||||
$(Q)echo "CLEAN locally installed npm dependencies"
|
$(Q)./manage build_msg WARN $@ is deprecated use docker.build
|
||||||
$(Q)rm -rf \
|
|
||||||
./node_modules \
|
|
||||||
./package-lock.json \
|
|
||||||
./searx/static/themes/oscar/package-lock.json \
|
|
||||||
./searx/static/themes/oscar/node_modules \
|
|
||||||
./searx/static/themes/simple/package-lock.json \
|
|
||||||
./searx/static/themes/simple/node_modules
|
|
||||||
|
|
||||||
# build themes
|
themes: themes.all
|
||||||
# ------------
|
$(Q)./manage build_msg WARN $@ is deprecated use themes.all
|
||||||
|
|
||||||
PHONY += themes themes.oscar themes.simple
|
|
||||||
themes: buildenv themes.oscar themes.simple
|
|
||||||
|
|
||||||
quiet_cmd_lessc = LESSC $3
|
|
||||||
cmd_lessc = PATH="$$(npm bin):$$PATH" \
|
|
||||||
lessc --clean-css="--s1 --advanced --compatibility=ie9" "searx/static/$2" "searx/static/$3"
|
|
||||||
|
|
||||||
quiet_cmd_grunt = GRUNT $2
|
|
||||||
cmd_grunt = PATH="$$(npm bin):$$PATH" \
|
|
||||||
grunt --gruntfile "$2"
|
|
||||||
|
|
||||||
themes.oscar: node.env
|
|
||||||
$(Q)echo '[!] build oscar theme'
|
|
||||||
$(call cmd,grunt,searx/static/themes/oscar/gruntfile.js)
|
|
||||||
|
|
||||||
themes.simple: node.env
|
|
||||||
$(Q)echo '[!] build simple theme'
|
|
||||||
$(call cmd,grunt,searx/static/themes/simple/gruntfile.js)
|
|
||||||
|
|
||||||
|
|
||||||
# docker
|
|
||||||
# ------
|
|
||||||
|
|
||||||
PHONY += docker
|
|
||||||
docker: buildenv
|
|
||||||
$(Q)./manage.sh docker_build
|
|
||||||
|
|
||||||
docker.push: buildenv
|
|
||||||
$(Q)./manage.sh docker_build push
|
|
||||||
|
|
||||||
# gecko
|
|
||||||
# -----
|
|
||||||
|
|
||||||
PHONY += gecko.driver
|
|
||||||
gecko.driver:
|
|
||||||
$(PY_ENV_ACT); ./manage.sh install_geckodriver
|
|
||||||
|
|
||||||
# search.checker
|
|
||||||
# --------------
|
|
||||||
|
|
||||||
search.checker: pyenvinstall
|
|
||||||
$(Q)$(PY_ENV_ACT); searx-checker -v
|
|
||||||
|
|
||||||
ENGINE_TARGETS=$(patsubst searx/engines/%.py,search.checker.%,$(wildcard searx/engines/[!_]*.py))
|
|
||||||
|
|
||||||
$(ENGINE_TARGETS): pyenvinstall
|
|
||||||
$(Q)$(PY_ENV_ACT); searx-checker -v "$(subst _, ,$(patsubst search.checker.%,%,$@))"
|
|
||||||
|
|
||||||
|
|
||||||
# test
|
|
||||||
# ----
|
|
||||||
|
|
||||||
PHONY += test test.sh test.pylint test.pep8 test.unit test.coverage test.robot
|
|
||||||
test: buildenv test.pylint test.pep8 test.unit gecko.driver test.robot
|
|
||||||
|
|
||||||
PYLINT_FILES=\
|
|
||||||
searx/preferences.py \
|
|
||||||
searx/testing.py \
|
|
||||||
searx/engines/gigablast.py \
|
|
||||||
searx/engines/deviantart.py \
|
|
||||||
searx/engines/digg.py \
|
|
||||||
searx/engines/google.py \
|
|
||||||
searx/engines/google_news.py \
|
|
||||||
searx/engines/google_videos.py \
|
|
||||||
searx/engines/google_images.py \
|
|
||||||
searx/engines/mediathekviewweb.py \
|
|
||||||
searx/engines/meilisearch.py \
|
|
||||||
searx/engines/solidtorrents.py \
|
|
||||||
searx/engines/solr.py \
|
|
||||||
searx/engines/springer.py \
|
|
||||||
searx/engines/google_scholar.py \
|
|
||||||
searx/engines/yahoo_news.py \
|
|
||||||
searx/engines/apkmirror.py \
|
|
||||||
searx/engines/artic.py \
|
|
||||||
searx_extra/update/update_external_bangs.py
|
|
||||||
|
|
||||||
test.pylint: pyenvinstall
|
|
||||||
$(call cmd,pylint,$(PYLINT_FILES))
|
|
||||||
$(call cmd,pylint,\
|
|
||||||
--disable=$(PYLINT_SEARX_DISABLE_OPTION) \
|
|
||||||
--additional-builtins=$(PYLINT_ADDITIONAL_BUILTINS_FOR_ENGINES) \
|
|
||||||
searx/engines \
|
|
||||||
)
|
|
||||||
$(call cmd,pylint,\
|
|
||||||
--disable=$(PYLINT_SEARX_DISABLE_OPTION) \
|
|
||||||
--ignore=searx/engines \
|
|
||||||
searx tests \
|
|
||||||
)
|
|
||||||
|
|
||||||
# ignored rules:
|
|
||||||
# E402 module level import not at top of file
|
|
||||||
# W503 line break before binary operator
|
|
||||||
|
|
||||||
# ubu1604: uses shellcheck v0.3.7 (from 04/2015), no longer supported!
|
|
||||||
test.sh:
|
|
||||||
shellcheck -x -s bash utils/brand.env
|
|
||||||
shellcheck -x utils/lib.sh
|
|
||||||
shellcheck -x utils/filtron.sh
|
|
||||||
shellcheck -x utils/searx.sh
|
|
||||||
shellcheck -x utils/morty.sh
|
|
||||||
shellcheck -x utils/lxc.sh
|
|
||||||
shellcheck -x utils/lxc-searx.env
|
|
||||||
shellcheck -x .config.sh
|
|
||||||
|
|
||||||
test.pep8: pyenvinstall
|
|
||||||
@echo "TEST pycodestyle (formerly pep8)"
|
|
||||||
$(Q)$(PY_ENV_ACT); pycodestyle --exclude='searx/static, searx/languages.py, $(foreach f,$(PYLINT_FILES),$(f),)' \
|
|
||||||
--max-line-length=120 --ignore "E117,E252,E402,E722,E741,W503,W504,W605" searx tests
|
|
||||||
|
|
||||||
test.unit: pyenvinstall
|
|
||||||
@echo "TEST tests/unit"
|
|
||||||
$(Q)$(PY_ENV_ACT); python -m nose2 -s tests/unit
|
|
||||||
|
|
||||||
test.coverage: pyenvinstall
|
|
||||||
@echo "TEST unit test coverage"
|
|
||||||
$(Q)$(PY_ENV_ACT); \
|
|
||||||
python -m nose2 -C --log-capture --with-coverage --coverage searx -s tests/unit \
|
|
||||||
&& coverage report \
|
|
||||||
&& coverage html \
|
|
||||||
|
|
||||||
test.robot: pyenvinstall gecko.driver
|
|
||||||
@echo "TEST robot"
|
|
||||||
$(Q)$(PY_ENV_ACT); PYTHONPATH=. python searx/testing.py robot
|
|
||||||
|
|
||||||
test.clean:
|
|
||||||
@echo "CLEAN intermediate test stuff"
|
|
||||||
$(Q)rm -rf geckodriver.log .coverage coverage/
|
|
||||||
|
|
||||||
|
|
||||||
# travis
|
|
||||||
# ------
|
|
||||||
|
|
||||||
PHONY += ci.test
|
|
||||||
ci.test:
|
|
||||||
$(PY_ENV_BIN)/python -c "import yaml" || make clean
|
|
||||||
$(MAKE) test
|
|
||||||
|
|
||||||
travis.codecov:
|
|
||||||
$(Q)$(PY_ENV_BIN)/python -m pip install codecov
|
|
||||||
|
|
||||||
.PHONY: $(PHONY)
|
|
||||||
|
|
|
@ -49,9 +49,9 @@ Build docs
|
||||||
- dvisvgm_
|
- dvisvgm_
|
||||||
|
|
||||||
Most of the sphinx requirements are installed from :origin:`setup.py` and the
|
Most of the sphinx requirements are installed from :origin:`setup.py` and the
|
||||||
docs can be build from scratch with ``make docs``. For better math and image
|
docs can be build from scratch with ``make docs.html``. For better math and
|
||||||
processing additional packages are needed. The XeTeX_ needed not only for PDF
|
image processing additional packages are needed. The XeTeX_ needed not only for
|
||||||
creation, its also needed for :ref:`math` when HTML output is build.
|
PDF creation, its also needed for :ref:`math` when HTML output is build.
|
||||||
|
|
||||||
To be able to do :ref:`sphinx:math-support` without CDNs, the math are rendered
|
To be able to do :ref:`sphinx:math-support` without CDNs, the math are rendered
|
||||||
as images (``sphinx.ext.imgmath`` extension).
|
as images (``sphinx.ext.imgmath`` extension).
|
||||||
|
@ -64,7 +64,7 @@ to ``imgmath``:
|
||||||
:start-after: # sphinx.ext.imgmath setup
|
:start-after: # sphinx.ext.imgmath setup
|
||||||
:end-before: # sphinx.ext.imgmath setup END
|
:end-before: # sphinx.ext.imgmath setup END
|
||||||
|
|
||||||
If your docs build (``make docs``) shows warnings like this::
|
If your docs build (``make docs.html``) shows warnings like this::
|
||||||
|
|
||||||
WARNING: dot(1) not found, for better output quality install \
|
WARNING: dot(1) not found, for better output quality install \
|
||||||
graphviz from https://www.graphviz.org
|
graphviz from https://www.graphviz.org
|
||||||
|
|
|
@ -51,7 +51,7 @@ It's also possible to build searx from the embedded Dockerfile.
|
||||||
|
|
||||||
git clone https://github.com/searx/searx.git
|
git clone https://github.com/searx/searx.git
|
||||||
cd searx
|
cd searx
|
||||||
make docker
|
make docker.build
|
||||||
|
|
||||||
|
|
||||||
Public instance
|
Public instance
|
||||||
|
|
|
@ -31,7 +31,7 @@ might fail in some aspects we should not overlook.
|
||||||
|
|
||||||
The environment in which we run all our development processes matters!
|
The environment in which we run all our development processes matters!
|
||||||
|
|
||||||
The :ref:`makefile` and the :ref:`make pyenv` encapsulate a lot for us, but they
|
The :ref:`makefile` and the :ref:`make install` encapsulate a lot for us, but they
|
||||||
do not have access to all prerequisites. For example, there may have
|
do not have access to all prerequisites. For example, there may have
|
||||||
dependencies on packages that are installed on the developer's desktop, but
|
dependencies on packages that are installed on the developer's desktop, but
|
||||||
usually are not preinstalled on a server or client system. Another examples
|
usually are not preinstalled on a server or client system. Another examples
|
||||||
|
@ -356,7 +356,7 @@ daily usage:
|
||||||
.. code:: sh
|
.. code:: sh
|
||||||
|
|
||||||
$ sudo -H ./utils/lxc.sh cmd searx-archlinux \
|
$ sudo -H ./utils/lxc.sh cmd searx-archlinux \
|
||||||
make docs
|
make docs.html
|
||||||
|
|
||||||
.. _blog-lxcdev-202006 abstract:
|
.. _blog-lxcdev-202006 abstract:
|
||||||
|
|
||||||
|
@ -407,7 +407,7 @@ To get remarks from the suite of the archlinux container we can use:
|
||||||
...
|
...
|
||||||
[searx-archlinux] INFO: (eth0) filtron: http://10.174.184.156:4004/ http://10.174.184.156/searx
|
[searx-archlinux] INFO: (eth0) filtron: http://10.174.184.156:4004/ http://10.174.184.156/searx
|
||||||
[searx-archlinux] INFO: (eth0) morty: http://10.174.184.156:3000/
|
[searx-archlinux] INFO: (eth0) morty: http://10.174.184.156:3000/
|
||||||
[searx-archlinux] INFO: (eth0) docs-live: http://10.174.184.156:8080/
|
[searx-archlinux] INFO: (eth0) docs.live: http://10.174.184.156:8080/
|
||||||
[searx-archlinux] INFO: (eth0) IPv6: http://[fd42:573b:e0b3:e97e:216:3eff:fea5:9b65]
|
[searx-archlinux] INFO: (eth0) IPv6: http://[fd42:573b:e0b3:e97e:216:3eff:fea5:9b65]
|
||||||
...
|
...
|
||||||
|
|
||||||
|
|
|
@ -1,7 +1,7 @@
|
||||||
# -*- coding: utf-8 -*-
|
# -*- coding: utf-8 -*-
|
||||||
|
# SPDX-License-Identifier: AGPL-3.0-or-later
|
||||||
|
|
||||||
import sys, os
|
import sys, os
|
||||||
from sphinx_build_tools import load_sphinx_config
|
|
||||||
from pallets_sphinx_themes import ProjectLink
|
from pallets_sphinx_themes import ProjectLink
|
||||||
|
|
||||||
from searx import brand
|
from searx import brand
|
||||||
|
@ -128,9 +128,3 @@ html_show_sourcelink = False
|
||||||
latex_documents = [
|
latex_documents = [
|
||||||
(master_doc, "searx-{}.tex".format(VERSION_STRING), html_title, author, "manual")
|
(master_doc, "searx-{}.tex".format(VERSION_STRING), html_title, author, "manual")
|
||||||
]
|
]
|
||||||
|
|
||||||
# ------------------------------------------------------------------------------
|
|
||||||
# Since loadConfig overwrites settings from the global namespace, it has to be
|
|
||||||
# the last statement in the conf.py file
|
|
||||||
# ------------------------------------------------------------------------------
|
|
||||||
load_sphinx_config(globals())
|
|
||||||
|
|
|
@ -132,11 +132,11 @@ Here is an example which makes a complete rebuild:
|
||||||
|
|
||||||
.. code:: sh
|
.. code:: sh
|
||||||
|
|
||||||
$ make docs-clean docs
|
$ make docs.clean docs.html
|
||||||
...
|
...
|
||||||
The HTML pages are in dist/docs.
|
The HTML pages are in dist/docs.
|
||||||
|
|
||||||
.. _make docs-live:
|
.. _make docs.live:
|
||||||
|
|
||||||
live build
|
live build
|
||||||
----------
|
----------
|
||||||
|
@ -144,19 +144,19 @@ live build
|
||||||
.. _sphinx-autobuild:
|
.. _sphinx-autobuild:
|
||||||
https://github.com/executablebooks/sphinx-autobuild/blob/master/README.md
|
https://github.com/executablebooks/sphinx-autobuild/blob/master/README.md
|
||||||
|
|
||||||
.. sidebar:: docs-clean
|
.. sidebar:: docs.clean
|
||||||
|
|
||||||
It is recommended to assert a complete rebuild before deploying (use
|
It is recommended to assert a complete rebuild before deploying (use
|
||||||
``docs-clean``).
|
``docs.clean``).
|
||||||
|
|
||||||
Live build is like WYSIWYG. If you want to edit the documentation, its
|
Live build is like WYSIWYG. If you want to edit the documentation, its
|
||||||
recommended to use. The Makefile target ``docs-live`` builds the docs, opens
|
recommended to use. The Makefile target ``docs.live`` builds the docs, opens
|
||||||
URL in your favorite browser and rebuilds every time a reST file has been
|
URL in your favorite browser and rebuilds every time a reST file has been
|
||||||
changed.
|
changed.
|
||||||
|
|
||||||
.. code:: sh
|
.. code:: sh
|
||||||
|
|
||||||
$ make docs-live
|
$ make docs.live
|
||||||
...
|
...
|
||||||
The HTML pages are in dist/docs.
|
The HTML pages are in dist/docs.
|
||||||
... Serving on http://0.0.0.0:8000
|
... Serving on http://0.0.0.0:8000
|
||||||
|
@ -169,7 +169,7 @@ argument. E.g to find and use a free port, use:
|
||||||
|
|
||||||
.. code:: sh
|
.. code:: sh
|
||||||
|
|
||||||
$ SPHINXOPTS="--port 0" make docs-live
|
$ SPHINXOPTS="--port 0" make docs.live
|
||||||
...
|
...
|
||||||
... Serving on http://0.0.0.0:50593
|
... Serving on http://0.0.0.0:50593
|
||||||
...
|
...
|
||||||
|
@ -180,21 +180,10 @@ argument. E.g to find and use a free port, use:
|
||||||
deploy on github.io
|
deploy on github.io
|
||||||
-------------------
|
-------------------
|
||||||
|
|
||||||
To deploy documentation at :docs:`github.io <.>` use Makefile target
|
To deploy documentation at :docs:`github.io <.>` use Makefile target :ref:`make
|
||||||
:ref:`make gh-pages`, which will builds the documentation, clones searx into a sub
|
docs.gh-pages`, which builds the documentation and runs all the needed git add,
|
||||||
folder ``gh-pages``, cleans it, copies the doc build into and runs all the
|
commit and push:
|
||||||
needed git add, commit and push:
|
|
||||||
|
|
||||||
.. code:: sh
|
.. code:: sh
|
||||||
|
|
||||||
$ make docs-clean gh-pages
|
$ make docs.clean docs.gh-pages
|
||||||
...
|
|
||||||
SPHINX docs --> file://<...>/dist/docs
|
|
||||||
The HTML pages are in dist/docs.
|
|
||||||
...
|
|
||||||
Cloning into 'gh-pages' ...
|
|
||||||
...
|
|
||||||
cd gh-pages; git checkout gh-pages >/dev/null
|
|
||||||
Switched to a new branch 'gh-pages'
|
|
||||||
...
|
|
||||||
doc available at --> https://searx.github.io/searx
|
|
||||||
|
|
|
@ -1,33 +1,33 @@
|
||||||
.. _makefile:
|
.. _makefile:
|
||||||
|
|
||||||
================
|
========
|
||||||
Makefile Targets
|
Makefile
|
||||||
================
|
========
|
||||||
|
|
||||||
.. _gnu-make: https://www.gnu.org/software/make/manual/make.html#Introduction
|
.. _gnu-make: https://www.gnu.org/software/make/manual/make.html#Introduction
|
||||||
|
|
||||||
.. sidebar:: build environment
|
.. sidebar:: build environment
|
||||||
|
|
||||||
Before looking deeper at the targets, first read about :ref:`make pyenv`.
|
Before looking deeper at the targets, first read about :ref:`make
|
||||||
|
install`.
|
||||||
|
|
||||||
To install system requirements follow :ref:`buildhosts`.
|
To install system requirements follow :ref:`buildhosts`.
|
||||||
|
|
||||||
With the aim to simplify development cycles, started with :pull:`1756` a
|
All relevant build tasks are implemented in :origin:`manage.sh` and for CI or
|
||||||
``Makefile`` based boilerplate was added. If you are not familiar with
|
IDE integration a small ``Makefile`` wrapper is available. If you are not
|
||||||
Makefiles, we recommend to read gnu-make_ introduction.
|
familiar with Makefiles, we recommend to read gnu-make_ introduction.
|
||||||
|
|
||||||
The usage is simple, just type ``make {target-name}`` to *build* a target.
|
The usage is simple, just type ``make {target-name}`` to *build* a target.
|
||||||
Calling the ``help`` target gives a first overview (``make help``):
|
Calling the ``help`` target gives a first overview (``make help``):
|
||||||
|
|
||||||
.. program-output:: bash -c "cd ..; make --no-print-directory help"
|
.. program-output:: bash -c "cd ..; make --no-print-directory help"
|
||||||
|
|
||||||
|
|
||||||
.. contents:: Contents
|
.. contents:: Contents
|
||||||
:depth: 2
|
:depth: 2
|
||||||
:local:
|
:local:
|
||||||
:backlinks: entry
|
:backlinks: entry
|
||||||
|
|
||||||
.. _make pyenv:
|
.. _make install:
|
||||||
|
|
||||||
Python environment
|
Python environment
|
||||||
==================
|
==================
|
||||||
|
@ -36,31 +36,42 @@ Python environment
|
||||||
|
|
||||||
``source ./local/py3/bin/activate``
|
``source ./local/py3/bin/activate``
|
||||||
|
|
||||||
With Makefile we do no longer need to build up the virtualenv manually (as
|
We do no longer need to build up the virtualenv manually. Jump into your git
|
||||||
described in the :ref:`devquickstart` guide). Jump into your git working tree
|
working tree and release a ``make install`` to get a virtualenv with a
|
||||||
and release a ``make pyenv``:
|
*developer install* of searx (:origin:`setup.py`). ::
|
||||||
|
|
||||||
.. code:: sh
|
|
||||||
|
|
||||||
$ cd ~/searx-clone
|
$ cd ~/searx-clone
|
||||||
$ make pyenv
|
$ make install
|
||||||
PYENV usage: source ./local/py3/bin/activate
|
PYENV [virtualenv] installing ./requirements*.txt into local/py3
|
||||||
...
|
...
|
||||||
|
PYENV OK
|
||||||
|
PYENV [install] pip install -e 'searx[test]'
|
||||||
|
...
|
||||||
|
Successfully installed argparse-1.4.0 searx
|
||||||
|
BUILDENV INFO:searx:load the default settings from ./searx/settings.yml
|
||||||
|
BUILDENV INFO:searx:Initialisation done
|
||||||
|
BUILDENV build utils/brand.env
|
||||||
|
|
||||||
With target ``pyenv`` a development environment (aka virtualenv) was build up in
|
If you release ``make install`` multiple times the installation will only
|
||||||
``./local/py3/``. To make a *developer install* of searx (:origin:`setup.py`)
|
rebuild if the sha256 sum of the *requirement files* fails. With other words:
|
||||||
into this environment, use make target ``install``:
|
the check fails if you edit the requirements listed in
|
||||||
|
:origin:`requirements-dev.txt` and :origin:`requirements.txt`). ::
|
||||||
.. code:: sh
|
|
||||||
|
|
||||||
$ make install
|
$ make install
|
||||||
PYENV usage: source ./local/py3/bin/activate
|
PYENV OK
|
||||||
PYENV using virtualenv from ./local/py3
|
PYENV [virtualenv] requirements.sha256 failed
|
||||||
PYENV install .
|
[virtualenv] - 6cea6eb6def9e14a18bf32f8a3e... ./requirements-dev.txt
|
||||||
|
[virtualenv] - 471efef6c73558e391c3adb35f4... ./requirements.txt
|
||||||
You have never to think about intermediate targets like ``pyenv`` or
|
...
|
||||||
``install``, the ``Makefile`` chains them as requisites. Just run your main
|
PYENV [virtualenv] installing ./requirements*.txt into local/py3
|
||||||
target.
|
...
|
||||||
|
PYENV OK
|
||||||
|
PYENV [install] pip install -e 'searx[test]'
|
||||||
|
...
|
||||||
|
Successfully installed argparse-1.4.0 searx
|
||||||
|
BUILDENV INFO:searx:load the default settings from ./searx/settings.yml
|
||||||
|
BUILDENV INFO:searx:Initialisation done
|
||||||
|
BUILDENV build utils/brand.env
|
||||||
|
|
||||||
.. sidebar:: drop environment
|
.. sidebar:: drop environment
|
||||||
|
|
||||||
|
@ -68,10 +79,7 @@ target.
|
||||||
<make clean>` first.
|
<make clean>` first.
|
||||||
|
|
||||||
If you think, something goes wrong with your ./local environment or you change
|
If you think, something goes wrong with your ./local environment or you change
|
||||||
the :origin:`setup.py` file (or the requirements listed in
|
the :origin:`setup.py` file, you have to call :ref:`make clean`.
|
||||||
:origin:`requirements-dev.txt` and :origin:`requirements.txt`), you have to call
|
|
||||||
:ref:`make clean`.
|
|
||||||
|
|
||||||
|
|
||||||
.. _make run:
|
.. _make run:
|
||||||
|
|
||||||
|
@ -81,77 +89,44 @@ the :origin:`setup.py` file (or the requirements listed in
|
||||||
To get up a running a developer instance simply call ``make run``. This enables
|
To get up a running a developer instance simply call ``make run``. This enables
|
||||||
*debug* option in :origin:`searx/settings.yml`, starts a ``./searx/webapp.py``
|
*debug* option in :origin:`searx/settings.yml`, starts a ``./searx/webapp.py``
|
||||||
instance, disables *debug* option again and opens the URL in your favorite WEB
|
instance, disables *debug* option again and opens the URL in your favorite WEB
|
||||||
browser (:man:`xdg-open`):
|
browser (:man:`xdg-open`)::
|
||||||
|
|
||||||
.. code:: sh
|
$ make run
|
||||||
|
PYENV OK
|
||||||
$ make run
|
SEARX_DEBUG=1 ./manage.sh pyenv.cmd python ./searx/webapp.py
|
||||||
PYENV usage: source ./local/py3/bin/activate
|
...
|
||||||
PYENV install .
|
INFO:werkzeug: * Running on http://127.0.0.1:8888/ (Press CTRL+C to quit)
|
||||||
./local/py3/bin/python ./searx/webapp.py
|
|
||||||
...
|
|
||||||
INFO:werkzeug: * Running on http://127.0.0.1:8888/ (Press CTRL+C to quit)
|
|
||||||
...
|
|
||||||
|
|
||||||
.. _make clean:
|
.. _make clean:
|
||||||
|
|
||||||
``make clean``
|
``make clean``
|
||||||
==============
|
==============
|
||||||
|
|
||||||
Drop all intermediate files, all builds, but keep sources untouched. Includes
|
Drop all intermediate files, all builds, but keep sources untouched. Before
|
||||||
target ``pyclean`` which drops ./local environment. Before calling ``make
|
calling ``make clean`` stop all processes using :ref:`make install`. ::
|
||||||
clean`` stop all processes using :ref:`make pyenv`.
|
|
||||||
|
|
||||||
.. code:: sh
|
|
||||||
|
|
||||||
$ make clean
|
$ make clean
|
||||||
CLEAN pyclean
|
CLEAN pyenv
|
||||||
CLEAN clean
|
PYENV [virtualenv] drop ./local/py3
|
||||||
|
CLEAN docs -- ./build/docs ./dist/docs
|
||||||
|
CLEAN locally installed npm dependencies
|
||||||
|
CLEAN test stuff
|
||||||
|
CLEAN common files
|
||||||
|
|
||||||
.. _make docs:
|
.. _make docs:
|
||||||
|
|
||||||
``make docs docs-live docs-clean``
|
``make docs docs.autobuild docs.clean``
|
||||||
==================================
|
=======================================
|
||||||
|
|
||||||
We describe the usage of the ``doc*`` targets in the :ref:`How to contribute /
|
We describe the usage of the ``doc.*`` targets in the :ref:`How to contribute /
|
||||||
Documentation <contrib docs>` section. If you want to edit the documentation
|
Documentation <contrib docs>` section. If you want to edit the documentation
|
||||||
read our :ref:`make docs-live` section. If you are working in your own brand,
|
read our :ref:`make docs.live` section. If you are working in your own brand,
|
||||||
adjust your :ref:`settings global`.
|
adjust your :ref:`settings global`.
|
||||||
|
|
||||||
.. _make books:
|
.. _make docs.gh-pages:
|
||||||
|
|
||||||
``make books/{name}.html books/{name}.pdf``
|
``make docs.gh-pages``
|
||||||
===========================================
|
======================
|
||||||
|
|
||||||
.. _intersphinx: https://www.sphinx-doc.org/en/stable/ext/intersphinx.html
|
|
||||||
.. _XeTeX: https://tug.org/xetex/
|
|
||||||
|
|
||||||
.. sidebar:: info
|
|
||||||
|
|
||||||
To build PDF a XeTeX_ is needed, see :ref:`buildhosts`.
|
|
||||||
|
|
||||||
|
|
||||||
The ``books/{name}.*`` targets are building *books*. A *book* is a
|
|
||||||
sub-directory containing a ``conf.py`` file. One example is the user handbook
|
|
||||||
which can deployed separately (:origin:`docs/user/conf.py`). Such ``conf.py``
|
|
||||||
do inherit from :origin:`docs/conf.py` and overwrite values to fit *book's*
|
|
||||||
needs.
|
|
||||||
|
|
||||||
With the help of Intersphinx_ (:ref:`reST smart ref`) the links to searx’s
|
|
||||||
documentation outside of the book will be bound by the object inventory of
|
|
||||||
``DOCS_URL``. Take into account that URLs will be picked from the inventary at
|
|
||||||
documentation's build time.
|
|
||||||
|
|
||||||
Use ``make docs-help`` to see which books available:
|
|
||||||
|
|
||||||
.. program-output:: bash -c "cd ..; make --no-print-directory docs-help"
|
|
||||||
:ellipsis: 0,-6
|
|
||||||
|
|
||||||
|
|
||||||
.. _make gh-pages:
|
|
||||||
|
|
||||||
``make gh-pages``
|
|
||||||
=================
|
|
||||||
|
|
||||||
To deploy on github.io first adjust your :ref:`settings global`. For any
|
To deploy on github.io first adjust your :ref:`settings global`. For any
|
||||||
further read :ref:`deploy on github.io`.
|
further read :ref:`deploy on github.io`.
|
||||||
|
@ -161,37 +136,66 @@ further read :ref:`deploy on github.io`.
|
||||||
``make test``
|
``make test``
|
||||||
=============
|
=============
|
||||||
|
|
||||||
Runs a series of tests: ``test.pep8``, ``test.unit``, ``test.robot`` and does
|
Runs a series of tests: :ref:`make test.pylint`, ``test.pep8``, ``test.unit``
|
||||||
additional :ref:`pylint checks <make pylint>`. You can run tests selective,
|
and ``test.robot``. You can run tests selective, e.g.::
|
||||||
e.g.:
|
|
||||||
|
|
||||||
.. code:: sh
|
|
||||||
|
|
||||||
$ make test.pep8 test.unit test.sh
|
$ make test.pep8 test.unit test.sh
|
||||||
. ./local/py3/bin/activate; ./manage.sh pep8_check
|
TEST test.pep8 OK
|
||||||
[!] Running pep8 check
|
...
|
||||||
. ./local/py3/bin/activate; ./manage.sh unit_tests
|
TEST test.unit OK
|
||||||
[!] Running unit tests
|
...
|
||||||
|
TEST test.sh OK
|
||||||
|
|
||||||
.. _make pylint:
|
.. _make test.sh:
|
||||||
|
|
||||||
``make pylint``
|
``make test.sh``
|
||||||
===============
|
================
|
||||||
|
|
||||||
|
:ref:`sh lint` / if you have changed some bash scripting run this test before
|
||||||
|
commit.
|
||||||
|
|
||||||
|
.. _make test.pylint:
|
||||||
|
|
||||||
|
``make test.pylint``
|
||||||
|
====================
|
||||||
|
|
||||||
.. _Pylint: https://www.pylint.org/
|
.. _Pylint: https://www.pylint.org/
|
||||||
|
|
||||||
Before commiting its recommend to do some (more) linting. Pylint_ is known as
|
Pylint_ is known as one of the best source-code, bug and quality checker for the
|
||||||
one of the best source-code, bug and quality checker for the Python programming
|
Python programming language. The pylint profile we use at searx project is
|
||||||
language. Pylint_ is not yet a quality gate within our searx project (like
|
found in project's root folder :origin:`.pylintrc`.
|
||||||
:ref:`test.pep8 <make test>` it is), but Pylint_ can help to improve code
|
|
||||||
quality anyway. The pylint profile we use at searx project is found in
|
|
||||||
project's root folder :origin:`.pylintrc`.
|
|
||||||
|
|
||||||
Code quality is a ongoing process. Don't try to fix all messages from Pylint,
|
.. _make search.checker:
|
||||||
run Pylint and check if your changed lines are bringing up new messages. If so,
|
|
||||||
fix it. By this, code quality gets incremental better and if there comes the
|
``search.checker.{engine name}``
|
||||||
day, the linting is balanced out, we might decide to add Pylint as a quality
|
================================
|
||||||
gate.
|
|
||||||
|
To check all engines::
|
||||||
|
|
||||||
|
make search.checker
|
||||||
|
|
||||||
|
To check a engine with whitespace in the name like *google news* replace space
|
||||||
|
by underline::
|
||||||
|
|
||||||
|
make search.checker.google_news
|
||||||
|
|
||||||
|
To see HTTP requests and more use SEARX_DEBUG::
|
||||||
|
|
||||||
|
make SEARX_DEBUG=1 search.checker.google_news
|
||||||
|
|
||||||
|
.. _3xx: https://en.wikipedia.org/wiki/List_of_HTTP_status_codes#3xx_redirection
|
||||||
|
|
||||||
|
To filter out HTTP redirects (3xx_)::
|
||||||
|
|
||||||
|
make SEARX_DEBUG=1 search.checker.google_news | grep -A1 "HTTP/1.1\" 3[0-9][0-9]"
|
||||||
|
...
|
||||||
|
Engine google news Checking
|
||||||
|
https://news.google.com:443 "GET /search?q=life&hl=en&lr=lang_en&ie=utf8&oe=utf8&ceid=US%3Aen&gl=US HTTP/1.1" 302 0
|
||||||
|
https://news.google.com:443 "GET /search?q=life&hl=en-US&lr=lang_en&ie=utf8&oe=utf8&ceid=US:en&gl=US HTTP/1.1" 200 None
|
||||||
|
--
|
||||||
|
https://news.google.com:443 "GET /search?q=computer&hl=en&lr=lang_en&ie=utf8&oe=utf8&ceid=US%3Aen&gl=US HTTP/1.1" 302 0
|
||||||
|
https://news.google.com:443 "GET /search?q=computer&hl=en-US&lr=lang_en&ie=utf8&oe=utf8&ceid=US:en&gl=US HTTP/1.1" 200 None
|
||||||
|
--
|
||||||
|
|
||||||
|
|
||||||
``make pybuild``
|
``make pybuild``
|
||||||
|
@ -200,9 +204,7 @@ gate.
|
||||||
.. _PyPi: https://pypi.org/
|
.. _PyPi: https://pypi.org/
|
||||||
.. _twine: https://twine.readthedocs.io/en/latest/
|
.. _twine: https://twine.readthedocs.io/en/latest/
|
||||||
|
|
||||||
Build Python packages in ``./dist/py``.
|
Build Python packages in ``./dist/py``::
|
||||||
|
|
||||||
.. code:: sh
|
|
||||||
|
|
||||||
$ make pybuild
|
$ make pybuild
|
||||||
...
|
...
|
||||||
|
@ -210,9 +212,11 @@ Build Python packages in ``./dist/py``.
|
||||||
running sdist
|
running sdist
|
||||||
running egg_info
|
running egg_info
|
||||||
...
|
...
|
||||||
$ ls ./dist/py/
|
running bdist_wheel
|
||||||
searx-0.15.0-py3-none-any.whl searx-0.15.0.tar.gz
|
|
||||||
|
|
||||||
To upload packages to PyPi_, there is also a ``upload-pypi`` target. It needs
|
$ ls ./dist
|
||||||
twine_ to be installed. Since you are not the owner of :pypi:`searx` you will
|
searx-0.18.0-py3-none-any.whl searx-0.18.0.tar.gz
|
||||||
never need the latter.
|
|
||||||
|
To upload packages to PyPi_, there is also a ``pypi.upload`` target (to test use
|
||||||
|
``pypi.upload.test``). Since you are not the owner of :pypi:`searx` you will
|
||||||
|
never need to upload.
|
||||||
|
|
|
@ -15,8 +15,8 @@ generated and deployed at :docs:`github.io <.>`. For build prerequisites read
|
||||||
:ref:`docs build`.
|
:ref:`docs build`.
|
||||||
|
|
||||||
The source files of Searx's documentation are located at :origin:`docs`. Sphinx
|
The source files of Searx's documentation are located at :origin:`docs`. Sphinx
|
||||||
assumes source files to be encoded in UTF-8 by defaul. Run :ref:`make docs-live
|
assumes source files to be encoded in UTF-8 by defaul. Run :ref:`make docs.live
|
||||||
<make docs-live>` to build HTML while editing.
|
<make docs.live>` to build HTML while editing.
|
||||||
|
|
||||||
.. sidebar:: Further reading
|
.. sidebar:: Further reading
|
||||||
|
|
||||||
|
@ -1276,13 +1276,12 @@ Templating
|
||||||
|
|
||||||
.. sidebar:: Build environment
|
.. sidebar:: Build environment
|
||||||
|
|
||||||
All *generic-doc* tasks are running in the :ref:`build environment <make
|
All *generic-doc* tasks are running in the :ref:`make install`.
|
||||||
pyenv>`.
|
|
||||||
|
|
||||||
Templating is suitable for documentation which is created generic at the build
|
Templating is suitable for documentation which is created generic at the build
|
||||||
time. The sphinx-jinja_ extension evaluates jinja_ templates in the :ref:`build
|
time. The sphinx-jinja_ extension evaluates jinja_ templates in the :ref:`make
|
||||||
environment <make pyenv>` (with searx modules installed). We use this e.g. to
|
install` (with searx modules installed). We use this e.g. to build chapter:
|
||||||
build chapter: :ref:`engines generic`. Below the jinja directive from the
|
:ref:`engines generic`. Below the jinja directive from the
|
||||||
:origin:`docs/admin/engines.rst` is shown:
|
:origin:`docs/admin/engines.rst` is shown:
|
||||||
|
|
||||||
.. literalinclude:: ../admin/engines.rst
|
.. literalinclude:: ../admin/engines.rst
|
||||||
|
|
|
@ -1,21 +0,0 @@
|
||||||
# -*- coding: utf-8; mode: python -*-
|
|
||||||
"""Configuration for the Searx user handbook
|
|
||||||
"""
|
|
||||||
project = 'Searx User-HB'
|
|
||||||
version = release = VERSION_STRING
|
|
||||||
|
|
||||||
intersphinx_mapping['searx'] = (brand.DOCS_URL, None)
|
|
||||||
|
|
||||||
# Grouping the document tree into LaTeX files. List of tuples
|
|
||||||
# (source start file, target name, title,
|
|
||||||
# author, documentclass [howto, manual, or own class]).
|
|
||||||
latex_documents = [
|
|
||||||
('index' # startdocname
|
|
||||||
, 'searx-user-hb.tex' # targetname
|
|
||||||
, '' # take title from .rst
|
|
||||||
, author # author
|
|
||||||
, 'howto' # documentclass
|
|
||||||
, False # toctree_only
|
|
||||||
),
|
|
||||||
]
|
|
||||||
|
|
|
@ -119,15 +119,15 @@ of coffee).::
|
||||||
|
|
||||||
To build (live) documentation inside a archlinux_ container::
|
To build (live) documentation inside a archlinux_ container::
|
||||||
|
|
||||||
sudo -H ./utils/lxc.sh cmd searx-archlinux make docs-clean docs-live
|
sudo -H ./utils/lxc.sh cmd searx-archlinux make docs.clean docs.live
|
||||||
...
|
...
|
||||||
[I 200331 15:00:42 server:296] Serving on http://0.0.0.0:8080
|
[I 200331 15:00:42 server:296] Serving on http://0.0.0.0:8080
|
||||||
|
|
||||||
To get IP of the container and the port number *live docs* is listening::
|
To get IP of the container and the port number *live docs* is listening::
|
||||||
|
|
||||||
$ sudo ./utils/lxc.sh show suite | grep docs-live
|
$ sudo ./utils/lxc.sh show suite | grep docs.live
|
||||||
...
|
...
|
||||||
[searx-archlinux] INFO: (eth0) docs-live: http://n.n.n.12:8080/
|
[searx-archlinux] INFO: (eth0) docs.live: http://n.n.n.12:8080/
|
||||||
|
|
||||||
|
|
||||||
.. _lxc.sh help:
|
.. _lxc.sh help:
|
||||||
|
|
506
manage
Executable file
506
manage
Executable file
|
@ -0,0 +1,506 @@
|
||||||
|
#!/usr/bin/env bash
|
||||||
|
# -*- coding: utf-8; mode: sh indent-tabs-mode: nil -*-
|
||||||
|
# SPDX-License-Identifier: AGPL-3.0-or-later
|
||||||
|
# shellcheck disable=SC2031
|
||||||
|
|
||||||
|
# shellcheck source=utils/lib.sh
|
||||||
|
source "$(dirname "${BASH_SOURCE[0]}")/utils/lib.sh"
|
||||||
|
# shellcheck source=utils/brand.env
|
||||||
|
source "${REPO_ROOT}/utils/brand.env"
|
||||||
|
source_dot_config
|
||||||
|
|
||||||
|
# config
|
||||||
|
|
||||||
|
PYOBJECTS="searx"
|
||||||
|
PY_SETUP_EXTRAS='[test]'
|
||||||
|
NPM_PACKAGES="less@2.7 less-plugin-clean-css grunt-cli"
|
||||||
|
GECKODRIVER_VERSION="v0.28.0"
|
||||||
|
# SPHINXOPTS=
|
||||||
|
|
||||||
|
# These py files are linted by test.pylint(), all other files are linted by
|
||||||
|
# test.pep8()
|
||||||
|
PYLINT_FILES=(
|
||||||
|
searx/preferences.py
|
||||||
|
searx/testing.py
|
||||||
|
searx/engines/gigablast.py
|
||||||
|
searx/engines/deviantart.py
|
||||||
|
searx/engines/digg.py
|
||||||
|
searx/engines/google.py
|
||||||
|
searx/engines/google_news.py
|
||||||
|
searx/engines/google_videos.py
|
||||||
|
searx/engines/google_images.py
|
||||||
|
searx/engines/mediathekviewweb.py
|
||||||
|
searx/engines/meilisearch.py
|
||||||
|
searx/engines/solidtorrents.py
|
||||||
|
searx/engines/solr.py
|
||||||
|
searx/engines/springer.py
|
||||||
|
searx/engines/google_scholar.py
|
||||||
|
searx/engines/yahoo_news.py
|
||||||
|
searx/engines/apkmirror.py
|
||||||
|
searx/engines/artic.py
|
||||||
|
searx_extra/update/update_external_bangs.py
|
||||||
|
)
|
||||||
|
|
||||||
|
PYLINT_SEARX_DISABLE_OPTION="\
|
||||||
|
I,C,R,\
|
||||||
|
W0105,W0212,W0511,W0603,W0613,W0621,W0702,W0703,W1401,\
|
||||||
|
E1136"
|
||||||
|
PYLINT_ADDITIONAL_BUILTINS_FOR_ENGINES="supported_languages,language_aliases"
|
||||||
|
PYLINT_OPTIONS="-m pylint -j 0 --rcfile .pylintrc"
|
||||||
|
|
||||||
|
help() {
|
||||||
|
cat <<EOF
|
||||||
|
buildenv
|
||||||
|
rebuild ./utils/brand.env
|
||||||
|
babel.compile
|
||||||
|
pybabel compile ./searx/translations
|
||||||
|
data.*
|
||||||
|
all : update searx/languages.py and ./data/*
|
||||||
|
languages : update searx/data/engines_languages.json & searx/languages.py
|
||||||
|
useragents: update searx/data/useragents.json with the most recent versions of Firefox.
|
||||||
|
docs.*
|
||||||
|
html : build HTML documentation
|
||||||
|
live : autobuild HTML documentation while editing
|
||||||
|
gh-pages : deploy on gh-pages branch
|
||||||
|
prebuild : build reST include files (./${DOCS_BUILD}/includes)
|
||||||
|
clean : clean documentation build
|
||||||
|
docker
|
||||||
|
build : build docker image
|
||||||
|
push : build and push docker image
|
||||||
|
gecko.driver
|
||||||
|
download & install geckodriver if not already installed (required for
|
||||||
|
robot_tests)
|
||||||
|
node.*
|
||||||
|
env : download & install npm dependencies locally
|
||||||
|
clean : drop npm installations
|
||||||
|
py.*
|
||||||
|
build : Build python packages at ./${PYDIST}
|
||||||
|
clean : delete virtualenv and intermediate py files
|
||||||
|
pyenv.* :
|
||||||
|
install : developer install of searx into virtualenv
|
||||||
|
uninstall : uninstall developer installation
|
||||||
|
cmd ... : run command ... in virtualenv
|
||||||
|
OK : test if virtualenv is OK
|
||||||
|
pypi.upload:
|
||||||
|
Upload python packages to PyPi (to test use pypi.upload.test)
|
||||||
|
test.* :
|
||||||
|
pylint : lint PYLINT_FILES, searx/engines, searx & tests
|
||||||
|
pep8 : pycodestyle (pep8) for all files except PYLINT_FILES
|
||||||
|
unit : run unit tests
|
||||||
|
coverage : run unit tests with coverage
|
||||||
|
robot : run robot test
|
||||||
|
clean : clean intermediate test stuff
|
||||||
|
themes.* :
|
||||||
|
all : build all themes
|
||||||
|
oscar : build oscar theme
|
||||||
|
simple : build simple theme
|
||||||
|
EOF
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
if [ "$VERBOSE" = "1" ]; then
|
||||||
|
SPHINX_VERBOSE="-v"
|
||||||
|
PYLINT_VERBOSE="-v"
|
||||||
|
fi
|
||||||
|
|
||||||
|
# needed by sphinx-docs
|
||||||
|
export DOCS_BUILD
|
||||||
|
|
||||||
|
buildenv() {
|
||||||
|
SEARX_DEBUG=1 pyenv.cmd python utils/build_env.py 2>&1 \
|
||||||
|
| prefix_stdout "${_Blue}BUILDENV${_creset} "
|
||||||
|
return "${PIPESTATUS[0]}"
|
||||||
|
}
|
||||||
|
|
||||||
|
babel.compile() {
|
||||||
|
build_msg BABEL compile
|
||||||
|
pyenv.cmd pybabel compile -d "${REPO_ROOT}/searx/translations"
|
||||||
|
dump_return $?
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
data.all() {
|
||||||
|
data.languages
|
||||||
|
data.useragents
|
||||||
|
build_msg DATA "update searx/data/ahmia_blacklist.txt"
|
||||||
|
pyenv.cmd python searx_extra/update/update_ahmia_blacklist.py
|
||||||
|
build_msg DATA "update searx/data/wikidata_units.json"
|
||||||
|
pyenv.cmd python searx_extra/update/update_wikidata_units.py
|
||||||
|
build_msg DATA "update searx/data/currencies.json"
|
||||||
|
pyenv.cmd python searx_extra/update/update_currencies.py
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
data.languages() {
|
||||||
|
( set -e
|
||||||
|
build_msg ENGINES "fetch languages .."
|
||||||
|
pyenv.cmd python searx_extra/update/update_languages.py
|
||||||
|
build_msg ENGINES "update update searx/languages.py"
|
||||||
|
build_msg DATA "update searx/data/engines_languages.json"
|
||||||
|
)
|
||||||
|
dump_return $?
|
||||||
|
}
|
||||||
|
|
||||||
|
data.useragents() {
|
||||||
|
build_msg DATA "update searx/data/useragents.json"
|
||||||
|
pyenv.cmd python searx_extra/update/update_firefox_version.py
|
||||||
|
dump_return $?
|
||||||
|
}
|
||||||
|
|
||||||
|
docs.prebuild() {
|
||||||
|
build_msg DOCS "build ${DOCS_BUILD}/includes"
|
||||||
|
(
|
||||||
|
set -e
|
||||||
|
[ "$VERBOSE" = "1" ] && set -x
|
||||||
|
mkdir -p "${DOCS_BUILD}/includes"
|
||||||
|
./utils/searx.sh doc | cat > "${DOCS_BUILD}/includes/searx.rst"
|
||||||
|
./utils/filtron.sh doc | cat > "${DOCS_BUILD}/includes/filtron.rst"
|
||||||
|
./utils/morty.sh doc | cat > "${DOCS_BUILD}/includes/morty.rst"
|
||||||
|
)
|
||||||
|
dump_return $?
|
||||||
|
}
|
||||||
|
|
||||||
|
docker.push() {
|
||||||
|
docker.build push
|
||||||
|
}
|
||||||
|
|
||||||
|
# shellcheck disable=SC2119
|
||||||
|
docker.build() {
|
||||||
|
pyenv.install
|
||||||
|
|
||||||
|
build_msg DOCKER build
|
||||||
|
# run installation in a subprocess and activate pyenv
|
||||||
|
|
||||||
|
# See https://www.shellcheck.net/wiki/SC1001 and others ..
|
||||||
|
# shellcheck disable=SC2031,SC2230,SC2002,SC2236,SC2143,SC1001
|
||||||
|
( set -e
|
||||||
|
# shellcheck source=/dev/null
|
||||||
|
source "${PY_ENV_BIN}/activate"
|
||||||
|
|
||||||
|
# Check if it is a git repository
|
||||||
|
if [ ! -d .git ]; then
|
||||||
|
die 1 "This is not Git repository"
|
||||||
|
fi
|
||||||
|
if [ ! -x "$(which git)" ]; then
|
||||||
|
die 1 "git is not installed"
|
||||||
|
fi
|
||||||
|
|
||||||
|
if ! git remote get-url origin 2> /dev/null; then
|
||||||
|
die 1 "there is no remote origin"
|
||||||
|
fi
|
||||||
|
|
||||||
|
# This is a git repository
|
||||||
|
|
||||||
|
# "git describe" to get the Docker version (for example : v0.15.0-89-g0585788e)
|
||||||
|
# awk to remove the "v" and the "g"
|
||||||
|
SEARX_GIT_VERSION=$(git describe --match "v[0-9]*\.[0-9]*\.[0-9]*" HEAD 2>/dev/null | awk -F'-' '{OFS="-"; $1=substr($1, 2); if ($3) { $3=substr($3, 2); } print}')
|
||||||
|
|
||||||
|
# add the suffix "-dirty" if the repository has uncommited change
|
||||||
|
# /!\ HACK for searx/searx: ignore utils/brand.env
|
||||||
|
git update-index -q --refresh
|
||||||
|
if [ ! -z "$(git diff-index --name-only HEAD -- | grep -v 'utils/brand.env')" ]; then
|
||||||
|
SEARX_GIT_VERSION="${SEARX_GIT_VERSION}-dirty"
|
||||||
|
fi
|
||||||
|
|
||||||
|
# Get the last git commit id, will be added to the Searx version (see Dockerfile)
|
||||||
|
VERSION_GITCOMMIT=$(echo "$SEARX_GIT_VERSION" | cut -d- -f2-4)
|
||||||
|
build_msg DOCKER "Last commit : $VERSION_GITCOMMIT"
|
||||||
|
|
||||||
|
# Check consistency between the git tag and the searx/version.py file
|
||||||
|
# /! HACK : parse Python file with bash /!
|
||||||
|
# otherwise it is not possible build the docker image without all Python
|
||||||
|
# dependencies ( version.py loads __init__.py )
|
||||||
|
# SEARX_PYTHON_VERSION=$(python3 -c "import six; import searx.version; six.print_(searx.version.VERSION_STRING)")
|
||||||
|
SEARX_PYTHON_VERSION=$(cat searx/version.py | grep "\(VERSION_MAJOR\|VERSION_MINOR\|VERSION_BUILD\) =" | cut -d\= -f2 | sed -e 's/^[[:space:]]*//' | paste -sd "." -)
|
||||||
|
if [ "$(echo "$SEARX_GIT_VERSION" | cut -d- -f1)" != "$SEARX_PYTHON_VERSION" ]; then
|
||||||
|
err_msg "git tag: $SEARX_GIT_VERSION"
|
||||||
|
err_msg "searx/version.py: $SEARX_PYTHON_VERSION"
|
||||||
|
die 1 "Inconsistency between the last git tag and the searx/version.py file"
|
||||||
|
fi
|
||||||
|
|
||||||
|
# define the docker image name
|
||||||
|
GITHUB_USER=$(echo "${GIT_URL}" | sed 's/.*github\.com\/\([^\/]*\).*/\1/')
|
||||||
|
SEARX_IMAGE_NAME="${SEARX_IMAGE_NAME:-${GITHUB_USER:-searxng}/searxng}"
|
||||||
|
|
||||||
|
# build Docker image
|
||||||
|
build_msg DOCKER "Building image ${SEARX_IMAGE_NAME}:${SEARX_GIT_VERSION}"
|
||||||
|
sudo docker build \
|
||||||
|
--build-arg GIT_URL="${GIT_URL}" \
|
||||||
|
--build-arg SEARX_GIT_VERSION="${SEARX_GIT_VERSION}" \
|
||||||
|
--build-arg VERSION_GITCOMMIT="${VERSION_GITCOMMIT}" \
|
||||||
|
--build-arg LABEL_DATE="$(date -u +"%Y-%m-%dT%H:%M:%SZ")" \
|
||||||
|
--build-arg LABEL_VCS_REF="$(git rev-parse HEAD)" \
|
||||||
|
--build-arg LABEL_VCS_URL="${GIT_URL}" \
|
||||||
|
--build-arg TIMESTAMP_SETTINGS="$(git log -1 --format="%cd" --date=unix -- searx/settings.yml)" \
|
||||||
|
--build-arg TIMESTAMP_UWSGI="$(git log -1 --format="%cd" --date=unix -- dockerfiles/uwsgi.ini)" \
|
||||||
|
-t "${SEARX_IMAGE_NAME}:latest" -t "${SEARX_IMAGE_NAME}:${SEARX_GIT_VERSION}" .
|
||||||
|
|
||||||
|
if [ "$1" = "push" ]; then
|
||||||
|
sudo docker push "${SEARX_IMAGE_NAME}:latest"
|
||||||
|
sudo docker push "${SEARX_IMAGE_NAME}:${SEARX_GIT_VERSION}"
|
||||||
|
fi
|
||||||
|
)
|
||||||
|
dump_return $?
|
||||||
|
}
|
||||||
|
|
||||||
|
# shellcheck disable=SC2119
|
||||||
|
gecko.driver() {
|
||||||
|
pyenv.install
|
||||||
|
|
||||||
|
build_msg INSTALL "gecko.driver"
|
||||||
|
# run installation in a subprocess and activate pyenv
|
||||||
|
( set -e
|
||||||
|
# shellcheck source=/dev/null
|
||||||
|
source "${PY_ENV_BIN}/activate"
|
||||||
|
|
||||||
|
# TODO : check the current geckodriver version
|
||||||
|
geckodriver -V > /dev/null 2>&1 || NOTFOUND=1
|
||||||
|
set +e
|
||||||
|
if [ -z "$NOTFOUND" ]; then
|
||||||
|
build_msg INSTALL "geckodriver already installed"
|
||||||
|
return
|
||||||
|
fi
|
||||||
|
PLATFORM="$(python3 -c 'import platform; print(platform.system().lower(), platform.architecture()[0])')"
|
||||||
|
case "$PLATFORM" in
|
||||||
|
"linux 32bit" | "linux2 32bit") ARCH="linux32";;
|
||||||
|
"linux 64bit" | "linux2 64bit") ARCH="linux64";;
|
||||||
|
"windows 32 bit") ARCH="win32";;
|
||||||
|
"windows 64 bit") ARCH="win64";;
|
||||||
|
"mac 64bit") ARCH="macos";;
|
||||||
|
esac
|
||||||
|
GECKODRIVER_URL="https://github.com/mozilla/geckodriver/releases/download/$GECKODRIVER_VERSION/geckodriver-$GECKODRIVER_VERSION-$ARCH.tar.gz";
|
||||||
|
|
||||||
|
build_msg GECKO "Installing ${PY_ENV_BIN}/geckodriver from $GECKODRIVER_URL"
|
||||||
|
|
||||||
|
FILE="$(mktemp)"
|
||||||
|
wget -qO "$FILE" -- "$GECKODRIVER_URL" && tar xz -C "${PY_ENV_BIN}" -f "$FILE" geckodriver
|
||||||
|
rm -- "$FILE"
|
||||||
|
chmod 755 -- "${PY_ENV_BIN}/geckodriver"
|
||||||
|
)
|
||||||
|
dump_return $?
|
||||||
|
}
|
||||||
|
|
||||||
|
node.env() {
|
||||||
|
local err=0
|
||||||
|
pushd "${REPO_ROOT}" &> /dev/null
|
||||||
|
# shellcheck disable=SC2230
|
||||||
|
which npm &> /dev/null || die 1 'node.env - npm is not found!'
|
||||||
|
|
||||||
|
( set -e
|
||||||
|
# shellcheck disable=SC2030
|
||||||
|
PATH="$(npm bin):$PATH"
|
||||||
|
export PATH
|
||||||
|
|
||||||
|
build_msg INSTALL "npm install $NPM_PACKAGES"
|
||||||
|
# shellcheck disable=SC2086
|
||||||
|
npm install $NPM_PACKAGES
|
||||||
|
|
||||||
|
cd "${REPO_ROOT}/searx/static/themes/oscar"
|
||||||
|
build_msg INSTALL "($(pwd)) npm install"
|
||||||
|
npm install
|
||||||
|
|
||||||
|
build_msg INSTALL "($(pwd)) npm install"
|
||||||
|
cd "${REPO_ROOT}/searx/static/themes/simple"
|
||||||
|
npm install
|
||||||
|
)
|
||||||
|
err=$?
|
||||||
|
popd &> /dev/null
|
||||||
|
dump_return "$err"
|
||||||
|
}
|
||||||
|
|
||||||
|
node.clean() {
|
||||||
|
|
||||||
|
build_msg CLEAN "locally installed npm dependencies"
|
||||||
|
rm -rf \
|
||||||
|
./node_modules \
|
||||||
|
./package-lock.json \
|
||||||
|
./searx/static/themes/oscar/package-lock.json \
|
||||||
|
./searx/static/themes/oscar/node_modules \
|
||||||
|
./searx/static/themes/simple/package-lock.json \
|
||||||
|
./searx/static/themes/simple/node_modules
|
||||||
|
dump_return $?
|
||||||
|
}
|
||||||
|
|
||||||
|
py.build() {
|
||||||
|
build_msg BUILD "[pylint] python package ${PYDIST}"
|
||||||
|
pyenv.cmd python setup.py \
|
||||||
|
sdist -d "${PYDIST}" \
|
||||||
|
bdist_wheel --bdist-dir "${PYBUILD}" -d "${PYDIST}"
|
||||||
|
}
|
||||||
|
|
||||||
|
py.clean() {
|
||||||
|
build_msg CLEAN pyenv
|
||||||
|
( set -e
|
||||||
|
pyenv.drop
|
||||||
|
[ "$VERBOSE" = "1" ] && set -x
|
||||||
|
rm -rf "${PYDIST}" "${PYBUILD}" "${PY_ENV}" ./.tox ./*.egg-info
|
||||||
|
find . -name '*.pyc' -exec rm -f {} +
|
||||||
|
find . -name '*.pyo' -exec rm -f {} +
|
||||||
|
find . -name __pycache__ -exec rm -rf {} +
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
pyenv.check() {
|
||||||
|
cat <<EOF
|
||||||
|
import yaml
|
||||||
|
print('import yaml --> OK')
|
||||||
|
EOF
|
||||||
|
}
|
||||||
|
|
||||||
|
pyenv.install() {
|
||||||
|
|
||||||
|
if ! pyenv.OK; then
|
||||||
|
py.clean > /dev/null
|
||||||
|
fi
|
||||||
|
if pyenv.install.OK > /dev/null; then
|
||||||
|
return 0
|
||||||
|
fi
|
||||||
|
pyenv
|
||||||
|
pyenv.OK || die 42 "error while build pyenv (${PY_ENV_BIN})"
|
||||||
|
|
||||||
|
( set -e
|
||||||
|
build_msg PYENV "[install] pip install -e 'searx${PY_SETUP_EXTRAS}'"
|
||||||
|
"${PY_ENV_BIN}/python" -m pip install -e ".${PY_SETUP_EXTRAS}"
|
||||||
|
buildenv
|
||||||
|
) || die 42 "error while pip install (${PY_ENV_BIN})"
|
||||||
|
}
|
||||||
|
|
||||||
|
pyenv.uninstall() {
|
||||||
|
build_msg PYENV "[pyenv.uninstall] uninstall packages: ${PYOBJECTS}"
|
||||||
|
pyenv.cmd python setup.py develop --uninstall 2>&1 \
|
||||||
|
| prefix_stdout "${_Blue}PYENV ${_creset}[pyenv.uninstall] "
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
pypi.upload() {
|
||||||
|
py.clean
|
||||||
|
py.build
|
||||||
|
# https://github.com/pypa/twine
|
||||||
|
pyenv.cmd twine upload "${PYDIST}"/*
|
||||||
|
}
|
||||||
|
|
||||||
|
pypi.upload.test() {
|
||||||
|
py.clean
|
||||||
|
py.build
|
||||||
|
pyenv.cmd twine upload -r testpypi "${PYDIST}"/*
|
||||||
|
}
|
||||||
|
|
||||||
|
test.pylint() {
|
||||||
|
# shellcheck disable=SC2086
|
||||||
|
( set -e
|
||||||
|
build_msg TEST "[pylint] \$PYLINT_FILES"
|
||||||
|
pyenv.cmd python ${PYLINT_OPTIONS} ${PYLINT_VERBOSE} \
|
||||||
|
"${PYLINT_FILES[@]}"
|
||||||
|
|
||||||
|
build_msg TEST "[pylint] searx/engines"
|
||||||
|
pyenv.cmd python ${PYLINT_OPTIONS} ${PYLINT_VERBOSE} \
|
||||||
|
--disable="${PYLINT_SEARX_DISABLE_OPTION}" \
|
||||||
|
--additional-builtins="${PYLINT_ADDITIONAL_BUILTINS_FOR_ENGINES}" \
|
||||||
|
searx/engines
|
||||||
|
|
||||||
|
build_msg TEST "[pylint] searx tests"
|
||||||
|
pyenv.cmd python ${PYLINT_OPTIONS} ${PYLINT_VERBOSE} \
|
||||||
|
--disable="${PYLINT_SEARX_DISABLE_OPTION}" \
|
||||||
|
--ignore=searx/engines \
|
||||||
|
searx tests
|
||||||
|
)
|
||||||
|
dump_return $?
|
||||||
|
}
|
||||||
|
|
||||||
|
test.pep8() {
|
||||||
|
build_msg TEST 'pycodestyle (formerly pep8)'
|
||||||
|
local _exclude=""
|
||||||
|
printf -v _exclude '%s, ' "${PYLINT_FILES[@]}"
|
||||||
|
pyenv.cmd pycodestyle \
|
||||||
|
--exclude="searx/static, searx/languages.py, $_exclude " \
|
||||||
|
--max-line-length=120 \
|
||||||
|
--ignore "E117,E252,E402,E722,E741,W503,W504,W605" \
|
||||||
|
searx tests
|
||||||
|
dump_return $?
|
||||||
|
}
|
||||||
|
|
||||||
|
test.unit() {
|
||||||
|
build_msg TEST 'tests/unit'
|
||||||
|
pyenv.cmd python -m nose2 -s tests/unit
|
||||||
|
dump_return $?
|
||||||
|
}
|
||||||
|
|
||||||
|
test.coverage() {
|
||||||
|
build_msg TEST 'unit test coverage'
|
||||||
|
( set -e
|
||||||
|
pyenv.cmd python -m nose2 -C --log-capture --with-coverage --coverage searx -s tests/unit
|
||||||
|
pyenv.cmd coverage report
|
||||||
|
pyenv.cmd coverage html
|
||||||
|
)
|
||||||
|
dump_return $?
|
||||||
|
}
|
||||||
|
|
||||||
|
test.robot() {
|
||||||
|
build_msg TEST 'robot'
|
||||||
|
gecko.driver
|
||||||
|
PYTHONPATH=. pyenv.cmd python searx/testing.py robot
|
||||||
|
dump_return $?
|
||||||
|
}
|
||||||
|
|
||||||
|
test.clean() {
|
||||||
|
build_msg CLEAN "test stuff"
|
||||||
|
rm -rf geckodriver.log .coverage coverage/
|
||||||
|
dump_return $?
|
||||||
|
}
|
||||||
|
|
||||||
|
themes.all() {
|
||||||
|
( set -e
|
||||||
|
node.env
|
||||||
|
themes.oscar
|
||||||
|
themes.simple
|
||||||
|
)
|
||||||
|
dump_return $?
|
||||||
|
}
|
||||||
|
|
||||||
|
themes.oscar() {
|
||||||
|
local gruntfile=searx/static/themes/oscar/gruntfile.js
|
||||||
|
build_msg GRUNT "${gruntfile}"
|
||||||
|
PATH="$(npm bin):$PATH" grunt --gruntfile "${gruntfile}"
|
||||||
|
dump_return $?
|
||||||
|
}
|
||||||
|
|
||||||
|
themes.simple() {
|
||||||
|
local gruntfile=searx/static/themes/simple/gruntfile.js
|
||||||
|
build_msg GRUNT "${gruntfile}"
|
||||||
|
PATH="$(npm bin):$PATH" grunt --gruntfile "${gruntfile}"
|
||||||
|
dump_return $?
|
||||||
|
}
|
||||||
|
|
||||||
|
# shellcheck disable=SC2119
|
||||||
|
main() {
|
||||||
|
|
||||||
|
local _type
|
||||||
|
local cmd="$1"; shift
|
||||||
|
|
||||||
|
if [ "$cmd" == "" ]; then
|
||||||
|
help
|
||||||
|
err_msg "missing command"
|
||||||
|
return 42
|
||||||
|
fi
|
||||||
|
|
||||||
|
case "$cmd" in
|
||||||
|
--getenv) var="$1"; echo "${!var}";;
|
||||||
|
--help) help;;
|
||||||
|
--*)
|
||||||
|
help
|
||||||
|
err_msg "unknown option $cmd"
|
||||||
|
return 42
|
||||||
|
;;
|
||||||
|
*)
|
||||||
|
_type="$(type -t "$cmd")"
|
||||||
|
if [ "$_type" != 'function' ]; then
|
||||||
|
err_msg "unknown command $1 / use --help"
|
||||||
|
return 42
|
||||||
|
else
|
||||||
|
"$cmd" "$@"
|
||||||
|
fi
|
||||||
|
;;
|
||||||
|
esac
|
||||||
|
}
|
||||||
|
|
||||||
|
main "$@"
|
205
manage.sh
205
manage.sh
|
@ -1,205 +0,0 @@
|
||||||
#!/bin/sh
|
|
||||||
|
|
||||||
export LANG=C
|
|
||||||
|
|
||||||
BASE_DIR="$(dirname -- "`readlink -f -- "$0"`")"
|
|
||||||
|
|
||||||
cd -- "$BASE_DIR"
|
|
||||||
set -e
|
|
||||||
|
|
||||||
# subshell
|
|
||||||
PYTHONPATH="$BASE_DIR"
|
|
||||||
SEARX_DIR="$BASE_DIR/searx"
|
|
||||||
ACTION="$1"
|
|
||||||
|
|
||||||
. "${BASE_DIR}/utils/brand.env"
|
|
||||||
|
|
||||||
#
|
|
||||||
# Python
|
|
||||||
#
|
|
||||||
|
|
||||||
update_packages() {
|
|
||||||
pip install --upgrade pip
|
|
||||||
pip install --upgrade setuptools
|
|
||||||
pip install -Ur "$BASE_DIR/requirements.txt"
|
|
||||||
}
|
|
||||||
|
|
||||||
update_dev_packages() {
|
|
||||||
update_packages
|
|
||||||
pip install -Ur "$BASE_DIR/requirements-dev.txt"
|
|
||||||
}
|
|
||||||
|
|
||||||
install_geckodriver() {
|
|
||||||
echo '[!] Checking geckodriver'
|
|
||||||
# TODO : check the current geckodriver version
|
|
||||||
set -e
|
|
||||||
geckodriver -V > /dev/null 2>&1 || NOTFOUND=1
|
|
||||||
set +e
|
|
||||||
if [ -z "$NOTFOUND" ]; then
|
|
||||||
return
|
|
||||||
fi
|
|
||||||
GECKODRIVER_VERSION="v0.28.0"
|
|
||||||
PLATFORM="`python3 -c "import platform; print(platform.system().lower(), platform.architecture()[0])"`"
|
|
||||||
case "$PLATFORM" in
|
|
||||||
"linux 32bit" | "linux2 32bit") ARCH="linux32";;
|
|
||||||
"linux 64bit" | "linux2 64bit") ARCH="linux64";;
|
|
||||||
"windows 32 bit") ARCH="win32";;
|
|
||||||
"windows 64 bit") ARCH="win64";;
|
|
||||||
"mac 64bit") ARCH="macos";;
|
|
||||||
esac
|
|
||||||
GECKODRIVER_URL="https://github.com/mozilla/geckodriver/releases/download/$GECKODRIVER_VERSION/geckodriver-$GECKODRIVER_VERSION-$ARCH.tar.gz";
|
|
||||||
|
|
||||||
if [ -z "$1" ]; then
|
|
||||||
if [ -z "$VIRTUAL_ENV" ]; then
|
|
||||||
printf "geckodriver can't be installed because VIRTUAL_ENV is not set, you should download it from\n %s" "$GECKODRIVER_URL"
|
|
||||||
exit
|
|
||||||
else
|
|
||||||
GECKODRIVER_DIR="$VIRTUAL_ENV/bin"
|
|
||||||
fi
|
|
||||||
else
|
|
||||||
GECKODRIVER_DIR="$1"
|
|
||||||
mkdir -p -- "$GECKODRIVER_DIR"
|
|
||||||
fi
|
|
||||||
|
|
||||||
printf "Installing %s/geckodriver from\n %s" "$GECKODRIVER_DIR" "$GECKODRIVER_URL"
|
|
||||||
|
|
||||||
FILE="`mktemp`"
|
|
||||||
wget -qO "$FILE" -- "$GECKODRIVER_URL" && tar xz -C "$GECKODRIVER_DIR" -f "$FILE" geckodriver
|
|
||||||
rm -- "$FILE"
|
|
||||||
chmod 777 -- "$GECKODRIVER_DIR/geckodriver"
|
|
||||||
}
|
|
||||||
|
|
||||||
locales() {
|
|
||||||
pybabel compile -d "$SEARX_DIR/translations"
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
#
|
|
||||||
# Web
|
|
||||||
#
|
|
||||||
|
|
||||||
npm_path_setup() {
|
|
||||||
which npm || (printf 'Error: npm is not found\n'; exit 1)
|
|
||||||
export PATH="$(npm bin)":$PATH
|
|
||||||
}
|
|
||||||
|
|
||||||
npm_packages() {
|
|
||||||
npm_path_setup
|
|
||||||
|
|
||||||
echo '[!] install NPM packages'
|
|
||||||
cd -- "$BASE_DIR"
|
|
||||||
npm install less@2.7 less-plugin-clean-css grunt-cli
|
|
||||||
|
|
||||||
echo '[!] install NPM packages for oscar theme'
|
|
||||||
cd -- "$BASE_DIR/searx/static/themes/oscar"
|
|
||||||
npm install
|
|
||||||
|
|
||||||
echo '[!] install NPM packages for simple theme'
|
|
||||||
cd -- "$BASE_DIR/searx/static/themes/simple"
|
|
||||||
npm install
|
|
||||||
}
|
|
||||||
|
|
||||||
docker_build() {
|
|
||||||
# Check if it is a git repository
|
|
||||||
if [ ! -d .git ]; then
|
|
||||||
echo "This is not Git repository"
|
|
||||||
exit 1
|
|
||||||
fi
|
|
||||||
|
|
||||||
if [ ! -x "$(which git)" ]; then
|
|
||||||
echo "git is not installed"
|
|
||||||
exit 1
|
|
||||||
fi
|
|
||||||
|
|
||||||
if [ ! git remote get-url origin 2> /dev/null ]; then
|
|
||||||
echo "there is no remote origin"
|
|
||||||
exit 1
|
|
||||||
fi
|
|
||||||
|
|
||||||
# This is a git repository
|
|
||||||
|
|
||||||
# "git describe" to get the Docker version (for example : v0.15.0-89-g0585788e)
|
|
||||||
# awk to remove the "v" and the "g"
|
|
||||||
SEARX_GIT_VERSION=$(git describe --match "v[0-9]*\.[0-9]*\.[0-9]*" HEAD 2>/dev/null | awk -F'-' '{OFS="-"; $1=substr($1, 2); if ($3) { $3=substr($3, 2); } print}')
|
|
||||||
|
|
||||||
# add the suffix "-dirty" if the repository has uncommited change
|
|
||||||
# /!\ HACK for searx/searx: ignore utils/brand.env
|
|
||||||
git update-index -q --refresh
|
|
||||||
if [ ! -z "$(git diff-index --name-only HEAD -- | grep -v 'utils/brand.env')" ]; then
|
|
||||||
SEARX_GIT_VERSION="${SEARX_GIT_VERSION}-dirty"
|
|
||||||
fi
|
|
||||||
|
|
||||||
# Get the last git commit id, will be added to the Searx version (see Dockerfile)
|
|
||||||
VERSION_GITCOMMIT=$(echo $SEARX_GIT_VERSION | cut -d- -f2-4)
|
|
||||||
echo "Last commit : $VERSION_GITCOMMIT"
|
|
||||||
|
|
||||||
# Check consistency between the git tag and the searx/version.py file
|
|
||||||
# /!\ HACK : parse Python file with bash /!\
|
|
||||||
# otherwise it is not possible build the docker image without all Python dependencies ( version.py loads __init__.py )
|
|
||||||
# SEARX_PYTHON_VERSION=$(python3 -c "import six; import searx.version; six.print_(searx.version.VERSION_STRING)")
|
|
||||||
SEARX_PYTHON_VERSION=$(cat searx/version.py | grep "\(VERSION_MAJOR\|VERSION_MINOR\|VERSION_BUILD\) =" | cut -d\= -f2 | sed -e 's/^[[:space:]]*//' | paste -sd "." -)
|
|
||||||
if [ $(echo "$SEARX_GIT_VERSION" | cut -d- -f1) != "$SEARX_PYTHON_VERSION" ]; then
|
|
||||||
echo "Inconsistency between the last git tag and the searx/version.py file"
|
|
||||||
echo "git tag: $SEARX_GIT_VERSION"
|
|
||||||
echo "searx/version.py: $SEARX_PYTHON_VERSION"
|
|
||||||
exit 1
|
|
||||||
fi
|
|
||||||
|
|
||||||
# define the docker image name
|
|
||||||
GITHUB_USER=$(echo "${GIT_URL}" | sed 's/.*github\.com\/\([^\/]*\).*/\1/')
|
|
||||||
SEARX_IMAGE_NAME="${SEARX_IMAGE_NAME:-${GITHUB_USER:-searxng}/searxng}"
|
|
||||||
|
|
||||||
# build Docker image
|
|
||||||
echo "Building image ${SEARX_IMAGE_NAME}:${SEARX_GIT_VERSION}"
|
|
||||||
sudo docker build \
|
|
||||||
--build-arg GIT_URL="${GIT_URL}" \
|
|
||||||
--build-arg SEARX_GIT_VERSION="${SEARX_GIT_VERSION}" \
|
|
||||||
--build-arg VERSION_GITCOMMIT="${VERSION_GITCOMMIT}" \
|
|
||||||
--build-arg LABEL_DATE=$(date -u +"%Y-%m-%dT%H:%M:%SZ") \
|
|
||||||
--build-arg LABEL_VCS_REF=$(git rev-parse HEAD) \
|
|
||||||
--build-arg LABEL_VCS_URL="${GIT_URL}" \
|
|
||||||
--build-arg TIMESTAMP_SETTINGS=$(git log -1 --format="%cd" --date=unix -- searx/settings.yml) \
|
|
||||||
--build-arg TIMESTAMP_UWSGI=$(git log -1 --format="%cd" --date=unix -- dockerfiles/uwsgi.ini) \
|
|
||||||
-t ${SEARX_IMAGE_NAME}:latest -t ${SEARX_IMAGE_NAME}:${SEARX_GIT_VERSION} .
|
|
||||||
|
|
||||||
if [ "$1" = "push" ]; then
|
|
||||||
sudo docker push ${SEARX_IMAGE_NAME}:latest
|
|
||||||
sudo docker push ${SEARX_IMAGE_NAME}:${SEARX_GIT_VERSION}
|
|
||||||
fi
|
|
||||||
}
|
|
||||||
|
|
||||||
#
|
|
||||||
# Help
|
|
||||||
#
|
|
||||||
|
|
||||||
help() {
|
|
||||||
[ -z "$1" ] || printf 'Error: %s\n' "$1"
|
|
||||||
echo "Searx manage.sh help
|
|
||||||
|
|
||||||
Commands
|
|
||||||
========
|
|
||||||
help - This text
|
|
||||||
|
|
||||||
Build requirements
|
|
||||||
------------------
|
|
||||||
update_packages - Check & update production dependency changes
|
|
||||||
update_dev_packages - Check & update development and production dependency changes
|
|
||||||
install_geckodriver - Download & install geckodriver if not already installed (required for robot_tests)
|
|
||||||
npm_packages - Download & install npm dependencies
|
|
||||||
|
|
||||||
Build
|
|
||||||
-----
|
|
||||||
locales - Compile locales
|
|
||||||
|
|
||||||
Environment:
|
|
||||||
GIT_URL: ${GIT_URL}
|
|
||||||
ISSUE_URL: ${ISSUE_URL}
|
|
||||||
SEARX_URL: ${SEARX_URL}
|
|
||||||
DOCS_URL: ${DOCS_URL}
|
|
||||||
PUBLIC_INSTANCES: ${PUBLIC_INSTANCES}
|
|
||||||
"
|
|
||||||
}
|
|
||||||
|
|
||||||
[ "$(command -V "$ACTION" | grep ' function$')" = "" ] \
|
|
||||||
&& help "action not found" \
|
|
||||||
|| "$ACTION" "$2"
|
|
317
utils/lib.sh
317
utils/lib.sh
|
@ -86,7 +86,7 @@ set_terminal_colors() {
|
||||||
_Red='\e[0;31m'
|
_Red='\e[0;31m'
|
||||||
_Green='\e[0;32m'
|
_Green='\e[0;32m'
|
||||||
_Yellow='\e[0;33m'
|
_Yellow='\e[0;33m'
|
||||||
_Blue='\e[0;34m'
|
_Blue='\e[0;94m'
|
||||||
_Violet='\e[0;35m'
|
_Violet='\e[0;35m'
|
||||||
_Cyan='\e[0;36m'
|
_Cyan='\e[0;36m'
|
||||||
|
|
||||||
|
@ -95,12 +95,12 @@ set_terminal_colors() {
|
||||||
_BRed='\e[1;31m'
|
_BRed='\e[1;31m'
|
||||||
_BGreen='\e[1;32m'
|
_BGreen='\e[1;32m'
|
||||||
_BYellow='\e[1;33m'
|
_BYellow='\e[1;33m'
|
||||||
_BBlue='\e[1;34m'
|
_BBlue='\e[1;94m'
|
||||||
_BPurple='\e[1;35m'
|
_BPurple='\e[1;35m'
|
||||||
_BCyan='\e[1;36m'
|
_BCyan='\e[1;36m'
|
||||||
}
|
}
|
||||||
|
|
||||||
if [ ! -p /dev/stdout ]; then
|
if [ ! -p /dev/stdout ] && [ ! "$TERM" = 'dumb' ] && [ ! "$TERM" = 'unknown' ]; then
|
||||||
set_terminal_colors
|
set_terminal_colors
|
||||||
fi
|
fi
|
||||||
|
|
||||||
|
@ -152,6 +152,22 @@ err_msg() { echo -e "${_BRed}ERROR:${_creset} $*" >&2; }
|
||||||
warn_msg() { echo -e "${_BBlue}WARN:${_creset} $*" >&2; }
|
warn_msg() { echo -e "${_BBlue}WARN:${_creset} $*" >&2; }
|
||||||
info_msg() { echo -e "${_BYellow}INFO:${_creset} $*" >&2; }
|
info_msg() { echo -e "${_BYellow}INFO:${_creset} $*" >&2; }
|
||||||
|
|
||||||
|
build_msg() {
|
||||||
|
local tag="$1 "
|
||||||
|
shift
|
||||||
|
echo -e "${_Blue}${tag:0:10}${_creset}$*"
|
||||||
|
}
|
||||||
|
|
||||||
|
dump_return() {
|
||||||
|
|
||||||
|
# Use this as last command in your function to prompt an ERROR message if
|
||||||
|
# the exit code is not zero.
|
||||||
|
|
||||||
|
local err=$1
|
||||||
|
[ "$err" -ne "0" ] && err_msg "${FUNCNAME[1]} exit with error ($err)"
|
||||||
|
return "$err"
|
||||||
|
}
|
||||||
|
|
||||||
clean_stdin() {
|
clean_stdin() {
|
||||||
if [[ $(uname -s) != 'Darwin' ]]; then
|
if [[ $(uname -s) != 'Darwin' ]]; then
|
||||||
while read -r -n1 -t 0.1; do : ; done
|
while read -r -n1 -t 0.1; do : ; done
|
||||||
|
@ -496,6 +512,295 @@ service_is_available() {
|
||||||
return "$exit_val"
|
return "$exit_val"
|
||||||
}
|
}
|
||||||
|
|
||||||
|
# python
|
||||||
|
# ------
|
||||||
|
|
||||||
|
PY="${PY:=3}"
|
||||||
|
PYTHON="${PYTHON:=python$PY}"
|
||||||
|
PY_ENV="${PY_ENV:=local/py${PY}}"
|
||||||
|
PY_ENV_BIN="${PY_ENV}/bin"
|
||||||
|
PY_ENV_REQ="${PY_ENV_REQ:=${REPO_ROOT}/requirements*.txt}"
|
||||||
|
|
||||||
|
# List of python packages (folders) or modules (files) installed by command:
|
||||||
|
# pyenv.install
|
||||||
|
PYOBJECTS="${PYOBJECTS:=.}"
|
||||||
|
|
||||||
|
# folder where the python distribution takes place
|
||||||
|
PYDIST="${PYDIST:=dist}"
|
||||||
|
|
||||||
|
# folder where the intermediate build files take place
|
||||||
|
PYBUILD="${PYBUILD:=build/py${PY}}"
|
||||||
|
|
||||||
|
# https://www.python.org/dev/peps/pep-0508/#extras
|
||||||
|
#PY_SETUP_EXTRAS='[develop,test]'
|
||||||
|
PY_SETUP_EXTRAS="${PY_SETUP_EXTRAS:=[develop,test]}"
|
||||||
|
|
||||||
|
PIP_BOILERPLATE=( pip wheel setuptools )
|
||||||
|
|
||||||
|
# shellcheck disable=SC2120
|
||||||
|
pyenv() {
|
||||||
|
|
||||||
|
# usage: pyenv [vtenv_opts ...]
|
||||||
|
#
|
||||||
|
# vtenv_opts: see 'pip install --help'
|
||||||
|
#
|
||||||
|
# Builds virtualenv with 'requirements*.txt' (PY_ENV_REQ) installed. The
|
||||||
|
# virtualenv will be reused by validating sha256sum of the requirement
|
||||||
|
# files.
|
||||||
|
|
||||||
|
required_commands \
|
||||||
|
sha256sum "${PYTHON}" \
|
||||||
|
|| exit
|
||||||
|
|
||||||
|
local pip_req=()
|
||||||
|
|
||||||
|
if ! pyenv.OK > /dev/null; then
|
||||||
|
rm -f "${PY_ENV}/${PY_ENV_REQ}.sha256"
|
||||||
|
pyenv.drop > /dev/null
|
||||||
|
build_msg PYENV "[virtualenv] installing ${PY_ENV_REQ} into ${PY_ENV}"
|
||||||
|
|
||||||
|
"${PYTHON}" -m venv "$@" "${PY_ENV}"
|
||||||
|
"${PY_ENV_BIN}/python" -m pip install -U "${PIP_BOILERPLATE[@]}"
|
||||||
|
|
||||||
|
for i in ${PY_ENV_REQ}; do
|
||||||
|
pip_req=( "${pip_req[@]}" "-r" "$i" )
|
||||||
|
done
|
||||||
|
|
||||||
|
(
|
||||||
|
[ "$VERBOSE" = "1" ] && set -x
|
||||||
|
# shellcheck disable=SC2086
|
||||||
|
"${PY_ENV_BIN}/python" -m pip install "${pip_req[@]}" \
|
||||||
|
&& sha256sum ${PY_ENV_REQ} > "${PY_ENV}/requirements.sha256"
|
||||||
|
)
|
||||||
|
fi
|
||||||
|
pyenv.OK
|
||||||
|
}
|
||||||
|
|
||||||
|
_pyenv_OK=''
|
||||||
|
pyenv.OK() {
|
||||||
|
|
||||||
|
# probes if pyenv exists and runs the script from pyenv.check
|
||||||
|
|
||||||
|
[ "$_pyenv_OK" == "OK" ] && return 0
|
||||||
|
|
||||||
|
if [ ! -f "${PY_ENV_BIN}/python" ]; then
|
||||||
|
build_msg PYENV "[virtualenv] missing ${PY_ENV_BIN}/python"
|
||||||
|
return 1
|
||||||
|
fi
|
||||||
|
|
||||||
|
if [ ! -f "${PY_ENV}/requirements.sha256" ] \
|
||||||
|
|| ! sha256sum --check --status <"${PY_ENV}/requirements.sha256" 2>/dev/null; then
|
||||||
|
build_msg PYENV "[virtualenv] requirements.sha256 failed"
|
||||||
|
sed 's/^/ [virtualenv] - /' <"${PY_ENV}/requirements.sha256"
|
||||||
|
return 1
|
||||||
|
fi
|
||||||
|
|
||||||
|
pyenv.check \
|
||||||
|
| "${PY_ENV_BIN}/python" 2>&1 \
|
||||||
|
| prefix_stdout "${_Blue}PYENV ${_creset}[check] "
|
||||||
|
|
||||||
|
local err=${PIPESTATUS[1]}
|
||||||
|
if [ "$err" -ne "0" ]; then
|
||||||
|
build_msg PYENV "[check] python test failed"
|
||||||
|
return "$err"
|
||||||
|
fi
|
||||||
|
|
||||||
|
build_msg PYENV "OK"
|
||||||
|
_pyenv_OK="OK"
|
||||||
|
return 0
|
||||||
|
}
|
||||||
|
|
||||||
|
pyenv.drop() {
|
||||||
|
|
||||||
|
build_msg PYENV "[virtualenv] drop ${PY_ENV}"
|
||||||
|
rm -rf "${PY_ENV}"
|
||||||
|
_pyenv_OK=''
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
pyenv.check() {
|
||||||
|
|
||||||
|
# Prompts a python script with additional checks. Used by pyenv.OK to check
|
||||||
|
# if virtualenv is ready to install python objects. This function should be
|
||||||
|
# overwritten by the application script.
|
||||||
|
|
||||||
|
local imp=""
|
||||||
|
|
||||||
|
for i in "${PIP_BOILERPLATE[@]}"; do
|
||||||
|
imp="$imp, $i"
|
||||||
|
done
|
||||||
|
|
||||||
|
cat <<EOF
|
||||||
|
import ${imp#,*}
|
||||||
|
|
||||||
|
EOF
|
||||||
|
}
|
||||||
|
|
||||||
|
pyenv.install() {
|
||||||
|
|
||||||
|
if ! pyenv.OK; then
|
||||||
|
py.clean > /dev/null
|
||||||
|
fi
|
||||||
|
if ! pyenv.install.OK > /dev/null; then
|
||||||
|
build_msg PYENV "[install] ${PYOBJECTS}"
|
||||||
|
if ! pyenv.OK >/dev/null; then
|
||||||
|
pyenv
|
||||||
|
fi
|
||||||
|
for i in ${PYOBJECTS}; do
|
||||||
|
build_msg PYENV "[install] pip install -e '$i${PY_SETUP_EXTRAS}'"
|
||||||
|
"${PY_ENV_BIN}/python" -m pip install -e "$i${PY_SETUP_EXTRAS}"
|
||||||
|
done
|
||||||
|
fi
|
||||||
|
pyenv.install.OK
|
||||||
|
}
|
||||||
|
|
||||||
|
_pyenv_install_OK=''
|
||||||
|
pyenv.install.OK() {
|
||||||
|
|
||||||
|
[ "$_pyenv_install_OK" == "OK" ] && return 0
|
||||||
|
|
||||||
|
local imp=""
|
||||||
|
local err=""
|
||||||
|
|
||||||
|
if [ "." = "${PYOBJECTS}" ]; then
|
||||||
|
imp="import $(basename "$(pwd)")"
|
||||||
|
else
|
||||||
|
# shellcheck disable=SC2086
|
||||||
|
for i in ${PYOBJECTS}; do imp="$imp, $i"; done
|
||||||
|
imp="import ${imp#,*} "
|
||||||
|
fi
|
||||||
|
(
|
||||||
|
[ "$VERBOSE" = "1" ] && set -x
|
||||||
|
"${PY_ENV_BIN}/python" -c "import sys; sys.path.pop(0); $imp;" 2>/dev/null
|
||||||
|
)
|
||||||
|
|
||||||
|
err=$?
|
||||||
|
if [ "$err" -ne "0" ]; then
|
||||||
|
build_msg PYENV "[install] python installation test failed"
|
||||||
|
return "$err"
|
||||||
|
fi
|
||||||
|
|
||||||
|
build_msg PYENV "[install] OK"
|
||||||
|
_pyenv_install_OK="OK"
|
||||||
|
return 0
|
||||||
|
}
|
||||||
|
|
||||||
|
pyenv.uninstall() {
|
||||||
|
|
||||||
|
build_msg PYENV "[uninstall] ${PYOBJECTS}"
|
||||||
|
|
||||||
|
if [ "." = "${PYOBJECTS}" ]; then
|
||||||
|
pyenv.cmd python setup.py develop --uninstall 2>&1 \
|
||||||
|
| prefix_stdout "${_Blue}PYENV ${_creset}[pyenv.uninstall] "
|
||||||
|
else
|
||||||
|
pyenv.cmd python -m pip uninstall --yes ${PYOBJECTS} 2>&1 \
|
||||||
|
| prefix_stdout "${_Blue}PYENV ${_creset}[pyenv.uninstall] "
|
||||||
|
fi
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
pyenv.cmd() {
|
||||||
|
pyenv.install
|
||||||
|
( set -e
|
||||||
|
# shellcheck source=/dev/null
|
||||||
|
source "${PY_ENV_BIN}/activate"
|
||||||
|
[ "$VERBOSE" = "1" ] && set -x
|
||||||
|
"$@"
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
# Sphinx doc
|
||||||
|
# ----------
|
||||||
|
|
||||||
|
GH_PAGES="build/gh-pages"
|
||||||
|
DOCS_DIST="${DOCS_DIST:=dist/docs}"
|
||||||
|
DOCS_BUILD="${DOCS_BUILD:=build/docs}"
|
||||||
|
|
||||||
|
docs.html() {
|
||||||
|
build_msg SPHINX "HTML ./docs --> file://$(readlink -e "$(pwd)/$DOCS_DIST")"
|
||||||
|
pyenv.install
|
||||||
|
docs.prebuild
|
||||||
|
# shellcheck disable=SC2086
|
||||||
|
PATH="${PY_ENV_BIN}:${PATH}" pyenv.cmd sphinx-build \
|
||||||
|
${SPHINX_VERBOSE} ${SPHINXOPTS} \
|
||||||
|
-b html -c ./docs -d "${DOCS_BUILD}/.doctrees" ./docs "${DOCS_DIST}"
|
||||||
|
dump_return $?
|
||||||
|
}
|
||||||
|
|
||||||
|
docs.live() {
|
||||||
|
build_msg SPHINX "autobuild ./docs --> file://$(readlink -e "$(pwd)/$DOCS_DIST")"
|
||||||
|
pyenv.install
|
||||||
|
docs.prebuild
|
||||||
|
# shellcheck disable=SC2086
|
||||||
|
PATH="${PY_ENV_BIN}:${PATH}" pyenv.cmd sphinx-autobuild \
|
||||||
|
${SPHINX_VERBOSE} ${SPHINXOPTS} --open-browser --host 0.0.0.0 \
|
||||||
|
-b html -c ./docs -d "${DOCS_BUILD}/.doctrees" ./docs "${DOCS_DIST}"
|
||||||
|
dump_return $?
|
||||||
|
}
|
||||||
|
|
||||||
|
docs.clean() {
|
||||||
|
build_msg CLEAN "docs -- ${DOCS_BUILD} ${DOCS_DIST}"
|
||||||
|
# shellcheck disable=SC2115
|
||||||
|
rm -rf "${GH_PAGES}" "${DOCS_BUILD}" "${DOCS_DIST}"
|
||||||
|
dump_return $?
|
||||||
|
}
|
||||||
|
|
||||||
|
docs.prebuild() {
|
||||||
|
# Dummy function to run some actions before sphinx-doc build gets started.
|
||||||
|
# This finction needs to be overwritten by the application script.
|
||||||
|
true
|
||||||
|
dump_return $?
|
||||||
|
}
|
||||||
|
|
||||||
|
# shellcheck disable=SC2155
|
||||||
|
docs.gh-pages() {
|
||||||
|
|
||||||
|
# The commit history in the gh-pages branch makes no sense, the history only
|
||||||
|
# inflates the repository unnecessarily. Therefore a *new orphan* branch
|
||||||
|
# is created each time we deploy on the gh-pages branch.
|
||||||
|
|
||||||
|
docs.clean
|
||||||
|
docs.prebuild
|
||||||
|
docs.html
|
||||||
|
|
||||||
|
[ "$VERBOSE" = "1" ] && set -x
|
||||||
|
local head="$(git rev-parse HEAD)"
|
||||||
|
local branch="$(git name-rev --name-only HEAD)"
|
||||||
|
local remote="$(git config branch."${branch}".remote)"
|
||||||
|
local remote_url="$(git config remote."${remote}".url)"
|
||||||
|
|
||||||
|
build_msg GH-PAGES "prepare folder: ${GH_PAGES}"
|
||||||
|
build_msg GH-PAGES "remote of the gh-pages branch: ${remote} / ${remote_url}"
|
||||||
|
build_msg GH-PAGES "current branch: ${branch}"
|
||||||
|
|
||||||
|
# prepare the *orphan* gh-pages working tree
|
||||||
|
(
|
||||||
|
git worktree remove -f "${GH_PAGES}"
|
||||||
|
git branch -D gh-pages
|
||||||
|
) &> /dev/null || true
|
||||||
|
git worktree add --no-checkout "${GH_PAGES}" "${remote}/master"
|
||||||
|
|
||||||
|
pushd "${GH_PAGES}" &> /dev/null
|
||||||
|
git checkout --orphan gh-pages
|
||||||
|
git rm -rfq .
|
||||||
|
popd &> /dev/null
|
||||||
|
|
||||||
|
cp -r "${DOCS_DIST}"/* "${GH_PAGES}"/
|
||||||
|
touch "${GH_PAGES}/.nojekyll"
|
||||||
|
cat > "${GH_PAGES}/404.html" <<EOF
|
||||||
|
<html><head><META http-equiv='refresh' content='0;URL=index.html'></head></html>
|
||||||
|
EOF
|
||||||
|
|
||||||
|
pushd "${GH_PAGES}" &> /dev/null
|
||||||
|
git add --all .
|
||||||
|
git commit -q -m "gh-pages build from: ${branch}@${head} (${remote_url})"
|
||||||
|
git push -f "${remote}" gh-pages
|
||||||
|
popd &> /dev/null
|
||||||
|
|
||||||
|
set +x
|
||||||
|
build_msg GH-PAGES "deployed"
|
||||||
|
}
|
||||||
|
|
||||||
# golang
|
# golang
|
||||||
# ------
|
# ------
|
||||||
|
|
||||||
|
@ -1382,6 +1687,12 @@ LXC_ENV_FOLDER=
|
||||||
if in_container; then
|
if in_container; then
|
||||||
# shellcheck disable=SC2034
|
# shellcheck disable=SC2034
|
||||||
LXC_ENV_FOLDER="lxc-env/$(hostname)/"
|
LXC_ENV_FOLDER="lxc-env/$(hostname)/"
|
||||||
|
PY_ENV="${LXC_ENV_FOLDER}${PY_ENV}"
|
||||||
|
PY_ENV_BIN="${LXC_ENV_FOLDER}${PY_ENV_BIN}"
|
||||||
|
PYDIST="${LXC_ENV_FOLDER}${PYDIST}"
|
||||||
|
PYBUILD="${LXC_ENV_FOLDER}${PYBUILD}"
|
||||||
|
DOCS_DIST="${LXC_ENV_FOLDER}${DOCS_DIST}"
|
||||||
|
DOCS_BUILD="${LXC_ENV_FOLDER}${DOCS_BUILD}"
|
||||||
fi
|
fi
|
||||||
|
|
||||||
lxc_init_container_env() {
|
lxc_init_container_env() {
|
||||||
|
|
|
@ -142,7 +142,7 @@ main() {
|
||||||
local _usage="unknown or missing $1 command $2"
|
local _usage="unknown or missing $1 command $2"
|
||||||
|
|
||||||
# don't check prerequisite when in recursion
|
# don't check prerequisite when in recursion
|
||||||
if [[ ! $1 == __* ]]; then
|
if [[ ! $1 == __* ]] && [[ ! $1 == --help ]]; then
|
||||||
if ! in_container; then
|
if ! in_container; then
|
||||||
! required_commands lxc && lxd_info && exit 42
|
! required_commands lxc && lxd_info && exit 42
|
||||||
fi
|
fi
|
||||||
|
|
|
@ -1,4 +1,5 @@
|
||||||
# -*- coding: utf-8; mode: makefile-gmake -*-
|
# -*- coding: utf-8; mode: makefile-gmake -*-
|
||||||
|
# SPDX-License-Identifier: AGPL-3.0-or-later
|
||||||
|
|
||||||
ifeq (,$(wildcard /.lxcenv.mk))
|
ifeq (,$(wildcard /.lxcenv.mk))
|
||||||
PHONY += lxc-activate lxc-purge
|
PHONY += lxc-activate lxc-purge
|
||||||
|
@ -10,60 +11,26 @@ else
|
||||||
include /.lxcenv.mk
|
include /.lxcenv.mk
|
||||||
endif
|
endif
|
||||||
|
|
||||||
|
PHONY += make-help
|
||||||
ifeq (,$(wildcard /.lxcenv.mk))
|
ifeq (,$(wildcard /.lxcenv.mk))
|
||||||
make-help:
|
make-help:
|
||||||
else
|
else
|
||||||
make-help: lxc-help
|
make-help: lxc-help
|
||||||
endif
|
endif
|
||||||
|
@echo 'options:'
|
||||||
@echo ' make V=0|1 [targets] 0 => quiet build (default), 1 => verbose build'
|
@echo ' make V=0|1 [targets] 0 => quiet build (default), 1 => verbose build'
|
||||||
@echo ' make V=2 [targets] 2 => give reason for rebuild of target'
|
@echo ' make V=2 [targets] 2 => give reason for rebuild of target'
|
||||||
|
|
||||||
quiet_cmd_common_clean = CLEAN $@
|
|
||||||
cmd_common_clean = \
|
|
||||||
find . -name '*.orig' -exec rm -f {} + ;\
|
|
||||||
find . -name '*.rej' -exec rm -f {} + ;\
|
|
||||||
find . -name '*~' -exec rm -f {} + ;\
|
|
||||||
find . -name '*.bak' -exec rm -f {} + ;\
|
|
||||||
|
|
||||||
FMT = cat
|
|
||||||
ifeq ($(shell which fmt >/dev/null 2>&1; echo $$?), 0)
|
|
||||||
FMT = fmt
|
|
||||||
endif
|
|
||||||
|
|
||||||
# MS-Windows
|
|
||||||
#
|
|
||||||
# For a minimal *make-environment*, I'am using the gnu-tools from:
|
|
||||||
#
|
|
||||||
# - GNU MCU Eclipse Windows Build Tools, which brings 'make', 'rm' etc.
|
|
||||||
# https://github.com/gnu-mcu-eclipse/windows-build-tools/releases
|
|
||||||
#
|
|
||||||
# - git for Windows, which brings 'find', 'grep' etc.
|
|
||||||
# https://git-scm.com/download/win
|
|
||||||
|
|
||||||
|
|
||||||
# normpath
|
|
||||||
#
|
|
||||||
# System-dependent normalization of the path name
|
|
||||||
#
|
|
||||||
# usage: $(call normpath,/path/to/file)
|
|
||||||
|
|
||||||
normpath = $1
|
|
||||||
ifeq ($(OS),Windows_NT)
|
|
||||||
normpath = $(subst /,\,$1)
|
|
||||||
endif
|
|
||||||
|
|
||||||
|
|
||||||
# stolen from linux/Makefile
|
|
||||||
#
|
|
||||||
|
|
||||||
ifeq ("$(origin V)", "command line")
|
ifeq ("$(origin V)", "command line")
|
||||||
KBUILD_VERBOSE = $(V)
|
VERBOSE = $(V)
|
||||||
endif
|
endif
|
||||||
ifndef KBUILD_VERBOSE
|
ifndef VERBOSE
|
||||||
KBUILD_VERBOSE = 0
|
VERBOSE = 0
|
||||||
endif
|
endif
|
||||||
|
|
||||||
ifeq ($(KBUILD_VERBOSE),1)
|
export VERBOSE
|
||||||
|
|
||||||
|
ifeq ($(VERBOSE),1)
|
||||||
quiet =
|
quiet =
|
||||||
Q =
|
Q =
|
||||||
else
|
else
|
||||||
|
@ -75,14 +42,8 @@ endif
|
||||||
#
|
#
|
||||||
|
|
||||||
# Convenient variables
|
# Convenient variables
|
||||||
comma := ,
|
|
||||||
quote := "
|
|
||||||
#" this comment is only for emacs highlighting
|
|
||||||
squote := '
|
squote := '
|
||||||
#' this comment is only for emacs highlighting
|
#' this comment is only for emacs highlighting
|
||||||
empty :=
|
|
||||||
space := $(empty) $(empty)
|
|
||||||
space_escape := _-_SPACE_-_
|
|
||||||
|
|
||||||
# Find any prerequisites that is newer than target or that does not exist.
|
# Find any prerequisites that is newer than target or that does not exist.
|
||||||
# PHONY targets skipped in both cases.
|
# PHONY targets skipped in both cases.
|
||||||
|
@ -107,7 +68,7 @@ any-prereq = $(filter-out $(PHONY),$?) $(filter-out $(PHONY) $(wildcard $^),$^)
|
||||||
# (5) No dir/.target.cmd file (used to store command line)
|
# (5) No dir/.target.cmd file (used to store command line)
|
||||||
# (6) No dir/.target.cmd file and target not listed in $(targets)
|
# (6) No dir/.target.cmd file and target not listed in $(targets)
|
||||||
# This is a good hint that there is a bug in the kbuild file
|
# This is a good hint that there is a bug in the kbuild file
|
||||||
ifeq ($(KBUILD_VERBOSE),2)
|
ifeq ($(VERBOSE),2)
|
||||||
why = \
|
why = \
|
||||||
$(if $(filter $@, $(PHONY)),- due to target is PHONY, \
|
$(if $(filter $@, $(PHONY)),- due to target is PHONY, \
|
||||||
$(if $(wildcard $@), \
|
$(if $(wildcard $@), \
|
||||||
|
|
|
@ -1,269 +0,0 @@
|
||||||
# -*- coding: utf-8; mode: makefile-gmake -*-
|
|
||||||
|
|
||||||
# list of python packages (folders) or modules (files) of this build
|
|
||||||
PYOBJECTS ?=
|
|
||||||
|
|
||||||
SITE_PYTHON ?=$(dir $(abspath $(lastword $(MAKEFILE_LIST))))site-python
|
|
||||||
export PYTHONPATH := $(SITE_PYTHON):$$PYTHONPATH
|
|
||||||
export PY_ENV PYDIST PYBUILD
|
|
||||||
|
|
||||||
# folder where the python distribution takes place
|
|
||||||
PYDIST = ./$(LXC_ENV_FOLDER)dist
|
|
||||||
# folder where the python intermediate build files take place
|
|
||||||
PYBUILD = ./$(LXC_ENV_FOLDER)build
|
|
||||||
# python version to use
|
|
||||||
PY ?=3
|
|
||||||
# $(PYTHON) points to the python interpreter from the OS! The python from the
|
|
||||||
# OS is needed e.g. to create a virtualenv. For tasks inside the virtualenv the
|
|
||||||
# interpeter from '$(PY_ENV_BIN)/python' is used.
|
|
||||||
PYTHON ?= python$(PY)
|
|
||||||
PIP ?= pip$(PY)
|
|
||||||
PIP_INST ?= --user
|
|
||||||
|
|
||||||
# https://www.python.org/dev/peps/pep-0508/#extras
|
|
||||||
#PY_SETUP_EXTRAS ?= \[develop,test\]
|
|
||||||
PY_SETUP_EXTRAS ?=
|
|
||||||
|
|
||||||
PYDEBUG ?= --pdb
|
|
||||||
PYLINT_RC ?= .pylintrc
|
|
||||||
|
|
||||||
TEST_FOLDER ?= ./tests
|
|
||||||
TEST ?= .
|
|
||||||
|
|
||||||
PY_ENV = ./$(LXC_ENV_FOLDER)local/py$(PY)
|
|
||||||
PY_ENV_BIN = $(PY_ENV)/bin
|
|
||||||
PY_ENV_ACT = . $(PY_ENV_BIN)/activate
|
|
||||||
|
|
||||||
ifeq ($(OS),Windows_NT)
|
|
||||||
PYTHON = python
|
|
||||||
PY_ENV_BIN = $(PY_ENV)/Scripts
|
|
||||||
PY_ENV_ACT = $(PY_ENV_BIN)/activate
|
|
||||||
endif
|
|
||||||
|
|
||||||
VTENV_OPTS ?=
|
|
||||||
|
|
||||||
python-help::
|
|
||||||
@echo 'makefile.python:'
|
|
||||||
@echo ' pyenv | pyenv[un]install'
|
|
||||||
@echo ' build $(PY_ENV) & [un]install python objects'
|
|
||||||
@echo ' targts using pyenv $(PY_ENV):'
|
|
||||||
@echo ' pylint - run pylint *linting*'
|
|
||||||
@echo ' pytest - run *tox* test on python objects'
|
|
||||||
@echo ' pydebug - run tests within a PDB debug session'
|
|
||||||
@echo ' pybuild - build python packages ($(PYDIST) $(PYBUILD))'
|
|
||||||
@echo ' pyclean - clean intermediate python objects'
|
|
||||||
@echo ' targets using system users environment:'
|
|
||||||
@echo ' py[un]install - [un]install python objects in editable mode'
|
|
||||||
@echo ' upload-pypi - upload $(PYDIST)/* files to PyPi'
|
|
||||||
@echo 'options:'
|
|
||||||
@echo ' make PY=3.7 [targets] => to eval targets with python 3.7 ($(PY))'
|
|
||||||
@echo ' make PIP_INST= => to set/unset pip install options ($(PIP_INST))'
|
|
||||||
@echo ' make TEST=. => choose test from $(TEST_FOLDER) (default "." runs all)'
|
|
||||||
@echo ' make DEBUG= => target "debug": do not invoke PDB on errors'
|
|
||||||
@echo ' make PY_SETUP_EXTRAS => also install extras_require from setup.py \[develop,test\]'
|
|
||||||
@echo ' when using target "pydebug", set breakpoints within py-source by adding::'
|
|
||||||
@echo ' DEBUG()'
|
|
||||||
|
|
||||||
# ------------------------------------------------------------------------------
|
|
||||||
# OS requirements
|
|
||||||
# ------------------------------------------------------------------------------
|
|
||||||
|
|
||||||
PHONY += msg-python-exe python-exe
|
|
||||||
msg-python-exe:
|
|
||||||
@echo "\n $(PYTHON) is required\n\n\
|
|
||||||
Make sure you have $(PYTHON) installed, grab it from\n\
|
|
||||||
https://www.python.org or install it from your package\n\
|
|
||||||
manager. On debian based OS these requirements are\n\
|
|
||||||
installed by::\n\n\
|
|
||||||
sudo -H add-apt-repository ppa:deadsnakes/ppa\n\
|
|
||||||
sudo -H apt update\n\
|
|
||||||
sudo -H apt-get install $(PYTHON) $(PYTHON)-venv\n"
|
|
||||||
|
|
||||||
ifeq ($(shell which $(PYTHON) >/dev/null 2>&1; echo $$?), 1)
|
|
||||||
python-exe: msg-python-exe
|
|
||||||
$(error The '$(PYTHON)' command was not found)
|
|
||||||
else
|
|
||||||
python-exe:
|
|
||||||
@:
|
|
||||||
endif
|
|
||||||
|
|
||||||
msg-pip-exe:
|
|
||||||
@echo "\n $(PIP) is required\n\n\
|
|
||||||
Make sure you have updated pip installed, grab it from\n\
|
|
||||||
https://pip.pypa.io or install it from your package\n\
|
|
||||||
manager. On debian based OS these requirements are\n\
|
|
||||||
installed by::\n\n\
|
|
||||||
sudo -H apt-get install python$(PY)-pip\n" | $(FMT)
|
|
||||||
|
|
||||||
ifeq ($(shell which $(PIP) >/dev/null 2>&1; echo $$?), 1)
|
|
||||||
pip-exe: msg-pip-exe
|
|
||||||
$(error The '$(PIP)' command was not found)
|
|
||||||
else
|
|
||||||
pip-exe:
|
|
||||||
@:
|
|
||||||
endif
|
|
||||||
|
|
||||||
# ------------------------------------------------------------------------------
|
|
||||||
# commands
|
|
||||||
# ------------------------------------------------------------------------------
|
|
||||||
|
|
||||||
# $2 path to folder with setup.py, this uses pip from the OS
|
|
||||||
quiet_cmd_pyinstall = INSTALL $2
|
|
||||||
cmd_pyinstall = $(PIP) $(PIP_VERBOSE) install $(PIP_INST) -e $2$(PY_SETUP_EXTRAS)
|
|
||||||
|
|
||||||
# $2 path to folder with setup.py, this uses pip from pyenv (not OS!)
|
|
||||||
quiet_cmd_pyenvinstall = PYENV install $2
|
|
||||||
cmd_pyenvinstall = \
|
|
||||||
if ! cat $(PY_ENV)/requirements.sha256 2>/dev/null | sha256sum --check --status 2>/dev/null; then \
|
|
||||||
rm -f $(PY_ENV)/requirements.sha256; \
|
|
||||||
$(PY_ENV_BIN)/python -m pip $(PIP_VERBOSE) install -e $2$(PY_SETUP_EXTRAS) &&\
|
|
||||||
sha256sum requirements*.txt > $(PY_ENV)/requirements.sha256 ;\
|
|
||||||
else \
|
|
||||||
echo "PYENV $2 already installed"; \
|
|
||||||
fi
|
|
||||||
|
|
||||||
# Uninstall the package. Since pip does not uninstall the no longer needed
|
|
||||||
# depencies (something like autoremove) the depencies remain.
|
|
||||||
|
|
||||||
# $2 package name to uninstall, this uses pip from the OS.
|
|
||||||
quiet_cmd_pyuninstall = UNINSTALL $2
|
|
||||||
cmd_pyuninstall = $(PIP) $(PIP_VERBOSE) uninstall --yes $2
|
|
||||||
|
|
||||||
# $2 path to folder with setup.py, this uses pip from pyenv (not OS!)
|
|
||||||
quiet_cmd_pyenvuninstall = PYENV uninstall $2
|
|
||||||
cmd_pyenvuninstall = $(PY_ENV_BIN)/python -m pip $(PIP_VERBOSE) uninstall --yes $2
|
|
||||||
|
|
||||||
# $2 path to folder where virtualenv take place
|
|
||||||
quiet_cmd_virtualenv = PYENV usage: $ source ./$@/bin/activate
|
|
||||||
cmd_virtualenv = \
|
|
||||||
if [ -d "./$(PY_ENV)" -a -x "./$(PY_ENV_BIN)/python" ]; then \
|
|
||||||
echo "PYENV using virtualenv from $2"; \
|
|
||||||
else \
|
|
||||||
$(PYTHON) -m venv $(VTENV_OPTS) $2; \
|
|
||||||
$(PY_ENV_BIN)/python -m pip install $(PIP_VERBOSE) -U pip wheel setuptools; \
|
|
||||||
$(PY_ENV_BIN)/python -m pip install $(PIP_VERBOSE) -r requirements.txt; \
|
|
||||||
fi
|
|
||||||
|
|
||||||
# $2 path to lint
|
|
||||||
quiet_cmd_pylint = LINT $@
|
|
||||||
cmd_pylint = $(PY_ENV_BIN)/python -m pylint -j 0 --rcfile $(PYLINT_RC) $2
|
|
||||||
|
|
||||||
quiet_cmd_pytest = TEST $@
|
|
||||||
cmd_pytest = $(PY_ENV_BIN)/python -m tox -vv
|
|
||||||
|
|
||||||
# setuptools, pip, easy_install its a mess full of cracks, a documentation hell
|
|
||||||
# and broken by design ... all sucks, I really, really hate all this ... aaargh!
|
|
||||||
#
|
|
||||||
# About python packaging see `Python Packaging Authority`_. Most of the names
|
|
||||||
# here are mapped to ``setup(<name1>=..., <name2>=...)`` arguments in
|
|
||||||
# ``setup.py``. See `Packaging and distributing projects`_ about ``setup(...)``
|
|
||||||
# arguments. If this is all new for you, start with `PyPI Quick and Dirty`_.
|
|
||||||
#
|
|
||||||
# Further read:
|
|
||||||
#
|
|
||||||
# - pythonwheels_
|
|
||||||
# - setuptools_
|
|
||||||
# - packaging_
|
|
||||||
# - sdist_
|
|
||||||
# - installing_
|
|
||||||
#
|
|
||||||
# .. _`Python Packaging Authority`: https://www.pypa.io
|
|
||||||
# .. _`Packaging and distributing projects`: https://packaging.python.org/guides/distributing-packages-using-setuptools/
|
|
||||||
# .. _`PyPI Quick and Dirty`: https://hynek.me/articles/sharing-your-labor-of-love-pypi-quick-and-dirty/
|
|
||||||
# .. _pythonwheels: https://pythonwheels.com/
|
|
||||||
# .. _setuptools: https://setuptools.readthedocs.io/en/latest/setuptools.html
|
|
||||||
# .. _packaging: https://packaging.python.org/guides/distributing-packages-using-setuptools/#packaging-and-distributing-projects
|
|
||||||
# .. _sdist: https://packaging.python.org/guides/distributing-packages-using-setuptools/#source-distributions
|
|
||||||
# .. _bdist_wheel: https://packaging.python.org/guides/distributing-packages-using-setuptools/#pure-python-wheels
|
|
||||||
# .. _installing: https://packaging.python.org/tutorials/installing-packages/
|
|
||||||
#
|
|
||||||
quiet_cmd_pybuild = BUILD $@
|
|
||||||
cmd_pybuild = $(PY_ENV_BIN)/python setup.py \
|
|
||||||
sdist -d $(PYDIST) \
|
|
||||||
bdist_wheel --bdist-dir $(PYBUILD) -d $(PYDIST)
|
|
||||||
|
|
||||||
quiet_cmd_pyclean = CLEAN $@
|
|
||||||
# remove 'build' folder since bdist_wheel does not care the --bdist-dir
|
|
||||||
cmd_pyclean = \
|
|
||||||
rm -rf $(PYDIST) $(PYBUILD) $(PY_ENV) ./.tox *.egg-info ;\
|
|
||||||
find . -name '*.pyc' -exec rm -f {} + ;\
|
|
||||||
find . -name '*.pyo' -exec rm -f {} + ;\
|
|
||||||
find . -name __pycache__ -exec rm -rf {} +
|
|
||||||
|
|
||||||
# ------------------------------------------------------------------------------
|
|
||||||
# targets
|
|
||||||
# ------------------------------------------------------------------------------
|
|
||||||
|
|
||||||
# for installation use the pip from the OS!
|
|
||||||
PHONY += pyinstall
|
|
||||||
pyinstall: pip-exe
|
|
||||||
$(call cmd,pyinstall,.)
|
|
||||||
|
|
||||||
PHONY += pyuninstall
|
|
||||||
pyuninstall: pip-exe
|
|
||||||
$(call cmd,pyuninstall,$(PYOBJECTS))
|
|
||||||
|
|
||||||
# for installation use the pip from PY_ENV (not the OS)!
|
|
||||||
PHONY += pyenvinstall
|
|
||||||
pyenvinstall: $(PY_ENV)
|
|
||||||
$(call cmd,pyenvinstall,.)
|
|
||||||
|
|
||||||
PHONY += pyenvuninstall
|
|
||||||
pyenvuninstall: $(PY_ENV)
|
|
||||||
$(call cmd,pyenvuninstall,$(PYOBJECTS))
|
|
||||||
|
|
||||||
PHONY += pyclean
|
|
||||||
pyclean:
|
|
||||||
$(call cmd,pyclean)
|
|
||||||
|
|
||||||
# to build *local* environment, python from the OS is needed!
|
|
||||||
pyenv: $(PY_ENV)
|
|
||||||
$(PY_ENV): python-exe
|
|
||||||
$(call cmd,virtualenv,$(PY_ENV))
|
|
||||||
|
|
||||||
PHONY += pylint-exe
|
|
||||||
pylint-exe: $(PY_ENV)
|
|
||||||
@$(PY_ENV_BIN)/python -m pip $(PIP_VERBOSE) install pylint
|
|
||||||
|
|
||||||
PHONY += pylint
|
|
||||||
pylint: pylint-exe
|
|
||||||
$(call cmd,pylint,$(PYOBJECTS))
|
|
||||||
|
|
||||||
PHONY += pybuild
|
|
||||||
pybuild: $(PY_ENV)
|
|
||||||
$(call cmd,pybuild)
|
|
||||||
|
|
||||||
PHONY += pytest
|
|
||||||
pytest: $(PY_ENV)
|
|
||||||
$(call cmd,pytest)
|
|
||||||
|
|
||||||
PHONY += pydebug
|
|
||||||
# set breakpoint with:
|
|
||||||
# DEBUG()
|
|
||||||
# e.g. to run tests in debug mode in emacs use:
|
|
||||||
# 'M-x pdb' ... 'make pydebug'
|
|
||||||
pydebug: $(PY_ENV)
|
|
||||||
DEBUG=$(DEBUG) $(PY_ENV_BIN)/pytest $(DEBUG) -v $(TEST_FOLDER)/$(TEST)
|
|
||||||
|
|
||||||
# runs python interpreter from ./local/py<N>/bin/python
|
|
||||||
pyenv-python: pyenvinstall
|
|
||||||
$(PY_ENV_BIN)/python -i
|
|
||||||
|
|
||||||
# With 'dependency_links=' setuptools supports dependencies on packages hosted
|
|
||||||
# on other reposetories then PyPi, see "Packages Not On PyPI" [1]. The big
|
|
||||||
# drawback is, due to security reasons (I don't know where the security gate on
|
|
||||||
# PyPi is), this feature is not supported by pip [2]. Thats why an upload to
|
|
||||||
# PyPi is required and since uploads via setuptools is not recommended, we have
|
|
||||||
# to imstall / use twine ... its really a mess.
|
|
||||||
#
|
|
||||||
# [1] https://python-packaging.readthedocs.io/en/latest/dependencies.html#packages-not-on-pypi
|
|
||||||
# [2] https://github.com/pypa/pip/pull/1519
|
|
||||||
|
|
||||||
# https://github.com/pypa/twine
|
|
||||||
PHONY += upload-pypi upload-pypi-test
|
|
||||||
upload-pypi: pyclean pyenvinstall pybuild
|
|
||||||
@$(PY_ENV_BIN)/twine upload $(PYDIST)/*
|
|
||||||
|
|
||||||
upload-pypi-test: pyclean pyenvinstall pybuild
|
|
||||||
@$(PY_ENV_BIN)/twine upload -r testpypi $(PYDIST)/*
|
|
||||||
.PHONY: $(PHONY)
|
|
|
@ -1,199 +0,0 @@
|
||||||
# -*- coding: utf-8; mode: makefile-gmake -*-
|
|
||||||
|
|
||||||
export DOCS_FOLDER DOCS_BUILD DOCS_DIST BOOKS_FOLDER BOOKS_DIST
|
|
||||||
|
|
||||||
# You can set these variables from the command line.
|
|
||||||
SPHINXOPTS ?=
|
|
||||||
SPHINXBUILD ?= $(PY_ENV_BIN)/sphinx-build
|
|
||||||
SPHINX_CONF ?= conf.py
|
|
||||||
|
|
||||||
DOCS_FOLDER = ./docs
|
|
||||||
DOCS_BUILD = ./$(LXC_ENV_FOLDER)build/docs
|
|
||||||
DOCS_DIST = ./$(LXC_ENV_FOLDER)dist/docs
|
|
||||||
GH_PAGES ?= build/gh-pages
|
|
||||||
|
|
||||||
BOOKS_FOLDER = ./docs
|
|
||||||
BOOKS_DIST = ./$(LXC_ENV_FOLDER)dist/books
|
|
||||||
|
|
||||||
ifeq ($(KBUILD_VERBOSE),1)
|
|
||||||
SPHINX_VERBOSE = "-v"
|
|
||||||
else
|
|
||||||
SPHINX_VERBOSE =
|
|
||||||
endif
|
|
||||||
|
|
||||||
|
|
||||||
docs-help:
|
|
||||||
@echo 'makefile.sphinx:'
|
|
||||||
@echo ' docs-clean - clean intermediate doc objects'
|
|
||||||
@echo ' $(GH_PAGES) - create & upload github pages'
|
|
||||||
@echo ' sphinx-pdf - run sphinx latex & pdf targets'
|
|
||||||
@echo ''
|
|
||||||
@echo ' books/{name}.html : build only the HTML of document {name}'
|
|
||||||
@echo ' valid values for books/{name}.html are:'
|
|
||||||
@echo ' $(BOOKS_HTML)' | $(FMT)
|
|
||||||
@echo ' books/{name}.pdf : build only the PDF of document {name}'
|
|
||||||
@echo ' valid values for books/{name}.pdf are:'
|
|
||||||
@echo ' $(BOOKS_PDF) ' | $(FMT)
|
|
||||||
|
|
||||||
# ------------------------------------------------------------------------------
|
|
||||||
# requirements
|
|
||||||
# ------------------------------------------------------------------------------
|
|
||||||
|
|
||||||
PHONY += msg-texlive texlive
|
|
||||||
|
|
||||||
ifeq ($(shell which xelatex >/dev/null 2>&1; echo $$?), 1)
|
|
||||||
texlive: msg-TeXLive
|
|
||||||
$(error The 'xelatex' command was not found)
|
|
||||||
else
|
|
||||||
texlive:
|
|
||||||
@:
|
|
||||||
endif
|
|
||||||
|
|
||||||
msg-texlive:
|
|
||||||
$(Q)echo "\n\
|
|
||||||
The TeX/PDF output and the *math* extension require TexLive and latexmk:\n\n\
|
|
||||||
Make sure you have a updated TeXLive with XeTeX engine installed, grab it\n\
|
|
||||||
it from https://www.tug.org/texlive or install it from your package manager.\n\n\
|
|
||||||
Install latexmk from your package manager or visit https://ctan.org/pkg/latexmk\n\n\
|
|
||||||
Sphinx-doc produce (Xe)LaTeX files which might use additional TeX-packages\n\
|
|
||||||
and fonts. To process these LaTeX files, a TexLive installation with the\n\
|
|
||||||
additional packages is required. On debian based OS these requirements\n\
|
|
||||||
are installed by::\n\n\
|
|
||||||
sudo -H apt-get install\n\
|
|
||||||
latexmk\n\
|
|
||||||
texlive-base texlive-xetex texlive-latex-recommended\n\
|
|
||||||
texlive-extra-utils dvipng ttf-dejavu\n"
|
|
||||||
|
|
||||||
# ------------------------------------------------------------------------------
|
|
||||||
# commands
|
|
||||||
# ------------------------------------------------------------------------------
|
|
||||||
|
|
||||||
# $2 sphinx builder e.g. "html"
|
|
||||||
# $3 path where configuration file (conf.py) is located
|
|
||||||
# $4 sourcedir
|
|
||||||
# $5 dest subfolder e.g. "man" for man pages at $(DOCS_DIST)/man
|
|
||||||
|
|
||||||
quiet_cmd_sphinx = SPHINX $@ --> file://$(abspath $(DOCS_DIST)/$5)
|
|
||||||
cmd_sphinx = SPHINX_CONF=$(abspath $4/$(SPHINX_CONF))\
|
|
||||||
$(SPHINXBUILD) $(SPHINX_VERBOSE) $(SPHINXOPTS)\
|
|
||||||
-b $2 -c $3 -d $(DOCS_BUILD)/.doctrees $4 $(DOCS_DIST)/$5
|
|
||||||
|
|
||||||
quiet_cmd_sphinx_autobuild = SPHINX $@ --> file://$(abspath $(DOCS_DIST)/$5)
|
|
||||||
cmd_sphinx_autobuild = PATH="$(PY_ENV_BIN):$(PATH)" $(PY_ENV_BIN)/sphinx-autobuild $(SPHINX_VERBOSE) --open-browser --host 0.0.0.0 $(SPHINXOPTS)\
|
|
||||||
-b $2 -c $3 -d $(DOCS_BUILD)/.doctrees $4 $(DOCS_DIST)/$5
|
|
||||||
|
|
||||||
quiet_cmd_sphinx_clean = CLEAN $@
|
|
||||||
cmd_sphinx_clean = rm -rf $(DOCS_BUILD) $(DOCS_DIST) $(GH_PAGES)/* $(GH_PAGES)/.buildinfo
|
|
||||||
|
|
||||||
# ------------------------------------------------------------------------------
|
|
||||||
# targets
|
|
||||||
# ------------------------------------------------------------------------------
|
|
||||||
|
|
||||||
# build PDF of whole documentation in: $(DOCS_DIST)/pdf
|
|
||||||
|
|
||||||
PHONY += sphinx-pdf
|
|
||||||
sphinx-pdf: sphinx-latex
|
|
||||||
$(Q)cd $(DOCS_BUILD)/latex/; make all-pdf
|
|
||||||
$(Q)mkdir -p $(DOCS_DIST)/pdf
|
|
||||||
$(Q)cp $(DOCS_BUILD)/latex/*.pdf $(DOCS_DIST)/pdf
|
|
||||||
@echo "SPHINX *.pdf --> file://$(abspath $(DOCS_DIST)/pdf)"
|
|
||||||
|
|
||||||
PHONY += sphinx-latex
|
|
||||||
sphinx-latex: pyenvinstall texlive
|
|
||||||
$(SPHINXBUILD) $(SPHINX_VERBOSE) $(SPHINXOPTS)\
|
|
||||||
-b latex \
|
|
||||||
-c $(DOCS_FOLDER) \
|
|
||||||
-d $(DOCS_BUILD)/.doctrees \
|
|
||||||
$(DOCS_FOLDER) \
|
|
||||||
$(DOCS_BUILD)/latex
|
|
||||||
|
|
||||||
# Sphinx projects, we call them *books* (what is more common). Books are
|
|
||||||
# folders under $(BOOKS_FOLDER) containing a conf.py file. The HTML output goes
|
|
||||||
# to folder $(BOOKS_DIST)/<name> while PDF is placed (BOOKS_DIST)/<name>/pdf
|
|
||||||
|
|
||||||
BOOKS=$(patsubst $(BOOKS_FOLDER)/%/conf.py,books/%,$(wildcard $(BOOKS_FOLDER)/*/conf.py))
|
|
||||||
|
|
||||||
# fine grained targets
|
|
||||||
BOOKS_HTML = $(patsubst %,%.html,$(BOOKS))
|
|
||||||
BOOKS_CLEAN = $(patsubst %,%.clean,$(BOOKS))
|
|
||||||
BOOKS_LATEX = $(patsubst %,%.latex,$(BOOKS))
|
|
||||||
BOOKS_PDF = $(patsubst %,%.pdf,$(BOOKS))
|
|
||||||
BOOKS_LIVE = $(patsubst %,%.live,$(BOOKS))
|
|
||||||
|
|
||||||
$(BOOKS_DIST):
|
|
||||||
mkdir -p $(BOOKS_DIST)
|
|
||||||
|
|
||||||
PHONY += $(BOOKS_HTML)
|
|
||||||
$(BOOKS_HTML): pyenvinstall | $(BOOKS_DIST)
|
|
||||||
SPHINX_CONF=$(patsubst books/%.html,%,$@)/conf.py \
|
|
||||||
$(SPHINXBUILD) $(SPHINX_VERBOSE) $(SPHINXOPTS)\
|
|
||||||
-b html \
|
|
||||||
-c $(DOCS_FOLDER) \
|
|
||||||
-d $(DOCS_BUILD)/books/$(patsubst books/%.html,%,$@)/.doctrees \
|
|
||||||
$(BOOKS_FOLDER)/$(patsubst books/%.html,%,$@) \
|
|
||||||
$(BOOKS_DIST)/$(patsubst books/%.html,%,$@)
|
|
||||||
@echo "SPHINX $@ --> file://$(abspath $(BOOKS_DIST)/$(patsubst books/%.html,%,$@))"
|
|
||||||
|
|
||||||
PHONY += $(BOOKS_HTML)
|
|
||||||
$(BOOKS_LIVE): pyenvinstall | $(BOOKS_DIST)
|
|
||||||
PATH="$(PY_ENV_BIN):$(PATH)" \
|
|
||||||
SPHINX_CONF=$(patsubst books/%.live,%,$@)/conf.py \
|
|
||||||
$(PY_ENV_BIN)/sphinx-autobuild --poll -B --host 0.0.0.0 --port 8080 $(SPHINX_VERBOSE) $(SPHINXOPTS)\
|
|
||||||
-b html \
|
|
||||||
-c $(DOCS_FOLDER) \
|
|
||||||
-d $(DOCS_BUILD)/books/$(patsubst books/%.live,%,$@)/.doctrees \
|
|
||||||
$(BOOKS_FOLDER)/$(patsubst books/%.live,%,$@) \
|
|
||||||
$(BOOKS_DIST)/$(patsubst books/%.live,%,$@)
|
|
||||||
|
|
||||||
$(BOOKS_PDF): %.pdf : %.latex
|
|
||||||
$(Q)cd $(DOCS_BUILD)/latex/$(patsubst books/%.pdf,%,$@); make all-pdf
|
|
||||||
$(Q)mkdir -p $(BOOKS_DIST)/$(patsubst books/%.pdf,%,$@)/pdf
|
|
||||||
$(Q)cp -v $(DOCS_BUILD)/latex/$(patsubst books/%.pdf,%,$@)/*.pdf $(BOOKS_DIST)/$(patsubst books/%.pdf,%,$@)/pdf
|
|
||||||
@echo "SPHINX $@ --> file://$(abspath $(BOOKS_DIST)/$(patsubst books/%.pdf,%,$@))/pdf"
|
|
||||||
|
|
||||||
PHONY += $(BOOKS_LATEX)
|
|
||||||
$(BOOKS_LATEX): pyenvinstall | $(BOOKS_DIST)
|
|
||||||
SPHINX_CONF=$(patsubst books/%.latex,%,$@)/conf.py \
|
|
||||||
$(SPHINXBUILD) $(SPHINX_VERBOSE) $(SPHINXOPTS)\
|
|
||||||
-b latex \
|
|
||||||
-c $(DOCS_FOLDER) \
|
|
||||||
-d $(DOCS_BUILD)/books/$(patsubst books/%.latex,%,$@)/.doctrees \
|
|
||||||
$(BOOKS_FOLDER)/$(patsubst books/%.latex,%,$@) \
|
|
||||||
$(DOCS_BUILD)/latex/$(patsubst books/%.latex,%,$@)
|
|
||||||
@echo "SPHINX $@ --> file://$(abspath $(DOCS_BUILD)/latex/$(patsubst books/%.latex,%,$@))"
|
|
||||||
|
|
||||||
$(BOOKS_CLEAN):
|
|
||||||
$(Q)rm -rf $(BOOKS_DIST)/$(patsubst books/%.clean,%,$@) \
|
|
||||||
$(DOCS_BUILD)/books/$(patsubst books/%.clean,%,$@) \
|
|
||||||
$(DOCS_BUILD)/latex/$(patsubst books/%.clean,%,$@)
|
|
||||||
|
|
||||||
# github pages
|
|
||||||
PHONY += prepare-gh-pages
|
|
||||||
prepare-gh-pages:
|
|
||||||
cp -r $(DOCS_DIST)/* $(GH_PAGES)/
|
|
||||||
touch $(GH_PAGES)/.nojekyll
|
|
||||||
echo "<html><head><META http-equiv='refresh' content='0;URL=index.html'></head></html>" > $(GH_PAGES)/404.html
|
|
||||||
|
|
||||||
PHONY += gh-pages
|
|
||||||
gh-pages: docs-clean docs
|
|
||||||
- git worktree remove -f $(GH_PAGES) || exit 0
|
|
||||||
- git branch -D gh-pages || exit 0
|
|
||||||
git worktree add --no-checkout $(GH_PAGES) master
|
|
||||||
cd $(GH_PAGES); git checkout --orphan gh-pages && git rm -rfq .
|
|
||||||
$(MAKE) prepare-gh-pages
|
|
||||||
cd $(GH_PAGES);\
|
|
||||||
git add --all . ;\
|
|
||||||
git commit -q -m "make gh-pages: from $(shell git config --get remote.origin.url)@$(shell git rev-parse HEAD)" ;\
|
|
||||||
git push -f origin gh-pages
|
|
||||||
|
|
||||||
PHONY += ci-gh-pages
|
|
||||||
ci-gh-pages: docs-clean docs
|
|
||||||
rm -Rf $(GH_PAGES)
|
|
||||||
mkdir -p $(GH_PAGES)
|
|
||||||
$(MAKE) prepare-gh-pages
|
|
||||||
|
|
||||||
PHONY += docs-clean
|
|
||||||
docs-clean: $(BOOKS_CLEAN)
|
|
||||||
$(call cmd,sphinx_clean)
|
|
||||||
|
|
||||||
.PHONY: $(PHONY)
|
|
|
@ -418,9 +418,9 @@ install_settings() {
|
||||||
err_msg "you have to install searx first"
|
err_msg "you have to install searx first"
|
||||||
exit 42
|
exit 42
|
||||||
fi
|
fi
|
||||||
mkdir -p "$(dirname ${SEARX_SETTINGS_PATH})"
|
mkdir -p "$(dirname "${SEARX_SETTINGS_PATH}")"
|
||||||
|
|
||||||
if [[ ! -f ${SEARX_SETTINGS_PATH} ]]; then
|
if [[ ! -f "${SEARX_SETTINGS_PATH}" ]]; then
|
||||||
info_msg "install settings ${SEARX_SETTINGS_TEMPLATE}"
|
info_msg "install settings ${SEARX_SETTINGS_TEMPLATE}"
|
||||||
info_msg " --> ${SEARX_SETTINGS_PATH}"
|
info_msg " --> ${SEARX_SETTINGS_PATH}"
|
||||||
cp "${SEARX_SETTINGS_TEMPLATE}" "${SEARX_SETTINGS_PATH}"
|
cp "${SEARX_SETTINGS_TEMPLATE}" "${SEARX_SETTINGS_PATH}"
|
||||||
|
@ -481,7 +481,7 @@ pyenv_is_available() {
|
||||||
create_pyenv() {
|
create_pyenv() {
|
||||||
rst_title "Create virtualenv (python)" section
|
rst_title "Create virtualenv (python)" section
|
||||||
echo
|
echo
|
||||||
if [[ ! -f "${SEARX_SRC}/manage.sh" ]]; then
|
if [[ ! -f "${SEARX_SRC}/manage" ]]; then
|
||||||
err_msg "to create pyenv for searx, searx has to be cloned first"
|
err_msg "to create pyenv for searx, searx has to be cloned first"
|
||||||
return 42
|
return 42
|
||||||
fi
|
fi
|
||||||
|
|
|
@ -1,48 +0,0 @@
|
||||||
# -*- coding: utf-8; mode: python -*-
|
|
||||||
"""Implement some sphinx-build tools.
|
|
||||||
|
|
||||||
"""
|
|
||||||
|
|
||||||
import os
|
|
||||||
import sys
|
|
||||||
from sphinx.util.pycompat import execfile_
|
|
||||||
|
|
||||||
# ------------------------------------------------------------------------------
|
|
||||||
def load_sphinx_config(namespace):
|
|
||||||
# ------------------------------------------------------------------------------
|
|
||||||
|
|
||||||
u"""Load an additional configuration file into *namespace*.
|
|
||||||
|
|
||||||
The name of the configuration file is taken from the environment
|
|
||||||
``SPHINX_CONF``. The external configuration file extends (or overwrites) the
|
|
||||||
configuration values from the origin ``conf.py``. With this you are able to
|
|
||||||
maintain *build themes*. To your docs/conf.py add::
|
|
||||||
|
|
||||||
from sphinx_build_tools import load_sphinx_config
|
|
||||||
...
|
|
||||||
|
|
||||||
# Since loadConfig overwrites settings from the global namespace, it has to be
|
|
||||||
# the last statement in the conf.py file
|
|
||||||
|
|
||||||
load_sphinx_config(globals())
|
|
||||||
|
|
||||||
"""
|
|
||||||
|
|
||||||
config_file = os.environ.get("SPHINX_CONF", None)
|
|
||||||
if (config_file is not None
|
|
||||||
and os.path.normpath(namespace["__file__"]) != os.path.normpath(config_file) ):
|
|
||||||
config_file = os.path.abspath(config_file)
|
|
||||||
|
|
||||||
if os.path.isfile(config_file):
|
|
||||||
sys.stdout.write(
|
|
||||||
"load additional sphinx-config: %s\n"
|
|
||||||
% config_file)
|
|
||||||
config = namespace.copy()
|
|
||||||
config['__file__'] = config_file
|
|
||||||
execfile_(config_file, config)
|
|
||||||
del config['__file__']
|
|
||||||
namespace.update(config)
|
|
||||||
else:
|
|
||||||
sys.stderr.write(
|
|
||||||
"WARNING: additional sphinx-config not found: %s\n"
|
|
||||||
% config_file)
|
|
Loading…
Reference in a new issue