diff --git a/searx/engines/etools.py b/searx/engines/etools.py
deleted file mode 100644
index 08bc63cd8..000000000
--- a/searx/engines/etools.py
+++ /dev/null
@@ -1,58 +0,0 @@
-# SPDX-License-Identifier: AGPL-3.0-or-later
-"""
- eTools (Web)
-"""
-
-from lxml import html
-from urllib.parse import quote
-from searx.utils import extract_text, eval_xpath
-
-# about
-about = {
-    "website": 'https://www.etools.ch',
-    "wikidata_id": None,
-    "official_api_documentation": None,
-    "use_official_api": False,
-    "require_api_key": False,
-    "results": 'HTML',
-}
-
-categories = ['general', 'web']
-paging = False
-safesearch = True
-
-base_url = 'https://www.etools.ch'
-search_path = (
-    # fmt: off
-    '/searchAdvancedSubmit.do'
-    '?query={search_term}'
-    '&pageResults=20'
-    '&safeSearch={safesearch}'
-    # fmt: on
-)
-
-
-def request(query, params):
-    if params['safesearch']:
-        safesearch = 'true'
-    else:
-        safesearch = 'false'
-
-    params['url'] = base_url + search_path.format(search_term=quote(query), safesearch=safesearch)
-
-    return params
-
-
-def response(resp):
-    results = []
-
-    dom = html.fromstring(resp.text)
-
-    for result in eval_xpath(dom, '//table[@class="result"]//td[@class="record"]'):
-        url = eval_xpath(result, './a/@href')[0]
-        title = extract_text(eval_xpath(result, './a//text()'))
-        content = extract_text(eval_xpath(result, './/div[@class="text"]//text()'))
-
-        results.append({'url': url, 'title': title, 'content': content})
-
-    return results
diff --git a/searx/settings.yml b/searx/settings.yml
index f1d6fa1b1..3b4b0e3b4 100644
--- a/searx/settings.yml
+++ b/searx/settings.yml
@@ -479,13 +479,6 @@ engines:
     timeout: 3.0
     disabled: true
 
-  - name: etools
-    engine: etools
-    shortcut: eto
-    disabled: true
-    additional_tests:
-      rosebud: *test_rosebud
-
   - name: etymonline
     engine: xpath
     paging: true