| 
									
										
										
										
											2022-01-29 11:16:28 +01:00
										 |  |  | # SPDX-License-Identifier: AGPL-3.0-or-later | 
					
						
							|  |  |  | """Utility functions for the engines
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | """
 | 
					
						
							| 
									
										
										
										
											2024-05-24 01:21:58 +02:00
										 |  |  | 
 | 
					
						
							|  |  |  | from __future__ import annotations | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2015-01-11 13:26:40 +01:00
										 |  |  | import re | 
					
						
							| 
									
										
										
										
											2020-10-05 12:50:08 +02:00
										 |  |  | import importlib | 
					
						
							| 
									
										
										
										
											2022-01-30 22:14:12 +01:00
										 |  |  | import importlib.util | 
					
						
							| 
									
										
										
										
											2023-09-09 12:18:39 +02:00
										 |  |  | import json | 
					
						
							| 
									
										
										
										
											2022-01-30 22:14:12 +01:00
										 |  |  | import types | 
					
						
							| 
									
										
										
										
											2015-01-11 13:26:40 +01:00
										 |  |  | 
 | 
					
						
							| 
									
										
										
										
											2022-01-29 11:16:28 +01:00
										 |  |  | from typing import Optional, Union, Any, Set, List, Dict, MutableMapping, Tuple, Callable | 
					
						
							| 
									
										
										
										
											2017-12-01 20:45:24 +01:00
										 |  |  | from numbers import Number | 
					
						
							| 
									
										
										
										
											2016-11-19 17:51:19 +01:00
										 |  |  | from os.path import splitext, join | 
					
						
							| 
									
										
										
										
											2014-04-25 01:46:40 +02:00
										 |  |  | from random import choice | 
					
						
							| 
									
										
										
										
											2020-08-06 17:42:46 +02:00
										 |  |  | from html.parser import HTMLParser | 
					
						
							| 
									
										
										
										
											2023-10-18 14:34:18 +02:00
										 |  |  | from html import escape | 
					
						
							| 
									
										
										
										
											2024-09-15 00:28:35 +02:00
										 |  |  | from urllib.parse import urljoin, urlparse, parse_qs, urlencode | 
					
						
							| 
									
										
										
										
											2023-09-08 08:40:22 +02:00
										 |  |  | from markdown_it import MarkdownIt | 
					
						
							| 
									
										
										
										
											2020-10-02 18:13:56 +02:00
										 |  |  | 
 | 
					
						
							|  |  |  | from lxml import html | 
					
						
							| 
									
										
										
										
											2024-04-08 08:50:14 +02:00
										 |  |  | from lxml.etree import ElementBase, XPath, XPathError, XPathSyntaxError | 
					
						
							| 
									
										
										
										
											2020-10-02 18:13:56 +02:00
										 |  |  | 
 | 
					
						
							| 
									
										
										
										
											2018-03-01 05:30:48 +01:00
										 |  |  | from searx import settings | 
					
						
							| 
									
										
										
										
											2022-12-11 16:45:47 +01:00
										 |  |  | from searx.data import USER_AGENTS, data_dir | 
					
						
							| 
									
										
										
										
											2021-07-27 18:37:46 +02:00
										 |  |  | from searx.version import VERSION_TAG | 
					
						
							| 
									
										
										
										
											2022-10-10 19:31:22 +02:00
										 |  |  | from searx.sxng_locales import sxng_locales | 
					
						
							| 
									
										
										
										
											2020-11-26 15:12:11 +01:00
										 |  |  | from searx.exceptions import SearxXPathSyntaxException, SearxEngineXPathException | 
					
						
							| 
									
										
										
										
											2015-01-11 13:26:40 +01:00
										 |  |  | from searx import logger | 
					
						
							| 
									
										
										
										
											2014-11-18 11:37:42 +01:00
										 |  |  | 
 | 
					
						
							| 
									
										
										
										
											2015-01-11 13:26:40 +01:00
										 |  |  | 
 | 
					
						
							|  |  |  | logger = logger.getChild('utils') | 
					
						
							| 
									
										
										
										
											2014-01-10 23:38:08 +01:00
										 |  |  | 
 | 
					
						
							| 
									
										
										
										
											2022-01-29 11:16:28 +01:00
										 |  |  | XPathSpecType = Union[str, XPath] | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | _BLOCKED_TAGS = ('script', 'style') | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | _ECMA_UNESCAPE4_RE = re.compile(r'%u([0-9a-fA-F]{4})', re.UNICODE) | 
					
						
							|  |  |  | _ECMA_UNESCAPE2_RE = re.compile(r'%([0-9a-fA-F]{2})', re.UNICODE) | 
					
						
							| 
									
										
										
										
											2015-01-01 14:13:56 +01:00
										 |  |  | 
 | 
					
						
							| 
									
										
										
										
											2023-09-09 12:18:39 +02:00
										 |  |  | _JS_QUOTE_KEYS_RE = re.compile(r'([\{\s,])(\w+)(:)') | 
					
						
							|  |  |  | _JS_VOID_RE = re.compile(r'void\s+[0-9]+|void\s*\([0-9]+\)') | 
					
						
							| 
									
										
										
										
											2023-09-15 20:57:03 +02:00
										 |  |  | _JS_DECIMAL_RE = re.compile(r":\s*\.") | 
					
						
							| 
									
										
										
										
											2023-09-09 12:18:39 +02:00
										 |  |  | 
 | 
					
						
							| 
									
										
										
										
											2022-01-30 22:14:12 +01:00
										 |  |  | _XPATH_CACHE: Dict[str, XPath] = {} | 
					
						
							|  |  |  | _LANG_TO_LC_CACHE: Dict[str, Dict[str, str]] = {} | 
					
						
							| 
									
										
										
										
											2019-07-18 21:32:17 +02:00
										 |  |  | 
 | 
					
						
							| 
									
										
										
										
											2024-03-11 14:06:26 +01:00
										 |  |  | _FASTTEXT_MODEL: Optional["fasttext.FastText._FastText"] = None  # type: ignore | 
					
						
							| 
									
										
										
										
											2024-10-09 11:59:31 +02:00
										 |  |  | """fasttext model to predict language of a search term""" | 
					
						
							| 
									
										
										
										
											2022-12-11 16:45:47 +01:00
										 |  |  | 
 | 
					
						
							| 
									
										
										
										
											2022-10-10 19:31:22 +02:00
										 |  |  | SEARCH_LANGUAGE_CODES = frozenset([searxng_locale[0].split('-')[0] for searxng_locale in sxng_locales]) | 
					
						
							|  |  |  | """Languages supported by most searxng engines (:py:obj:`searx.sxng_locales.sxng_locales`).""" | 
					
						
							| 
									
										
										
										
											2022-12-16 21:28:57 +01:00
										 |  |  | 
 | 
					
						
							| 
									
										
										
										
											2014-01-19 22:59:01 +01:00
										 |  |  | 
 | 
					
						
							| 
									
										
										
										
											2022-01-29 11:16:28 +01:00
										 |  |  | class _NotSetClass:  # pylint: disable=too-few-public-methods | 
					
						
							|  |  |  |     """Internal class for this module, do not create instance of this class.
 | 
					
						
							|  |  |  |     Replace the None value, allow explicitly pass None as a function argument"""
 | 
					
						
							| 
									
										
										
										
											2020-11-26 15:12:11 +01:00
										 |  |  | 
 | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2022-01-29 11:16:28 +01:00
										 |  |  | _NOTSET = _NotSetClass() | 
					
						
							| 
									
										
										
										
											2020-11-26 15:12:11 +01:00
										 |  |  | 
 | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2022-01-29 11:16:28 +01:00
										 |  |  | def searx_useragent() -> str: | 
					
						
							| 
									
										
										
										
											2020-10-02 18:17:01 +02:00
										 |  |  |     """Return the searx User Agent""" | 
					
						
							| 
									
										
										
										
											2014-12-29 21:31:04 +01:00
										 |  |  |     return 'searx/{searx_version} {suffix}'.format( | 
					
						
							| 
									
										
										
										
											2021-12-27 09:26:22 +01:00
										 |  |  |         searx_version=VERSION_TAG, suffix=settings['outgoing']['useragent_suffix'] | 
					
						
							|  |  |  |     ).strip() | 
					
						
							| 
									
										
										
										
											2014-10-19 12:41:04 +02:00
										 |  |  | 
 | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2022-07-28 13:02:56 +02:00
										 |  |  | def gen_useragent(os_string: Optional[str] = None) -> str: | 
					
						
							| 
									
										
										
										
											2020-10-02 18:17:01 +02:00
										 |  |  |     """Return a random browser User Agent
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     See searx/data/useragents.json | 
					
						
							|  |  |  |     """
 | 
					
						
							| 
									
										
										
										
											2022-01-29 11:16:28 +01:00
										 |  |  |     return USER_AGENTS['ua'].format(os=os_string or choice(USER_AGENTS['os']), version=choice(USER_AGENTS['versions'])) | 
					
						
							| 
									
										
										
										
											2017-05-28 15:46:45 +02:00
										 |  |  | 
 | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2022-01-29 11:16:28 +01:00
										 |  |  | class _HTMLTextExtractorException(Exception): | 
					
						
							|  |  |  |     """Internal exception raised when the HTML is invalid""" | 
					
						
							| 
									
										
										
										
											2020-09-11 10:23:56 +02:00
										 |  |  | 
 | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2023-10-18 14:34:18 +02:00
										 |  |  | class _HTMLTextExtractor(HTMLParser): | 
					
						
							| 
									
										
										
										
											2022-01-29 11:16:28 +01:00
										 |  |  |     """Internal class to extract text from HTML""" | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2013-11-08 23:44:26 +01:00
										 |  |  |     def __init__(self): | 
					
						
							|  |  |  |         HTMLParser.__init__(self) | 
					
						
							| 
									
										
										
										
											2014-01-19 22:59:01 +01:00
										 |  |  |         self.result = [] | 
					
						
							| 
									
										
										
										
											2015-01-01 14:13:56 +01:00
										 |  |  |         self.tags = [] | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     def handle_starttag(self, tag, attrs): | 
					
						
							|  |  |  |         self.tags.append(tag) | 
					
						
							| 
									
										
										
										
											2022-04-15 13:37:27 +02:00
										 |  |  |         if tag == 'br': | 
					
						
							|  |  |  |             self.result.append(' ') | 
					
						
							| 
									
										
										
										
											2015-01-01 14:13:56 +01:00
										 |  |  | 
 | 
					
						
							|  |  |  |     def handle_endtag(self, tag): | 
					
						
							| 
									
										
										
										
											2015-01-22 17:43:45 +01:00
										 |  |  |         if not self.tags: | 
					
						
							|  |  |  |             return | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2015-01-01 14:13:56 +01:00
										 |  |  |         if tag != self.tags[-1]: | 
					
						
							| 
									
										
										
										
											2022-01-29 11:16:28 +01:00
										 |  |  |             raise _HTMLTextExtractorException() | 
					
						
							| 
									
										
										
										
											2015-01-22 17:43:45 +01:00
										 |  |  | 
 | 
					
						
							| 
									
										
										
										
											2015-01-01 14:13:56 +01:00
										 |  |  |         self.tags.pop() | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     def is_valid_tag(self): | 
					
						
							| 
									
										
										
										
											2022-01-29 11:16:28 +01:00
										 |  |  |         return not self.tags or self.tags[-1] not in _BLOCKED_TAGS | 
					
						
							| 
									
										
										
										
											2013-11-08 23:44:26 +01:00
										 |  |  | 
 | 
					
						
							| 
									
										
										
										
											2020-11-16 09:43:23 +01:00
										 |  |  |     def handle_data(self, data): | 
					
						
							| 
									
										
										
										
											2015-01-01 14:13:56 +01:00
										 |  |  |         if not self.is_valid_tag(): | 
					
						
							|  |  |  |             return | 
					
						
							| 
									
										
										
										
											2020-11-16 09:43:23 +01:00
										 |  |  |         self.result.append(data) | 
					
						
							| 
									
										
										
										
											2013-11-08 23:44:26 +01:00
										 |  |  | 
 | 
					
						
							| 
									
										
										
										
											2020-11-16 09:43:23 +01:00
										 |  |  |     def handle_charref(self, name): | 
					
						
							| 
									
										
										
										
											2015-01-01 14:13:56 +01:00
										 |  |  |         if not self.is_valid_tag(): | 
					
						
							|  |  |  |             return | 
					
						
							| 
									
										
										
										
											2020-11-16 09:43:23 +01:00
										 |  |  |         if name[0] in ('x', 'X'): | 
					
						
							|  |  |  |             codepoint = int(name[1:], 16) | 
					
						
							| 
									
										
										
										
											2014-01-20 02:31:20 +01:00
										 |  |  |         else: | 
					
						
							| 
									
										
										
										
											2020-11-16 09:43:23 +01:00
										 |  |  |             codepoint = int(name) | 
					
						
							| 
									
										
										
										
											2020-08-06 17:42:46 +02:00
										 |  |  |         self.result.append(chr(codepoint)) | 
					
						
							| 
									
										
										
										
											2013-11-08 23:44:26 +01:00
										 |  |  | 
 | 
					
						
							|  |  |  |     def handle_entityref(self, name): | 
					
						
							| 
									
										
										
										
											2015-01-01 14:13:56 +01:00
										 |  |  |         if not self.is_valid_tag(): | 
					
						
							|  |  |  |             return | 
					
						
							| 
									
										
										
										
											2014-10-19 12:41:04 +02:00
										 |  |  |         # codepoint = htmlentitydefs.name2codepoint[name] | 
					
						
							| 
									
										
										
										
											2020-08-06 17:42:46 +02:00
										 |  |  |         # self.result.append(chr(codepoint)) | 
					
						
							| 
									
										
										
										
											2013-11-18 16:47:20 +01:00
										 |  |  |         self.result.append(name) | 
					
						
							| 
									
										
										
										
											2013-11-08 23:44:26 +01:00
										 |  |  | 
 | 
					
						
							|  |  |  |     def get_text(self): | 
					
						
							| 
									
										
										
										
											2020-08-06 17:42:46 +02:00
										 |  |  |         return ''.join(self.result).strip() | 
					
						
							| 
									
										
										
										
											2013-11-08 23:44:26 +01:00
										 |  |  | 
 | 
					
						
							| 
									
										
										
										
											2023-10-18 14:34:18 +02:00
										 |  |  |     def error(self, message): | 
					
						
							|  |  |  |         # error handle is needed in <py3.10 | 
					
						
							|  |  |  |         # https://github.com/python/cpython/pull/8562/files | 
					
						
							|  |  |  |         raise AssertionError(message) | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2014-01-19 22:59:01 +01:00
										 |  |  | 
 | 
					
						
							| 
									
										
										
										
											2022-01-29 11:16:28 +01:00
										 |  |  | def html_to_text(html_str: str) -> str: | 
					
						
							| 
									
										
										
										
											2020-10-02 18:17:01 +02:00
										 |  |  |     """Extract text from a HTML string
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     Args: | 
					
						
							|  |  |  |         * html_str (str): string HTML | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     Returns: | 
					
						
							|  |  |  |         * str: extracted text | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     Examples: | 
					
						
							|  |  |  |         >>> html_to_text('Example <span id="42">#2</span>') | 
					
						
							|  |  |  |         'Example #2' | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |         >>> html_to_text('<style>.span { color: red; }</style><span>Example</span>') | 
					
						
							|  |  |  |         'Example' | 
					
						
							| 
									
										
										
										
											2023-10-18 14:34:18 +02:00
										 |  |  | 
 | 
					
						
							|  |  |  |         >>> html_to_text(r'regexp: (?<![a-zA-Z]') | 
					
						
							|  |  |  |         'regexp: (?<![a-zA-Z]' | 
					
						
							| 
									
										
										
										
											2020-10-02 18:17:01 +02:00
										 |  |  |     """
 | 
					
						
							| 
									
										
										
										
											2022-04-15 13:37:27 +02:00
										 |  |  |     html_str = html_str.replace('\n', ' ').replace('\r', ' ') | 
					
						
							| 
									
										
										
										
											2020-10-02 18:17:01 +02:00
										 |  |  |     html_str = ' '.join(html_str.split()) | 
					
						
							| 
									
										
										
										
											2022-01-29 11:16:28 +01:00
										 |  |  |     s = _HTMLTextExtractor() | 
					
						
							| 
									
										
										
										
											2020-09-11 10:23:56 +02:00
										 |  |  |     try: | 
					
						
							| 
									
										
										
										
											2020-10-02 18:17:01 +02:00
										 |  |  |         s.feed(html_str) | 
					
						
							| 
									
										
										
										
											2023-10-18 14:34:18 +02:00
										 |  |  |     except AssertionError: | 
					
						
							|  |  |  |         s = _HTMLTextExtractor() | 
					
						
							|  |  |  |         s.feed(escape(html_str, quote=True)) | 
					
						
							| 
									
										
										
										
											2022-01-29 11:16:28 +01:00
										 |  |  |     except _HTMLTextExtractorException: | 
					
						
							| 
									
										
										
										
											2020-10-02 18:17:01 +02:00
										 |  |  |         logger.debug("HTMLTextExtractor: invalid HTML\n%s", html_str) | 
					
						
							| 
									
										
										
										
											2013-11-08 23:44:26 +01:00
										 |  |  |     return s.get_text() | 
					
						
							| 
									
										
										
										
											2013-11-15 18:55:18 +01:00
										 |  |  | 
 | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2023-09-08 08:40:22 +02:00
										 |  |  | def markdown_to_text(markdown_str: str) -> str: | 
					
						
							|  |  |  |     """Extract text from a Markdown string
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     Args: | 
					
						
							|  |  |  |         * markdown_str (str): string Markdown | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     Returns: | 
					
						
							|  |  |  |         * str: extracted text | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     Examples: | 
					
						
							|  |  |  |         >>> markdown_to_text('[example](https://example.com)') | 
					
						
							|  |  |  |         'example' | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |         >>> markdown_to_text('## Headline') | 
					
						
							|  |  |  |         'Headline' | 
					
						
							|  |  |  |     """
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     html_str = ( | 
					
						
							|  |  |  |         MarkdownIt("commonmark", {"typographer": True}).enable(["replacements", "smartquotes"]).render(markdown_str) | 
					
						
							|  |  |  |     ) | 
					
						
							|  |  |  |     return html_to_text(html_str) | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2022-01-30 22:14:12 +01:00
										 |  |  | def extract_text(xpath_results, allow_none: bool = False) -> Optional[str]: | 
					
						
							| 
									
										
										
										
											2020-10-02 18:17:01 +02:00
										 |  |  |     """Extract text from a lxml result
 | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2021-12-27 09:26:22 +01:00
										 |  |  |     * if xpath_results is list, extract the text from each result and concat the list | 
					
						
							|  |  |  |     * if xpath_results is a xml element, extract all the text node from it | 
					
						
							|  |  |  |       ( text_content() method from lxml ) | 
					
						
							|  |  |  |     * if xpath_results is a string element, then it's already done | 
					
						
							| 
									
										
										
										
											2020-10-02 18:17:01 +02:00
										 |  |  |     """
 | 
					
						
							| 
									
										
										
										
											2020-11-26 15:12:11 +01:00
										 |  |  |     if isinstance(xpath_results, list): | 
					
						
							| 
									
										
										
										
											2020-10-02 18:13:56 +02:00
										 |  |  |         # it's list of result : concat everything using recursive call | 
					
						
							|  |  |  |         result = '' | 
					
						
							|  |  |  |         for e in xpath_results: | 
					
						
							| 
									
										
										
										
											2022-01-29 11:16:28 +01:00
										 |  |  |             result = result + (extract_text(e) or '') | 
					
						
							| 
									
										
										
										
											2020-10-02 18:13:56 +02:00
										 |  |  |         return result.strip() | 
					
						
							| 
									
										
										
										
											2022-01-29 11:16:28 +01:00
										 |  |  |     if isinstance(xpath_results, ElementBase): | 
					
						
							| 
									
										
										
										
											2020-10-02 18:13:56 +02:00
										 |  |  |         # it's a element | 
					
						
							| 
									
										
										
										
											2022-01-29 11:16:28 +01:00
										 |  |  |         text: str = html.tostring(xpath_results, encoding='unicode', method='text', with_tail=False) | 
					
						
							| 
									
										
										
										
											2020-10-02 18:13:56 +02:00
										 |  |  |         text = text.strip().replace('\n', ' ') | 
					
						
							|  |  |  |         return ' '.join(text.split()) | 
					
						
							| 
									
										
										
										
											2024-04-08 08:50:14 +02:00
										 |  |  |     if isinstance(xpath_results, (str, Number, bool)): | 
					
						
							| 
									
										
										
										
											2020-11-26 15:12:11 +01:00
										 |  |  |         return str(xpath_results) | 
					
						
							| 
									
										
										
										
											2022-01-29 11:16:28 +01:00
										 |  |  |     if xpath_results is None and allow_none: | 
					
						
							| 
									
										
										
										
											2020-11-26 15:12:11 +01:00
										 |  |  |         return None | 
					
						
							| 
									
										
										
										
											2022-01-29 11:16:28 +01:00
										 |  |  |     if xpath_results is None and not allow_none: | 
					
						
							| 
									
										
										
										
											2020-11-26 15:12:11 +01:00
										 |  |  |         raise ValueError('extract_text(None, allow_none=False)') | 
					
						
							| 
									
										
										
										
											2022-01-29 11:16:28 +01:00
										 |  |  |     raise ValueError('unsupported type') | 
					
						
							| 
									
										
										
										
											2020-10-02 18:13:56 +02:00
										 |  |  | 
 | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2022-01-29 11:16:28 +01:00
										 |  |  | def normalize_url(url: str, base_url: str) -> str: | 
					
						
							| 
									
										
										
										
											2020-10-03 10:02:50 +02:00
										 |  |  |     """Normalize URL: add protocol, join URL with base_url, add trailing slash if there is no path
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     Args: | 
					
						
							|  |  |  |         * url (str): Relative URL | 
					
						
							|  |  |  |         * base_url (str): Base URL, it must be an absolute URL. | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     Example: | 
					
						
							|  |  |  |         >>> normalize_url('https://example.com', 'http://example.com/') | 
					
						
							|  |  |  |         'https://example.com/' | 
					
						
							|  |  |  |         >>> normalize_url('//example.com', 'http://example.com/') | 
					
						
							|  |  |  |         'http://example.com/' | 
					
						
							|  |  |  |         >>> normalize_url('//example.com', 'https://example.com/') | 
					
						
							|  |  |  |         'https://example.com/' | 
					
						
							|  |  |  |         >>> normalize_url('/path?a=1', 'https://example.com') | 
					
						
							|  |  |  |         'https://example.com/path?a=1' | 
					
						
							|  |  |  |         >>> normalize_url('', 'https://example.com') | 
					
						
							|  |  |  |         'https://example.com/' | 
					
						
							|  |  |  |         >>> normalize_url('/test', '/path') | 
					
						
							| 
									
										
										
										
											2020-11-26 15:12:11 +01:00
										 |  |  |         raise ValueError | 
					
						
							| 
									
										
										
										
											2020-10-03 10:02:50 +02:00
										 |  |  | 
 | 
					
						
							|  |  |  |     Raises: | 
					
						
							|  |  |  |         * lxml.etree.ParserError | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     Returns: | 
					
						
							|  |  |  |         * str: normalized URL | 
					
						
							|  |  |  |     """
 | 
					
						
							|  |  |  |     if url.startswith('//'): | 
					
						
							|  |  |  |         # add http or https to this kind of url //example.com/ | 
					
						
							|  |  |  |         parsed_search_url = urlparse(base_url) | 
					
						
							|  |  |  |         url = '{0}:{1}'.format(parsed_search_url.scheme or 'http', url) | 
					
						
							|  |  |  |     elif url.startswith('/'): | 
					
						
							|  |  |  |         # fix relative url to the search engine | 
					
						
							|  |  |  |         url = urljoin(base_url, url) | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     # fix relative urls that fall through the crack | 
					
						
							|  |  |  |     if '://' not in url: | 
					
						
							|  |  |  |         url = urljoin(base_url, url) | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2020-10-02 18:17:01 +02:00
										 |  |  |     parsed_url = urlparse(url) | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     # add a / at this end of the url if there is no path | 
					
						
							|  |  |  |     if not parsed_url.netloc: | 
					
						
							| 
									
										
										
										
											2020-11-26 15:12:11 +01:00
										 |  |  |         raise ValueError('Cannot parse url') | 
					
						
							| 
									
										
										
										
											2020-10-02 18:17:01 +02:00
										 |  |  |     if not parsed_url.path: | 
					
						
							|  |  |  |         url += '/' | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     return url | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2022-01-29 11:16:28 +01:00
										 |  |  | def extract_url(xpath_results, base_url) -> str: | 
					
						
							| 
									
										
										
										
											2020-10-02 18:17:01 +02:00
										 |  |  |     """Extract and normalize URL from lxml Element
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     Args: | 
					
						
							|  |  |  |         * xpath_results (Union[List[html.HtmlElement], html.HtmlElement]): lxml Element(s) | 
					
						
							| 
									
										
										
										
											2020-10-03 10:02:50 +02:00
										 |  |  |         * base_url (str): Base URL | 
					
						
							| 
									
										
										
										
											2020-10-02 18:17:01 +02:00
										 |  |  | 
 | 
					
						
							|  |  |  |     Example: | 
					
						
							|  |  |  |         >>> def f(s, search_url): | 
					
						
							|  |  |  |         >>>    return searx.utils.extract_url(html.fromstring(s), search_url) | 
					
						
							|  |  |  |         >>> f('<span id="42">https://example.com</span>', 'http://example.com/') | 
					
						
							|  |  |  |         'https://example.com/' | 
					
						
							|  |  |  |         >>> f('https://example.com', 'http://example.com/') | 
					
						
							|  |  |  |         'https://example.com/' | 
					
						
							|  |  |  |         >>> f('//example.com', 'http://example.com/') | 
					
						
							|  |  |  |         'http://example.com/' | 
					
						
							|  |  |  |         >>> f('//example.com', 'https://example.com/') | 
					
						
							|  |  |  |         'https://example.com/' | 
					
						
							|  |  |  |         >>> f('/path?a=1', 'https://example.com') | 
					
						
							|  |  |  |         'https://example.com/path?a=1' | 
					
						
							|  |  |  |         >>> f('', 'https://example.com') | 
					
						
							|  |  |  |         raise lxml.etree.ParserError | 
					
						
							|  |  |  |         >>> searx.utils.extract_url([], 'https://example.com') | 
					
						
							| 
									
										
										
										
											2020-11-26 15:12:11 +01:00
										 |  |  |         raise ValueError | 
					
						
							| 
									
										
										
										
											2020-10-02 18:17:01 +02:00
										 |  |  | 
 | 
					
						
							|  |  |  |     Raises: | 
					
						
							| 
									
										
										
										
											2020-11-26 15:12:11 +01:00
										 |  |  |         * ValueError | 
					
						
							| 
									
										
										
										
											2020-10-02 18:17:01 +02:00
										 |  |  |         * lxml.etree.ParserError | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     Returns: | 
					
						
							|  |  |  |         * str: normalized URL | 
					
						
							|  |  |  |     """
 | 
					
						
							| 
									
										
										
										
											2020-10-02 18:13:56 +02:00
										 |  |  |     if xpath_results == []: | 
					
						
							| 
									
										
										
										
											2020-11-26 15:12:11 +01:00
										 |  |  |         raise ValueError('Empty url resultset') | 
					
						
							| 
									
										
										
										
											2020-10-02 18:13:56 +02:00
										 |  |  | 
 | 
					
						
							| 
									
										
										
										
											2020-10-03 10:02:50 +02:00
										 |  |  |     url = extract_text(xpath_results) | 
					
						
							| 
									
										
										
										
											2022-01-30 22:14:12 +01:00
										 |  |  |     if url: | 
					
						
							|  |  |  |         return normalize_url(url, base_url) | 
					
						
							|  |  |  |     raise ValueError('URL not found') | 
					
						
							| 
									
										
										
										
											2020-10-02 18:13:56 +02:00
										 |  |  | 
 | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2022-09-27 17:01:00 +02:00
										 |  |  | def dict_subset(dictionary: MutableMapping, properties: Set[str]) -> Dict: | 
					
						
							| 
									
										
										
										
											2020-10-02 18:17:01 +02:00
										 |  |  |     """Extract a subset of a dict
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     Examples: | 
					
						
							|  |  |  |         >>> dict_subset({'A': 'a', 'B': 'b', 'C': 'c'}, ['A', 'C']) | 
					
						
							|  |  |  |         {'A': 'a', 'C': 'c'} | 
					
						
							|  |  |  |         >>> >> dict_subset({'A': 'a', 'B': 'b', 'C': 'c'}, ['A', 'D']) | 
					
						
							|  |  |  |         {'A': 'a'} | 
					
						
							|  |  |  |     """
 | 
					
						
							| 
									
										
										
										
											2022-09-27 17:01:00 +02:00
										 |  |  |     return {k: dictionary[k] for k in properties if k in dictionary} | 
					
						
							| 
									
										
										
										
											2015-01-29 19:44:52 +01:00
										 |  |  | 
 | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2024-02-20 10:51:58 +01:00
										 |  |  | def humanize_bytes(size, precision=2): | 
					
						
							|  |  |  |     """Determine the *human readable* value of bytes on 1024 base (1KB=1024B).""" | 
					
						
							|  |  |  |     s = ['B ', 'KB', 'MB', 'GB', 'TB'] | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     x = len(s) | 
					
						
							|  |  |  |     p = 0 | 
					
						
							|  |  |  |     while size > 1024 and p < x: | 
					
						
							|  |  |  |         p += 1 | 
					
						
							|  |  |  |         size = size / 1024.0 | 
					
						
							|  |  |  |     return "%.*f %s" % (precision, size, s[p]) | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2024-07-20 21:27:12 +02:00
										 |  |  | def humanize_number(size, precision=0): | 
					
						
							|  |  |  |     """Determine the *human readable* value of a decimal number.""" | 
					
						
							|  |  |  |     s = ['', 'K', 'M', 'B', 'T'] | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     x = len(s) | 
					
						
							|  |  |  |     p = 0 | 
					
						
							|  |  |  |     while size > 1000 and p < x: | 
					
						
							|  |  |  |         p += 1 | 
					
						
							|  |  |  |         size = size / 1000.0 | 
					
						
							|  |  |  |     return "%.*f%s" % (precision, size, s[p]) | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2022-01-29 11:16:28 +01:00
										 |  |  | def convert_str_to_int(number_str: str) -> int: | 
					
						
							| 
									
										
										
										
											2020-10-02 18:17:01 +02:00
										 |  |  |     """Convert number_str to int or 0 if number_str is not a number.""" | 
					
						
							| 
									
										
										
										
											2016-10-11 19:31:42 +02:00
										 |  |  |     if number_str.isdigit(): | 
					
						
							|  |  |  |         return int(number_str) | 
					
						
							| 
									
										
										
										
											2022-01-29 11:16:28 +01:00
										 |  |  |     return 0 | 
					
						
							| 
									
										
										
										
											2016-10-11 19:31:42 +02:00
										 |  |  | 
 | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2024-05-24 01:21:58 +02:00
										 |  |  | def extr(txt: str, begin: str, end: str, default: str = ""): | 
					
						
							|  |  |  |     """Extract the string between ``begin`` and ``end`` from ``txt``
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     :param txt:     String to search in | 
					
						
							|  |  |  |     :param begin:   First string to be searched for | 
					
						
							|  |  |  |     :param end:     Second string to be searched for after ``begin`` | 
					
						
							|  |  |  |     :param default: Default value if one of ``begin`` or ``end`` is not | 
					
						
							|  |  |  |                     found.  Defaults to an empty string. | 
					
						
							|  |  |  |     :return: The string between the two search-strings ``begin`` and ``end``. | 
					
						
							|  |  |  |              If at least one of ``begin`` or ``end`` is not found, the value of | 
					
						
							|  |  |  |              ``default`` is returned. | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     Examples: | 
					
						
							|  |  |  |       >>> extr("abcde", "a", "e") | 
					
						
							|  |  |  |       "bcd" | 
					
						
							|  |  |  |       >>> extr("abcde", "a", "z", deafult="nothing") | 
					
						
							|  |  |  |       "nothing" | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     """
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     # From https://github.com/mikf/gallery-dl/blob/master/gallery_dl/text.py#L129 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     try: | 
					
						
							|  |  |  |         first = txt.index(begin) + len(begin) | 
					
						
							|  |  |  |         return txt[first : txt.index(end, first)] | 
					
						
							|  |  |  |     except ValueError: | 
					
						
							|  |  |  |         return default | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2022-01-29 11:16:28 +01:00
										 |  |  | def int_or_zero(num: Union[List[str], str]) -> int: | 
					
						
							| 
									
										
										
										
											2020-10-02 18:17:01 +02:00
										 |  |  |     """Convert num to int or 0. num can be either a str or a list.
 | 
					
						
							|  |  |  |     If num is a list, the first element is converted to int (or return 0 if the list is empty). | 
					
						
							|  |  |  |     If num is a str, see convert_str_to_int | 
					
						
							|  |  |  |     """
 | 
					
						
							| 
									
										
										
										
											2017-09-04 20:05:04 +02:00
										 |  |  |     if isinstance(num, list): | 
					
						
							|  |  |  |         if len(num) < 1: | 
					
						
							|  |  |  |             return 0 | 
					
						
							|  |  |  |         num = num[0] | 
					
						
							|  |  |  |     return convert_str_to_int(num) | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2022-01-29 11:16:28 +01:00
										 |  |  | def is_valid_lang(lang) -> Optional[Tuple[bool, str, str]]: | 
					
						
							| 
									
										
										
										
											2020-10-02 18:17:01 +02:00
										 |  |  |     """Return language code and name if lang describe a language.
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     Examples: | 
					
						
							|  |  |  |         >>> is_valid_lang('zz') | 
					
						
							| 
									
										
										
										
											2022-01-29 11:16:28 +01:00
										 |  |  |         None | 
					
						
							| 
									
										
										
										
											2020-10-02 18:17:01 +02:00
										 |  |  |         >>> is_valid_lang('uk') | 
					
						
							|  |  |  |         (True, 'uk', 'ukrainian') | 
					
						
							|  |  |  |         >>> is_valid_lang(b'uk') | 
					
						
							|  |  |  |         (True, 'uk', 'ukrainian') | 
					
						
							|  |  |  |         >>> is_valid_lang('en') | 
					
						
							|  |  |  |         (True, 'en', 'english') | 
					
						
							|  |  |  |         >>> searx.utils.is_valid_lang('Español') | 
					
						
							|  |  |  |         (True, 'es', 'spanish') | 
					
						
							|  |  |  |         >>> searx.utils.is_valid_lang('Spanish') | 
					
						
							|  |  |  |         (True, 'es', 'spanish') | 
					
						
							|  |  |  |     """
 | 
					
						
							| 
									
										
										
										
											2020-09-08 16:08:37 +02:00
										 |  |  |     if isinstance(lang, bytes): | 
					
						
							|  |  |  |         lang = lang.decode() | 
					
						
							| 
									
										
										
										
											2021-12-27 09:26:22 +01:00
										 |  |  |     is_abbr = len(lang) == 2 | 
					
						
							| 
									
										
										
										
											2020-09-08 16:08:37 +02:00
										 |  |  |     lang = lang.lower() | 
					
						
							| 
									
										
										
										
											2016-09-06 16:43:48 +02:00
										 |  |  |     if is_abbr: | 
					
						
							| 
									
										
										
										
											2022-10-10 19:31:22 +02:00
										 |  |  |         for l in sxng_locales: | 
					
						
							| 
									
										
										
										
											2019-10-16 14:52:57 +02:00
										 |  |  |             if l[0][:2] == lang: | 
					
						
							| 
									
										
										
										
											2017-06-15 10:51:09 +02:00
										 |  |  |                 return (True, l[0][:2], l[3].lower()) | 
					
						
							| 
									
										
										
										
											2022-01-29 11:16:28 +01:00
										 |  |  |         return None | 
					
						
							| 
									
										
										
										
											2022-10-10 19:31:22 +02:00
										 |  |  |     for l in sxng_locales: | 
					
						
							| 
									
										
										
										
											2022-01-29 11:16:28 +01:00
										 |  |  |         if l[1].lower() == lang or l[3].lower() == lang: | 
					
						
							|  |  |  |             return (True, l[0][:2], l[3].lower()) | 
					
						
							|  |  |  |     return None | 
					
						
							| 
									
										
										
										
											2016-11-19 17:51:19 +01:00
										 |  |  | 
 | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2022-01-30 22:14:12 +01:00
										 |  |  | def load_module(filename: str, module_dir: str) -> types.ModuleType: | 
					
						
							| 
									
										
										
										
											2016-11-19 17:51:19 +01:00
										 |  |  |     modname = splitext(filename)[0] | 
					
						
							| 
									
										
										
										
											2022-01-30 22:14:12 +01:00
										 |  |  |     modpath = join(module_dir, filename) | 
					
						
							| 
									
										
										
										
											2020-10-05 12:50:08 +02:00
										 |  |  |     # and https://docs.python.org/3/library/importlib.html#importing-a-source-file-directly | 
					
						
							| 
									
										
										
										
											2022-01-30 22:14:12 +01:00
										 |  |  |     spec = importlib.util.spec_from_file_location(modname, modpath) | 
					
						
							|  |  |  |     if not spec: | 
					
						
							|  |  |  |         raise ValueError(f"Error loading '{modpath}' module") | 
					
						
							| 
									
										
										
										
											2020-10-05 12:50:08 +02:00
										 |  |  |     module = importlib.util.module_from_spec(spec) | 
					
						
							| 
									
										
										
										
											2022-01-30 22:14:12 +01:00
										 |  |  |     if not spec.loader: | 
					
						
							|  |  |  |         raise ValueError(f"Error loading '{modpath}' module") | 
					
						
							| 
									
										
										
										
											2020-10-05 12:50:08 +02:00
										 |  |  |     spec.loader.exec_module(module) | 
					
						
							| 
									
										
										
										
											2016-11-19 17:51:19 +01:00
										 |  |  |     return module | 
					
						
							| 
									
										
										
										
											2017-07-20 15:44:02 +02:00
										 |  |  | 
 | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2022-01-29 11:16:28 +01:00
										 |  |  | def to_string(obj: Any) -> str: | 
					
						
							| 
									
										
										
										
											2020-10-02 18:17:01 +02:00
										 |  |  |     """Convert obj to its string representation.""" | 
					
						
							| 
									
										
										
										
											2020-08-06 17:42:46 +02:00
										 |  |  |     if isinstance(obj, str): | 
					
						
							| 
									
										
										
										
											2017-12-01 20:45:24 +01:00
										 |  |  |         return obj | 
					
						
							|  |  |  |     if hasattr(obj, '__str__'): | 
					
						
							| 
									
										
										
										
											2022-06-03 15:41:52 +02:00
										 |  |  |         return str(obj) | 
					
						
							| 
									
										
										
										
											2022-01-29 11:16:28 +01:00
										 |  |  |     return repr(obj) | 
					
						
							| 
									
										
										
										
											2019-08-02 13:37:13 +02:00
										 |  |  | 
 | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2022-01-29 11:16:28 +01:00
										 |  |  | def ecma_unescape(string: str) -> str: | 
					
						
							| 
									
										
										
										
											2020-10-02 18:17:01 +02:00
										 |  |  |     """Python implementation of the unescape javascript function
 | 
					
						
							| 
									
										
										
										
											2019-08-02 13:37:13 +02:00
										 |  |  | 
 | 
					
						
							|  |  |  |     https://www.ecma-international.org/ecma-262/6.0/#sec-unescape-string | 
					
						
							|  |  |  |     https://developer.mozilla.org/fr/docs/Web/JavaScript/Reference/Objets_globaux/unescape | 
					
						
							| 
									
										
										
										
											2020-10-02 18:17:01 +02:00
										 |  |  | 
 | 
					
						
							|  |  |  |     Examples: | 
					
						
							|  |  |  |         >>> ecma_unescape('%u5409') | 
					
						
							|  |  |  |         '吉' | 
					
						
							|  |  |  |         >>> ecma_unescape('%20') | 
					
						
							|  |  |  |         ' ' | 
					
						
							|  |  |  |         >>> ecma_unescape('%F3') | 
					
						
							|  |  |  |         'ó' | 
					
						
							| 
									
										
										
										
											2019-08-02 13:37:13 +02:00
										 |  |  |     """
 | 
					
						
							|  |  |  |     # "%u5409" becomes "吉" | 
					
						
							| 
									
										
										
										
											2022-01-29 11:16:28 +01:00
										 |  |  |     string = _ECMA_UNESCAPE4_RE.sub(lambda e: chr(int(e.group(1), 16)), string) | 
					
						
							| 
									
										
										
										
											2019-08-02 13:37:13 +02:00
										 |  |  |     # "%20" becomes " ", "%F3" becomes "ó" | 
					
						
							| 
									
										
										
										
											2022-01-29 11:16:28 +01:00
										 |  |  |     string = _ECMA_UNESCAPE2_RE.sub(lambda e: chr(int(e.group(1), 16)), string) | 
					
						
							|  |  |  |     return string | 
					
						
							| 
									
										
										
										
											2019-09-23 17:14:32 +02:00
										 |  |  | 
 | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2022-01-30 22:14:12 +01:00
										 |  |  | def get_string_replaces_function(replaces: Dict[str, str]) -> Callable[[str], str]: | 
					
						
							| 
									
										
										
										
											2020-10-26 19:25:28 +01:00
										 |  |  |     rep = {re.escape(k): v for k, v in replaces.items()} | 
					
						
							|  |  |  |     pattern = re.compile("|".join(rep.keys())) | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2022-01-29 11:16:28 +01:00
										 |  |  |     def func(text): | 
					
						
							| 
									
										
										
										
											2020-10-26 19:25:28 +01:00
										 |  |  |         return pattern.sub(lambda m: rep[re.escape(m.group(0))], text) | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2022-01-29 11:16:28 +01:00
										 |  |  |     return func | 
					
						
							| 
									
										
										
										
											2020-10-26 19:25:28 +01:00
										 |  |  | 
 | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2022-01-29 11:16:28 +01:00
										 |  |  | def get_engine_from_settings(name: str) -> Dict: | 
					
						
							| 
									
										
										
										
											2019-09-23 17:14:32 +02:00
										 |  |  |     """Return engine configuration from settings.yml of a given engine name""" | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     if 'engines' not in settings: | 
					
						
							|  |  |  |         return {} | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2019-09-30 14:27:13 +02:00
										 |  |  |     for engine in settings['engines']: | 
					
						
							| 
									
										
										
										
											2019-09-23 17:14:32 +02:00
										 |  |  |         if 'name' not in engine: | 
					
						
							|  |  |  |             continue | 
					
						
							|  |  |  |         if name == engine['name']: | 
					
						
							|  |  |  |             return engine | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     return {} | 
					
						
							| 
									
										
										
										
											2019-11-15 09:31:37 +01:00
										 |  |  | 
 | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2022-01-29 11:16:28 +01:00
										 |  |  | def get_xpath(xpath_spec: XPathSpecType) -> XPath: | 
					
						
							| 
									
										
										
										
											2020-10-02 18:17:01 +02:00
										 |  |  |     """Return cached compiled XPath
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     There is no thread lock. | 
					
						
							|  |  |  |     Worst case scenario, xpath_str is compiled more than one time. | 
					
						
							| 
									
										
										
										
											2020-11-26 15:12:11 +01:00
										 |  |  | 
 | 
					
						
							|  |  |  |     Args: | 
					
						
							|  |  |  |         * xpath_spec (str|lxml.etree.XPath): XPath as a str or lxml.etree.XPath | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     Returns: | 
					
						
							|  |  |  |         * result (bool, float, list, str): Results. | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     Raises: | 
					
						
							|  |  |  |         * TypeError: Raise when xpath_spec is neither a str nor a lxml.etree.XPath | 
					
						
							|  |  |  |         * SearxXPathSyntaxException: Raise when there is a syntax error in the XPath | 
					
						
							|  |  |  |     """
 | 
					
						
							|  |  |  |     if isinstance(xpath_spec, str): | 
					
						
							| 
									
										
										
										
											2022-01-29 11:16:28 +01:00
										 |  |  |         result = _XPATH_CACHE.get(xpath_spec, None) | 
					
						
							| 
									
										
										
										
											2020-11-26 15:12:11 +01:00
										 |  |  |         if result is None: | 
					
						
							|  |  |  |             try: | 
					
						
							|  |  |  |                 result = XPath(xpath_spec) | 
					
						
							|  |  |  |             except XPathSyntaxError as e: | 
					
						
							| 
									
										
										
										
											2020-12-17 09:57:57 +01:00
										 |  |  |                 raise SearxXPathSyntaxException(xpath_spec, str(e.msg)) from e | 
					
						
							| 
									
										
										
										
											2022-01-29 11:16:28 +01:00
										 |  |  |             _XPATH_CACHE[xpath_spec] = result | 
					
						
							| 
									
										
										
										
											2020-11-26 15:12:11 +01:00
										 |  |  |         return result | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     if isinstance(xpath_spec, XPath): | 
					
						
							|  |  |  |         return xpath_spec | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     raise TypeError('xpath_spec must be either a str or a lxml.etree.XPath') | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2022-01-29 11:16:28 +01:00
										 |  |  | def eval_xpath(element: ElementBase, xpath_spec: XPathSpecType): | 
					
						
							| 
									
										
										
										
											2020-11-26 15:12:11 +01:00
										 |  |  |     """Equivalent of element.xpath(xpath_str) but compile xpath_str once for all.
 | 
					
						
							|  |  |  |     See https://lxml.de/xpathxslt.html#xpath-return-values | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     Args: | 
					
						
							|  |  |  |         * element (ElementBase): [description] | 
					
						
							|  |  |  |         * xpath_spec (str|lxml.etree.XPath): XPath as a str or lxml.etree.XPath | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     Returns: | 
					
						
							|  |  |  |         * result (bool, float, list, str): Results. | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     Raises: | 
					
						
							|  |  |  |         * TypeError: Raise when xpath_spec is neither a str nor a lxml.etree.XPath | 
					
						
							|  |  |  |         * SearxXPathSyntaxException: Raise when there is a syntax error in the XPath | 
					
						
							|  |  |  |         * SearxEngineXPathException: Raise when the XPath can't be evaluated. | 
					
						
							| 
									
										
										
										
											2020-10-02 18:17:01 +02:00
										 |  |  |     """
 | 
					
						
							| 
									
										
										
										
											2020-11-26 15:12:11 +01:00
										 |  |  |     xpath = get_xpath(xpath_spec) | 
					
						
							|  |  |  |     try: | 
					
						
							|  |  |  |         return xpath(element) | 
					
						
							|  |  |  |     except XPathError as e: | 
					
						
							|  |  |  |         arg = ' '.join([str(i) for i in e.args]) | 
					
						
							| 
									
										
										
										
											2020-12-17 09:57:57 +01:00
										 |  |  |         raise SearxEngineXPathException(xpath_spec, arg) from e | 
					
						
							| 
									
										
										
										
											2020-11-26 15:12:11 +01:00
										 |  |  | 
 | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2022-07-28 13:02:56 +02:00
										 |  |  | def eval_xpath_list(element: ElementBase, xpath_spec: XPathSpecType, min_len: Optional[int] = None): | 
					
						
							| 
									
										
										
										
											2020-11-26 15:12:11 +01:00
										 |  |  |     """Same as eval_xpath, check if the result is a list
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     Args: | 
					
						
							|  |  |  |         * element (ElementBase): [description] | 
					
						
							|  |  |  |         * xpath_spec (str|lxml.etree.XPath): XPath as a str or lxml.etree.XPath | 
					
						
							|  |  |  |         * min_len (int, optional): [description]. Defaults to None. | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     Raises: | 
					
						
							|  |  |  |         * TypeError: Raise when xpath_spec is neither a str nor a lxml.etree.XPath | 
					
						
							|  |  |  |         * SearxXPathSyntaxException: Raise when there is a syntax error in the XPath | 
					
						
							|  |  |  |         * SearxEngineXPathException: raise if the result is not a list | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     Returns: | 
					
						
							|  |  |  |         * result (bool, float, list, str): Results. | 
					
						
							|  |  |  |     """
 | 
					
						
							|  |  |  |     result = eval_xpath(element, xpath_spec) | 
					
						
							|  |  |  |     if not isinstance(result, list): | 
					
						
							|  |  |  |         raise SearxEngineXPathException(xpath_spec, 'the result is not a list') | 
					
						
							|  |  |  |     if min_len is not None and min_len > len(result): | 
					
						
							|  |  |  |         raise SearxEngineXPathException(xpath_spec, 'len(xpath_str) < ' + str(min_len)) | 
					
						
							| 
									
										
										
										
											2019-11-15 09:31:37 +01:00
										 |  |  |     return result | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2022-01-29 11:16:28 +01:00
										 |  |  | def eval_xpath_getindex(elements: ElementBase, xpath_spec: XPathSpecType, index: int, default=_NOTSET): | 
					
						
							| 
									
										
										
										
											2020-11-26 15:12:11 +01:00
										 |  |  |     """Call eval_xpath_list then get one element using the index parameter.
 | 
					
						
							| 
									
										
										
										
											2023-09-15 09:53:03 +02:00
										 |  |  |     If the index does not exist, either raise an exception is default is not set, | 
					
						
							| 
									
										
										
										
											2020-11-26 15:12:11 +01:00
										 |  |  |     other return the default value (can be None). | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     Args: | 
					
						
							|  |  |  |         * elements (ElementBase): lxml element to apply the xpath. | 
					
						
							|  |  |  |         * xpath_spec (str|lxml.etree.XPath): XPath as a str or lxml.etree.XPath. | 
					
						
							|  |  |  |         * index (int): index to get | 
					
						
							|  |  |  |         * default (Object, optional): Defaults if index doesn't exist. | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     Raises: | 
					
						
							|  |  |  |         * TypeError: Raise when xpath_spec is neither a str nor a lxml.etree.XPath | 
					
						
							|  |  |  |         * SearxXPathSyntaxException: Raise when there is a syntax error in the XPath | 
					
						
							|  |  |  |         * SearxEngineXPathException: if the index is not found. Also see eval_xpath. | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     Returns: | 
					
						
							|  |  |  |         * result (bool, float, list, str): Results. | 
					
						
							|  |  |  |     """
 | 
					
						
							|  |  |  |     result = eval_xpath_list(elements, xpath_spec) | 
					
						
							| 
									
										
										
										
											2022-01-29 11:16:28 +01:00
										 |  |  |     if -len(result) <= index < len(result): | 
					
						
							| 
									
										
										
										
											2020-11-26 15:12:11 +01:00
										 |  |  |         return result[index] | 
					
						
							| 
									
										
										
										
											2022-01-29 11:16:28 +01:00
										 |  |  |     if default == _NOTSET: | 
					
						
							| 
									
										
										
										
											2020-11-26 15:12:11 +01:00
										 |  |  |         # raise an SearxEngineXPathException instead of IndexError | 
					
						
							|  |  |  |         # to record xpath_spec | 
					
						
							|  |  |  |         raise SearxEngineXPathException(xpath_spec, 'index ' + str(index) + ' not found') | 
					
						
							|  |  |  |     return default | 
					
						
							| 
									
										
										
										
											2022-12-11 16:45:47 +01:00
										 |  |  | 
 | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2024-03-11 14:06:26 +01:00
										 |  |  | def _get_fasttext_model() -> "fasttext.FastText._FastText":  # type: ignore | 
					
						
							| 
									
										
										
										
											2022-12-11 16:45:47 +01:00
										 |  |  |     global _FASTTEXT_MODEL  # pylint: disable=global-statement | 
					
						
							|  |  |  |     if _FASTTEXT_MODEL is None: | 
					
						
							| 
									
										
										
										
											2022-12-26 09:32:58 +01:00
										 |  |  |         import fasttext  # pylint: disable=import-outside-toplevel | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |         # Monkey patch: prevent fasttext from showing a (useless) warning when loading a model. | 
					
						
							|  |  |  |         fasttext.FastText.eprint = lambda x: None | 
					
						
							| 
									
										
										
										
											2022-12-11 16:45:47 +01:00
										 |  |  |         _FASTTEXT_MODEL = fasttext.load_model(str(data_dir / 'lid.176.ftz')) | 
					
						
							|  |  |  |     return _FASTTEXT_MODEL | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2024-09-15 00:28:35 +02:00
										 |  |  | def get_embeded_stream_url(url): | 
					
						
							|  |  |  |     """
 | 
					
						
							|  |  |  |     Converts a standard video URL into its embed format. Supported services include Youtube, | 
					
						
							|  |  |  |     Facebook, Instagram, TikTok, and Dailymotion. | 
					
						
							|  |  |  |     """
 | 
					
						
							|  |  |  |     parsed_url = urlparse(url) | 
					
						
							|  |  |  |     iframe_src = None | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     # YouTube | 
					
						
							|  |  |  |     if parsed_url.netloc in ['www.youtube.com', 'youtube.com'] and parsed_url.path == '/watch' and parsed_url.query: | 
					
						
							|  |  |  |         video_id = parse_qs(parsed_url.query).get('v', []) | 
					
						
							|  |  |  |         if video_id: | 
					
						
							|  |  |  |             iframe_src = 'https://www.youtube-nocookie.com/embed/' + video_id[0] | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     # Facebook | 
					
						
							|  |  |  |     elif parsed_url.netloc in ['www.facebook.com', 'facebook.com']: | 
					
						
							|  |  |  |         encoded_href = urlencode({'href': url}) | 
					
						
							|  |  |  |         iframe_src = 'https://www.facebook.com/plugins/video.php?allowfullscreen=true&' + encoded_href | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     # Instagram | 
					
						
							|  |  |  |     elif parsed_url.netloc in ['www.instagram.com', 'instagram.com'] and parsed_url.path.startswith('/p/'): | 
					
						
							|  |  |  |         if parsed_url.path.endswith('/'): | 
					
						
							|  |  |  |             iframe_src = url + 'embed' | 
					
						
							|  |  |  |         else: | 
					
						
							|  |  |  |             iframe_src = url + '/embed' | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     # TikTok | 
					
						
							|  |  |  |     elif ( | 
					
						
							|  |  |  |         parsed_url.netloc in ['www.tiktok.com', 'tiktok.com'] | 
					
						
							|  |  |  |         and parsed_url.path.startswith('/@') | 
					
						
							|  |  |  |         and '/video/' in parsed_url.path | 
					
						
							|  |  |  |     ): | 
					
						
							|  |  |  |         path_parts = parsed_url.path.split('/video/') | 
					
						
							|  |  |  |         video_id = path_parts[1] | 
					
						
							|  |  |  |         iframe_src = 'https://www.tiktok.com/embed/' + video_id | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     # Dailymotion | 
					
						
							|  |  |  |     elif parsed_url.netloc in ['www.dailymotion.com', 'dailymotion.com'] and parsed_url.path.startswith('/video/'): | 
					
						
							|  |  |  |         path_parts = parsed_url.path.split('/') | 
					
						
							|  |  |  |         if len(path_parts) == 3: | 
					
						
							|  |  |  |             video_id = path_parts[2] | 
					
						
							|  |  |  |             iframe_src = 'https://www.dailymotion.com/embed/video/' + video_id | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     return iframe_src | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2022-12-16 21:28:57 +01:00
										 |  |  | def detect_language(text: str, threshold: float = 0.3, only_search_languages: bool = False) -> Optional[str]: | 
					
						
							| 
									
										
										
										
											2023-01-30 08:53:48 +01:00
										 |  |  |     """Detect the language of the ``text`` parameter.
 | 
					
						
							| 
									
										
										
										
											2022-12-16 21:28:57 +01:00
										 |  |  | 
 | 
					
						
							| 
									
										
										
										
											2023-01-30 08:53:48 +01:00
										 |  |  |     :param str text: The string whose language is to be detected. | 
					
						
							| 
									
										
										
										
											2022-12-16 21:28:57 +01:00
										 |  |  | 
 | 
					
						
							| 
									
										
										
										
											2023-01-30 08:53:48 +01:00
										 |  |  |     :param float threshold: Threshold filters the returned labels by a threshold | 
					
						
							|  |  |  |         on probability.  A choice of 0.3 will return labels with at least 0.3 | 
					
						
							|  |  |  |         probability. | 
					
						
							| 
									
										
										
										
											2022-12-16 21:28:57 +01:00
										 |  |  | 
 | 
					
						
							| 
									
										
										
										
											2023-01-30 08:53:48 +01:00
										 |  |  |     :param bool only_search_languages: If ``True``, returns only supported | 
					
						
							|  |  |  |         SearXNG search languages.  see :py:obj:`searx.languages` | 
					
						
							| 
									
										
										
										
											2022-12-16 21:28:57 +01:00
										 |  |  | 
 | 
					
						
							| 
									
										
										
										
											2023-01-30 08:53:48 +01:00
										 |  |  |     :rtype: str, None | 
					
						
							|  |  |  |     :returns: | 
					
						
							|  |  |  |         The detected language code or ``None``. See below. | 
					
						
							| 
									
										
										
										
											2022-12-16 21:28:57 +01:00
										 |  |  | 
 | 
					
						
							| 
									
										
										
										
											2023-01-30 08:53:48 +01:00
										 |  |  |     :raises ValueError: If ``text`` is not a string. | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     The language detection is done by using `a fork`_ of the fastText_ library | 
					
						
							|  |  |  |     (`python fasttext`_). fastText_ distributes the `language identification | 
					
						
							|  |  |  |     model`_, for reference: | 
					
						
							| 
									
										
										
										
											2022-12-16 21:28:57 +01:00
										 |  |  | 
 | 
					
						
							|  |  |  |     - `FastText.zip: Compressing text classification models`_ | 
					
						
							|  |  |  |     - `Bag of Tricks for Efficient Text Classification`_ | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2023-01-30 08:53:48 +01:00
										 |  |  |     The `language identification model`_ support the language codes | 
					
						
							|  |  |  |     (ISO-639-3):: | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |         af als am an ar arz as ast av az azb ba bar bcl be bg bh bn bo bpy br bs | 
					
						
							|  |  |  |         bxr ca cbk ce ceb ckb co cs cv cy da de diq dsb dty dv el eml en eo es | 
					
						
							|  |  |  |         et eu fa fi fr frr fy ga gd gl gn gom gu gv he hi hif hr hsb ht hu hy ia | 
					
						
							|  |  |  |         id ie ilo io is it ja jbo jv ka kk km kn ko krc ku kv kw ky la lb lez li | 
					
						
							|  |  |  |         lmo lo lrc lt lv mai mg mhr min mk ml mn mr mrj ms mt mwl my myv mzn nah | 
					
						
							|  |  |  |         nap nds ne new nl nn no oc or os pa pam pfl pl pms pnb ps pt qu rm ro ru | 
					
						
							|  |  |  |         rue sa sah sc scn sco sd sh si sk sl so sq sr su sv sw ta te tg th tk tl | 
					
						
							|  |  |  |         tr tt tyv ug uk ur uz vec vep vi vls vo wa war wuu xal xmf yi yo yue zh | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     By using ``only_search_languages=True`` the `language identification model`_ | 
					
						
							|  |  |  |     is harmonized with the SearXNG's language (locale) model.  General | 
					
						
							|  |  |  |     conditions of SearXNG's locale model are: | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     a. SearXNG's locale of a query is passed to the | 
					
						
							|  |  |  |        :py:obj:`searx.locales.get_engine_locale` to get a language and/or region | 
					
						
							|  |  |  |        code that is used by an engine. | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     b. Most of SearXNG's engines do not support all the languages from `language | 
					
						
							|  |  |  |        identification model`_ and there is also a discrepancy in the ISO-639-3 | 
					
						
							| 
									
										
										
										
											2023-09-15 09:53:03 +02:00
										 |  |  |        (fasttext) and ISO-639-2 (SearXNG)handling.  Further more, in SearXNG the | 
					
						
							| 
									
										
										
										
											2023-01-30 08:53:48 +01:00
										 |  |  |        locales like ``zh-TH`` (``zh-CN``) are mapped to ``zh_Hant`` | 
					
						
							|  |  |  |        (``zh_Hans``) while the `language identification model`_ reduce both to | 
					
						
							|  |  |  |        ``zh``. | 
					
						
							| 
									
										
										
										
											2022-12-16 21:28:57 +01:00
										 |  |  | 
 | 
					
						
							|  |  |  |     .. _a fork: https://github.com/searxng/fasttext-predict | 
					
						
							|  |  |  |     .. _fastText: https://fasttext.cc/ | 
					
						
							|  |  |  |     .. _python fasttext: https://pypi.org/project/fasttext/ | 
					
						
							|  |  |  |     .. _language identification model: https://fasttext.cc/docs/en/language-identification.html | 
					
						
							|  |  |  |     .. _Bag of Tricks for Efficient Text Classification: https://arxiv.org/abs/1607.01759 | 
					
						
							|  |  |  |     .. _`FastText.zip: Compressing text classification models`: https://arxiv.org/abs/1612.03651 | 
					
						
							| 
									
										
										
										
											2023-01-30 08:53:48 +01:00
										 |  |  | 
 | 
					
						
							| 
									
										
										
										
											2022-12-16 21:28:57 +01:00
										 |  |  |     """
 | 
					
						
							| 
									
										
										
										
											2022-12-11 16:45:47 +01:00
										 |  |  |     if not isinstance(text, str): | 
					
						
							|  |  |  |         raise ValueError('text must a str') | 
					
						
							|  |  |  |     r = _get_fasttext_model().predict(text.replace('\n', ' '), k=1, threshold=threshold) | 
					
						
							| 
									
										
										
										
											2022-12-16 21:28:57 +01:00
										 |  |  |     if isinstance(r, tuple) and len(r) == 2 and len(r[0]) > 0 and len(r[1]) > 0: | 
					
						
							|  |  |  |         language = r[0][0].split('__label__')[1] | 
					
						
							|  |  |  |         if only_search_languages and language not in SEARCH_LANGUAGE_CODES: | 
					
						
							|  |  |  |             return None | 
					
						
							|  |  |  |         return language | 
					
						
							| 
									
										
										
										
											2022-12-11 16:45:47 +01:00
										 |  |  |     return None | 
					
						
							| 
									
										
										
										
											2023-09-09 12:18:39 +02:00
										 |  |  | 
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | def js_variable_to_python(js_variable): | 
					
						
							|  |  |  |     """Convert a javascript variable into JSON and then load the value
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     It does not deal with all cases, but it is good enough for now. | 
					
						
							|  |  |  |     chompjs has a better implementation. | 
					
						
							|  |  |  |     """
 | 
					
						
							|  |  |  |     # when in_string is not None, it contains the character that has opened the string | 
					
						
							|  |  |  |     # either simple quote or double quote | 
					
						
							|  |  |  |     in_string = None | 
					
						
							|  |  |  |     # cut the string: | 
					
						
							|  |  |  |     # r"""{ a:"f\"irst", c:'sec"ond'}""" | 
					
						
							|  |  |  |     # becomes | 
					
						
							|  |  |  |     # ['{ a:', '"', 'f\\', '"', 'irst', '"', ', c:', "'", 'sec', '"', 'ond', "'", '}'] | 
					
						
							|  |  |  |     parts = re.split(r'(["\'])', js_variable) | 
					
						
							|  |  |  |     # previous part (to check the escape character antislash) | 
					
						
							|  |  |  |     previous_p = "" | 
					
						
							|  |  |  |     for i, p in enumerate(parts): | 
					
						
							|  |  |  |         # parse characters inside a ECMA string | 
					
						
							|  |  |  |         if in_string: | 
					
						
							|  |  |  |             # we are in a JS string: replace the colon by a temporary character | 
					
						
							|  |  |  |             # so quote_keys_regex doesn't have to deal with colon inside the JS strings | 
					
						
							|  |  |  |             parts[i] = parts[i].replace(':', chr(1)) | 
					
						
							|  |  |  |             if in_string == "'": | 
					
						
							|  |  |  |                 # the JS string is delimited by simple quote. | 
					
						
							|  |  |  |                 # This is not supported by JSON. | 
					
						
							|  |  |  |                 # simple quote delimited string are converted to double quote delimited string | 
					
						
							|  |  |  |                 # here, inside a JS string, we escape the double quote | 
					
						
							|  |  |  |                 parts[i] = parts[i].replace('"', r'\"') | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2023-09-15 09:53:03 +02:00
										 |  |  |         # deal with delimiters and escape character | 
					
						
							| 
									
										
										
										
											2023-09-09 12:18:39 +02:00
										 |  |  |         if not in_string and p in ('"', "'"): | 
					
						
							|  |  |  |             # we are not in string | 
					
						
							|  |  |  |             # but p is double or simple quote | 
					
						
							|  |  |  |             # that's the start of a new string | 
					
						
							|  |  |  |             # replace simple quote by double quote | 
					
						
							|  |  |  |             # (JSON doesn't support simple quote) | 
					
						
							|  |  |  |             parts[i] = '"' | 
					
						
							|  |  |  |             in_string = p | 
					
						
							|  |  |  |             continue | 
					
						
							|  |  |  |         if p == in_string: | 
					
						
							|  |  |  |             # we are in a string and the current part MAY close the string | 
					
						
							|  |  |  |             if len(previous_p) > 0 and previous_p[-1] == '\\': | 
					
						
							|  |  |  |                 # there is an antislash just before: the ECMA string continue | 
					
						
							|  |  |  |                 continue | 
					
						
							|  |  |  |             # the current p close the string | 
					
						
							|  |  |  |             # replace simple quote by double quote | 
					
						
							|  |  |  |             parts[i] = '"' | 
					
						
							|  |  |  |             in_string = None | 
					
						
							| 
									
										
										
										
											2023-09-15 20:57:03 +02:00
										 |  |  | 
 | 
					
						
							| 
									
										
										
										
											2023-09-09 12:18:39 +02:00
										 |  |  |         if not in_string: | 
					
						
							|  |  |  |             # replace void 0 by null | 
					
						
							|  |  |  |             # https://developer.mozilla.org/en-US/docs/Web/JavaScript/Reference/Operators/void | 
					
						
							|  |  |  |             # we are sure there is no string in p | 
					
						
							|  |  |  |             parts[i] = _JS_VOID_RE.sub("null", p) | 
					
						
							|  |  |  |         # update previous_p | 
					
						
							|  |  |  |         previous_p = p | 
					
						
							|  |  |  |     # join the string | 
					
						
							|  |  |  |     s = ''.join(parts) | 
					
						
							| 
									
										
										
										
											2023-09-15 09:53:03 +02:00
										 |  |  |     # add quote around the key | 
					
						
							| 
									
										
										
										
											2023-09-09 12:18:39 +02:00
										 |  |  |     # { a: 12 } | 
					
						
							|  |  |  |     # becomes | 
					
						
							|  |  |  |     # { "a": 12 } | 
					
						
							|  |  |  |     s = _JS_QUOTE_KEYS_RE.sub(r'\1"\2"\3', s) | 
					
						
							| 
									
										
										
										
											2023-09-15 20:57:03 +02:00
										 |  |  |     s = _JS_DECIMAL_RE.sub(":0.", s) | 
					
						
							| 
									
										
										
										
											2023-09-09 12:18:39 +02:00
										 |  |  |     # replace the surogate character by colon | 
					
						
							|  |  |  |     s = s.replace(chr(1), ':') | 
					
						
							|  |  |  |     # load the JSON and return the result | 
					
						
							|  |  |  |     return json.loads(s) |