2022-04-13 20:29:07 +03:00
|
|
|
from app.models.config import Config
|
Improve public instance session management (#480)
This introduces a new approach to handling user sessions, which should
allow for users to set more reliable config settings on public instances.
Previously, when a user with cookies disabled would update their config,
this would modify the app's default config file, which would in turn
cause new users to inherit these settings when visiting the app for the
first time and cause users to inherit these settings when their current
session cookie expired (which was after 30 days by default I believe).
There was also some half-baked logic for determining on the backend
whether or not a user had cookies disabled, which lead to some issues
with out of control session file creation by Flask.
Now, when a user visits the site, their initial request is forwarded to
a session/<session id> endpoint, and during that subsequent request
their current session id is matched against the one found in the url. If
the ids match, the user has cookies enabled. If not, their original
request is modified with a 'cookies_disabled' query param that tells
Flask not to bother trying to set up a new session for that user, and
instead just use the app's fallback Fernet key for encryption and the
default config.
Since attempting to create a session for a user with cookies disabled
creates a new session file, there is now also a clean-up routine included
in the new session decorator, which will remove all sessions that don't
include a valid key in the dict. NOTE!!! This means that current user
sessions on public instances will be cleared once this update is merged
in. In the long run that's a good thing though, since this will allow session
mgmt to be a lot more reliable overall for users regardless of their cookie
preference.
Individual user sessions still use a unique Fernet key for encrypting queries,
but users with cookies disabled will use the default app key for encryption
and decryption.
Sessions are also now (semi)permanent and have a lifetime of 1 year.
2021-11-18 05:35:30 +03:00
|
|
|
from app.models.endpoint import Endpoint
|
2023-05-19 21:15:15 +03:00
|
|
|
from app.utils.misc import list_to_dict
|
2023-03-22 21:29:05 +03:00
|
|
|
from bs4 import BeautifulSoup, NavigableString
|
2022-02-07 20:47:25 +03:00
|
|
|
import copy
|
2022-04-13 20:29:07 +03:00
|
|
|
from flask import current_app
|
2021-11-02 00:34:59 +03:00
|
|
|
import html
|
2020-12-06 01:01:21 +03:00
|
|
|
import os
|
2020-07-26 20:53:59 +03:00
|
|
|
import urllib.parse as urlparse
|
|
|
|
from urllib.parse import parse_qs
|
2021-10-26 23:59:23 +03:00
|
|
|
import re
|
2023-03-07 21:28:55 +03:00
|
|
|
import warnings
|
|
|
|
|
2020-07-26 20:53:59 +03:00
|
|
|
SKIP_ARGS = ['ref_src', 'utm']
|
2023-11-01 23:07:45 +03:00
|
|
|
SKIP_PREFIX = ['//www.', '//mobile.', '//m.']
|
2021-02-20 23:04:32 +03:00
|
|
|
GOOG_STATIC = 'www.gstatic.com'
|
2022-04-19 00:27:45 +03:00
|
|
|
G_M_LOGO_URL = 'https://www.gstatic.com/m/images/icons/googleg.gif'
|
2020-07-26 20:53:59 +03:00
|
|
|
GOOG_IMG = '/images/branding/searchlogo/1x/googlelogo'
|
|
|
|
LOGO_URL = GOOG_IMG + '_desk'
|
2020-12-18 00:06:47 +03:00
|
|
|
BLANK_B64 = ('data:image/png;base64,'
|
|
|
|
'iVBORw0KGgoAAAANSUhEUgAAAAoAAAAKCAQAAAAnOwc2AAAAD0lEQVR42mNkw'
|
|
|
|
'AIYh7IgAAVVAAuInjI5AAAAAElFTkSuQmCC')
|
2020-07-26 20:53:59 +03:00
|
|
|
|
2020-12-18 00:06:47 +03:00
|
|
|
# Ad keywords
|
2020-07-26 20:53:59 +03:00
|
|
|
BLACKLIST = [
|
2022-02-26 09:02:58 +03:00
|
|
|
'ad', 'ads', 'anuncio', 'annuncio', 'annonce', 'Anzeige', '广告', '廣告',
|
|
|
|
'Reklama', 'Реклама', 'Anunț', '광고', 'annons', 'Annonse', 'Iklan',
|
|
|
|
'広告', 'Augl.', 'Mainos', 'Advertentie', 'إعلان', 'Գովազդ', 'विज्ञापन',
|
|
|
|
'Reklam', 'آگهی', 'Reklāma', 'Reklaam', 'Διαφήμιση', 'מודעה', 'Hirdetés',
|
2022-11-28 20:17:16 +03:00
|
|
|
'Anúncio', 'Quảng cáo','โฆษณา', 'sponsored', 'patrocinado', 'gesponsert'
|
2020-07-26 20:53:59 +03:00
|
|
|
]
|
|
|
|
|
|
|
|
SITE_ALTS = {
|
Use farside.link for frontend alternatives in results (#560)
* Integrate Farside into Whoogle
When instances are ratelimited (when a captcha is returned instead of
the user's search results) the user can now hop to a new instance via
Farside, a new backend service that redirects users to working instances
of a particular frontend. In this case, it presents a user with a
Farside link to a new Whoogle (or Searx) instance instead, so that the
user can resume their search.
For the generated Farside->Whoogle link, the generated link includes the
user's current Whoogle configuration settings as URL params, to ensure a
more seamless transition between instances. This doesn't translate to
the Farside->Searx link, but potentially could with some changes.
* Expand conversion of config<->url params
Config settings can now be translated to and from URL params using a
predetermined set of "safe" keys (i.e. config settings that easily
translate to URL params).
* Allow jumping instances via Farside when ratelimited
When instances are ratelimited (when a captcha is returned instead of
the user's search results) the user can now hop to a new instance via
Farside, a new backend service that redirects users to working instances
of a particular frontend. In this case, it presents a user with a
Farside link to a new Whoogle (or Searx) instance instead, so that the
user can resume their search.
For the generated Farside->Whoogle link, the generated link includes the
user's current Whoogle configuration settings as URL params, to ensure a
more seamless transition between instances. This doesn't translate to
the Farside->Searx link, but potentially could with some changes.
Closes #554
Closes #559
2021-12-09 03:27:33 +03:00
|
|
|
'twitter.com': os.getenv('WHOOGLE_ALT_TW', 'farside.link/nitter'),
|
|
|
|
'youtube.com': os.getenv('WHOOGLE_ALT_YT', 'farside.link/invidious'),
|
|
|
|
'reddit.com': os.getenv('WHOOGLE_ALT_RD', 'farside.link/libreddit'),
|
2021-10-24 08:23:37 +03:00
|
|
|
**dict.fromkeys([
|
2022-12-05 23:28:29 +03:00
|
|
|
'medium.com',
|
2021-10-24 08:23:37 +03:00
|
|
|
'levelup.gitconnected.com'
|
2022-01-14 19:59:03 +03:00
|
|
|
], os.getenv('WHOOGLE_ALT_MD', 'farside.link/scribe')),
|
2022-03-21 19:08:52 +03:00
|
|
|
'imgur.com': os.getenv('WHOOGLE_ALT_IMG', 'farside.link/rimgo'),
|
2022-08-01 20:49:09 +03:00
|
|
|
'wikipedia.org': os.getenv('WHOOGLE_ALT_WIKI', 'farside.link/wikiless'),
|
|
|
|
'imdb.com': os.getenv('WHOOGLE_ALT_IMDB', 'farside.link/libremdb'),
|
|
|
|
'quora.com': os.getenv('WHOOGLE_ALT_QUORA', 'farside.link/quetre')
|
2020-07-26 20:53:59 +03:00
|
|
|
}
|
|
|
|
|
2023-05-19 21:15:15 +03:00
|
|
|
# Include custom site redirects from WHOOGLE_REDIRECTS
|
|
|
|
SITE_ALTS.update(list_to_dict(re.split(',|:', os.getenv('WHOOGLE_REDIRECTS', ''))))
|
|
|
|
|
2020-07-26 20:53:59 +03:00
|
|
|
|
2023-01-09 22:54:41 +03:00
|
|
|
def contains_cjko(s: str) -> bool:
|
|
|
|
"""This function check whether or not a string contains Chinese, Japanese,
|
|
|
|
or Korean characters. It employs regex and uses the u escape sequence to
|
|
|
|
match any character in a set of Unicode ranges.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
s (str): string to be checked
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
bool: True if the input s contains the characters and False otherwise
|
|
|
|
"""
|
|
|
|
unicode_ranges = ('\u4e00-\u9fff' # Chinese characters
|
|
|
|
'\u3040-\u309f' # Japanese hiragana
|
|
|
|
'\u30a0-\u30ff' # Japanese katakana
|
|
|
|
'\u4e00-\u9faf' # Japanese kanji
|
|
|
|
'\uac00-\ud7af' # Korean hangul syllables
|
|
|
|
'\u1100-\u11ff' # Korean hangul jamo
|
|
|
|
)
|
|
|
|
return bool(re.search(fr'[{unicode_ranges}]', s))
|
|
|
|
|
|
|
|
|
2021-10-26 23:59:23 +03:00
|
|
|
def bold_search_terms(response: str, query: str) -> BeautifulSoup:
|
|
|
|
"""Wraps all search terms in bold tags (<b>). If any terms are wrapped
|
|
|
|
in quotes, only that exact phrase will be made bold.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
response: The initial response body for the query
|
|
|
|
query: The original search query
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
BeautifulSoup: modified soup object with bold items
|
|
|
|
"""
|
|
|
|
response = BeautifulSoup(response, 'html.parser')
|
|
|
|
|
|
|
|
def replace_any_case(element: NavigableString, target_word: str) -> None:
|
|
|
|
# Replace all instances of the word, but maintaining the same case in
|
|
|
|
# the replacement
|
2021-10-27 01:15:24 +03:00
|
|
|
if len(element) == len(target_word):
|
|
|
|
return
|
|
|
|
|
2022-10-31 21:45:44 +03:00
|
|
|
# Ensure target word is escaped for regex
|
|
|
|
target_word = re.escape(target_word)
|
|
|
|
|
2023-01-09 22:54:41 +03:00
|
|
|
# Check if the word contains Chinese, Japanese, or Korean characters
|
|
|
|
if contains_cjko(target_word):
|
|
|
|
reg_pattern = fr'((?![{{}}<>-]){target_word}(?![{{}}<>-]))'
|
|
|
|
else:
|
|
|
|
reg_pattern = fr'\b((?![{{}}<>-]){target_word}(?![{{}}<>-]))\b'
|
|
|
|
|
2022-10-26 19:23:39 +03:00
|
|
|
if re.match('.*[@_!#$%^&*()<>?/\|}{~:].*', target_word) or (
|
2021-10-28 21:54:27 +03:00
|
|
|
element.parent and element.parent.name == 'style'):
|
2021-10-27 19:50:21 +03:00
|
|
|
return
|
|
|
|
|
2021-11-02 00:34:59 +03:00
|
|
|
element.replace_with(BeautifulSoup(
|
2023-01-09 22:54:41 +03:00
|
|
|
re.sub(reg_pattern,
|
2021-10-27 01:17:38 +03:00
|
|
|
r'<b>\1</b>',
|
2022-11-23 03:34:25 +03:00
|
|
|
element,
|
2021-11-02 00:34:59 +03:00
|
|
|
flags=re.I), 'html.parser')
|
2021-10-26 23:59:23 +03:00
|
|
|
)
|
|
|
|
|
|
|
|
# Split all words out of query, grouping the ones wrapped in quotes
|
|
|
|
for word in re.split(r'\s+(?=[^"]*(?:"[^"]*"[^"]*)*$)', query):
|
2022-10-26 19:23:39 +03:00
|
|
|
word = re.sub(r'[@_!#$%^&*()<>?/\|}{~:]+', '', word)
|
2021-10-26 23:59:23 +03:00
|
|
|
target = response.find_all(
|
|
|
|
text=re.compile(r'' + re.escape(word), re.I))
|
|
|
|
for nav_str in target:
|
|
|
|
replace_any_case(nav_str, word)
|
|
|
|
|
|
|
|
return response
|
|
|
|
|
|
|
|
|
2021-03-08 20:22:04 +03:00
|
|
|
def has_ad_content(element: str) -> bool:
|
|
|
|
"""Inspects an HTML element for ad related content
|
|
|
|
|
|
|
|
Args:
|
|
|
|
element: The HTML element to inspect
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
bool: True/False for the element containing an ad
|
|
|
|
|
|
|
|
"""
|
2022-02-26 09:02:58 +03:00
|
|
|
element_str = ''.join(filter(str.isalpha, element))
|
|
|
|
return (element_str.upper() in (value.upper() for value in BLACKLIST)
|
2021-03-08 20:38:40 +03:00
|
|
|
or 'ⓘ' in element)
|
2020-07-26 20:53:59 +03:00
|
|
|
|
|
|
|
|
2021-03-08 20:22:04 +03:00
|
|
|
def get_first_link(soup: BeautifulSoup) -> str:
|
|
|
|
"""Retrieves the first result link from the query response
|
|
|
|
|
|
|
|
Args:
|
|
|
|
soup: The BeautifulSoup response body
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
str: A str link to the first result
|
|
|
|
|
|
|
|
"""
|
2024-04-19 21:40:06 +03:00
|
|
|
first_link = ''
|
|
|
|
orig_details = []
|
|
|
|
|
|
|
|
# Temporarily remove details so we don't grab those links
|
|
|
|
for details in soup.find_all('details'):
|
|
|
|
temp_details = soup.new_tag('removed_details')
|
|
|
|
orig_details.append(details.replace_with(temp_details))
|
|
|
|
|
2020-07-26 20:53:59 +03:00
|
|
|
# Replace hrefs with only the intended destination (no "utm" type tags)
|
|
|
|
for a in soup.find_all('a', href=True):
|
|
|
|
# Return the first search result URL
|
2024-04-19 21:40:06 +03:00
|
|
|
if a['href'].startswith('http://') or a['href'].startswith('https://'):
|
|
|
|
first_link = a['href']
|
|
|
|
break
|
|
|
|
|
|
|
|
# Add the details back
|
|
|
|
for orig_detail, details in zip(orig_details, soup.find_all('removed_details')):
|
|
|
|
details.replace_with(orig_detail)
|
|
|
|
|
|
|
|
return first_link
|
2020-07-26 20:53:59 +03:00
|
|
|
|
|
|
|
|
2023-11-01 23:07:45 +03:00
|
|
|
def get_site_alt(link: str, site_alts: dict = SITE_ALTS) -> str:
|
2021-03-08 20:22:04 +03:00
|
|
|
"""Returns an alternative to a particular site, if one is configured
|
|
|
|
|
|
|
|
Args:
|
2023-11-01 23:07:45 +03:00
|
|
|
link: A string result URL to check against the site_alts map
|
|
|
|
site_alts: A map of site alternatives to replace with. defaults to SITE_ALTS
|
2021-03-08 20:22:04 +03:00
|
|
|
|
|
|
|
Returns:
|
|
|
|
str: An updated (or ignored) result link
|
|
|
|
|
|
|
|
"""
|
2021-10-16 21:22:00 +03:00
|
|
|
# Need to replace full hostname with alternative to encapsulate
|
|
|
|
# subdomains as well
|
2022-07-13 19:25:51 +03:00
|
|
|
parsed_link = urlparse.urlparse(link)
|
2022-12-08 20:54:21 +03:00
|
|
|
|
|
|
|
# Extract subdomain separately from the domain+tld. The subdomain
|
|
|
|
# is used for wikiless translations.
|
|
|
|
split_host = parsed_link.netloc.split('.')
|
|
|
|
subdomain = split_host[0] if len(split_host) > 2 else ''
|
|
|
|
hostname = '.'.join(split_host[-2:])
|
2021-03-28 20:27:08 +03:00
|
|
|
|
2022-11-23 03:34:25 +03:00
|
|
|
# The full scheme + hostname is used when comparing against the list of
|
|
|
|
# available alternative services, due to how Medium links are constructed.
|
|
|
|
# (i.e. for medium.com: "https://something.medium.com" should match,
|
|
|
|
# "https://medium.com/..." should match, but "philomedium.com" should not)
|
|
|
|
hostcomp = f'{parsed_link.scheme}://{hostname}'
|
|
|
|
|
2023-11-01 23:07:45 +03:00
|
|
|
for site_key in site_alts.keys():
|
2022-12-05 23:28:29 +03:00
|
|
|
site_alt = f'{parsed_link.scheme}://{site_key}'
|
2023-11-01 23:07:45 +03:00
|
|
|
if not hostname or site_alt not in hostcomp or not site_alts[site_key]:
|
2020-07-26 20:53:59 +03:00
|
|
|
continue
|
|
|
|
|
2022-07-06 18:49:43 +03:00
|
|
|
# Wikipedia -> Wikiless replacements require the subdomain (if it's
|
|
|
|
# a 2-char language code) to be passed as a URL param to Wikiless
|
|
|
|
# in order to preserve the language setting.
|
2022-07-13 19:25:51 +03:00
|
|
|
params = ''
|
2022-12-08 20:54:21 +03:00
|
|
|
if 'wikipedia' in hostname and len(subdomain) == 2:
|
|
|
|
hostname = f'{subdomain}.{hostname}'
|
|
|
|
params = f'?lang={subdomain}'
|
2023-02-05 02:36:16 +03:00
|
|
|
elif 'medium' in hostname and len(subdomain) > 0:
|
|
|
|
hostname = f'{subdomain}.{hostname}'
|
2022-07-13 19:25:51 +03:00
|
|
|
|
2023-11-01 23:07:45 +03:00
|
|
|
parsed_alt = urlparse.urlparse(site_alts[site_key])
|
|
|
|
link = link.replace(hostname, site_alts[site_key]) + params
|
2022-07-13 19:31:55 +03:00
|
|
|
# If a scheme is specified in the alternative, this results in a
|
|
|
|
# replaced link that looks like "https://http://altservice.tld".
|
|
|
|
# In this case, we can remove the original scheme from the result
|
|
|
|
# and use the one specified for the alt.
|
2022-07-13 19:25:51 +03:00
|
|
|
if parsed_alt.scheme:
|
|
|
|
link = '//'.join(link.split('//')[1:])
|
2022-07-06 18:49:43 +03:00
|
|
|
|
2021-10-11 23:25:21 +03:00
|
|
|
for prefix in SKIP_PREFIX:
|
2023-01-04 20:10:32 +03:00
|
|
|
if parsed_alt.scheme:
|
2023-11-01 23:07:45 +03:00
|
|
|
# If a scheme is specified, remove everything before the
|
|
|
|
# first occurence of it
|
|
|
|
link = f'{parsed_alt.scheme}{link.split(parsed_alt.scheme, 1)[-1]}'
|
2023-01-04 20:10:32 +03:00
|
|
|
else:
|
2023-11-01 23:07:45 +03:00
|
|
|
# Otherwise, replace the first occurrence of the prefix
|
|
|
|
link = link.replace(prefix, '//', 1)
|
2020-07-26 20:53:59 +03:00
|
|
|
break
|
|
|
|
|
2021-01-24 01:43:53 +03:00
|
|
|
return link
|
2020-07-26 20:53:59 +03:00
|
|
|
|
|
|
|
|
2021-03-08 20:22:04 +03:00
|
|
|
def filter_link_args(link: str) -> str:
|
|
|
|
"""Filters out unnecessary URL args from a result link
|
|
|
|
|
|
|
|
Args:
|
|
|
|
link: The string result link to check for extraneous URL params
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
str: An updated (or ignored) result link
|
|
|
|
|
|
|
|
"""
|
|
|
|
parsed_link = urlparse.urlparse(link)
|
2020-07-26 20:53:59 +03:00
|
|
|
link_args = parse_qs(parsed_link.query)
|
|
|
|
safe_args = {}
|
|
|
|
|
|
|
|
if len(link_args) == 0 and len(parsed_link) > 0:
|
2021-03-08 20:22:04 +03:00
|
|
|
return link
|
2020-07-26 20:53:59 +03:00
|
|
|
|
|
|
|
for arg in link_args.keys():
|
|
|
|
if arg in SKIP_ARGS:
|
|
|
|
continue
|
|
|
|
|
|
|
|
safe_args[arg] = link_args[arg]
|
|
|
|
|
|
|
|
# Remove original link query and replace with filtered args
|
2021-03-08 20:22:04 +03:00
|
|
|
link = link.replace(parsed_link.query, '')
|
2020-07-26 20:53:59 +03:00
|
|
|
if len(safe_args) > 0:
|
2021-03-08 20:22:04 +03:00
|
|
|
link = link + urlparse.urlencode(safe_args, doseq=True)
|
2020-07-26 20:53:59 +03:00
|
|
|
else:
|
2021-03-08 20:22:04 +03:00
|
|
|
link = link.replace('?', '')
|
|
|
|
|
|
|
|
return link
|
|
|
|
|
|
|
|
|
|
|
|
def append_nojs(result: BeautifulSoup) -> None:
|
|
|
|
"""Appends a no-Javascript alternative for a search result
|
2020-07-26 20:53:59 +03:00
|
|
|
|
2021-03-08 20:22:04 +03:00
|
|
|
Args:
|
|
|
|
result: The search result to append a no-JS link to
|
2020-07-26 20:53:59 +03:00
|
|
|
|
2021-03-08 20:22:04 +03:00
|
|
|
Returns:
|
|
|
|
None
|
2020-07-26 20:53:59 +03:00
|
|
|
|
2021-03-08 20:22:04 +03:00
|
|
|
"""
|
2020-12-30 02:43:42 +03:00
|
|
|
nojs_link = BeautifulSoup(features='html.parser').new_tag('a')
|
2022-04-27 23:25:14 +03:00
|
|
|
nojs_link['href'] = f'{Endpoint.window}?nojs=1&location=' + result['href']
|
2021-11-02 01:03:47 +03:00
|
|
|
nojs_link.string = ' NoJS Link'
|
2021-03-08 20:22:04 +03:00
|
|
|
result.append(nojs_link)
|
2021-10-21 19:42:31 +03:00
|
|
|
|
|
|
|
|
2022-04-13 20:29:07 +03:00
|
|
|
def append_anon_view(result: BeautifulSoup, config: Config) -> None:
|
|
|
|
"""Appends an 'anonymous view' for a search result, where all site
|
|
|
|
contents are viewed through Whoogle as a proxy.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
result: The search result to append an anon view link to
|
|
|
|
nojs: Remove Javascript from Anonymous View
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
None
|
|
|
|
|
|
|
|
"""
|
|
|
|
av_link = BeautifulSoup(features='html.parser').new_tag('a')
|
|
|
|
nojs = 'nojs=1' if config.nojs else 'nojs=0'
|
|
|
|
location = f'location={result["href"]}'
|
2022-04-27 23:25:14 +03:00
|
|
|
av_link['href'] = f'{Endpoint.window}?{nojs}&{location}'
|
2022-04-13 20:29:07 +03:00
|
|
|
translation = current_app.config['TRANSLATIONS'][
|
|
|
|
config.get_localization_lang()
|
|
|
|
]
|
2022-04-20 23:11:29 +03:00
|
|
|
av_link.string = f'{translation["anon-view"]}'
|
2022-04-13 20:29:07 +03:00
|
|
|
av_link['class'] = 'anon-view'
|
|
|
|
result.append(av_link)
|
|
|
|
|
2021-12-07 08:56:13 +03:00
|
|
|
def check_currency(response: str) -> dict:
|
|
|
|
"""Check whether the results have currency conversion
|
|
|
|
|
|
|
|
Args:
|
|
|
|
response: Search query Result
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
dict: Consists of currency names and values
|
|
|
|
|
|
|
|
"""
|
|
|
|
soup = BeautifulSoup(response, 'html.parser')
|
|
|
|
currency_link = soup.find('a', {'href': 'https://g.co/gfd'})
|
|
|
|
if currency_link:
|
|
|
|
while 'class' not in currency_link.attrs or \
|
|
|
|
'ZINbbc' not in currency_link.attrs['class']:
|
2022-06-16 19:28:06 +03:00
|
|
|
if currency_link.parent:
|
|
|
|
currency_link = currency_link.parent
|
|
|
|
else:
|
|
|
|
return {}
|
2021-12-07 08:56:13 +03:00
|
|
|
currency_link = currency_link.find_all(class_='BNeawe')
|
|
|
|
currency1 = currency_link[0].text
|
|
|
|
currency2 = currency_link[1].text
|
|
|
|
currency1 = currency1.rstrip('=').split(' ', 1)
|
|
|
|
currency2 = currency2.split(' ', 1)
|
2022-02-18 02:33:44 +03:00
|
|
|
|
|
|
|
# Handle differences in currency formatting
|
|
|
|
# i.e. "5.000" vs "5,000"
|
2021-12-07 08:56:13 +03:00
|
|
|
if currency2[0][-3] == ',':
|
|
|
|
currency1[0] = currency1[0].replace('.', '')
|
|
|
|
currency1[0] = currency1[0].replace(',', '.')
|
|
|
|
currency2[0] = currency2[0].replace('.', '')
|
|
|
|
currency2[0] = currency2[0].replace(',', '.')
|
|
|
|
else:
|
|
|
|
currency1[0] = currency1[0].replace(',', '')
|
|
|
|
currency2[0] = currency2[0].replace(',', '')
|
2022-02-18 02:33:44 +03:00
|
|
|
|
|
|
|
currency1_value = float(re.sub(r'[^\d\.]', '', currency1[0]))
|
|
|
|
currency1_label = currency1[1]
|
|
|
|
|
|
|
|
currency2_value = float(re.sub(r'[^\d\.]', '', currency2[0]))
|
|
|
|
currency2_label = currency2[1]
|
|
|
|
|
|
|
|
return {'currencyValue1': currency1_value,
|
|
|
|
'currencyLabel1': currency1_label,
|
|
|
|
'currencyValue2': currency2_value,
|
|
|
|
'currencyLabel2': currency2_label
|
2021-12-07 08:56:13 +03:00
|
|
|
}
|
|
|
|
return {}
|
|
|
|
|
|
|
|
|
|
|
|
def add_currency_card(soup: BeautifulSoup,
|
|
|
|
conversion_details: dict) -> BeautifulSoup:
|
|
|
|
"""Adds the currency conversion boxes
|
|
|
|
to response of the search query
|
|
|
|
|
|
|
|
Args:
|
|
|
|
soup: Parsed search result
|
|
|
|
conversion_details: Dictionary of currency
|
|
|
|
related information
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
BeautifulSoup
|
|
|
|
"""
|
|
|
|
# Element before which the code will be changed
|
|
|
|
# (This is the 'disclaimer' link)
|
|
|
|
element1 = soup.find('a', {'href': 'https://g.co/gfd'})
|
|
|
|
|
|
|
|
while 'class' not in element1.attrs or \
|
|
|
|
'nXE3Ob' not in element1.attrs['class']:
|
|
|
|
element1 = element1.parent
|
|
|
|
|
|
|
|
# Creating the conversion factor
|
|
|
|
conversion_factor = (conversion_details['currencyValue1'] /
|
|
|
|
conversion_details['currencyValue2'])
|
|
|
|
|
|
|
|
# Creating a new div for the input boxes
|
|
|
|
conversion_box = soup.new_tag('div')
|
|
|
|
conversion_box['class'] = 'conversion_box'
|
|
|
|
|
|
|
|
# Currency to be converted from
|
|
|
|
input_box1 = soup.new_tag('input')
|
|
|
|
input_box1['id'] = 'cb1'
|
|
|
|
input_box1['type'] = 'number'
|
|
|
|
input_box1['class'] = 'cb'
|
|
|
|
input_box1['value'] = conversion_details['currencyValue1']
|
|
|
|
input_box1['oninput'] = f'convert(1, 2, {1 / conversion_factor})'
|
|
|
|
|
|
|
|
label_box1 = soup.new_tag('label')
|
|
|
|
label_box1['for'] = 'cb1'
|
|
|
|
label_box1['class'] = 'cb_label'
|
|
|
|
label_box1.append(conversion_details['currencyLabel1'])
|
|
|
|
|
|
|
|
br = soup.new_tag('br')
|
|
|
|
|
|
|
|
# Currency to be converted to
|
|
|
|
input_box2 = soup.new_tag('input')
|
|
|
|
input_box2['id'] = 'cb2'
|
|
|
|
input_box2['type'] = 'number'
|
|
|
|
input_box2['class'] = 'cb'
|
|
|
|
input_box2['value'] = conversion_details['currencyValue2']
|
|
|
|
input_box2['oninput'] = f'convert(2, 1, {conversion_factor})'
|
|
|
|
|
|
|
|
label_box2 = soup.new_tag('label')
|
|
|
|
label_box2['for'] = 'cb2'
|
|
|
|
label_box2['class'] = 'cb_label'
|
|
|
|
label_box2.append(conversion_details['currencyLabel2'])
|
|
|
|
|
|
|
|
conversion_box.append(input_box1)
|
|
|
|
conversion_box.append(label_box1)
|
|
|
|
conversion_box.append(br)
|
|
|
|
conversion_box.append(input_box2)
|
|
|
|
conversion_box.append(label_box2)
|
|
|
|
|
|
|
|
element1.insert_before(conversion_box)
|
|
|
|
return soup
|
2022-02-07 20:47:25 +03:00
|
|
|
|
|
|
|
|
|
|
|
def get_tabs_content(tabs: dict,
|
|
|
|
full_query: str,
|
|
|
|
search_type: str,
|
2022-09-22 23:14:56 +03:00
|
|
|
preferences: str,
|
2022-02-07 20:47:25 +03:00
|
|
|
translation: dict) -> dict:
|
|
|
|
"""Takes the default tabs content and updates it according to the query.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
tabs: The default content for the tabs
|
|
|
|
full_query: The original search query
|
|
|
|
search_type: The current search_type
|
|
|
|
translation: The translation to get the names of the tabs
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
dict: contains the name, the href and if the tab is selected or not
|
|
|
|
"""
|
2023-09-09 01:44:04 +03:00
|
|
|
map_query = full_query
|
|
|
|
if '-site:' in full_query:
|
|
|
|
block_idx = full_query.index('-site:')
|
|
|
|
map_query = map_query[:block_idx]
|
2022-02-07 20:47:25 +03:00
|
|
|
tabs = copy.deepcopy(tabs)
|
|
|
|
for tab_id, tab_content in tabs.items():
|
|
|
|
# update name to desired language
|
|
|
|
if tab_id in translation:
|
|
|
|
tab_content['name'] = translation[tab_id]
|
|
|
|
|
|
|
|
# update href with query
|
|
|
|
query = full_query.replace(f'&tbm={search_type}', '')
|
|
|
|
|
|
|
|
if tab_content['tbm'] is not None:
|
|
|
|
query = f"{query}&tbm={tab_content['tbm']}"
|
2022-10-31 21:45:44 +03:00
|
|
|
|
2022-09-22 23:14:56 +03:00
|
|
|
if preferences:
|
|
|
|
query = f"{query}&preferences={preferences}"
|
2022-02-07 20:47:25 +03:00
|
|
|
|
2023-09-09 01:44:04 +03:00
|
|
|
tab_content['href'] = tab_content['href'].format(
|
|
|
|
query=query,
|
|
|
|
map_query=map_query)
|
2022-02-07 20:47:25 +03:00
|
|
|
|
|
|
|
# update if selected tab (default all tab is selected)
|
|
|
|
if tab_content['tbm'] == search_type:
|
|
|
|
tabs['all']['selected'] = False
|
|
|
|
tab_content['selected'] = True
|
|
|
|
return tabs
|