
The ":past <duration>" query string filtering isn't used anymore since adding the option to filter by time in the result view.
158 lines
5.1 KiB
Python
158 lines
5.1 KiB
Python
from bs4 import BeautifulSoup
|
|
from app.filter import Filter
|
|
from app.models.config import Config
|
|
from app.models.endpoint import Endpoint
|
|
from app.utils import results
|
|
from app.utils.session import generate_key
|
|
from datetime import datetime
|
|
from dateutil.parser import ParserError, parse
|
|
from urllib.parse import urlparse
|
|
|
|
from test.conftest import demo_config
|
|
|
|
|
|
def get_search_results(data):
|
|
secret_key = generate_key()
|
|
soup = Filter(user_key=secret_key, config=Config(**demo_config)).clean(
|
|
BeautifulSoup(data, 'html.parser'))
|
|
|
|
main_divs = soup.find('div', {'id': 'main'})
|
|
assert len(main_divs) > 1
|
|
|
|
result_divs = []
|
|
for div in main_divs:
|
|
# Result divs should only have 1 inner div
|
|
if (len(list(div.children)) != 1
|
|
or not div.findChild()
|
|
or 'div' not in div.findChild().name):
|
|
continue
|
|
|
|
result_divs.append(div)
|
|
|
|
return result_divs
|
|
|
|
|
|
def test_get_results(client):
|
|
rv = client.get(f'/{Endpoint.search}?q=test')
|
|
assert rv._status_code == 200
|
|
|
|
# Depending on the search, there can be more
|
|
# than 10 result divs
|
|
results = get_search_results(rv.data)
|
|
assert len(results) >= 10
|
|
assert len(results) <= 15
|
|
|
|
|
|
def test_post_results(client):
|
|
rv = client.post(f'/{Endpoint.search}', data=dict(q='test'))
|
|
assert rv._status_code == 302
|
|
|
|
|
|
def test_translate_search(client):
|
|
rv = client.get(f'/{Endpoint.search}?q=translate hola')
|
|
assert rv._status_code == 200
|
|
|
|
# Pretty weak test, but better than nothing
|
|
str_data = str(rv.data)
|
|
assert 'iframe' in str_data
|
|
assert '/auto/en/ hola' in str_data
|
|
|
|
|
|
def test_block_results(client):
|
|
rv = client.get(f'/{Endpoint.search}?q=pinterest')
|
|
assert rv._status_code == 200
|
|
|
|
has_pinterest = False
|
|
for link in BeautifulSoup(rv.data, 'html.parser').find_all('a', href=True):
|
|
if 'pinterest.com' in urlparse(link['href']).netloc:
|
|
has_pinterest = True
|
|
break
|
|
|
|
assert has_pinterest
|
|
|
|
demo_config['block'] = 'pinterest.com'
|
|
rv = client.post(f'/{Endpoint.config}', data=demo_config)
|
|
assert rv._status_code == 302
|
|
|
|
rv = client.get(f'/{Endpoint.search}?q=pinterest')
|
|
assert rv._status_code == 200
|
|
|
|
for link in BeautifulSoup(rv.data, 'html.parser').find_all('a', href=True):
|
|
result_site = urlparse(link['href']).netloc
|
|
if not result_site:
|
|
continue
|
|
assert result_site not in 'pinterest.com'
|
|
|
|
|
|
def test_view_my_ip(client):
|
|
rv = client.get(f'/{Endpoint.search}?q=my ip address')
|
|
assert rv._status_code == 200
|
|
|
|
# Pretty weak test, but better than nothing
|
|
str_data = str(rv.data)
|
|
assert 'Your public IP address' in str_data
|
|
assert '127.0.0.1' in str_data
|
|
|
|
|
|
def test_recent_results(client):
|
|
times = {
|
|
'tbs=qdr:y': 365,
|
|
'tbs=qdr:m': 31,
|
|
'tbs=qdr:w': 7
|
|
}
|
|
|
|
for time, num_days in times.items():
|
|
rv = client.get(f'/{Endpoint.search}?q=test&' + time)
|
|
result_divs = get_search_results(rv.data)
|
|
|
|
current_date = datetime.now()
|
|
for div in [_ for _ in result_divs if _.find('span')]:
|
|
date_span = div.find('span').decode_contents()
|
|
if not date_span or len(date_span) > 15 or len(date_span) < 7:
|
|
continue
|
|
|
|
try:
|
|
date = parse(date_span)
|
|
# Date can have a little bit of wiggle room
|
|
assert (current_date - date).days <= (num_days + 5)
|
|
except ParserError:
|
|
pass
|
|
|
|
|
|
def test_leading_slash_search(client):
|
|
# Ensure searches with a leading slash are interpreted
|
|
# correctly as queries and not endpoints
|
|
q = '/test'
|
|
rv = client.get(f'/{Endpoint.search}?q={q}')
|
|
assert rv._status_code == 200
|
|
|
|
soup = Filter(
|
|
user_key=generate_key(),
|
|
config=Config(**demo_config),
|
|
query=q
|
|
).clean(BeautifulSoup(rv.data, 'html.parser'))
|
|
|
|
for link in soup.find_all('a', href=True):
|
|
if 'start=' not in link['href']:
|
|
continue
|
|
|
|
assert link['href'].startswith(f'{Endpoint.search}')
|
|
|
|
|
|
def test_site_alt_prefix_skip():
|
|
# Ensure prefixes are skipped correctly for site alts
|
|
|
|
# default silte_alts (farside.link)
|
|
assert results.get_site_alt(link = 'https://www.reddit.com') == 'https://farside.link/libreddit'
|
|
assert results.get_site_alt(link = 'https://www.twitter.com') == 'https://farside.link/nitter'
|
|
assert results.get_site_alt(link = 'https://www.youtube.com') == 'https://farside.link/invidious'
|
|
|
|
test_site_alts = {
|
|
'reddit.com': 'reddit.endswithmobile.domain',
|
|
'twitter.com': 'https://twitter.endswithm.domain',
|
|
'youtube.com': 'http://yt.endswithwww.domain',
|
|
}
|
|
# Domains with part of SKIP_PREFIX in them
|
|
assert results.get_site_alt(link = 'https://www.reddit.com', site_alts = test_site_alts) == 'https://reddit.endswithmobile.domain'
|
|
assert results.get_site_alt(link = 'https://www.twitter.com', site_alts = test_site_alts) == 'https://twitter.endswithm.domain'
|
|
assert results.get_site_alt(link = 'https://www.youtube.com', site_alts = test_site_alts) == 'http://yt.endswithwww.domain'
|