forked from GoogleChrome/related-website-sets
-
Notifications
You must be signed in to change notification settings - Fork 0
/
RwsCheck.py
570 lines (512 loc) · 25.2 KB
/
RwsCheck.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
# Copyright 2022 Google LLC
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# https://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import json
import requests
from RwsSet import RwsSet
from jsonschema import validate
from urllib.request import urlopen
from urllib.request import Request
from publicsuffix2 import PublicSuffixList
WELL_KNOWN = "/.well-known/related-website-set.json"
class RwsCheck:
"""Stores and runs checks on the list of rws sites
Attributes:
rws_sites: A json file read from canonical_sites that should contain all
submitted related website sets
etlds: A string of effective top level domains read from public suffix list
icanns: A set of domains associated with country codes
schema: Static. Stores schema for format the canonical_sites should follow
error_list: Stores all exceptions and issues generated by the checks. This
allows the issues to be shared in full when iterated through
without any given check failing halfway through and not
catching other issues.
"""
def __init__(self, rws_sites: json, etlds: PublicSuffixList, icanns: set):
"""Stores the input from canonical_sites, effective_tld_names.dat, and
ICANN_domains into the RwsCheck object"""
self.rws_sites = rws_sites
self.etlds = etlds
self.icanns = icanns
self.error_list = []
def validate_schema(self, schema_file):
"""Validates the canonical sites list
Calls the validate function from the jsonschema package on the input
from canonical_sites against our predertermined schema
Args:
self
Returns:
None
Raises:
jsonschema.exceptions.ValidationError if the schema does not match
the format stored in SCHEMA
"""
with open(schema_file) as f:
SCHEMA = json.loads(f.read())
validate(self.rws_sites, schema = SCHEMA)
def load_sets(self):
"""Loads sets from the JSON file into a dictionary of primary->RwsSet
Loads the sets from rws_list into check_sets, a dictionary of
string->RwsSet, where the key is the primary of the RwsSet
If any given primary is listed multiple times, will append an error to
the error_list for any primary past the first
Args:
None
Returns:
Dict[string, RwsSet]
"""
check_sets = {}
load_sets_errors = []
for rwset in self.rws_sites['sets']:
primary = rwset.get('primary')
ccTLDs = rwset.get('ccTLDs')
associated_sites = rwset.get('associatedSites')
service_sites = rwset.get('serviceSites')
if primary in check_sets.keys():
load_sets_errors.append(
f"{primary} is already a primary of another site")
else:
check_sets[primary] = RwsSet(
ccTLDs, primary, associated_sites, service_sites)
self.error_list += load_sets_errors
return check_sets
def has_all_rationales(self, check_sets):
"""Checks for the presence of all rationaleBySite elements in schema
Reads the associated sites and service sites from all RwsSets, and
checks that they have a corresponding entry in their Rws'
rationaleBysites field. If any given site does not have a rationale,
or the field is not present when it should be, appends an error to the
error_list
Args:
Dict[string, RwsSet]
Returns:
None
"""
for rwset in self.rws_sites['sets']:
if rwset['primary'] not in check_sets:
continue
sites = rwset.get("associatedSites", []) + rwset.get("serviceSites", [])
rationales = rwset.get('rationaleBySite', None)
if sites and rationales!=None:
for site in sites:
if site not in rationales.keys():
self.error_list.append(
f"There is no provided rationale for {site}")
if sites!=None and rationales == None:
self.error_list.append(
"A rationaleBySite field is required for this set, but"
+ " none is provided. ")
def check_exclusivity(self, check_sets):
"""This method checks for exclusivity of each field in a set of RwsSets
Ensures that no RwsSets intersect, e.g. a primary of one set cannot be
an associated site of another, nor can it be the primary of another set
etc. If any sets intersect, information about the intersections is
added to the error_list.
Args:
check_sets: Dict[string, RwsSet]
Returns:
None
"""
site_list = set()
for primary, rws in check_sets.items():
# Check the primary
if primary in site_list:
self.error_list.append(
"This primary is already registered in another related website"
+ f" set: {primary}")
else:
site_list.update(primary)
# Check the associated sites
associated_overlap = set(rws.associated_sites) & site_list
if associated_overlap:
self.error_list.append(
"These associated sites are already registered in " +
f"another related website set: {associated_overlap}")
else:
site_list.update(rws.associated_sites)
# Check the service sites
service_overlap = set(rws.service_sites) & site_list
if service_overlap:
self.error_list.append(
"These service sites are already registered in another"
+ f" related website set: {service_overlap}")
else:
site_list.update(rws.service_sites)
# Check the ccTLDs
for aliases in rws.ccTLDs.values():
alias_overlap = set(aliases) & site_list
if alias_overlap:
self.error_list.append(
"These ccTLD sites are already registered in "
+ f"another related website set: {alias_overlap}")
else:
site_list.update(aliases)
def url_is_https(self, site):
"""A function that checks for https://
Reads a domain name and returns whether or not it begins with https://
Args:
site: string corresponding to a domain name
Returns:
boolean with truth value if the domain name begins with https://
"""
return site.startswith("https://")
def find_non_https_urls(self, check_sets):
"""Checks for https:// in all sites.
Calls url_is_https on all sites in each RwsSet contained in check_sets,
and appends errors to the error list for any that return false
Args:
check_sets: Dict[string, RwsSet]
Returns:
None
"""
for primary, curr_set in check_sets.items():
# Apply to the primary
if not self.url_is_https(primary):
self.error_list.append(
f"The provided primary site does not begin with https:// {primary}")
# Apply to the country codes
for aliased_site, aliases in curr_set.ccTLDs.items():
if not self.url_is_https(aliased_site):
self.error_list.append(
f"The provided aliased site does not begin with https:// {aliased_site}")
for alias in aliases:
if not self.url_is_https(alias):
self.error_list.append(
f"The provided alias site does not begin with https:// {alias}")
# Apply to associated sites
for associated_site in curr_set.associated_sites:
if not self.url_is_https(associated_site):
self.error_list.append(
f"The provided associated site does not begin with https:// {associated_site}")
# Apply to service sites
for service_site in curr_set.service_sites:
if not self.url_is_https(service_site):
self.error_list.append(
f"The provided service site does not begin with https:// {service_site}")
def is_eTLD_Plus1(self, site):
"""A helper function for checking if a domain is etld+1 compliant
calls get_public suffix from the publicsuffix2 package on the provided
domain name, returns true if the domain name contains a public suffix,
else false
Args:
site: a string corresponding to a domain name
Returns:
boolean with truth value dependent on value of get_public_suffix
"""
assert site is not None
site = site.removeprefix("https://")
is_etldp1_or_etld = self.etlds.get_sld(site, strict=True) == site
is_etld = self.etlds.get_tld(site, strict=True) == site
return is_etldp1_or_etld and not is_etld
def find_invalid_eTLD_Plus1(self, check_sets):
"""Checks if all domains are etld+1 compliant
Calls is_eTLD_Plus1 on all sites in each RwsSet contained in check_sets
and appends errors to the error list for any that return false
Args:
check_sets: Dict[string, RwsSet]
Returns:
None
"""
for primary, curr_set in check_sets.items():
# Apply to the primary
if not self.is_eTLD_Plus1(primary):
self.error_list.append(
f"The provided primary site is not an eTLD+1: {primary}")
# Apply to the country codes
for aliased_site, aliases in curr_set.ccTLDs.items():
if not self.is_eTLD_Plus1(aliased_site):
self.error_list.append(
f"The provided aliased site is not an eTLD+1: {aliased_site}")
for alias in aliases:
if not self.is_eTLD_Plus1(alias):
self.error_list.append(
f"The provided alias site is not an eTLD+1: {alias}")
# Apply to associated sites
for associated_site in curr_set.associated_sites:
if not self.is_eTLD_Plus1(associated_site):
self.error_list.append(
f"The provided associated site is not an eTLD+1: {associated_site}")
# Apply to service sites
for service_site in curr_set.service_sites:
if not self.is_eTLD_Plus1(service_site):
self.error_list.append(
f"The provided service site is not an eTLD+1: {service_site}")
def open_and_load_json(self, url):
"""Calls urlopen and returns json from a site
Calls urlopena and json.load on a domain. Returns the json object.
This functionality is separated out here to make testing easier.
Args:
url: a domain that we want to load the json from
"""
req = Request(url=url, headers={'User-Agent': 'Chrome'})
with urlopen(req) as json_file:
return json.load(json_file)
def check_list_sites(self, primary, site_list):
"""Checks that sites in a given list have the correct primary on their
well-known page
Calls urlopen on a given list of sites, reads their json, and adds any
sites that do not contain the passed in primary as their listed primary
to the error list. Also catches and adds any exceptions when trying to
open or read the url
Args:
primary: the domain name of the primary site
site_list: a list of domain names to access
Returns:
None
"""
for site in site_list:
url = site + WELL_KNOWN
try:
json_schema = self.open_and_load_json(url)
if 'primary' not in json_schema.keys():
self.error_list.append(
"The listed associated site site did not have primary"
+ f" as a key in its {WELL_KNOWN} file: {site}")
elif json_schema['primary'] != primary:
self.error_list.append("The listed associated site "
+ f"did not have {primary} listed as its primary: {site}")
except Exception as inst:
self.error_list.append(
f"Experienced an error when trying to access {url}; "
+ f"error was: {inst}")
def check_well_known_list(self, field, list1, list2):
"""Checks that 2 lists for a given field match each other
Applies a symmetric diff to list1 and list2 and returns an empty list
if the 2 fields are symmetric. Otherwise returns a list with a single
string to be used as error text, containing the field, list1, list2,
and their symmetric diff.
Args:
field: string
list1: list[string]
list2: list[string]
Returns:
list[string]
"""
if list1 == list2:
return []
diff = sorted(set(list1) ^ set(list2))
return [f"Encountered an inequality between the PR submission and the {WELL_KNOWN} file:\n" +
f"\t{field} was {list1} in the PR, and {list2} in the well-known.\n" +
f"\tDiff was: {diff}."]
def find_invalid_well_known(self, check_sets):
"""Checks for and validates well-known pages for RWS sets
Checks for a ./well-known page for related website sets under each
domain, and checks that the format of the file aligns with the provided
pages in the canonical list.
Calls check_list_sites on all ccTLDs, associated, and service sites.
Appends to the error_list whenever a site is unreachable, an incorrect
format, or its contents do no match what is expected.
Args:
check_sets: Dict[string, RwsSet]
Returns:
None
"""
# Check the schema to ensure consistency
for primary, curr_rws_set in check_sets.items():
# First we check the primary sites
url = primary + WELL_KNOWN
# Read the well-known files and check them against the schema we
# have stored
try:
json_schema = self.open_and_load_json(url)
well_known_set = RwsSet(
json_schema.get('ccTLDs'),
json_schema.get('primary'),
json_schema.get('associatedSites'),
json_schema.get('serviceSites'))
if well_known_set.primary != curr_rws_set.primary:
self.error_list.append(f"The {WELL_KNOWN} set's primary ({well_known_set.primary}) did not equal " +
f"the PR set's primary ({curr_rws_set.primary})")
self.error_list.extend(self.check_well_known_list(
"associatedSites",
curr_rws_set.associated_sites,
well_known_set.associated_sites
)
)
self.error_list.extend(self.check_well_known_list(
"serviceSites",
curr_rws_set.service_sites,
well_known_set.service_sites
)
)
for aliased_site in curr_rws_set.ccTLDs | well_known_set.ccTLDs:
self.error_list.extend(self.check_well_known_list(
aliased_site + " alias list",
curr_rws_set.ccTLDs.get(aliased_site, []),
well_known_set.ccTLDs.get(aliased_site, [])
)
)
except Exception as inst:
self.error_list.append(
f"Experienced an error when trying to access {url}; error was: {inst}")
# Check the member sites.
self.check_list_sites(
primary,
curr_rws_set.associated_sites +
curr_rws_set.service_sites +
[alias
for aliases in curr_rws_set.ccTLDs.values()
for alias in aliases
]
)
def find_invalid_removal(self, subtracted_sets):
"""Checks that any sets being removed were properly removed by owner
Checks that the /.well-known page for the primary of any RWS removed
from the list returns an error 404.
Args:
subtracted_sets: Dict[string, RwsSet]
Returns:
None"""
for primary in subtracted_sets:
url = primary + WELL_KNOWN
try:
r = requests.get(url, timeout=10)
if r.status_code != 404:
self.error_list.append(f"The set associated with {primary}"
+ f" was removed from the list, but {url} does not return error 404.")
except Exception as inst:
self.error_list.append(f"Unexpected error when accessing {url}; Received error: {inst}")
def find_invalid_alias_eSLDs(self, check_sets):
"""Checks that eSLDs match their alias, and that country codes are
members of icann
Reads the ccTLDs and makes sure that they match their equivalent sites,
and that their eTLDs are part of ICANN's list of country codes.
If either of these is not the case, appends an error to the error_list.
Note: A site may list a variant with "com" as its eTLD IFF the site
being aliased has an eTLD on ICANN's list of countrycodes.
Args:
check_sets: Dict[string, RwsSet]
Returns:
None
"""
for primary, curr_set in check_sets.items():
for aliased_site, aliases in curr_set.ccTLDs.items():
# first check if the aliased site is actually anywhere else
# in the rws
if not curr_set.includes(aliased_site, False):
self.error_list.append(
f"The aliased site {aliased_site} contained within the ccTLDs must be a " +
"primary, associated site, or service site " +
f"within the firsty pary set for {primary}")
# check the validity of the aliases
aliased_eSLD, aliased_tld = (aliased_site.split(".")[0],
aliased_site.split(".")[-1])
if aliased_tld in self.icanns:
icann_check = self.icanns.union({"com"})
else:
icann_check = self.icanns
variants = [(site, site.split(".")[0], site.split(".")[-1])
for site in aliases]
for site, eSLD, tld in variants:
if eSLD != aliased_eSLD:
self.error_list.append(
f"The following top level domain must match: {aliased_site}, but is instead: {site}")
if tld not in icann_check:
self.error_list.append(
f"The provided country code: {tld}, in: {site} is not a ICANN registered country code")
def find_robots_txt(self, check_sets):
"""Checks service sites to see if they have a robots.txt subdomain.
Iterates through all service_sites in each RwsSet provided, and makes
a get request to site/robots.txt for each. This request should return
an error 4xx, 5xx, or a timeout error. If it does not, and the page
does exist, then it is expected that the site contains a X-Robots-Tag
in its header. If none of these conditions is met, an error is appended
to the error list.
Args:
check_sets: Dict[string, RwsSet]
Returns:
None
"""
exception_retries = "Max retries exceeded with url: /robots.txt"
exception_timeout = "Read timed out. (read timeout=10)"
for curr_set in check_sets.values():
for service_site in curr_set.service_sites:
try:
r_service = requests.get(service_site, timeout=10, allow_redirects=False)
if 'X-Robots-Tag' not in r_service.headers:
self.error_list.append(f"The service site {service_site} does not have an X-Robots-Tag in its "
+ "header")
else:
robots_tag = r_service.headers['X-Robots-Tag']
if ':' in robots_tag:
self.error_list.append(f"The service site {service_site} contains an 'X-Robots-Tag' " +
"that does not meet the policy requirements")
elif 'none' not in robots_tag and 'noindex' not in robots_tag:
self.error_list.append(f"The service site {service_site} does not have a " +
"'noindex' or 'none' tag in its header"
)
except Exception as inst:
if exception_retries not in str(inst):
if exception_timeout not in str(inst):
self.error_list.append(
f"Unexpected error for service site: {service_site}; Received error: {inst}")
def find_ads_txt(self, check_sets):
"""Checks to see if service sites have an ads.txt subdomain.
Iterates through all service_sites in each RwsSet provided, and makes
a get request to site/ads.txt for each. Appends errors to the error
list for any that do not return an error 4xx or 5xx or if the site
does not cause a timeout error.
Args:
check_sets: Dict[string, RwsSet]
Returns:
None
"""
exception_retries = "Max retries exceeded with url: /ads.txt"
exception_timeout = "Read timed out. (read timeout=10)"
for curr_set in check_sets.values():
for service_site in curr_set.service_sites:
ads_site = service_site + "/ads.txt"
try:
r = requests.get(ads_site, timeout=10)
if r.status_code == 200:
self.error_list.append(f"The service site {service_site} has an ads.txt file, this violates "
+ "the policies for service sites")
except Exception as inst:
if exception_retries not in str(inst):
if exception_timeout not in str(inst):
self.error_list.append(
f"Unexpected error for service site: {service_site}\n" +
f"Received error: {inst}")
def check_for_service_redirect(self, check_sets):
"""Checks to see if service sites redirect to another site
or return a user/server error.
Makes a get request to all service sites in each RwsSet contained in
check_sets, and appends errors to the error list for any that do not
return an error 4xx or 5xx or if the site does not cause a timeout
error.
Args:
check_sets: Dict[string, RwsSet]
Returns:
None
"""
exception_retries = "Max retries exceeded with url: /"
exception_timeout = "Read timed out. (read timeout=10)"
for curr_set in check_sets.values():
for service_site in curr_set.service_sites:
try:
r = requests.get(service_site, timeout=10)
# We want the request status_code to be a 4xx or 5xx, raise
# an exception if it's outside that range
if r.status_code < 400 or r.status_code >= 600:
# If a get request to a service site successfully
# connects to that site, we expect it to be a redirect
# If it is not a redirect, we raise an exception
if r.url == service_site or r.url == service_site+"/":
self.error_list.append(
f"The service site must not be an endpoint: {service_site}")
except Exception as inst:
if exception_retries not in str(inst):
if exception_timeout not in str(inst):
self.error_list.append(f"Unexpected error for "
+ f"service site: {service_site}\n"
+ f"Received error: {inst}")