home
·
contact
·
privacy
projects
/
berlin-corona-table
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
Remove redundant whitespace.
[berlin-corona-table]
/
scrape.py
diff --git
a/scrape.py
b/scrape.py
index 41a0db9d3c44a5781a66d7dbf02be6200c2df34c..e449b998b2821f4fe1fe9539fc1fbf28b2898b9c 100755
(executable)
--- a/
scrape.py
+++ b/
scrape.py
@@
-104,12
+104,10
@@
fixes = {
# compromise to keep as many surrounding numbers stable as possible.
datetime.datetime(2020, 3, 26): {
'SZ': {
# compromise to keep as many surrounding numbers stable as possible.
datetime.datetime(2020, 3, 26): {
'SZ': {
- 'growth': 12,
- 'total': 132
+ 'growth': 12
},
'sum': {
},
'sum': {
- 'growth': 286,
- 'total': 1931
+ 'growth': 286
}
},
# Here the official total is 220, while the summation of district
}
},
# Here the official total is 220, while the summation of district
@@
-121,7
+119,7
@@
fixes = {
},
}
},
}
-# Scan navigation bar for maximum pagination value.
+# Scan navigation bar for maximum pagination value.
url = url_prefix + pm_dir
with urllib.request.urlopen(url) as response:
html = response.read()
url = url_prefix + pm_dir
with urllib.request.urlopen(url) as response:
html = response.read()
@@
-147,7
+145,7
@@
for i in range(max_page):
continue
day_urls += [link['href']]
continue
day_urls += [link['href']]
-# Collect infection data.
+# Collect infection data.
data = {}
first_run = True
districts_sorted = []
data = {}
first_run = True
districts_sorted = []