|
21 | 21 | import html.entities
|
22 | 22 | import logging
|
23 | 23 | import re
|
24 |
| -import urllib |
| 24 | +import requests |
25 | 25 |
|
26 | 26 |
|
27 | 27 | log = logging.getLogger(__name__)
|
@@ -105,7 +105,8 @@ def __call__(self):
|
105 | 105 | review_count = 0
|
106 | 106 | for url in self._getTargetURLs():
|
107 | 107 | try:
|
108 |
| - sehepunkte_xml = urllib.request.urlopen(url).read() # nosec B310 |
| 108 | + response = requests.get(url) |
| 109 | + sehepunkte_xml = response.content |
109 | 110 | data.append(sehepunkte_parser.parse(sehepunkte_xml))
|
110 | 111 | except OSError:
|
111 | 112 | pass # The library takes care of logging a failure
|
@@ -133,7 +134,7 @@ def __call__(self):
|
133 | 134 | return "Success"
|
134 | 135 |
|
135 | 136 | def _getTargetURLs(self):
|
136 |
| - base = "http://www.sehepunkte.de/export/sehepunkte_%s.xml" |
| 137 | + base = "https://www.sehepunkte.de/export/sehepunkte_%s.xml" |
137 | 138 | now = datetime.datetime.now()
|
138 | 139 | past_months = int(self.request.get("past_months", 1))
|
139 | 140 | for idx in reversed(range(past_months + 1)):
|
@@ -225,7 +226,8 @@ def setter(mapper):
|
225 | 226 |
|
226 | 227 | def _extractAndSanitizeHTML(self, review):
|
227 | 228 | # XXX check scheme? (bandit)
|
228 |
| - html = urllib.request.urlopen(review["canonical_uri"]).read() # nosec B310 |
| 229 | + response = requests.get(review["canonical_uri"]) |
| 230 | + html = response.content |
229 | 231 | soup = BeautifulSoup(html, "lxml")
|
230 | 232 | dirt = soup.findAll("div", {"class": "box"})
|
231 | 233 | for div in dirt:
|
|
0 commit comments