bookmark.icon = None
error, http_status_code, redirect_to, headers, content = \
- await self.get_url(bookmark, bookmark.href, True)
+ await self.get_url(bookmark.href, bookmark.charset)
if error is not None:
bookmark.error = error
icon_error, \
icon_status_code, icon_redirect_to, \
icon_headers, icon_data = \
- await self.get_url(bookmark, _icon_url)
+ await self.get_url(_icon_url)
if icon_error:
raise IOError("No icon: " + icon_error)
break
finally:
self.finish_check_url(bookmark)
- async def get_url(self, bookmark, url, accept_charset=False):
+ async def get_url(self, url, accept_charset=None):
split_results = urlsplit(url)
url_proto = split_results.scheme
url_host = split_results.hostname
if use_proxy and url_host in self.proxy_ok:
self.log(' Immediately trying with the proxy')
error, http_status_code, redirect_to, headers, content = \
- await self.get(bookmark, url,
- accept_charset=accept_charset,
+ await self.get(url, accept_charset=accept_charset,
use_proxy=True)
else:
error, http_status_code, redirect_to, headers, content = \
- await self.get(bookmark, url,
- accept_charset=accept_charset)
+ await self.get(url, accept_charset=accept_charset)
if error is not None and (
not url_host.startswith('localhost') and
not url_host.startswith('127.')
if use_proxy and http_status_code != 404:
self.log(' Retrying with the proxy...')
error, http_status_code, redirect_to, headers, content = \
- await self.get(bookmark, url,
- accept_charset=accept_charset,
+ await self.get(url, accept_charset=accept_charset,
use_proxy=True)
if error is None:
self.proxy_ok.add(url_host)
def version_str(self):
return 'aiohttp/%s' % aiohttp.__version__
- async def get(self, bookmark, url, accept_charset=False, use_proxy=False):
+ async def get(self, url, accept_charset=None, use_proxy=False):
if url.startswith('ftp://'):
error, body = await _get_ftp(
url, timeout=self.ftp_timeout,
return error, None, None, None, None
return None, None, None, None, body
- if accept_charset and bookmark.charset:
+ if accept_charset:
headers = request_headers.copy()
- headers['Accept-Charset'] = bookmark.charset
+ headers['Accept-Charset'] = accept_charset
else:
headers = request_headers
def version_str(self):
return str(pycurl.version)
- async def get(self, bookmark, url, accept_charset=False, use_proxy=False):
- if accept_charset and bookmark.charset:
+ async def get(self, url, accept_charset=None, use_proxy=False):
+ if accept_charset:
headers = request_headers.copy()
- headers['Accept-Charset'] = bookmark.charset
+ headers['Accept-Charset'] = accept_charset
else:
headers = request_headers
headers = ['%s: %s' % (k, v) for k, v in headers.items()]
try:
url.encode('ascii')
except UnicodeEncodeError:
- url = encode_url(url, bookmark.charset)
+ url = encode_url(url, accept_charset)
curl.setopt(pycurl.URL, url)
try:
curl.perform()
current_href.set(bookmark.href)
await self.check_bookmark_async(bookmark)
- async def get_url(self, bookmark, url, accept_charset=False):
- if bookmark.href not in self.logs:
- self.logs[bookmark.href] = []
- current_href.set(bookmark.href)
+ async def get_url(self, url, accept_charset=None):
+ if url not in self.logs:
+ self.logs[url] = []
+ current_href.set(url)
return await super(robot_multiaio, self).get_url(
- bookmark, url, accept_charset=accept_charset)
+ url, accept_charset=accept_charset)
def wait(self):
self.loop.run_until_complete(self.wait_async())
def version_str(self):
return 'python-requests urllib3/%s' % urllib3.__version__
- async def get(self, bookmark, url, accept_charset=False, use_proxy=False):
+ async def get(self, url, accept_charset=None, use_proxy=False):
if url.startswith('ftp://'):
error, welcome, body = _get_ftp(url, self.timeout)
if error is not None:
self.welcome = welcome
return None, None, None, None, body
- if accept_charset and bookmark.charset:
+ if accept_charset:
headers = request_headers.copy()
- headers['Accept-Charset'] = bookmark.charset
+ headers['Accept-Charset'] = accept_charset
else:
headers = request_headers