X-Git-Url: https://git.phdru.name/?p=bookmarks_db.git;a=blobdiff_plain;f=Robots%2Fbkmk_rurllib.py;h=e0d5c2cba9cfdfae19d56a72057da8cd3c0867f5;hp=efade79a328d9c8724877cf1819b5b86570c66e3;hb=1322d04c660325532fe6e2bd0aa10ab8a4693440;hpb=8fedded7827ae1d79b31ec38d8e3146bb1a7bea1 diff --git a/Robots/bkmk_rurllib.py b/Robots/bkmk_rurllib.py index efade79..e0d5c2c 100644 --- a/Robots/bkmk_rurllib.py +++ b/Robots/bkmk_rurllib.py @@ -13,9 +13,15 @@ __all__ = ['robot_urllib'] import sys, os import time, urllib -from Robots.bkmk_robot_base import robot_base, RedirectException, get_error +from Robots.bkmk_robot_base import robot_base, get_error +class RedirectException(Exception): + def __init__(self, errcode, newurl): + Exception.__init__(self) + self.errcode = errcode + self.newurl = newurl + class MyURLopener(urllib.URLopener): # Error 301 -- relocated (permanently) def http_error_301(self, url, fp, errcode, errmsg, headers, data=None): @@ -48,17 +54,17 @@ class MyURLopener(urllib.URLopener): urllib._urlopener = MyURLopener() # Fake headers to pretend this is a real browser -_version = "Mozilla/5.0 (X11; U; Linux 2.6 i686; en) Gecko/20001221 Firefox/2.0.0" -urllib._urlopener.addheaders[0] = ('User-Agent', _version) -_version = "bookmarks_db (Python %d.%d.%d; urllib/%s)" % ( +_user_agent = "Mozilla/5.0 (X11; U; Linux 2.6 i686; en) Gecko/20001221 Firefox/2.0.0" +urllib._urlopener.addheaders[0] = ('User-Agent', _user_agent) +_x_user_agent = "bookmarks_db (Python %d.%d.%d; urllib/%s)" % ( sys.version_info[0], sys.version_info[1], sys.version_info[2], urllib.__version__) -urllib._urlopener.addheader('X-User-Agent', _version) +urllib._urlopener.addheader('X-User-Agent', _x_user_agent) urllib._urlopener.addheader('Referer', '') -urllib._urlopener.addheader('Connection', 'close') urllib._urlopener.addheader('Accept', '*/*') urllib._urlopener.addheader('Accept-Language', 'ru,en') urllib._urlopener.addheader('Cache-Control', 'max-age=300') +urllib._urlopener.addheader('Connection', 'close') urllib_ftpwrapper = urllib.ftpwrapper @@ -74,33 +80,38 @@ urllib.ftpwrapper = myftpwrapper class robot_urllib(robot_base): - def urlretrieve(self, bookmark, url, accept_charset=False): + def get(self, bookmark, url, accept_charset=False): try: # Set fake referer to the base URL urllib._urlopener.addheaders[2] = ('Referer', url) if accept_charset and bookmark.charset: urllib._urlopener.addheader('Accept-Charset', bookmark.charset) - fname, headers = urllib.urlretrieve(url) - if accept_charset and bookmark.charset: - del urllib._urlopener.addheaders[-1] + try: + fname, headers = urllib.urlretrieve(url) + finally: + if accept_charset and bookmark.charset: + del urllib._urlopener.addheaders[-1] # Remove Accept-Charset infile = open(fname, 'rb') content = infile.read() infile.close() - return headers, content, None + return None, None, None, headers, content + + except RedirectException, e: + return None, e.errcode, e.newurl, None, None - except IOError, msg: - if (msg[0] == "http error") and (msg[1] == -1): + except IOError, e: + if (e[0] == "http error") and (e[1] == -1): error = None bookmark.no_error = "The server did not return any header - it is not an error, actually" self.log(' no headers: %s' % bookmark.no_error) else: - error = get_error(msg) + error = get_error(e) self.log(' Error: %s' % error) - return None, None, error + return error, None, None, None, None def get_ftp_welcome(self): global ftpcache_key @@ -109,5 +120,6 @@ class robot_urllib(robot_base): # If there are - ftpcache_key in prev line is invalid. return _welcome - def cleanup(self): + def finish_check_url(self, bookmark): + robot_base.finish_check_url(self, bookmark) urllib.urlcleanup()