import urllib2, httplib import re import urlparse import traceback import feedparser import time from utils import log as logging from apps.rss_feeds.models import MFeedPage from utils.feed_functions import timelimit, mail_feed_error_to_admin HEADERS = { 'User-Agent': 'NewsBlur Page Fetcher - http://www.newsblur.com', 'Connection': 'close', } BROKEN_PAGES = [ 'tag:', 'info:', 'uuid:', '[]', ] class PageImporter(object): def __init__(self, url, feed): self.url = url self.feed = feed @timelimit(15) def fetch_page(self): if not self.url: self.save_no_page() return try: if self.url.startswith('http'): request = urllib2.Request(self.url, headers=HEADERS) response = urllib2.urlopen(request) time.sleep(0.01) # Grrr, GIL. data = response.read() elif any(self.url.startswith(s) for s in BROKEN_PAGES): self.save_no_page() return else: data = open(self.url, 'r').read() html = self.rewrite_page(data) self.save_page(html) except (ValueError, urllib2.URLError, httplib.BadStatusLine, httplib.InvalidURL), e: self.feed.save_page_history(401, "Bad URL", e) fp = feedparser.parse(self.feed.feed_address) self.feed.feed_link = fp.feed.get('link', "") self.feed.save() except (urllib2.HTTPError), e: self.feed.save_page_history(e.code, e.msg, e.fp.read()) return except (httplib.IncompleteRead), e: self.feed.save_page_history(500, "IncompleteRead", e) return except Exception, e: logging.debug('[%d] ! -------------------------' % (self.feed.id,)) tb = traceback.format_exc() logging.debug(tb) logging.debug('[%d] ! -------------------------' % (self.feed.id,)) self.feed.save_page_history(500, "Error", tb) mail_feed_error_to_admin(self.feed, e) return self.feed.save_page_history(200, "OK") def save_no_page(self): self.feed.has_page = False self.feed.save() self.feed.save_page_history(404, "Feed has no original page.") def rewrite_page(self, response): BASE_RE = re.compile(r'
)', re.I) base_code = u'