2009-08-13 03:26:12 +00:00
|
|
|
import urllib2
|
|
|
|
import logging
|
|
|
|
import re
|
2009-09-16 02:34:04 +00:00
|
|
|
import multiprocessing
|
2009-08-13 03:26:12 +00:00
|
|
|
|
|
|
|
class PageImporter(object):
|
|
|
|
|
|
|
|
def __init__(self, url, feed):
|
|
|
|
self.url = url
|
|
|
|
self.feed = feed
|
2009-09-16 02:34:04 +00:00
|
|
|
self.lock = multiprocessing.Lock()
|
2009-08-13 03:26:12 +00:00
|
|
|
|
|
|
|
def fetch_page(self):
|
|
|
|
request = urllib2.Request(self.url)
|
|
|
|
|
|
|
|
try:
|
|
|
|
response = urllib2.urlopen(request)
|
2009-08-15 15:10:21 +00:00
|
|
|
except urllib2.HTTPError, e:
|
2009-08-13 03:26:12 +00:00
|
|
|
logging.error('The server couldn\'t fulfill the request. Error: %s' % e.code)
|
2009-08-15 15:10:21 +00:00
|
|
|
except urllib2.URLError, e:
|
2009-08-13 03:26:12 +00:00
|
|
|
logging.error('Failed to reach server. Reason: %s' % e.reason)
|
|
|
|
else:
|
|
|
|
data = response.read()
|
|
|
|
html = data
|
|
|
|
html = self.rewrite_page(html)
|
|
|
|
self.save_page(html)
|
|
|
|
|
|
|
|
def rewrite_page(self, response):
|
|
|
|
base_code = u'<base href="%s" />' % (self.feed.feed_link,)
|
|
|
|
try:
|
2009-08-26 03:12:55 +00:00
|
|
|
html = re.sub(r'<head(.*?\>)', r'<head\1 '+base_code, response)
|
2009-08-13 03:26:12 +00:00
|
|
|
except:
|
|
|
|
response = response.decode('latin1').encode('utf-8')
|
2009-08-26 03:12:55 +00:00
|
|
|
html = re.sub(r'<head(.*?\>)', r'<head\1 '+base_code, response)
|
2009-08-13 03:26:12 +00:00
|
|
|
|
|
|
|
return html
|
|
|
|
|
|
|
|
def save_page(self, html):
|
|
|
|
self.feed.page_data = html
|
2009-09-16 04:00:37 +00:00
|
|
|
self.lock.acquire()
|
|
|
|
try:
|
|
|
|
self.feed.save()
|
|
|
|
finally:
|
|
|
|
self.lock.release()
|