2010-07-06 16:37:49 -04:00
|
|
|
import datetime
|
2012-07-16 18:11:18 -07:00
|
|
|
import time
|
2013-10-07 13:36:10 -07:00
|
|
|
import re
|
2012-07-16 18:11:18 -07:00
|
|
|
import redis
|
2010-08-16 15:45:35 -04:00
|
|
|
from utils import log as logging
|
2010-10-23 13:06:28 -04:00
|
|
|
from utils import json_functions as json
|
2010-11-09 09:55:44 -05:00
|
|
|
from django.db import models, IntegrityError
|
2013-06-27 15:21:13 -07:00
|
|
|
from django.db.models import Q
|
2010-09-19 11:30:18 -04:00
|
|
|
from django.conf import settings
|
2009-06-16 03:08:55 +00:00
|
|
|
from django.contrib.auth.models import User
|
2013-05-02 16:49:44 -07:00
|
|
|
from django.core.cache import cache
|
2011-11-05 16:25:04 -07:00
|
|
|
from mongoengine.queryset import OperationError
|
2011-02-06 15:04:21 -05:00
|
|
|
from apps.reader.managers import UserSubscriptionManager
|
2011-01-17 23:20:25 -05:00
|
|
|
from apps.rss_feeds.models import Feed, MStory, DuplicateFeed
|
2010-08-22 18:34:40 -04:00
|
|
|
from apps.analyzer.models import MClassifierFeed, MClassifierAuthor, MClassifierTag, MClassifierTitle
|
2010-01-21 13:12:29 -05:00
|
|
|
from apps.analyzer.models import apply_classifier_titles, apply_classifier_feeds, apply_classifier_authors, apply_classifier_tags
|
2013-06-28 19:09:34 -07:00
|
|
|
from utils.feed_functions import add_object_to_folder, chunks
|
2010-01-21 13:12:29 -05:00
|
|
|
|
2009-06-16 03:08:55 +00:00
|
|
|
class UserSubscription(models.Model):
|
2010-07-20 20:23:49 -04:00
|
|
|
"""
|
2013-06-27 17:21:01 -07:00
|
|
|
A feed which a user has subscribed to. Carries all of the cached information
|
2010-07-20 20:23:49 -04:00
|
|
|
about the subscription, including unread counts of the three primary scores.
|
|
|
|
|
|
|
|
Also has a dirty flag (needs_unread_recalc) which means that the unread counts
|
|
|
|
are not accurate and need to be calculated with `self.calculate_feed_scores()`.
|
|
|
|
"""
|
2010-09-24 18:22:12 -04:00
|
|
|
UNREAD_CUTOFF = datetime.datetime.utcnow() - datetime.timedelta(days=settings.DAYS_OF_UNREAD)
|
2011-01-07 16:26:17 -05:00
|
|
|
|
2010-09-16 14:35:20 -04:00
|
|
|
user = models.ForeignKey(User, related_name='subscriptions')
|
|
|
|
feed = models.ForeignKey(Feed, related_name='subscribers')
|
2010-12-11 15:26:45 -05:00
|
|
|
user_title = models.CharField(max_length=255, null=True, blank=True)
|
2010-09-28 18:53:57 -04:00
|
|
|
active = models.BooleanField(default=False)
|
2010-09-19 11:30:18 -04:00
|
|
|
last_read_date = models.DateTimeField(default=UNREAD_CUTOFF)
|
|
|
|
mark_read_date = models.DateTimeField(default=UNREAD_CUTOFF)
|
2010-01-21 13:12:29 -05:00
|
|
|
unread_count_neutral = models.IntegerField(default=0)
|
|
|
|
unread_count_positive = models.IntegerField(default=0)
|
|
|
|
unread_count_negative = models.IntegerField(default=0)
|
2010-09-28 18:53:57 -04:00
|
|
|
unread_count_updated = models.DateTimeField(default=datetime.datetime.now)
|
2011-01-15 18:41:41 -05:00
|
|
|
oldest_unread_story_date = models.DateTimeField(default=datetime.datetime.now)
|
2010-04-08 18:36:48 -04:00
|
|
|
needs_unread_recalc = models.BooleanField(default=False)
|
2010-07-20 20:23:49 -04:00
|
|
|
feed_opens = models.IntegerField(default=0)
|
2010-08-01 19:12:42 -04:00
|
|
|
is_trained = models.BooleanField(default=False)
|
2011-02-06 15:04:21 -05:00
|
|
|
|
|
|
|
objects = UserSubscriptionManager()
|
2009-06-16 03:08:55 +00:00
|
|
|
|
|
|
|
def __unicode__(self):
|
2012-07-23 13:08:55 -07:00
|
|
|
return '[%s (%s): %s (%s)] ' % (self.user.username, self.user.pk,
|
|
|
|
self.feed.feed_title, self.feed.pk)
|
2012-01-18 17:39:00 -08:00
|
|
|
|
|
|
|
class Meta:
|
|
|
|
unique_together = ("user", "feed")
|
2010-12-31 10:34:31 -05:00
|
|
|
|
2011-11-29 09:43:16 -08:00
|
|
|
def canonical(self, full=False, include_favicon=True, classifiers=None):
|
2011-04-04 12:01:29 -04:00
|
|
|
feed = self.feed.canonical(full=full, include_favicon=include_favicon)
|
2011-03-04 12:27:31 -05:00
|
|
|
feed['feed_title'] = self.user_title or feed['feed_title']
|
|
|
|
feed['ps'] = self.unread_count_positive
|
|
|
|
feed['nt'] = self.unread_count_neutral
|
|
|
|
feed['ng'] = self.unread_count_negative
|
|
|
|
feed['active'] = self.active
|
2011-04-25 20:53:29 -04:00
|
|
|
feed['feed_opens'] = self.feed_opens
|
2012-05-21 20:08:27 -07:00
|
|
|
feed['subscribed'] = True
|
2011-11-29 09:43:16 -08:00
|
|
|
if classifiers:
|
|
|
|
feed['classifiers'] = classifiers
|
2011-03-15 23:42:27 -04:00
|
|
|
if not self.active and self.user.profile.is_premium:
|
|
|
|
feed['active'] = True
|
|
|
|
self.active = True
|
|
|
|
self.save()
|
2011-02-06 15:04:21 -05:00
|
|
|
|
|
|
|
return feed
|
|
|
|
|
2010-10-25 20:20:59 -04:00
|
|
|
def save(self, *args, **kwargs):
|
2012-05-14 15:01:30 -07:00
|
|
|
user_title_max = self._meta.get_field('user_title').max_length
|
|
|
|
if self.user_title and len(self.user_title) > user_title_max:
|
|
|
|
self.user_title = self.user_title[:user_title_max]
|
2011-03-15 23:42:27 -04:00
|
|
|
if not self.active and self.user.profile.is_premium:
|
|
|
|
self.active = True
|
2010-11-09 09:55:44 -05:00
|
|
|
try:
|
2010-10-25 20:20:59 -04:00
|
|
|
super(UserSubscription, self).save(*args, **kwargs)
|
2010-11-09 09:55:44 -05:00
|
|
|
except IntegrityError:
|
2012-01-26 09:32:24 -08:00
|
|
|
duplicate_feeds = DuplicateFeed.objects.filter(duplicate_feed_id=self.feed_id)
|
2011-11-01 09:25:59 -07:00
|
|
|
for duplicate_feed in duplicate_feeds:
|
|
|
|
already_subscribed = UserSubscription.objects.filter(user=self.user, feed=duplicate_feed.feed)
|
|
|
|
if not already_subscribed:
|
|
|
|
self.feed = duplicate_feed.feed
|
|
|
|
super(UserSubscription, self).save(*args, **kwargs)
|
|
|
|
break
|
2011-10-28 10:29:11 -07:00
|
|
|
else:
|
|
|
|
self.delete()
|
2013-06-27 15:21:13 -07:00
|
|
|
|
|
|
|
@classmethod
|
2013-06-27 17:21:01 -07:00
|
|
|
def subs_for_feeds(cls, user_id, feed_ids=None, read_filter="unread"):
|
|
|
|
usersubs = cls.objects
|
|
|
|
if read_filter == "unread":
|
|
|
|
usersubs = usersubs.filter(Q(unread_count_neutral__gt=0) |
|
|
|
|
Q(unread_count_positive__gt=0))
|
2013-06-27 15:21:13 -07:00
|
|
|
if not feed_ids:
|
2013-06-27 17:21:01 -07:00
|
|
|
usersubs = usersubs.filter(user=user_id,
|
2013-06-28 18:30:20 -07:00
|
|
|
active=True).only('feed', 'mark_read_date', 'is_trained')
|
2013-06-27 15:21:13 -07:00
|
|
|
else:
|
2013-06-27 17:21:01 -07:00
|
|
|
usersubs = usersubs.filter(user=user_id,
|
|
|
|
active=True,
|
2013-06-28 18:30:20 -07:00
|
|
|
feed__in=feed_ids).only('feed', 'mark_read_date', 'is_trained')
|
2013-06-27 17:21:01 -07:00
|
|
|
|
|
|
|
return usersubs
|
|
|
|
|
|
|
|
@classmethod
|
|
|
|
def story_hashes(cls, user_id, feed_ids=None, usersubs=None, read_filter="unread", order="newest",
|
2014-05-05 14:33:54 -07:00
|
|
|
include_timestamps=False, group_by_feed=True, cutoff_date=None,
|
|
|
|
across_all_feeds=True):
|
2013-06-27 17:21:01 -07:00
|
|
|
r = redis.Redis(connection_pool=settings.REDIS_STORY_HASH_POOL)
|
2013-06-27 15:21:13 -07:00
|
|
|
pipeline = r.pipeline()
|
2013-07-15 13:35:17 -07:00
|
|
|
story_hashes = {} if group_by_feed else []
|
2013-06-27 17:21:01 -07:00
|
|
|
|
2014-05-05 14:33:54 -07:00
|
|
|
if not feed_ids and not across_all_feeds:
|
|
|
|
return story_hashes
|
|
|
|
|
2013-06-27 17:21:01 -07:00
|
|
|
if not usersubs:
|
|
|
|
usersubs = cls.subs_for_feeds(user_id, feed_ids=feed_ids, read_filter=read_filter)
|
|
|
|
feed_ids = [sub.feed_id for sub in usersubs]
|
2013-07-01 16:50:09 -07:00
|
|
|
if not feed_ids:
|
2013-07-15 13:35:17 -07:00
|
|
|
return story_hashes
|
2013-06-27 17:21:01 -07:00
|
|
|
|
2013-06-27 16:23:29 -07:00
|
|
|
read_dates = dict((us.feed_id, int(us.mark_read_date.strftime('%s'))) for us in usersubs)
|
|
|
|
current_time = int(time.time() + 60*60*24)
|
2013-09-16 16:42:49 -07:00
|
|
|
if not cutoff_date:
|
|
|
|
cutoff_date = datetime.datetime.now() - datetime.timedelta(days=settings.DAYS_OF_STORY_HASHES)
|
|
|
|
unread_timestamp = int(time.mktime(cutoff_date.timetuple()))-1000
|
2013-06-28 19:09:34 -07:00
|
|
|
feed_counter = 0
|
|
|
|
|
|
|
|
for feed_id_group in chunks(feed_ids, 20):
|
|
|
|
pipeline = r.pipeline()
|
2013-06-28 19:10:59 -07:00
|
|
|
for feed_id in feed_id_group:
|
2013-06-28 19:09:34 -07:00
|
|
|
stories_key = 'F:%s' % feed_id
|
|
|
|
sorted_stories_key = 'zF:%s' % feed_id
|
|
|
|
read_stories_key = 'RS:%s:%s' % (user_id, feed_id)
|
|
|
|
unread_stories_key = 'U:%s:%s' % (user_id, feed_id)
|
|
|
|
unread_ranked_stories_key = 'zU:%s:%s' % (user_id, feed_id)
|
2013-07-01 08:21:53 -07:00
|
|
|
expire_unread_stories_key = False
|
2013-06-27 16:23:29 -07:00
|
|
|
|
2013-06-28 19:09:34 -07:00
|
|
|
max_score = current_time
|
|
|
|
if read_filter == 'unread':
|
|
|
|
# +1 for the intersection b/w zF and F, which carries an implicit score of 1.
|
|
|
|
min_score = read_dates[feed_id] + 1
|
|
|
|
pipeline.sdiffstore(unread_stories_key, stories_key, read_stories_key)
|
2013-07-01 08:21:53 -07:00
|
|
|
expire_unread_stories_key = True
|
2013-06-28 19:09:34 -07:00
|
|
|
else:
|
|
|
|
min_score = unread_timestamp
|
|
|
|
unread_stories_key = stories_key
|
2013-06-27 16:23:29 -07:00
|
|
|
|
2013-06-28 19:09:34 -07:00
|
|
|
if order == 'oldest':
|
|
|
|
byscorefunc = pipeline.zrangebyscore
|
|
|
|
else:
|
|
|
|
byscorefunc = pipeline.zrevrangebyscore
|
|
|
|
min_score, max_score = max_score, min_score
|
2013-06-27 16:23:29 -07:00
|
|
|
|
2013-06-28 19:09:34 -07:00
|
|
|
pipeline.zinterstore(unread_ranked_stories_key, [sorted_stories_key, unread_stories_key])
|
|
|
|
byscorefunc(unread_ranked_stories_key, min_score, max_score, withscores=include_timestamps)
|
2013-10-07 13:36:10 -07:00
|
|
|
pipeline.delete(unread_ranked_stories_key)
|
2013-07-01 08:21:53 -07:00
|
|
|
if expire_unread_stories_key:
|
|
|
|
pipeline.delete(unread_stories_key)
|
2013-10-07 13:36:10 -07:00
|
|
|
|
2013-06-27 15:21:13 -07:00
|
|
|
|
2013-06-28 19:09:34 -07:00
|
|
|
results = pipeline.execute()
|
2013-06-27 17:21:01 -07:00
|
|
|
|
2013-06-28 19:09:34 -07:00
|
|
|
for hashes in results:
|
|
|
|
if not isinstance(hashes, list): continue
|
|
|
|
if group_by_feed:
|
|
|
|
story_hashes[feed_ids[feed_counter]] = hashes
|
|
|
|
feed_counter += 1
|
|
|
|
else:
|
|
|
|
story_hashes.extend(hashes)
|
2013-10-07 13:36:10 -07:00
|
|
|
|
2013-06-27 16:23:29 -07:00
|
|
|
return story_hashes
|
2012-07-24 17:16:01 -07:00
|
|
|
|
2013-09-16 16:42:49 -07:00
|
|
|
def get_stories(self, offset=0, limit=6, order='newest', read_filter='all', withscores=False,
|
2013-12-05 15:10:09 -08:00
|
|
|
hashes_only=False, cutoff_date=None, default_cutoff_date=None):
|
2013-04-29 15:27:22 -07:00
|
|
|
r = redis.Redis(connection_pool=settings.REDIS_STORY_HASH_POOL)
|
2013-10-07 13:36:10 -07:00
|
|
|
rt = redis.Redis(connection_pool=settings.REDIS_STORY_HASH_TEMP_POOL)
|
2012-07-18 18:34:19 -07:00
|
|
|
ignore_user_stories = False
|
2013-04-29 15:27:22 -07:00
|
|
|
|
2012-07-16 18:11:18 -07:00
|
|
|
stories_key = 'F:%s' % (self.feed_id)
|
|
|
|
read_stories_key = 'RS:%s:%s' % (self.user_id, self.feed_id)
|
|
|
|
unread_stories_key = 'U:%s:%s' % (self.user_id, self.feed_id)
|
2012-07-16 20:49:43 -07:00
|
|
|
|
2013-05-09 16:16:14 -07:00
|
|
|
unread_ranked_stories_key = 'z%sU:%s:%s' % ('h' if hashes_only else '',
|
2013-05-02 16:49:44 -07:00
|
|
|
self.user_id, self.feed_id)
|
2013-10-07 13:36:10 -07:00
|
|
|
if withscores or not offset or not rt.exists(unread_ranked_stories_key):
|
|
|
|
rt.delete(unread_ranked_stories_key)
|
2012-12-06 17:33:56 -08:00
|
|
|
if not r.exists(stories_key):
|
2013-06-25 15:41:42 -07:00
|
|
|
# print " ---> No stories on feed: %s" % self
|
2012-12-06 17:33:56 -08:00
|
|
|
return []
|
2013-10-07 13:36:10 -07:00
|
|
|
elif read_filter == 'all' or not r.exists(read_stories_key):
|
2012-12-06 17:33:56 -08:00
|
|
|
ignore_user_stories = True
|
|
|
|
unread_stories_key = stories_key
|
|
|
|
else:
|
|
|
|
r.sdiffstore(unread_stories_key, stories_key, read_stories_key)
|
|
|
|
sorted_stories_key = 'zF:%s' % (self.feed_id)
|
|
|
|
r.zinterstore(unread_ranked_stories_key, [sorted_stories_key, unread_stories_key])
|
2013-10-07 13:36:10 -07:00
|
|
|
if not ignore_user_stories:
|
|
|
|
r.delete(unread_stories_key)
|
|
|
|
|
|
|
|
dump = r.dump(unread_ranked_stories_key)
|
2013-10-07 15:00:12 -07:00
|
|
|
if dump:
|
2013-10-07 16:55:21 -07:00
|
|
|
pipeline = rt.pipeline()
|
|
|
|
pipeline.delete(unread_ranked_stories_key)
|
2013-10-08 10:05:28 -07:00
|
|
|
pipeline.restore(unread_ranked_stories_key, 1*60*60*1000, dump)
|
2013-10-07 16:55:21 -07:00
|
|
|
pipeline.execute()
|
2013-10-07 15:00:12 -07:00
|
|
|
r.delete(unread_ranked_stories_key)
|
2012-07-16 18:11:18 -07:00
|
|
|
|
2013-10-07 13:36:10 -07:00
|
|
|
current_time = int(time.time() + 60*60*24)
|
2013-09-16 16:42:49 -07:00
|
|
|
if not cutoff_date:
|
2013-12-05 14:25:15 -08:00
|
|
|
if read_filter == "unread":
|
|
|
|
cutoff_date = self.mark_read_date
|
2013-12-05 15:10:09 -08:00
|
|
|
elif default_cutoff_date:
|
|
|
|
cutoff_date = default_cutoff_date
|
2013-12-05 14:25:15 -08:00
|
|
|
else:
|
|
|
|
cutoff_date = datetime.datetime.now() - datetime.timedelta(days=settings.DAYS_OF_UNREAD)
|
2013-09-16 16:42:49 -07:00
|
|
|
|
2012-07-18 18:34:19 -07:00
|
|
|
if order == 'oldest':
|
2013-10-07 13:36:10 -07:00
|
|
|
byscorefunc = rt.zrangebyscore
|
2013-01-10 12:51:39 -08:00
|
|
|
if read_filter == 'unread':
|
2013-12-05 14:25:15 -08:00
|
|
|
min_score = int(time.mktime(cutoff_date.timetuple())) + 1
|
2012-08-02 11:04:11 -07:00
|
|
|
else:
|
2013-12-05 14:25:15 -08:00
|
|
|
min_score = int(time.mktime(cutoff_date.timetuple())) - 1000
|
2012-07-18 18:34:19 -07:00
|
|
|
max_score = current_time
|
|
|
|
else:
|
2013-10-07 13:36:10 -07:00
|
|
|
byscorefunc = rt.zrevrangebyscore
|
2012-07-18 18:34:19 -07:00
|
|
|
min_score = current_time
|
2012-09-04 11:46:41 -07:00
|
|
|
if read_filter == 'unread':
|
|
|
|
# +1 for the intersection b/w zF and F, which carries an implicit score of 1.
|
2013-12-05 14:25:15 -08:00
|
|
|
max_score = int(time.mktime(cutoff_date.timetuple())) + 1
|
2012-09-04 11:46:41 -07:00
|
|
|
else:
|
|
|
|
max_score = 0
|
2013-06-25 15:41:42 -07:00
|
|
|
|
2013-06-15 12:48:35 -07:00
|
|
|
if settings.DEBUG and False:
|
2013-10-07 13:36:10 -07:00
|
|
|
debug_stories = rt.zrevrange(unread_ranked_stories_key, 0, -1, withscores=True)
|
2012-09-04 11:46:41 -07:00
|
|
|
print " ---> Unread all stories (%s - %s) %s stories: %s" % (
|
|
|
|
min_score,
|
|
|
|
max_score,
|
|
|
|
len(debug_stories),
|
|
|
|
debug_stories)
|
2012-07-18 18:34:19 -07:00
|
|
|
story_ids = byscorefunc(unread_ranked_stories_key, min_score,
|
2012-12-06 17:33:56 -08:00
|
|
|
max_score, start=offset, num=500,
|
|
|
|
withscores=withscores)[:limit]
|
2013-10-07 13:36:10 -07:00
|
|
|
|
2013-06-16 21:39:38 -07:00
|
|
|
if withscores:
|
|
|
|
story_ids = [(s[0], int(s[1])) for s in story_ids]
|
|
|
|
|
2013-05-09 16:16:14 -07:00
|
|
|
if withscores or hashes_only:
|
2012-09-04 11:46:41 -07:00
|
|
|
return story_ids
|
|
|
|
elif story_ids:
|
2012-11-27 16:44:10 -08:00
|
|
|
story_date_order = "%sstory_date" % ('' if order == 'oldest' else '-')
|
2013-04-29 15:27:22 -07:00
|
|
|
mstories = MStory.objects(story_hash__in=story_ids).order_by(story_date_order)
|
2012-09-04 11:46:41 -07:00
|
|
|
stories = Feed.format_stories(mstories)
|
|
|
|
return stories
|
|
|
|
else:
|
|
|
|
return []
|
2012-07-16 18:11:18 -07:00
|
|
|
|
|
|
|
@classmethod
|
2013-06-27 17:21:01 -07:00
|
|
|
def feed_stories(cls, user_id, feed_ids=None, offset=0, limit=6,
|
2013-10-02 13:08:19 -07:00
|
|
|
order='newest', read_filter='all', usersubs=None, cutoff_date=None,
|
|
|
|
all_feed_ids=None):
|
2013-10-07 13:36:10 -07:00
|
|
|
rt = redis.Redis(connection_pool=settings.REDIS_STORY_HASH_TEMP_POOL)
|
2014-05-05 14:33:54 -07:00
|
|
|
across_all_feeds = False
|
2012-07-18 18:34:19 -07:00
|
|
|
|
|
|
|
if order == 'oldest':
|
2013-10-07 13:36:10 -07:00
|
|
|
range_func = rt.zrange
|
2012-07-18 18:34:19 -07:00
|
|
|
else:
|
2013-10-07 13:36:10 -07:00
|
|
|
range_func = rt.zrevrange
|
2013-09-06 12:42:39 -07:00
|
|
|
|
2014-05-05 14:33:54 -07:00
|
|
|
if feed_ids is None:
|
|
|
|
across_all_feeds = True
|
2013-09-06 12:42:39 -07:00
|
|
|
feed_ids = []
|
2013-10-02 13:08:19 -07:00
|
|
|
if not all_feed_ids:
|
|
|
|
all_feed_ids = [f for f in feed_ids]
|
2013-09-06 12:42:39 -07:00
|
|
|
|
|
|
|
# feeds_string = ""
|
2013-10-02 13:08:19 -07:00
|
|
|
feeds_string = ','.join(str(f) for f in sorted(all_feed_ids))[:30]
|
|
|
|
ranked_stories_keys = 'zU:%s:feeds:%s' % (user_id, feeds_string)
|
2013-09-06 12:42:39 -07:00
|
|
|
unread_ranked_stories_keys = 'zhU:%s:feeds:%s' % (user_id, feeds_string)
|
2013-10-07 13:36:10 -07:00
|
|
|
stories_cached = rt.exists(ranked_stories_keys)
|
|
|
|
unreads_cached = True if read_filter == "unread" else rt.exists(unread_ranked_stories_keys)
|
2013-06-28 12:01:22 -07:00
|
|
|
if offset and stories_cached and unreads_cached:
|
2013-05-02 16:49:44 -07:00
|
|
|
story_hashes = range_func(ranked_stories_keys, offset, limit)
|
2013-06-27 17:21:01 -07:00
|
|
|
if read_filter == "unread":
|
|
|
|
unread_story_hashes = story_hashes
|
|
|
|
else:
|
2013-06-28 12:01:22 -07:00
|
|
|
unread_story_hashes = range_func(unread_ranked_stories_keys, 0, offset+limit)
|
2013-05-02 16:49:44 -07:00
|
|
|
return story_hashes, unread_story_hashes
|
2012-07-16 19:13:32 -07:00
|
|
|
else:
|
2013-10-07 13:36:10 -07:00
|
|
|
rt.delete(ranked_stories_keys)
|
|
|
|
rt.delete(unread_ranked_stories_keys)
|
2013-05-02 16:49:44 -07:00
|
|
|
|
2013-06-27 17:53:47 -07:00
|
|
|
story_hashes = cls.story_hashes(user_id, feed_ids=feed_ids,
|
|
|
|
read_filter=read_filter, order=order,
|
|
|
|
include_timestamps=True,
|
2013-10-07 10:51:28 -07:00
|
|
|
group_by_feed=False,
|
|
|
|
usersubs=usersubs,
|
2014-05-05 14:33:54 -07:00
|
|
|
cutoff_date=cutoff_date,
|
|
|
|
across_all_feeds=across_all_feeds)
|
2013-06-28 15:07:05 -07:00
|
|
|
if not story_hashes:
|
|
|
|
return [], []
|
|
|
|
|
2013-10-07 13:36:10 -07:00
|
|
|
pipeline = rt.pipeline()
|
2013-06-28 19:17:07 -07:00
|
|
|
for story_hash_group in chunks(story_hashes, 100):
|
2013-10-07 13:36:10 -07:00
|
|
|
pipeline.zadd(ranked_stories_keys, **dict(story_hash_group))
|
|
|
|
pipeline.execute()
|
2013-05-02 16:49:44 -07:00
|
|
|
story_hashes = range_func(ranked_stories_keys, offset, limit)
|
2013-06-27 17:21:01 -07:00
|
|
|
|
|
|
|
if read_filter == "unread":
|
|
|
|
unread_feed_story_hashes = story_hashes
|
2013-10-07 13:48:44 -07:00
|
|
|
rt.zunionstore(unread_ranked_stories_keys, [ranked_stories_keys])
|
2013-06-27 17:21:01 -07:00
|
|
|
else:
|
2013-06-27 17:53:47 -07:00
|
|
|
unread_story_hashes = cls.story_hashes(user_id, feed_ids=feed_ids,
|
|
|
|
read_filter="unread", order=order,
|
|
|
|
include_timestamps=True,
|
2013-09-16 16:42:49 -07:00
|
|
|
group_by_feed=False,
|
|
|
|
cutoff_date=cutoff_date)
|
2013-06-28 18:30:20 -07:00
|
|
|
if unread_story_hashes:
|
2013-06-28 19:25:24 -07:00
|
|
|
for unread_story_hash_group in chunks(unread_story_hashes, 100):
|
2013-10-07 13:36:10 -07:00
|
|
|
rt.zadd(unread_ranked_stories_keys, **dict(unread_story_hash_group))
|
2013-06-27 17:21:01 -07:00
|
|
|
unread_feed_story_hashes = range_func(unread_ranked_stories_keys, offset, limit)
|
|
|
|
|
2013-10-07 13:36:10 -07:00
|
|
|
rt.expire(ranked_stories_keys, 60*60)
|
|
|
|
rt.expire(unread_ranked_stories_keys, 60*60)
|
2012-07-16 18:11:18 -07:00
|
|
|
|
2013-05-02 16:49:44 -07:00
|
|
|
return story_hashes, unread_feed_story_hashes
|
2012-07-16 18:11:18 -07:00
|
|
|
|
2011-01-21 20:29:19 -05:00
|
|
|
@classmethod
|
2012-08-12 20:34:30 -07:00
|
|
|
def add_subscription(cls, user, feed_address, folder=None, bookmarklet=False, auto_active=True,
|
|
|
|
skip_fetch=False):
|
2011-01-21 20:29:19 -05:00
|
|
|
feed = None
|
2011-01-22 11:22:14 -05:00
|
|
|
us = None
|
2011-01-21 20:29:19 -05:00
|
|
|
|
2012-07-29 00:49:34 -07:00
|
|
|
logging.user(user, "~FRAdding URL: ~SB%s (in %s) %s" % (feed_address, folder,
|
|
|
|
"~FCAUTO-ADD" if not auto_active else ""))
|
2011-01-21 20:29:19 -05:00
|
|
|
|
2011-02-08 22:07:59 -05:00
|
|
|
feed = Feed.get_feed_from_url(feed_address)
|
2011-01-22 11:22:14 -05:00
|
|
|
|
2011-01-21 20:29:19 -05:00
|
|
|
if not feed:
|
|
|
|
code = -1
|
2011-01-22 21:42:58 -05:00
|
|
|
if bookmarklet:
|
|
|
|
message = "This site does not have an RSS feed. Nothing is linked to from this page."
|
|
|
|
else:
|
2011-02-08 22:07:59 -05:00
|
|
|
message = "This address does not point to an RSS feed or a website with an RSS feed."
|
2011-01-21 20:29:19 -05:00
|
|
|
else:
|
2011-01-22 18:25:16 -05:00
|
|
|
us, subscription_created = cls.objects.get_or_create(
|
2011-01-21 20:29:19 -05:00
|
|
|
feed=feed,
|
|
|
|
user=user,
|
|
|
|
defaults={
|
|
|
|
'needs_unread_recalc': True,
|
2012-03-19 15:46:59 -07:00
|
|
|
'active': auto_active,
|
2011-01-21 20:29:19 -05:00
|
|
|
}
|
|
|
|
)
|
|
|
|
code = 1
|
|
|
|
message = ""
|
|
|
|
|
2011-02-14 01:15:45 -05:00
|
|
|
if us:
|
2011-01-21 20:29:19 -05:00
|
|
|
user_sub_folders_object, created = UserSubscriptionFolders.objects.get_or_create(
|
|
|
|
user=user,
|
|
|
|
defaults={'folders': '[]'}
|
|
|
|
)
|
|
|
|
if created:
|
|
|
|
user_sub_folders = []
|
|
|
|
else:
|
|
|
|
user_sub_folders = json.decode(user_sub_folders_object.folders)
|
2011-01-23 02:13:55 -05:00
|
|
|
user_sub_folders = add_object_to_folder(feed.pk, folder, user_sub_folders)
|
2011-01-21 20:29:19 -05:00
|
|
|
user_sub_folders_object.folders = json.encode(user_sub_folders)
|
|
|
|
user_sub_folders_object.save()
|
2012-03-19 16:35:56 -07:00
|
|
|
|
2012-07-29 00:49:34 -07:00
|
|
|
if auto_active or user.profile.is_premium:
|
2012-04-11 12:14:18 -07:00
|
|
|
us.active = True
|
2012-07-29 00:49:34 -07:00
|
|
|
us.save()
|
2011-01-21 20:29:19 -05:00
|
|
|
|
2012-08-12 20:34:30 -07:00
|
|
|
if not skip_fetch and feed.last_update < datetime.datetime.utcnow() - datetime.timedelta(days=1):
|
2012-04-11 12:14:18 -07:00
|
|
|
feed = feed.update()
|
2012-04-16 11:21:52 -07:00
|
|
|
|
|
|
|
from apps.social.models import MActivity
|
2012-04-11 15:20:57 -07:00
|
|
|
MActivity.new_feed_subscription(user_id=user.pk, feed_id=feed.pk, feed_title=feed.title)
|
2012-12-10 18:36:54 -08:00
|
|
|
|
2012-04-11 12:14:18 -07:00
|
|
|
feed.setup_feed_for_premium_subscribers()
|
2012-03-19 16:35:56 -07:00
|
|
|
|
2011-01-22 11:22:14 -05:00
|
|
|
return code, message, us
|
2012-03-26 11:04:05 -07:00
|
|
|
|
|
|
|
@classmethod
|
2013-07-31 11:17:04 -07:00
|
|
|
def feeds_with_updated_counts(cls, user, feed_ids=None, check_fetch_status=False, force=False):
|
2012-03-26 11:04:05 -07:00
|
|
|
feeds = {}
|
|
|
|
|
|
|
|
# Get subscriptions for user
|
|
|
|
user_subs = cls.objects.select_related('feed').filter(user=user, active=True)
|
|
|
|
feed_ids = [f for f in feed_ids if f and not f.startswith('river')]
|
|
|
|
if feed_ids:
|
|
|
|
user_subs = user_subs.filter(feed__in=feed_ids)
|
|
|
|
|
|
|
|
for i, sub in enumerate(user_subs):
|
|
|
|
# Count unreads if subscription is stale.
|
2013-07-31 11:17:04 -07:00
|
|
|
if (force or
|
|
|
|
sub.needs_unread_recalc or
|
2013-09-16 16:42:49 -07:00
|
|
|
sub.unread_count_updated < user.profile.unread_cutoff or
|
|
|
|
sub.oldest_unread_story_date < user.profile.unread_cutoff):
|
2013-07-31 11:17:04 -07:00
|
|
|
sub = sub.calculate_feed_scores(silent=True, force=force)
|
2012-03-26 11:04:05 -07:00
|
|
|
if not sub: continue # TODO: Figure out the correct sub and give it a new feed_id
|
2011-01-21 20:29:19 -05:00
|
|
|
|
2012-03-30 16:03:07 -07:00
|
|
|
feed_id = sub.feed_id
|
2012-03-26 11:04:05 -07:00
|
|
|
feeds[feed_id] = {
|
|
|
|
'ps': sub.unread_count_positive,
|
|
|
|
'nt': sub.unread_count_neutral,
|
|
|
|
'ng': sub.unread_count_negative,
|
|
|
|
'id': feed_id,
|
|
|
|
}
|
2012-03-30 16:03:07 -07:00
|
|
|
if not sub.feed.fetched_once or check_fetch_status:
|
2012-05-22 17:39:21 -07:00
|
|
|
feeds[feed_id]['fetched_once'] = sub.feed.fetched_once
|
|
|
|
feeds[feed_id]['not_yet_fetched'] = not sub.feed.fetched_once # Legacy. Dammit.
|
2012-03-26 11:04:05 -07:00
|
|
|
if sub.feed.favicon_fetching:
|
|
|
|
feeds[feed_id]['favicon_fetching'] = True
|
|
|
|
if sub.feed.has_feed_exception or sub.feed.has_page_exception:
|
|
|
|
feeds[feed_id]['has_exception'] = True
|
|
|
|
feeds[feed_id]['exception_type'] = 'feed' if sub.feed.has_feed_exception else 'page'
|
|
|
|
feeds[feed_id]['feed_address'] = sub.feed.feed_address
|
|
|
|
feeds[feed_id]['exception_code'] = sub.feed.exception_code
|
|
|
|
|
|
|
|
return feeds
|
2013-07-02 12:02:41 -07:00
|
|
|
|
2013-07-02 12:10:34 -07:00
|
|
|
def trim_read_stories(self, r=None):
|
2013-07-02 12:02:41 -07:00
|
|
|
if not r:
|
|
|
|
r = redis.Redis(connection_pool=settings.REDIS_STORY_HASH_POOL)
|
|
|
|
|
|
|
|
read_stories_key = "RS:%s:%s" % (self.user_id, self.feed_id)
|
|
|
|
stale_story_hashes = r.sdiff(read_stories_key, "F:%s" % self.feed_id)
|
|
|
|
if not stale_story_hashes:
|
|
|
|
return
|
2012-03-26 11:04:05 -07:00
|
|
|
|
2013-07-02 12:47:49 -07:00
|
|
|
logging.user(self.user, "~FBTrimming ~FR%s~FB read stories (~SB%s~SN)..." % (len(stale_story_hashes), self.feed_id))
|
2013-07-02 12:02:41 -07:00
|
|
|
r.srem(read_stories_key, *stale_story_hashes)
|
|
|
|
r.srem("RS:%s" % self.feed_id, *stale_story_hashes)
|
|
|
|
|
2013-10-07 13:36:10 -07:00
|
|
|
@classmethod
|
|
|
|
def trim_user_read_stories(self, user_id):
|
2014-01-21 13:34:53 -08:00
|
|
|
user = User.objects.get(pk=user_id)
|
2013-10-07 13:36:10 -07:00
|
|
|
r = redis.Redis(connection_pool=settings.REDIS_STORY_HASH_POOL)
|
2013-10-07 14:46:07 -07:00
|
|
|
subs = UserSubscription.objects.filter(user_id=user_id).only('feed')
|
2013-10-07 14:47:36 -07:00
|
|
|
if not subs: return
|
2014-01-21 13:34:53 -08:00
|
|
|
|
|
|
|
key = "RS:%s" % user_id
|
2013-10-07 13:36:10 -07:00
|
|
|
feeds = [f.feed_id for f in subs]
|
2014-01-21 13:34:53 -08:00
|
|
|
old_rs = r.smembers(key)
|
2013-10-07 14:47:36 -07:00
|
|
|
old_count = len(old_rs)
|
2014-01-21 15:00:47 -08:00
|
|
|
if not old_count:
|
2014-01-21 15:05:47 -08:00
|
|
|
logging.user(user, "~FBTrimming all read stories, ~SBnone found~SN.")
|
2014-01-21 15:00:47 -08:00
|
|
|
return
|
2014-01-21 13:34:53 -08:00
|
|
|
|
|
|
|
r.sunionstore("%s:backup" % key, key)
|
|
|
|
r.expire("%s:backup" % key, 60*60*24)
|
|
|
|
r.sunionstore(key, *["%s:%s" % (key, f) for f in feeds])
|
|
|
|
new_rs = r.smembers(key)
|
2013-10-07 13:36:10 -07:00
|
|
|
|
|
|
|
missing_rs = []
|
2014-01-21 13:51:13 -08:00
|
|
|
missing_count = 0
|
2013-10-07 13:36:10 -07:00
|
|
|
feed_re = re.compile(r'(\d+):.*?')
|
2014-01-21 13:44:16 -08:00
|
|
|
for i, rs in enumerate(old_rs):
|
2014-01-21 13:49:00 -08:00
|
|
|
if i and i % 1000 == 0:
|
2014-01-21 13:53:31 -08:00
|
|
|
if missing_rs:
|
|
|
|
r.sadd(key, *missing_rs)
|
2014-01-21 13:51:13 -08:00
|
|
|
missing_count += len(missing_rs)
|
2014-01-21 13:49:00 -08:00
|
|
|
missing_rs = []
|
2013-10-07 13:57:13 -07:00
|
|
|
found = feed_re.search(rs)
|
|
|
|
if not found:
|
|
|
|
print " ---> Not found: %s" % rs
|
|
|
|
continue
|
|
|
|
rs_feed_id = found.groups()[0]
|
2013-10-07 13:36:10 -07:00
|
|
|
if int(rs_feed_id) not in feeds:
|
|
|
|
missing_rs.append(rs)
|
2014-01-21 13:53:31 -08:00
|
|
|
|
|
|
|
if missing_rs:
|
|
|
|
r.sadd(key, *missing_rs)
|
2014-01-21 13:51:13 -08:00
|
|
|
missing_count += len(missing_rs)
|
2013-10-07 13:36:10 -07:00
|
|
|
new_count = len(new_rs)
|
|
|
|
new_total = new_count + missing_count
|
2014-03-13 16:53:20 -07:00
|
|
|
logging.user(user, "~FBTrimming ~FR%s~FB/%s (~SB%s sub'ed ~SN+ ~SB%s unsub'ed~SN saved)" %
|
2013-10-07 13:36:10 -07:00
|
|
|
(old_count - new_total, old_count, new_count, missing_count))
|
|
|
|
|
|
|
|
|
2013-09-13 15:02:10 -07:00
|
|
|
def mark_feed_read(self, cutoff_date=None):
|
2013-02-11 08:40:32 -08:00
|
|
|
if (self.unread_count_negative == 0
|
|
|
|
and self.unread_count_neutral == 0
|
|
|
|
and self.unread_count_positive == 0
|
|
|
|
and not self.needs_unread_recalc):
|
|
|
|
return
|
|
|
|
|
2013-09-13 15:02:10 -07:00
|
|
|
recount = True
|
2011-01-07 16:26:17 -05:00
|
|
|
# Use the latest story to get last read time.
|
2013-09-13 15:02:10 -07:00
|
|
|
if cutoff_date:
|
|
|
|
cutoff_date = cutoff_date + datetime.timedelta(seconds=1)
|
2010-06-14 01:01:21 -04:00
|
|
|
else:
|
2013-10-02 16:05:22 -07:00
|
|
|
latest_story = MStory.objects(story_feed_id=self.feed.pk)\
|
|
|
|
.order_by('-story_date').only('story_date').limit(1)
|
2013-09-13 15:02:10 -07:00
|
|
|
if latest_story and len(latest_story) >= 1:
|
|
|
|
cutoff_date = (latest_story[0]['story_date']
|
|
|
|
+ datetime.timedelta(seconds=1))
|
|
|
|
else:
|
|
|
|
cutoff_date = datetime.datetime.utcnow()
|
|
|
|
recount = False
|
2012-04-24 17:40:34 -07:00
|
|
|
|
2013-09-13 15:02:10 -07:00
|
|
|
self.last_read_date = cutoff_date
|
|
|
|
self.mark_read_date = cutoff_date
|
|
|
|
self.oldest_unread_story_date = cutoff_date
|
|
|
|
if not recount:
|
|
|
|
self.unread_count_negative = 0
|
|
|
|
self.unread_count_positive = 0
|
|
|
|
self.unread_count_neutral = 0
|
2013-09-13 17:21:05 -07:00
|
|
|
self.unread_count_updated = datetime.datetime.utcnow()
|
2013-09-13 15:02:10 -07:00
|
|
|
self.needs_unread_recalc = False
|
|
|
|
else:
|
|
|
|
self.needs_unread_recalc = True
|
2012-07-18 18:34:19 -07:00
|
|
|
|
2009-06-16 03:08:55 +00:00
|
|
|
self.save()
|
2011-11-05 16:25:04 -07:00
|
|
|
|
2013-05-13 14:22:49 -07:00
|
|
|
return True
|
|
|
|
|
2013-12-05 14:25:15 -08:00
|
|
|
def mark_newer_stories_read(self, cutoff_date):
|
|
|
|
if (self.unread_count_negative == 0
|
|
|
|
and self.unread_count_neutral == 0
|
|
|
|
and self.unread_count_positive == 0
|
|
|
|
and not self.needs_unread_recalc):
|
|
|
|
return
|
|
|
|
|
|
|
|
cutoff_date = cutoff_date - datetime.timedelta(seconds=1)
|
|
|
|
story_hashes = self.get_stories(limit=500, order="newest", cutoff_date=cutoff_date,
|
|
|
|
read_filter="unread", hashes_only=True)
|
|
|
|
data = self.mark_story_ids_as_read(story_hashes)
|
|
|
|
return data
|
|
|
|
|
|
|
|
|
2013-05-14 17:31:08 -07:00
|
|
|
def mark_story_ids_as_read(self, story_hashes, request=None):
|
|
|
|
data = dict(code=0, payload=story_hashes)
|
2011-11-05 16:25:04 -07:00
|
|
|
|
|
|
|
if not request:
|
|
|
|
request = self.user
|
|
|
|
|
|
|
|
if not self.needs_unread_recalc:
|
|
|
|
self.needs_unread_recalc = True
|
|
|
|
self.save()
|
|
|
|
|
2013-05-14 17:31:08 -07:00
|
|
|
if len(story_hashes) > 1:
|
|
|
|
logging.user(request, "~FYRead %s stories in feed: %s" % (len(story_hashes), self.feed))
|
2011-11-05 16:25:04 -07:00
|
|
|
else:
|
|
|
|
logging.user(request, "~FYRead story in feed: %s" % (self.feed))
|
|
|
|
|
2013-05-14 17:31:08 -07:00
|
|
|
for story_hash in set(story_hashes):
|
|
|
|
RUserStory.mark_read(self.user_id, self.feed_id, story_hash)
|
2013-05-02 16:49:44 -07:00
|
|
|
|
2011-11-05 16:25:04 -07:00
|
|
|
return data
|
2009-06-16 03:08:55 +00:00
|
|
|
|
2014-03-28 13:49:03 -07:00
|
|
|
def invert_read_stories_after_unread_story(self, story, request=None):
|
|
|
|
data = dict(code=1)
|
|
|
|
if story.story_date > self.mark_read_date:
|
|
|
|
return data
|
|
|
|
|
|
|
|
# Story is outside the mark as read range, so invert all stories before.
|
|
|
|
newer_stories = MStory.objects(story_feed_id=story.story_feed_id,
|
|
|
|
story_date__gte=story.story_date,
|
|
|
|
story_date__lte=self.mark_read_date
|
|
|
|
).only('story_hash')
|
|
|
|
newer_stories = [s.story_hash for s in newer_stories]
|
|
|
|
self.mark_read_date = story.story_date - datetime.timedelta(minutes=1)
|
|
|
|
self.needs_unread_recalc = True
|
|
|
|
self.save()
|
|
|
|
|
|
|
|
# Mark stories as read only after the mark_read_date has been moved, otherwise
|
|
|
|
# these would be ignored.
|
|
|
|
data = self.mark_story_ids_as_read(newer_stories, request=request)
|
|
|
|
|
|
|
|
return data
|
|
|
|
|
2013-05-03 17:07:45 -07:00
|
|
|
def calculate_feed_scores(self, silent=False, stories=None, force=False):
|
2011-10-10 20:46:13 -07:00
|
|
|
# now = datetime.datetime.strptime("2009-07-06 22:30:03", "%Y-%m-%d %H:%M:%S")
|
|
|
|
now = datetime.datetime.now()
|
2013-06-21 14:15:13 -07:00
|
|
|
oldest_unread_story_date = now
|
2013-09-16 16:42:49 -07:00
|
|
|
|
|
|
|
if self.user.profile.last_seen_on < self.user.profile.unread_cutoff and not force:
|
2010-09-23 16:33:06 -04:00
|
|
|
# if not silent:
|
|
|
|
# logging.info(' ---> [%s] SKIPPING Computing scores: %s (1 week+)' % (self.user, self.feed))
|
2013-06-21 12:30:06 -07:00
|
|
|
return self
|
|
|
|
ong = self.unread_count_negative
|
|
|
|
ont = self.unread_count_neutral
|
|
|
|
ops = self.unread_count_positive
|
2010-08-19 10:43:07 -04:00
|
|
|
|
2013-05-03 17:07:45 -07:00
|
|
|
# if not self.feed.fetched_once:
|
|
|
|
# if not silent:
|
|
|
|
# logging.info(' ---> [%s] NOT Computing scores: %s' % (self.user, self.feed))
|
|
|
|
# self.needs_unread_recalc = False
|
|
|
|
# self.save()
|
|
|
|
# return
|
2011-10-10 20:46:13 -07:00
|
|
|
|
2010-01-21 13:12:29 -05:00
|
|
|
feed_scores = dict(negative=0, neutral=0, positive=0)
|
2010-02-17 00:53:05 -05:00
|
|
|
|
|
|
|
# Two weeks in age. If mark_read_date is older, mark old stories as read.
|
2013-09-16 16:42:49 -07:00
|
|
|
date_delta = self.user.profile.unread_cutoff
|
2010-01-21 13:12:29 -05:00
|
|
|
if date_delta < self.mark_read_date:
|
|
|
|
date_delta = self.mark_read_date
|
2010-02-17 00:53:05 -05:00
|
|
|
else:
|
|
|
|
self.mark_read_date = date_delta
|
2012-10-29 12:25:28 -07:00
|
|
|
|
2013-06-21 14:15:13 -07:00
|
|
|
if self.is_trained:
|
|
|
|
if not stories:
|
|
|
|
stories = cache.get('S:%s' % self.feed_id)
|
2013-05-03 09:43:21 -07:00
|
|
|
|
2013-10-08 19:42:08 -07:00
|
|
|
unread_story_hashes = self.story_hashes(user_id=self.user_id, feed_ids=[self.feed_id],
|
2013-10-11 20:02:16 -07:00
|
|
|
usersubs=[self],
|
2013-10-08 19:42:08 -07:00
|
|
|
read_filter='unread', group_by_feed=False,
|
|
|
|
cutoff_date=self.user.profile.unread_cutoff)
|
2013-06-21 14:15:13 -07:00
|
|
|
|
|
|
|
if not stories:
|
|
|
|
stories_db = MStory.objects(story_hash__in=unread_story_hashes)
|
|
|
|
stories = Feed.format_stories(stories_db, self.feed_id)
|
|
|
|
|
|
|
|
unread_stories = []
|
|
|
|
for story in stories:
|
|
|
|
if story['story_date'] < date_delta:
|
|
|
|
continue
|
|
|
|
if story['story_hash'] in unread_story_hashes:
|
|
|
|
unread_stories.append(story)
|
|
|
|
if story['story_date'] < oldest_unread_story_date:
|
|
|
|
oldest_unread_story_date = story['story_date']
|
2012-10-29 12:25:28 -07:00
|
|
|
|
2013-06-21 14:15:13 -07:00
|
|
|
# if not silent:
|
|
|
|
# logging.info(' ---> [%s] Format stories: %s' % (self.user, datetime.datetime.now() - now))
|
2010-08-21 23:49:36 -04:00
|
|
|
|
2013-06-21 14:15:13 -07:00
|
|
|
classifier_feeds = list(MClassifierFeed.objects(user_id=self.user_id, feed_id=self.feed_id, social_user_id=0))
|
|
|
|
classifier_authors = list(MClassifierAuthor.objects(user_id=self.user_id, feed_id=self.feed_id))
|
|
|
|
classifier_titles = list(MClassifierTitle.objects(user_id=self.user_id, feed_id=self.feed_id))
|
|
|
|
classifier_tags = list(MClassifierTag.objects(user_id=self.user_id, feed_id=self.feed_id))
|
|
|
|
|
|
|
|
if (not len(classifier_feeds) and
|
|
|
|
not len(classifier_authors) and
|
|
|
|
not len(classifier_titles) and
|
|
|
|
not len(classifier_tags)):
|
|
|
|
self.is_trained = False
|
|
|
|
|
|
|
|
# if not silent:
|
|
|
|
# logging.info(' ---> [%s] Classifiers: %s (%s)' % (self.user, datetime.datetime.now() - now, classifier_feeds.count() + classifier_authors.count() + classifier_tags.count() + classifier_titles.count()))
|
2010-09-18 16:15:59 -04:00
|
|
|
|
2013-06-21 14:15:13 -07:00
|
|
|
scores = {
|
|
|
|
'feed': apply_classifier_feeds(classifier_feeds, self.feed),
|
|
|
|
}
|
2010-01-21 13:12:29 -05:00
|
|
|
|
2013-06-21 14:15:13 -07:00
|
|
|
for story in unread_stories:
|
|
|
|
scores.update({
|
|
|
|
'author' : apply_classifier_authors(classifier_authors, story),
|
|
|
|
'tags' : apply_classifier_tags(classifier_tags, story),
|
|
|
|
'title' : apply_classifier_titles(classifier_titles, story),
|
|
|
|
})
|
2010-01-21 13:12:29 -05:00
|
|
|
|
2013-06-21 14:15:13 -07:00
|
|
|
max_score = max(scores['author'], scores['tags'], scores['title'])
|
|
|
|
min_score = min(scores['author'], scores['tags'], scores['title'])
|
|
|
|
if max_score > 0:
|
2010-10-13 18:18:37 -04:00
|
|
|
feed_scores['positive'] += 1
|
2013-06-21 14:15:13 -07:00
|
|
|
elif min_score < 0:
|
2010-10-13 18:18:37 -04:00
|
|
|
feed_scores['negative'] += 1
|
|
|
|
else:
|
2013-06-21 14:15:13 -07:00
|
|
|
if scores['feed'] > 0:
|
|
|
|
feed_scores['positive'] += 1
|
|
|
|
elif scores['feed'] < 0:
|
|
|
|
feed_scores['negative'] += 1
|
|
|
|
else:
|
|
|
|
feed_scores['neutral'] += 1
|
|
|
|
else:
|
2013-10-08 19:42:08 -07:00
|
|
|
unread_story_hashes = self.story_hashes(user_id=self.user_id, feed_ids=[self.feed_id],
|
2013-10-11 20:02:16 -07:00
|
|
|
usersubs=[self],
|
2013-10-08 19:42:08 -07:00
|
|
|
read_filter='unread', group_by_feed=False,
|
|
|
|
include_timestamps=True,
|
2014-03-28 13:56:47 -07:00
|
|
|
cutoff_date=date_delta)
|
2013-10-08 19:42:08 -07:00
|
|
|
|
2013-06-21 14:15:13 -07:00
|
|
|
feed_scores['neutral'] = len(unread_story_hashes)
|
|
|
|
if feed_scores['neutral']:
|
|
|
|
oldest_unread_story_date = datetime.datetime.fromtimestamp(unread_story_hashes[-1][1])
|
2010-01-21 13:12:29 -05:00
|
|
|
|
2014-03-28 14:00:36 -07:00
|
|
|
if not silent or settings.DEBUG:
|
2013-06-21 14:24:34 -07:00
|
|
|
logging.user(self.user, '~FBUnread count (~SB%s~SN%s): ~SN(~FC%s~FB/~FC%s~FB/~FC%s~FB) ~SBto~SN (~FC%s~FB/~FC%s~FB/~FC%s~FB)' % (self.feed_id, '/~FMtrained~FB' if self.is_trained else '', ong, ont, ops, feed_scores['negative'], feed_scores['neutral'], feed_scores['positive']))
|
2012-06-18 15:59:31 -07:00
|
|
|
|
2010-01-21 13:12:29 -05:00
|
|
|
self.unread_count_positive = feed_scores['positive']
|
|
|
|
self.unread_count_neutral = feed_scores['neutral']
|
|
|
|
self.unread_count_negative = feed_scores['negative']
|
2011-01-07 16:26:17 -05:00
|
|
|
self.unread_count_updated = datetime.datetime.now()
|
2011-01-15 18:41:41 -05:00
|
|
|
self.oldest_unread_story_date = oldest_unread_story_date
|
2010-04-08 18:36:48 -04:00
|
|
|
self.needs_unread_recalc = False
|
2009-08-29 19:34:42 +00:00
|
|
|
|
|
|
|
self.save()
|
2011-03-17 22:15:29 -04:00
|
|
|
|
2011-11-11 11:26:34 -08:00
|
|
|
if (self.unread_count_positive == 0 and
|
2013-05-16 11:21:20 -07:00
|
|
|
self.unread_count_neutral == 0):
|
2011-11-11 11:26:34 -08:00
|
|
|
self.mark_feed_read()
|
2010-07-06 16:37:49 -04:00
|
|
|
|
2011-10-10 20:46:13 -07:00
|
|
|
if not silent:
|
2013-01-04 16:44:41 -08:00
|
|
|
logging.user(self.user, '~FC~SNComputing scores: %s (~SB%s~SN/~SB%s~SN/~SB%s~SN)' % (self.feed, feed_scores['negative'], feed_scores['neutral'], feed_scores['positive']))
|
2013-07-02 12:08:22 -07:00
|
|
|
|
2013-07-02 12:10:34 -07:00
|
|
|
self.trim_read_stories()
|
2013-07-02 12:08:22 -07:00
|
|
|
|
2011-11-28 18:01:39 -08:00
|
|
|
return self
|
2011-11-16 10:00:03 -08:00
|
|
|
|
|
|
|
def switch_feed(self, new_feed, old_feed):
|
|
|
|
# Rewrite feed in subscription folders
|
|
|
|
try:
|
|
|
|
user_sub_folders = UserSubscriptionFolders.objects.get(user=self.user)
|
|
|
|
except Exception, e:
|
|
|
|
logging.info(" *** ---> UserSubscriptionFolders error: %s" % e)
|
|
|
|
return
|
|
|
|
|
|
|
|
logging.info(" ===> %s " % self.user)
|
2013-05-09 14:56:51 -07:00
|
|
|
|
2011-11-29 09:43:16 -08:00
|
|
|
# Switch read stories
|
2013-05-14 17:31:08 -07:00
|
|
|
RUserStory.switch_feed(user_id=self.user_id, old_feed_id=old_feed.pk,
|
|
|
|
new_feed_id=new_feed.pk)
|
2013-05-09 14:56:51 -07:00
|
|
|
|
2011-11-29 09:43:16 -08:00
|
|
|
def switch_feed_for_classifier(model):
|
2012-01-26 09:32:24 -08:00
|
|
|
duplicates = model.objects(feed_id=old_feed.pk, user_id=self.user_id)
|
2011-11-29 09:43:16 -08:00
|
|
|
if duplicates.count():
|
|
|
|
logging.info(" ---> Switching %s %s" % (duplicates.count(), model))
|
|
|
|
for duplicate in duplicates:
|
|
|
|
duplicate.feed_id = new_feed.pk
|
2012-12-12 15:41:44 -08:00
|
|
|
if duplicate.social_user_id is None:
|
|
|
|
duplicate.social_user_id = 0
|
2011-11-29 09:43:16 -08:00
|
|
|
try:
|
|
|
|
duplicate.save()
|
|
|
|
pass
|
|
|
|
except (IntegrityError, OperationError):
|
|
|
|
logging.info(" !!!!> %s already exists" % duplicate)
|
|
|
|
duplicate.delete()
|
|
|
|
|
|
|
|
switch_feed_for_classifier(MClassifierTitle)
|
|
|
|
switch_feed_for_classifier(MClassifierAuthor)
|
|
|
|
switch_feed_for_classifier(MClassifierFeed)
|
|
|
|
switch_feed_for_classifier(MClassifierTag)
|
2013-05-09 14:56:51 -07:00
|
|
|
|
|
|
|
# Switch to original feed for the user subscription
|
|
|
|
self.feed = new_feed
|
|
|
|
self.needs_unread_recalc = True
|
|
|
|
try:
|
2013-05-14 17:31:08 -07:00
|
|
|
UserSubscription.objects.get(user=self.user, feed=new_feed)
|
2013-05-09 14:56:51 -07:00
|
|
|
except UserSubscription.DoesNotExist:
|
|
|
|
self.save()
|
|
|
|
user_sub_folders.rewrite_feed(new_feed, old_feed)
|
|
|
|
else:
|
|
|
|
# except (IntegrityError, OperationError):
|
|
|
|
logging.info(" !!!!> %s already subscribed" % self.user)
|
|
|
|
self.delete()
|
|
|
|
return
|
2012-03-19 21:29:19 -07:00
|
|
|
|
|
|
|
@classmethod
|
|
|
|
def collect_orphan_feeds(cls, user):
|
|
|
|
us = cls.objects.filter(user=user)
|
2012-03-20 10:32:17 -07:00
|
|
|
try:
|
|
|
|
usf = UserSubscriptionFolders.objects.get(user=user)
|
|
|
|
except UserSubscriptionFolders.DoesNotExist:
|
|
|
|
return
|
2012-03-19 21:29:19 -07:00
|
|
|
us_feed_ids = set([sub.feed_id for sub in us])
|
|
|
|
folders = json.decode(usf.folders)
|
2011-11-29 09:43:16 -08:00
|
|
|
|
2012-03-19 21:29:19 -07:00
|
|
|
def collect_ids(folders, found_ids):
|
|
|
|
for item in folders:
|
|
|
|
# print ' --> %s' % item
|
|
|
|
if isinstance(item, int):
|
|
|
|
# print ' --> Adding feed: %s' % item
|
|
|
|
found_ids.add(item)
|
|
|
|
elif isinstance(item, dict):
|
|
|
|
# print ' --> Descending folder dict: %s' % item.values()
|
|
|
|
found_ids.update(collect_ids(item.values(), found_ids))
|
|
|
|
elif isinstance(item, list):
|
|
|
|
# print ' --> Descending folder list: %s' % len(item)
|
|
|
|
found_ids.update(collect_ids(item, found_ids))
|
|
|
|
# print ' --> Returning: %s' % found_ids
|
|
|
|
return found_ids
|
|
|
|
found_ids = collect_ids(folders, set())
|
2012-03-20 10:32:17 -07:00
|
|
|
diff = len(us_feed_ids) - len(found_ids)
|
|
|
|
if diff > 0:
|
|
|
|
logging.info(" ---> Collecting orphans on %s. %s feeds with %s orphans" % (user.username, len(us_feed_ids), diff))
|
|
|
|
orphan_ids = us_feed_ids - found_ids
|
|
|
|
folders.extend(list(orphan_ids))
|
|
|
|
usf.folders = json.encode(folders)
|
|
|
|
usf.save()
|
2012-03-19 21:29:19 -07:00
|
|
|
|
2012-03-20 10:33:15 -07:00
|
|
|
|
2013-05-02 16:49:44 -07:00
|
|
|
class RUserStory:
|
|
|
|
|
2013-07-17 14:49:54 -07:00
|
|
|
@classmethod
|
2013-09-06 13:20:08 -07:00
|
|
|
def mark_story_hashes_read(cls, user_id, story_hashes, r=None, s=None):
|
2013-07-17 14:49:54 -07:00
|
|
|
if not r:
|
|
|
|
r = redis.Redis(connection_pool=settings.REDIS_STORY_HASH_POOL)
|
2013-09-06 13:20:08 -07:00
|
|
|
if not s:
|
|
|
|
s = redis.Redis(connection_pool=settings.REDIS_POOL)
|
2013-08-14 14:32:50 -07:00
|
|
|
# if not r2:
|
|
|
|
# r2 = redis.Redis(connection_pool=settings.REDIS_STORY_HASH_POOL2)
|
2013-07-17 14:49:54 -07:00
|
|
|
|
|
|
|
p = r.pipeline()
|
2013-08-14 14:32:50 -07:00
|
|
|
# p2 = r2.pipeline()
|
2013-07-17 15:01:19 -07:00
|
|
|
feed_ids = set()
|
|
|
|
friend_ids = set()
|
2013-07-17 14:49:54 -07:00
|
|
|
|
|
|
|
if not isinstance(story_hashes, list):
|
|
|
|
story_hashes = [story_hashes]
|
|
|
|
|
|
|
|
for story_hash in story_hashes:
|
|
|
|
feed_id, _ = MStory.split_story_hash(story_hash)
|
2013-07-17 15:01:19 -07:00
|
|
|
feed_ids.add(feed_id)
|
2013-07-17 14:49:54 -07:00
|
|
|
|
|
|
|
# Find other social feeds with this story to update their counts
|
|
|
|
friend_key = "F:%s:F" % (user_id)
|
|
|
|
share_key = "S:%s" % (story_hash)
|
2013-09-06 13:20:08 -07:00
|
|
|
friends_with_shares = [int(f) for f in s.sinter(share_key, friend_key)]
|
2013-07-17 15:01:19 -07:00
|
|
|
friend_ids.update(friends_with_shares)
|
2013-08-14 14:32:50 -07:00
|
|
|
cls.mark_read(user_id, feed_id, story_hash, social_user_ids=friends_with_shares, r=p)
|
2013-07-17 14:49:54 -07:00
|
|
|
|
|
|
|
p.execute()
|
2013-08-14 14:32:50 -07:00
|
|
|
# p2.execute()
|
2013-07-17 15:01:19 -07:00
|
|
|
|
2013-07-17 15:32:08 -07:00
|
|
|
return list(feed_ids), list(friend_ids)
|
2014-03-24 14:44:21 -07:00
|
|
|
|
|
|
|
@classmethod
|
|
|
|
def mark_story_hash_unread(cls, user_id, story_hash, r=None, s=None):
|
|
|
|
if not r:
|
|
|
|
r = redis.Redis(connection_pool=settings.REDIS_STORY_HASH_POOL)
|
|
|
|
if not s:
|
|
|
|
s = redis.Redis(connection_pool=settings.REDIS_POOL)
|
|
|
|
# if not r2:
|
|
|
|
# r2 = redis.Redis(connection_pool=settings.REDIS_STORY_HASH_POOL2)
|
|
|
|
|
|
|
|
friend_ids = set()
|
|
|
|
feed_id, _ = MStory.split_story_hash(story_hash)
|
|
|
|
|
|
|
|
# Find other social feeds with this story to update their counts
|
|
|
|
friend_key = "F:%s:F" % (user_id)
|
|
|
|
share_key = "S:%s" % (story_hash)
|
|
|
|
friends_with_shares = [int(f) for f in s.sinter(share_key, friend_key)]
|
|
|
|
friend_ids.update(friends_with_shares)
|
|
|
|
cls.mark_unread(user_id, feed_id, story_hash, social_user_ids=friends_with_shares, r=r)
|
|
|
|
|
|
|
|
return feed_id, list(friend_ids)
|
2013-07-17 15:01:19 -07:00
|
|
|
|
2013-05-02 16:49:44 -07:00
|
|
|
@classmethod
|
2013-08-14 14:32:50 -07:00
|
|
|
def mark_read(cls, user_id, story_feed_id, story_hash, social_user_ids=None, r=None):
|
2013-05-02 16:49:44 -07:00
|
|
|
if not r:
|
|
|
|
r = redis.Redis(connection_pool=settings.REDIS_STORY_HASH_POOL)
|
2013-08-14 14:32:50 -07:00
|
|
|
# if not r2:
|
|
|
|
# r2 = redis.Redis(connection_pool=settings.REDIS_STORY_HASH_POOL2)
|
2013-05-02 16:49:44 -07:00
|
|
|
|
2013-06-04 15:54:36 -07:00
|
|
|
story_hash = MStory.ensure_story_hash(story_hash, story_feed_id=story_feed_id)
|
2013-05-03 16:48:17 -07:00
|
|
|
|
2013-05-02 16:49:44 -07:00
|
|
|
if not story_hash: return
|
|
|
|
|
2013-07-05 13:27:34 -07:00
|
|
|
def redis_commands(key):
|
|
|
|
r.sadd(key, story_hash)
|
2013-08-14 14:32:50 -07:00
|
|
|
# r2.sadd(key, story_hash)
|
2013-09-16 16:42:49 -07:00
|
|
|
r.expire(key, settings.DAYS_OF_STORY_HASHES*24*60*60)
|
|
|
|
# r2.expire(key, settings.DAYS_OF_STORY_HASHES*24*60*60)
|
2013-07-05 13:27:34 -07:00
|
|
|
|
2013-05-02 16:49:44 -07:00
|
|
|
all_read_stories_key = 'RS:%s' % (user_id)
|
2013-07-05 13:27:34 -07:00
|
|
|
redis_commands(all_read_stories_key)
|
2013-05-02 16:49:44 -07:00
|
|
|
|
|
|
|
read_story_key = 'RS:%s:%s' % (user_id, story_feed_id)
|
2013-07-05 13:27:34 -07:00
|
|
|
redis_commands(read_story_key)
|
|
|
|
|
|
|
|
if social_user_ids:
|
|
|
|
for social_user_id in social_user_ids:
|
|
|
|
social_read_story_key = 'RS:%s:B:%s' % (user_id, social_user_id)
|
|
|
|
redis_commands(social_read_story_key)
|
2013-05-02 16:49:44 -07:00
|
|
|
|
|
|
|
@staticmethod
|
2014-03-24 14:44:21 -07:00
|
|
|
def story_can_be_marked_read_by_user(story, user):
|
|
|
|
message = None
|
|
|
|
if story.story_date < user.profile.unread_cutoff:
|
|
|
|
if user.profile.is_premium:
|
|
|
|
message = "Story is more than %s days old, cannot mark as unread." % (
|
|
|
|
settings.DAYS_OF_UNREAD)
|
|
|
|
elif story.story_date > user.profile.unread_cutoff_premium:
|
|
|
|
message = "Story is more than %s days old. Premiums can mark unread up to 30 days." % (
|
|
|
|
settings.DAYS_OF_UNREAD_FREE)
|
|
|
|
else:
|
|
|
|
message = "Story is more than %s days old, cannot mark as unread." % (
|
|
|
|
settings.DAYS_OF_UNREAD_FREE)
|
|
|
|
return message
|
|
|
|
|
|
|
|
@staticmethod
|
|
|
|
def mark_unread(user_id, story_feed_id, story_hash, social_user_ids=None, r=None):
|
|
|
|
if not r:
|
|
|
|
r = redis.Redis(connection_pool=settings.REDIS_STORY_HASH_POOL)
|
|
|
|
# r2 = redis.Redis(connection_pool=settings.REDIS_STORY_HASH_POOL2)
|
2013-07-05 13:27:34 -07:00
|
|
|
|
2014-03-24 14:44:21 -07:00
|
|
|
story_hash = MStory.ensure_story_hash(story_hash, story_feed_id=story_feed_id)
|
|
|
|
|
|
|
|
if not story_hash: return
|
|
|
|
|
|
|
|
def redis_commands(key):
|
|
|
|
r.srem(key, story_hash)
|
|
|
|
# r2.srem(key, story_hash)
|
|
|
|
r.expire(key, settings.DAYS_OF_STORY_HASHES*24*60*60)
|
|
|
|
# r2.expire(key, settings.DAYS_OF_STORY_HASHES*24*60*60)
|
2013-07-05 13:27:34 -07:00
|
|
|
|
2014-03-24 14:44:21 -07:00
|
|
|
all_read_stories_key = 'RS:%s' % (user_id)
|
|
|
|
redis_commands(all_read_stories_key)
|
|
|
|
|
|
|
|
read_story_key = 'RS:%s:%s' % (user_id, story_feed_id)
|
|
|
|
redis_commands(read_story_key)
|
|
|
|
|
|
|
|
if social_user_ids:
|
|
|
|
for social_user_id in social_user_ids:
|
|
|
|
social_read_story_key = 'RS:%s:B:%s' % (user_id, social_user_id)
|
|
|
|
redis_commands(social_read_story_key)
|
2013-05-02 16:49:44 -07:00
|
|
|
|
2013-05-09 14:56:51 -07:00
|
|
|
@staticmethod
|
|
|
|
def get_stories(user_id, feed_id, r=None):
|
|
|
|
if not r:
|
|
|
|
r = redis.Redis(connection_pool=settings.REDIS_STORY_HASH_POOL)
|
|
|
|
story_hashes = r.smembers("RS:%s:%s" % (user_id, feed_id))
|
|
|
|
return story_hashes
|
|
|
|
|
|
|
|
@classmethod
|
|
|
|
def switch_feed(cls, user_id, old_feed_id, new_feed_id):
|
|
|
|
r = redis.Redis(connection_pool=settings.REDIS_STORY_HASH_POOL)
|
2013-08-14 14:32:50 -07:00
|
|
|
# r2 = redis.Redis(connection_pool=settings.REDIS_STORY_HASH_POOL2)
|
2013-05-09 14:56:51 -07:00
|
|
|
p = r.pipeline()
|
2013-08-14 14:32:50 -07:00
|
|
|
# p2 = r2.pipeline()
|
2013-05-09 14:56:51 -07:00
|
|
|
story_hashes = cls.get_stories(user_id, old_feed_id, r=r)
|
2013-07-01 21:59:09 -07:00
|
|
|
|
2013-05-09 14:56:51 -07:00
|
|
|
for story_hash in story_hashes:
|
2013-06-04 18:10:20 -07:00
|
|
|
_, hash_story = MStory.split_story_hash(story_hash)
|
2013-05-09 14:56:51 -07:00
|
|
|
new_story_hash = "%s:%s" % (new_feed_id, hash_story)
|
2013-07-01 22:39:16 -07:00
|
|
|
read_feed_key = "RS:%s:%s" % (user_id, new_feed_id)
|
|
|
|
p.sadd(read_feed_key, new_story_hash)
|
2013-08-14 14:32:50 -07:00
|
|
|
# p2.sadd(read_feed_key, new_story_hash)
|
2013-09-16 16:42:49 -07:00
|
|
|
p.expire(read_feed_key, settings.DAYS_OF_STORY_HASHES*24*60*60)
|
|
|
|
# p2.expire(read_feed_key, settings.DAYS_OF_STORY_HASHES*24*60*60)
|
2013-07-01 22:39:16 -07:00
|
|
|
|
|
|
|
read_user_key = "RS:%s" % (user_id)
|
|
|
|
p.sadd(read_user_key, new_story_hash)
|
2013-08-14 14:32:50 -07:00
|
|
|
# p2.sadd(read_user_key, new_story_hash)
|
2013-09-16 16:42:49 -07:00
|
|
|
p.expire(read_user_key, settings.DAYS_OF_STORY_HASHES*24*60*60)
|
|
|
|
# p2.expire(read_user_key, settings.DAYS_OF_STORY_HASHES*24*60*60)
|
2013-05-09 14:56:51 -07:00
|
|
|
|
|
|
|
p.execute()
|
2013-08-14 14:32:50 -07:00
|
|
|
# p2.execute()
|
2013-05-09 14:56:51 -07:00
|
|
|
|
|
|
|
if len(story_hashes) > 0:
|
|
|
|
logging.info(" ---> %s read stories" % len(story_hashes))
|
2011-01-07 16:26:17 -05:00
|
|
|
|
|
|
|
@classmethod
|
2013-05-10 16:11:30 -07:00
|
|
|
def switch_hash(cls, feed_id, old_hash, new_hash):
|
|
|
|
r = redis.Redis(connection_pool=settings.REDIS_STORY_HASH_POOL)
|
2013-08-14 14:32:50 -07:00
|
|
|
# r2 = redis.Redis(connection_pool=settings.REDIS_STORY_HASH_POOL2)
|
2013-05-10 16:11:30 -07:00
|
|
|
p = r.pipeline()
|
2013-08-14 14:32:50 -07:00
|
|
|
# p2 = r2.pipeline()
|
2013-09-16 16:42:49 -07:00
|
|
|
UNREAD_CUTOFF = datetime.datetime.now() - datetime.timedelta(days=settings.DAYS_OF_STORY_HASHES)
|
2013-07-01 21:59:09 -07:00
|
|
|
|
2013-05-28 10:23:36 -07:00
|
|
|
usersubs = UserSubscription.objects.filter(feed_id=feed_id, last_read_date__gte=UNREAD_CUTOFF)
|
|
|
|
logging.info(" ---> ~SB%s usersubs~SN to switch read story hashes..." % len(usersubs))
|
|
|
|
for sub in usersubs:
|
|
|
|
rs_key = "RS:%s:%s" % (sub.user.pk, feed_id)
|
2013-05-10 16:11:30 -07:00
|
|
|
read = r.sismember(rs_key, old_hash)
|
|
|
|
if read:
|
|
|
|
p.sadd(rs_key, new_hash)
|
2013-08-14 14:32:50 -07:00
|
|
|
# p2.sadd(rs_key, new_hash)
|
2013-09-16 16:42:49 -07:00
|
|
|
p.expire(rs_key, settings.DAYS_OF_STORY_HASHES*24*60*60)
|
|
|
|
# p2.expire(rs_key, settings.DAYS_OF_STORY_HASHES*24*60*60)
|
2013-07-01 22:39:16 -07:00
|
|
|
|
|
|
|
read_user_key = "RS:%s" % sub.user.pk
|
|
|
|
p.sadd(read_user_key, new_hash)
|
2013-08-14 14:32:50 -07:00
|
|
|
# p2.sadd(read_user_key, new_hash)
|
2013-09-16 16:42:49 -07:00
|
|
|
p.expire(read_user_key, settings.DAYS_OF_STORY_HASHES*24*60*60)
|
|
|
|
# p2.expire(read_user_key, settings.DAYS_OF_STORY_HASHES*24*60*60)
|
2013-04-08 17:02:03 -07:00
|
|
|
|
2013-05-10 16:11:30 -07:00
|
|
|
p.execute()
|
2013-08-14 14:32:50 -07:00
|
|
|
# p2.execute()
|
2013-07-01 19:17:10 -07:00
|
|
|
|
2014-04-03 12:44:04 -07:00
|
|
|
@classmethod
|
|
|
|
def read_story_count(cls, user_id):
|
|
|
|
r = redis.Redis(connection_pool=settings.REDIS_STORY_HASH_POOL)
|
|
|
|
key = "RS:%s" % user_id
|
|
|
|
count = r.scard(key)
|
|
|
|
return count
|
2012-12-17 17:14:47 -08:00
|
|
|
|
2009-06-16 03:08:55 +00:00
|
|
|
class UserSubscriptionFolders(models.Model):
|
2010-07-20 20:23:49 -04:00
|
|
|
"""
|
|
|
|
A JSON list of folders and feeds for while a user has subscribed. The list
|
|
|
|
is a recursive descent of feeds and folders in folders. Used to layout
|
|
|
|
the feeds and folders in the Reader's feed navigation pane.
|
|
|
|
"""
|
2010-11-04 22:47:18 -04:00
|
|
|
user = models.ForeignKey(User, unique=True)
|
2010-05-11 22:49:41 -04:00
|
|
|
folders = models.TextField(default="[]")
|
2009-06-16 03:08:55 +00:00
|
|
|
|
|
|
|
def __unicode__(self):
|
2010-02-11 01:28:47 -05:00
|
|
|
return "[%s]: %s" % (self.user, len(self.folders),)
|
2009-06-16 03:08:55 +00:00
|
|
|
|
|
|
|
class Meta:
|
|
|
|
verbose_name_plural = "folders"
|
2010-06-12 21:20:06 -04:00
|
|
|
verbose_name = "folder"
|
2011-03-21 10:15:18 -04:00
|
|
|
|
2012-07-20 20:37:35 -07:00
|
|
|
def compact(self):
|
|
|
|
folders = json.decode(self.folders)
|
|
|
|
|
|
|
|
def _compact(folder):
|
|
|
|
new_folder = []
|
|
|
|
for item in folder:
|
|
|
|
if isinstance(item, int) and item not in new_folder:
|
|
|
|
new_folder.append(item)
|
|
|
|
elif isinstance(item, dict):
|
|
|
|
for f_k, f_v in item.items():
|
|
|
|
new_folder.append({f_k: _compact(f_v)})
|
|
|
|
return new_folder
|
|
|
|
|
|
|
|
new_folders = _compact(folders)
|
|
|
|
logging.info(" ---> Compacting from %s to %s" % (folders, new_folders))
|
|
|
|
new_folders = json.encode(new_folders)
|
|
|
|
logging.info(" ---> Compacting from %s to %s" % (len(self.folders), len(new_folders)))
|
|
|
|
self.folders = new_folders
|
|
|
|
self.save()
|
|
|
|
|
2011-03-21 10:15:18 -04:00
|
|
|
def add_folder(self, parent_folder, folder):
|
|
|
|
if self.folders:
|
|
|
|
user_sub_folders = json.decode(self.folders)
|
|
|
|
else:
|
|
|
|
user_sub_folders = []
|
|
|
|
obj = {folder: []}
|
|
|
|
user_sub_folders = add_object_to_folder(obj, parent_folder, user_sub_folders)
|
|
|
|
self.folders = json.encode(user_sub_folders)
|
|
|
|
self.save()
|
2013-11-06 11:12:41 -08:00
|
|
|
|
|
|
|
def arranged_folders(self):
|
|
|
|
user_sub_folders = json.decode(self.folders)
|
|
|
|
def _arrange_folder(folder):
|
|
|
|
folder_feeds = []
|
|
|
|
folder_folders = []
|
|
|
|
for item in folder:
|
|
|
|
if isinstance(item, int):
|
|
|
|
folder_feeds.append(item)
|
|
|
|
elif isinstance(item, dict):
|
|
|
|
for f_k, f_v in item.items():
|
|
|
|
arranged_folder = _arrange_folder(f_v)
|
|
|
|
folder_folders.append({f_k: arranged_folder})
|
|
|
|
|
|
|
|
arranged_folder = folder_feeds + folder_folders
|
|
|
|
return arranged_folder
|
2010-09-16 10:35:36 -04:00
|
|
|
|
2013-11-06 11:12:41 -08:00
|
|
|
return _arrange_folder(user_sub_folders)
|
|
|
|
|
2014-01-27 15:29:55 -08:00
|
|
|
def flatten_folders(self, feeds=None):
|
|
|
|
folders = json.decode(self.folders)
|
|
|
|
flat_folders = {" ": []}
|
|
|
|
|
|
|
|
def _flatten_folders(items, parent_folder="", depth=0):
|
|
|
|
for item in items:
|
|
|
|
if isinstance(item, int) and ((not feeds) or (feeds and item in feeds)):
|
|
|
|
if not parent_folder:
|
|
|
|
parent_folder = ' '
|
|
|
|
if parent_folder in flat_folders:
|
|
|
|
flat_folders[parent_folder].append(item)
|
|
|
|
else:
|
|
|
|
flat_folders[parent_folder] = [item]
|
|
|
|
elif isinstance(item, dict):
|
|
|
|
for folder_name in item:
|
|
|
|
folder = item[folder_name]
|
|
|
|
flat_folder_name = "%s%s%s" % (
|
|
|
|
parent_folder if parent_folder and parent_folder != ' ' else "",
|
|
|
|
" - " if parent_folder and parent_folder != ' ' else "",
|
|
|
|
folder_name
|
|
|
|
)
|
|
|
|
flat_folders[flat_folder_name] = []
|
|
|
|
_flatten_folders(folder, flat_folder_name, depth+1)
|
|
|
|
|
|
|
|
_flatten_folders(folders)
|
|
|
|
|
|
|
|
return flat_folders
|
|
|
|
|
2011-11-07 20:50:46 -08:00
|
|
|
def delete_feed(self, feed_id, in_folder, commit_delete=True):
|
2010-09-16 10:35:36 -04:00
|
|
|
def _find_feed_in_folders(old_folders, folder_name='', multiples_found=False, deleted=False):
|
|
|
|
new_folders = []
|
|
|
|
for k, folder in enumerate(old_folders):
|
|
|
|
if isinstance(folder, int):
|
2013-08-05 15:30:56 -07:00
|
|
|
if (folder == feed_id and in_folder is not None and (
|
2010-09-16 10:35:36 -04:00
|
|
|
(folder_name != in_folder) or
|
|
|
|
(folder_name == in_folder and deleted))):
|
|
|
|
multiples_found = True
|
2011-02-23 13:46:47 -05:00
|
|
|
logging.user(self.user, "~FB~SBDeleting feed, and a multiple has been found in '%s'" % (folder_name))
|
2013-08-05 15:30:56 -07:00
|
|
|
if (folder == feed_id and
|
|
|
|
(folder_name == in_folder or in_folder is None) and
|
|
|
|
not deleted):
|
2011-02-23 13:46:47 -05:00
|
|
|
logging.user(self.user, "~FBDelete feed: %s'th item: %s folders/feeds" % (
|
|
|
|
k, len(old_folders)
|
2010-09-16 10:35:36 -04:00
|
|
|
))
|
|
|
|
deleted = True
|
|
|
|
else:
|
|
|
|
new_folders.append(folder)
|
|
|
|
elif isinstance(folder, dict):
|
|
|
|
for f_k, f_v in folder.items():
|
|
|
|
nf, multiples_found, deleted = _find_feed_in_folders(f_v, f_k, multiples_found, deleted)
|
|
|
|
new_folders.append({f_k: nf})
|
|
|
|
|
|
|
|
return new_folders, multiples_found, deleted
|
2011-12-05 09:26:02 -08:00
|
|
|
|
2013-11-06 11:12:41 -08:00
|
|
|
user_sub_folders = self.arranged_folders()
|
2010-09-16 10:35:36 -04:00
|
|
|
user_sub_folders, multiples_found, deleted = _find_feed_in_folders(user_sub_folders)
|
|
|
|
self.folders = json.encode(user_sub_folders)
|
|
|
|
self.save()
|
2010-07-20 20:23:49 -04:00
|
|
|
|
2011-11-07 20:50:46 -08:00
|
|
|
if not multiples_found and deleted and commit_delete:
|
2010-11-09 09:55:44 -05:00
|
|
|
try:
|
|
|
|
user_sub = UserSubscription.objects.get(user=self.user, feed=feed_id)
|
|
|
|
except Feed.DoesNotExist:
|
|
|
|
duplicate_feed = DuplicateFeed.objects.filter(duplicate_feed_id=feed_id)
|
|
|
|
if duplicate_feed:
|
|
|
|
try:
|
|
|
|
user_sub = UserSubscription.objects.get(user=self.user,
|
|
|
|
feed=duplicate_feed[0].feed)
|
|
|
|
except Feed.DoesNotExist:
|
|
|
|
return
|
2011-11-02 09:08:58 -07:00
|
|
|
if user_sub:
|
|
|
|
user_sub.delete()
|
2010-09-16 10:35:36 -04:00
|
|
|
|
2011-11-08 09:20:10 -08:00
|
|
|
def delete_folder(self, folder_to_delete, in_folder, feed_ids_in_folder, commit_delete=True):
|
|
|
|
def _find_folder_in_folders(old_folders, folder_name, feeds_to_delete, deleted_folder=None):
|
2010-09-16 10:35:36 -04:00
|
|
|
new_folders = []
|
|
|
|
for k, folder in enumerate(old_folders):
|
|
|
|
if isinstance(folder, int):
|
2010-09-22 10:12:38 -04:00
|
|
|
new_folders.append(folder)
|
|
|
|
if folder in feeds_to_delete:
|
|
|
|
feeds_to_delete.remove(folder)
|
2010-09-16 10:35:36 -04:00
|
|
|
elif isinstance(folder, dict):
|
|
|
|
for f_k, f_v in folder.items():
|
2013-08-05 15:30:56 -07:00
|
|
|
if f_k == folder_to_delete and (folder_name == in_folder or in_folder is None):
|
2011-02-23 13:46:47 -05:00
|
|
|
logging.user(self.user, "~FBDeleting folder '~SB%s~SN' in '%s': %s" % (f_k, folder_name, folder))
|
2011-11-08 09:20:10 -08:00
|
|
|
deleted_folder = folder
|
2010-09-22 10:12:38 -04:00
|
|
|
else:
|
2011-11-08 09:20:10 -08:00
|
|
|
nf, feeds_to_delete, deleted_folder = _find_folder_in_folders(f_v, f_k, feeds_to_delete, deleted_folder)
|
2010-09-22 10:12:38 -04:00
|
|
|
new_folders.append({f_k: nf})
|
2010-09-16 10:35:36 -04:00
|
|
|
|
2011-11-08 09:20:10 -08:00
|
|
|
return new_folders, feeds_to_delete, deleted_folder
|
2010-09-16 10:35:36 -04:00
|
|
|
|
|
|
|
user_sub_folders = json.decode(self.folders)
|
2011-11-08 09:20:10 -08:00
|
|
|
user_sub_folders, feeds_to_delete, deleted_folder = _find_folder_in_folders(user_sub_folders, '', feed_ids_in_folder)
|
2010-09-16 10:35:36 -04:00
|
|
|
self.folders = json.encode(user_sub_folders)
|
|
|
|
self.save()
|
2011-12-05 09:26:02 -08:00
|
|
|
|
2011-11-08 09:20:10 -08:00
|
|
|
if commit_delete:
|
2012-05-23 21:07:01 -07:00
|
|
|
UserSubscription.objects.filter(user=self.user, feed__in=feeds_to_delete).delete()
|
2011-11-08 09:20:10 -08:00
|
|
|
|
|
|
|
return deleted_folder
|
2010-12-11 17:16:12 -05:00
|
|
|
|
|
|
|
def rename_folder(self, folder_to_rename, new_folder_name, in_folder):
|
|
|
|
def _find_folder_in_folders(old_folders, folder_name):
|
|
|
|
new_folders = []
|
|
|
|
for k, folder in enumerate(old_folders):
|
|
|
|
if isinstance(folder, int):
|
|
|
|
new_folders.append(folder)
|
|
|
|
elif isinstance(folder, dict):
|
|
|
|
for f_k, f_v in folder.items():
|
|
|
|
nf = _find_folder_in_folders(f_v, f_k)
|
|
|
|
if f_k == folder_to_rename and folder_name == in_folder:
|
2011-02-23 13:46:47 -05:00
|
|
|
logging.user(self.user, "~FBRenaming folder '~SB%s~SN' in '%s' to: ~SB%s" % (
|
|
|
|
f_k, folder_name, new_folder_name))
|
2010-12-11 17:16:12 -05:00
|
|
|
f_k = new_folder_name
|
|
|
|
new_folders.append({f_k: nf})
|
|
|
|
|
|
|
|
return new_folders
|
|
|
|
|
|
|
|
user_sub_folders = json.decode(self.folders)
|
|
|
|
user_sub_folders = _find_folder_in_folders(user_sub_folders, '')
|
|
|
|
self.folders = json.encode(user_sub_folders)
|
|
|
|
self.save()
|
2011-11-07 20:50:46 -08:00
|
|
|
|
|
|
|
def move_feed_to_folder(self, feed_id, in_folder=None, to_folder=None):
|
2011-12-04 13:55:57 -08:00
|
|
|
logging.user(self.user, "~FBMoving feed '~SB%s~SN' in '%s' to: ~SB%s" % (
|
|
|
|
feed_id, in_folder, to_folder))
|
2011-11-07 20:50:46 -08:00
|
|
|
user_sub_folders = json.decode(self.folders)
|
|
|
|
self.delete_feed(feed_id, in_folder, commit_delete=False)
|
|
|
|
user_sub_folders = json.decode(self.folders)
|
|
|
|
user_sub_folders = add_object_to_folder(int(feed_id), to_folder, user_sub_folders)
|
|
|
|
self.folders = json.encode(user_sub_folders)
|
|
|
|
self.save()
|
|
|
|
|
|
|
|
return self
|
2011-11-08 09:20:10 -08:00
|
|
|
|
|
|
|
def move_folder_to_folder(self, folder_name, in_folder=None, to_folder=None):
|
2011-12-04 13:55:57 -08:00
|
|
|
logging.user(self.user, "~FBMoving folder '~SB%s~SN' in '%s' to: ~SB%s" % (
|
|
|
|
folder_name, in_folder, to_folder))
|
2011-11-08 09:20:10 -08:00
|
|
|
user_sub_folders = json.decode(self.folders)
|
|
|
|
deleted_folder = self.delete_folder(folder_name, in_folder, [], commit_delete=False)
|
|
|
|
user_sub_folders = json.decode(self.folders)
|
|
|
|
user_sub_folders = add_object_to_folder(deleted_folder, to_folder, user_sub_folders)
|
|
|
|
self.folders = json.encode(user_sub_folders)
|
|
|
|
self.save()
|
|
|
|
|
|
|
|
return self
|
2011-11-16 10:00:03 -08:00
|
|
|
|
|
|
|
def rewrite_feed(self, original_feed, duplicate_feed):
|
|
|
|
def rewrite_folders(folders, original_feed, duplicate_feed):
|
|
|
|
new_folders = []
|
|
|
|
|
|
|
|
for k, folder in enumerate(folders):
|
|
|
|
if isinstance(folder, int):
|
|
|
|
if folder == duplicate_feed.pk:
|
|
|
|
# logging.info(" ===> Rewrote %s'th item: %s" % (k+1, folders))
|
|
|
|
new_folders.append(original_feed.pk)
|
|
|
|
else:
|
|
|
|
new_folders.append(folder)
|
|
|
|
elif isinstance(folder, dict):
|
|
|
|
for f_k, f_v in folder.items():
|
|
|
|
new_folders.append({f_k: rewrite_folders(f_v, original_feed, duplicate_feed)})
|
|
|
|
|
|
|
|
return new_folders
|
|
|
|
|
|
|
|
folders = json.decode(self.folders)
|
|
|
|
folders = rewrite_folders(folders, original_feed, duplicate_feed)
|
|
|
|
self.folders = json.encode(folders)
|
|
|
|
self.save()
|
2012-12-12 16:05:28 -08:00
|
|
|
|
|
|
|
def flat(self):
|
|
|
|
folders = json.decode(self.folders)
|
|
|
|
|
|
|
|
def _flat(folder, feeds=None):
|
|
|
|
if not feeds:
|
|
|
|
feeds = []
|
|
|
|
for item in folder:
|
|
|
|
if isinstance(item, int) and item not in feeds:
|
|
|
|
feeds.append(item)
|
|
|
|
elif isinstance(item, dict):
|
|
|
|
for f_k, f_v in item.items():
|
|
|
|
feeds.extend(_flat(f_v))
|
|
|
|
return feeds
|
|
|
|
|
|
|
|
return _flat(folders)
|
2012-12-12 16:09:48 -08:00
|
|
|
|
|
|
|
@classmethod
|
|
|
|
def add_all_missing_feeds(cls):
|
2012-12-12 16:36:36 -08:00
|
|
|
usf = cls.objects.all().order_by('pk')
|
2012-12-12 16:09:48 -08:00
|
|
|
total = usf.count()
|
|
|
|
|
|
|
|
for i, f in enumerate(usf):
|
|
|
|
print "%s/%s: %s" % (i, total, f)
|
|
|
|
f.add_missing_feeds()
|
|
|
|
|
2012-12-12 16:05:28 -08:00
|
|
|
def add_missing_feeds(self):
|
|
|
|
all_feeds = self.flat()
|
|
|
|
subs = [us.feed_id for us in
|
|
|
|
UserSubscription.objects.filter(user=self.user).only('feed')]
|
|
|
|
|
2013-07-15 12:10:51 -07:00
|
|
|
missing_subs = set(all_feeds) - set(subs)
|
|
|
|
if missing_subs:
|
|
|
|
logging.debug(" ---> %s is missing %s subs. Adding %s..." % (
|
|
|
|
self.user, len(missing_subs), missing_subs))
|
|
|
|
for feed_id in missing_subs:
|
2012-12-12 16:09:48 -08:00
|
|
|
feed = Feed.get_by_id(feed_id)
|
|
|
|
if feed:
|
2013-03-25 13:15:14 -07:00
|
|
|
us, _ = UserSubscription.objects.get_or_create(user=self.user, feed=feed, defaults={
|
|
|
|
'needs_unread_recalc': True
|
|
|
|
})
|
|
|
|
if not us.needs_unread_recalc:
|
|
|
|
us.needs_unread_recalc = True
|
|
|
|
us.save()
|
2011-11-16 10:00:03 -08:00
|
|
|
|
2013-07-15 12:10:51 -07:00
|
|
|
missing_folder_feeds = set(subs) - set(all_feeds)
|
|
|
|
if missing_folder_feeds:
|
|
|
|
user_sub_folders = json.decode(self.folders)
|
|
|
|
logging.debug(" ---> %s is missing %s folder feeds. Adding %s..." % (
|
|
|
|
self.user, len(missing_folder_feeds), missing_folder_feeds))
|
|
|
|
for feed_id in missing_folder_feeds:
|
|
|
|
feed = Feed.get_by_id(feed_id)
|
|
|
|
if feed and feed.pk == feed_id:
|
2013-07-15 12:11:39 -07:00
|
|
|
user_sub_folders = add_object_to_folder(feed_id, "", user_sub_folders)
|
2013-07-15 12:10:51 -07:00
|
|
|
self.folders = json.encode(user_sub_folders)
|
|
|
|
self.save()
|
|
|
|
|
2010-07-20 20:23:49 -04:00
|
|
|
|
2010-06-12 21:20:06 -04:00
|
|
|
class Feature(models.Model):
|
2010-07-20 20:23:49 -04:00
|
|
|
"""
|
|
|
|
Simple blog-like feature board shown to all users on the home page.
|
|
|
|
"""
|
2010-06-12 21:20:06 -04:00
|
|
|
description = models.TextField(default="")
|
|
|
|
date = models.DateTimeField(default=datetime.datetime.now)
|
|
|
|
|
|
|
|
def __unicode__(self):
|
|
|
|
return "[%s] %s" % (self.date, self.description[:50])
|
|
|
|
|
|
|
|
class Meta:
|
2010-07-20 20:23:49 -04:00
|
|
|
ordering = ["-date"]
|