Clean up reblog tracking keys, related improvements (#5428)
* Clean up reblog-tracking sets from FeedManager Builds on #5419, with a few minor optimizations and cleanup of sets after they are no longer needed. * Update tests, fix multiply-reblogged case Previously, we would have lost the fact that a given status was reblogged if the displayed reblog of it was removed, now we don't. Also added tests to make sure FeedManager#trim cleans up our reblog tracking keys, fixed up FeedCleanupScheduler to use the right loop, and fixed the test for it.
This commit is contained in:
parent
a2b600428c
commit
554c2fd8af
|
@ -56,7 +56,17 @@ class FeedManager
|
||||||
falloff_rank = FeedManager::REBLOG_FALLOFF - 1
|
falloff_rank = FeedManager::REBLOG_FALLOFF - 1
|
||||||
falloff_range = redis.zrevrange(timeline_key, falloff_rank, falloff_rank, with_scores: true)
|
falloff_range = redis.zrevrange(timeline_key, falloff_rank, falloff_rank, with_scores: true)
|
||||||
falloff_score = falloff_range&.first&.last&.to_i || 0
|
falloff_score = falloff_range&.first&.last&.to_i || 0
|
||||||
redis.zremrangebyscore(reblog_key, 0, falloff_score)
|
|
||||||
|
# Get any reblogs we might have to clean up after.
|
||||||
|
redis.zrangebyscore(reblog_key, 0, falloff_score).each do |reblogged_id|
|
||||||
|
# Remove it from the set of reblogs we're tracking *first* to avoid races.
|
||||||
|
redis.zrem(reblog_key, reblogged_id)
|
||||||
|
# Just drop any set we might have created to track additional reblogs.
|
||||||
|
# This means that if this reblog is deleted, we won't automatically insert
|
||||||
|
# another reblog, but also that any new reblog can be inserted into the
|
||||||
|
# feed.
|
||||||
|
redis.del(key(type, account_id, "reblogs:#{reblogged_id}"))
|
||||||
|
end
|
||||||
end
|
end
|
||||||
|
|
||||||
def push_update_required?(timeline_type, account_id)
|
def push_update_required?(timeline_type, account_id)
|
||||||
|
@ -177,23 +187,28 @@ class FeedManager
|
||||||
reblog_key = key(timeline_type, account.id, 'reblogs')
|
reblog_key = key(timeline_type, account.id, 'reblogs')
|
||||||
|
|
||||||
if status.reblog?
|
if status.reblog?
|
||||||
reblog_set_key = key(timeline_type, account.id, "reblogs:#{status.reblog_of_id}")
|
|
||||||
|
|
||||||
# If the original status or a reblog of it is within
|
# If the original status or a reblog of it is within
|
||||||
# REBLOG_FALLOFF statuses from the top, do not re-insert it into
|
# REBLOG_FALLOFF statuses from the top, do not re-insert it into
|
||||||
# the feed
|
# the feed
|
||||||
rank = redis.zrevrank(timeline_key, status.reblog_of_id)
|
rank = redis.zrevrank(timeline_key, status.reblog_of_id)
|
||||||
|
|
||||||
redis.sadd(reblog_set_key, status.reblog_of_id) unless rank.nil?
|
|
||||||
redis.sadd(reblog_set_key, status.id)
|
|
||||||
|
|
||||||
return false if !rank.nil? && rank < FeedManager::REBLOG_FALLOFF
|
return false if !rank.nil? && rank < FeedManager::REBLOG_FALLOFF
|
||||||
|
|
||||||
reblog_rank = redis.zrevrank(reblog_key, status.reblog_of_id)
|
reblog_rank = redis.zrevrank(reblog_key, status.reblog_of_id)
|
||||||
return false unless reblog_rank.nil?
|
if reblog_rank.nil?
|
||||||
|
# This is not something we've already seen reblogged, so we
|
||||||
redis.zadd(timeline_key, status.id, status.id)
|
# can just add it to the feed (and note that we're
|
||||||
redis.zadd(reblog_key, status.id, status.reblog_of_id)
|
# reblogging it).
|
||||||
|
redis.zadd(timeline_key, status.id, status.id)
|
||||||
|
redis.zadd(reblog_key, status.id, status.reblog_of_id)
|
||||||
|
else
|
||||||
|
# Another reblog of the same status was already in the
|
||||||
|
# REBLOG_FALLOFF most recent statuses, so we note that this
|
||||||
|
# is an "extra" reblog, by storing it in reblog_set_key.
|
||||||
|
reblog_set_key = key(timeline_type, account.id, "reblogs:#{status.reblog_of_id}")
|
||||||
|
redis.sadd(reblog_set_key, status.id)
|
||||||
|
return false
|
||||||
|
end
|
||||||
else
|
else
|
||||||
redis.zadd(timeline_key, status.id, status.id)
|
redis.zadd(timeline_key, status.id, status.id)
|
||||||
end
|
end
|
||||||
|
@ -207,23 +222,21 @@ class FeedManager
|
||||||
# do so if appropriate.
|
# do so if appropriate.
|
||||||
def remove_from_feed(timeline_type, account, status)
|
def remove_from_feed(timeline_type, account, status)
|
||||||
timeline_key = key(timeline_type, account.id)
|
timeline_key = key(timeline_type, account.id)
|
||||||
reblog_key = key(timeline_type, account.id, 'reblogs')
|
|
||||||
|
|
||||||
if status.reblog?
|
if status.reblog?
|
||||||
# 1. If the reblogging status is not in the feed, stop.
|
# 1. If the reblogging status is not in the feed, stop.
|
||||||
status_rank = redis.zrevrank(timeline_key, status.id)
|
status_rank = redis.zrevrank(timeline_key, status.id)
|
||||||
return false if status_rank.nil?
|
return false if status_rank.nil?
|
||||||
|
|
||||||
# 2. Remove the reblogged status from the `:reblogs` zset.
|
# 2. Remove reblog from set of this status's reblogs.
|
||||||
redis.zrem(reblog_key, status.reblog_of_id)
|
|
||||||
|
|
||||||
# 3. Remove reblog from set of this status's reblogs, and
|
|
||||||
# re-insert another reblog or original into the feed if
|
|
||||||
# one remains in the set
|
|
||||||
reblog_set_key = key(timeline_type, account.id, "reblogs:#{status.reblog_of_id}")
|
reblog_set_key = key(timeline_type, account.id, "reblogs:#{status.reblog_of_id}")
|
||||||
|
|
||||||
redis.srem(reblog_set_key, status.id)
|
redis.srem(reblog_set_key, status.id)
|
||||||
other_reblog = redis.srandmember(reblog_set_key)
|
# 3. Re-insert another reblog or original into the feed if one
|
||||||
|
# remains in the set. We could pick a random element, but this
|
||||||
|
# set should generally be small, and it seems ideal to show the
|
||||||
|
# oldest potential such reblog.
|
||||||
|
other_reblog = redis.smembers(reblog_set_key).map(&:to_i).sort.first
|
||||||
|
|
||||||
redis.zadd(timeline_key, other_reblog, other_reblog) if other_reblog
|
redis.zadd(timeline_key, other_reblog, other_reblog) if other_reblog
|
||||||
|
|
||||||
|
|
|
@ -5,18 +5,36 @@ class Scheduler::FeedCleanupScheduler
|
||||||
include Sidekiq::Worker
|
include Sidekiq::Worker
|
||||||
|
|
||||||
def perform
|
def perform
|
||||||
|
reblogged_id_sets = {}
|
||||||
|
feedmanager = FeedManager.instance
|
||||||
|
|
||||||
redis.pipelined do
|
redis.pipelined do
|
||||||
inactive_users.each do |account_id|
|
inactive_user_ids.each do |account_id|
|
||||||
redis.del(FeedManager.instance.key(:home, account_id))
|
redis.del(feedmanager.key(:home, account_id))
|
||||||
redis.del(FeedManager.instance.key(:home, account_id, 'reblogs'))
|
reblog_key = feedmanager.key(:home, account_id, 'reblogs')
|
||||||
|
# We collect a future for this: we don't block while getting
|
||||||
|
# it, but we can iterate over it later.
|
||||||
|
reblogged_id_sets[account_id] = redis.zrange(reblog_key, 0, -1)
|
||||||
|
redis.del(reblog_key)
|
||||||
|
end
|
||||||
|
end
|
||||||
|
|
||||||
|
# Remove all of the reblog tracking keys we just removed the
|
||||||
|
# references to.
|
||||||
|
redis.pipelined do
|
||||||
|
reblogged_id_sets.each do |account_id, future|
|
||||||
|
future.value.each do |reblogged_id|
|
||||||
|
reblog_set_key = feedmanager.key(:home, account_id, "reblogs:#{reblogged_id}")
|
||||||
|
redis.del(reblog_set_key)
|
||||||
|
end
|
||||||
end
|
end
|
||||||
end
|
end
|
||||||
end
|
end
|
||||||
|
|
||||||
private
|
private
|
||||||
|
|
||||||
def inactive_users
|
def inactive_user_ids
|
||||||
@inactive_users ||= User.confirmed.inactive.pluck(:account_id)
|
@inactive_user_ids ||= User.confirmed.inactive.pluck(:account_id)
|
||||||
end
|
end
|
||||||
|
|
||||||
def redis
|
def redis
|
||||||
|
|
|
@ -211,6 +211,22 @@ RSpec.describe FeedManager do
|
||||||
expect(FeedManager.instance.push('type', account, reblogs.last)).to be false
|
expect(FeedManager.instance.push('type', account, reblogs.last)).to be false
|
||||||
end
|
end
|
||||||
|
|
||||||
|
it 'does not save a new reblog of a multiply-reblogged-then-unreblogged status' do
|
||||||
|
account = Fabricate(:account)
|
||||||
|
reblogged = Fabricate(:status)
|
||||||
|
reblogs = 3.times.map { Fabricate(:status, reblog: reblogged) }
|
||||||
|
|
||||||
|
# Accept the reblogs
|
||||||
|
FeedManager.instance.push('type', account, reblogs[0])
|
||||||
|
FeedManager.instance.push('type', account, reblogs[1])
|
||||||
|
|
||||||
|
# Unreblog the first one
|
||||||
|
FeedManager.instance.unpush('type', account, reblogs[0])
|
||||||
|
|
||||||
|
# The last reblog should still be ignored
|
||||||
|
expect(FeedManager.instance.push('type', account, reblogs.last)).to be false
|
||||||
|
end
|
||||||
|
|
||||||
it 'saves a new reblog of a long-ago-reblogged status' do
|
it 'saves a new reblog of a long-ago-reblogged status' do
|
||||||
account = Fabricate(:account)
|
account = Fabricate(:account)
|
||||||
reblogged = Fabricate(:status)
|
reblogged = Fabricate(:status)
|
||||||
|
@ -230,6 +246,38 @@ RSpec.describe FeedManager do
|
||||||
end
|
end
|
||||||
end
|
end
|
||||||
|
|
||||||
|
describe '#trim' do
|
||||||
|
let(:receiver) { Fabricate(:account) }
|
||||||
|
|
||||||
|
it 'cleans up reblog tracking keys' do
|
||||||
|
reblogged = Fabricate(:status)
|
||||||
|
status = Fabricate(:status, reblog: reblogged)
|
||||||
|
another_status = Fabricate(:status, reblog: reblogged)
|
||||||
|
reblogs_key = FeedManager.instance.key('type', receiver.id, 'reblogs')
|
||||||
|
reblog_set_key = FeedManager.instance.key('type', receiver.id, "reblogs:#{reblogged.id}")
|
||||||
|
|
||||||
|
FeedManager.instance.push('type', receiver, status)
|
||||||
|
FeedManager.instance.push('type', receiver, another_status)
|
||||||
|
|
||||||
|
# We should have a tracking set and an entry in reblogs.
|
||||||
|
expect(Redis.current.exists(reblog_set_key)).to be true
|
||||||
|
expect(Redis.current.zrange(reblogs_key, 0, -1)).to eq [reblogged.id.to_s]
|
||||||
|
|
||||||
|
# Push everything off the end of the feed.
|
||||||
|
FeedManager::MAX_ITEMS.times do
|
||||||
|
FeedManager.instance.push('type', receiver, Fabricate(:status))
|
||||||
|
end
|
||||||
|
|
||||||
|
# `trim` should be called automatically, but do it anyway, as
|
||||||
|
# we're testing `trim`, not side effects of `push`.
|
||||||
|
FeedManager.instance.trim('type', receiver.id)
|
||||||
|
|
||||||
|
# We should not have any reblog tracking data.
|
||||||
|
expect(Redis.current.exists(reblog_set_key)).to be false
|
||||||
|
expect(Redis.current.zrange(reblogs_key, 0, -1)).to be_empty
|
||||||
|
end
|
||||||
|
end
|
||||||
|
|
||||||
describe '#unpush' do
|
describe '#unpush' do
|
||||||
let(:receiver) { Fabricate(:account) }
|
let(:receiver) { Fabricate(:account) }
|
||||||
|
|
||||||
|
@ -265,20 +313,22 @@ RSpec.describe FeedManager do
|
||||||
expect(Redis.current.zrange("feed:type:#{receiver.id}", 0, -1)).to be_empty
|
expect(Redis.current.zrange("feed:type:#{receiver.id}", 0, -1)).to be_empty
|
||||||
end
|
end
|
||||||
|
|
||||||
it 'leaves a reblogged status if another reblog was in feed' do
|
it 'leaves a multiply-reblogged status if another reblog was in feed' do
|
||||||
reblogged = Fabricate(:status)
|
reblogged = Fabricate(:status)
|
||||||
status = Fabricate(:status, reblog: reblogged)
|
reblogs = 3.times.map { Fabricate(:status, reblog: reblogged) }
|
||||||
another_status = Fabricate(:status, reblog: reblogged)
|
|
||||||
|
|
||||||
FeedManager.instance.push('type', receiver, status)
|
reblogs.each do |reblog|
|
||||||
FeedManager.instance.push('type', receiver, another_status)
|
FeedManager.instance.push('type', receiver, reblog)
|
||||||
|
end
|
||||||
|
|
||||||
# The reblogging status should show up under normal conditions.
|
# The reblogging status should show up under normal conditions.
|
||||||
expect(Redis.current.zrange("feed:type:#{receiver.id}", 0, -1)).to eq [status.id.to_s]
|
expect(Redis.current.zrange("feed:type:#{receiver.id}", 0, -1)).to eq [reblogs.first.id.to_s]
|
||||||
|
|
||||||
FeedManager.instance.unpush('type', receiver, status)
|
reblogs[0...-1].each do |reblog|
|
||||||
|
FeedManager.instance.unpush('type', receiver, reblog)
|
||||||
|
end
|
||||||
|
|
||||||
expect(Redis.current.zrange("feed:type:#{receiver.id}", 0, -1)).to eq [another_status.id.to_s]
|
expect(Redis.current.zrange("feed:type:#{receiver.id}", 0, -1)).to eq [reblogs.last.id.to_s]
|
||||||
end
|
end
|
||||||
|
|
||||||
it 'sends push updates' do
|
it 'sends push updates' do
|
||||||
|
|
|
@ -9,14 +9,18 @@ describe Scheduler::FeedCleanupScheduler do
|
||||||
it 'clears feeds of inactives' do
|
it 'clears feeds of inactives' do
|
||||||
Redis.current.zadd(feed_key_for(inactive_user), 1, 1)
|
Redis.current.zadd(feed_key_for(inactive_user), 1, 1)
|
||||||
Redis.current.zadd(feed_key_for(active_user), 1, 1)
|
Redis.current.zadd(feed_key_for(active_user), 1, 1)
|
||||||
|
Redis.current.zadd(feed_key_for(inactive_user, 'reblogs'), 2, 2)
|
||||||
|
Redis.current.sadd(feed_key_for(inactive_user, 'reblogs:2'), 3)
|
||||||
|
|
||||||
subject.perform
|
subject.perform
|
||||||
|
|
||||||
expect(Redis.current.zcard(feed_key_for(inactive_user))).to eq 0
|
expect(Redis.current.zcard(feed_key_for(inactive_user))).to eq 0
|
||||||
expect(Redis.current.zcard(feed_key_for(active_user))).to eq 1
|
expect(Redis.current.zcard(feed_key_for(active_user))).to eq 1
|
||||||
|
expect(Redis.current.exists(feed_key_for(inactive_user, 'reblogs'))).to be false
|
||||||
|
expect(Redis.current.exists(feed_key_for(inactive_user, 'reblogs:2'))).to be false
|
||||||
end
|
end
|
||||||
|
|
||||||
def feed_key_for(user)
|
def feed_key_for(user, subtype = nil)
|
||||||
FeedManager.instance.key(:home, user.account_id)
|
FeedManager.instance.key(:home, user.account_id, subtype)
|
||||||
end
|
end
|
||||||
end
|
end
|
||||||
|
|
Reference in a new issue