##// END OF EJS Templates
Limit only first posting by this session. Assume messages created by spam bots...
Limit only first posting by this session. Assume messages created by spam bots can be removed afterwards

File last commit:

r1489:9709b2fb default
r1494:89a50a1d default
Show More
thread.py
322 lines | 9.9 KiB | text/x-python | PythonLexer
import logging
from adjacent import Client
from django.db.models import Count, Sum, QuerySet, Q
from django.utils import timezone
from django.db import models, transaction
from boards.models import STATUS_BUMPLIMIT, STATUS_ACTIVE, STATUS_ARCHIVE
from boards import settings
import boards
from boards.utils import cached_result, datetime_to_epoch
from boards.models.post import Post
from boards.models.tag import Tag
FAV_THREAD_NO_UPDATES = -1
__author__ = 'neko259'
logger = logging.getLogger(__name__)
WS_NOTIFICATION_TYPE_NEW_POST = 'new_post'
WS_NOTIFICATION_TYPE = 'notification_type'
WS_CHANNEL_THREAD = "thread:"
STATUS_CHOICES = (
(STATUS_ACTIVE, STATUS_ACTIVE),
(STATUS_BUMPLIMIT, STATUS_BUMPLIMIT),
(STATUS_ARCHIVE, STATUS_ARCHIVE),
)
class ThreadManager(models.Manager):
def process_oldest_threads(self):
"""
Preserves maximum thread count. If there are too many threads,
archive or delete the old ones.
"""
threads = Thread.objects.exclude(status=STATUS_ARCHIVE).order_by('-bump_time')
thread_count = threads.count()
max_thread_count = settings.get_int('Messages', 'MaxThreadCount')
if thread_count > max_thread_count:
num_threads_to_delete = thread_count - max_thread_count
old_threads = threads[thread_count - num_threads_to_delete:]
for thread in old_threads:
if settings.get_bool('Storage', 'ArchiveThreads'):
self._archive_thread(thread)
else:
thread.delete()
logger.info('Processed %d old threads' % num_threads_to_delete)
def _archive_thread(self, thread):
thread.status = STATUS_ARCHIVE
thread.last_edit_time = timezone.now()
thread.update_posts_time()
thread.save(update_fields=['last_edit_time', 'status'])
def get_new_posts(self, datas):
query = None
# TODO Use classes instead of dicts
for data in datas:
if data['last_id'] != FAV_THREAD_NO_UPDATES:
q = (Q(id=data['op'].get_thread_id())
& Q(multi_replies__id__gt=data['last_id']))
if query is None:
query = q
else:
query = query | q
if query is not None:
return self.filter(query).annotate(
new_post_count=Count('multi_replies'))
def get_new_post_count(self, datas):
new_posts = self.get_new_posts(datas)
return new_posts.aggregate(total_count=Count('multi_replies'))\
['total_count'] if new_posts else 0
def get_thread_max_posts():
return settings.get_int('Messages', 'MaxPostsPerThread')
class Thread(models.Model):
objects = ThreadManager()
class Meta:
app_label = 'boards'
tags = models.ManyToManyField('Tag', related_name='thread_tags')
bump_time = models.DateTimeField(db_index=True)
last_edit_time = models.DateTimeField()
max_posts = models.IntegerField(default=get_thread_max_posts)
status = models.CharField(max_length=50, default=STATUS_ACTIVE,
choices=STATUS_CHOICES)
monochrome = models.BooleanField(default=False)
def get_tags(self) -> QuerySet:
"""
Gets a sorted tag list.
"""
return self.tags.order_by('name')
def bump(self):
"""
Bumps (moves to up) thread if possible.
"""
if self.can_bump():
self.bump_time = self.last_edit_time
self.update_bump_status()
logger.info('Bumped thread %d' % self.id)
def has_post_limit(self) -> bool:
return self.max_posts > 0
def update_bump_status(self, exclude_posts=None):
if self.has_post_limit() and self.get_reply_count() >= self.max_posts:
self.status = STATUS_BUMPLIMIT
self.update_posts_time(exclude_posts=exclude_posts)
def _get_cache_key(self):
return [datetime_to_epoch(self.last_edit_time)]
@cached_result(key_method=_get_cache_key)
def get_reply_count(self) -> int:
return self.get_replies().count()
@cached_result(key_method=_get_cache_key)
def get_images_count(self) -> int:
return self.get_replies().annotate(images_count=Count(
'images')).aggregate(Sum('images_count'))['images_count__sum']
def can_bump(self) -> bool:
"""
Checks if the thread can be bumped by replying to it.
"""
return self.get_status() == STATUS_ACTIVE
def get_last_replies(self) -> QuerySet:
"""
Gets several last replies, not including opening post
"""
last_replies_count = settings.get_int('View', 'LastRepliesCount')
if last_replies_count > 0:
reply_count = self.get_reply_count()
if reply_count > 0:
reply_count_to_show = min(last_replies_count,
reply_count - 1)
replies = self.get_replies()
last_replies = replies[reply_count - reply_count_to_show:]
return last_replies
def get_skipped_replies_count(self) -> int:
"""
Gets number of posts between opening post and last replies.
"""
reply_count = self.get_reply_count()
last_replies_count = min(settings.get_int('View', 'LastRepliesCount'),
reply_count - 1)
return reply_count - last_replies_count - 1
def get_replies(self, view_fields_only=False) -> QuerySet:
"""
Gets sorted thread posts
"""
query = self.multi_replies.order_by('pub_time').prefetch_related(
'images', 'thread', 'threads', 'attachments')
if view_fields_only:
query = query.defer('poster_ip')
return query.all()
def get_top_level_replies(self) -> QuerySet:
return self.get_replies().exclude(refposts__threads__in=[self])
def get_replies_with_images(self, view_fields_only=False) -> QuerySet:
"""
Gets replies that have at least one image attached
"""
return self.get_replies(view_fields_only).annotate(images_count=Count(
'images')).filter(images_count__gt=0)
def get_opening_post(self, only_id=False) -> Post:
"""
Gets the first post of the thread
"""
query = self.get_replies().filter(opening=True)
if only_id:
query = query.only('id')
opening_post = query.first()
return opening_post
@cached_result()
def get_opening_post_id(self) -> int:
"""
Gets ID of the first thread post.
"""
return self.get_opening_post(only_id=True).id
def get_pub_time(self):
"""
Gets opening post's pub time because thread does not have its own one.
"""
return self.get_opening_post().pub_time
def __str__(self):
return 'T#{}/{}'.format(self.id, self.get_opening_post_id())
def get_tag_url_list(self) -> list:
return boards.models.Tag.objects.get_tag_url_list(self.get_tags())
def update_posts_time(self, exclude_posts=None):
last_edit_time = self.last_edit_time
for post in self.multi_replies.all():
if exclude_posts is None or post not in exclude_posts:
# Manual update is required because uids are generated on save
post.last_edit_time = last_edit_time
post.save(update_fields=['last_edit_time'])
post.get_threads().update(last_edit_time=last_edit_time)
def notify_clients(self):
if not settings.get_bool('External', 'WebsocketsEnabled'):
return
client = Client()
channel_name = WS_CHANNEL_THREAD + str(self.get_opening_post_id())
client.publish(channel_name, {
WS_NOTIFICATION_TYPE: WS_NOTIFICATION_TYPE_NEW_POST,
})
client.send()
def get_absolute_url(self):
return self.get_opening_post().get_absolute_url()
def get_required_tags(self):
return self.get_tags().filter(required=True)
def get_replies_newer(self, post_id):
return self.get_replies().filter(id__gt=post_id)
def is_archived(self):
return self.get_status() == STATUS_ARCHIVE
def get_status(self):
return self.status
def is_monochrome(self):
return self.monochrome
# If tags have parent, add them to the tag list
@transaction.atomic
def refresh_tags(self):
for tag in self.get_tags().all():
parents = tag.get_all_parents()
if len(parents) > 0:
self.tags.add(*parents)
def get_reply_tree(self):
replies = self.get_replies().prefetch_related('refposts')
tree = []
for reply in replies:
parents = reply.refposts.all()
found_parent = False
searching_for_index = False
if len(parents) > 0:
index = 0
parent_depth = 0
indexes_to_insert = []
for depth, element in tree:
index += 1
# If this element is next after parent on the same level,
# insert child before it
if searching_for_index and depth <= parent_depth:
indexes_to_insert.append((index - 1, parent_depth))
searching_for_index = False
if element in parents:
found_parent = True
searching_for_index = True
parent_depth = depth
if not found_parent:
tree.append((0, reply))
else:
if searching_for_index:
tree.append((parent_depth + 1, reply))
offset = 0
for last_index, parent_depth in indexes_to_insert:
tree.insert(last_index + offset, (parent_depth + 1, reply))
offset += 1
return tree