takahe/activities/models/post.py

1218 lines
45 KiB
Python
Raw Normal View History

import datetime
import json
import logging
import mimetypes
import ssl
from collections.abc import Iterable
from typing import Optional
from urllib.parse import urlparse
2022-11-11 22:04:43 -08:00
2022-11-13 17:42:47 -08:00
import httpx
2022-11-11 21:02:43 -08:00
import urlman
from django.conf import settings
2022-11-28 20:47:31 -08:00
from django.contrib.postgres.indexes import GinIndex
from django.contrib.postgres.search import SearchVector
from django.db import models, transaction
from django.db.utils import IntegrityError
from django.template import loader
from django.template.defaultfilters import linebreaks_filter
from django.utils import timezone
from pyld.jsonld import JsonLdError
2022-11-11 21:02:43 -08:00
2022-12-14 23:50:54 -08:00
from activities.models.emoji import Emoji
2022-11-11 22:04:43 -08:00
from activities.models.fan_out import FanOut
from activities.models.hashtag import Hashtag, HashtagStates
from activities.models.post_types import (
PostTypeData,
PostTypeDataDecoder,
PostTypeDataEncoder,
QuestionData,
)
from core.exceptions import ActivityPubFormatError
from core.html import ContentRenderer, FediverseHtmlParser
from core.ld import (
canonicalise,
format_ld_date,
get_list,
get_value_or_map,
parse_ld_date,
)
from core.snowflake import Snowflake
from stator.exceptions import TryAgainLater
2022-11-11 21:02:43 -08:00
from stator.models import State, StateField, StateGraph, StatorModel
2023-01-08 09:36:18 -08:00
from users.models.follow import FollowStates
2023-03-14 14:35:40 -07:00
from users.models.hashtag_follow import HashtagFollow
2022-12-04 20:22:20 -08:00
from users.models.identity import Identity, IdentityStates
from users.models.inbox_message import InboxMessage
2022-12-04 20:13:33 -08:00
from users.models.system_actor import SystemActor
2022-11-11 21:02:43 -08:00
class PostStates(StateGraph):
new = State(try_interval=300)
fanned_out = State(externally_progressed=True)
deleted = State(try_interval=300)
deleted_fanned_out = State(delete_after=86400)
2022-11-11 21:02:43 -08:00
2022-11-27 10:09:46 -08:00
edited = State(try_interval=300)
edited_fanned_out = State(externally_progressed=True)
2022-11-11 21:02:43 -08:00
new.transitions_to(fanned_out)
fanned_out.transitions_to(deleted_fanned_out)
fanned_out.transitions_to(deleted)
2022-11-27 10:09:46 -08:00
fanned_out.transitions_to(edited)
deleted.transitions_to(deleted_fanned_out)
2022-11-27 10:09:46 -08:00
edited.transitions_to(edited_fanned_out)
edited_fanned_out.transitions_to(edited)
edited_fanned_out.transitions_to(deleted)
2022-11-11 21:02:43 -08:00
@classmethod
def targets_fan_out(cls, post: "Post", type_: str) -> None:
# Fan out to each target
for follow in post.get_targets():
FanOut.objects.create(
2022-11-16 05:53:39 -08:00
identity=follow,
2022-11-28 20:41:36 -08:00
type=type_,
2022-11-16 05:53:39 -08:00
subject_post=post,
)
2022-11-28 20:41:36 -08:00
@classmethod
def handle_new(cls, instance: "Post"):
2022-11-28 20:41:36 -08:00
"""
Creates all needed fan-out objects for a new Post.
"""
# Only fan out if the post was published in the last day or it's local
# (we don't want to fan out anything older that that which is remote)
if instance.local or (timezone.now() - instance.published) < datetime.timedelta(
days=1
):
cls.targets_fan_out(instance, FanOut.Types.post)
instance.ensure_hashtags()
return cls.fanned_out
@classmethod
def handle_fanned_out(cls, instance: "Post"):
"""
For remote posts, sees if we can delete them every so often.
"""
2023-10-01 09:49:10 -07:00
# Skip all of this if the horizon is zero
if settings.SETUP.REMOTE_PRUNE_HORIZON <= 0:
return
# To be a candidate for deletion, a post must be remote and old enough
if instance.local:
return
if instance.created > timezone.now() - datetime.timedelta(
days=settings.SETUP.REMOTE_PRUNE_HORIZON
):
return
# It must have no local interactions
if instance.interactions.filter(identity__local=True).exists():
return
# OK, delete it!
instance.delete()
return cls.deleted_fanned_out
@classmethod
def handle_deleted(cls, instance: "Post"):
"""
Creates all needed fan-out objects needed to delete a Post.
"""
cls.targets_fan_out(instance, FanOut.Types.post_deleted)
return cls.deleted_fanned_out
2022-11-11 21:02:43 -08:00
2022-11-27 10:09:46 -08:00
@classmethod
def handle_edited(cls, instance: "Post"):
2022-11-27 10:09:46 -08:00
"""
Creates all needed fan-out objects for an edited Post.
"""
cls.targets_fan_out(instance, FanOut.Types.post_edited)
instance.ensure_hashtags()
2022-11-27 10:09:46 -08:00
return cls.edited_fanned_out
2022-11-11 21:02:43 -08:00
2022-11-28 20:41:36 -08:00
class PostQuerySet(models.QuerySet):
2022-12-11 11:37:28 -08:00
def not_hidden(self):
query = self.exclude(
state__in=[PostStates.deleted, PostStates.deleted_fanned_out]
)
return query
2022-11-28 21:14:52 -08:00
def public(self, include_replies: bool = False):
query = self.filter(
visibility__in=[
Post.Visibilities.public,
Post.Visibilities.local_only,
],
)
if not include_replies:
return query.filter(in_reply_to__isnull=True)
return query
2022-11-28 20:41:36 -08:00
def local_public(self, include_replies: bool = False):
query = self.filter(
visibility__in=[
Post.Visibilities.public,
Post.Visibilities.local_only,
],
local=True,
2022-11-28 20:41:36 -08:00
)
if not include_replies:
return query.filter(in_reply_to__isnull=True)
return query
2022-12-11 11:37:28 -08:00
def unlisted(self, include_replies: bool = False):
query = self.filter(
visibility__in=[
Post.Visibilities.public,
Post.Visibilities.local_only,
Post.Visibilities.unlisted,
],
)
if not include_replies:
return query.filter(in_reply_to__isnull=True)
return query
def visible_to(self, identity: Identity | None, include_replies: bool = False):
if identity is None:
return self.unlisted(include_replies=include_replies)
2022-12-20 01:59:06 -08:00
query = self.filter(
models.Q(
visibility__in=[
Post.Visibilities.public,
Post.Visibilities.local_only,
Post.Visibilities.unlisted,
]
)
| models.Q(
visibility=Post.Visibilities.followers,
author__inbound_follows__source=identity,
)
| models.Q(
mentions=identity,
)
| models.Q(author=identity)
2022-12-20 01:59:06 -08:00
).distinct()
if not include_replies:
return query.filter(in_reply_to__isnull=True)
return query
2022-11-28 20:41:36 -08:00
def tagged_with(self, hashtag: str | Hashtag):
if isinstance(hashtag, str):
tag_q = models.Q(hashtags__contains=hashtag)
else:
tag_q = models.Q(hashtags__contains=hashtag.hashtag)
if hashtag.aliases:
for alias in hashtag.aliases:
tag_q |= models.Q(hashtags__contains=alias)
return self.filter(tag_q)
class PostManager(models.Manager):
def get_queryset(self):
return PostQuerySet(self.model, using=self._db)
2022-12-11 11:37:28 -08:00
def not_hidden(self):
return self.get_queryset().not_hidden()
2022-11-28 21:14:52 -08:00
def public(self, include_replies: bool = False):
return self.get_queryset().public(include_replies=include_replies)
2022-11-28 20:41:36 -08:00
def local_public(self, include_replies: bool = False):
return self.get_queryset().local_public(include_replies=include_replies)
2022-12-11 11:37:28 -08:00
def unlisted(self, include_replies: bool = False):
return self.get_queryset().unlisted(include_replies=include_replies)
2022-11-28 20:41:36 -08:00
def tagged_with(self, hashtag: str | Hashtag):
return self.get_queryset().tagged_with(hashtag=hashtag)
2022-11-11 21:02:43 -08:00
class Post(StatorModel):
"""
A post (status, toot) that is either local or remote.
"""
class Visibilities(models.IntegerChoices):
public = 0
2022-11-26 09:09:31 -08:00
local_only = 4
2022-11-11 21:02:43 -08:00
unlisted = 1
followers = 2
mentioned = 3
class Types(models.TextChoices):
article = "Article"
audio = "Audio"
event = "Event"
image = "Image"
note = "Note"
page = "Page"
question = "Question"
video = "Video"
id = models.BigIntegerField(primary_key=True, default=Snowflake.generate_post)
2022-11-11 21:02:43 -08:00
# The author (attributedTo) of the post
author = models.ForeignKey(
"users.Identity",
2022-11-20 11:59:06 -08:00
on_delete=models.CASCADE,
2022-11-11 21:02:43 -08:00
related_name="posts",
)
# The state the post is in
state = StateField(PostStates)
# If it is our post or not
local = models.BooleanField()
# The canonical object ID
object_uri = models.CharField(max_length=2048, blank=True, null=True, unique=True)
2022-11-11 21:02:43 -08:00
# Who should be able to see this Post
visibility = models.IntegerField(
choices=Visibilities.choices,
default=Visibilities.public,
)
# The main (HTML) content
content = models.TextField()
type = models.CharField(
max_length=20,
choices=Types.choices,
default=Types.note,
)
type_data = models.JSONField(
blank=True, null=True, encoder=PostTypeDataEncoder, decoder=PostTypeDataDecoder
)
2022-11-11 21:02:43 -08:00
# If the contents of the post are sensitive, and the summary (content
# warning) to show if it is
sensitive = models.BooleanField(default=False)
summary = models.TextField(blank=True, null=True)
# The public, web URL of this Post on the original server
url = models.CharField(max_length=2048, blank=True, null=True)
2022-11-11 21:02:43 -08:00
# The Post it is replying to as an AP ID URI
# (as otherwise we'd have to pull entire threads to use IDs)
2023-03-05 09:34:58 -08:00
in_reply_to = models.CharField(max_length=500, blank=True, null=True, db_index=True)
2022-11-11 21:02:43 -08:00
# The identities the post is directly to (who can see it if not public)
to = models.ManyToManyField(
"users.Identity",
related_name="posts_to",
blank=True,
)
# The identities mentioned in the post
mentions = models.ManyToManyField(
"users.Identity",
related_name="posts_mentioning",
blank=True,
)
2022-11-16 05:53:39 -08:00
# Hashtags in the post
hashtags = models.JSONField(blank=True, null=True)
2022-11-16 05:53:39 -08:00
2022-12-14 23:50:54 -08:00
emojis = models.ManyToManyField(
"activities.Emoji",
related_name="posts_using_emoji",
blank=True,
)
# Like/Boost/etc counts
stats = models.JSONField(blank=True, null=True)
2022-11-11 22:04:43 -08:00
# When the post was originally created (as opposed to when we received it)
2022-11-13 17:42:47 -08:00
published = models.DateTimeField(default=timezone.now)
2022-11-11 22:04:43 -08:00
2022-11-16 21:23:32 -08:00
# If the post has been edited after initial publication
edited = models.DateTimeField(blank=True, null=True)
2022-11-11 21:02:43 -08:00
created = models.DateTimeField(auto_now_add=True)
updated = models.DateTimeField(auto_now=True)
2022-11-28 20:41:36 -08:00
objects = PostManager()
2022-11-28 20:47:31 -08:00
class Meta:
indexes = [
GinIndex(fields=["hashtags"], name="hashtags_gin"),
GinIndex(
SearchVector("content", config="english"),
name="content_vector_gin",
),
models.Index(
fields=["visibility", "local", "published"],
name="ix_post_local_public_published",
),
models.Index(
fields=["visibility", "local", "created"],
name="ix_post_local_public_created",
),
]
2022-11-28 20:47:31 -08:00
2022-11-11 21:02:43 -08:00
class urls(urlman.Urls):
2022-11-11 22:04:43 -08:00
view = "{self.author.urls.view}posts/{self.id}/"
object_uri = "{self.author.actor_uri}posts/{self.id}/"
action_like = "{view}like/"
action_unlike = "{view}unlike/"
action_boost = "{view}boost/"
action_unboost = "{view}unboost/"
2023-03-11 10:17:20 -08:00
action_bookmark = "{view}bookmark/"
action_unbookmark = "{view}unbookmark/"
2022-11-24 21:31:45 -08:00
action_delete = "{view}delete/"
2022-11-27 10:09:46 -08:00
action_edit = "{view}edit/"
2022-12-17 13:45:31 -08:00
action_report = "{view}report/"
2022-11-24 14:17:32 -08:00
action_reply = "/compose/?reply_to={self.id}"
admin_edit = "/djadmin/activities/post/{self.id}/change/"
2022-11-11 22:04:43 -08:00
def get_scheme(self, url):
return "https"
def get_hostname(self, url):
return self.instance.author.domain.uri_domain
2022-11-11 21:02:43 -08:00
def __str__(self):
return f"{self.author} #{self.id}"
2022-11-16 22:00:10 -08:00
def get_absolute_url(self):
return self.urls.view
def absolute_object_uri(self):
"""
Returns an object URI that is always absolute, for sending out to
other servers.
"""
if self.local:
return self.author.absolute_profile_uri() + f"posts/{self.id}/"
else:
return self.object_uri
2022-11-24 14:17:32 -08:00
def in_reply_to_post(self) -> Optional["Post"]:
"""
Returns the actual Post object we're replying to, if we can find it
"""
2022-12-11 11:37:28 -08:00
if self.in_reply_to is None:
return None
2022-11-24 14:17:32 -08:00
return (
Post.objects.filter(object_uri=self.in_reply_to)
.select_related("author")
.first()
)
2022-11-21 18:59:27 -08:00
### Content cleanup and extraction ###
def clean_type_data(self, value):
PostTypeData.parse_obj(value)
2022-11-21 18:59:27 -08:00
def _safe_content_note(self, *, local: bool = True):
return ContentRenderer(local=local).render_post(self.content, self)
def _safe_content_question(self, *, local: bool = True):
if local:
context = {
"post": self,
"sanitized_content": self._safe_content_note(local=local),
"local_display": local,
}
return loader.render_to_string("activities/_type_question.html", context)
else:
return ContentRenderer(local=local).render_post(self.content, self)
def _safe_content_typed(self, *, local: bool = True):
context = {
"post": self,
"sanitized_content": self._safe_content_note(local=local),
"local_display": local,
}
return loader.render_to_string(
(
f"activities/_type_{self.type.lower()}.html",
"activities/_type_unknown.html",
),
context,
)
def safe_content(self, *, local: bool = True):
func = getattr(
self, f"_safe_content_{self.type.lower()}", self._safe_content_typed
)
if callable(func):
return func(local=local)
return self._safe_content_note(local=local) # fallback
def safe_content_local(self):
"""
Returns the content formatted for local display
"""
return self.safe_content(local=True)
def safe_content_remote(self):
"""
Returns the content formatted for remote consumption
"""
return self.safe_content(local=False)
2022-11-11 21:02:43 -08:00
def summary_class(self) -> str:
"""
Returns a CSS class name to identify this summary value
"""
if not self.summary:
return ""
return "summary-{self.id}"
@property
def stats_with_defaults(self):
"""
Returns the stats dict with counts of likes/etc. in it
"""
return {
"likes": self.stats.get("likes", 0) if self.stats else 0,
"boosts": self.stats.get("boosts", 0) if self.stats else 0,
"replies": self.stats.get("replies", 0) if self.stats else 0,
}
2022-11-27 11:09:08 -08:00
### Local creation/editing ###
2022-11-11 21:02:43 -08:00
@classmethod
2022-11-13 15:14:38 -08:00
def create_local(
2022-11-16 05:53:39 -08:00
cls,
author: Identity,
content: str,
summary: str | None = None,
2022-12-11 11:37:28 -08:00
sensitive: bool = False,
2022-11-16 05:53:39 -08:00
visibility: int = Visibilities.public,
2022-11-24 14:17:32 -08:00
reply_to: Optional["Post"] = None,
attachments: list | None = None,
question: dict | None = None,
2022-11-13 15:14:38 -08:00
) -> "Post":
with transaction.atomic():
# Find mentions in this post
2022-11-27 11:09:08 -08:00
mentions = cls.mentions_from_content(content, author)
2022-11-24 14:17:32 -08:00
if reply_to:
mentions.add(reply_to.author)
2022-11-26 09:09:31 -08:00
# Maintain local-only for replies
if reply_to.visibility == reply_to.Visibilities.local_only:
visibility = reply_to.Visibilities.local_only
2022-12-14 23:50:54 -08:00
# Find emoji in this post
2022-12-15 10:56:48 -08:00
emojis = Emoji.emojis_from_content(content, None)
# Strip all unwanted HTML and apply linebreaks filter, grabbing hashtags on the way
parser = FediverseHtmlParser(linebreaks_filter(content), find_hashtags=True)
content = parser.html
hashtags = (
sorted([tag[: Hashtag.MAXIMUM_LENGTH] for tag in parser.hashtags])
or None
)
# Make the Post object
post = cls.objects.create(
author=author,
content=content,
summary=summary or None,
2022-12-11 11:37:28 -08:00
sensitive=bool(summary) or sensitive,
local=True,
2022-11-16 05:53:39 -08:00
visibility=visibility,
2022-11-28 20:41:36 -08:00
hashtags=hashtags,
2022-11-24 14:17:32 -08:00
in_reply_to=reply_to.object_uri if reply_to else None,
)
post.object_uri = post.urls.object_uri
post.url = post.absolute_object_uri()
post.mentions.set(mentions)
2022-12-14 23:50:54 -08:00
post.emojis.set(emojis)
2022-12-01 17:46:49 -08:00
if attachments:
post.attachments.set(attachments)
if question:
post.type = question["type"]
post.type_data = PostTypeData(__root__=question).__root__
post.save()
# Recalculate parent stats for replies
if reply_to:
reply_to.calculate_stats()
2022-11-11 21:02:43 -08:00
return post
2022-11-27 11:09:08 -08:00
def edit_local(
self,
content: str,
summary: str | None = None,
2023-02-13 18:40:10 -08:00
sensitive: bool | None = None,
2022-11-27 11:09:08 -08:00
visibility: int = Visibilities.public,
attachments: list | None = None,
attachment_attributes: list | None = None,
2022-11-27 11:09:08 -08:00
):
with transaction.atomic():
# Strip all HTML and apply linebreaks filter
2023-01-29 17:13:21 -08:00
parser = FediverseHtmlParser(linebreaks_filter(content), find_hashtags=True)
self.content = parser.html
self.hashtags = (
sorted([tag[: Hashtag.MAXIMUM_LENGTH] for tag in parser.hashtags])
or None
)
2022-11-27 11:09:08 -08:00
self.summary = summary or None
2023-02-13 18:40:10 -08:00
self.sensitive = bool(summary) if sensitive is None else sensitive
2022-11-27 11:09:08 -08:00
self.visibility = visibility
self.edited = timezone.now()
self.mentions.set(self.mentions_from_content(content, self.author))
2022-12-15 10:56:48 -08:00
self.emojis.set(Emoji.emojis_from_content(content, None))
2022-12-01 17:46:49 -08:00
self.attachments.set(attachments or [])
2022-11-27 11:09:08 -08:00
self.save()
for attrs in attachment_attributes or []:
attachment = next(
(a for a in attachments or [] if str(a.id) == attrs.id), None
)
if attachment is None:
continue
attachment.name = attrs.description
attachment.save()
self.transition_perform(PostStates.edited)
2022-11-27 11:09:08 -08:00
@classmethod
def mentions_from_content(cls, content, author) -> set[Identity]:
mention_hits = FediverseHtmlParser(content, find_mentions=True).mentions
2022-11-27 11:09:08 -08:00
mentions = set()
for handle in mention_hits:
2022-12-04 19:22:24 -08:00
handle = handle.lower()
2022-11-27 11:09:08 -08:00
if "@" in handle:
username, domain = handle.split("@", 1)
else:
username = handle
domain = author.domain_id
identity = Identity.by_username_and_domain(
username=username,
domain=domain,
fetch=True,
)
if identity is not None:
mentions.add(identity)
return mentions
def ensure_hashtags(self) -> None:
2022-11-28 20:41:36 -08:00
"""
Ensure any of the already parsed hashtags from this Post
have a corresponding Hashtag record.
"""
# Ensure hashtags
if self.hashtags:
for hashtag in self.hashtags:
tag, _ = Hashtag.objects.get_or_create(
hashtag=hashtag[: Hashtag.MAXIMUM_LENGTH],
2022-11-28 20:41:36 -08:00
)
tag.transition_perform(HashtagStates.outdated)
2022-11-28 20:41:36 -08:00
def calculate_stats(self, save=True):
"""
Recalculates our stats dict
"""
from activities.models import PostInteraction, PostInteractionStates
self.stats = {
"likes": self.interactions.filter(
type=PostInteraction.Types.like,
state__in=PostInteractionStates.group_active(),
).count(),
"boosts": self.interactions.filter(
type=PostInteraction.Types.boost,
state__in=PostInteractionStates.group_active(),
).count(),
"replies": Post.objects.filter(in_reply_to=self.object_uri).count(),
}
if save:
self.save()
def calculate_type_data(self, save=True):
"""
Recalculate type_data (used mostly for poll votes)
"""
from activities.models import PostInteraction
if self.local and isinstance(self.type_data, QuestionData):
self.type_data.voter_count = (
self.interactions.filter(
type=PostInteraction.Types.vote,
)
.values("identity")
.distinct()
.count()
)
for option in self.type_data.options:
option.votes = self.interactions.filter(
type=PostInteraction.Types.vote,
value=option.name,
).count()
if save:
self.save()
2022-11-11 22:04:43 -08:00
### ActivityPub (outbound) ###
2022-11-11 21:02:43 -08:00
def to_ap(self) -> dict:
2022-11-11 22:04:43 -08:00
"""
Returns the AP JSON for this object
"""
2023-07-22 09:38:22 -07:00
self.author.ensure_uris()
2022-11-11 22:04:43 -08:00
value = {
2022-12-20 06:20:11 -08:00
"to": [],
"cc": [],
"type": self.type,
2022-11-11 22:04:43 -08:00
"id": self.object_uri,
2022-11-13 17:42:47 -08:00
"published": format_ld_date(self.published),
2022-11-11 22:04:43 -08:00
"attributedTo": self.author.actor_uri,
"content": self.safe_content_remote(),
2022-12-15 17:58:11 -08:00
"sensitive": self.sensitive,
"url": self.absolute_object_uri(),
"tag": [],
2022-12-01 17:46:49 -08:00
"attachment": [],
2022-11-11 22:04:43 -08:00
}
if self.type == Post.Types.question and self.type_data:
value[self.type_data.mode] = [
{
"name": option.name,
"type": option.type,
"replies": {"type": "Collection", "totalItems": option.votes},
}
for option in self.type_data.options
]
value["toot:votersCount"] = self.type_data.voter_count
if self.type_data.end_time:
value["endTime"] = format_ld_date(self.type_data.end_time)
2022-11-11 22:04:43 -08:00
if self.summary:
value["summary"] = self.summary
2022-11-24 14:17:32 -08:00
if self.in_reply_to:
value["inReplyTo"] = self.in_reply_to
2022-11-27 10:09:46 -08:00
if self.edited:
value["updated"] = format_ld_date(self.edited)
2022-12-20 06:20:11 -08:00
# Targeting
if self.visibility == self.Visibilities.public:
value["to"].append("as:Public")
2022-12-20 06:20:11 -08:00
elif self.visibility == self.Visibilities.unlisted:
value["cc"].append("as:Public")
2023-07-22 09:38:22 -07:00
elif (
self.visibility == self.Visibilities.followers and self.author.followers_uri
):
value["to"].append(self.author.followers_uri)
# Mentions
for mention in self.mentions.all():
2022-12-14 23:50:54 -08:00
value["tag"].append(mention.to_ap_tag())
value["cc"].append(mention.actor_uri)
2022-12-15 16:00:00 -08:00
# Hashtags
for hashtag in self.hashtags or []:
value["tag"].append(
{
"href": f"https://{self.author.domain.uri_domain}/tags/{hashtag}/",
"name": f"#{hashtag}",
2022-12-15 17:58:11 -08:00
"type": "Hashtag",
2022-12-15 16:00:00 -08:00
}
)
2022-12-14 23:50:54 -08:00
# Emoji
for emoji in self.emojis.all():
value["tag"].append(emoji.to_ap_tag())
2022-12-01 17:46:49 -08:00
# Attachments
for attachment in self.attachments.all():
value["attachment"].append(attachment.to_ap())
# Remove fields if they're empty
2022-12-20 06:20:11 -08:00
for field in ["to", "cc", "tag", "attachment"]:
2022-12-01 17:46:49 -08:00
if not value[field]:
del value[field]
2022-11-11 22:04:43 -08:00
return value
def to_create_ap(self):
"""
Returns the AP JSON to create this object
"""
object = self.to_ap()
2022-11-11 22:04:43 -08:00
return {
2022-12-20 06:20:11 -08:00
"to": object.get("to", []),
"cc": object.get("cc", []),
2022-11-11 22:04:43 -08:00
"type": "Create",
"id": self.object_uri + "#create",
"actor": self.author.actor_uri,
"object": object,
2022-11-11 22:04:43 -08:00
}
2022-11-27 10:09:46 -08:00
def to_update_ap(self):
"""
Returns the AP JSON to update this object
"""
object = self.to_ap()
return {
2022-12-20 06:20:11 -08:00
"to": object.get("to", []),
2022-11-27 10:09:46 -08:00
"cc": object.get("cc", []),
"type": "Update",
"id": self.object_uri + "#update",
"actor": self.author.actor_uri,
"object": object,
}
def to_delete_ap(self):
"""
Returns the AP JSON to create this object
"""
object = self.to_ap()
return {
2022-12-20 06:20:11 -08:00
"to": object.get("to", []),
"cc": object.get("cc", []),
"type": "Delete",
"id": self.object_uri + "#delete",
"actor": self.author.actor_uri,
"object": object,
}
def get_targets(self) -> Iterable[Identity]:
"""
Returns a list of Identities that need to see posts and their changes
"""
targets = set()
for mention in self.mentions.all():
targets.add(mention)
2023-03-14 14:35:40 -07:00
# Then, if it's not mentions only, also deliver to followers and all hashtag followers
if self.visibility != Post.Visibilities.mentioned:
for follower in self.author.inbound_follows.filter(
2023-01-08 09:36:18 -08:00
state__in=FollowStates.group_active()
).select_related("source"):
targets.add(follower.source)
2023-03-14 14:35:40 -07:00
if self.hashtags:
for follow in HashtagFollow.objects.by_hashtags(
2023-03-14 14:35:40 -07:00
self.hashtags
).prefetch_related("identity"):
targets.add(follow.identity)
# If it's a reply, always include the original author if we know them
reply_post = self.in_reply_to_post()
if reply_post:
targets.add(reply_post.author)
# And if it's a reply to one of our own, we have to re-fan-out to
# the original author's followers
if reply_post.author.local:
for follower in reply_post.author.inbound_follows.filter(
2023-01-08 09:36:18 -08:00
state__in=FollowStates.group_active()
).select_related("source"):
targets.add(follower.source)
2022-11-26 09:09:31 -08:00
# If this is a remote post or local-only, filter to only include
# local identities
if not self.local or self.visibility == Post.Visibilities.local_only:
targets = {target for target in targets if target.local}
# If it's a local post, include the author
2022-11-26 09:09:31 -08:00
if self.local:
targets.add(self.author)
2023-01-15 12:35:45 -08:00
# Fetch the author's full blocks and remove them as targets
blocks = (
self.author.outbound_blocks.active()
.filter(mute=False)
.select_related("target")
)
for block in blocks:
try:
targets.remove(block.target)
except KeyError:
pass
2023-01-08 12:46:40 -08:00
# Now dedupe the targets based on shared inboxes (we only keep one per
# shared inbox)
deduped_targets = set()
shared_inboxes = set()
for target in targets:
if target.local or not target.shared_inbox_uri:
deduped_targets.add(target)
elif target.shared_inbox_uri not in shared_inboxes:
shared_inboxes.add(target.shared_inbox_uri)
deduped_targets.add(target)
else:
# Their shared inbox is already being sent to
pass
return deduped_targets
2022-11-11 22:04:43 -08:00
### ActivityPub (inbound) ###
2022-11-11 21:02:43 -08:00
@classmethod
def by_ap(cls, data, create=False, update=False, fetch_author=False) -> "Post":
2022-11-11 21:02:43 -08:00
"""
Retrieves a Post instance by its ActivityPub JSON object.
Optionally creates one if it's not present.
Raises DoesNotExist if it's not found and create is False,
or it's from a blocked domain.
2022-11-11 21:02:43 -08:00
"""
try:
# Ensure data has the primary fields of all Posts
if (
not isinstance(data["id"], str)
or not isinstance(data["attributedTo"], str)
or not isinstance(data["type"], str)
):
raise TypeError()
# Ensure the domain of the object's actor and ID match to prevent injection
if urlparse(data["id"]).hostname != urlparse(data["attributedTo"]).hostname:
raise ValueError("Object's ID domain is different to its author")
except (TypeError, KeyError) as ex:
raise cls.DoesNotExist(
"Object data is not a recognizable ActivityPub object"
) from ex
2022-11-11 21:02:43 -08:00
# Do we have one with the right ID?
2022-11-12 20:14:21 -08:00
created = False
2022-11-11 21:02:43 -08:00
try:
post: Post = cls.objects.select_related("author__domain").get(
2022-12-14 23:50:54 -08:00
object_uri=data["id"]
)
2022-11-11 21:02:43 -08:00
except cls.DoesNotExist:
if create:
# Resolve the author
author = Identity.by_actor_uri(data["attributedTo"], create=create)
# If the author is not fetched yet, try again later
if author.domain is None:
if fetch_author:
if not author.fetch_actor() or author.domain is None:
raise TryAgainLater()
else:
raise TryAgainLater()
# If the post is from a blocked domain, stop and drop
2023-07-22 09:54:36 -07:00
if author.domain.recursively_blocked():
raise cls.DoesNotExist("Post is from a blocked domain")
# parallelism may cause another simultaneous worker thread
# to try to create the same post - so watch for that and
# try to avoid failing the entire transaction
try:
# wrapped in a transaction to avoid breaking the outer
# transaction
with transaction.atomic():
post = cls.objects.create(
object_uri=data["id"],
author=author,
content="",
local=False,
type=data["type"],
)
created = True
except IntegrityError:
# despite previous checks, a parallel thread managed
# to create the same object already
raise TryAgainLater()
2022-11-11 21:02:43 -08:00
else:
raise cls.DoesNotExist(f"No post with ID {data['id']}", data)
2022-11-12 20:14:21 -08:00
if update or created:
post.type = data["type"]
post.url = data.get("url", data["id"])
if post.type in (cls.Types.article, cls.Types.question):
post.type_data = PostTypeData(__root__=data).__root__
try:
# apparently sometimes posts (Pages?) in the fediverse
# don't have content, but this shouldn't be a total failure
post.content = get_value_or_map(data, "content", "contentMap")
except ActivityPubFormatError as err:
logging.warning(f"{err} on {post.url}")
post.content = None
# Document types have names, not summaries
post.summary = data.get("summary") or data.get("name")
if not post.content and post.summary:
post.content = post.summary
post.summary = None
2022-12-16 15:38:52 -08:00
post.sensitive = data.get("sensitive", False)
2022-11-16 22:00:10 -08:00
post.published = parse_ld_date(data.get("published"))
post.edited = parse_ld_date(data.get("updated"))
post.in_reply_to = data.get("inReplyTo")
2022-11-16 05:53:39 -08:00
# Mentions and hashtags
post.hashtags = []
for tag in get_list(data, "tag"):
tag_type = tag["type"].lower()
if tag_type == "mention":
2022-11-16 05:53:39 -08:00
mention_identity = Identity.by_actor_uri(tag["href"], create=True)
post.mentions.add(mention_identity)
elif tag_type in ["_:hashtag", "hashtag"]:
# kbin produces tags with 'tag' instead of 'name'
if "tag" in tag and "name" not in tag:
name = get_value_or_map(tag, "tag", "tagMap")
else:
name = get_value_or_map(tag, "name", "nameMap")
post.hashtags.append(
name.lower().lstrip("#")[: Hashtag.MAXIMUM_LENGTH]
)
elif tag_type in ["toot:emoji", "emoji"]:
2022-12-14 23:50:54 -08:00
emoji = Emoji.by_ap_tag(post.author.domain, tag, create=True)
post.emojis.add(emoji)
2022-11-16 05:53:39 -08:00
else:
# Various ActivityPub implementations and proposals introduced tag
# types, e.g. Edition in Bookwyrm and Link in fep-e232 Object Links
# it should be safe to ignore (and log) them before a full support
pass
2022-11-16 05:53:39 -08:00
# Visibility and to
2022-12-20 06:20:11 -08:00
# (a post is public if it's to:public, otherwise it's unlisted if
# it's cc:public, otherwise it's more limited)
to = [x.lower() for x in get_list(data, "to")]
cc = [x.lower() for x in get_list(data, "cc")]
post.visibility = Post.Visibilities.mentioned
if "public" in to or "as:public" in to:
post.visibility = Post.Visibilities.public
elif "public" in cc or "as:public" in cc:
post.visibility = Post.Visibilities.unlisted
2023-07-22 09:38:22 -07:00
elif post.author.followers_uri in to:
post.visibility = Post.Visibilities.followers
2022-11-16 22:00:10 -08:00
# Attachments
# These have no IDs, so we have to wipe them each time
post.attachments.all().delete()
for attachment in get_list(data, "attachment"):
if "url" not in attachment and "href" in attachment:
# Links have hrefs, while other Objects have urls
attachment["url"] = attachment["href"]
2022-12-18 10:00:21 -08:00
if "focalPoint" in attachment:
try:
focal_x, focal_y = attachment["focalPoint"]
except (ValueError, TypeError):
focal_x, focal_y = None, None
2022-11-16 22:00:10 -08:00
else:
focal_x, focal_y = None, None
mimetype = attachment.get("mediaType")
if not mimetype or not isinstance(mimetype, str):
if "url" not in attachment:
raise ActivityPubFormatError(
f"No URL present on attachment in {post.url}"
)
mimetype, _ = mimetypes.guess_type(attachment["url"])
if not mimetype:
mimetype = "application/octet-stream"
2022-11-16 22:00:10 -08:00
post.attachments.create(
remote_url=attachment["url"],
mimetype=mimetype,
2022-11-16 22:00:10 -08:00
name=attachment.get("name"),
width=attachment.get("width"),
height=attachment.get("height"),
2022-12-18 10:00:21 -08:00
blurhash=attachment.get("blurhash"),
2022-11-16 22:00:10 -08:00
focal_x=focal_x,
focal_y=focal_y,
)
# Calculate stats in case we have existing replies
post.calculate_stats(save=False)
with transaction.atomic():
# if we don't commit the transaction here, there's a chance
# the parent fetch below goes into an infinite loop
post.save()
# Potentially schedule a fetch of the reply parent, and recalculate
# its stats if it's here already.
if post.in_reply_to:
try:
parent = cls.by_object_uri(post.in_reply_to)
except cls.DoesNotExist:
try:
cls.ensure_object_uri(post.in_reply_to, reason=post.object_uri)
except ValueError:
logging.warning(
f"Cannot fetch ancestor of Post={post.pk}, ancestor_uri={post.in_reply_to}"
)
else:
parent.calculate_stats()
2022-11-12 20:14:21 -08:00
return post
2022-11-11 21:02:43 -08:00
2022-11-13 17:42:47 -08:00
@classmethod
def by_object_uri(cls, object_uri, fetch=False) -> "Post":
2022-11-13 17:42:47 -08:00
"""
Gets the post by URI - either looking up locally, or fetching
from the other end if it's not here.
"""
try:
return cls.objects.get(object_uri=object_uri)
except cls.DoesNotExist:
if fetch:
2022-12-04 20:13:33 -08:00
try:
response = SystemActor().signed_request(
2022-12-04 20:13:33 -08:00
method="get", uri=object_uri
)
except (httpx.HTTPError, ssl.SSLCertVerificationError, ValueError):
2022-12-04 20:13:33 -08:00
raise cls.DoesNotExist(f"Could not fetch {object_uri}")
if response.status_code in [404, 410]:
raise cls.DoesNotExist(f"No post at {object_uri}")
if response.status_code >= 500:
raise cls.DoesNotExist(f"Server error fetching {object_uri}")
if response.status_code >= 400:
raise cls.DoesNotExist(
f"Error fetching post from {object_uri}: {response.status_code}",
{response.content},
2022-11-13 17:42:47 -08:00
)
try:
post = cls.by_ap(
canonicalise(response.json(), include_security=True),
create=True,
update=True,
fetch_author=True,
)
except (json.JSONDecodeError, ValueError, JsonLdError) as err:
raise cls.DoesNotExist(
f"Invalid ld+json response for {object_uri}"
) from err
2022-12-04 20:22:20 -08:00
# We may need to fetch the author too
if post.author.state == IdentityStates.outdated:
post.author.fetch_actor()
2022-12-04 20:22:20 -08:00
return post
2022-12-04 20:13:33 -08:00
else:
raise cls.DoesNotExist(f"Cannot find Post with URI {object_uri}")
2022-11-13 17:42:47 -08:00
@classmethod
def ensure_object_uri(cls, object_uri: str, reason: str | None = None):
"""
Sees if the post is in our local set, and if not, schedules a fetch
for it (in the background)
"""
if not object_uri or "://" not in object_uri:
raise ValueError("URI missing or invalid")
try:
cls.by_object_uri(object_uri)
except cls.DoesNotExist:
InboxMessage.create_internal(
{
"type": "FetchPost",
"object": object_uri,
"reason": reason,
}
)
2022-11-11 21:02:43 -08:00
@classmethod
def handle_create_ap(cls, data):
"""
Handles an incoming create request
"""
with transaction.atomic():
# Ensure the Create actor is the Post's attributedTo
if data["actor"] != data["object"]["attributedTo"]:
raise ValueError("Create actor does not match its Post object", data)
# Create it, stator will fan it out locally
cls.by_ap(data["object"], create=True, update=True, fetch_author=True)
2022-11-13 18:03:43 -08:00
2022-11-16 21:23:32 -08:00
@classmethod
def handle_update_ap(cls, data):
"""
Handles an incoming update request
"""
with transaction.atomic():
# Ensure the Create actor is the Post's attributedTo
if data["actor"] != data["object"]["attributedTo"]:
raise ValueError("Create actor does not match its Post object", data)
# Find it and update it
try:
cls.by_ap(data["object"], create=False, update=True)
except cls.DoesNotExist:
# We don't have a copy - assume we got a delete first and ignore.
pass
2022-11-16 21:23:32 -08:00
@classmethod
def handle_delete_ap(cls, data):
"""
2022-11-25 18:33:46 -08:00
Handles an incoming delete request
"""
with transaction.atomic():
# Is this an embedded object or plain ID?
if isinstance(data["object"], str):
object_uri = data["object"]
else:
object_uri = data["object"]["id"]
# Find our post by ID if we have one
try:
post = cls.by_object_uri(object_uri)
except cls.DoesNotExist:
# It's already been deleted
return
# Ensure the actor on the request authored the post
if not post.author.actor_uri == data["actor"]:
raise ValueError("Actor on delete does not match object")
post.delete()
@classmethod
def handle_fetch_internal(cls, data):
"""
Handles an internal fetch-request inbox message
"""
try:
uri = data["object"]
if "://" in uri:
cls.by_object_uri(uri, fetch=True)
except (cls.DoesNotExist, KeyError):
pass
### OpenGraph API ###
def to_opengraph_dict(self) -> dict:
return {
"og:title": f"{self.author.name} (@{self.author.handle})",
"og:type": "article",
"og:published_time": (self.published or self.created).isoformat(),
"og:modified_time": (
self.edited or self.published or self.created
).isoformat(),
"og:description": (self.summary or self.safe_content_local()),
"og:image:url": self.author.local_icon_url().absolute,
"og:image:height": 85,
"og:image:width": 85,
}
2022-12-10 23:25:48 -08:00
### Mastodon API ###
2023-03-11 10:17:20 -08:00
def to_mastodon_json(self, interactions=None, bookmarks=None, identity=None):
2022-12-10 23:25:48 -08:00
reply_parent = None
if self.in_reply_to:
# Load the PK and author.id explicitly to prevent a SELECT on the entire author Identity
reply_parent = (
Post.objects.filter(object_uri=self.in_reply_to)
.only("pk", "author_id")
.first()
)
2022-12-11 23:54:51 -08:00
visibility_mapping = {
self.Visibilities.public: "public",
self.Visibilities.unlisted: "unlisted",
self.Visibilities.followers: "private",
self.Visibilities.mentioned: "direct",
self.Visibilities.local_only: "public",
}
2022-12-11 11:37:28 -08:00
value = {
2022-12-10 23:25:48 -08:00
"id": self.pk,
"uri": self.object_uri,
"created_at": format_ld_date(self.published),
"account": self.author.to_mastodon_json(include_counts=False),
2022-12-10 23:25:48 -08:00
"content": self.safe_content_remote(),
2022-12-11 23:54:51 -08:00
"visibility": visibility_mapping[self.visibility],
2022-12-10 23:25:48 -08:00
"sensitive": self.sensitive,
"spoiler_text": self.summary or "",
"media_attachments": [
attachment.to_mastodon_json() for attachment in self.attachments.all()
],
"mentions": [
mention.to_mastodon_mention_json() for mention in self.mentions.all()
2022-12-10 23:25:48 -08:00
],
"tags": (
[
{
"name": tag,
"url": f"https://{self.author.domain.uri_domain}/tags/{tag}/",
}
for tag in self.hashtags
]
2022-12-10 23:25:48 -08:00
if self.hashtags
else []
),
# Filter in the list comp rather than query because the common case is no emoji in the resultset
# When filter is on emojis like `emojis.usable()` it causes a query that is not cached by prefetch_related
"emojis": [
emoji.to_mastodon_json()
for emoji in self.emojis.all()
if emoji.is_usable
],
"reblogs_count": self.stats_with_defaults["boosts"],
"favourites_count": self.stats_with_defaults["likes"],
"replies_count": self.stats_with_defaults["replies"],
2022-12-10 23:25:48 -08:00
"url": self.absolute_object_uri(),
"in_reply_to_id": reply_parent.pk if reply_parent else None,
"in_reply_to_account_id": (
reply_parent.author_id if reply_parent else None
),
2022-12-10 23:25:48 -08:00
"reblog": None,
"poll": self.type_data.to_mastodon_json(self, identity)
if isinstance(self.type_data, QuestionData)
else None,
2022-12-10 23:25:48 -08:00
"card": None,
"language": None,
"text": self.safe_content_remote(),
2022-12-10 23:25:48 -08:00
"edited_at": format_ld_date(self.edited) if self.edited else None,
}
2022-12-11 11:37:28 -08:00
if interactions:
value["favourited"] = self.pk in interactions.get("like", [])
value["reblogged"] = self.pk in interactions.get("boost", [])
2023-05-13 09:01:27 -07:00
value["pinned"] = self.pk in interactions.get("pin", [])
2023-03-11 10:17:20 -08:00
if bookmarks:
value["bookmarked"] = self.pk in bookmarks
2022-12-11 11:37:28 -08:00
return value