takahe/activities/models/post.py

1193 lines
44 KiB
Python
Raw Normal View History

import datetime
import json
import mimetypes
import ssl
from collections.abc import Iterable
from typing import Optional
from urllib.parse import urlparse
2022-11-11 22:04:43 -08:00
2022-11-13 17:42:47 -08:00
import httpx
2022-11-11 21:02:43 -08:00
import urlman
2022-11-28 20:47:31 -08:00
from django.contrib.postgres.indexes import GinIndex
from django.contrib.postgres.search import SearchVector
from django.db import models, transaction
from django.db.utils import IntegrityError
from django.template import loader
from django.template.defaultfilters import linebreaks_filter
from django.utils import timezone
from pyld.jsonld import JsonLdError
2022-11-11 21:02:43 -08:00
2022-12-14 23:50:54 -08:00
from activities.models.emoji import Emoji
2022-11-11 22:04:43 -08:00
from activities.models.fan_out import FanOut
from activities.models.hashtag import Hashtag, HashtagStates
from activities.models.post_types import (
PostTypeData,
PostTypeDataDecoder,
PostTypeDataEncoder,
QuestionData,
)
from core.exceptions import ActivityPubFormatError, capture_message
from core.html import ContentRenderer, FediverseHtmlParser
from core.ld import (
canonicalise,
format_ld_date,
get_list,
get_value_or_map,
parse_ld_date,
)
from core.snowflake import Snowflake
from stator.exceptions import TryAgainLater
2022-11-11 21:02:43 -08:00
from stator.models import State, StateField, StateGraph, StatorModel
2023-01-08 09:36:18 -08:00
from users.models.follow import FollowStates
2023-03-14 14:35:40 -07:00
from users.models.hashtag_follow import HashtagFollow
2022-12-04 20:22:20 -08:00
from users.models.identity import Identity, IdentityStates
from users.models.inbox_message import InboxMessage
2022-12-04 20:13:33 -08:00
from users.models.system_actor import SystemActor
2022-11-11 21:02:43 -08:00
class PostStates(StateGraph):
new = State(try_interval=300)
fanned_out = State(externally_progressed=True)
deleted = State(try_interval=300)
deleted_fanned_out = State(delete_after=24 * 60 * 60)
2022-11-11 21:02:43 -08:00
2022-11-27 10:09:46 -08:00
edited = State(try_interval=300)
edited_fanned_out = State(externally_progressed=True)
2022-11-11 21:02:43 -08:00
new.transitions_to(fanned_out)
fanned_out.transitions_to(deleted)
2022-11-27 10:09:46 -08:00
fanned_out.transitions_to(edited)
deleted.transitions_to(deleted_fanned_out)
2022-11-27 10:09:46 -08:00
edited.transitions_to(edited_fanned_out)
edited_fanned_out.transitions_to(edited)
edited_fanned_out.transitions_to(deleted)
2022-11-11 21:02:43 -08:00
@classmethod
def targets_fan_out(cls, post: "Post", type_: str) -> None:
# Fan out to each target
for follow in post.get_targets():
FanOut.objects.create(
2022-11-16 05:53:39 -08:00
identity=follow,
2022-11-28 20:41:36 -08:00
type=type_,
2022-11-16 05:53:39 -08:00
subject_post=post,
)
2022-11-28 20:41:36 -08:00
@classmethod
def handle_new(cls, instance: "Post"):
2022-11-28 20:41:36 -08:00
"""
Creates all needed fan-out objects for a new Post.
"""
# Only fan out if the post was published in the last day or it's local
# (we don't want to fan out anything older that that which is remote)
if instance.local or (timezone.now() - instance.published) < datetime.timedelta(
days=1
):
cls.targets_fan_out(instance, FanOut.Types.post)
instance.ensure_hashtags()
return cls.fanned_out
@classmethod
def handle_deleted(cls, instance: "Post"):
"""
Creates all needed fan-out objects needed to delete a Post.
"""
cls.targets_fan_out(instance, FanOut.Types.post_deleted)
return cls.deleted_fanned_out
2022-11-11 21:02:43 -08:00
2022-11-27 10:09:46 -08:00
@classmethod
def handle_edited(cls, instance: "Post"):
2022-11-27 10:09:46 -08:00
"""
Creates all needed fan-out objects for an edited Post.
"""
cls.targets_fan_out(instance, FanOut.Types.post_edited)
instance.ensure_hashtags()
2022-11-27 10:09:46 -08:00
return cls.edited_fanned_out
2022-11-11 21:02:43 -08:00
2022-11-28 20:41:36 -08:00
class PostQuerySet(models.QuerySet):
2022-12-11 11:37:28 -08:00
def not_hidden(self):
query = self.exclude(
state__in=[PostStates.deleted, PostStates.deleted_fanned_out]
)
return query
2022-11-28 21:14:52 -08:00
def public(self, include_replies: bool = False):
query = self.filter(
visibility__in=[
Post.Visibilities.public,
Post.Visibilities.local_only,
],
)
if not include_replies:
return query.filter(in_reply_to__isnull=True)
return query
2022-11-28 20:41:36 -08:00
def local_public(self, include_replies: bool = False):
query = self.filter(
visibility__in=[
Post.Visibilities.public,
Post.Visibilities.local_only,
],
local=True,
2022-11-28 20:41:36 -08:00
)
if not include_replies:
return query.filter(in_reply_to__isnull=True)
return query
2022-12-11 11:37:28 -08:00
def unlisted(self, include_replies: bool = False):
query = self.filter(
visibility__in=[
Post.Visibilities.public,
Post.Visibilities.local_only,
Post.Visibilities.unlisted,
],
)
if not include_replies:
return query.filter(in_reply_to__isnull=True)
return query
def visible_to(self, identity: Identity | None, include_replies: bool = False):
if identity is None:
return self.unlisted(include_replies=include_replies)
2022-12-20 01:59:06 -08:00
query = self.filter(
models.Q(
visibility__in=[
Post.Visibilities.public,
Post.Visibilities.local_only,
Post.Visibilities.unlisted,
]
)
| models.Q(
visibility=Post.Visibilities.followers,
author__inbound_follows__source=identity,
)
| models.Q(
mentions=identity,
)
| models.Q(author=identity)
2022-12-20 01:59:06 -08:00
).distinct()
if not include_replies:
return query.filter(in_reply_to__isnull=True)
return query
2022-11-28 20:41:36 -08:00
def tagged_with(self, hashtag: str | Hashtag):
if isinstance(hashtag, str):
tag_q = models.Q(hashtags__contains=hashtag)
else:
tag_q = models.Q(hashtags__contains=hashtag.hashtag)
if hashtag.aliases:
for alias in hashtag.aliases:
tag_q |= models.Q(hashtags__contains=alias)
return self.filter(tag_q)
class PostManager(models.Manager):
def get_queryset(self):
return PostQuerySet(self.model, using=self._db)
2022-12-11 11:37:28 -08:00
def not_hidden(self):
return self.get_queryset().not_hidden()
2022-11-28 21:14:52 -08:00
def public(self, include_replies: bool = False):
return self.get_queryset().public(include_replies=include_replies)
2022-11-28 20:41:36 -08:00
def local_public(self, include_replies: bool = False):
return self.get_queryset().local_public(include_replies=include_replies)
2022-12-11 11:37:28 -08:00
def unlisted(self, include_replies: bool = False):
return self.get_queryset().unlisted(include_replies=include_replies)
2022-11-28 20:41:36 -08:00
def tagged_with(self, hashtag: str | Hashtag):
return self.get_queryset().tagged_with(hashtag=hashtag)
2022-11-11 21:02:43 -08:00
class Post(StatorModel):
"""
A post (status, toot) that is either local or remote.
"""
class Visibilities(models.IntegerChoices):
public = 0
2022-11-26 09:09:31 -08:00
local_only = 4
2022-11-11 21:02:43 -08:00
unlisted = 1
followers = 2
mentioned = 3
class Types(models.TextChoices):
article = "Article"
audio = "Audio"
event = "Event"
image = "Image"
note = "Note"
page = "Page"
question = "Question"
video = "Video"
id = models.BigIntegerField(primary_key=True, default=Snowflake.generate_post)
2022-11-11 21:02:43 -08:00
# The author (attributedTo) of the post
author = models.ForeignKey(
"users.Identity",
2022-11-20 11:59:06 -08:00
on_delete=models.CASCADE,
2022-11-11 21:02:43 -08:00
related_name="posts",
)
# The state the post is in
state = StateField(PostStates)
# If it is our post or not
local = models.BooleanField()
# The canonical object ID
object_uri = models.CharField(max_length=2048, blank=True, null=True, unique=True)
2022-11-11 21:02:43 -08:00
# Who should be able to see this Post
visibility = models.IntegerField(
choices=Visibilities.choices,
default=Visibilities.public,
)
# The main (HTML) content
content = models.TextField()
type = models.CharField(
max_length=20,
choices=Types.choices,
default=Types.note,
)
type_data = models.JSONField(
blank=True, null=True, encoder=PostTypeDataEncoder, decoder=PostTypeDataDecoder
)
2022-11-11 21:02:43 -08:00
# If the contents of the post are sensitive, and the summary (content
# warning) to show if it is
sensitive = models.BooleanField(default=False)
summary = models.TextField(blank=True, null=True)
# The public, web URL of this Post on the original server
url = models.CharField(max_length=2048, blank=True, null=True)
2022-11-11 21:02:43 -08:00
# The Post it is replying to as an AP ID URI
# (as otherwise we'd have to pull entire threads to use IDs)
2023-03-05 09:34:58 -08:00
in_reply_to = models.CharField(max_length=500, blank=True, null=True, db_index=True)
2022-11-11 21:02:43 -08:00
# The identities the post is directly to (who can see it if not public)
to = models.ManyToManyField(
"users.Identity",
related_name="posts_to",
blank=True,
)
# The identities mentioned in the post
mentions = models.ManyToManyField(
"users.Identity",
related_name="posts_mentioning",
blank=True,
)
2022-11-16 05:53:39 -08:00
# Hashtags in the post
hashtags = models.JSONField(blank=True, null=True)
2022-11-16 05:53:39 -08:00
2022-12-14 23:50:54 -08:00
emojis = models.ManyToManyField(
"activities.Emoji",
related_name="posts_using_emoji",
blank=True,
)
# Like/Boost/etc counts
stats = models.JSONField(blank=True, null=True)
2022-11-11 22:04:43 -08:00
# When the post was originally created (as opposed to when we received it)
2022-11-13 17:42:47 -08:00
published = models.DateTimeField(default=timezone.now)
2022-11-11 22:04:43 -08:00
2022-11-16 21:23:32 -08:00
# If the post has been edited after initial publication
edited = models.DateTimeField(blank=True, null=True)
2022-11-11 21:02:43 -08:00
created = models.DateTimeField(auto_now_add=True)
updated = models.DateTimeField(auto_now=True)
2022-11-28 20:41:36 -08:00
objects = PostManager()
2022-11-28 20:47:31 -08:00
class Meta:
indexes = [
GinIndex(fields=["hashtags"], name="hashtags_gin"),
GinIndex(
SearchVector("content", config="english"),
name="content_vector_gin",
),
models.Index(
fields=["visibility", "local", "published"],
name="ix_post_local_public_published",
),
models.Index(
fields=["visibility", "local", "created"],
name="ix_post_local_public_created",
),
]
2022-11-28 20:47:31 -08:00
2022-11-11 21:02:43 -08:00
class urls(urlman.Urls):
2022-11-11 22:04:43 -08:00
view = "{self.author.urls.view}posts/{self.id}/"
object_uri = "{self.author.actor_uri}posts/{self.id}/"
action_like = "{view}like/"
action_unlike = "{view}unlike/"
action_boost = "{view}boost/"
action_unboost = "{view}unboost/"
2023-03-11 10:17:20 -08:00
action_bookmark = "{view}bookmark/"
action_unbookmark = "{view}unbookmark/"
2022-11-24 21:31:45 -08:00
action_delete = "{view}delete/"
2022-11-27 10:09:46 -08:00
action_edit = "{view}edit/"
2022-12-17 13:45:31 -08:00
action_report = "{view}report/"
2022-11-24 14:17:32 -08:00
action_reply = "/compose/?reply_to={self.id}"
admin_edit = "/djadmin/activities/post/{self.id}/change/"
2022-11-11 22:04:43 -08:00
def get_scheme(self, url):
return "https"
def get_hostname(self, url):
return self.instance.author.domain.uri_domain
2022-11-11 21:02:43 -08:00
def __str__(self):
return f"{self.author} #{self.id}"
2022-11-16 22:00:10 -08:00
def get_absolute_url(self):
return self.urls.view
def absolute_object_uri(self):
"""
Returns an object URI that is always absolute, for sending out to
other servers.
"""
if self.local:
return self.author.absolute_profile_uri() + f"posts/{self.id}/"
else:
return self.object_uri
2022-11-24 14:17:32 -08:00
def in_reply_to_post(self) -> Optional["Post"]:
"""
Returns the actual Post object we're replying to, if we can find it
"""
2022-12-11 11:37:28 -08:00
if self.in_reply_to is None:
return None
2022-11-24 14:17:32 -08:00
return (
Post.objects.filter(object_uri=self.in_reply_to)
.select_related("author")
.first()
)
2022-11-21 18:59:27 -08:00
### Content cleanup and extraction ###
def clean_type_data(self, value):
PostTypeData.parse_obj(value)
2022-11-21 18:59:27 -08:00
def _safe_content_note(self, *, local: bool = True):
return ContentRenderer(local=local).render_post(self.content, self)
def _safe_content_question(self, *, local: bool = True):
if local:
context = {
"post": self,
"sanitized_content": self._safe_content_note(local=local),
"local_display": local,
}
return loader.render_to_string("activities/_type_question.html", context)
else:
return ContentRenderer(local=local).render_post(self.content, self)
def _safe_content_typed(self, *, local: bool = True):
context = {
"post": self,
"sanitized_content": self._safe_content_note(local=local),
"local_display": local,
}
return loader.render_to_string(
(
f"activities/_type_{self.type.lower()}.html",
"activities/_type_unknown.html",
),
context,
)
def safe_content(self, *, local: bool = True):
func = getattr(
self, f"_safe_content_{self.type.lower()}", self._safe_content_typed
)
if callable(func):
return func(local=local)
return self._safe_content_note(local=local) # fallback
def safe_content_local(self):
"""
Returns the content formatted for local display
"""
return self.safe_content(local=True)
def safe_content_remote(self):
"""
Returns the content formatted for remote consumption
"""
return self.safe_content(local=False)
2022-11-11 21:02:43 -08:00
def summary_class(self) -> str:
"""
Returns a CSS class name to identify this summary value
"""
if not self.summary:
return ""
return "summary-{self.id}"
@property
def stats_with_defaults(self):
"""
Returns the stats dict with counts of likes/etc. in it
"""
return {
"likes": self.stats.get("likes", 0) if self.stats else 0,
"boosts": self.stats.get("boosts", 0) if self.stats else 0,
"replies": self.stats.get("replies", 0) if self.stats else 0,
}
2022-11-27 11:09:08 -08:00
### Local creation/editing ###
2022-11-11 21:02:43 -08:00
@classmethod
2022-11-13 15:14:38 -08:00
def create_local(
2022-11-16 05:53:39 -08:00
cls,
author: Identity,
content: str,
summary: str | None = None,
2022-12-11 11:37:28 -08:00
sensitive: bool = False,
2022-11-16 05:53:39 -08:00
visibility: int = Visibilities.public,
2022-11-24 14:17:32 -08:00
reply_to: Optional["Post"] = None,
attachments: list | None = None,
question: dict | None = None,
2022-11-13 15:14:38 -08:00
) -> "Post":
with transaction.atomic():
# Find mentions in this post
2022-11-27 11:09:08 -08:00
mentions = cls.mentions_from_content(content, author)
2022-11-24 14:17:32 -08:00
if reply_to:
mentions.add(reply_to.author)
2022-11-26 09:09:31 -08:00
# Maintain local-only for replies
if reply_to.visibility == reply_to.Visibilities.local_only:
visibility = reply_to.Visibilities.local_only
2022-12-14 23:50:54 -08:00
# Find emoji in this post
2022-12-15 10:56:48 -08:00
emojis = Emoji.emojis_from_content(content, None)
# Strip all unwanted HTML and apply linebreaks filter, grabbing hashtags on the way
parser = FediverseHtmlParser(linebreaks_filter(content), find_hashtags=True)
content = parser.html
hashtags = (
sorted([tag[: Hashtag.MAXIMUM_LENGTH] for tag in parser.hashtags])
or None
)
# Make the Post object
post = cls.objects.create(
author=author,
content=content,
summary=summary or None,
2022-12-11 11:37:28 -08:00
sensitive=bool(summary) or sensitive,
local=True,
2022-11-16 05:53:39 -08:00
visibility=visibility,
2022-11-28 20:41:36 -08:00
hashtags=hashtags,
2022-11-24 14:17:32 -08:00
in_reply_to=reply_to.object_uri if reply_to else None,
)
post.object_uri = post.urls.object_uri
post.url = post.absolute_object_uri()
post.mentions.set(mentions)
2022-12-14 23:50:54 -08:00
post.emojis.set(emojis)
2022-12-01 17:46:49 -08:00
if attachments:
post.attachments.set(attachments)
if question:
post.type = question["type"]
post.type_data = PostTypeData(__root__=question).__root__
post.save()
# Recalculate parent stats for replies
if reply_to:
reply_to.calculate_stats()
2022-11-11 21:02:43 -08:00
return post
2022-11-27 11:09:08 -08:00
def edit_local(
self,
content: str,
summary: str | None = None,
2023-02-13 18:40:10 -08:00
sensitive: bool | None = None,
2022-11-27 11:09:08 -08:00
visibility: int = Visibilities.public,
attachments: list | None = None,
attachment_attributes: list | None = None,
2022-11-27 11:09:08 -08:00
):
with transaction.atomic():
# Strip all HTML and apply linebreaks filter
2023-01-29 17:13:21 -08:00
parser = FediverseHtmlParser(linebreaks_filter(content), find_hashtags=True)
self.content = parser.html
self.hashtags = (
sorted([tag[: Hashtag.MAXIMUM_LENGTH] for tag in parser.hashtags])
or None
)
2022-11-27 11:09:08 -08:00
self.summary = summary or None
2023-02-13 18:40:10 -08:00
self.sensitive = bool(summary) if sensitive is None else sensitive
2022-11-27 11:09:08 -08:00
self.visibility = visibility
self.edited = timezone.now()
self.mentions.set(self.mentions_from_content(content, self.author))
2022-12-15 10:56:48 -08:00
self.emojis.set(Emoji.emojis_from_content(content, None))
2022-12-01 17:46:49 -08:00
self.attachments.set(attachments or [])
2022-11-27 11:09:08 -08:00
self.save()
for attrs in attachment_attributes or []:
attachment = next(
(a for a in attachments or [] if str(a.id) == attrs.id), None
)
if attachment is None:
continue
attachment.name = attrs.description
attachment.save()
self.transition_perform(PostStates.edited)
2022-11-27 11:09:08 -08:00
@classmethod
def mentions_from_content(cls, content, author) -> set[Identity]:
mention_hits = FediverseHtmlParser(content, find_mentions=True).mentions
2022-11-27 11:09:08 -08:00
mentions = set()
for handle in mention_hits:
2022-12-04 19:22:24 -08:00
handle = handle.lower()
2022-11-27 11:09:08 -08:00
if "@" in handle:
username, domain = handle.split("@", 1)
else:
username = handle
domain = author.domain_id
identity = Identity.by_username_and_domain(
username=username,
domain=domain,
fetch=True,
)
if identity is not None:
mentions.add(identity)
return mentions
def ensure_hashtags(self) -> None:
2022-11-28 20:41:36 -08:00
"""
Ensure any of the already parsed hashtags from this Post
have a corresponding Hashtag record.
"""
# Ensure hashtags
if self.hashtags:
for hashtag in self.hashtags:
tag, _ = Hashtag.objects.get_or_create(
hashtag=hashtag[: Hashtag.MAXIMUM_LENGTH],
2022-11-28 20:41:36 -08:00
)
tag.transition_perform(HashtagStates.outdated)
2022-11-28 20:41:36 -08:00
def calculate_stats(self, save=True):
"""
Recalculates our stats dict
"""
from activities.models import PostInteraction, PostInteractionStates
self.stats = {
"likes": self.interactions.filter(
type=PostInteraction.Types.like,
state__in=PostInteractionStates.group_active(),
).count(),
"boosts": self.interactions.filter(
type=PostInteraction.Types.boost,
state__in=PostInteractionStates.group_active(),
).count(),
"replies": Post.objects.filter(in_reply_to=self.object_uri).count(),
}
if save:
self.save()
def calculate_type_data(self, save=True):
"""
Recalculate type_data (used mostly for poll votes)
"""
from activities.models import PostInteraction
if self.local and isinstance(self.type_data, QuestionData):
self.type_data.voter_count = (
self.interactions.filter(
type=PostInteraction.Types.vote,
)
.values("identity")
.distinct()
.count()
)
for option in self.type_data.options:
option.votes = self.interactions.filter(
type=PostInteraction.Types.vote,
value=option.name,
).count()
if save:
self.save()
2022-11-11 22:04:43 -08:00
### ActivityPub (outbound) ###
2022-11-11 21:02:43 -08:00
def to_ap(self) -> dict:
2022-11-11 22:04:43 -08:00
"""
Returns the AP JSON for this object
"""
2023-07-22 09:38:22 -07:00
self.author.ensure_uris()
2022-11-11 22:04:43 -08:00
value = {
2022-12-20 06:20:11 -08:00
"to": [],
"cc": [],
"type": self.type,
2022-11-11 22:04:43 -08:00
"id": self.object_uri,
2022-11-13 17:42:47 -08:00
"published": format_ld_date(self.published),
2022-11-11 22:04:43 -08:00
"attributedTo": self.author.actor_uri,
"content": self.safe_content_remote(),
2022-12-15 17:58:11 -08:00
"sensitive": self.sensitive,
"url": self.absolute_object_uri(),
"tag": [],
2022-12-01 17:46:49 -08:00
"attachment": [],
2022-11-11 22:04:43 -08:00
}
if self.type == Post.Types.question and self.type_data:
value[self.type_data.mode] = [
{
"name": option.name,
"type": option.type,
"replies": {"type": "Collection", "totalItems": option.votes},
}
for option in self.type_data.options
]
value["toot:votersCount"] = self.type_data.voter_count
if self.type_data.end_time:
value["endTime"] = format_ld_date(self.type_data.end_time)
2022-11-11 22:04:43 -08:00
if self.summary:
value["summary"] = self.summary
2022-11-24 14:17:32 -08:00
if self.in_reply_to:
value["inReplyTo"] = self.in_reply_to
2022-11-27 10:09:46 -08:00
if self.edited:
value["updated"] = format_ld_date(self.edited)
2022-12-20 06:20:11 -08:00
# Targeting
if self.visibility == self.Visibilities.public:
value["to"].append("as:Public")
2022-12-20 06:20:11 -08:00
elif self.visibility == self.Visibilities.unlisted:
value["cc"].append("as:Public")
2023-07-22 09:38:22 -07:00
elif (
self.visibility == self.Visibilities.followers and self.author.followers_uri
):
value["to"].append(self.author.followers_uri)
# Mentions
for mention in self.mentions.all():
2022-12-14 23:50:54 -08:00
value["tag"].append(mention.to_ap_tag())
value["cc"].append(mention.actor_uri)
2022-12-15 16:00:00 -08:00
# Hashtags
for hashtag in self.hashtags or []:
value["tag"].append(
{
"href": f"https://{self.author.domain.uri_domain}/tags/{hashtag}/",
"name": f"#{hashtag}",
2022-12-15 17:58:11 -08:00
"type": "Hashtag",
2022-12-15 16:00:00 -08:00
}
)
2022-12-14 23:50:54 -08:00
# Emoji
for emoji in self.emojis.all():
value["tag"].append(emoji.to_ap_tag())
2022-12-01 17:46:49 -08:00
# Attachments
for attachment in self.attachments.all():
value["attachment"].append(attachment.to_ap())
# Remove fields if they're empty
2022-12-20 06:20:11 -08:00
for field in ["to", "cc", "tag", "attachment"]:
2022-12-01 17:46:49 -08:00
if not value[field]:
del value[field]
2022-11-11 22:04:43 -08:00
return value
def to_create_ap(self):
"""
Returns the AP JSON to create this object
"""
object = self.to_ap()
2022-11-11 22:04:43 -08:00
return {
2022-12-20 06:20:11 -08:00
"to": object.get("to", []),
"cc": object.get("cc", []),
2022-11-11 22:04:43 -08:00
"type": "Create",
"id": self.object_uri + "#create",
"actor": self.author.actor_uri,
"object": object,
2022-11-11 22:04:43 -08:00
}
2022-11-27 10:09:46 -08:00
def to_update_ap(self):
"""
Returns the AP JSON to update this object
"""
object = self.to_ap()
return {
2022-12-20 06:20:11 -08:00
"to": object.get("to", []),
2022-11-27 10:09:46 -08:00
"cc": object.get("cc", []),
"type": "Update",
"id": self.object_uri + "#update",
"actor": self.author.actor_uri,
"object": object,
}
def to_delete_ap(self):
"""
Returns the AP JSON to create this object
"""
object = self.to_ap()
return {
2022-12-20 06:20:11 -08:00
"to": object.get("to", []),
"cc": object.get("cc", []),
"type": "Delete",
"id": self.object_uri + "#delete",
"actor": self.author.actor_uri,
"object": object,
}
def get_targets(self) -> Iterable[Identity]:
"""
Returns a list of Identities that need to see posts and their changes
"""
targets = set()
for mention in self.mentions.all():
targets.add(mention)
2023-03-14 14:35:40 -07:00
# Then, if it's not mentions only, also deliver to followers and all hashtag followers
if self.visibility != Post.Visibilities.mentioned:
for follower in self.author.inbound_follows.filter(
2023-01-08 09:36:18 -08:00
state__in=FollowStates.group_active()
).select_related("source"):
targets.add(follower.source)
2023-03-14 14:35:40 -07:00
if self.hashtags:
for follow in HashtagFollow.objects.by_hashtags(
2023-03-14 14:35:40 -07:00
self.hashtags
).prefetch_related("identity"):
targets.add(follow.identity)
# If it's a reply, always include the original author if we know them
reply_post = self.in_reply_to_post()
if reply_post:
targets.add(reply_post.author)
# And if it's a reply to one of our own, we have to re-fan-out to
# the original author's followers
if reply_post.author.local:
for follower in reply_post.author.inbound_follows.filter(
2023-01-08 09:36:18 -08:00
state__in=FollowStates.group_active()
).select_related("source"):
targets.add(follower.source)
2022-11-26 09:09:31 -08:00
# If this is a remote post or local-only, filter to only include
# local identities
if not self.local or self.visibility == Post.Visibilities.local_only:
targets = {target for target in targets if target.local}
# If it's a local post, include the author
2022-11-26 09:09:31 -08:00
if self.local:
targets.add(self.author)
2023-01-15 12:35:45 -08:00
# Fetch the author's full blocks and remove them as targets
blocks = (
self.author.outbound_blocks.active()
.filter(mute=False)
.select_related("target")
)
for block in blocks:
try:
targets.remove(block.target)
except KeyError:
pass
2023-01-08 12:46:40 -08:00
# Now dedupe the targets based on shared inboxes (we only keep one per
# shared inbox)
deduped_targets = set()
shared_inboxes = set()
for target in targets:
if target.local or not target.shared_inbox_uri:
deduped_targets.add(target)
elif target.shared_inbox_uri not in shared_inboxes:
shared_inboxes.add(target.shared_inbox_uri)
deduped_targets.add(target)
else:
# Their shared inbox is already being sent to
pass
return deduped_targets
2022-11-11 22:04:43 -08:00
### ActivityPub (inbound) ###
2022-11-11 21:02:43 -08:00
@classmethod
def by_ap(cls, data, create=False, update=False, fetch_author=False) -> "Post":
2022-11-11 21:02:43 -08:00
"""
Retrieves a Post instance by its ActivityPub JSON object.
Optionally creates one if it's not present.
Raises DoesNotExist if it's not found and create is False,
or it's from a blocked domain.
2022-11-11 21:02:43 -08:00
"""
try:
# Ensure data has the primary fields of all Posts
if (
not isinstance(data["id"], str)
or not isinstance(data["attributedTo"], str)
or not isinstance(data["type"], str)
):
raise TypeError()
# Ensure the domain of the object's actor and ID match to prevent injection
if urlparse(data["id"]).hostname != urlparse(data["attributedTo"]).hostname:
raise ValueError("Object's ID domain is different to its author")
except (TypeError, KeyError) as ex:
raise cls.DoesNotExist(
"Object data is not a recognizable ActivityPub object"
) from ex
2022-11-11 21:02:43 -08:00
# Do we have one with the right ID?
2022-11-12 20:14:21 -08:00
created = False
2022-11-11 21:02:43 -08:00
try:
post: Post = cls.objects.select_related("author__domain").get(
2022-12-14 23:50:54 -08:00
object_uri=data["id"]
)
2022-11-11 21:02:43 -08:00
except cls.DoesNotExist:
if create:
# Resolve the author
author = Identity.by_actor_uri(data["attributedTo"], create=create)
# If the author is not fetched yet, try again later
if author.domain is None:
if fetch_author:
author.fetch_actor()
# perhaps the entire "try again" logic below
# could be replaced with TryAgainLater for
# _all_ fetches, to let it handle pinned posts?
if author.domain is None:
raise TryAgainLater()
else:
raise TryAgainLater()
# If the post is from a blocked domain, stop and drop
2023-07-22 09:54:36 -07:00
if author.domain.recursively_blocked():
raise cls.DoesNotExist("Post is from a blocked domain")
try:
# try again, because fetch_actor() also fetches pinned posts
post = cls.objects.select_related("author__domain").get(
object_uri=data["id"]
)
except cls.DoesNotExist:
# finally, create a stub
try:
post = cls.objects.create(
object_uri=data["id"],
author=author,
content="",
local=False,
type=data["type"],
)
created = True
except IntegrityError as dupe:
# there's still some kind of race condition here
# it's far more rare, but sometimes we fire an
# IntegrityError on activities_post_object_uri_key
# this transaction is now aborted and anything following
# in the caller function will fail in the database.
raise TryAgainLater() from dupe
2022-11-11 21:02:43 -08:00
else:
raise cls.DoesNotExist(f"No post with ID {data['id']}", data)
2022-11-12 20:14:21 -08:00
if update or created:
post.type = data["type"]
if post.type in (cls.Types.article, cls.Types.question):
post.type_data = PostTypeData(__root__=data).__root__
try:
# apparently sometimes posts (Pages?) in the fediverse
# don't have content?!
post.content = get_value_or_map(data, "content", "contentMap")
except KeyError:
post.content = None
# Document types have names, not summaries
post.summary = data.get("summary") or data.get("name")
if not post.content and post.summary:
post.content = post.summary
post.summary = None
2022-12-16 15:38:52 -08:00
post.sensitive = data.get("sensitive", False)
post.url = data.get("url", data["id"])
2022-11-16 22:00:10 -08:00
post.published = parse_ld_date(data.get("published"))
post.edited = parse_ld_date(data.get("updated"))
post.in_reply_to = data.get("inReplyTo")
2022-11-16 05:53:39 -08:00
# Mentions and hashtags
post.hashtags = []
for tag in get_list(data, "tag"):
tag_type = tag["type"].lower()
if tag_type == "mention":
2022-11-16 05:53:39 -08:00
mention_identity = Identity.by_actor_uri(tag["href"], create=True)
post.mentions.add(mention_identity)
elif tag_type in ["_:hashtag", "hashtag"]:
# kbin produces tags with 'tag' instead of 'name'
if "tag" in tag and "name" not in tag:
name = get_value_or_map(tag, "tag", "tagMap")
else:
name = get_value_or_map(tag, "name", "nameMap")
post.hashtags.append(
name.lower().lstrip("#")[: Hashtag.MAXIMUM_LENGTH]
)
elif tag_type in ["toot:emoji", "emoji"]:
2022-12-14 23:50:54 -08:00
emoji = Emoji.by_ap_tag(post.author.domain, tag, create=True)
post.emojis.add(emoji)
elif tag_type == "edition":
# Bookwyrm Edition is similar to hashtags. There should be a link to
# the book in the Note's content and a post attachment of the cover
# image. No special processing should be needed for ingest.
pass
2022-11-16 05:53:39 -08:00
else:
raise ValueError(f"Unknown tag type {tag['type']}")
# Visibility and to
2022-12-20 06:20:11 -08:00
# (a post is public if it's to:public, otherwise it's unlisted if
# it's cc:public, otherwise it's more limited)
to = [x.lower() for x in get_list(data, "to")]
cc = [x.lower() for x in get_list(data, "cc")]
post.visibility = Post.Visibilities.mentioned
if "public" in to or "as:public" in to:
post.visibility = Post.Visibilities.public
elif "public" in cc or "as:public" in cc:
post.visibility = Post.Visibilities.unlisted
2023-07-22 09:38:22 -07:00
elif post.author.followers_uri in to:
post.visibility = Post.Visibilities.followers
2022-11-16 22:00:10 -08:00
# Attachments
# These have no IDs, so we have to wipe them each time
post.attachments.all().delete()
for attachment in get_list(data, "attachment"):
2022-12-18 10:00:21 -08:00
if "focalPoint" in attachment:
try:
focal_x, focal_y = attachment["focalPoint"]
except (ValueError, TypeError):
focal_x, focal_y = None, None
2022-11-16 22:00:10 -08:00
else:
focal_x, focal_y = None, None
mimetype = attachment.get("mediaType")
if not mimetype or not isinstance(mimetype, str):
if "url" not in attachment:
raise ActivityPubFormatError("No URL present on attachment")
mimetype, _ = mimetypes.guess_type(attachment["url"])
if not mimetype:
mimetype = "application/octet-stream"
2022-11-16 22:00:10 -08:00
post.attachments.create(
remote_url=attachment["url"],
mimetype=mimetype,
2022-11-16 22:00:10 -08:00
name=attachment.get("name"),
width=attachment.get("width"),
height=attachment.get("height"),
2022-12-18 10:00:21 -08:00
blurhash=attachment.get("blurhash"),
2022-11-16 22:00:10 -08:00
focal_x=focal_x,
focal_y=focal_y,
)
# Calculate stats in case we have existing replies
post.calculate_stats(save=False)
2022-11-12 20:14:21 -08:00
post.save()
# Potentially schedule a fetch of the reply parent, and recalculate
# its stats if it's here already.
if post.in_reply_to:
try:
parent = cls.by_object_uri(post.in_reply_to)
except cls.DoesNotExist:
try:
cls.ensure_object_uri(post.in_reply_to, reason=post.object_uri)
except ValueError:
capture_message(
f"Cannot fetch ancestor of Post={post.pk}, ancestor_uri={post.in_reply_to}"
)
else:
parent.calculate_stats()
2022-11-12 20:14:21 -08:00
return post
2022-11-11 21:02:43 -08:00
2022-11-13 17:42:47 -08:00
@classmethod
def by_object_uri(cls, object_uri, fetch=False) -> "Post":
2022-11-13 17:42:47 -08:00
"""
Gets the post by URI - either looking up locally, or fetching
from the other end if it's not here.
"""
try:
return cls.objects.get(object_uri=object_uri)
except cls.DoesNotExist:
if fetch:
2022-12-04 20:13:33 -08:00
try:
response = SystemActor().signed_request(
2022-12-04 20:13:33 -08:00
method="get", uri=object_uri
)
except (httpx.HTTPError, ssl.SSLCertVerificationError):
2022-12-04 20:13:33 -08:00
raise cls.DoesNotExist(f"Could not fetch {object_uri}")
if response.status_code in [404, 410]:
raise cls.DoesNotExist(f"No post at {object_uri}")
if response.status_code >= 500:
raise cls.DoesNotExist(f"Server error fetching {object_uri}")
if response.status_code >= 400:
raise cls.DoesNotExist(
f"Error fetching post from {object_uri}: {response.status_code}",
{response.content},
2022-11-13 17:42:47 -08:00
)
try:
post = cls.by_ap(
canonicalise(response.json(), include_security=True),
create=True,
update=True,
fetch_author=True,
)
except (json.JSONDecodeError, ValueError, JsonLdError) as err:
raise cls.DoesNotExist(
f"Invalid ld+json response for {object_uri}"
) from err
2022-12-04 20:22:20 -08:00
# We may need to fetch the author too
if post.author.state == IdentityStates.outdated:
post.author.fetch_actor()
2022-12-04 20:22:20 -08:00
return post
2022-12-04 20:13:33 -08:00
else:
raise cls.DoesNotExist(f"Cannot find Post with URI {object_uri}")
2022-11-13 17:42:47 -08:00
@classmethod
def ensure_object_uri(cls, object_uri: str, reason: str | None = None):
"""
Sees if the post is in our local set, and if not, schedules a fetch
for it (in the background)
"""
if not object_uri or "://" not in object_uri:
raise ValueError("URI missing or invalid")
try:
cls.by_object_uri(object_uri)
except cls.DoesNotExist:
InboxMessage.create_internal(
{
"type": "FetchPost",
"object": object_uri,
"reason": reason,
}
)
2022-11-11 21:02:43 -08:00
@classmethod
def handle_create_ap(cls, data):
"""
Handles an incoming create request
"""
with transaction.atomic():
# Ensure the Create actor is the Post's attributedTo
if data["actor"] != data["object"]["attributedTo"]:
raise ValueError("Create actor does not match its Post object", data)
# Create it, stator will fan it out locally
cls.by_ap(data["object"], create=True, update=True)
2022-11-13 18:03:43 -08:00
2022-11-16 21:23:32 -08:00
@classmethod
def handle_update_ap(cls, data):
"""
Handles an incoming update request
"""
with transaction.atomic():
# Ensure the Create actor is the Post's attributedTo
if data["actor"] != data["object"]["attributedTo"]:
raise ValueError("Create actor does not match its Post object", data)
# Find it and update it
try:
cls.by_ap(data["object"], create=False, update=True)
except cls.DoesNotExist:
# We don't have a copy - assume we got a delete first and ignore.
pass
2022-11-16 21:23:32 -08:00
@classmethod
def handle_delete_ap(cls, data):
"""
2022-11-25 18:33:46 -08:00
Handles an incoming delete request
"""
with transaction.atomic():
# Is this an embedded object or plain ID?
if isinstance(data["object"], str):
object_uri = data["object"]
else:
object_uri = data["object"]["id"]
# Find our post by ID if we have one
try:
post = cls.by_object_uri(object_uri)
except cls.DoesNotExist:
# It's already been deleted
return
# Ensure the actor on the request authored the post
if not post.author.actor_uri == data["actor"]:
raise ValueError("Actor on delete does not match object")
post.delete()
@classmethod
def handle_fetch_internal(cls, data):
"""
Handles an internal fetch-request inbox message
"""
try:
uri = data["object"]
if "://" in uri:
cls.by_object_uri(uri, fetch=True)
except (cls.DoesNotExist, KeyError):
pass
### OpenGraph API ###
def to_opengraph_dict(self) -> dict:
return {
"og:title": f"{self.author.name} (@{self.author.handle})",
"og:type": "article",
"og:published_time": (self.published or self.created).isoformat(),
"og:modified_time": (
self.edited or self.published or self.created
).isoformat(),
"og:description": (self.summary or self.safe_content_local()),
"og:image:url": self.author.local_icon_url().absolute,
"og:image:height": 85,
"og:image:width": 85,
}
2022-12-10 23:25:48 -08:00
### Mastodon API ###
2023-03-11 10:17:20 -08:00
def to_mastodon_json(self, interactions=None, bookmarks=None, identity=None):
2022-12-10 23:25:48 -08:00
reply_parent = None
if self.in_reply_to:
# Load the PK and author.id explicitly to prevent a SELECT on the entire author Identity
reply_parent = (
Post.objects.filter(object_uri=self.in_reply_to)
.only("pk", "author_id")
.first()
)
2022-12-11 23:54:51 -08:00
visibility_mapping = {
self.Visibilities.public: "public",
self.Visibilities.unlisted: "unlisted",
self.Visibilities.followers: "private",
self.Visibilities.mentioned: "direct",
self.Visibilities.local_only: "public",
}
2022-12-11 11:37:28 -08:00
value = {
2022-12-10 23:25:48 -08:00
"id": self.pk,
"uri": self.object_uri,
"created_at": format_ld_date(self.published),
"account": self.author.to_mastodon_json(include_counts=False),
2022-12-10 23:25:48 -08:00
"content": self.safe_content_remote(),
2022-12-11 23:54:51 -08:00
"visibility": visibility_mapping[self.visibility],
2022-12-10 23:25:48 -08:00
"sensitive": self.sensitive,
"spoiler_text": self.summary or "",
"media_attachments": [
attachment.to_mastodon_json() for attachment in self.attachments.all()
],
"mentions": [
mention.to_mastodon_mention_json() for mention in self.mentions.all()
2022-12-10 23:25:48 -08:00
],
"tags": (
[
{
"name": tag,
"url": f"https://{self.author.domain.uri_domain}/tags/{tag}/",
}
for tag in self.hashtags
]
2022-12-10 23:25:48 -08:00
if self.hashtags
else []
),
# Filter in the list comp rather than query because the common case is no emoji in the resultset
# When filter is on emojis like `emojis.usable()` it causes a query that is not cached by prefetch_related
"emojis": [
emoji.to_mastodon_json()
for emoji in self.emojis.all()
if emoji.is_usable
],
"reblogs_count": self.stats_with_defaults["boosts"],
"favourites_count": self.stats_with_defaults["likes"],
"replies_count": self.stats_with_defaults["replies"],
2022-12-10 23:25:48 -08:00
"url": self.absolute_object_uri(),
"in_reply_to_id": reply_parent.pk if reply_parent else None,
"in_reply_to_account_id": (
reply_parent.author_id if reply_parent else None
),
2022-12-10 23:25:48 -08:00
"reblog": None,
"poll": self.type_data.to_mastodon_json(self, identity)
if isinstance(self.type_data, QuestionData)
else None,
2022-12-10 23:25:48 -08:00
"card": None,
"language": None,
"text": self.safe_content_remote(),
2022-12-10 23:25:48 -08:00
"edited_at": format_ld_date(self.edited) if self.edited else None,
}
2022-12-11 11:37:28 -08:00
if interactions:
value["favourited"] = self.pk in interactions.get("like", [])
value["reblogged"] = self.pk in interactions.get("boost", [])
2023-05-13 09:01:27 -07:00
value["pinned"] = self.pk in interactions.get("pin", [])
2023-03-11 10:17:20 -08:00
if bookmarks:
value["bookmarked"] = self.pk in bookmarks
2022-12-11 11:37:28 -08:00
return value