2022-11-21 18:59:27 -08:00
|
|
|
import re
|
2022-11-13 15:14:38 -08:00
|
|
|
from typing import Dict, Optional
|
2022-11-11 22:04:43 -08:00
|
|
|
|
2022-11-13 17:42:47 -08:00
|
|
|
import httpx
|
2022-11-11 21:02:43 -08:00
|
|
|
import urlman
|
2022-11-15 17:30:30 -08:00
|
|
|
from django.db import models, transaction
|
2022-11-21 20:18:13 -08:00
|
|
|
from django.template.defaultfilters import linebreaks_filter
|
2022-11-11 22:04:43 -08:00
|
|
|
from django.utils import timezone
|
2022-11-21 18:59:27 -08:00
|
|
|
from django.utils.safestring import mark_safe
|
2022-11-11 21:02:43 -08:00
|
|
|
|
2022-11-11 22:04:43 -08:00
|
|
|
from activities.models.fan_out import FanOut
|
2022-11-11 21:02:43 -08:00
|
|
|
from activities.models.timeline_event import TimelineEvent
|
2022-11-21 20:18:13 -08:00
|
|
|
from core.html import sanitize_post, strip_html
|
2022-11-16 05:53:39 -08:00
|
|
|
from core.ld import canonicalise, format_ld_date, get_list, parse_ld_date
|
2022-11-11 21:02:43 -08:00
|
|
|
from stator.models import State, StateField, StateGraph, StatorModel
|
|
|
|
from users.models.follow import Follow
|
|
|
|
from users.models.identity import Identity
|
|
|
|
|
|
|
|
|
|
|
|
class PostStates(StateGraph):
|
|
|
|
new = State(try_interval=300)
|
|
|
|
fanned_out = State()
|
|
|
|
|
|
|
|
new.transitions_to(fanned_out)
|
|
|
|
|
|
|
|
@classmethod
|
|
|
|
async def handle_new(cls, instance: "Post"):
|
|
|
|
"""
|
|
|
|
Creates all needed fan-out objects for a new Post.
|
|
|
|
"""
|
2022-11-16 05:53:39 -08:00
|
|
|
post = await instance.afetch_full()
|
|
|
|
# Non-local posts should not be here
|
2022-11-17 07:21:42 -08:00
|
|
|
# TODO: This seems to keep happening. Work out how?
|
2022-11-16 05:53:39 -08:00
|
|
|
if not post.local:
|
2022-11-17 07:21:42 -08:00
|
|
|
print(f"Trying to run handle_new on a non-local post {post.pk}!")
|
|
|
|
return cls.fanned_out
|
2022-11-16 05:53:39 -08:00
|
|
|
# Build list of targets - mentions always included
|
|
|
|
targets = set()
|
|
|
|
async for mention in post.mentions.all():
|
|
|
|
targets.add(mention)
|
|
|
|
# Then, if it's not mentions only, also deliver to followers
|
|
|
|
if post.visibility != Post.Visibilities.mentioned:
|
|
|
|
async for follower in post.author.inbound_follows.select_related("source"):
|
|
|
|
targets.add(follower.source)
|
|
|
|
# Fan out to each one
|
|
|
|
for follow in targets:
|
|
|
|
await FanOut.objects.acreate(
|
|
|
|
identity=follow,
|
|
|
|
type=FanOut.Types.post,
|
|
|
|
subject_post=post,
|
|
|
|
)
|
|
|
|
# And one for themselves if they're local
|
2022-11-22 17:51:01 -08:00
|
|
|
# (most views will do this at time of post, but it's idempotent)
|
2022-11-16 05:53:39 -08:00
|
|
|
if post.author.local:
|
|
|
|
await FanOut.objects.acreate(
|
|
|
|
identity_id=post.author_id,
|
|
|
|
type=FanOut.Types.post,
|
|
|
|
subject_post=post,
|
|
|
|
)
|
2022-11-11 22:04:43 -08:00
|
|
|
return cls.fanned_out
|
2022-11-11 21:02:43 -08:00
|
|
|
|
|
|
|
|
|
|
|
class Post(StatorModel):
|
|
|
|
"""
|
|
|
|
A post (status, toot) that is either local or remote.
|
|
|
|
"""
|
|
|
|
|
|
|
|
class Visibilities(models.IntegerChoices):
|
|
|
|
public = 0
|
|
|
|
unlisted = 1
|
|
|
|
followers = 2
|
|
|
|
mentioned = 3
|
|
|
|
|
|
|
|
# The author (attributedTo) of the post
|
|
|
|
author = models.ForeignKey(
|
|
|
|
"users.Identity",
|
2022-11-20 11:59:06 -08:00
|
|
|
on_delete=models.CASCADE,
|
2022-11-11 21:02:43 -08:00
|
|
|
related_name="posts",
|
|
|
|
)
|
|
|
|
|
|
|
|
# The state the post is in
|
|
|
|
state = StateField(PostStates)
|
|
|
|
|
|
|
|
# If it is our post or not
|
|
|
|
local = models.BooleanField()
|
|
|
|
|
|
|
|
# The canonical object ID
|
2022-11-12 20:14:21 -08:00
|
|
|
object_uri = models.CharField(max_length=500, blank=True, null=True, unique=True)
|
2022-11-11 21:02:43 -08:00
|
|
|
|
|
|
|
# Who should be able to see this Post
|
|
|
|
visibility = models.IntegerField(
|
|
|
|
choices=Visibilities.choices,
|
|
|
|
default=Visibilities.public,
|
|
|
|
)
|
|
|
|
|
|
|
|
# The main (HTML) content
|
|
|
|
content = models.TextField()
|
|
|
|
|
|
|
|
# If the contents of the post are sensitive, and the summary (content
|
|
|
|
# warning) to show if it is
|
|
|
|
sensitive = models.BooleanField(default=False)
|
|
|
|
summary = models.TextField(blank=True, null=True)
|
|
|
|
|
|
|
|
# The public, web URL of this Post on the original server
|
|
|
|
url = models.CharField(max_length=500, blank=True, null=True)
|
|
|
|
|
|
|
|
# The Post it is replying to as an AP ID URI
|
|
|
|
# (as otherwise we'd have to pull entire threads to use IDs)
|
|
|
|
in_reply_to = models.CharField(max_length=500, blank=True, null=True)
|
|
|
|
|
|
|
|
# The identities the post is directly to (who can see it if not public)
|
|
|
|
to = models.ManyToManyField(
|
|
|
|
"users.Identity",
|
|
|
|
related_name="posts_to",
|
|
|
|
blank=True,
|
|
|
|
)
|
|
|
|
|
|
|
|
# The identities mentioned in the post
|
|
|
|
mentions = models.ManyToManyField(
|
|
|
|
"users.Identity",
|
|
|
|
related_name="posts_mentioning",
|
|
|
|
blank=True,
|
|
|
|
)
|
|
|
|
|
2022-11-16 05:53:39 -08:00
|
|
|
# Hashtags in the post
|
2022-11-16 20:42:25 -08:00
|
|
|
hashtags = models.JSONField(blank=True, null=True)
|
2022-11-16 05:53:39 -08:00
|
|
|
|
2022-11-11 22:04:43 -08:00
|
|
|
# When the post was originally created (as opposed to when we received it)
|
2022-11-13 17:42:47 -08:00
|
|
|
published = models.DateTimeField(default=timezone.now)
|
2022-11-11 22:04:43 -08:00
|
|
|
|
2022-11-16 21:23:32 -08:00
|
|
|
# If the post has been edited after initial publication
|
|
|
|
edited = models.DateTimeField(blank=True, null=True)
|
|
|
|
|
2022-11-11 21:02:43 -08:00
|
|
|
created = models.DateTimeField(auto_now_add=True)
|
|
|
|
updated = models.DateTimeField(auto_now=True)
|
|
|
|
|
|
|
|
class urls(urlman.Urls):
|
2022-11-11 22:04:43 -08:00
|
|
|
view = "{self.author.urls.view}posts/{self.id}/"
|
2022-11-15 17:30:30 -08:00
|
|
|
object_uri = "{self.author.actor_uri}posts/{self.id}/"
|
|
|
|
action_like = "{view}like/"
|
|
|
|
action_unlike = "{view}unlike/"
|
|
|
|
action_boost = "{view}boost/"
|
|
|
|
action_unboost = "{view}unboost/"
|
2022-11-11 22:04:43 -08:00
|
|
|
|
|
|
|
def get_scheme(self, url):
|
|
|
|
return "https"
|
|
|
|
|
|
|
|
def get_hostname(self, url):
|
|
|
|
return self.instance.author.domain.uri_domain
|
2022-11-11 21:02:43 -08:00
|
|
|
|
|
|
|
def __str__(self):
|
|
|
|
return f"{self.author} #{self.id}"
|
|
|
|
|
2022-11-16 22:00:10 -08:00
|
|
|
def get_absolute_url(self):
|
|
|
|
return self.urls.view
|
|
|
|
|
2022-11-21 20:18:13 -08:00
|
|
|
def absolute_object_uri(self):
|
|
|
|
"""
|
|
|
|
Returns an object URI that is always absolute, for sending out to
|
|
|
|
other servers.
|
|
|
|
"""
|
|
|
|
if self.local:
|
|
|
|
return self.author.absolute_profile_uri() + f"posts/{self.id}/"
|
|
|
|
else:
|
|
|
|
return self.object_uri
|
|
|
|
|
2022-11-21 18:59:27 -08:00
|
|
|
### Content cleanup and extraction ###
|
|
|
|
|
|
|
|
mention_regex = re.compile(
|
2022-11-21 21:17:22 -08:00
|
|
|
r"(^|[^\w\d\-_])@([\w\d\-_]+(?:@[\w\d\-_]+\.[\w\d\-_\.]+)?)"
|
2022-11-21 18:59:27 -08:00
|
|
|
)
|
|
|
|
|
2022-11-21 20:18:13 -08:00
|
|
|
def linkify_mentions(self, content, local=False):
|
2022-11-21 18:59:27 -08:00
|
|
|
"""
|
2022-11-21 20:18:13 -08:00
|
|
|
Links mentions _in the context of the post_ - as in, using the mentions
|
|
|
|
property as the only source (as we might be doing this without other
|
|
|
|
DB access allowed)
|
2022-11-21 18:59:27 -08:00
|
|
|
"""
|
|
|
|
|
2022-11-21 20:18:13 -08:00
|
|
|
possible_matches = {}
|
|
|
|
for mention in self.mentions.all():
|
|
|
|
if local:
|
|
|
|
url = str(mention.urls.view)
|
|
|
|
else:
|
|
|
|
url = mention.absolute_profile_uri()
|
|
|
|
possible_matches[mention.username] = url
|
|
|
|
possible_matches[f"{mention.username}@{mention.domain_id}"] = url
|
|
|
|
|
2022-11-21 18:59:27 -08:00
|
|
|
def replacer(match):
|
|
|
|
precursor = match.group(1)
|
|
|
|
handle = match.group(2)
|
2022-11-21 20:18:13 -08:00
|
|
|
if handle in possible_matches:
|
|
|
|
return f'{precursor}<a href="{possible_matches[handle]}">@{handle}</a>'
|
2022-11-21 18:59:27 -08:00
|
|
|
else:
|
|
|
|
return match.group()
|
|
|
|
|
|
|
|
return mark_safe(self.mention_regex.sub(replacer, content))
|
|
|
|
|
2022-11-21 20:18:13 -08:00
|
|
|
def safe_content_local(self):
|
|
|
|
"""
|
|
|
|
Returns the content formatted for local display
|
|
|
|
"""
|
|
|
|
return self.linkify_mentions(sanitize_post(self.content), local=True)
|
|
|
|
|
|
|
|
def safe_content_remote(self):
|
|
|
|
"""
|
|
|
|
Returns the content formatted for remote consumption
|
|
|
|
"""
|
2022-11-21 18:59:27 -08:00
|
|
|
return self.linkify_mentions(sanitize_post(self.content))
|
2022-11-11 21:02:43 -08:00
|
|
|
|
2022-11-11 22:04:43 -08:00
|
|
|
### Async helpers ###
|
|
|
|
|
|
|
|
async def afetch_full(self):
|
|
|
|
"""
|
|
|
|
Returns a version of the object with all relations pre-loaded
|
|
|
|
"""
|
2022-11-21 20:18:13 -08:00
|
|
|
return (
|
|
|
|
await Post.objects.select_related("author", "author__domain")
|
|
|
|
.prefetch_related("mentions", "mentions__domain")
|
|
|
|
.aget(pk=self.pk)
|
2022-11-11 22:04:43 -08:00
|
|
|
)
|
|
|
|
|
2022-11-11 21:02:43 -08:00
|
|
|
### Local creation ###
|
|
|
|
|
|
|
|
@classmethod
|
2022-11-13 15:14:38 -08:00
|
|
|
def create_local(
|
2022-11-16 05:53:39 -08:00
|
|
|
cls,
|
|
|
|
author: Identity,
|
|
|
|
content: str,
|
|
|
|
summary: Optional[str] = None,
|
|
|
|
visibility: int = Visibilities.public,
|
2022-11-13 15:14:38 -08:00
|
|
|
) -> "Post":
|
2022-11-15 17:30:30 -08:00
|
|
|
with transaction.atomic():
|
2022-11-21 20:18:13 -08:00
|
|
|
# Find mentions in this post
|
|
|
|
mention_hits = cls.mention_regex.findall(content)
|
|
|
|
mentions = set()
|
|
|
|
for precursor, handle in mention_hits:
|
|
|
|
if "@" in handle:
|
|
|
|
username, domain = handle.split("@", 1)
|
|
|
|
else:
|
|
|
|
username = handle
|
|
|
|
domain = author.domain_id
|
|
|
|
identity = Identity.by_username_and_domain(
|
|
|
|
username=username,
|
|
|
|
domain=domain,
|
|
|
|
fetch=True,
|
|
|
|
)
|
|
|
|
if identity is not None:
|
|
|
|
mentions.add(identity)
|
|
|
|
# Strip all HTML and apply linebreaks filter
|
|
|
|
content = linebreaks_filter(strip_html(content))
|
|
|
|
# Make the Post object
|
2022-11-15 17:30:30 -08:00
|
|
|
post = cls.objects.create(
|
|
|
|
author=author,
|
|
|
|
content=content,
|
|
|
|
summary=summary or None,
|
|
|
|
sensitive=bool(summary),
|
|
|
|
local=True,
|
2022-11-16 05:53:39 -08:00
|
|
|
visibility=visibility,
|
2022-11-15 17:30:30 -08:00
|
|
|
)
|
|
|
|
post.object_uri = post.urls.object_uri
|
2022-11-21 20:18:13 -08:00
|
|
|
post.url = post.absolute_object_uri()
|
|
|
|
post.mentions.set(mentions)
|
2022-11-15 17:30:30 -08:00
|
|
|
post.save()
|
2022-11-11 21:02:43 -08:00
|
|
|
return post
|
|
|
|
|
2022-11-11 22:04:43 -08:00
|
|
|
### ActivityPub (outbound) ###
|
2022-11-11 21:02:43 -08:00
|
|
|
|
2022-11-11 22:04:43 -08:00
|
|
|
def to_ap(self) -> Dict:
|
|
|
|
"""
|
|
|
|
Returns the AP JSON for this object
|
|
|
|
"""
|
|
|
|
value = {
|
2022-11-21 20:18:13 -08:00
|
|
|
"to": "as:Public",
|
|
|
|
"cc": [],
|
2022-11-11 22:04:43 -08:00
|
|
|
"type": "Note",
|
|
|
|
"id": self.object_uri,
|
2022-11-13 17:42:47 -08:00
|
|
|
"published": format_ld_date(self.published),
|
2022-11-11 22:04:43 -08:00
|
|
|
"attributedTo": self.author.actor_uri,
|
2022-11-21 20:18:13 -08:00
|
|
|
"content": self.safe_content_remote(),
|
2022-11-11 22:04:43 -08:00
|
|
|
"as:sensitive": self.sensitive,
|
2022-11-21 20:18:13 -08:00
|
|
|
"url": self.absolute_object_uri(),
|
|
|
|
"tag": [],
|
2022-11-11 22:04:43 -08:00
|
|
|
}
|
|
|
|
if self.summary:
|
|
|
|
value["summary"] = self.summary
|
2022-11-21 20:18:13 -08:00
|
|
|
# Mentions
|
|
|
|
for mention in self.mentions.all():
|
|
|
|
value["tag"].append(
|
|
|
|
{
|
|
|
|
"href": mention.actor_uri,
|
|
|
|
"name": "@" + mention.handle,
|
|
|
|
"type": "Mention",
|
|
|
|
}
|
|
|
|
)
|
|
|
|
value["cc"].append(mention.actor_uri)
|
|
|
|
# Remove tag and cc if they're empty
|
|
|
|
if not value["cc"]:
|
|
|
|
del value["cc"]
|
|
|
|
if not value["tag"]:
|
|
|
|
del value["tag"]
|
2022-11-11 22:04:43 -08:00
|
|
|
return value
|
|
|
|
|
|
|
|
def to_create_ap(self):
|
|
|
|
"""
|
|
|
|
Returns the AP JSON to create this object
|
|
|
|
"""
|
2022-11-21 20:18:13 -08:00
|
|
|
object = self.to_ap()
|
2022-11-11 22:04:43 -08:00
|
|
|
return {
|
2022-11-21 20:18:13 -08:00
|
|
|
"to": object["to"],
|
|
|
|
"cc": object.get("cc", []),
|
2022-11-11 22:04:43 -08:00
|
|
|
"type": "Create",
|
|
|
|
"id": self.object_uri + "#create",
|
|
|
|
"actor": self.author.actor_uri,
|
2022-11-21 20:18:13 -08:00
|
|
|
"object": object,
|
2022-11-11 22:04:43 -08:00
|
|
|
}
|
|
|
|
|
|
|
|
### ActivityPub (inbound) ###
|
2022-11-11 21:02:43 -08:00
|
|
|
|
|
|
|
@classmethod
|
2022-11-12 20:14:21 -08:00
|
|
|
def by_ap(cls, data, create=False, update=False) -> "Post":
|
2022-11-11 21:02:43 -08:00
|
|
|
"""
|
|
|
|
Retrieves a Post instance by its ActivityPub JSON object.
|
|
|
|
|
|
|
|
Optionally creates one if it's not present.
|
|
|
|
Raises KeyError if it's not found and create is False.
|
|
|
|
"""
|
|
|
|
# Do we have one with the right ID?
|
2022-11-12 20:14:21 -08:00
|
|
|
created = False
|
2022-11-11 21:02:43 -08:00
|
|
|
try:
|
2022-11-12 20:14:21 -08:00
|
|
|
post = cls.objects.get(object_uri=data["id"])
|
2022-11-11 21:02:43 -08:00
|
|
|
except cls.DoesNotExist:
|
|
|
|
if create:
|
|
|
|
# Resolve the author
|
|
|
|
author = Identity.by_actor_uri(data["attributedTo"], create=create)
|
2022-11-12 20:14:21 -08:00
|
|
|
post = cls.objects.create(
|
|
|
|
object_uri=data["id"],
|
2022-11-11 21:02:43 -08:00
|
|
|
author=author,
|
2022-11-21 19:00:35 -08:00
|
|
|
content=data["content"],
|
2022-11-11 21:02:43 -08:00
|
|
|
local=False,
|
|
|
|
)
|
2022-11-12 20:14:21 -08:00
|
|
|
created = True
|
2022-11-11 21:02:43 -08:00
|
|
|
else:
|
|
|
|
raise KeyError(f"No post with ID {data['id']}", data)
|
2022-11-12 20:14:21 -08:00
|
|
|
if update or created:
|
2022-11-21 19:00:35 -08:00
|
|
|
post.content = data["content"]
|
2022-11-16 22:00:10 -08:00
|
|
|
post.summary = data.get("summary")
|
2022-11-12 20:14:21 -08:00
|
|
|
post.sensitive = data.get("as:sensitive", False)
|
2022-11-16 22:00:10 -08:00
|
|
|
post.url = data.get("url")
|
|
|
|
post.published = parse_ld_date(data.get("published"))
|
|
|
|
post.edited = parse_ld_date(data.get("updated"))
|
|
|
|
post.in_reply_to = data.get("inReplyTo")
|
2022-11-16 05:53:39 -08:00
|
|
|
# Mentions and hashtags
|
|
|
|
post.hashtags = []
|
|
|
|
for tag in get_list(data, "tag"):
|
|
|
|
if tag["type"].lower() == "mention":
|
|
|
|
mention_identity = Identity.by_actor_uri(tag["href"], create=True)
|
|
|
|
post.mentions.add(mention_identity)
|
|
|
|
elif tag["type"].lower() == "as:hashtag":
|
|
|
|
post.hashtags.append(tag["name"].lstrip("#"))
|
2022-11-16 21:23:32 -08:00
|
|
|
elif tag["type"].lower() == "http://joinmastodon.org/ns#emoji":
|
|
|
|
# TODO: Handle incoming emoji
|
|
|
|
pass
|
2022-11-16 05:53:39 -08:00
|
|
|
else:
|
|
|
|
raise ValueError(f"Unknown tag type {tag['type']}")
|
|
|
|
# Visibility and to
|
|
|
|
# (a post is public if it's ever to/cc as:Public, otherwise we
|
|
|
|
# regard it as unlisted for now)
|
|
|
|
targets = get_list(data, "to") + get_list(data, "cc")
|
|
|
|
post.visibility = Post.Visibilities.unlisted
|
|
|
|
for target in targets:
|
|
|
|
if target.lower() == "as:public":
|
|
|
|
post.visibility = Post.Visibilities.public
|
2022-11-16 22:00:10 -08:00
|
|
|
# Attachments
|
|
|
|
# These have no IDs, so we have to wipe them each time
|
|
|
|
post.attachments.all().delete()
|
|
|
|
for attachment in get_list(data, "attachment"):
|
|
|
|
if "http://joinmastodon.org/ns#focalPoint" in attachment:
|
|
|
|
focal_x, focal_y = attachment[
|
|
|
|
"http://joinmastodon.org/ns#focalPoint"
|
|
|
|
]["@list"]
|
|
|
|
else:
|
|
|
|
focal_x, focal_y = None, None
|
|
|
|
post.attachments.create(
|
|
|
|
remote_url=attachment["url"],
|
|
|
|
mimetype=attachment["mediaType"],
|
|
|
|
name=attachment.get("name"),
|
|
|
|
width=attachment.get("width"),
|
|
|
|
height=attachment.get("height"),
|
|
|
|
blurhash=attachment.get("http://joinmastodon.org/ns#blurhash"),
|
|
|
|
focal_x=focal_x,
|
|
|
|
focal_y=focal_y,
|
|
|
|
)
|
2022-11-12 20:14:21 -08:00
|
|
|
post.save()
|
|
|
|
return post
|
2022-11-11 21:02:43 -08:00
|
|
|
|
2022-11-13 17:42:47 -08:00
|
|
|
@classmethod
|
|
|
|
def by_object_uri(cls, object_uri, fetch=False):
|
|
|
|
"""
|
|
|
|
Gets the post by URI - either looking up locally, or fetching
|
|
|
|
from the other end if it's not here.
|
|
|
|
"""
|
|
|
|
try:
|
|
|
|
return cls.objects.get(object_uri=object_uri)
|
|
|
|
except cls.DoesNotExist:
|
|
|
|
if fetch:
|
|
|
|
# Go grab the data from the URI
|
|
|
|
response = httpx.get(
|
|
|
|
object_uri,
|
|
|
|
headers={"Accept": "application/json"},
|
|
|
|
follow_redirects=True,
|
|
|
|
)
|
|
|
|
if 200 <= response.status_code < 300:
|
|
|
|
return cls.by_ap(
|
|
|
|
canonicalise(response.json(), include_security=True),
|
|
|
|
create=True,
|
|
|
|
update=True,
|
|
|
|
)
|
2022-11-15 17:30:30 -08:00
|
|
|
raise cls.DoesNotExist(f"Cannot find Post with URI {object_uri}")
|
2022-11-13 17:42:47 -08:00
|
|
|
|
2022-11-11 21:02:43 -08:00
|
|
|
@classmethod
|
|
|
|
def handle_create_ap(cls, data):
|
|
|
|
"""
|
|
|
|
Handles an incoming create request
|
|
|
|
"""
|
2022-11-16 20:42:25 -08:00
|
|
|
with transaction.atomic():
|
|
|
|
# Ensure the Create actor is the Post's attributedTo
|
|
|
|
if data["actor"] != data["object"]["attributedTo"]:
|
|
|
|
raise ValueError("Create actor does not match its Post object", data)
|
|
|
|
# Create it
|
|
|
|
post = cls.by_ap(data["object"], create=True, update=True)
|
2022-11-16 22:00:10 -08:00
|
|
|
# Make timeline events for followers if it's not a reply
|
|
|
|
# TODO: _do_ show replies to people we follow somehow
|
|
|
|
if not post.in_reply_to:
|
|
|
|
for follow in Follow.objects.filter(
|
|
|
|
target=post.author, source__local=True
|
|
|
|
):
|
|
|
|
TimelineEvent.add_post(follow.source, post)
|
2022-11-16 20:42:25 -08:00
|
|
|
# Make timeline events for mentions if they're local
|
|
|
|
for mention in post.mentions.all():
|
|
|
|
if mention.local:
|
|
|
|
TimelineEvent.add_mentioned(mention, post)
|
|
|
|
# Force it into fanned_out as it's not ours
|
|
|
|
post.transition_perform(PostStates.fanned_out)
|
2022-11-13 18:03:43 -08:00
|
|
|
|
2022-11-16 21:23:32 -08:00
|
|
|
@classmethod
|
|
|
|
def handle_update_ap(cls, data):
|
|
|
|
"""
|
|
|
|
Handles an incoming update request
|
|
|
|
"""
|
|
|
|
with transaction.atomic():
|
|
|
|
# Ensure the Create actor is the Post's attributedTo
|
|
|
|
if data["actor"] != data["object"]["attributedTo"]:
|
|
|
|
raise ValueError("Create actor does not match its Post object", data)
|
|
|
|
# Find it and update it
|
|
|
|
cls.by_ap(data["object"], create=False, update=True)
|
|
|
|
|
2022-11-15 17:30:30 -08:00
|
|
|
@classmethod
|
|
|
|
def handle_delete_ap(cls, data):
|
|
|
|
"""
|
|
|
|
Handles an incoming create request
|
|
|
|
"""
|
2022-11-16 20:42:25 -08:00
|
|
|
with transaction.atomic():
|
|
|
|
# Find our post by ID if we have one
|
|
|
|
try:
|
|
|
|
post = cls.by_object_uri(data["object"]["id"])
|
|
|
|
except cls.DoesNotExist:
|
|
|
|
# It's already been deleted
|
|
|
|
return
|
|
|
|
# Ensure the actor on the request authored the post
|
|
|
|
if not post.author.actor_uri == data["actor"]:
|
|
|
|
raise ValueError("Actor on delete does not match object")
|
|
|
|
post.delete()
|
2022-11-15 17:30:30 -08:00
|
|
|
|
2022-11-13 18:03:43 -08:00
|
|
|
def debug_fetch(self):
|
|
|
|
"""
|
|
|
|
Fetches the Post from its original URL again and updates us with it
|
|
|
|
"""
|
|
|
|
response = httpx.get(
|
|
|
|
self.object_uri,
|
|
|
|
headers={"Accept": "application/json"},
|
|
|
|
follow_redirects=True,
|
|
|
|
)
|
|
|
|
if 200 <= response.status_code < 300:
|
|
|
|
return self.by_ap(
|
|
|
|
canonicalise(response.json(), include_security=True),
|
|
|
|
update=True,
|
|
|
|
)
|