use crate::{
- activities::verify_person_in_community,
- context::lemmy_context,
- fetcher::object_id::ObjectId,
- migrations::CommentInReplyToMigration,
- objects::{create_tombstone, FromApub, Source, ToApub},
+ activities::{verify_is_public, verify_person_in_community},
+ check_apub_id_valid_with_strictness,
+ local_instance,
+ mentions::collect_non_local_mentions,
+ objects::{read_from_string_or_source, verify_is_remote_object},
+ protocol::{objects::note::Note, Source},
PostOrComment,
};
-use activitystreams::{
- base::AnyBase,
- object::{kind::NoteType, Tombstone},
- primitives::OneOrMany,
- unparsed::Unparsed,
-};
-use anyhow::{anyhow, Context};
-use chrono::{DateTime, FixedOffset};
-use lemmy_api_common::blocking;
-use lemmy_apub_lib::{
- traits::ActorType,
- values::{MediaTypeHtml, MediaTypeMarkdown, PublicUrl},
- verify::verify_domains_match,
+use activitypub_federation::{
+ core::object_id::ObjectId,
+ deser::values::MediaTypeMarkdownOrHtml,
+ traits::ApubObject,
+ utils::verify_domains_match,
};
+use activitystreams_kinds::{object::NoteType, public};
+use chrono::NaiveDateTime;
+use lemmy_api_common::utils::blocking;
use lemmy_db_schema::{
- newtypes::CommentId,
source::{
comment::{Comment, CommentForm},
community::Community,
post::Post,
},
traits::Crud,
- DbPool,
};
use lemmy_utils::{
- location_info,
- utils::{convert_datetime, remove_slurs},
- LemmyError,
+ error::LemmyError,
+ utils::{convert_datetime, markdown_to_html, remove_slurs},
};
use lemmy_websocket::LemmyContext;
-use serde::{Deserialize, Serialize};
-use serde_with::skip_serializing_none;
use std::ops::Deref;
use url::Url;
-#[skip_serializing_none]
-#[derive(Clone, Debug, Deserialize, Serialize)]
-#[serde(rename_all = "camelCase")]
-pub struct Note {
- #[serde(rename = "@context")]
- context: OneOrMany<AnyBase>,
- r#type: NoteType,
- id: Url,
- pub(crate) attributed_to: ObjectId<Person>,
- /// Indicates that the object is publicly readable. Unlike [`Post.to`], this one doesn't contain
- /// the community ID, as it would be incompatible with Pleroma (and we can get the community from
- /// the post in [`in_reply_to`]).
- to: PublicUrl,
- content: String,
- media_type: MediaTypeHtml,
- source: Source,
- in_reply_to: CommentInReplyToMigration,
- published: DateTime<FixedOffset>,
- updated: Option<DateTime<FixedOffset>>,
- #[serde(flatten)]
- unparsed: Unparsed,
-}
+#[derive(Clone, Debug)]
+pub struct ApubComment(Comment);
-impl Note {
- pub(crate) fn id_unchecked(&self) -> &Url {
- &self.id
+impl Deref for ApubComment {
+ type Target = Comment;
+ fn deref(&self) -> &Self::Target {
+ &self.0
}
- pub(crate) fn id(&self, expected_domain: &Url) -> Result<&Url, LemmyError> {
- verify_domains_match(&self.id, expected_domain)?;
- Ok(&self.id)
+}
+
+impl From<Comment> for ApubComment {
+ fn from(c: Comment) -> Self {
+ ApubComment(c)
}
+}
- pub(crate) async fn get_parents(
- &self,
- context: &LemmyContext,
- request_counter: &mut i32,
- ) -> Result<(Post, Option<CommentId>), LemmyError> {
- match &self.in_reply_to {
- CommentInReplyToMigration::Old(in_reply_to) => {
- // This post, or the parent comment might not yet exist on this server yet, fetch them.
- let post_id = in_reply_to.get(0).context(location_info!())?;
- let post_id = ObjectId::new(post_id.clone());
- let post = Box::pin(post_id.dereference(context, request_counter)).await?;
-
- // The 2nd item, if it exists, is the parent comment apub_id
- // Nested comments will automatically get fetched recursively
- let parent_id: Option<CommentId> = match in_reply_to.get(1) {
- Some(comment_id) => {
- let comment_id = ObjectId::<Comment>::new(comment_id.clone());
- let parent_comment = Box::pin(comment_id.dereference(context, request_counter)).await?;
-
- Some(parent_comment.id)
- }
- None => None,
- };
-
- Ok((post, parent_id))
- }
- CommentInReplyToMigration::New(in_reply_to) => {
- let parent = Box::pin(in_reply_to.dereference(context, request_counter).await?);
- match parent.deref() {
- PostOrComment::Post(p) => {
- // Workaround because I cant figure ut how to get the post out of the box (and we dont
- // want to stackoverflow in a deep comment hierarchy).
- let post_id = p.id;
- let post = blocking(context.pool(), move |conn| Post::read(conn, post_id)).await??;
- Ok((post, None))
- }
- PostOrComment::Comment(c) => {
- let post_id = c.post_id;
- let post = blocking(context.pool(), move |conn| Post::read(conn, post_id)).await??;
- Ok((post, Some(c.id)))
- }
- }
- }
- }
+#[async_trait::async_trait(?Send)]
+impl ApubObject for ApubComment {
+ type DataType = LemmyContext;
+ type ApubType = Note;
+ type DbType = Comment;
+ type Error = LemmyError;
+
+ fn last_refreshed_at(&self) -> Option<NaiveDateTime> {
+ None
}
- pub(crate) async fn verify(
- &self,
+ #[tracing::instrument(skip_all)]
+ async fn read_from_apub_id(
+ object_id: Url,
context: &LemmyContext,
- request_counter: &mut i32,
- ) -> Result<(), LemmyError> {
- let (post, _parent_comment_id) = self.get_parents(context, request_counter).await?;
- let community_id = post.community_id;
- let community = blocking(context.pool(), move |conn| {
- Community::read(conn, community_id)
- })
- .await??;
+ ) -> Result<Option<Self>, LemmyError> {
+ Ok(
+ blocking(context.pool(), move |conn| {
+ Comment::read_from_apub_id(conn, object_id)
+ })
+ .await??
+ .map(Into::into),
+ )
+ }
- if post.locked {
- return Err(anyhow!("Post is locked").into());
+ #[tracing::instrument(skip_all)]
+ async fn delete(self, context: &LemmyContext) -> Result<(), LemmyError> {
+ if !self.deleted {
+ blocking(context.pool(), move |conn| {
+ Comment::update_deleted(conn, self.id, true)
+ })
+ .await??;
}
- verify_domains_match(self.attributed_to.inner(), &self.id)?;
- verify_person_in_community(
- &self.attributed_to,
- &ObjectId::new(community.actor_id()),
- context,
- request_counter,
- )
- .await?;
Ok(())
}
-}
-#[async_trait::async_trait(?Send)]
-impl ToApub for Comment {
- type ApubType = Note;
-
- async fn to_apub(&self, pool: &DbPool) -> Result<Note, LemmyError> {
+ #[tracing::instrument(skip_all)]
+ async fn into_apub(self, context: &LemmyContext) -> Result<Note, LemmyError> {
let creator_id = self.creator_id;
- let creator = blocking(pool, move |conn| Person::read(conn, creator_id)).await??;
+ let creator = blocking(context.pool(), move |conn| Person::read(conn, creator_id)).await??;
let post_id = self.post_id;
- let post = blocking(pool, move |conn| Post::read(conn, post_id)).await??;
-
- // Add a vector containing some important info to the "in_reply_to" field
- // [post_ap_id, Option(parent_comment_ap_id)]
- let mut in_reply_to_vec = vec![post.ap_id.into_inner()];
-
- if let Some(parent_id) = self.parent_id {
- let parent_comment = blocking(pool, move |conn| Comment::read(conn, parent_id)).await??;
+ let post = blocking(context.pool(), move |conn| Post::read(conn, post_id)).await??;
+ let community_id = post.community_id;
+ let community = blocking(context.pool(), move |conn| {
+ Community::read(conn, community_id)
+ })
+ .await??;
- in_reply_to_vec.push(parent_comment.ap_id.into_inner());
- }
+ let in_reply_to = if let Some(comment_id) = self.parent_comment_id() {
+ let parent_comment =
+ blocking(context.pool(), move |conn| Comment::read(conn, comment_id)).await??;
+ ObjectId::<PostOrComment>::new(parent_comment.ap_id)
+ } else {
+ ObjectId::<PostOrComment>::new(post.ap_id)
+ };
+ let maa =
+ collect_non_local_mentions(&self, ObjectId::new(community.actor_id), context, &mut 0).await?;
let note = Note {
- context: lemmy_context(),
r#type: NoteType::Note,
- id: self.ap_id.to_owned().into_inner(),
+ id: ObjectId::new(self.ap_id.clone()),
attributed_to: ObjectId::new(creator.actor_id),
- to: PublicUrl::Public,
- content: self.content.clone(),
- media_type: MediaTypeHtml::Html,
- source: Source {
- content: self.content.clone(),
- media_type: MediaTypeMarkdown::Markdown,
- },
- in_reply_to: CommentInReplyToMigration::Old(in_reply_to_vec),
- published: convert_datetime(self.published),
+ to: vec![public()],
+ cc: maa.ccs,
+ content: markdown_to_html(&self.content),
+ media_type: Some(MediaTypeMarkdownOrHtml::Html),
+ source: Some(Source::new(self.content.clone())),
+ in_reply_to,
+ published: Some(convert_datetime(self.published)),
updated: self.updated.map(convert_datetime),
- unparsed: Default::default(),
+ tag: maa.tags,
+ distinguished: Some(self.distinguished),
};
Ok(note)
}
- fn to_tombstone(&self) -> Result<Tombstone, LemmyError> {
- create_tombstone(
- self.deleted,
- self.ap_id.to_owned().into(),
- self.updated,
- NoteType::Note,
+ #[tracing::instrument(skip_all)]
+ async fn verify(
+ note: &Note,
+ expected_domain: &Url,
+ context: &LemmyContext,
+ request_counter: &mut i32,
+ ) -> Result<(), LemmyError> {
+ verify_domains_match(note.id.inner(), expected_domain)?;
+ verify_domains_match(note.attributed_to.inner(), note.id.inner())?;
+ verify_is_public(¬e.to, ¬e.cc)?;
+ let (post, _) = note.get_parents(context, request_counter).await?;
+ let community_id = post.community_id;
+ let community = blocking(context.pool(), move |conn| {
+ Community::read(conn, community_id)
+ })
+ .await??;
+ check_apub_id_valid_with_strictness(note.id.inner(), community.local, context.settings())?;
+ verify_is_remote_object(note.id.inner(), context.settings())?;
+ verify_person_in_community(
+ ¬e.attributed_to,
+ &community.into(),
+ context,
+ request_counter,
)
+ .await?;
+ if post.locked {
+ return Err(LemmyError::from_message("Post is locked"));
+ }
+ Ok(())
}
-}
-
-#[async_trait::async_trait(?Send)]
-impl FromApub for Comment {
- type ApubType = Note;
/// Converts a `Note` to `Comment`.
///
/// If the parent community, post and comment(s) are not known locally, these are also fetched.
+ #[tracing::instrument(skip_all)]
async fn from_apub(
- note: &Note,
+ note: Note,
context: &LemmyContext,
- expected_domain: &Url,
request_counter: &mut i32,
- ) -> Result<Comment, LemmyError> {
- let ap_id = Some(note.id(expected_domain)?.clone().into());
+ ) -> Result<ApubComment, LemmyError> {
let creator = note
.attributed_to
- .dereference(context, request_counter)
+ .dereference(context, local_instance(context), request_counter)
.await?;
- let (post, parent_comment_id) = note.get_parents(context, request_counter).await?;
- if post.locked {
- return Err(anyhow!("Post is locked").into());
- }
+ let (post, parent_comment) = note.get_parents(context, request_counter).await?;
- let content = ¬e.source.content;
- let content_slurs_removed = remove_slurs(content, &context.settings().slur_regex());
+ let content = read_from_string_or_source(¬e.content, ¬e.media_type, ¬e.source);
+ let content_slurs_removed = remove_slurs(&content, &context.settings().slur_regex());
let form = CommentForm {
creator_id: creator.id,
post_id: post.id,
- parent_id: parent_comment_id,
content: content_slurs_removed,
removed: None,
- read: None,
- published: Some(note.published.naive_local()),
- updated: note.updated.map(|u| u.to_owned().naive_local()),
+ published: note.published.map(|u| u.naive_local()),
+ updated: note.updated.map(|u| u.naive_local()),
deleted: None,
- ap_id,
+ ap_id: Some(note.id.into()),
+ distinguished: note.distinguished,
local: Some(false),
};
- Ok(blocking(context.pool(), move |conn| Comment::upsert(conn, &form)).await??)
+ let parent_comment_path = parent_comment.map(|t| t.0.path);
+ let comment = blocking(context.pool(), move |conn| {
+ Comment::create(conn, &form, parent_comment_path.as_ref())
+ })
+ .await??;
+ Ok(comment.into())
+ }
+}
+
+#[cfg(test)]
+pub(crate) mod tests {
+ use super::*;
+ use crate::{
+ objects::{
+ community::{tests::parse_lemmy_community, ApubCommunity},
+ instance::ApubSite,
+ person::{tests::parse_lemmy_person, ApubPerson},
+ post::ApubPost,
+ tests::init_context,
+ },
+ protocol::tests::file_to_json_object,
+ };
+ use assert_json_diff::assert_json_include;
+ use html2md::parse_html;
+ use lemmy_db_schema::source::site::Site;
+ use serial_test::serial;
+
+ async fn prepare_comment_test(
+ url: &Url,
+ context: &LemmyContext,
+ ) -> (ApubPerson, ApubCommunity, ApubPost, ApubSite) {
+ let (person, site) = parse_lemmy_person(context).await;
+ let community = parse_lemmy_community(context).await;
+ let post_json = file_to_json_object("assets/lemmy/objects/page.json").unwrap();
+ ApubPost::verify(&post_json, url, context, &mut 0)
+ .await
+ .unwrap();
+ let post = ApubPost::from_apub(post_json, context, &mut 0)
+ .await
+ .unwrap();
+ (person, community, post, site)
+ }
+
+ fn cleanup(data: (ApubPerson, ApubCommunity, ApubPost, ApubSite), context: &LemmyContext) {
+ Post::delete(&*context.pool().get().unwrap(), data.2.id).unwrap();
+ Community::delete(&*context.pool().get().unwrap(), data.1.id).unwrap();
+ Person::delete(&*context.pool().get().unwrap(), data.0.id).unwrap();
+ Site::delete(&*context.pool().get().unwrap(), data.3.id).unwrap();
+ }
+
+ #[actix_rt::test]
+ #[serial]
+ pub(crate) async fn test_parse_lemmy_comment() {
+ let context = init_context();
+ let url = Url::parse("https://enterprise.lemmy.ml/comment/38741").unwrap();
+ let data = prepare_comment_test(&url, &context).await;
+
+ let json: Note = file_to_json_object("assets/lemmy/objects/note.json").unwrap();
+ let mut request_counter = 0;
+ ApubComment::verify(&json, &url, &context, &mut request_counter)
+ .await
+ .unwrap();
+ let comment = ApubComment::from_apub(json.clone(), &context, &mut request_counter)
+ .await
+ .unwrap();
+
+ assert_eq!(comment.ap_id, url.into());
+ assert_eq!(comment.content.len(), 14);
+ assert!(!comment.local);
+ assert_eq!(request_counter, 0);
+
+ let comment_id = comment.id;
+ let to_apub = comment.into_apub(&context).await.unwrap();
+ assert_json_include!(actual: json, expected: to_apub);
+
+ Comment::delete(&*context.pool().get().unwrap(), comment_id).unwrap();
+ cleanup(data, &context);
+ }
+
+ #[actix_rt::test]
+ #[serial]
+ async fn test_parse_pleroma_comment() {
+ let context = init_context();
+ let url = Url::parse("https://enterprise.lemmy.ml/comment/38741").unwrap();
+ let data = prepare_comment_test(&url, &context).await;
+
+ let pleroma_url =
+ Url::parse("https://queer.hacktivis.me/objects/8d4973f4-53de-49cd-8c27-df160e16a9c2")
+ .unwrap();
+ let person_json = file_to_json_object("assets/pleroma/objects/person.json").unwrap();
+ ApubPerson::verify(&person_json, &pleroma_url, &context, &mut 0)
+ .await
+ .unwrap();
+ ApubPerson::from_apub(person_json, &context, &mut 0)
+ .await
+ .unwrap();
+ let json = file_to_json_object("assets/pleroma/objects/note.json").unwrap();
+ let mut request_counter = 0;
+ ApubComment::verify(&json, &pleroma_url, &context, &mut request_counter)
+ .await
+ .unwrap();
+ let comment = ApubComment::from_apub(json, &context, &mut request_counter)
+ .await
+ .unwrap();
+
+ assert_eq!(comment.ap_id, pleroma_url.into());
+ assert_eq!(comment.content.len(), 64);
+ assert!(!comment.local);
+ assert_eq!(request_counter, 0);
+
+ Comment::delete(&*context.pool().get().unwrap(), comment.id).unwrap();
+ cleanup(data, &context);
+ }
+
+ #[actix_rt::test]
+ #[serial]
+ async fn test_html_to_markdown_sanitize() {
+ let parsed = parse_html("<script></script><b>hello</b>");
+ assert_eq!(parsed, "**hello**");
}
}