]> Untitled Git - lemmy.git/blobdiff - crates/apub/src/objects/comment.rs
Cache & Optimize Woodpecker CI (#3450)
[lemmy.git] / crates / apub / src / objects / comment.rs
index 1748af791a2216434a710472ed75428b357460c8..2954de0968cab4af742a911b161cfabfdb889c00 100644 (file)
 use crate::{
-  activities::verify_person_in_community,
-  context::lemmy_context,
-  fetcher::object_id::ObjectId,
-  migrations::CommentInReplyToMigration,
-  objects::{create_tombstone, FromApub, Source, ToApub},
-  PostOrComment,
-};
-use activitystreams::{
-  base::AnyBase,
-  object::{kind::NoteType, Tombstone},
-  primitives::OneOrMany,
-  unparsed::Unparsed,
+  activities::{verify_is_public, verify_person_in_community},
+  check_apub_id_valid_with_strictness,
+  mentions::collect_non_local_mentions,
+  objects::{read_from_string_or_source, verify_is_remote_object},
+  protocol::{
+    objects::{note::Note, LanguageTag},
+    InCommunity,
+    Source,
+  },
 };
-use anyhow::{anyhow, Context};
-use chrono::{DateTime, FixedOffset};
-use lemmy_api_common::blocking;
-use lemmy_apub_lib::{
-  traits::ActorType,
-  values::{MediaTypeHtml, MediaTypeMarkdown, PublicUrl},
-  verify::verify_domains_match,
+use activitypub_federation::{
+  config::Data,
+  kinds::{object::NoteType, public},
+  protocol::{values::MediaTypeMarkdownOrHtml, verification::verify_domains_match},
+  traits::Object,
 };
+use chrono::NaiveDateTime;
+use lemmy_api_common::{context::LemmyContext, utils::local_site_opt_to_slur_regex};
 use lemmy_db_schema::{
-  newtypes::CommentId,
   source::{
-    comment::{Comment, CommentForm},
+    comment::{Comment, CommentInsertForm, CommentUpdateForm},
     community::Community,
+    local_site::LocalSite,
     person::Person,
     post::Post,
   },
   traits::Crud,
-  DbPool,
 };
 use lemmy_utils::{
-  location_info,
-  utils::{convert_datetime, remove_slurs},
-  LemmyError,
+  error::{LemmyError, LemmyErrorType},
+  utils::{markdown::markdown_to_html, slurs::remove_slurs, time::convert_datetime},
 };
-use lemmy_websocket::LemmyContext;
-use serde::{Deserialize, Serialize};
-use serde_with::skip_serializing_none;
 use std::ops::Deref;
 use url::Url;
 
-#[skip_serializing_none]
-#[derive(Clone, Debug, Deserialize, Serialize)]
-#[serde(rename_all = "camelCase")]
-pub struct Note {
-  #[serde(rename = "@context")]
-  context: OneOrMany<AnyBase>,
-  r#type: NoteType,
-  id: Url,
-  pub(crate) attributed_to: ObjectId<Person>,
-  /// Indicates that the object is publicly readable. Unlike [`Post.to`], this one doesn't contain
-  /// the community ID, as it would be incompatible with Pleroma (and we can get the community from
-  /// the post in [`in_reply_to`]).
-  to: PublicUrl,
-  content: String,
-  media_type: MediaTypeHtml,
-  source: Source,
-  in_reply_to: CommentInReplyToMigration,
-  published: DateTime<FixedOffset>,
-  updated: Option<DateTime<FixedOffset>>,
-  #[serde(flatten)]
-  unparsed: Unparsed,
-}
+#[derive(Clone, Debug)]
+pub struct ApubComment(pub(crate) Comment);
 
-impl Note {
-  pub(crate) fn id_unchecked(&self) -> &Url {
-    &self.id
+impl Deref for ApubComment {
+  type Target = Comment;
+  fn deref(&self) -> &Self::Target {
+    &self.0
   }
-  pub(crate) fn id(&self, expected_domain: &Url) -> Result<&Url, LemmyError> {
-    verify_domains_match(&self.id, expected_domain)?;
-    Ok(&self.id)
+}
+
+impl From<Comment> for ApubComment {
+  fn from(c: Comment) -> Self {
+    ApubComment(c)
   }
+}
 
-  pub(crate) async fn get_parents(
-    &self,
-    context: &LemmyContext,
-    request_counter: &mut i32,
-  ) -> Result<(Post, Option<CommentId>), LemmyError> {
-    match &self.in_reply_to {
-      CommentInReplyToMigration::Old(in_reply_to) => {
-        // This post, or the parent comment might not yet exist on this server yet, fetch them.
-        let post_id = in_reply_to.get(0).context(location_info!())?;
-        let post_id = ObjectId::new(post_id.clone());
-        let post = Box::pin(post_id.dereference(context, request_counter)).await?;
-
-        // The 2nd item, if it exists, is the parent comment apub_id
-        // Nested comments will automatically get fetched recursively
-        let parent_id: Option<CommentId> = match in_reply_to.get(1) {
-          Some(comment_id) => {
-            let comment_id = ObjectId::<Comment>::new(comment_id.clone());
-            let parent_comment = Box::pin(comment_id.dereference(context, request_counter)).await?;
-
-            Some(parent_comment.id)
-          }
-          None => None,
-        };
-
-        Ok((post, parent_id))
-      }
-      CommentInReplyToMigration::New(in_reply_to) => {
-        let parent = Box::pin(in_reply_to.dereference(context, request_counter).await?);
-        match parent.deref() {
-          PostOrComment::Post(p) => {
-            // Workaround because I cant figure ut how to get the post out of the box (and we dont
-            // want to stackoverflow in a deep comment hierarchy).
-            let post_id = p.id;
-            let post = blocking(context.pool(), move |conn| Post::read(conn, post_id)).await??;
-            Ok((post, None))
-          }
-          PostOrComment::Comment(c) => {
-            let post_id = c.post_id;
-            let post = blocking(context.pool(), move |conn| Post::read(conn, post_id)).await??;
-            Ok((post, Some(c.id)))
-          }
-        }
-      }
-    }
+#[async_trait::async_trait]
+impl Object for ApubComment {
+  type DataType = LemmyContext;
+  type Kind = Note;
+  type Error = LemmyError;
+
+  fn last_refreshed_at(&self) -> Option<NaiveDateTime> {
+    None
   }
 
-  pub(crate) async fn verify(
-    &self,
-    context: &LemmyContext,
-    request_counter: &mut i32,
-  ) -> Result<(), LemmyError> {
-    let (post, _parent_comment_id) = self.get_parents(context, request_counter).await?;
-    let community_id = post.community_id;
-    let community = blocking(context.pool(), move |conn| {
-      Community::read(conn, community_id)
-    })
-    .await??;
+  #[tracing::instrument(skip_all)]
+  async fn read_from_id(
+    object_id: Url,
+    context: &Data<Self::DataType>,
+  ) -> Result<Option<Self>, LemmyError> {
+    Ok(
+      Comment::read_from_apub_id(&mut context.pool(), object_id)
+        .await?
+        .map(Into::into),
+    )
+  }
 
-    if post.locked {
-      return Err(anyhow!("Post is locked").into());
+  #[tracing::instrument(skip_all)]
+  async fn delete(self, context: &Data<Self::DataType>) -> Result<(), LemmyError> {
+    if !self.deleted {
+      let form = CommentUpdateForm::builder().deleted(Some(true)).build();
+      Comment::update(&mut context.pool(), self.id, &form).await?;
     }
-    verify_domains_match(self.attributed_to.inner(), &self.id)?;
-    verify_person_in_community(
-      &self.attributed_to,
-      &ObjectId::new(community.actor_id()),
-      context,
-      request_counter,
-    )
-    .await?;
     Ok(())
   }
-}
-
-#[async_trait::async_trait(?Send)]
-impl ToApub for Comment {
-  type ApubType = Note;
 
-  async fn to_apub(&self, pool: &DbPool) -> Result<Note, LemmyError> {
+  #[tracing::instrument(skip_all)]
+  async fn into_json(self, context: &Data<Self::DataType>) -> Result<Note, LemmyError> {
     let creator_id = self.creator_id;
-    let creator = blocking(pool, move |conn| Person::read(conn, creator_id)).await??;
+    let creator = Person::read(&mut context.pool(), creator_id).await?;
 
     let post_id = self.post_id;
-    let post = blocking(pool, move |conn| Post::read(conn, post_id)).await??;
-
-    // Add a vector containing some important info to the "in_reply_to" field
-    // [post_ap_id, Option(parent_comment_ap_id)]
-    let mut in_reply_to_vec = vec![post.ap_id.into_inner()];
-
-    if let Some(parent_id) = self.parent_id {
-      let parent_comment = blocking(pool, move |conn| Comment::read(conn, parent_id)).await??;
+    let post = Post::read(&mut context.pool(), post_id).await?;
+    let community_id = post.community_id;
+    let community = Community::read(&mut context.pool(), community_id).await?;
 
-      in_reply_to_vec.push(parent_comment.ap_id.into_inner());
-    }
+    let in_reply_to = if let Some(comment_id) = self.parent_comment_id() {
+      let parent_comment = Comment::read(&mut context.pool(), comment_id).await?;
+      parent_comment.ap_id.into()
+    } else {
+      post.ap_id.into()
+    };
+    let language = LanguageTag::new_single(self.language_id, &mut context.pool()).await?;
+    let maa = collect_non_local_mentions(&self, community.actor_id.clone().into(), context).await?;
 
     let note = Note {
-      context: lemmy_context(),
       r#type: NoteType::Note,
-      id: self.ap_id.to_owned().into_inner(),
-      attributed_to: ObjectId::new(creator.actor_id),
-      to: PublicUrl::Public,
-      content: self.content.clone(),
-      media_type: MediaTypeHtml::Html,
-      source: Source {
-        content: self.content.clone(),
-        media_type: MediaTypeMarkdown::Markdown,
-      },
-      in_reply_to: CommentInReplyToMigration::Old(in_reply_to_vec),
-      published: convert_datetime(self.published),
+      id: self.ap_id.clone().into(),
+      attributed_to: creator.actor_id.into(),
+      to: vec![public()],
+      cc: maa.ccs,
+      content: markdown_to_html(&self.content),
+      media_type: Some(MediaTypeMarkdownOrHtml::Html),
+      source: Some(Source::new(self.content.clone())),
+      in_reply_to,
+      published: Some(convert_datetime(self.published)),
       updated: self.updated.map(convert_datetime),
-      unparsed: Default::default(),
+      tag: maa.tags,
+      distinguished: Some(self.distinguished),
+      language,
+      audience: Some(community.actor_id.into()),
     };
 
     Ok(note)
   }
 
-  fn to_tombstone(&self) -> Result<Tombstone, LemmyError> {
-    create_tombstone(
-      self.deleted,
-      self.ap_id.to_owned().into(),
-      self.updated,
-      NoteType::Note,
-    )
-  }
-}
+  #[tracing::instrument(skip_all)]
+  async fn verify(
+    note: &Note,
+    expected_domain: &Url,
+    context: &Data<LemmyContext>,
+  ) -> Result<(), LemmyError> {
+    verify_domains_match(note.id.inner(), expected_domain)?;
+    verify_domains_match(note.attributed_to.inner(), note.id.inner())?;
+    verify_is_public(&note.to, &note.cc)?;
+    let community = note.community(context).await?;
 
-#[async_trait::async_trait(?Send)]
-impl FromApub for Comment {
-  type ApubType = Note;
+    check_apub_id_valid_with_strictness(note.id.inner(), community.local, context).await?;
+    verify_is_remote_object(note.id.inner(), context.settings())?;
+    verify_person_in_community(&note.attributed_to, &community, context).await?;
+    let (post, _) = note.get_parents(context).await?;
+    if post.locked {
+      return Err(LemmyErrorType::PostIsLocked)?;
+    }
+    Ok(())
+  }
 
   /// Converts a `Note` to `Comment`.
   ///
   /// If the parent community, post and comment(s) are not known locally, these are also fetched.
-  async fn from_apub(
-    note: &Note,
-    context: &LemmyContext,
-    expected_domain: &Url,
-    request_counter: &mut i32,
-  ) -> Result<Comment, LemmyError> {
-    let ap_id = Some(note.id(expected_domain)?.clone().into());
-    let creator = note
-      .attributed_to
-      .dereference(context, request_counter)
-      .await?;
-    let (post, parent_comment_id) = note.get_parents(context, request_counter).await?;
-    if post.locked {
-      return Err(anyhow!("Post is locked").into());
-    }
+  #[tracing::instrument(skip_all)]
+  async fn from_json(note: Note, context: &Data<LemmyContext>) -> Result<ApubComment, LemmyError> {
+    let creator = note.attributed_to.dereference(context).await?;
+    let (post, parent_comment) = note.get_parents(context).await?;
 
-    let content = &note.source.content;
-    let content_slurs_removed = remove_slurs(content, &context.settings().slur_regex());
+    let content = read_from_string_or_source(&note.content, &note.media_type, &note.source);
 
-    let form = CommentForm {
+    let local_site = LocalSite::read(&mut context.pool()).await.ok();
+    let slur_regex = &local_site_opt_to_slur_regex(&local_site);
+    let content_slurs_removed = remove_slurs(&content, slur_regex);
+    let language_id =
+      LanguageTag::to_language_id_single(note.language, &mut context.pool()).await?;
+
+    let form = CommentInsertForm {
       creator_id: creator.id,
       post_id: post.id,
-      parent_id: parent_comment_id,
       content: content_slurs_removed,
       removed: None,
-      read: None,
-      published: Some(note.published.naive_local()),
-      updated: note.updated.map(|u| u.to_owned().naive_local()),
-      deleted: None,
-      ap_id,
+      published: note.published.map(|u| u.naive_local()),
+      updated: note.updated.map(|u| u.naive_local()),
+      deleted: Some(false),
+      ap_id: Some(note.id.into()),
+      distinguished: note.distinguished,
       local: Some(false),
+      language_id,
     };
-    Ok(blocking(context.pool(), move |conn| Comment::upsert(conn, &form)).await??)
+    let parent_comment_path = parent_comment.map(|t| t.0.path);
+    let comment = Comment::create(&mut context.pool(), &form, parent_comment_path.as_ref()).await?;
+    Ok(comment.into())
+  }
+}
+
+#[cfg(test)]
+pub(crate) mod tests {
+  #![allow(clippy::unwrap_used)]
+  #![allow(clippy::indexing_slicing)]
+
+  use super::*;
+  use crate::{
+    objects::{
+      community::{tests::parse_lemmy_community, ApubCommunity},
+      instance::ApubSite,
+      person::{tests::parse_lemmy_person, ApubPerson},
+      post::ApubPost,
+      tests::init_context,
+    },
+    protocol::tests::file_to_json_object,
+  };
+  use assert_json_diff::assert_json_include;
+  use html2md::parse_html;
+  use lemmy_db_schema::source::site::Site;
+  use serial_test::serial;
+
+  async fn prepare_comment_test(
+    url: &Url,
+    context: &Data<LemmyContext>,
+  ) -> (ApubPerson, ApubCommunity, ApubPost, ApubSite) {
+    // use separate counter so this doesnt affect tests
+    let context2 = context.reset_request_count();
+    let (person, site) = parse_lemmy_person(&context2).await;
+    let community = parse_lemmy_community(&context2).await;
+    let post_json = file_to_json_object("assets/lemmy/objects/page.json").unwrap();
+    ApubPost::verify(&post_json, url, &context2).await.unwrap();
+    let post = ApubPost::from_json(post_json, &context2).await.unwrap();
+    (person, community, post, site)
+  }
+
+  async fn cleanup(data: (ApubPerson, ApubCommunity, ApubPost, ApubSite), context: &LemmyContext) {
+    Post::delete(&mut context.pool(), data.2.id).await.unwrap();
+    Community::delete(&mut context.pool(), data.1.id)
+      .await
+      .unwrap();
+    Person::delete(&mut context.pool(), data.0.id)
+      .await
+      .unwrap();
+    Site::delete(&mut context.pool(), data.3.id).await.unwrap();
+    LocalSite::delete(&mut context.pool()).await.unwrap();
+  }
+
+  #[tokio::test]
+  #[serial]
+  pub(crate) async fn test_parse_lemmy_comment() {
+    let context = init_context().await;
+    let url = Url::parse("https://enterprise.lemmy.ml/comment/38741").unwrap();
+    let data = prepare_comment_test(&url, &context).await;
+
+    let json: Note = file_to_json_object("assets/lemmy/objects/note.json").unwrap();
+    ApubComment::verify(&json, &url, &context).await.unwrap();
+    let comment = ApubComment::from_json(json.clone(), &context)
+      .await
+      .unwrap();
+
+    assert_eq!(comment.ap_id, url.into());
+    assert_eq!(comment.content.len(), 14);
+    assert!(!comment.local);
+    assert_eq!(context.request_count(), 0);
+
+    let comment_id = comment.id;
+    let to_apub = comment.into_json(&context).await.unwrap();
+    assert_json_include!(actual: json, expected: to_apub);
+
+    Comment::delete(&mut context.pool(), comment_id)
+      .await
+      .unwrap();
+    cleanup(data, &context).await;
+  }
+
+  #[tokio::test]
+  #[serial]
+  async fn test_parse_pleroma_comment() {
+    let context = init_context().await;
+    let url = Url::parse("https://enterprise.lemmy.ml/comment/38741").unwrap();
+    let data = prepare_comment_test(&url, &context).await;
+
+    let pleroma_url =
+      Url::parse("https://queer.hacktivis.me/objects/8d4973f4-53de-49cd-8c27-df160e16a9c2")
+        .unwrap();
+    let person_json = file_to_json_object("assets/pleroma/objects/person.json").unwrap();
+    ApubPerson::verify(&person_json, &pleroma_url, &context)
+      .await
+      .unwrap();
+    ApubPerson::from_json(person_json, &context).await.unwrap();
+    let json = file_to_json_object("assets/pleroma/objects/note.json").unwrap();
+    ApubComment::verify(&json, &pleroma_url, &context)
+      .await
+      .unwrap();
+    let comment = ApubComment::from_json(json, &context).await.unwrap();
+
+    assert_eq!(comment.ap_id, pleroma_url.into());
+    assert_eq!(comment.content.len(), 64);
+    assert!(!comment.local);
+    assert_eq!(context.request_count(), 1);
+
+    Comment::delete(&mut context.pool(), comment.id)
+      .await
+      .unwrap();
+    cleanup(data, &context).await;
+  }
+
+  #[tokio::test]
+  #[serial]
+  async fn test_html_to_markdown_sanitize() {
+    let parsed = parse_html("<script></script><b>hello</b>");
+    assert_eq!(parsed, "**hello**");
   }
 }