X-Git-Url: http://these/git/?a=blobdiff_plain;f=crates%2Fapub%2Fsrc%2Fobjects%2Fcomment.rs;h=2954de0968cab4af742a911b161cfabfdb889c00;hb=92568956353f21649ed9aff68b42699c9d036f30;hp=e27817d6fd418dbf1e6ce32230e47747d631efc1;hpb=8708ad1b44e58dfbd8c555f9bc270daadbfa01af;p=lemmy.git diff --git a/crates/apub/src/objects/comment.rs b/crates/apub/src/objects/comment.rs index e27817d6..2954de09 100644 --- a/crates/apub/src/objects/comment.rs +++ b/crates/apub/src/objects/comment.rs @@ -1,242 +1,293 @@ use crate::{ - activities::verify_person_in_community, - extensions::context::lemmy_context, - fetcher::object_id::ObjectId, - migrations::CommentInReplyToMigration, - objects::{create_tombstone, FromApub, Source, ToApub}, - ActorType, - PostOrComment, -}; -use activitystreams::{ - base::AnyBase, - object::{kind::NoteType, Tombstone}, - primitives::OneOrMany, - unparsed::Unparsed, + activities::{verify_is_public, verify_person_in_community}, + check_apub_id_valid_with_strictness, + mentions::collect_non_local_mentions, + objects::{read_from_string_or_source, verify_is_remote_object}, + protocol::{ + objects::{note::Note, LanguageTag}, + InCommunity, + Source, + }, }; -use anyhow::{anyhow, Context}; -use chrono::{DateTime, FixedOffset}; -use lemmy_api_common::blocking; -use lemmy_apub_lib::{ - values::{MediaTypeHtml, MediaTypeMarkdown, PublicUrl}, - verify_domains_match, +use activitypub_federation::{ + config::Data, + kinds::{object::NoteType, public}, + protocol::{values::MediaTypeMarkdownOrHtml, verification::verify_domains_match}, + traits::Object, }; -use lemmy_db_queries::{source::comment::Comment_, Crud, DbPool}; +use chrono::NaiveDateTime; +use lemmy_api_common::{context::LemmyContext, utils::local_site_opt_to_slur_regex}; use lemmy_db_schema::{ source::{ - comment::{Comment, CommentForm}, + comment::{Comment, CommentInsertForm, CommentUpdateForm}, community::Community, + local_site::LocalSite, person::Person, post::Post, }, - CommentId, + traits::Crud, }; use lemmy_utils::{ - location_info, - utils::{convert_datetime, remove_slurs}, - LemmyError, + error::{LemmyError, LemmyErrorType}, + utils::{markdown::markdown_to_html, slurs::remove_slurs, time::convert_datetime}, }; -use lemmy_websocket::LemmyContext; -use serde::{Deserialize, Serialize}; -use serde_with::skip_serializing_none; use std::ops::Deref; use url::Url; -#[skip_serializing_none] -#[derive(Clone, Debug, Deserialize, Serialize)] -#[serde(rename_all = "camelCase")] -pub struct Note { - #[serde(rename = "@context")] - context: OneOrMany, - r#type: NoteType, - id: Url, - pub(crate) attributed_to: ObjectId, - /// Indicates that the object is publicly readable. Unlike [`Post.to`], this one doesn't contain - /// the community ID, as it would be incompatible with Pleroma (and we can get the community from - /// the post in [`in_reply_to`]). - to: PublicUrl, - content: String, - media_type: MediaTypeHtml, - source: Source, - in_reply_to: CommentInReplyToMigration, - published: DateTime, - updated: Option>, - #[serde(flatten)] - unparsed: Unparsed, -} +#[derive(Clone, Debug)] +pub struct ApubComment(pub(crate) Comment); -impl Note { - pub(crate) fn id_unchecked(&self) -> &Url { - &self.id +impl Deref for ApubComment { + type Target = Comment; + fn deref(&self) -> &Self::Target { + &self.0 } - pub(crate) fn id(&self, expected_domain: &Url) -> Result<&Url, LemmyError> { - verify_domains_match(&self.id, expected_domain)?; - Ok(&self.id) +} + +impl From for ApubComment { + fn from(c: Comment) -> Self { + ApubComment(c) } +} - async fn get_parents( - &self, - context: &LemmyContext, - request_counter: &mut i32, - ) -> Result<(Post, Option), LemmyError> { - match &self.in_reply_to { - CommentInReplyToMigration::Old(in_reply_to) => { - // This post, or the parent comment might not yet exist on this server yet, fetch them. - let post_id = in_reply_to.get(0).context(location_info!())?; - let post_id = ObjectId::new(post_id.clone()); - let post = Box::pin(post_id.dereference(context, request_counter)).await?; - - // The 2nd item, if it exists, is the parent comment apub_id - // Nested comments will automatically get fetched recursively - let parent_id: Option = match in_reply_to.get(1) { - Some(comment_id) => { - let comment_id = ObjectId::::new(comment_id.clone()); - let parent_comment = Box::pin(comment_id.dereference(context, request_counter)).await?; - - Some(parent_comment.id) - } - None => None, - }; - - Ok((post, parent_id)) - } - CommentInReplyToMigration::New(in_reply_to) => { - let parent = Box::pin(in_reply_to.dereference(context, request_counter).await?); - match parent.deref() { - PostOrComment::Post(p) => { - // Workaround because I cant figure ut how to get the post out of the box (and we dont - // want to stackoverflow in a deep comment hierarchy). - let post_id = p.id; - let post = blocking(context.pool(), move |conn| Post::read(conn, post_id)).await??; - Ok((post, None)) - } - PostOrComment::Comment(c) => { - let post_id = c.post_id; - let post = blocking(context.pool(), move |conn| Post::read(conn, post_id)).await??; - Ok((post, Some(c.id))) - } - } - } - } +#[async_trait::async_trait] +impl Object for ApubComment { + type DataType = LemmyContext; + type Kind = Note; + type Error = LemmyError; + + fn last_refreshed_at(&self) -> Option { + None } - pub(crate) async fn verify( - &self, - context: &LemmyContext, - request_counter: &mut i32, - ) -> Result<(), LemmyError> { - let (post, _parent_comment_id) = self.get_parents(context, request_counter).await?; - let community_id = post.community_id; - let community = blocking(context.pool(), move |conn| { - Community::read(conn, community_id) - }) - .await??; + #[tracing::instrument(skip_all)] + async fn read_from_id( + object_id: Url, + context: &Data, + ) -> Result, LemmyError> { + Ok( + Comment::read_from_apub_id(&mut context.pool(), object_id) + .await? + .map(Into::into), + ) + } - if post.locked { - return Err(anyhow!("Post is locked").into()); + #[tracing::instrument(skip_all)] + async fn delete(self, context: &Data) -> Result<(), LemmyError> { + if !self.deleted { + let form = CommentUpdateForm::builder().deleted(Some(true)).build(); + Comment::update(&mut context.pool(), self.id, &form).await?; } - verify_domains_match(self.attributed_to.inner(), &self.id)?; - verify_person_in_community( - &self.attributed_to, - &ObjectId::new(community.actor_id()), - context, - request_counter, - ) - .await?; Ok(()) } -} - -#[async_trait::async_trait(?Send)] -impl ToApub for Comment { - type ApubType = Note; - async fn to_apub(&self, pool: &DbPool) -> Result { + #[tracing::instrument(skip_all)] + async fn into_json(self, context: &Data) -> Result { let creator_id = self.creator_id; - let creator = blocking(pool, move |conn| Person::read(conn, creator_id)).await??; + let creator = Person::read(&mut context.pool(), creator_id).await?; let post_id = self.post_id; - let post = blocking(pool, move |conn| Post::read(conn, post_id)).await??; - - // Add a vector containing some important info to the "in_reply_to" field - // [post_ap_id, Option(parent_comment_ap_id)] - let mut in_reply_to_vec = vec![post.ap_id.into_inner()]; - - if let Some(parent_id) = self.parent_id { - let parent_comment = blocking(pool, move |conn| Comment::read(conn, parent_id)).await??; + let post = Post::read(&mut context.pool(), post_id).await?; + let community_id = post.community_id; + let community = Community::read(&mut context.pool(), community_id).await?; - in_reply_to_vec.push(parent_comment.ap_id.into_inner()); - } + let in_reply_to = if let Some(comment_id) = self.parent_comment_id() { + let parent_comment = Comment::read(&mut context.pool(), comment_id).await?; + parent_comment.ap_id.into() + } else { + post.ap_id.into() + }; + let language = LanguageTag::new_single(self.language_id, &mut context.pool()).await?; + let maa = collect_non_local_mentions(&self, community.actor_id.clone().into(), context).await?; let note = Note { - context: lemmy_context(), r#type: NoteType::Note, - id: self.ap_id.to_owned().into_inner(), - attributed_to: ObjectId::new(creator.actor_id), - to: PublicUrl::Public, - content: self.content.clone(), - media_type: MediaTypeHtml::Html, - source: Source { - content: self.content.clone(), - media_type: MediaTypeMarkdown::Markdown, - }, - in_reply_to: CommentInReplyToMigration::Old(in_reply_to_vec), - published: convert_datetime(self.published), + id: self.ap_id.clone().into(), + attributed_to: creator.actor_id.into(), + to: vec![public()], + cc: maa.ccs, + content: markdown_to_html(&self.content), + media_type: Some(MediaTypeMarkdownOrHtml::Html), + source: Some(Source::new(self.content.clone())), + in_reply_to, + published: Some(convert_datetime(self.published)), updated: self.updated.map(convert_datetime), - unparsed: Default::default(), + tag: maa.tags, + distinguished: Some(self.distinguished), + language, + audience: Some(community.actor_id.into()), }; Ok(note) } - fn to_tombstone(&self) -> Result { - create_tombstone( - self.deleted, - self.ap_id.to_owned().into(), - self.updated, - NoteType::Note, - ) - } -} + #[tracing::instrument(skip_all)] + async fn verify( + note: &Note, + expected_domain: &Url, + context: &Data, + ) -> Result<(), LemmyError> { + verify_domains_match(note.id.inner(), expected_domain)?; + verify_domains_match(note.attributed_to.inner(), note.id.inner())?; + verify_is_public(¬e.to, ¬e.cc)?; + let community = note.community(context).await?; -#[async_trait::async_trait(?Send)] -impl FromApub for Comment { - type ApubType = Note; + check_apub_id_valid_with_strictness(note.id.inner(), community.local, context).await?; + verify_is_remote_object(note.id.inner(), context.settings())?; + verify_person_in_community(¬e.attributed_to, &community, context).await?; + let (post, _) = note.get_parents(context).await?; + if post.locked { + return Err(LemmyErrorType::PostIsLocked)?; + } + Ok(()) + } /// Converts a `Note` to `Comment`. /// /// If the parent community, post and comment(s) are not known locally, these are also fetched. - async fn from_apub( - note: &Note, - context: &LemmyContext, - expected_domain: &Url, - request_counter: &mut i32, - ) -> Result { - let ap_id = Some(note.id(expected_domain)?.clone().into()); - let creator = note - .attributed_to - .dereference(context, request_counter) - .await?; - let (post, parent_comment_id) = note.get_parents(context, request_counter).await?; - if post.locked { - return Err(anyhow!("Post is locked").into()); - } + #[tracing::instrument(skip_all)] + async fn from_json(note: Note, context: &Data) -> Result { + let creator = note.attributed_to.dereference(context).await?; + let (post, parent_comment) = note.get_parents(context).await?; - let content = ¬e.source.content; - let content_slurs_removed = remove_slurs(content, &context.settings().slur_regex()); + let content = read_from_string_or_source(¬e.content, ¬e.media_type, ¬e.source); - let form = CommentForm { + let local_site = LocalSite::read(&mut context.pool()).await.ok(); + let slur_regex = &local_site_opt_to_slur_regex(&local_site); + let content_slurs_removed = remove_slurs(&content, slur_regex); + let language_id = + LanguageTag::to_language_id_single(note.language, &mut context.pool()).await?; + + let form = CommentInsertForm { creator_id: creator.id, post_id: post.id, - parent_id: parent_comment_id, content: content_slurs_removed, removed: None, - read: None, - published: Some(note.published.naive_local()), - updated: note.updated.map(|u| u.to_owned().naive_local()), - deleted: None, - ap_id, + published: note.published.map(|u| u.naive_local()), + updated: note.updated.map(|u| u.naive_local()), + deleted: Some(false), + ap_id: Some(note.id.into()), + distinguished: note.distinguished, local: Some(false), + language_id, }; - Ok(blocking(context.pool(), move |conn| Comment::upsert(conn, &form)).await??) + let parent_comment_path = parent_comment.map(|t| t.0.path); + let comment = Comment::create(&mut context.pool(), &form, parent_comment_path.as_ref()).await?; + Ok(comment.into()) + } +} + +#[cfg(test)] +pub(crate) mod tests { + #![allow(clippy::unwrap_used)] + #![allow(clippy::indexing_slicing)] + + use super::*; + use crate::{ + objects::{ + community::{tests::parse_lemmy_community, ApubCommunity}, + instance::ApubSite, + person::{tests::parse_lemmy_person, ApubPerson}, + post::ApubPost, + tests::init_context, + }, + protocol::tests::file_to_json_object, + }; + use assert_json_diff::assert_json_include; + use html2md::parse_html; + use lemmy_db_schema::source::site::Site; + use serial_test::serial; + + async fn prepare_comment_test( + url: &Url, + context: &Data, + ) -> (ApubPerson, ApubCommunity, ApubPost, ApubSite) { + // use separate counter so this doesnt affect tests + let context2 = context.reset_request_count(); + let (person, site) = parse_lemmy_person(&context2).await; + let community = parse_lemmy_community(&context2).await; + let post_json = file_to_json_object("assets/lemmy/objects/page.json").unwrap(); + ApubPost::verify(&post_json, url, &context2).await.unwrap(); + let post = ApubPost::from_json(post_json, &context2).await.unwrap(); + (person, community, post, site) + } + + async fn cleanup(data: (ApubPerson, ApubCommunity, ApubPost, ApubSite), context: &LemmyContext) { + Post::delete(&mut context.pool(), data.2.id).await.unwrap(); + Community::delete(&mut context.pool(), data.1.id) + .await + .unwrap(); + Person::delete(&mut context.pool(), data.0.id) + .await + .unwrap(); + Site::delete(&mut context.pool(), data.3.id).await.unwrap(); + LocalSite::delete(&mut context.pool()).await.unwrap(); + } + + #[tokio::test] + #[serial] + pub(crate) async fn test_parse_lemmy_comment() { + let context = init_context().await; + let url = Url::parse("https://enterprise.lemmy.ml/comment/38741").unwrap(); + let data = prepare_comment_test(&url, &context).await; + + let json: Note = file_to_json_object("assets/lemmy/objects/note.json").unwrap(); + ApubComment::verify(&json, &url, &context).await.unwrap(); + let comment = ApubComment::from_json(json.clone(), &context) + .await + .unwrap(); + + assert_eq!(comment.ap_id, url.into()); + assert_eq!(comment.content.len(), 14); + assert!(!comment.local); + assert_eq!(context.request_count(), 0); + + let comment_id = comment.id; + let to_apub = comment.into_json(&context).await.unwrap(); + assert_json_include!(actual: json, expected: to_apub); + + Comment::delete(&mut context.pool(), comment_id) + .await + .unwrap(); + cleanup(data, &context).await; + } + + #[tokio::test] + #[serial] + async fn test_parse_pleroma_comment() { + let context = init_context().await; + let url = Url::parse("https://enterprise.lemmy.ml/comment/38741").unwrap(); + let data = prepare_comment_test(&url, &context).await; + + let pleroma_url = + Url::parse("https://queer.hacktivis.me/objects/8d4973f4-53de-49cd-8c27-df160e16a9c2") + .unwrap(); + let person_json = file_to_json_object("assets/pleroma/objects/person.json").unwrap(); + ApubPerson::verify(&person_json, &pleroma_url, &context) + .await + .unwrap(); + ApubPerson::from_json(person_json, &context).await.unwrap(); + let json = file_to_json_object("assets/pleroma/objects/note.json").unwrap(); + ApubComment::verify(&json, &pleroma_url, &context) + .await + .unwrap(); + let comment = ApubComment::from_json(json, &context).await.unwrap(); + + assert_eq!(comment.ap_id, pleroma_url.into()); + assert_eq!(comment.content.len(), 64); + assert!(!comment.local); + assert_eq!(context.request_count(), 1); + + Comment::delete(&mut context.pool(), comment.id) + .await + .unwrap(); + cleanup(data, &context).await; + } + + #[tokio::test] + #[serial] + async fn test_html_to_markdown_sanitize() { + let parsed = parse_html("hello"); + assert_eq!(parsed, "**hello**"); } }