pull/3420/head
dull b 2023-06-30 03:58:51 +00:00
parent a8adf2dfb7
commit 2c479314ce
90 changed files with 2162 additions and 398 deletions

2
.gitattributes vendored 100644
View File

@ -0,0 +1,2 @@
# Normalize EOL for all files that Git considers text files.
* text=auto eol=lf

View File

@ -60,6 +60,9 @@ pipeline:
-D clippy::unused_self
-A clippy::uninlined_format_args
-D clippy::get_first
-D clippy::explicit_into_iter_loop
-D clippy::explicit_iter_loop
-D clippy::needless_collect
- cargo clippy --workspace --features console --
-D clippy::unwrap_used
-D clippy::indexing_slicing

209
Cargo.lock generated
View File

@ -14,9 +14,9 @@ dependencies = [
[[package]]
name = "activitypub_federation"
version = "0.4.4"
version = "0.4.5"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "27540f6c4b72c91176610ed5279061a021387f972c7c6f42c41032b78a808267"
checksum = "4ab3ac148d9c0b4163a6d41040c17de7558a42224b9ecbd4e8f033aef6c254d9"
dependencies = [
"activitystreams-kinds",
"actix-web",
@ -122,10 +122,12 @@ dependencies = [
"ahash 0.8.3",
"base64 0.21.2",
"bitflags 1.3.2",
"brotli",
"bytes",
"bytestring",
"derive_more",
"encoding_rs",
"flate2",
"futures-core",
"h2",
"http",
@ -143,6 +145,7 @@ dependencies = [
"tokio",
"tokio-util 0.7.4",
"tracing",
"zstd",
]
[[package]]
@ -215,7 +218,7 @@ dependencies = [
"futures-util",
"mio",
"num_cpus",
"socket2",
"socket2 0.4.9",
"tokio",
"tracing",
]
@ -245,7 +248,7 @@ dependencies = [
"http",
"log",
"pin-project-lite",
"tokio-rustls",
"tokio-rustls 0.23.4",
"tokio-util 0.7.4",
"webpki-roots",
]
@ -297,7 +300,7 @@ dependencies = [
"serde_json",
"serde_urlencoded",
"smallvec",
"socket2",
"socket2 0.4.9",
"time 0.3.15",
"url",
]
@ -361,6 +364,21 @@ dependencies = [
"memchr",
]
[[package]]
name = "alloc-no-stdlib"
version = "2.0.4"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "cc7bb162ec39d46ab1ca8c77bf72e890535becd1751bb45f64c597edb4c8c6b3"
[[package]]
name = "alloc-stdlib"
version = "0.2.2"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "94fb8275041c72129eb51b7d0322c29b8387a0386127718b096429201a5d6ece"
dependencies = [
"alloc-no-stdlib",
]
[[package]]
name = "android-tzdata"
version = "0.1.1"
@ -496,7 +514,7 @@ dependencies = [
"percent-encoding",
"pin-project-lite",
"rand 0.8.5",
"rustls",
"rustls 0.20.7",
"serde",
"serde_json",
"serde_urlencoded",
@ -714,6 +732,27 @@ dependencies = [
"cipher",
]
[[package]]
name = "brotli"
version = "3.3.4"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "a1a0b1dbcc8ae29329621f8d4f0d835787c1c38bb1401979b49d13b0b305ff68"
dependencies = [
"alloc-no-stdlib",
"alloc-stdlib",
"brotli-decompressor",
]
[[package]]
name = "brotli-decompressor"
version = "2.3.4"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "4b6561fd3f895a11e8f72af2cb7d22e08366bebc2b6b57f7744c4bda27034744"
dependencies = [
"alloc-no-stdlib",
"alloc-stdlib",
]
[[package]]
name = "bumpalo"
version = "3.11.1"
@ -766,6 +805,9 @@ name = "cc"
version = "1.0.73"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "2fff2a6927b3bb87f9595d67196a70493f627687a71d87a0d692242c33f58c11"
dependencies = [
"jobserver",
]
[[package]]
name = "cesu8"
@ -1328,6 +1370,7 @@ dependencies = [
"itoa",
"pq-sys",
"serde_json",
"uuid",
]
[[package]]
@ -2262,7 +2305,7 @@ dependencies = [
"httpdate",
"itoa",
"pin-project-lite",
"socket2",
"socket2 0.4.9",
"tokio",
"tower-service",
"tracing",
@ -2470,6 +2513,15 @@ version = "0.3.0"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "8eaf4bc02d17cbdd7ff4c7438cafcdf7fb9a4613313ad11b4f8fefe7d3fa0130"
[[package]]
name = "jobserver"
version = "0.1.26"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "936cfd212a0155903bcbc060e316fb6cc7cbf2e1907329391ebadc1fe0ce77c2"
dependencies = [
"libc",
]
[[package]]
name = "js-sys"
version = "0.3.60"
@ -2518,7 +2570,7 @@ checksum = "e2abad23fbc42b3700f2f279844dc832adb2b2eb069b2df918f455c4e18cc646"
[[package]]
name = "lemmy_api"
version = "0.18.0"
version = "0.18.1-rc.4"
dependencies = [
"actix-web",
"anyhow",
@ -2526,6 +2578,7 @@ dependencies = [
"base64 0.13.1",
"bcrypt",
"captcha",
"chrono",
"lemmy_api_common",
"lemmy_db_schema",
"lemmy_db_views",
@ -2541,9 +2594,8 @@ dependencies = [
[[package]]
name = "lemmy_api_common"
version = "0.18.0"
version = "0.18.1-rc.4"
dependencies = [
"actix-rt",
"actix-web",
"anyhow",
"chrono",
@ -2561,6 +2613,7 @@ dependencies = [
"rosetta-i18n",
"serde",
"serde_with",
"tokio",
"tracing",
"ts-rs",
"url",
@ -2570,12 +2623,13 @@ dependencies = [
[[package]]
name = "lemmy_api_crud"
version = "0.18.0"
version = "0.18.1-rc.4"
dependencies = [
"activitypub_federation",
"actix-web",
"async-trait",
"bcrypt",
"chrono",
"lemmy_api_common",
"lemmy_db_schema",
"lemmy_db_views",
@ -2584,15 +2638,15 @@ dependencies = [
"serde",
"tracing",
"url",
"uuid",
"webmention",
]
[[package]]
name = "lemmy_apub"
version = "0.18.0"
version = "0.18.1-rc.4"
dependencies = [
"activitypub_federation",
"actix-rt",
"actix-web",
"anyhow",
"assert-json-diff",
@ -2620,6 +2674,7 @@ dependencies = [
"sha2",
"strum_macros",
"task-local-extensions",
"tokio",
"tracing",
"url",
"uuid",
@ -2627,7 +2682,7 @@ dependencies = [
[[package]]
name = "lemmy_db_schema"
version = "0.18.0"
version = "0.18.1-rc.4"
dependencies = [
"activitypub_federation",
"async-trait",
@ -2640,9 +2695,11 @@ dependencies = [
"diesel-derive-newtype",
"diesel_ltree",
"diesel_migrations",
"futures-util",
"lemmy_utils",
"once_cell",
"regex",
"rustls 0.21.2",
"serde",
"serde_json",
"serde_with",
@ -2651,15 +2708,18 @@ dependencies = [
"strum",
"strum_macros",
"tokio",
"tokio-postgres",
"tokio-postgres-rustls",
"tracing",
"ts-rs",
"typed-builder",
"url",
"uuid",
]
[[package]]
name = "lemmy_db_views"
version = "0.18.0"
version = "0.18.1-rc.4"
dependencies = [
"diesel",
"diesel-async",
@ -2676,7 +2736,7 @@ dependencies = [
[[package]]
name = "lemmy_db_views_actor"
version = "0.18.0"
version = "0.18.1-rc.4"
dependencies = [
"diesel",
"diesel-async",
@ -2689,7 +2749,7 @@ dependencies = [
[[package]]
name = "lemmy_db_views_moderator"
version = "0.18.0"
version = "0.18.1-rc.4"
dependencies = [
"diesel",
"diesel-async",
@ -2701,7 +2761,7 @@ dependencies = [
[[package]]
name = "lemmy_routes"
version = "0.18.0"
version = "0.18.1-rc.4"
dependencies = [
"activitypub_federation",
"actix-web",
@ -2726,16 +2786,18 @@ dependencies = [
[[package]]
name = "lemmy_server"
version = "0.18.0"
version = "0.18.1-rc.4"
dependencies = [
"activitypub_federation",
"actix-cors",
"actix-web",
"chrono",
"clokwerk",
"console-subscriber",
"diesel",
"diesel-async",
"doku",
"futures-util",
"lemmy_api",
"lemmy_api_common",
"lemmy_api_crud",
@ -2749,9 +2811,12 @@ dependencies = [
"reqwest",
"reqwest-middleware",
"reqwest-tracing",
"rustls 0.21.2",
"serde",
"serde_json",
"tokio",
"tokio-postgres",
"tokio-postgres-rustls",
"tracing",
"tracing-actix-web 0.6.2",
"tracing-error",
@ -2763,7 +2828,7 @@ dependencies = [
[[package]]
name = "lemmy_utils"
version = "0.18.0"
version = "0.18.1-rc.4"
dependencies = [
"actix-web",
"anyhow",
@ -2820,7 +2885,7 @@ dependencies = [
"nom 7.1.1",
"once_cell",
"quoted_printable",
"socket2",
"socket2 0.4.9",
]
[[package]]
@ -3932,11 +3997,11 @@ dependencies = [
[[package]]
name = "postgres-protocol"
version = "0.6.4"
version = "0.6.5"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "878c6cbf956e03af9aa8204b407b9cbf47c072164800aa918c516cd4b056c50c"
checksum = "78b7fa9f396f51dffd61546fd8573ee20592287996568e6175ceb0f8699ad75d"
dependencies = [
"base64 0.13.1",
"base64 0.21.2",
"byteorder",
"bytes",
"fallible-iterator",
@ -4495,6 +4560,28 @@ dependencies = [
"webpki",
]
[[package]]
name = "rustls"
version = "0.21.2"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "e32ca28af694bc1bbf399c33a516dbdf1c90090b8ab23c2bc24f834aa2247f5f"
dependencies = [
"log",
"ring",
"rustls-webpki",
"sct",
]
[[package]]
name = "rustls-webpki"
version = "0.100.1"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "d6207cd5ed3d8dca7816f8f3725513a34609c0c765bf652b8c3cb4cfd87db46b"
dependencies = [
"ring",
"untrusted",
]
[[package]]
name = "rustversion"
version = "1.0.9"
@ -4859,6 +4946,16 @@ dependencies = [
"winapi",
]
[[package]]
name = "socket2"
version = "0.5.3"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "2538b18701741680e0322a2302176d3253a35388e2e62f172f64f4f16605f877"
dependencies = [
"libc",
"windows-sys 0.48.0",
]
[[package]]
name = "spin"
version = "0.5.2"
@ -5304,7 +5401,7 @@ dependencies = [
"parking_lot 0.12.1",
"pin-project-lite",
"signal-hook-registry",
"socket2",
"socket2 0.4.9",
"tokio-macros",
"tracing",
"windows-sys 0.48.0",
@ -5343,9 +5440,9 @@ dependencies = [
[[package]]
name = "tokio-postgres"
version = "0.7.7"
version = "0.7.8"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "29a12c1b3e0704ae7dfc25562629798b29c72e6b1d0a681b6f29ab4ae5e7f7bf"
checksum = "6e89f6234aa8fd43779746012fcf53603cdb91fdd8399aa0de868c2d56b6dde1"
dependencies = [
"async-trait",
"byteorder",
@ -5360,22 +5457,46 @@ dependencies = [
"pin-project-lite",
"postgres-protocol",
"postgres-types",
"socket2",
"socket2 0.5.3",
"tokio",
"tokio-util 0.7.4",
]
[[package]]
name = "tokio-postgres-rustls"
version = "0.10.0"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "dd5831152cb0d3f79ef5523b357319ba154795d64c7078b2daa95a803b54057f"
dependencies = [
"futures",
"ring",
"rustls 0.21.2",
"tokio",
"tokio-postgres",
"tokio-rustls 0.24.1",
]
[[package]]
name = "tokio-rustls"
version = "0.23.4"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "c43ee83903113e03984cb9e5cebe6c04a5116269e900e3ddba8f068a62adda59"
dependencies = [
"rustls",
"rustls 0.20.7",
"tokio",
"webpki",
]
[[package]]
name = "tokio-rustls"
version = "0.24.1"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "c28327cf380ac148141087fbfb9de9d7bd4e84ab5d2c28fbc911d753de8a7081"
dependencies = [
"rustls 0.21.2",
"tokio",
]
[[package]]
name = "tokio-stream"
version = "0.1.11"
@ -6460,3 +6581,33 @@ name = "zeroize"
version = "1.5.7"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "c394b5bd0c6f669e7275d9c20aa90ae064cb22e75a1cad54e1b34088034b149f"
[[package]]
name = "zstd"
version = "0.12.3+zstd.1.5.2"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "76eea132fb024e0e13fd9c2f5d5d595d8a967aa72382ac2f9d39fcc95afd0806"
dependencies = [
"zstd-safe",
]
[[package]]
name = "zstd-safe"
version = "6.0.5+zstd.1.5.4"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "d56d9e60b4b1758206c238a10165fbcae3ca37b01744e394c463463f6529d23b"
dependencies = [
"libc",
"zstd-sys",
]
[[package]]
name = "zstd-sys"
version = "2.0.8+zstd.1.5.5"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "5556e6ee25d32df2586c098bbfa278803692a20d0ab9565e049480d52707ec8c"
dependencies = [
"cc",
"libc",
"pkg-config",
]

View File

@ -1,5 +1,5 @@
[workspace.package]
version = "0.18.0"
version = "0.18.1-rc.4"
edition = "2021"
description = "A link aggregator for the fediverse"
license = "AGPL-3.0"
@ -49,23 +49,23 @@ members = [
]
[workspace.dependencies]
lemmy_api = { version = "=0.18.0", path = "./crates/api" }
lemmy_api_crud = { version = "=0.18.0", path = "./crates/api_crud" }
lemmy_apub = { version = "=0.18.0", path = "./crates/apub" }
lemmy_utils = { version = "=0.18.0", path = "./crates/utils" }
lemmy_db_schema = { version = "=0.18.0", path = "./crates/db_schema" }
lemmy_api_common = { version = "=0.18.0", path = "./crates/api_common" }
lemmy_routes = { version = "=0.18.0", path = "./crates/routes" }
lemmy_db_views = { version = "=0.18.0", path = "./crates/db_views" }
lemmy_db_views_actor = { version = "=0.18.0", path = "./crates/db_views_actor" }
lemmy_db_views_moderator = { version = "=0.18.0", path = "./crates/db_views_moderator" }
lemmy_api = { version = "=0.18.1-rc.4", path = "./crates/api" }
lemmy_api_crud = { version = "=0.18.1-rc.4", path = "./crates/api_crud" }
lemmy_apub = { version = "=0.18.1-rc.4", path = "./crates/apub" }
lemmy_utils = { version = "=0.18.1-rc.4", path = "./crates/utils" }
lemmy_db_schema = { version = "=0.18.1-rc.4", path = "./crates/db_schema" }
lemmy_api_common = { version = "=0.18.1-rc.4", path = "./crates/api_common" }
lemmy_routes = { version = "=0.18.1-rc.4", path = "./crates/routes" }
lemmy_db_views = { version = "=0.18.1-rc.4", path = "./crates/db_views" }
lemmy_db_views_actor = { version = "=0.18.1-rc.4", path = "./crates/db_views_actor" }
lemmy_db_views_moderator = { version = "=0.18.1-rc.4", path = "./crates/db_views_moderator" }
activitypub_federation = { version = "0.4.4", default-features = false, features = ["actix-web"] }
diesel = "2.1.0"
diesel_migrations = "2.1.0"
diesel-async = "0.3.1"
serde = { version = "1.0.164", features = ["derive"] }
serde_with = "1.14.0"
actix-web = { version = "4.3.1", default-features = false, features = ["macros", "rustls"] }
actix-web = { version = "4.3.1", default-features = false, features = ["macros", "rustls", "compress-brotli", "compress-gzip", "compress-zstd"] }
tracing = "0.1.37"
tracing-actix-web = { version = "0.6.2", default-features = false }
tracing-error = "0.2.0"
@ -89,7 +89,7 @@ anyhow = "1.0.71"
diesel_ltree = "0.3.0"
typed-builder = "0.10.0"
serial_test = "0.9.0"
tokio = "1.28.2"
tokio = { version = "1.28.2", features = ["full"] }
sha2 = "0.10.6"
regex = "1.8.4"
once_cell = "1.18.0"
@ -100,13 +100,16 @@ strum_macros = "0.24.3"
itertools = "0.10.5"
futures = "0.3.28"
http = "0.2.9"
actix-rt = { version = "2.8.0", default-features = false }
percent-encoding = "2.3.0"
rosetta-i18n = "0.1.2"
rand = "0.8.5"
opentelemetry = { version = "0.17.0", features = ["rt-tokio"] }
tracing-opentelemetry = { version = "0.17.4" }
ts-rs = { version = "6.2", features = ["serde-compat", "format", "chrono-impl"] }
rustls = { version ="0.21.2", features = ["dangerous_configuration"]}
futures-util = "0.3.28"
tokio-postgres = "0.7.8"
tokio-postgres-rustls = "0.10.0"
[dependencies]
lemmy_api = { workspace = true }
@ -140,3 +143,8 @@ opentelemetry-otlp = { version = "0.10.0", optional = true }
pict-rs = { version = "0.4.0-rc.3", optional = true }
tokio.workspace = true
actix-cors = "0.6.4"
rustls = { workspace = true }
futures-util = { workspace = true }
tokio-postgres = { workspace = true }
tokio-postgres-rustls = { workspace = true }
chrono = { workspace = true }

View File

@ -123,6 +123,7 @@ Each Lemmy server can set its own moderation policy; appointing site-wide admins
- [Jerboa - A native Android app made by Lemmy's developers](https://github.com/dessalines/jerboa)
- [Mlem - A Lemmy client for iOS](https://github.com/buresdv/Mlem)
- [Lemoa - A Gtk client for Lemmy on Linux](https://github.com/lemmy-gtk/lemoa)
- [Liftoff - A Lemmy for Windows , Linux and Android ](https://github.com/liftoff-app/liftoff)
### Libraries

View File

@ -76,4 +76,8 @@
port: 8536
# Whether the site is available over TLS. Needs to be true for federation to work.
tls_enabled: true
# The number of activitypub federation workers that can be in-flight concurrently
worker_count: 0
# The number of activitypub federation retry workers that can be in-flight concurrently
retry_count: 0
}

View File

@ -29,6 +29,7 @@ async-trait = { workspace = true }
captcha = { workspace = true }
anyhow = { workspace = true }
tracing = { workspace = true }
chrono = { workspace = true }
[dev-dependencies]
serial_test = { workspace = true }

View File

@ -47,7 +47,7 @@ impl Perform for BanFromCommunity {
community_id,
)
.await?;
is_valid_body_field(&data.reason)?;
is_valid_body_field(&data.reason, false)?;
let community_user_ban_form = CommunityPersonBanForm {
community_id: data.community_id,

View File

@ -1,4 +1,5 @@
use actix_web::web::Data;
use captcha::Captcha;
use lemmy_api_common::{context::LemmyContext, utils::local_site_to_slur_regex};
use lemmy_db_schema::source::local_site::LocalSite;
use lemmy_utils::{error::LemmyError, utils::slurs::check_slurs};
@ -20,6 +21,21 @@ pub trait Perform {
async fn perform(&self, context: &Data<LemmyContext>) -> Result<Self::Response, LemmyError>;
}
/// Converts the captcha to a base64 encoded wav audio file
pub(crate) fn captcha_as_wav_base64(captcha: &Captcha) -> String {
let letters = captcha.as_wav();
let mut concat_letters: Vec<u8> = Vec::new();
for letter in letters {
let bytes = letter.unwrap_or_default();
concat_letters.extend(bytes);
}
// Convert to base64
base64::encode(concat_letters)
}
/// Check size of report and remove whitespace
pub(crate) fn check_report_reason(reason: &str, local_site: &LocalSite) -> Result<(), LemmyError> {
let slur_regex = &local_site_to_slur_regex(local_site);

View File

@ -30,7 +30,7 @@ impl Perform for BanPerson {
// Make sure user is an admin
is_admin(&local_user_view)?;
is_valid_body_field(&data.reason)?;
is_valid_body_field(&data.reason, false)?;
let ban = data.ban;
let banned_person_id = data.person_id;

View File

@ -0,0 +1,50 @@
use crate::{captcha_as_wav_base64, Perform};
use actix_web::web::Data;
use captcha::{gen, Difficulty};
use lemmy_api_common::{
context::LemmyContext,
person::{CaptchaResponse, GetCaptcha, GetCaptchaResponse},
};
use lemmy_db_schema::source::{
captcha_answer::{CaptchaAnswer, CaptchaAnswerForm},
local_site::LocalSite,
};
use lemmy_utils::error::LemmyError;
#[async_trait::async_trait(?Send)]
impl Perform for GetCaptcha {
type Response = GetCaptchaResponse;
#[tracing::instrument(skip(context))]
async fn perform(&self, context: &Data<LemmyContext>) -> Result<Self::Response, LemmyError> {
let local_site = LocalSite::read(context.pool()).await?;
if !local_site.captcha_enabled {
return Ok(GetCaptchaResponse { ok: None });
}
let captcha = gen(match local_site.captcha_difficulty.as_str() {
"easy" => Difficulty::Easy,
"hard" => Difficulty::Hard,
_ => Difficulty::Medium,
});
let answer = captcha.chars_as_string();
let png = captcha.as_base64().expect("failed to generate captcha");
let wav = captcha_as_wav_base64(&captcha);
let captcha_form: CaptchaAnswerForm = CaptchaAnswerForm { answer };
// Stores the captcha item in the db
let captcha = CaptchaAnswer::insert(context.pool(), &captcha_form).await?;
Ok(GetCaptchaResponse {
ok: Some(CaptchaResponse {
png,
wav,
uuid: captcha.uuid.to_string(),
}),
})
}
}

View File

@ -3,6 +3,7 @@ mod ban_person;
mod block;
mod change_password;
mod change_password_after_reset;
mod get_captcha;
mod list_banned;
mod login;
mod notifications;

View File

@ -5,6 +5,7 @@ use lemmy_api_common::{
person::{PasswordReset, PasswordResetResponse},
utils::send_password_reset_email,
};
use lemmy_db_schema::source::password_reset_request::PasswordResetRequest;
use lemmy_db_views::structs::LocalUserView;
use lemmy_utils::error::LemmyError;
@ -25,6 +26,16 @@ impl Perform for PasswordReset {
.await
.map_err(|e| LemmyError::from_error_message(e, "couldnt_find_that_username_or_email"))?;
// Check for too many attempts (to limit potential abuse)
let recent_resets_count = PasswordResetRequest::get_recent_password_resets_count(
context.pool(),
local_user_view.local_user.id,
)
.await?;
if recent_resets_count >= 3 {
return Err(LemmyError::from_message("password_reset_limit_reached"));
}
// Email the pure token to the user.
send_password_reset_email(
&local_user_view,

View File

@ -3,7 +3,7 @@ use actix_web::web::Data;
use lemmy_api_common::{
context::LemmyContext,
site::{PurgeComment, PurgeItemResponse},
utils::{is_top_admin, local_user_view_from_jwt},
utils::{is_admin, local_user_view_from_jwt},
};
use lemmy_db_schema::{
source::{
@ -23,8 +23,8 @@ impl Perform for PurgeComment {
let data: &Self = self;
let local_user_view = local_user_view_from_jwt(&data.auth, context).await?;
// Only let the top admin purge an item
is_top_admin(&mut *context.conn().await?, local_user_view.person.id).await?;
// Only let admin purge an item
is_admin(&local_user_view)?;
let comment_id = data.comment_id;

View File

@ -4,7 +4,7 @@ use lemmy_api_common::{
context::LemmyContext,
request::purge_image_from_pictrs,
site::{PurgeCommunity, PurgeItemResponse},
utils::{is_top_admin, local_user_view_from_jwt, purge_image_posts_for_community},
utils::{is_admin, local_user_view_from_jwt, purge_image_posts_for_community},
};
use lemmy_db_schema::{
source::{
@ -24,8 +24,8 @@ impl Perform for PurgeCommunity {
let data: &Self = self;
let local_user_view = local_user_view_from_jwt(&data.auth, context).await?;
// Only let the top admin purge an item
is_top_admin(&mut *context.conn().await?, local_user_view.person.id).await?;
// Only let admin purge an item
is_admin(&local_user_view)?;
let community_id = data.community_id;

View File

@ -4,7 +4,7 @@ use lemmy_api_common::{
context::LemmyContext,
request::purge_image_from_pictrs,
site::{PurgeItemResponse, PurgePerson},
utils::{is_top_admin, local_user_view_from_jwt, purge_image_posts_for_person},
utils::{is_admin, local_user_view_from_jwt, purge_image_posts_for_person},
};
use lemmy_db_schema::{
source::{
@ -24,8 +24,8 @@ impl Perform for PurgePerson {
let data: &Self = self;
let local_user_view = local_user_view_from_jwt(&data.auth, context).await?;
// Only let the top admin purge an item
is_top_admin(&mut *context.conn().await?, local_user_view.person.id).await?;
// Only let admin purge an item
is_admin(&local_user_view)?;
// Read the person to get their images
let person_id = data.person_id;

View File

@ -4,7 +4,7 @@ use lemmy_api_common::{
context::LemmyContext,
request::purge_image_from_pictrs,
site::{PurgeItemResponse, PurgePost},
utils::{is_top_admin, local_user_view_from_jwt},
utils::{is_admin, local_user_view_from_jwt},
};
use lemmy_db_schema::{
source::{
@ -24,8 +24,8 @@ impl Perform for PurgePost {
let data: &Self = self;
let local_user_view = local_user_view_from_jwt(&data.auth, context).await?;
// Only let the top admin purge an item
is_top_admin(&mut *context.conn().await?, local_user_view.person.id).await?;
// Only let admin purge an item
is_admin(&local_user_view)?;
let post_id = data.post_id;

View File

@ -33,12 +33,12 @@ reqwest-middleware = { workspace = true, optional = true }
regex = { workspace = true }
rosetta-i18n = { workspace = true, optional = true }
percent-encoding = { workspace = true, optional = true }
webpage = { version = "1.6.0", default-features = false, features = ["serde"], optional = true }
webpage = { version = "1.6", default-features = false, features = ["serde"], optional = true }
encoding = { version = "0.2.33", optional = true }
anyhow = { workspace = true }
futures = { workspace = true }
uuid = { workspace = true }
actix-rt = { workspace = true }
tokio = { workspace = true }
reqwest = { workspace = true }
ts-rs = { workspace = true, optional = true }
actix-web = { workspace = true }

View File

@ -24,3 +24,10 @@ As you can see, each API endpoint needs a parameter type ( GetPosts), path (/pos
For a real example of a Lemmy API client, look at [lemmyBB](https://github.com/LemmyNet/lemmyBB/tree/main/src/api).
Lemmy also provides a websocket API. You can find the full websocket code in [this file](https://github.com/LemmyNet/lemmy/blob/main/src/api_routes_websocket.rs).
## Generate TypeScript bindings
TypeScript bindings (API types) can be generated by running `cargo test --features full`.
The ts files be generated into a `bindings` folder.
This crate uses [`ts_rs`](https://docs.rs/ts-rs/6.2.1/ts_rs/#traits) macros `derive(TS)` and `ts(export)` to attribute types for binding generating.

View File

@ -103,7 +103,6 @@ pub async fn send_local_notifs(
for mention in mentions
.iter()
.filter(|m| m.is_local(&context.settings().hostname) && m.name.ne(&person.name))
.collect::<Vec<&MentionData>>()
{
let mention_name = mention.name.clone();
let user_view = LocalUserView::read_from_name(&mut *context.conn().await?, &mention_name).await;

View File

@ -76,6 +76,7 @@ pub struct CommunityResponse {
pub struct ListCommunities {
pub type_: Option<ListingType>,
pub sort: Option<SortType>,
pub show_nsfw: Option<bool>,
pub page: Option<i64>,
pub limit: Option<i64>,
pub auth: Option<Sensitive<String>>,

View File

@ -27,12 +27,12 @@ pub async fn fetch_site_metadata(
// https://github.com/LemmyNet/lemmy/issues/1964
let html_bytes = response.bytes().await.map_err(LemmyError::from)?.to_vec();
let tags = html_to_site_metadata(&html_bytes)?;
let tags = html_to_site_metadata(&html_bytes, url)?;
Ok(tags)
}
fn html_to_site_metadata(html_bytes: &[u8]) -> Result<SiteMetadata, LemmyError> {
fn html_to_site_metadata(html_bytes: &[u8], url: &Url) -> Result<SiteMetadata, LemmyError> {
let html = String::from_utf8_lossy(html_bytes);
// Make sure the first line is doctype html
@ -81,12 +81,14 @@ fn html_to_site_metadata(html_bytes: &[u8]) -> Result<SiteMetadata, LemmyError>
.opengraph
.images
.first()
.and_then(|ogo| Url::parse(&ogo.url).ok());
// join also works if the target URL is absolute
.and_then(|ogo| url.join(&ogo.url).ok());
let og_embed_url = page
.opengraph
.videos
.first()
.and_then(|v| Url::parse(&v.url).ok());
// join also works if the target URL is absolute
.and_then(|v| url.join(&v.url).ok());
Ok(SiteMetadata {
title: og_title.or(page_title),
@ -266,12 +268,17 @@ pub fn build_user_agent(settings: &Settings) -> String {
#[cfg(test)]
mod tests {
use crate::request::{build_user_agent, fetch_site_metadata, SiteMetadata};
use crate::request::{
build_user_agent,
fetch_site_metadata,
html_to_site_metadata,
SiteMetadata,
};
use lemmy_utils::settings::SETTINGS;
use url::Url;
// These helped with testing
#[actix_rt::test]
#[tokio::test]
async fn test_site_metadata() {
let settings = &SETTINGS.clone();
let client = reqwest::Client::builder()
@ -305,4 +312,46 @@ mod tests {
// let res_other = fetch_pictshare("https://upload.wikimedia.org/wikipedia/en/2/27/The_Mandalorian_logo.jpgaoeu");
// assert!(res_other.is_err());
// }
#[test]
fn test_resolve_image_url() {
// url that lists the opengraph fields
let url = Url::parse("https://example.com/one/two.html").unwrap();
// root relative url
let html_bytes = b"<!DOCTYPE html><html><head><meta property='og:image' content='/image.jpg'></head><body></body></html>";
let metadata = html_to_site_metadata(html_bytes, &url).expect("Unable to parse metadata");
assert_eq!(
metadata.image,
Some(Url::parse("https://example.com/image.jpg").unwrap().into())
);
// base relative url
let html_bytes = b"<!DOCTYPE html><html><head><meta property='og:image' content='image.jpg'></head><body></body></html>";
let metadata = html_to_site_metadata(html_bytes, &url).expect("Unable to parse metadata");
assert_eq!(
metadata.image,
Some(
Url::parse("https://example.com/one/image.jpg")
.unwrap()
.into()
)
);
// absolute url
let html_bytes = b"<!DOCTYPE html><html><head><meta property='og:image' content='https://cdn.host.com/image.jpg'></head><body></body></html>";
let metadata = html_to_site_metadata(html_bytes, &url).expect("Unable to parse metadata");
assert_eq!(
metadata.image,
Some(Url::parse("https://cdn.host.com/image.jpg").unwrap().into())
);
// protocol relative url
let html_bytes = b"<!DOCTYPE html><html><head><meta property='og:image' content='//example.com/image.jpg'></head><body></body></html>";
let metadata = html_to_site_metadata(html_bytes, &url).expect("Unable to parse metadata");
assert_eq!(
metadata.image,
Some(Url::parse("https://example.com/image.jpg").unwrap().into())
);
}
}

View File

@ -177,7 +177,6 @@ pub struct CreateSite {
pub rate_limit_search_per_second: Option<i32>,
pub federation_enabled: Option<bool>,
pub federation_debug: Option<bool>,
pub federation_worker_count: Option<i32>,
pub captcha_enabled: Option<bool>,
pub captcha_difficulty: Option<String>,
pub allowed_instances: Option<Vec<String>>,
@ -250,8 +249,6 @@ pub struct EditSite {
pub federation_enabled: Option<bool>,
/// Enables federation debugging.
pub federation_debug: Option<bool>,
/// The number of federation workers.
pub federation_worker_count: Option<i32>,
/// Whether to enable captchas for signups.
pub captcha_enabled: Option<bool>,
/// The captcha difficulty. Can be easy, medium, or hard

View File

@ -32,7 +32,6 @@ use lemmy_db_views_actor::structs::{
CommunityModeratorView,
CommunityPersonBanView,
CommunityView,
PersonView,
};
use lemmy_utils::{
claims::Claims,
@ -79,18 +78,6 @@ pub async fn is_mod_or_admin_opt(
}
}
pub async fn is_top_admin(conn: &mut DbConn, person_id: PersonId) -> Result<(), LemmyError> {
let admins = PersonView::admins(conn).await?;
let top_admin = admins
.first()
.ok_or_else(|| LemmyError::from_message("no admins"))?;
if top_admin.person.id != person_id {
return Err(LemmyError::from_message("not_top_admin"));
}
Ok(())
}
pub fn is_admin(local_user_view: &LocalUserView) -> Result<(), LemmyError> {
if !local_user_view.person.admin {
return Err(LemmyError::from_message("not_an_admin"));
@ -316,15 +303,6 @@ pub fn password_length_check(pass: &str) -> Result<(), LemmyError> {
}
}
/// Checks the site description length
pub fn site_description_length_check(description: &str) -> Result<(), LemmyError> {
if description.len() > 150 {
Err(LemmyError::from_message("site_description_length_overflow"))
} else {
Ok(())
}
}
/// Checks for a honeypot. If this field is filled, fail the rest of the function
pub fn honeypot_check(honeypot: &Option<String>) -> Result<(), LemmyError> {
if honeypot.is_some() && honeypot != &Some(String::new()) {

View File

@ -22,3 +22,5 @@ tracing = { workspace = true }
url = { workspace = true }
async-trait = { workspace = true }
webmention = "0.4.0"
chrono = { workspace = true }
uuid = { workspace = true }

View File

@ -49,7 +49,7 @@ impl PerformCrud for CreateComment {
&data.content.clone(),
&local_site_to_slur_regex(&local_site),
);
is_valid_body_field(&Some(content_slurs_removed.clone()))?;
is_valid_body_field(&Some(content_slurs_removed.clone()), false)?;
// Check for a community ban
let post_id = data.post_id;
@ -207,7 +207,7 @@ impl PerformCrud for CreateComment {
pub fn check_comment_depth(comment: &Comment) -> Result<(), LemmyError> {
let path = &comment.path.0;
let length = path.split('.').collect::<Vec<&str>>().len();
let length = path.split('.').count();
if length > MAX_COMMENT_DEPTH_LIMIT {
Err(LemmyError::from_message("max_comment_depth_reached"))
} else {

View File

@ -64,7 +64,7 @@ impl PerformCrud for EditComment {
.as_ref()
.map(|c| remove_slurs(c, &local_site_to_slur_regex(&local_site)));
is_valid_body_field(&content_slurs_removed)?;
is_valid_body_field(&content_slurs_removed, false)?;
let comment_id = data.comment_id;
let form = CommentUpdateForm::builder()

View File

@ -67,7 +67,7 @@ impl PerformCrud for CreateCommunity {
check_slurs_opt(&data.description, &slur_regex)?;
is_valid_actor_name(&data.name, local_site.actor_name_max_length as usize)?;
is_valid_body_field(&data.description)?;
is_valid_body_field(&data.description, false)?;
// Double check for duplicate community actor_ids
let community_actor_id = generate_local_apub_endpoint(

View File

@ -27,6 +27,7 @@ impl PerformCrud for ListCommunities {
let sort = data.sort;
let listing_type = data.type_;
let show_nsfw = data.show_nsfw;
let page = data.page;
let limit = data.limit;
let local_user = local_user_view.map(|l| l.local_user);
@ -34,6 +35,7 @@ impl PerformCrud for ListCommunities {
let communities = CommunityQuery::builder()
.conn(&mut conn)
.listing_type(listing_type)
.show_nsfw(show_nsfw)
.sort(sort)
.local_user(local_user.as_ref())
.page(page)

View File

@ -39,7 +39,7 @@ impl PerformCrud for EditCommunity {
let slur_regex = local_site_to_slur_regex(&local_site);
check_slurs_opt(&data.title, &slur_regex)?;
check_slurs_opt(&data.description, &slur_regex)?;
is_valid_body_field(&data.description)?;
is_valid_body_field(&data.description, false)?;
// Verify its a mod (only mods can edit it)
let community_id = data.community_id;

View File

@ -57,7 +57,7 @@ impl PerformCrud for CreatePost {
let url = data_url.map(clean_url_params).map(Into::into); // TODO no good way to handle a "clear"
is_valid_post_title(&data.name)?;
is_valid_body_field(&data.body)?;
is_valid_body_field(&data.body, true)?;
check_community_ban(
local_user_view.person.id,

View File

@ -49,7 +49,7 @@ impl PerformCrud for EditPost {
is_valid_post_title(name)?;
}
is_valid_body_field(&data.body)?;
is_valid_body_field(&data.body, true)?;
let post_id = data.post_id;
let orig_post = Post::read(&mut *context.conn().await?, post_id).await?;

View File

@ -43,7 +43,7 @@ impl PerformCrud for CreatePrivateMessage {
&data.content.clone(),
&local_site_to_slur_regex(&local_site),
);
is_valid_body_field(&Some(content_slurs_removed.clone()))?;
is_valid_body_field(&Some(content_slurs_removed.clone()), false)?;
check_person_block(
local_user_view.person.id,

View File

@ -42,7 +42,7 @@ impl PerformCrud for EditPrivateMessage {
// Doing the update
let content_slurs_removed = remove_slurs(&data.content, &local_site_to_slur_regex(&local_site));
is_valid_body_field(&Some(content_slurs_removed.clone()))?;
is_valid_body_field(&Some(content_slurs_removed.clone()), false)?;
let private_message_id = data.private_message_id;
PrivateMessage::update(

View File

@ -1,4 +1,7 @@
use crate::{site::check_application_question, PerformCrud};
use crate::{
site::{application_question_check, site_default_post_listing_type_check},
PerformCrud,
};
use activitypub_federation::http_signatures::generate_actor_keypair;
use actix_web::web::Data;
use lemmy_api_common::{
@ -8,9 +11,7 @@ use lemmy_api_common::{
generate_site_inbox_url,
is_admin,
local_site_rate_limit_to_rate_limit_config,
local_site_to_slur_regex,
local_user_view_from_jwt,
site_description_length_check,
},
};
use lemmy_db_schema::{
@ -26,10 +27,16 @@ use lemmy_db_schema::{
};
use lemmy_db_views::structs::SiteView;
use lemmy_utils::{
error::LemmyError,
error::{LemmyError, LemmyResult},
utils::{
slurs::{check_slurs, check_slurs_opt},
validation::{check_site_visibility_valid, is_valid_body_field},
validation::{
build_and_check_regex,
check_site_visibility_valid,
is_valid_body_field,
site_description_length_check,
site_name_length_check,
},
},
};
use url::Url;
@ -42,56 +49,23 @@ impl PerformCrud for CreateSite {
async fn perform(&self, context: &Data<LemmyContext>) -> Result<SiteResponse, LemmyError> {
let data: &CreateSite = self;
let local_user_view = local_user_view_from_jwt(&data.auth, context).await?;
let local_site = LocalSite::read(&mut *context.conn().await?).await?;
if local_site.site_setup {
return Err(LemmyError::from_message("site_already_exists"));
};
let local_user_view = local_user_view_from_jwt(&data.auth, context).await?;
// Make sure user is an admin
// Make sure user is an admin; other types of users should not create site data...
is_admin(&local_user_view)?;
check_site_visibility_valid(
local_site.private_instance,
local_site.federation_enabled,
&data.private_instance,
&data.federation_enabled,
)?;
let sidebar = diesel_option_overwrite(&data.sidebar);
let description = diesel_option_overwrite(&data.description);
let icon = diesel_option_overwrite_to_url(&data.icon)?;
let banner = diesel_option_overwrite_to_url(&data.banner)?;
let slur_regex = local_site_to_slur_regex(&local_site);
check_slurs(&data.name, &slur_regex)?;
check_slurs_opt(&data.description, &slur_regex)?;
if let Some(Some(desc)) = &description {
site_description_length_check(desc)?;
}
is_valid_body_field(&data.sidebar)?;
let application_question = diesel_option_overwrite(&data.application_question);
check_application_question(
&application_question,
data
.registration_mode
.unwrap_or(local_site.registration_mode),
)?;
validate_create_payload(&local_site, data)?;
let actor_id: DbUrl = Url::parse(&context.settings().get_protocol_and_hostname())?.into();
let inbox_url = Some(generate_site_inbox_url(&actor_id)?);
let keypair = generate_actor_keypair()?;
let site_form = SiteUpdateForm::builder()
.name(Some(data.name.clone()))
.sidebar(sidebar)
.description(description)
.icon(icon)
.banner(banner)
.sidebar(diesel_option_overwrite(&data.sidebar))
.description(diesel_option_overwrite(&data.description))
.icon(diesel_option_overwrite_to_url(&data.icon)?)
.banner(diesel_option_overwrite_to_url(&data.banner)?)
.actor_id(Some(actor_id))
.last_refreshed_at(Some(naive_now()))
.inbox_url(inbox_url)
@ -111,7 +85,7 @@ impl PerformCrud for CreateSite {
.enable_nsfw(data.enable_nsfw)
.community_creation_admin_only(data.community_creation_admin_only)
.require_email_verification(data.require_email_verification)
.application_question(application_question)
.application_question(diesel_option_overwrite(&data.application_question))
.private_instance(data.private_instance)
.default_theme(data.default_theme.clone())
.default_post_listing_type(data.default_post_listing_type)
@ -122,7 +96,6 @@ impl PerformCrud for CreateSite {
.slur_filter_regex(diesel_option_overwrite(&data.slur_filter_regex))
.actor_name_max_length(data.actor_name_max_length)
.federation_enabled(data.federation_enabled)
.federation_worker_count(data.federation_worker_count)
.captcha_enabled(data.captcha_enabled)
.captcha_difficulty(data.captcha_difficulty.clone())
.build();
@ -165,3 +138,449 @@ impl PerformCrud for CreateSite {
})
}
}
fn validate_create_payload(local_site: &LocalSite, create_site: &CreateSite) -> LemmyResult<()> {
// Make sure the site hasn't already been set up...
if local_site.site_setup {
return Err(LemmyError::from_message("site_already_exists"));
};
// Check that the slur regex compiles, and returns the regex if valid...
// Prioritize using new slur regex from the request; if not provided, use the existing regex.
let slur_regex = build_and_check_regex(
&create_site
.slur_filter_regex
.as_deref()
.or(local_site.slur_filter_regex.as_deref()),
)?;
site_name_length_check(&create_site.name)?;
check_slurs(&create_site.name, &slur_regex)?;
if let Some(desc) = &create_site.description {
site_description_length_check(desc)?;
check_slurs_opt(&create_site.description, &slur_regex)?;
}
site_default_post_listing_type_check(&create_site.default_post_listing_type)?;
check_site_visibility_valid(
local_site.private_instance,
local_site.federation_enabled,
&create_site.private_instance,
&create_site.federation_enabled,
)?;
// Ensure that the sidebar has fewer than the max num characters...
is_valid_body_field(&create_site.sidebar, false)?;
application_question_check(
&local_site.application_question,
&create_site.application_question,
create_site
.registration_mode
.unwrap_or(local_site.registration_mode),
)
}
#[cfg(test)]
mod tests {
use crate::site::create::validate_create_payload;
use lemmy_api_common::site::CreateSite;
use lemmy_db_schema::{source::local_site::LocalSite, ListingType, RegistrationMode};
#[test]
fn test_validate_invalid_create_payload() {
let invalid_payloads = [
(
"CreateSite attempted on set up LocalSite",
"site_already_exists",
&generate_local_site(
true,
None::<String>,
true,
false,
None::<String>,
RegistrationMode::Open,
),
&generate_create_site(
String::from("site_name"),
None::<String>,
None::<String>,
None::<ListingType>,
None::<String>,
None::<bool>,
None::<bool>,
None::<String>,
None::<RegistrationMode>,
),
),
(
"CreateSite name matches LocalSite slur filter",
"slurs",
&generate_local_site(
false,
Some(String::from("(foo|bar)")),
true,
false,
None::<String>,
RegistrationMode::Open,
),
&generate_create_site(
String::from("foo site_name"),
None::<String>,
None::<String>,
None::<ListingType>,
None::<String>,
None::<bool>,
None::<bool>,
None::<String>,
None::<RegistrationMode>,
),
),
(
"CreateSite name matches new slur filter",
"slurs",
&generate_local_site(
false,
Some(String::from("(foo|bar)")),
true,
false,
None::<String>,
RegistrationMode::Open,
),
&generate_create_site(
String::from("zeta site_name"),
None::<String>,
None::<String>,
None::<ListingType>,
Some(String::from("(zeta|alpha)")),
None::<bool>,
None::<bool>,
None::<String>,
None::<RegistrationMode>,
),
),
(
"CreateSite listing type is Subscribed, which is invalid",
"invalid_default_post_listing_type",
&generate_local_site(
false,
None::<String>,
true,
false,
None::<String>,
RegistrationMode::Open,
),
&generate_create_site(
String::from("site_name"),
None::<String>,
None::<String>,
Some(ListingType::Subscribed),
None::<String>,
None::<bool>,
None::<bool>,
None::<String>,
None::<RegistrationMode>,
),
),
(
"CreateSite is both private and federated",
"cant_enable_private_instance_and_federation_together",
&generate_local_site(
false,
None::<String>,
true,
false,
None::<String>,
RegistrationMode::Open,
),
&generate_create_site(
String::from("site_name"),
None::<String>,
None::<String>,
None::<ListingType>,
None::<String>,
Some(true),
Some(true),
None::<String>,
None::<RegistrationMode>,
),
),
(
"LocalSite is private, but CreateSite also makes it federated",
"cant_enable_private_instance_and_federation_together",
&generate_local_site(
false,
None::<String>,
true,
false,
None::<String>,
RegistrationMode::Open,
),
&generate_create_site(
String::from("site_name"),
None::<String>,
None::<String>,
None::<ListingType>,
None::<String>,
None::<bool>,
Some(true),
None::<String>,
None::<RegistrationMode>,
),
),
(
"CreateSite requires application, but neither it nor LocalSite has an application question",
"application_question_required",
&generate_local_site(
false,
None::<String>,
true,
false,
None::<String>,
RegistrationMode::Open,
),
&generate_create_site(
String::from("site_name"),
None::<String>,
None::<String>,
None::<ListingType>,
None::<String>,
None::<bool>,
None::<bool>,
None::<String>,
Some(RegistrationMode::RequireApplication),
),
),
];
invalid_payloads.iter().enumerate().for_each(
|(
idx,
&(reason, expected_err, local_site, create_site),
)| {
match validate_create_payload(
local_site,
create_site,
) {
Ok(_) => {
panic!(
"Got Ok, but validation should have failed with error: {} for reason: {}. invalid_payloads.nth({})",
expected_err, reason, idx
)
}
Err(error) => {
assert!(
error.message.eq(&Some(String::from(expected_err))),
"Got Err {:?}, but should have failed with message: {} for reason: {}. invalid_payloads.nth({})",
error.message,
expected_err,
reason,
idx
)
}
}
},
);
}
#[test]
fn test_validate_valid_create_payload() {
let valid_payloads = [
(
"No changes between LocalSite and CreateSite",
&generate_local_site(
false,
None::<String>,
true,
false,
None::<String>,
RegistrationMode::Open,
),
&generate_create_site(
String::from("site_name"),
None::<String>,
None::<String>,
None::<ListingType>,
None::<String>,
None::<bool>,
None::<bool>,
None::<String>,
None::<RegistrationMode>,
),
),
(
"CreateSite allows clearing and changing values",
&generate_local_site(
false,
None::<String>,
true,
false,
None::<String>,
RegistrationMode::Open,
),
&generate_create_site(
String::from("site_name"),
Some(String::new()),
Some(String::new()),
Some(ListingType::All),
Some(String::new()),
Some(false),
Some(true),
Some(String::new()),
Some(RegistrationMode::Open),
),
),
(
"CreateSite clears existing slur filter regex",
&generate_local_site(
false,
Some(String::from("(foo|bar)")),
true,
false,
None::<String>,
RegistrationMode::Open,
),
&generate_create_site(
String::from("foo site_name"),
None::<String>,
None::<String>,
None::<ListingType>,
Some(String::new()),
None::<bool>,
None::<bool>,
None::<String>,
None::<RegistrationMode>,
),
),
(
"LocalSite has application question and CreateSite now requires applications,",
&generate_local_site(
false,
None::<String>,
true,
false,
Some(String::from("question")),
RegistrationMode::Open,
),
&generate_create_site(
String::from("site_name"),
None::<String>,
None::<String>,
None::<ListingType>,
None::<String>,
None::<bool>,
None::<bool>,
None::<String>,
Some(RegistrationMode::RequireApplication),
),
),
];
valid_payloads
.iter()
.enumerate()
.for_each(|(idx, &(reason, local_site, edit_site))| {
assert!(
validate_create_payload(local_site, edit_site).is_ok(),
"Got Err, but should have got Ok for reason: {}. valid_payloads.nth({})",
reason,
idx
);
})
}
fn generate_local_site(
site_setup: bool,
site_slur_filter_regex: Option<String>,
site_is_private: bool,
site_is_federated: bool,
site_application_question: Option<String>,
site_registration_mode: RegistrationMode,
) -> LocalSite {
LocalSite {
id: Default::default(),
site_id: Default::default(),
site_setup,
enable_downvotes: false,
enable_nsfw: false,
community_creation_admin_only: false,
require_email_verification: false,
application_question: site_application_question,
private_instance: site_is_private,
default_theme: String::new(),
default_post_listing_type: ListingType::All,
legal_information: None,
hide_modlog_mod_names: false,
application_email_admins: false,
slur_filter_regex: site_slur_filter_regex,
actor_name_max_length: 0,
federation_enabled: site_is_federated,
captcha_enabled: false,
captcha_difficulty: String::new(),
published: Default::default(),
updated: None,
registration_mode: site_registration_mode,
reports_email_admins: false,
}
}
// Allow the test helper function to have too many arguments.
// It's either this or generate the entire struct each time for testing.
#[allow(clippy::too_many_arguments)]
fn generate_create_site(
site_name: String,
site_description: Option<String>,
site_sidebar: Option<String>,
site_listing_type: Option<ListingType>,
site_slur_filter_regex: Option<String>,
site_is_private: Option<bool>,
site_is_federated: Option<bool>,
site_application_question: Option<String>,
site_registration_mode: Option<RegistrationMode>,
) -> CreateSite {
CreateSite {
name: site_name,
sidebar: site_sidebar,
description: site_description,
icon: None,
banner: None,
enable_downvotes: None,
enable_nsfw: None,
community_creation_admin_only: None,
require_email_verification: None,
application_question: site_application_question,
private_instance: site_is_private,
default_theme: None,
default_post_listing_type: site_listing_type,
legal_information: None,
application_email_admins: None,
hide_modlog_mod_names: None,
discussion_languages: None,
slur_filter_regex: site_slur_filter_regex,
actor_name_max_length: None,
rate_limit_message: None,
rate_limit_message_per_second: None,
rate_limit_post: None,
rate_limit_post_per_second: None,
rate_limit_register: None,
rate_limit_register_per_second: None,
rate_limit_image: None,
rate_limit_image_per_second: None,
rate_limit_comment: None,
rate_limit_comment_per_second: None,
rate_limit_search: None,
rate_limit_search_per_second: None,
federation_enabled: site_is_federated,
federation_debug: None,
captcha_enabled: None,
captcha_difficulty: None,
allowed_instances: None,
blocked_instances: None,
taglines: None,
registration_mode: site_registration_mode,
auth: Default::default(),
}
}
}

View File

@ -1,19 +1,95 @@
use lemmy_db_schema::RegistrationMode;
use lemmy_utils::error::LemmyError;
use lemmy_db_schema::{ListingType, RegistrationMode};
use lemmy_utils::error::{LemmyError, LemmyResult};
mod create;
mod read;
mod update;
pub fn check_application_question(
application_question: &Option<Option<String>>,
/// Checks whether the default post listing type is valid for a site.
pub fn site_default_post_listing_type_check(
default_post_listing_type: &Option<ListingType>,
) -> LemmyResult<()> {
if let Some(listing_type) = default_post_listing_type {
// Only allow all or local as default listing types...
if listing_type != &ListingType::All && listing_type != &ListingType::Local {
Err(LemmyError::from_message(
"invalid_default_post_listing_type",
))
} else {
Ok(())
}
} else {
Ok(())
}
}
/// Checks whether the application question and registration mode align.
pub fn application_question_check(
current_application_question: &Option<String>,
new_application_question: &Option<String>,
registration_mode: RegistrationMode,
) -> Result<(), LemmyError> {
) -> LemmyResult<()> {
let has_no_question: bool =
current_application_question.is_none() && new_application_question.is_none();
let is_nullifying_question: bool = new_application_question == &Some(String::new());
if registration_mode == RegistrationMode::RequireApplication
&& application_question.as_ref().unwrap_or(&None).is_none()
&& (has_no_question || is_nullifying_question)
{
Err(LemmyError::from_message("application_question_required"))
} else {
Ok(())
}
}
#[cfg(test)]
mod tests {
use crate::site::{application_question_check, site_default_post_listing_type_check};
use lemmy_db_schema::{ListingType, RegistrationMode};
#[test]
fn test_site_default_post_listing_type_check() {
assert!(site_default_post_listing_type_check(&None::<ListingType>).is_ok());
assert!(site_default_post_listing_type_check(&Some(ListingType::All)).is_ok());
assert!(site_default_post_listing_type_check(&Some(ListingType::Local)).is_ok());
assert!(site_default_post_listing_type_check(&Some(ListingType::Subscribed)).is_err());
}
#[test]
fn test_application_question_check() {
assert!(
application_question_check(&Some(String::from("q")), &Some(String::new()), RegistrationMode::RequireApplication).is_err(),
"Expected application to be invalid because an application is required, current question: {:?}, new question: {:?}",
"q",
String::new(),
);
assert!(
application_question_check(&None, &None, RegistrationMode::RequireApplication).is_err(),
"Expected application to be invalid because an application is required, current question: {:?}, new question: {:?}",
None::<String>,
None::<String>
);
assert!(
application_question_check(&None, &None, RegistrationMode::Open).is_ok(),
"Expected application to be valid because no application required, current question: {:?}, new question: {:?}, mode: {:?}",
None::<String>,
None::<String>,
RegistrationMode::Open
);
assert!(
application_question_check(&None, &Some(String::from("q")), RegistrationMode::RequireApplication).is_ok(),
"Expected application to be valid because new application provided, current question: {:?}, new question: {:?}, mode: {:?}",
None::<String>,
Some(String::from("q")),
RegistrationMode::RequireApplication
);
assert!(
application_question_check(&Some(String::from("q")), &None, RegistrationMode::RequireApplication).is_ok(),
"Expected application to be valid because application existed, current question: {:?}, new question: {:?}, mode: {:?}",
Some(String::from("q")),
None::<String>,
RegistrationMode::RequireApplication
);
}
}

View File

@ -1,15 +1,12 @@
use crate::{site::check_application_question, PerformCrud};
use crate::{
site::{application_question_check, site_default_post_listing_type_check},
PerformCrud,
};
use actix_web::web::Data;
use lemmy_api_common::{
context::LemmyContext,
site::{EditSite, SiteResponse},
utils::{
is_admin,
local_site_rate_limit_to_rate_limit_config,
local_site_to_slur_regex,
local_user_view_from_jwt,
site_description_length_check,
},
utils::{is_admin, local_site_rate_limit_to_rate_limit_config, local_user_view_from_jwt},
};
use lemmy_db_schema::{
source::{
@ -24,15 +21,20 @@ use lemmy_db_schema::{
},
traits::Crud,
utils::{diesel_option_overwrite, diesel_option_overwrite_to_url, naive_now},
ListingType,
RegistrationMode,
};
use lemmy_db_views::structs::SiteView;
use lemmy_utils::{
error::LemmyError,
error::{LemmyError, LemmyResult},
utils::{
slurs::check_slurs_opt,
validation::{check_site_visibility_valid, is_valid_body_field},
validation::{
build_and_check_regex,
check_site_visibility_valid,
is_valid_body_field,
site_description_length_check,
site_name_length_check,
},
},
};
@ -48,43 +50,10 @@ impl PerformCrud for EditSite {
let local_site = site_view.local_site;
let site = site_view.site;
// Make sure user is an admin
// Make sure user is an admin; other types of users should not update site data...
is_admin(&local_user_view)?;
check_site_visibility_valid(
local_site.private_instance,
local_site.federation_enabled,
&data.private_instance,
&data.federation_enabled,
)?;
let slur_regex = local_site_to_slur_regex(&local_site);
check_slurs_opt(&data.name, &slur_regex)?;
check_slurs_opt(&data.description, &slur_regex)?;
if let Some(desc) = &data.description {
site_description_length_check(desc)?;
}
is_valid_body_field(&data.sidebar)?;
let application_question = diesel_option_overwrite(&data.application_question);
check_application_question(
&application_question,
data
.registration_mode
.unwrap_or(local_site.registration_mode),
)?;
if let Some(listing_type) = &data.default_post_listing_type {
// only allow all or local as default listing types
if listing_type != &ListingType::All && listing_type != &ListingType::Local {
return Err(LemmyError::from_message(
"invalid_default_post_listing_type",
));
}
}
validate_update_payload(&local_site, data)?;
if let Some(discussion_languages) = data.discussion_languages.clone() {
SiteLanguage::update(
@ -95,9 +64,8 @@ impl PerformCrud for EditSite {
.await?;
}
let name = data.name.clone();
let site_form = SiteUpdateForm::builder()
.name(name)
.name(data.name.clone())
.sidebar(diesel_option_overwrite(&data.sidebar))
.description(diesel_option_overwrite(&data.description))
.icon(diesel_option_overwrite_to_url(&data.icon)?)
@ -117,7 +85,7 @@ impl PerformCrud for EditSite {
.enable_nsfw(data.enable_nsfw)
.community_creation_admin_only(data.community_creation_admin_only)
.require_email_verification(data.require_email_verification)
.application_question(application_question)
.application_question(diesel_option_overwrite(&data.application_question))
.private_instance(data.private_instance)
.default_theme(data.default_theme.clone())
.default_post_listing_type(data.default_post_listing_type)
@ -128,7 +96,6 @@ impl PerformCrud for EditSite {
.slur_filter_regex(diesel_option_overwrite(&data.slur_filter_regex))
.actor_name_max_length(data.actor_name_max_length)
.federation_enabled(data.federation_enabled)
.federation_worker_count(data.federation_worker_count)
.captcha_enabled(data.captcha_enabled)
.captcha_difficulty(data.captcha_difficulty.clone())
.reports_email_admins(data.reports_email_admins)
@ -211,3 +178,411 @@ impl PerformCrud for EditSite {
Ok(res)
}
}
fn validate_update_payload(local_site: &LocalSite, edit_site: &EditSite) -> LemmyResult<()> {
// Check that the slur regex compiles, and return the regex if valid...
// Prioritize using new slur regex from the request; if not provided, use the existing regex.
let slur_regex = build_and_check_regex(
&edit_site
.slur_filter_regex
.as_deref()
.or(local_site.slur_filter_regex.as_deref()),
)?;
if let Some(name) = &edit_site.name {
// The name doesn't need to be updated, but if provided it cannot be blanked out...
site_name_length_check(name)?;
check_slurs_opt(&edit_site.name, &slur_regex)?;
}
if let Some(desc) = &edit_site.description {
site_description_length_check(desc)?;
check_slurs_opt(&edit_site.description, &slur_regex)?;
}
site_default_post_listing_type_check(&edit_site.default_post_listing_type)?;
check_site_visibility_valid(
local_site.private_instance,
local_site.federation_enabled,
&edit_site.private_instance,
&edit_site.federation_enabled,
)?;
// Ensure that the sidebar has fewer than the max num characters...
is_valid_body_field(&edit_site.sidebar, false)?;
application_question_check(
&local_site.application_question,
&edit_site.application_question,
edit_site
.registration_mode
.unwrap_or(local_site.registration_mode),
)
}
#[cfg(test)]
mod tests {
use crate::site::update::validate_update_payload;
use lemmy_api_common::site::EditSite;
use lemmy_db_schema::{source::local_site::LocalSite, ListingType, RegistrationMode};
#[test]
fn test_validate_invalid_update_payload() {
let invalid_payloads = [
(
"EditSite name matches LocalSite slur filter",
"slurs",
&generate_local_site(
Some(String::from("(foo|bar)")),
true,
false,
None::<String>,
RegistrationMode::Open,
),
&generate_edit_site(
Some(String::from("foo site_name")),
None::<String>,
None::<String>,
None::<ListingType>,
None::<String>,
None::<bool>,
None::<bool>,
None::<String>,
None::<RegistrationMode>,
),
),
(
"EditSite name matches new slur filter",
"slurs",
&generate_local_site(
Some(String::from("(foo|bar)")),
true,
false,
None::<String>,
RegistrationMode::Open,
),
&generate_edit_site(
Some(String::from("zeta site_name")),
None::<String>,
None::<String>,
None::<ListingType>,
Some(String::from("(zeta|alpha)")),
None::<bool>,
None::<bool>,
None::<String>,
None::<RegistrationMode>,
),
),
(
"EditSite listing type is Subscribed, which is invalid",
"invalid_default_post_listing_type",
&generate_local_site(
None::<String>,
true,
false,
None::<String>,
RegistrationMode::Open,
),
&generate_edit_site(
Some(String::from("site_name")),
None::<String>,
None::<String>,
Some(ListingType::Subscribed),
None::<String>,
None::<bool>,
None::<bool>,
None::<String>,
None::<RegistrationMode>,
),
),
(
"EditSite is both private and federated",
"cant_enable_private_instance_and_federation_together",
&generate_local_site(
None::<String>,
true,
false,
None::<String>,
RegistrationMode::Open,
),
&generate_edit_site(
Some(String::from("site_name")),
None::<String>,
None::<String>,
None::<ListingType>,
None::<String>,
Some(true),
Some(true),
None::<String>,
None::<RegistrationMode>,
),
),
(
"LocalSite is private, but EditSite also makes it federated",
"cant_enable_private_instance_and_federation_together",
&generate_local_site(
None::<String>,
true,
false,
None::<String>,
RegistrationMode::Open,
),
&generate_edit_site(
Some(String::from("site_name")),
None::<String>,
None::<String>,
None::<ListingType>,
None::<String>,
None::<bool>,
Some(true),
None::<String>,
None::<RegistrationMode>,
),
),
(
"EditSite requires application, but neither it nor LocalSite has an application question",
"application_question_required",
&generate_local_site(
None::<String>,
true,
false,
None::<String>,
RegistrationMode::Open,
),
&generate_edit_site(
Some(String::from("site_name")),
None::<String>,
None::<String>,
None::<ListingType>,
None::<String>,
None::<bool>,
None::<bool>,
None::<String>,
Some(RegistrationMode::RequireApplication),
),
),
];
invalid_payloads.iter().enumerate().for_each(
|(
idx,
&(reason, expected_err, local_site, edit_site),
)| {
match validate_update_payload(local_site, edit_site) {
Ok(_) => {
panic!(
"Got Ok, but validation should have failed with error: {} for reason: {}. invalid_payloads.nth({})",
expected_err, reason, idx
)
}
Err(error) => {
assert!(
error.message.eq(&Some(String::from(expected_err))),
"Got Err {:?}, but should have failed with message: {} for reason: {}. invalid_payloads.nth({})",
error.message,
expected_err,
reason,
idx
)
}
}
},
);
}
#[test]
fn test_validate_valid_update_payload() {
let valid_payloads = [
(
"No changes between LocalSite and EditSite",
&generate_local_site(
None::<String>,
true,
false,
None::<String>,
RegistrationMode::Open,
),
&generate_edit_site(
None::<String>,
None::<String>,
None::<String>,
None::<ListingType>,
None::<String>,
None::<bool>,
None::<bool>,
None::<String>,
None::<RegistrationMode>,
),
),
(
"EditSite allows clearing and changing values",
&generate_local_site(
None::<String>,
true,
false,
None::<String>,
RegistrationMode::Open,
),
&generate_edit_site(
Some(String::from("site_name")),
Some(String::new()),
Some(String::new()),
Some(ListingType::All),
Some(String::new()),
Some(false),
Some(true),
Some(String::new()),
Some(RegistrationMode::Open),
),
),
(
"EditSite name passes slur filter regex",
&generate_local_site(
Some(String::from("(foo|bar)")),
true,
false,
None::<String>,
RegistrationMode::Open,
),
&generate_edit_site(
Some(String::from("foo site_name")),
None::<String>,
None::<String>,
None::<ListingType>,
Some(String::new()),
None::<bool>,
None::<bool>,
None::<String>,
None::<RegistrationMode>,
),
),
(
"LocalSite has application question and EditSite now requires applications,",
&generate_local_site(
None::<String>,
true,
false,
Some(String::from("question")),
RegistrationMode::Open,
),
&generate_edit_site(
Some(String::from("site_name")),
None::<String>,
None::<String>,
None::<ListingType>,
None::<String>,
None::<bool>,
None::<bool>,
None::<String>,
Some(RegistrationMode::RequireApplication),
),
),
];
valid_payloads
.iter()
.enumerate()
.for_each(|(idx, &(reason, local_site, edit_site))| {
assert!(
validate_update_payload(local_site, edit_site).is_ok(),
"Got Err, but should have got Ok for reason: {}. valid_payloads.nth({})",
reason,
idx
);
})
}
fn generate_local_site(
site_slur_filter_regex: Option<String>,
site_is_private: bool,
site_is_federated: bool,
site_application_question: Option<String>,
site_registration_mode: RegistrationMode,
) -> LocalSite {
LocalSite {
id: Default::default(),
site_id: Default::default(),
site_setup: true,
enable_downvotes: false,
enable_nsfw: false,
community_creation_admin_only: false,
require_email_verification: false,
application_question: site_application_question,
private_instance: site_is_private,
default_theme: String::new(),
default_post_listing_type: ListingType::All,
legal_information: None,
hide_modlog_mod_names: false,
application_email_admins: false,
slur_filter_regex: site_slur_filter_regex,
actor_name_max_length: 0,
federation_enabled: site_is_federated,
captcha_enabled: false,
captcha_difficulty: String::new(),
published: Default::default(),
updated: None,
registration_mode: site_registration_mode,
reports_email_admins: false,
}
}
// Allow the test helper function to have too many arguments.
// It's either this or generate the entire struct each time for testing.
#[allow(clippy::too_many_arguments)]
fn generate_edit_site(
site_name: Option<String>,
site_description: Option<String>,
site_sidebar: Option<String>,
site_listing_type: Option<ListingType>,
site_slur_filter_regex: Option<String>,
site_is_private: Option<bool>,
site_is_federated: Option<bool>,
site_application_question: Option<String>,
site_registration_mode: Option<RegistrationMode>,
) -> EditSite {
EditSite {
name: site_name,
sidebar: site_sidebar,
description: site_description,
icon: None,
banner: None,
enable_downvotes: None,
enable_nsfw: None,
community_creation_admin_only: None,
require_email_verification: None,
application_question: site_application_question,
private_instance: site_is_private,
default_theme: None,
default_post_listing_type: site_listing_type,
legal_information: None,
application_email_admins: None,
hide_modlog_mod_names: None,
discussion_languages: None,
slur_filter_regex: site_slur_filter_regex,
actor_name_max_length: None,
rate_limit_message: None,
rate_limit_message_per_second: None,
rate_limit_post: None,
rate_limit_post_per_second: None,
rate_limit_register: None,
rate_limit_register_per_second: None,
rate_limit_image: None,
rate_limit_image_per_second: None,
rate_limit_comment: None,
rate_limit_comment_per_second: None,
rate_limit_search: None,
rate_limit_search_per_second: None,
federation_enabled: site_is_federated,
federation_debug: None,
captcha_enabled: None,
captcha_difficulty: None,
allowed_instances: None,
blocked_instances: None,
taglines: None,
registration_mode: site_registration_mode,
reports_email_admins: None,
auth: Default::default(),
}
}
}

View File

@ -19,6 +19,7 @@ use lemmy_api_common::{
use lemmy_db_schema::{
aggregates::structs::PersonAggregates,
source::{
captcha_answer::{CaptchaAnswer, CheckCaptchaAnswer},
local_user::{LocalUser, LocalUserInsertForm},
person::{Person, PersonInsertForm},
registration_application::{RegistrationApplication, RegistrationApplicationInsertForm},
@ -71,6 +72,25 @@ impl PerformCrud for Register {
return Err(LemmyError::from_message("passwords_dont_match"));
}
if local_site.site_setup && local_site.captcha_enabled {
if let Some(captcha_uuid) = &data.captcha_uuid {
let uuid = uuid::Uuid::parse_str(captcha_uuid)?;
let check = CaptchaAnswer::check_captcha(
context.pool(),
CheckCaptchaAnswer {
uuid,
answer: data.captcha_answer.clone().unwrap_or_default(),
},
)
.await?;
if !check {
return Err(LemmyError::from_message("captcha_incorrect"));
}
} else {
return Err(LemmyError::from_message("captcha_incorrect"));
}
}
let slur_regex = local_site_to_slur_regex(&local_site);
check_slurs(&data.username, &slur_regex)?;
check_slurs_opt(&data.answer, &slur_regex)?;

View File

@ -25,7 +25,7 @@ chrono = { workspace = true }
serde_json = { workspace = true }
serde = { workspace = true }
actix-web = { workspace = true }
actix-rt = { workspace = true }
tokio = {workspace = true}
tracing = { workspace = true }
strum_macros = { workspace = true }
url = { workspace = true }

View File

@ -43,12 +43,11 @@ pub(crate) async fn send_activity_in_community(
// send to user followers
if !is_mod_action {
inboxes.append(
inboxes.extend(
&mut PersonFollower::list_followers(&mut *context.conn().await?, actor.id)
.await?
.into_iter()
.map(|p| ApubPerson(p).shared_inbox_or_inbox())
.collect(),
.map(|p| ApubPerson(p).shared_inbox_or_inbox()),
);
}

View File

@ -121,7 +121,7 @@ mod tests {
};
use serial_test::serial;
#[actix_rt::test]
#[tokio::test]
#[serial]
async fn test_parse_lemmy_community_moderators() {
let context = init_context().await;

View File

@ -48,7 +48,7 @@ where
Ok(
HttpResponse::Ok()
.content_type(FEDERATION_CONTENT_TYPE)
.content_type("application/json")
.content_type("application/activity+json")
.body(json),
)
}
@ -61,7 +61,7 @@ fn create_apub_tombstone_response<T: Into<Url>>(id: T) -> LemmyResult<HttpRespon
HttpResponse::Gone()
.content_type(FEDERATION_CONTENT_TYPE)
.status(StatusCode::GONE)
.content_type("application/json")
.content_type("application/activity+json")
.body(json),
)
}

View File

@ -11,10 +11,7 @@ use lemmy_db_schema::{
traits::Crud,
utils::DbConn,
};
use lemmy_utils::{
error::LemmyError,
utils::mention::{scrape_text_for_mentions, MentionData},
};
use lemmy_utils::{error::LemmyError, utils::mention::scrape_text_for_mentions};
use serde::{Deserialize, Serialize};
use serde_json::Value;
use url::Url;
@ -67,10 +64,9 @@ pub async fn collect_non_local_mentions(
let mentions = scrape_text_for_mentions(&comment.content)
.into_iter()
// Filter only the non-local ones
.filter(|m| !m.is_local(&context.settings().hostname))
.collect::<Vec<MentionData>>();
.filter(|m| !m.is_local(&context.settings().hostname));
for mention in &mentions {
for mention in mentions {
let identifier = format!("{}@{}", mention.name, mention.domain);
let person = webfinger_resolve_actor::<LemmyContext, ApubPerson>(&identifier, context).await;
if let Ok(person) = person {

View File

@ -239,7 +239,7 @@ pub(crate) mod tests {
.unwrap();
}
#[actix_rt::test]
#[tokio::test]
#[serial]
pub(crate) async fn test_parse_lemmy_comment() {
let context = init_context().await;
@ -267,7 +267,7 @@ pub(crate) mod tests {
cleanup(data, &context).await;
}
#[actix_rt::test]
#[tokio::test]
#[serial]
async fn test_parse_pleroma_comment() {
let context = init_context().await;
@ -299,7 +299,7 @@ pub(crate) mod tests {
cleanup(data, &context).await;
}
#[actix_rt::test]
#[tokio::test]
#[serial]
async fn test_html_to_markdown_sanitize() {
let parsed = parse_html("<script></script><b>hello</b>");

View File

@ -141,19 +141,16 @@ impl Object for ApubCommunity {
// Fetching mods and outbox is not necessary for Lemmy to work, so ignore errors. Besides,
// we need to ignore these errors so that tests can work entirely offline.
group
.outbox
.dereference(&community, context)
.await
.map_err(|e| debug!("{}", e))
.ok();
let fetch_outbox = group.outbox.dereference(&community, context);
if let Some(moderators) = group.attributed_to {
moderators
.dereference(&community, context)
.await
.map_err(|e| debug!("{}", e))
.ok();
let fetch_moderators = moderators.dereference(&community, context);
// Fetch mods and outbox in parallel
let res = tokio::join!(fetch_outbox, fetch_moderators);
res.0.map_err(|e| debug!("{}", e)).ok();
res.1.map_err(|e| debug!("{}", e)).ok();
} else {
fetch_outbox.await.map_err(|e| debug!("{}", e)).ok();
}
Ok(community)
@ -243,7 +240,7 @@ pub(crate) mod tests {
community
}
#[actix_rt::test]
#[tokio::test]
#[serial]
async fn test_parse_lemmy_community() {
let context = init_context().await;

View File

@ -222,7 +222,7 @@ pub(crate) mod tests {
site
}
#[actix_rt::test]
#[tokio::test]
#[serial]
async fn test_parse_lemmy_instance() {
let context = init_context().await;

View File

@ -223,7 +223,7 @@ pub(crate) mod tests {
(person, site)
}
#[actix_rt::test]
#[tokio::test]
#[serial]
async fn test_parse_lemmy_person() {
let context = init_context().await;
@ -236,7 +236,7 @@ pub(crate) mod tests {
cleanup((person, site), &context).await;
}
#[actix_rt::test]
#[tokio::test]
#[serial]
async fn test_parse_pleroma_person() {
let context = init_context().await;

View File

@ -282,7 +282,7 @@ mod tests {
use lemmy_db_schema::source::site::Site;
use serial_test::serial;
#[actix_rt::test]
#[tokio::test]
#[serial]
async fn test_parse_lemmy_post() {
let context = init_context().await;

View File

@ -193,7 +193,7 @@ mod tests {
.unwrap();
}
#[actix_rt::test]
#[tokio::test]
#[serial]
async fn test_parse_lemmy_pm() {
let context = init_context().await;
@ -221,7 +221,7 @@ mod tests {
cleanup(data, &context).await;
}
#[actix_rt::test]
#[tokio::test]
#[serial]
async fn test_parse_pleroma_pm() {
let context = init_context().await;

View File

@ -29,7 +29,7 @@ serde_json = { workspace = true, optional = true }
activitypub_federation = { workspace = true, optional = true }
lemmy_utils = { workspace = true, optional = true }
bcrypt = { workspace = true, optional = true }
diesel = { workspace = true, features = ["postgres","chrono", "serde_json"], optional = true }
diesel = { workspace = true, features = ["postgres","chrono", "serde_json", "uuid"], optional = true }
diesel-derive-newtype = { workspace = true, optional = true }
diesel-derive-enum = { workspace = true, optional = true }
diesel_migrations = { workspace = true, optional = true }
@ -43,8 +43,12 @@ async-trait = { workspace = true }
tokio = { workspace = true }
tracing = { workspace = true }
deadpool = { version = "0.9.5", features = ["rt_tokio_1"], optional = true }
ts-rs = { workspace = true, optional = true }
ts-rs = { workspace = true, optional = true }
rustls = { workspace = true }
futures-util = { workspace = true }
tokio-postgres = { workspace = true }
tokio-postgres-rustls = { workspace = true }
uuid = { workspace = true, features = ["v4"] }
[dev-dependencies]
serial_test = { workspace = true }

View File

@ -19,8 +19,8 @@ index 255c6422..f2ccf5e2 100644
#[derive(diesel::sql_types::SqlType)]
#[diesel(postgres_type(name = "sort_type_enum"))]
@@ -67,13 +63,13 @@ diesel::table! {
when_ -> Timestamp,
@@ -76,13 +76,13 @@ diesel::table! {
published -> Timestamp,
}
}

View File

@ -0,0 +1,118 @@
use crate::{
schema::captcha_answer::dsl::{answer, captcha_answer, uuid},
source::captcha_answer::{CaptchaAnswer, CaptchaAnswerForm, CheckCaptchaAnswer},
utils::{functions::lower, get_conn, DbPool},
};
use diesel::{
delete,
dsl::exists,
insert_into,
result::Error,
select,
ExpressionMethods,
QueryDsl,
};
use diesel_async::RunQueryDsl;
impl CaptchaAnswer {
pub async fn insert(pool: &DbPool, captcha: &CaptchaAnswerForm) -> Result<Self, Error> {
let conn = &mut get_conn(pool).await?;
insert_into(captcha_answer)
.values(captcha)
.get_result::<Self>(conn)
.await
}
pub async fn check_captcha(pool: &DbPool, to_check: CheckCaptchaAnswer) -> Result<bool, Error> {
let conn = &mut get_conn(pool).await?;
// fetch requested captcha
let captcha_exists = select(exists(
captcha_answer
.filter((uuid).eq(to_check.uuid))
.filter(lower(answer).eq(to_check.answer.to_lowercase().clone())),
))
.get_result::<bool>(conn)
.await?;
// delete checked captcha
delete(captcha_answer.filter(uuid.eq(to_check.uuid)))
.execute(conn)
.await?;
Ok(captcha_exists)
}
}
#[cfg(test)]
mod tests {
use crate::{
source::captcha_answer::{CaptchaAnswer, CaptchaAnswerForm, CheckCaptchaAnswer},
utils::build_db_pool_for_tests,
};
use serial_test::serial;
#[tokio::test]
#[serial]
async fn test_captcha_happy_path() {
let pool = &build_db_pool_for_tests().await;
let inserted = CaptchaAnswer::insert(
pool,
&CaptchaAnswerForm {
answer: "XYZ".to_string(),
},
)
.await
.expect("should not fail to insert captcha");
let result = CaptchaAnswer::check_captcha(
pool,
CheckCaptchaAnswer {
uuid: inserted.uuid,
answer: "xyz".to_string(),
},
)
.await;
assert!(result.is_ok());
assert!(result.unwrap());
}
#[tokio::test]
#[serial]
async fn test_captcha_repeat_answer_fails() {
let pool = &build_db_pool_for_tests().await;
let inserted = CaptchaAnswer::insert(
pool,
&CaptchaAnswerForm {
answer: "XYZ".to_string(),
},
)
.await
.expect("should not fail to insert captcha");
let _result = CaptchaAnswer::check_captcha(
pool,
CheckCaptchaAnswer {
uuid: inserted.uuid,
answer: "xyz".to_string(),
},
)
.await;
let result_repeat = CaptchaAnswer::check_captcha(
pool,
CheckCaptchaAnswer {
uuid: inserted.uuid,
answer: "xyz".to_string(),
},
)
.await;
assert!(result_repeat.is_ok());
assert!(!result_repeat.unwrap());
}
}

View File

@ -11,7 +11,7 @@ use crate::{
CommentUpdateForm,
},
traits::{Crud, Likeable, Saveable},
utils::{naive_now, DbConn},
utils::{get_conn, naive_now, DbConn, DELETED_REPLACEMENT_TEXT},
};
use diesel::{
dsl::{insert_into, sql_query},
@ -30,7 +30,7 @@ impl Comment {
) -> Result<Vec<Self>, Error> {
diesel::update(comment.filter(creator_id.eq(for_creator_id)))
.set((
content.eq("*Permananently Deleted*"),
content.eq(DELETED_REPLACEMENT_TEXT),
deleted.eq(true),
updated.eq(naive_now()),
))
@ -94,8 +94,7 @@ impl Comment {
// left join comment c2 on c2.path <@ c.path and c2.path != c.path
// group by c.id
let path_split = parent_path.0.split('.').collect::<Vec<&str>>();
let parent_id = path_split.get(1);
let parent_id = parent_path.0.split('.').nth(1);
if let Some(parent_id) = parent_id {
let top_parent = format!("0.{}", parent_id);

View File

@ -1,5 +1,6 @@
pub mod activity;
pub mod actor_language;
pub mod captcha_answer;
pub mod comment;
pub mod comment_reply;
pub mod comment_report;

View File

@ -1,6 +1,11 @@
use crate::{
newtypes::LocalUserId,
schema::password_reset_request::dsl::{password_reset_request, published, token_encrypted},
schema::password_reset_request::dsl::{
local_user_id,
password_reset_request,
published,
token_encrypted,
},
source::password_reset_request::{PasswordResetRequest, PasswordResetRequestForm},
traits::Crud,
utils::DbConn,
@ -73,6 +78,19 @@ impl PasswordResetRequest {
.first::<Self>(conn)
.await
}
pub async fn get_recent_password_resets_count(
pool: &DbPool,
user_id: LocalUserId,
) -> Result<i64, Error> {
let conn = &mut get_conn(pool).await?;
password_reset_request
.filter(local_user_id.eq(user_id))
.filter(published.gt(now - 1.days()))
.count()
.get_result(conn)
.await
}
}
fn bytes_to_hex(bytes: Vec<u8>) -> String {

View File

@ -27,7 +27,7 @@ use crate::{
PostUpdateForm,
},
traits::{Crud, Likeable, Readable, Saveable},
utils::{naive_now, DbConn, FETCH_LIMIT_MAX},
utils::{get_conn, naive_now, DbConn, DbPool, DELETED_REPLACEMENT_TEXT, FETCH_LIMIT_MAX},
};
use ::url::Url;
use diesel::{dsl::insert_into, result::Error, ExpressionMethods, QueryDsl, TextExpressionMethods};
@ -108,9 +108,9 @@ impl Post {
diesel::update(post.filter(creator_id.eq(for_creator_id)))
.set((
name.eq(perma_deleted),
url.eq(perma_deleted_url),
body.eq(perma_deleted),
name.eq(DELETED_REPLACEMENT_TEXT),
url.eq(Option::<&str>::None),
body.eq(DELETED_REPLACEMENT_TEXT),
deleted.eq(true),
updated.eq(naive_now()),
))

View File

@ -26,6 +26,7 @@ pub mod impls;
pub mod newtypes;
#[cfg(feature = "full")]
#[rustfmt::skip]
#[allow(clippy::wildcard_imports)]
pub mod schema;
pub mod source;
#[cfg(feature = "full")]
@ -62,6 +63,9 @@ pub enum SortType {
TopHour,
TopSixHour,
TopTwelveHour,
TopThreeMonths,
TopSixMonths,
TopNineMonths,
}
#[derive(EnumString, Display, Debug, Serialize, Deserialize, Clone, Copy)]

View File

@ -65,7 +65,16 @@ diesel::table! {
}
diesel::table! {
use diesel::sql_types::{Bool, Int4, Nullable, Text, Timestamp, Varchar};
captcha_answer (id) {
id -> Int4,
uuid -> Uuid,
answer -> Text,
published -> Timestamp,
}
}
diesel::table! {
use diesel::sql_types::*;
use diesel_ltree::sql_types::Ltree;
comment (id) {
@ -317,7 +326,7 @@ diesel::table! {
}
diesel::table! {
use diesel::sql_types::{Bool, Int4, Nullable, Text, Timestamp, Varchar};
use diesel::sql_types::*;
use super::sql_types::ListingTypeEnum;
use super::sql_types::RegistrationModeEnum;
@ -339,7 +348,6 @@ diesel::table! {
slur_filter_regex -> Nullable<Text>,
actor_name_max_length -> Int4,
federation_enabled -> Bool,
federation_worker_count -> Int4,
captcha_enabled -> Bool,
#[max_length = 255]
captcha_difficulty -> Varchar,
@ -372,7 +380,7 @@ diesel::table! {
}
diesel::table! {
use diesel::sql_types::{Bool, Int4, Nullable, Text, Timestamp, Varchar};
use diesel::sql_types::*;
use super::sql_types::SortTypeEnum;
use super::sql_types::ListingTypeEnum;
@ -382,8 +390,7 @@ diesel::table! {
password_encrypted -> Text,
email -> Nullable<Text>,
show_nsfw -> Bool,
#[max_length = 20]
theme -> Varchar,
theme -> Text,
default_sort_type -> SortTypeEnum,
default_listing_type -> ListingTypeEnum,
#[max_length = 20]
@ -916,6 +923,7 @@ diesel::allow_tables_to_appear_in_same_query!(
admin_purge_community,
admin_purge_person,
admin_purge_post,
captcha_answer,
comment,
comment_aggregates,
comment_like,

View File

@ -0,0 +1,33 @@
#[cfg(feature = "full")]
use crate::schema::captcha_answer;
use serde::{Deserialize, Serialize};
use serde_with::skip_serializing_none;
use uuid::Uuid;
#[skip_serializing_none]
#[derive(Clone, PartialEq, Eq, Debug, Serialize, Deserialize)]
#[cfg_attr(feature = "full", derive(Queryable))]
#[cfg_attr(feature = "full", diesel(table_name = captcha_answer))]
pub struct CaptchaAnswer {
pub id: i32,
pub uuid: Uuid,
pub answer: String,
pub published: chrono::NaiveDateTime,
}
#[skip_serializing_none]
#[derive(Clone, PartialEq, Eq, Debug, Serialize, Deserialize)]
#[cfg_attr(feature = "full", derive(Queryable))]
#[cfg_attr(feature = "full", diesel(table_name = captcha_answer))]
pub struct CheckCaptchaAnswer {
pub uuid: Uuid,
pub answer: String,
}
#[skip_serializing_none]
#[derive(Clone, PartialEq, Eq, Debug, Serialize, Deserialize)]
#[cfg_attr(feature = "full", derive(Insertable, AsChangeset))]
#[cfg_attr(feature = "full", diesel(table_name = captcha_answer))]
pub struct CaptchaAnswerForm {
pub answer: String,
}

View File

@ -50,8 +50,6 @@ pub struct LocalSite {
pub actor_name_max_length: i32,
/// Whether federation is enabled.
pub federation_enabled: bool,
/// The number of concurrent federation http workers.
pub federation_worker_count: i32,
/// Whether captcha is enabled.
pub captcha_enabled: bool,
/// The captcha difficulty.
@ -85,7 +83,6 @@ pub struct LocalSiteInsertForm {
pub slur_filter_regex: Option<String>,
pub actor_name_max_length: Option<i32>,
pub federation_enabled: Option<bool>,
pub federation_worker_count: Option<i32>,
pub captcha_enabled: Option<bool>,
pub captcha_difficulty: Option<String>,
pub registration_mode: Option<RegistrationMode>,
@ -112,7 +109,6 @@ pub struct LocalSiteUpdateForm {
pub slur_filter_regex: Option<Option<String>>,
pub actor_name_max_length: Option<i32>,
pub federation_enabled: Option<bool>,
pub federation_worker_count: Option<i32>,
pub captcha_enabled: Option<bool>,
pub captcha_difficulty: Option<String>,
pub registration_mode: Option<RegistrationMode>,

View File

@ -1,6 +1,7 @@
#[cfg(feature = "full")]
pub mod activity;
pub mod actor_language;
pub mod captcha_answer;
pub mod comment;
pub mod comment_reply;
pub mod comment_report;

View File

@ -12,7 +12,7 @@ use diesel::{
backend::Backend,
deserialize::FromSql,
pg::Pg,
result::{Error as DieselError, Error::QueryBuilderError},
result::{ConnectionError, ConnectionResult, Error as DieselError, Error::QueryBuilderError},
serialize::{Output, ToSql},
sql_types::Text,
PgConnection,
@ -25,11 +25,21 @@ use diesel_async::{
},
};
use diesel_migrations::EmbeddedMigrations;
use futures_util::{future::BoxFuture, FutureExt};
use lemmy_utils::{error::LemmyError, settings::structs::Settings};
use once_cell::sync::Lazy;
use regex::Regex;
use std::{env, env::VarError, time::Duration};
use tracing::info;
use rustls::{
client::{ServerCertVerified, ServerCertVerifier},
ServerName,
};
use std::{
env,
env::VarError,
sync::Arc,
time::{Duration, SystemTime},
};
use tracing::{error, info};
use url::Url;
const FETCH_LIMIT_DEFAULT: i64 = 10;
@ -138,7 +148,15 @@ pub fn diesel_option_overwrite_to_url_create(
async fn build_db_pool_settings_opt(settings: Option<&Settings>) -> Result<DbPool, LemmyError> {
let db_url = get_database_url(settings);
let pool_size = settings.map(|s| s.database.pool_size).unwrap_or(5);
let manager = AsyncDieselConnectionManager::<AsyncPgConnection>::new(&db_url);
// We only support TLS with sslmode=require currently
let tls_enabled = db_url.contains("sslmode=require");
let manager = if tls_enabled {
// diesel-async does not support any TLS connections out of the box, so we need to manually
// provide a setup function which handles creating the connection
AsyncDieselConnectionManager::<AsyncPgConnection>::new_with_setup(&db_url, establish_connection)
} else {
AsyncDieselConnectionManager::<AsyncPgConnection>::new(&db_url)
};
let pool = Pool::builder(manager)
.max_size(pool_size)
.wait_timeout(POOL_TIMEOUT)
@ -155,6 +173,44 @@ async fn build_db_pool_settings_opt(settings: Option<&Settings>) -> Result<DbPoo
Ok(pool)
}
fn establish_connection(config: &str) -> BoxFuture<ConnectionResult<AsyncPgConnection>> {
let fut = async {
let rustls_config = rustls::ClientConfig::builder()
.with_safe_defaults()
.with_custom_certificate_verifier(Arc::new(NoCertVerifier {}))
.with_no_client_auth();
let tls = tokio_postgres_rustls::MakeRustlsConnect::new(rustls_config);
let (client, conn) = tokio_postgres::connect(config, tls)
.await
.map_err(|e| ConnectionError::BadConnection(e.to_string()))?;
tokio::spawn(async move {
if let Err(e) = conn.await {
error!("Database connection failed: {e}");
}
});
AsyncPgConnection::try_from(client).await
};
fut.boxed()
}
struct NoCertVerifier {}
impl ServerCertVerifier for NoCertVerifier {
fn verify_server_cert(
&self,
_end_entity: &rustls::Certificate,
_intermediates: &[rustls::Certificate],
_server_name: &ServerName,
_scts: &mut dyn Iterator<Item = &[u8]>,
_ocsp_response: &[u8],
_now: SystemTime,
) -> Result<ServerCertVerified, rustls::Error> {
// Will verify all (even invalid) certs without any checks (sslmode=require)
Ok(ServerCertVerified::assertion())
}
}
pub const MIGRATIONS: EmbeddedMigrations = embed_migrations!();
pub fn run_migrations(db_url: &str) {
@ -211,7 +267,10 @@ pub fn post_to_comment_sort_type(sort: SortType) -> CommentSortType {
| SortType::TopAll
| SortType::TopWeek
| SortType::TopYear
| SortType::TopMonth => CommentSortType::Top,
| SortType::TopMonth
| SortType::TopThreeMonths
| SortType::TopSixMonths
| SortType::TopNineMonths => CommentSortType::Top,
}
}
@ -230,6 +289,8 @@ pub mod functions {
sql_function!(fn lower(x: Text) -> Text);
}
pub const DELETED_REPLACEMENT_TEXT: &str = "*Permanently Deleted*";
impl ToSql<Text, Pg> for DbUrl {
fn to_sql(&self, out: &mut Output<Pg>) -> diesel::serialize::Result {
<std::string::String as ToSql<Text, Pg>>::to_sql(&self.0.to_string(), &mut out.reborrow())

View File

@ -428,6 +428,18 @@ impl<'a> PostQuery<'a> {
.filter(post_aggregates::published.gt(now - 12.hours()))
.then_order_by(post_aggregates::score.desc())
.then_order_by(post_aggregates::published.desc()),
SortType::TopThreeMonths => query
.filter(post_aggregates::published.gt(now - 3.months()))
.then_order_by(post_aggregates::score.desc())
.then_order_by(post_aggregates::published.desc()),
SortType::TopSixMonths => query
.filter(post_aggregates::published.gt(now - 6.months()))
.then_order_by(post_aggregates::score.desc())
.then_order_by(post_aggregates::published.desc()),
SortType::TopNineMonths => query
.filter(post_aggregates::published.gt(now - 9.months()))
.then_order_by(post_aggregates::score.desc())
.then_order_by(post_aggregates::published.desc()),
};
let (limit, offset) = limit_and_offset(self.page, self.limit)?;

View File

@ -125,6 +125,7 @@ pub struct CommunityQuery<'a> {
local_user: Option<&'a LocalUser>,
search_term: Option<String>,
is_mod_or_admin: Option<bool>,
show_nsfw: Option<bool>,
page: Option<i64>,
limit: Option<i64>,
}
@ -202,8 +203,8 @@ impl<'a> CommunityQuery<'a> {
query = query.filter(community_block::person_id.is_null());
query = query.filter(community::nsfw.eq(false).or(local_user::show_nsfw.eq(true)));
} else {
// No person in request, only show nsfw communities if show_nsfw passed into request
if !self.local_user.map(|l| l.show_nsfw).unwrap_or(false) {
// No person in request, only show nsfw communities if show_nsfw is passed into request
if !self.show_nsfw.unwrap_or(false) {
query = query.filter(community::nsfw.eq(false));
}
}

View File

@ -119,6 +119,15 @@ impl<'a> PersonQuery<'a> {
SortType::TopTwelveHour => query
.filter(person::published.gt(now - 12.hours()))
.order_by(person_aggregates::comment_score.desc()),
SortType::TopThreeMonths => query
.filter(person::published.gt(now - 3.months()))
.order_by(person_aggregates::comment_score.desc()),
SortType::TopSixMonths => query
.filter(person::published.gt(now - 6.months()))
.order_by(person_aggregates::comment_score.desc()),
SortType::TopNineMonths => query
.filter(person::published.gt(now - 9.months()))
.order_by(person_aggregates::comment_score.desc()),
};
let (limit, offset) = limit_and_offset(self.page, self.limit)?;

View File

@ -482,7 +482,6 @@ fn create_post_items(
i.pub_date(dt.to_rfc2822());
let post_url = format!("{}/post/{}", protocol_and_hostname, p.post.id);
i.link(post_url.clone());
i.comments(post_url.clone());
let guid = GuidBuilder::default()
.permalink(true)
@ -506,6 +505,9 @@ fn create_post_items(
if let Some(url) = p.post.url {
let link_html = format!("<br><a href=\"{url}\">{url}</a>");
description.push_str(&link_html);
i.link(url.to_string());
} else {
i.link(post_url.clone());
}
if let Some(body) = p.post.body {

View File

@ -49,7 +49,6 @@ enum-map = "2.5"
[dev-dependencies]
reqwest = { workspace = true }
tokio = { workspace = true, features = ["macros"] }
[build-dependencies]
rosetta-build = "0.1.2"

View File

@ -0,0 +1,16 @@
use doku::json::{AutoComments, CommentsStyle, Formatting, ObjectsStyle};
use lemmy_utils::settings::structs::Settings;
fn main() {
let fmt = Formatting {
auto_comments: AutoComments::none(),
comments_style: CommentsStyle {
separator: "#".to_owned(),
},
objects_style: ObjectsStyle {
surround_keys_with_quotes: false,
use_comma_as_separator: false,
},
..Default::default()
};
println!("{}", doku::to_json_fmt_val(&fmt, &Settings::default()));
}

View File

@ -39,6 +39,12 @@ pub struct Settings {
#[default(None)]
#[doku(skip)]
pub opentelemetry_url: Option<Url>,
/// The number of activitypub federation workers that can be in-flight concurrently
#[default(0)]
pub worker_count: usize,
/// The number of activitypub federation retry workers that can be in-flight concurrently
#[default(0)]
pub retry_count: usize,
}
#[derive(Debug, Deserialize, Serialize, Clone, SmartDefault, Document)]

View File

@ -1,7 +1,7 @@
use crate::error::{LemmyError, LemmyResult};
use itertools::Itertools;
use once_cell::sync::Lazy;
use regex::Regex;
use regex::{Regex, RegexBuilder};
use totp_rs::{Secret, TOTP};
use url::Url;
@ -17,8 +17,13 @@ static CLEAN_URL_PARAMS_REGEX: Lazy<Regex> = Lazy::new(|| {
Regex::new(r"^utm_source|utm_medium|utm_campaign|utm_term|utm_content|gclid|gclsrc|dclid|fbclid$")
.expect("compile regex")
});
const BODY_MAX_LENGTH: usize = 10000;
const POST_BODY_MAX_LENGTH: usize = 50000;
const BIO_MAX_LENGTH: usize = 300;
const SITE_NAME_MAX_LENGTH: usize = 20;
const SITE_NAME_MIN_LENGTH: usize = 1;
const SITE_DESCRIPTION_MAX_LENGTH: usize = 150;
fn has_newline(name: &str) -> bool {
name.contains('\n')
@ -68,9 +73,14 @@ pub fn is_valid_post_title(title: &str) -> LemmyResult<()> {
}
/// This could be post bodies, comments, or any description field
pub fn is_valid_body_field(body: &Option<String>) -> LemmyResult<()> {
pub fn is_valid_body_field(body: &Option<String>, post: bool) -> LemmyResult<()> {
if let Some(body) = body {
let check = body.chars().count() <= BODY_MAX_LENGTH;
let check = if post {
body.chars().count() <= POST_BODY_MAX_LENGTH
} else {
body.chars().count() <= BODY_MAX_LENGTH
};
if !check {
Err(LemmyError::from_message("invalid_body_field"))
} else {
@ -82,14 +92,83 @@ pub fn is_valid_body_field(body: &Option<String>) -> LemmyResult<()> {
}
pub fn is_valid_bio_field(bio: &str) -> LemmyResult<()> {
let check = bio.chars().count() <= BIO_MAX_LENGTH;
if !check {
Err(LemmyError::from_message("bio_length_overflow"))
max_length_check(bio, BIO_MAX_LENGTH, String::from("bio_length_overflow"))
}
/// Checks the site name length, the limit as defined in the DB.
pub fn site_name_length_check(name: &str) -> LemmyResult<()> {
min_max_length_check(
name,
SITE_NAME_MIN_LENGTH,
SITE_NAME_MAX_LENGTH,
String::from("site_name_required"),
String::from("site_name_length_overflow"),
)
}
/// Checks the site description length, the limit as defined in the DB.
pub fn site_description_length_check(description: &str) -> LemmyResult<()> {
max_length_check(
description,
SITE_DESCRIPTION_MAX_LENGTH,
String::from("site_description_length_overflow"),
)
}
fn max_length_check(item: &str, max_length: usize, msg: String) -> LemmyResult<()> {
if item.len() > max_length {
Err(LemmyError::from_message(&msg))
} else {
Ok(())
}
}
fn min_max_length_check(
item: &str,
min_length: usize,
max_length: usize,
min_msg: String,
max_msg: String,
) -> LemmyResult<()> {
if item.len() > max_length {
Err(LemmyError::from_message(&max_msg))
} else if item.len() < min_length {
Err(LemmyError::from_message(&min_msg))
} else {
Ok(())
}
}
/// Attempts to build a regex and check it for common errors before inserting into the DB.
pub fn build_and_check_regex(regex_str_opt: &Option<&str>) -> LemmyResult<Option<Regex>> {
regex_str_opt.map_or_else(
|| Ok(None::<Regex>),
|regex_str| {
if regex_str.is_empty() {
// If the proposed regex is empty, return as having no regex at all; this is the same
// behavior that happens downstream before the write to the database.
return Ok(None::<Regex>);
}
RegexBuilder::new(regex_str)
.case_insensitive(true)
.build()
.map_err(|e| LemmyError::from_error_message(e, "invalid_regex"))
.and_then(|regex| {
// NOTE: It is difficult to know, in the universe of user-crafted regex, which ones
// may match against any string text. To keep it simple, we'll match the regex
// against an innocuous string - a single number - which should help catch a regex
// that accidentally matches against all strings.
if regex.is_match("1") {
return Err(LemmyError::from_message("permissive_regex"));
}
Ok(Some(regex))
})
},
)
}
pub fn clean_url_params(url: &Url) -> Url {
let mut url_out = url.clone();
if url.query().is_some() {
@ -171,13 +250,20 @@ pub fn check_site_visibility_valid(
mod tests {
use super::build_totp_2fa;
use crate::utils::validation::{
build_and_check_regex,
check_site_visibility_valid,
clean_url_params,
generate_totp_2fa_secret,
is_valid_actor_name,
is_valid_bio_field,
is_valid_display_name,
is_valid_matrix_id,
is_valid_post_title,
site_description_length_check,
site_name_length_check,
BIO_MAX_LENGTH,
SITE_DESCRIPTION_MAX_LENGTH,
SITE_NAME_MAX_LENGTH,
};
use url::Url;
@ -246,6 +332,126 @@ mod tests {
assert!(totp.is_ok());
}
#[test]
fn test_valid_site_name() {
let valid_names = [
(0..SITE_NAME_MAX_LENGTH).map(|_| 'A').collect::<String>(),
String::from("A"),
];
let invalid_names = [
(
&(0..SITE_NAME_MAX_LENGTH + 1)
.map(|_| 'A')
.collect::<String>(),
"site_name_length_overflow",
),
(&String::new(), "site_name_required"),
];
valid_names.iter().for_each(|valid_name| {
assert!(
site_name_length_check(valid_name).is_ok(),
"Expected {} of length {} to be Ok.",
valid_name,
valid_name.len()
)
});
invalid_names
.iter()
.for_each(|&(invalid_name, expected_err)| {
let result = site_name_length_check(invalid_name);
assert!(result.is_err());
assert!(
result
.unwrap_err()
.message
.eq(&Some(String::from(expected_err))),
"Testing {}, expected error {}",
invalid_name,
expected_err
);
});
}
#[test]
fn test_valid_bio() {
assert!(is_valid_bio_field(&(0..BIO_MAX_LENGTH).map(|_| 'A').collect::<String>()).is_ok());
let invalid_result =
is_valid_bio_field(&(0..BIO_MAX_LENGTH + 1).map(|_| 'A').collect::<String>());
assert!(
invalid_result.is_err()
&& invalid_result
.unwrap_err()
.message
.eq(&Some(String::from("bio_length_overflow")))
);
}
#[test]
fn test_valid_site_description() {
assert!(site_description_length_check(
&(0..SITE_DESCRIPTION_MAX_LENGTH)
.map(|_| 'A')
.collect::<String>()
)
.is_ok());
let invalid_result = site_description_length_check(
&(0..SITE_DESCRIPTION_MAX_LENGTH + 1)
.map(|_| 'A')
.collect::<String>(),
);
assert!(
invalid_result.is_err()
&& invalid_result
.unwrap_err()
.message
.eq(&Some(String::from("site_description_length_overflow")))
);
}
#[test]
fn test_valid_slur_regex() {
let valid_regexes = [&None, &Some(""), &Some("(foo|bar)")];
valid_regexes.iter().for_each(|regex| {
let result = build_and_check_regex(regex);
assert!(result.is_ok(), "Testing regex: {:?}", regex);
});
}
#[test]
fn test_too_permissive_slur_regex() {
let match_everything_regexes = [
(&Some("["), "invalid_regex"),
(&Some("(foo|bar|)"), "permissive_regex"),
(&Some(".*"), "permissive_regex"),
];
match_everything_regexes
.iter()
.for_each(|&(regex_str, expected_err)| {
let result = build_and_check_regex(regex_str);
assert!(result.is_err());
assert!(
result
.unwrap_err()
.message
.eq(&Some(String::from(expected_err))),
"Testing regex {:?}, expected error {}",
regex_str,
expected_err
);
});
}
#[test]
fn test_check_site_visibility_valid() {
assert!(check_site_visibility_valid(true, true, &None, &None).is_err());

@ -1 +1 @@
Subproject commit 7fc71d0860bbe5c6d620ec27112350ffe5b9229c
Subproject commit 5a9d44656e2658ab7cb2dbec3fd1bfaf57654533

View File

@ -29,7 +29,7 @@ RUN \
FROM alpine:3 as lemmy
# Install libpq for postgres
RUN apk add libpq
RUN apk add --no-cache libpq
# Copy resources
COPY --from=builder /app/lemmy_server /app/lemmy

View File

@ -6,20 +6,9 @@ x-logging: &default-logging
max-size: "50m"
max-file: 4
networks:
# communication to web and clients
lemmyexternalproxy:
# communication between lemmy services
lemmyinternal:
driver: bridge
internal: true
services:
proxy:
image: nginx:1-alpine
networks:
- lemmyinternal
- lemmyexternalproxy
ports:
# actual and only port facing any connection from outside
# Note, change the left number if port 1236 is already in use on your system
@ -45,9 +34,6 @@ services:
# RUST_RELEASE_MODE: release
# this hostname is used in nginx reverse proxy and also for lemmy ui to connect to the backend, do not change
hostname: lemmy
networks:
- lemmyinternal
- lemmyexternalproxy
restart: always
environment:
- RUST_LOG="warn,lemmy_server=debug,lemmy_api=debug,lemmy_api_common=debug,lemmy_api_crud=debug,lemmy_apub=debug,lemmy_db_schema=debug,lemmy_db_views=debug,lemmy_db_views_actor=debug,lemmy_db_views_moderator=debug,lemmy_routes=debug,lemmy_utils=debug,lemmy_websocket=debug"
@ -67,14 +53,12 @@ services:
# build:
# context: ../../lemmy-ui
# dockerfile: dev.dockerfile
networks:
- lemmyinternal
environment:
# this needs to match the hostname defined in the lemmy service
- LEMMY_UI_LEMMY_INTERNAL_HOST=lemmy:8536
# set the outside hostname here
- LEMMY_UI_LEMMY_EXTERNAL_HOST=localhost:1236
- LEMMY_HTTPS=false
- LEMMY_UI_HTTPS=false
- LEMMY_UI_DEBUG=true
depends_on:
- lemmy
@ -88,8 +72,6 @@ services:
hostname: pictrs
# we can set options to pictrs like this, here we set max. image size and forced format for conversion
# entrypoint: /sbin/tini -- /usr/local/bin/pict-rs -p /mnt -m 4 --image-format webp
networks:
- lemmyinternal
environment:
- PICTRS_OPENTELEMETRY_URL=http://otel:4137
- PICTRS__API_KEY=API_KEY
@ -126,10 +108,6 @@ services:
"-c",
"track_activity_query_size=1048576",
]
networks:
- lemmyinternal
# adding the external facing network to allow direct db access for devs
- lemmyexternalproxy
ports:
# use a different port so it doesnt conflict with potential postgres db running on the host
- "5433:5432"

View File

@ -0,0 +1 @@
alter table local_site add column federation_worker_count int default 64 not null;

View File

@ -0,0 +1 @@
alter table local_site drop column federation_worker_count;

View File

@ -0,0 +1,14 @@
-- update the default sort type
update local_user set default_sort_type = 'TopDay' where default_sort_type in ('TopThreeMonths', 'TopSixMonths', 'TopNineMonths');
-- rename the old enum
alter type sort_type_enum rename to sort_type_enum__;
-- create the new enum
CREATE TYPE sort_type_enum AS ENUM ('Active', 'Hot', 'New', 'Old', 'TopDay', 'TopWeek', 'TopMonth', 'TopYear', 'TopAll', 'MostComments', 'NewComments', 'TopHour', 'TopSixHour', 'TopTwelveHour');
-- alter all you enum columns
alter table local_user
alter column default_sort_type type sort_type_enum using default_sort_type::text::sort_type_enum;
-- drop the old enum
drop type sort_type_enum__;

View File

@ -0,0 +1,4 @@
-- Update the enums
ALTER TYPE sort_type_enum ADD VALUE 'TopThreeMonths';
ALTER TYPE sort_type_enum ADD VALUE 'TopSixMonths';
ALTER TYPE sort_type_enum ADD VALUE 'TopNineMonths';

View File

@ -0,0 +1 @@
drop table captcha_answer;

View File

@ -0,0 +1,6 @@
create table captcha_answer (
id serial primary key,
uuid uuid not null unique default gen_random_uuid(),
answer text not null,
published timestamp not null default now()
);

View File

@ -0,0 +1,2 @@
alter table only local_user alter column theme TYPE character varying(20);
alter table only local_user alter column theme set default 'browser'::character varying;

View File

@ -0,0 +1,2 @@
alter table only local_user alter column theme type text;
alter table only local_user alter column theme set default 'browser'::text;

View File

@ -0,0 +1,2 @@
drop index idx_comment_aggregates_published;
drop index idx_community_aggregates_published;

View File

@ -0,0 +1,4 @@
-- Add indexes on published column (needed for hot_rank updates)
create index idx_community_aggregates_published on community_aggregates (published desc);
create index idx_comment_aggregates_published on comment_aggregates (published desc);

View File

@ -14,7 +14,10 @@ cargo clippy --workspace --fix --allow-staged --allow-dirty --tests --all-target
-D clippy::manual_string_new -D clippy::redundant_closure_for_method_calls \
-D clippy::unused_self \
-A clippy::uninlined_format_args \
-D clippy::get_first
-D clippy::get_first \
-D clippy::explicit_into_iter_loop \
-D clippy::explicit_iter_loop \
-D clippy::needless_collect
cargo clippy --workspace --features console -- \
-D clippy::unwrap_used \

View File

@ -6,19 +6,23 @@ set -e
new_tag="$1"
third_semver=$(echo $new_tag | cut -d "." -f 3)
# Goto the upper route
CWD="$(cd -P -- "$(dirname -- "${BASH_SOURCE[0]}")" && pwd -P)"
cd $CWD/../
# The ansible and docker installs should only update for non release-candidates
# IE, when the third semver is a number, not '2-rc'
if [ ! -z "${third_semver##*[!0-9]*}" ]; then
pushd ../docker
sed -i "s/dessalines\/lemmy:.*/dessalines\/lemmy:$new_tag/" ../docker-compose.yml
sed -i "s/dessalines\/lemmy-ui:.*/dessalines\/lemmy-ui:$new_tag/" ../docker-compose.yml
sed -i "s/dessalines\/lemmy-ui:.*/dessalines\/lemmy-ui:$new_tag/" ../federation/docker-compose.yml
git add ../docker-compose.yml
git add ../federation/docker-compose.yml
pushd docker
sed -i "s/dessalines\/lemmy:.*/dessalines\/lemmy:$new_tag/" docker-compose.yml
sed -i "s/dessalines\/lemmy-ui:.*/dessalines\/lemmy-ui:$new_tag/" docker-compose.yml
sed -i "s/dessalines\/lemmy-ui:.*/dessalines\/lemmy-ui:$new_tag/" federation/docker-compose.yml
git add docker-compose.yml
git add federation/docker-compose.yml
popd
# Setting the version for Ansible
pushd ../../../lemmy-ansible
pushd ../lemmy-ansible
echo $new_tag > "VERSION"
git add "VERSION"
git commit -m"Updating VERSION"
@ -29,14 +33,16 @@ if [ ! -z "${third_semver##*[!0-9]*}" ]; then
fi
# Update crate versions
pushd ..
old_tag=$(grep version Cargo.toml | head -1 | cut -d'"' -f 2)
sed -i "s/{ version = \"=$old_tag\", path/{ version = \"=$new_tag\", path/g" Cargo.toml
sed -i "s/version = \"$old_tag\"/version = \"$new_tag\"/g" Cargo.toml
git add Cargo.toml
cargo check
git add Cargo.lock
popd
# Update the submodules
git submodule update --remote
git add crates/utils/translations
# The commit
git commit -m"Version $new_tag"

View File

@ -3,4 +3,4 @@ set -e
dest=${1-config/defaults.hjson}
cargo run -- --print-config-docs > "$dest"
cargo run --manifest-path crates/utils/Cargo.toml > "$dest"

View File

@ -38,6 +38,7 @@ use lemmy_api_common::{
ChangePassword,
DeleteAccount,
GetBannedPersons,
GetCaptcha,
GetPersonDetails,
GetPersonMentions,
GetReplies,
@ -272,6 +273,12 @@ pub fn config(cfg: &mut web::ServiceConfig, rate_limit: &RateLimitCell) {
.wrap(rate_limit.register())
.route(web::post().to(route_post_crud::<Register>)),
)
.service(
// Handle captcha separately
web::resource("/user/get_captcha")
.wrap(rate_limit.post())
.route(web::get().to(route_get::<GetCaptcha>)),
)
// User actions
.service(
web::scope("/user")

View File

@ -9,7 +9,6 @@ use crate::{code_migrations::run_advanced_migrations, root_span_builder::Quieter
use activitypub_federation::config::{FederationConfig, FederationMiddleware};
use actix_cors::Cors;
use actix_web::{middleware, web::Data, App, HttpServer, Result};
use doku::json::{AutoComments, CommentsStyle, Formatting, ObjectsStyle};
use lemmy_api_common::{
context::LemmyContext,
lemmy_db_views::structs::SiteView,
@ -25,11 +24,7 @@ use lemmy_db_schema::{
utils::{build_db_pool, get_conn, get_database_url, run_migrations},
};
use lemmy_routes::{feeds, images, nodeinfo, webfinger};
use lemmy_utils::{
error::LemmyError,
rate_limit::RateLimitCell,
settings::{structs::Settings, SETTINGS},
};
use lemmy_utils::{error::LemmyError, rate_limit::RateLimitCell, settings::SETTINGS};
use reqwest::Client;
use reqwest_middleware::ClientBuilder;
use reqwest_tracing::TracingMiddleware;
@ -47,21 +42,6 @@ pub(crate) const REQWEST_TIMEOUT: Duration = Duration::from_secs(10);
/// Placing the main function in lib.rs allows other crates to import it and embed Lemmy
pub async fn start_lemmy_server() -> Result<(), LemmyError> {
let args: Vec<String> = env::args().collect();
if args.get(1) == Some(&"--print-config-docs".to_string()) {
let fmt = Formatting {
auto_comments: AutoComments::none(),
comments_style: CommentsStyle {
separator: "#".to_owned(),
},
objects_style: ObjectsStyle {
surround_keys_with_quotes: false,
use_comma_as_separator: false,
},
..Default::default()
};
println!("{}", doku::to_json_fmt_val(&fmt, &Settings::default()));
return Ok(());
}
let scheduled_tasks_enabled = args.get(1) != Some(&"--disable-scheduled-tasks".to_string());
@ -140,24 +120,23 @@ pub async fn start_lemmy_server() -> Result<(), LemmyError> {
});
}
let settings_bind = settings.clone();
let federation_config = FederationConfig::builder()
.domain(settings.hostname.clone())
.app_data(context.clone())
.client(client.clone())
.http_fetch_limit(FEDERATION_HTTP_FETCH_LIMIT)
.worker_count(local_site.federation_worker_count as usize)
.worker_count(settings.worker_count)
.retry_count(settings.retry_count)
.debug(cfg!(debug_assertions))
.http_signature_compat(true)
.url_verifier(Box::new(VerifyUrlData(context.pool().clone())))
.build()
.await
.expect("configure federation");
.await?;
// Create Http server with websocket support
let settings_bind = settings.clone();
HttpServer::new(move || {
let context = context.clone();
let cors_config = if cfg!(debug_assertions) {
Cors::permissive()
} else {
@ -172,9 +151,10 @@ pub async fn start_lemmy_server() -> Result<(), LemmyError> {
// This is the default log format save for the usage of %{r}a over %a to guarantee to record the client's (forwarded) IP and not the last peer address, since the latter is frequently just a reverse proxy
"%{r}a '%r' %s %b '%{Referer}i' '%{User-Agent}i' %T",
))
.wrap(middleware::Compress::default())
.wrap(cors_config)
.wrap(TracingLogger::<QuieterRootSpanBuilder>::new())
.app_data(Data::new(context))
.app_data(Data::new(context.clone()))
.app_data(Data::new(rate_limit_cell.clone()))
.wrap(FederationMiddleware::new(federation_config.clone()))
// The routes

View File

@ -1,7 +1,7 @@
use lemmy_server::{init_logging, start_lemmy_server};
use lemmy_utils::{error::LemmyError, settings::SETTINGS};
#[actix_web::main]
#[tokio::main]
pub async fn main() -> Result<(), LemmyError> {
init_logging(&SETTINGS.opentelemetry_url)?;
#[cfg(not(feature = "embed-pictrs"))]

View File

@ -1,25 +1,21 @@
use chrono::NaiveDateTime;
use clokwerk::{Scheduler, TimeUnits as CTimeUnits};
use diesel::{
dsl::{now, IntervalDsl},
sql_types::{Integer, Timestamp},
Connection,
ExpressionMethods,
NullableExpressionMethods,
QueryDsl,
QueryableByName,
};
// Import week days and WeekDay
use diesel::{sql_query, PgConnection, RunQueryDsl};
use lemmy_api_common::context::LemmyContext;
use lemmy_db_schema::{
schema::{
activity,
comment_aggregates,
community_aggregates,
community_person_ban,
instance,
person,
post_aggregates,
},
schema::{activity, captcha_answer, comment, community_person_ban, instance, person, post},
source::instance::{Instance, InstanceForm},
utils::{functions::hot_rank, naive_now},
utils::{naive_now, DELETED_REPLACEMENT_TEXT},
};
use lemmy_routes::nodeinfo::NodeInfo;
use lemmy_utils::{error::LemmyError, REQWEST_TIMEOUT};
@ -46,13 +42,20 @@ pub fn setup(
update_banned_when_expired(&mut conn);
});
// Update hot ranks every 5 minutes
// Update hot ranks every 15 minutes
let url = db_url.clone();
scheduler.every(CTimeUnits::minutes(5)).run(move || {
scheduler.every(CTimeUnits::minutes(15)).run(move || {
let mut conn = PgConnection::establish(&url).expect("could not establish connection");
update_hot_ranks(&mut conn, true);
});
// Delete any captcha answers older than ten minutes, every ten minutes
let url = db_url.clone();
scheduler.every(CTimeUnits::minutes(10)).run(move || {
let mut conn = PgConnection::establish(&url).expect("could not establish connection");
delete_expired_captcha_answers(&mut conn);
});
// Clear old activities every week
let url = db_url.clone();
scheduler.every(CTimeUnits::weeks(1)).run(move || {
@ -66,6 +69,13 @@ pub fn setup(
context_1.settings_updated_channel().remove_older_than(hour);
});
// Overwrite deleted & removed posts and comments every day
let url = db_url.clone();
scheduler.every(CTimeUnits::days(1)).run(move || {
let mut conn = PgConnection::establish(&url).expect("could not establish connection");
overwrite_deleted_posts_and_comments(&mut conn);
});
// Update the Instance Software
scheduler.every(CTimeUnits::days(1)).run(move || {
let mut conn = PgConnection::establish(&db_url).expect("could not establish connection");
@ -86,67 +96,109 @@ fn startup_jobs(db_url: &str) {
update_hot_ranks(&mut conn, false);
update_banned_when_expired(&mut conn);
clear_old_activities(&mut conn);
overwrite_deleted_posts_and_comments(&mut conn);
}
/// Update the hot_rank columns for the aggregates tables
/// Runs in batches until all necessary rows are updated once
fn update_hot_ranks(conn: &mut PgConnection, last_week_only: bool) {
let mut post_update = diesel::update(post_aggregates::table).into_boxed();
let mut comment_update = diesel::update(comment_aggregates::table).into_boxed();
let mut community_update = diesel::update(community_aggregates::table).into_boxed();
// Only update for the last week of content
if last_week_only {
let process_start_time = if last_week_only {
info!("Updating hot ranks for last week...");
let last_week = now - diesel::dsl::IntervalDsl::weeks(1);
post_update = post_update.filter(post_aggregates::published.gt(last_week));
comment_update = comment_update.filter(comment_aggregates::published.gt(last_week));
community_update = community_update.filter(community_aggregates::published.gt(last_week));
naive_now() - chrono::Duration::days(7)
} else {
info!("Updating hot ranks for all history...");
}
NaiveDateTime::from_timestamp_opt(0, 0).expect("0 timestamp creation")
};
match post_update
.set((
post_aggregates::hot_rank.eq(hot_rank(post_aggregates::score, post_aggregates::published)),
post_aggregates::hot_rank_active.eq(hot_rank(
post_aggregates::score,
post_aggregates::newest_comment_time_necro,
)),
process_hot_ranks_in_batches(
conn,
"post_aggregates",
"SET hot_rank = hot_rank(a.score, a.published),
hot_rank_active = hot_rank(a.score, a.newest_comment_time_necro)",
process_start_time,
);
process_hot_ranks_in_batches(
conn,
"comment_aggregates",
"SET hot_rank = hot_rank(a.score, a.published)",
process_start_time,
);
process_hot_ranks_in_batches(
conn,
"community_aggregates",
"SET hot_rank = hot_rank(a.subscribers, a.published)",
process_start_time,
);
info!("Finished hot ranks update!");
}
#[derive(QueryableByName)]
struct HotRanksUpdateResult {
#[diesel(sql_type = Timestamp)]
published: NaiveDateTime,
}
/// Runs the hot rank update query in batches until all rows after `process_start_time` have been
/// processed.
/// In `set_clause`, "a" will refer to the current aggregates table.
/// Locked rows are skipped in order to prevent deadlocks (they will likely get updated on the next
/// run)
fn process_hot_ranks_in_batches(
conn: &mut PgConnection,
table_name: &str,
set_clause: &str,
process_start_time: NaiveDateTime,
) {
let update_batch_size = 1000; // Bigger batches than this tend to cause seq scans
let mut previous_batch_result = Some(process_start_time);
while let Some(previous_batch_last_published) = previous_batch_result {
// Raw `sql_query` is used as a performance optimization - Diesel does not support doing this
// in a single query (neither as a CTE, nor using a subquery)
let result = sql_query(format!(
r#"WITH batch AS (SELECT a.id
FROM {aggregates_table} a
WHERE a.published > $1
ORDER BY a.published
LIMIT $2
FOR UPDATE SKIP LOCKED)
UPDATE {aggregates_table} a {set_clause}
FROM batch WHERE a.id = batch.id RETURNING a.published;
"#,
aggregates_table = table_name,
set_clause = set_clause
))
.execute(conn)
{
Ok(_) => {}
Err(e) => {
error!("Failed to update post_aggregates hot_ranks: {}", e)
.bind::<Timestamp, _>(previous_batch_last_published)
.bind::<Integer, _>(update_batch_size)
.get_results::<HotRanksUpdateResult>(conn);
match result {
Ok(updated_rows) => previous_batch_result = updated_rows.last().map(|row| row.published),
Err(e) => {
error!("Failed to update {} hot_ranks: {}", table_name, e);
break;
}
}
}
info!(
"Finished process_hot_ranks_in_batches execution for {}",
table_name
);
}
match comment_update
.set(comment_aggregates::hot_rank.eq(hot_rank(
comment_aggregates::score,
comment_aggregates::published,
)))
.execute(conn)
{
Ok(_) => {}
Err(e) => {
error!("Failed to update comment_aggregates hot_ranks: {}", e)
}
}
match community_update
.set(community_aggregates::hot_rank.eq(hot_rank(
community_aggregates::subscribers,
community_aggregates::published,
)))
.execute(conn)
fn delete_expired_captcha_answers(conn: &mut PgConnection) {
match diesel::delete(
captcha_answer::table.filter(captcha_answer::published.lt(now - IntervalDsl::minutes(10))),
)
.execute(conn)
{
Ok(_) => {
info!("Done.");
}
Err(e) => {
error!("Failed to update community_aggregates hot_ranks: {}", e)
error!("Failed to clear old captcha answers: {}", e)
}
}
}
@ -166,6 +218,48 @@ fn clear_old_activities(conn: &mut PgConnection) {
}
}
/// overwrite posts and comments 30d after deletion
fn overwrite_deleted_posts_and_comments(conn: &mut PgConnection) {
info!("Overwriting deleted posts...");
match diesel::update(
post::table
.filter(post::deleted.eq(true))
.filter(post::updated.lt(now.nullable() - 1.months()))
.filter(post::body.ne(DELETED_REPLACEMENT_TEXT)),
)
.set((
post::body.eq(DELETED_REPLACEMENT_TEXT),
post::name.eq(DELETED_REPLACEMENT_TEXT),
))
.execute(conn)
{
Ok(_) => {
info!("Done.");
}
Err(e) => {
error!("Failed to overwrite deleted posts: {}", e)
}
}
info!("Overwriting deleted comments...");
match diesel::update(
comment::table
.filter(comment::deleted.eq(true))
.filter(comment::updated.lt(now.nullable() - 1.months()))
.filter(comment::content.ne(DELETED_REPLACEMENT_TEXT)),
)
.set(comment::content.eq(DELETED_REPLACEMENT_TEXT))
.execute(conn)
{
Ok(_) => {
info!("Done.");
}
Err(e) => {
error!("Failed to overwrite deleted comments: {}", e)
}
}
}
/// Re-calculate the site and community active counts every 12 hours
fn active_counts(conn: &mut PgConnection) {
info!("Updating active site and community aggregates ...");