2022-11-09 10:05:00 +00:00
|
|
|
|
use crate::{
|
|
|
|
|
diesel::Connection,
|
|
|
|
|
diesel_migrations::MigrationHarness,
|
|
|
|
|
newtypes::DbUrl,
|
|
|
|
|
CommentSortType,
|
|
|
|
|
SortType,
|
|
|
|
|
};
|
2023-03-21 15:03:05 +00:00
|
|
|
|
use activitypub_federation::{fetch::object_id::ObjectId, traits::Object};
|
2022-05-03 17:44:13 +00:00
|
|
|
|
use chrono::NaiveDateTime;
|
2023-03-03 09:30:03 +00:00
|
|
|
|
use deadpool::Runtime;
|
2022-05-03 17:44:13 +00:00
|
|
|
|
use diesel::{
|
|
|
|
|
backend::Backend,
|
|
|
|
|
deserialize::FromSql,
|
2022-09-26 14:09:32 +00:00
|
|
|
|
pg::Pg,
|
2023-06-26 08:25:38 +00:00
|
|
|
|
result::{ConnectionError, ConnectionResult, Error as DieselError, Error::QueryBuilderError},
|
2022-05-03 17:44:13 +00:00
|
|
|
|
serialize::{Output, ToSql},
|
|
|
|
|
sql_types::Text,
|
|
|
|
|
PgConnection,
|
|
|
|
|
};
|
2022-11-09 10:05:00 +00:00
|
|
|
|
use diesel_async::{
|
|
|
|
|
pg::AsyncPgConnection,
|
2023-02-28 21:45:37 +00:00
|
|
|
|
pooled_connection::{
|
|
|
|
|
deadpool::{Object as PooledConnection, Pool},
|
|
|
|
|
AsyncDieselConnectionManager,
|
|
|
|
|
},
|
2022-11-09 10:05:00 +00:00
|
|
|
|
};
|
2022-09-26 14:09:32 +00:00
|
|
|
|
use diesel_migrations::EmbeddedMigrations;
|
2023-06-26 08:25:38 +00:00
|
|
|
|
use futures_util::{future::BoxFuture, FutureExt};
|
2022-11-09 10:05:00 +00:00
|
|
|
|
use lemmy_utils::{error::LemmyError, settings::structs::Settings};
|
2022-05-03 17:44:13 +00:00
|
|
|
|
use once_cell::sync::Lazy;
|
|
|
|
|
use regex::Regex;
|
2023-06-26 08:25:38 +00:00
|
|
|
|
use rustls::{
|
|
|
|
|
client::{ServerCertVerified, ServerCertVerifier},
|
|
|
|
|
ServerName,
|
|
|
|
|
};
|
|
|
|
|
use std::{
|
|
|
|
|
env,
|
|
|
|
|
env::VarError,
|
|
|
|
|
sync::Arc,
|
|
|
|
|
time::{Duration, SystemTime},
|
|
|
|
|
};
|
|
|
|
|
use tracing::{error, info};
|
2022-05-03 17:44:13 +00:00
|
|
|
|
use url::Url;
|
|
|
|
|
|
2022-07-08 10:21:33 +00:00
|
|
|
|
const FETCH_LIMIT_DEFAULT: i64 = 10;
|
|
|
|
|
pub const FETCH_LIMIT_MAX: i64 = 50;
|
2023-03-03 09:30:03 +00:00
|
|
|
|
const POOL_TIMEOUT: Option<Duration> = Some(Duration::from_secs(5));
|
2022-07-08 10:21:33 +00:00
|
|
|
|
|
2022-11-09 10:05:00 +00:00
|
|
|
|
pub type DbPool = Pool<AsyncPgConnection>;
|
|
|
|
|
|
2023-02-28 21:45:37 +00:00
|
|
|
|
pub async fn get_conn(pool: &DbPool) -> Result<PooledConnection<AsyncPgConnection>, DieselError> {
|
2022-11-19 04:33:54 +00:00
|
|
|
|
pool.get().await.map_err(|e| QueryBuilderError(e.into()))
|
2022-11-09 10:05:00 +00:00
|
|
|
|
}
|
2022-05-03 17:44:13 +00:00
|
|
|
|
|
|
|
|
|
pub fn get_database_url_from_env() -> Result<String, VarError> {
|
|
|
|
|
env::var("LEMMY_DATABASE_URL")
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
pub fn fuzzy_search(q: &str) -> String {
|
|
|
|
|
let replaced = q.replace('%', "\\%").replace('_', "\\_").replace(' ', "%");
|
2023-01-30 19:17:24 +00:00
|
|
|
|
format!("%{replaced}%")
|
2022-05-03 17:44:13 +00:00
|
|
|
|
}
|
|
|
|
|
|
2022-07-08 10:21:33 +00:00
|
|
|
|
pub fn limit_and_offset(
|
|
|
|
|
page: Option<i64>,
|
|
|
|
|
limit: Option<i64>,
|
|
|
|
|
) -> Result<(i64, i64), diesel::result::Error> {
|
|
|
|
|
let page = match page {
|
|
|
|
|
Some(page) => {
|
|
|
|
|
if page < 1 {
|
|
|
|
|
return Err(QueryBuilderError("Page is < 1".into()));
|
|
|
|
|
} else {
|
|
|
|
|
page
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
None => 1,
|
|
|
|
|
};
|
|
|
|
|
let limit = match limit {
|
|
|
|
|
Some(limit) => {
|
|
|
|
|
if !(1..=FETCH_LIMIT_MAX).contains(&limit) {
|
|
|
|
|
return Err(QueryBuilderError(
|
2023-01-30 19:17:24 +00:00
|
|
|
|
format!("Fetch limit is > {FETCH_LIMIT_MAX}").into(),
|
2022-07-08 10:21:33 +00:00
|
|
|
|
));
|
|
|
|
|
} else {
|
|
|
|
|
limit
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
None => FETCH_LIMIT_DEFAULT,
|
|
|
|
|
};
|
2022-05-03 17:44:13 +00:00
|
|
|
|
let offset = limit * (page - 1);
|
2022-07-08 10:21:33 +00:00
|
|
|
|
Ok((limit, offset))
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
pub fn limit_and_offset_unlimited(page: Option<i64>, limit: Option<i64>) -> (i64, i64) {
|
|
|
|
|
let limit = limit.unwrap_or(FETCH_LIMIT_DEFAULT);
|
|
|
|
|
let offset = limit * (page.unwrap_or(1) - 1);
|
2022-05-03 17:44:13 +00:00
|
|
|
|
(limit, offset)
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
pub fn is_email_regex(test: &str) -> bool {
|
|
|
|
|
EMAIL_REGEX.is_match(test)
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
pub fn diesel_option_overwrite(opt: &Option<String>) -> Option<Option<String>> {
|
|
|
|
|
match opt {
|
|
|
|
|
// An empty string is an erase
|
|
|
|
|
Some(unwrapped) => {
|
|
|
|
|
if !unwrapped.eq("") {
|
2022-11-19 04:33:54 +00:00
|
|
|
|
Some(Some(unwrapped.clone()))
|
2022-05-03 17:44:13 +00:00
|
|
|
|
} else {
|
|
|
|
|
Some(None)
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
None => None,
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
pub fn diesel_option_overwrite_to_url(
|
|
|
|
|
opt: &Option<String>,
|
|
|
|
|
) -> Result<Option<Option<DbUrl>>, LemmyError> {
|
2022-11-19 04:33:54 +00:00
|
|
|
|
match opt.as_ref().map(std::string::String::as_str) {
|
2022-05-03 17:44:13 +00:00
|
|
|
|
// An empty string is an erase
|
|
|
|
|
Some("") => Ok(Some(None)),
|
|
|
|
|
Some(str_url) => match Url::parse(str_url) {
|
|
|
|
|
Ok(url) => Ok(Some(Some(url.into()))),
|
|
|
|
|
Err(e) => Err(LemmyError::from_error_message(e, "invalid_url")),
|
|
|
|
|
},
|
|
|
|
|
None => Ok(None),
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
2022-10-27 09:24:07 +00:00
|
|
|
|
pub fn diesel_option_overwrite_to_url_create(
|
|
|
|
|
opt: &Option<String>,
|
|
|
|
|
) -> Result<Option<DbUrl>, LemmyError> {
|
2022-11-19 04:33:54 +00:00
|
|
|
|
match opt.as_ref().map(std::string::String::as_str) {
|
2022-10-27 09:24:07 +00:00
|
|
|
|
// An empty string is nothing
|
|
|
|
|
Some("") => Ok(None),
|
|
|
|
|
Some(str_url) => match Url::parse(str_url) {
|
|
|
|
|
Ok(url) => Ok(Some(url.into())),
|
|
|
|
|
Err(e) => Err(LemmyError::from_error_message(e, "invalid_url")),
|
|
|
|
|
},
|
|
|
|
|
None => Ok(None),
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
2022-11-09 10:05:00 +00:00
|
|
|
|
async fn build_db_pool_settings_opt(settings: Option<&Settings>) -> Result<DbPool, LemmyError> {
|
|
|
|
|
let db_url = get_database_url(settings);
|
|
|
|
|
let pool_size = settings.map(|s| s.database.pool_size).unwrap_or(5);
|
2023-06-26 08:25:38 +00:00
|
|
|
|
// We only support TLS with sslmode=require currently
|
|
|
|
|
let tls_enabled = db_url.contains("sslmode=require");
|
|
|
|
|
let manager = if tls_enabled {
|
|
|
|
|
// diesel-async does not support any TLS connections out of the box, so we need to manually
|
|
|
|
|
// provide a setup function which handles creating the connection
|
|
|
|
|
AsyncDieselConnectionManager::<AsyncPgConnection>::new_with_setup(&db_url, establish_connection)
|
|
|
|
|
} else {
|
|
|
|
|
AsyncDieselConnectionManager::<AsyncPgConnection>::new(&db_url)
|
|
|
|
|
};
|
2023-03-03 09:30:03 +00:00
|
|
|
|
let pool = Pool::builder(manager)
|
|
|
|
|
.max_size(pool_size)
|
|
|
|
|
.wait_timeout(POOL_TIMEOUT)
|
|
|
|
|
.create_timeout(POOL_TIMEOUT)
|
|
|
|
|
.recycle_timeout(POOL_TIMEOUT)
|
|
|
|
|
.runtime(Runtime::Tokio1)
|
|
|
|
|
.build()?;
|
2022-11-09 10:05:00 +00:00
|
|
|
|
|
|
|
|
|
// If there's no settings, that means its a unit test, and migrations need to be run
|
|
|
|
|
if settings.is_none() {
|
|
|
|
|
run_migrations(&db_url);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
Ok(pool)
|
|
|
|
|
}
|
|
|
|
|
|
2023-06-26 08:25:38 +00:00
|
|
|
|
fn establish_connection(config: &str) -> BoxFuture<ConnectionResult<AsyncPgConnection>> {
|
|
|
|
|
let fut = async {
|
|
|
|
|
let rustls_config = rustls::ClientConfig::builder()
|
|
|
|
|
.with_safe_defaults()
|
|
|
|
|
.with_custom_certificate_verifier(Arc::new(NoCertVerifier {}))
|
|
|
|
|
.with_no_client_auth();
|
|
|
|
|
|
|
|
|
|
let tls = tokio_postgres_rustls::MakeRustlsConnect::new(rustls_config);
|
|
|
|
|
let (client, conn) = tokio_postgres::connect(config, tls)
|
|
|
|
|
.await
|
|
|
|
|
.map_err(|e| ConnectionError::BadConnection(e.to_string()))?;
|
|
|
|
|
tokio::spawn(async move {
|
|
|
|
|
if let Err(e) = conn.await {
|
|
|
|
|
error!("Database connection failed: {e}");
|
|
|
|
|
}
|
|
|
|
|
});
|
|
|
|
|
AsyncPgConnection::try_from(client).await
|
|
|
|
|
};
|
|
|
|
|
fut.boxed()
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
struct NoCertVerifier {}
|
|
|
|
|
|
|
|
|
|
impl ServerCertVerifier for NoCertVerifier {
|
|
|
|
|
fn verify_server_cert(
|
|
|
|
|
&self,
|
|
|
|
|
_end_entity: &rustls::Certificate,
|
|
|
|
|
_intermediates: &[rustls::Certificate],
|
|
|
|
|
_server_name: &ServerName,
|
|
|
|
|
_scts: &mut dyn Iterator<Item = &[u8]>,
|
|
|
|
|
_ocsp_response: &[u8],
|
|
|
|
|
_now: SystemTime,
|
|
|
|
|
) -> Result<ServerCertVerified, rustls::Error> {
|
|
|
|
|
// Will verify all (even invalid) certs without any checks (sslmode=require)
|
|
|
|
|
Ok(ServerCertVerified::assertion())
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
2022-09-26 14:09:32 +00:00
|
|
|
|
pub const MIGRATIONS: EmbeddedMigrations = embed_migrations!();
|
2022-05-03 17:44:13 +00:00
|
|
|
|
|
2022-11-09 10:05:00 +00:00
|
|
|
|
pub fn run_migrations(db_url: &str) {
|
|
|
|
|
// Needs to be a sync connection
|
2022-09-26 14:09:32 +00:00
|
|
|
|
let mut conn =
|
2023-02-05 17:30:10 +00:00
|
|
|
|
PgConnection::establish(db_url).unwrap_or_else(|e| panic!("Error connecting to {db_url}: {e}"));
|
2023-01-23 14:59:25 +00:00
|
|
|
|
info!("Running Database migrations (This may take a long time)...");
|
2022-09-26 14:09:32 +00:00
|
|
|
|
let _ = &mut conn
|
|
|
|
|
.run_pending_migrations(MIGRATIONS)
|
2023-02-05 17:30:10 +00:00
|
|
|
|
.unwrap_or_else(|e| panic!("Couldn't run DB Migrations: {e}"));
|
2023-01-23 14:59:25 +00:00
|
|
|
|
info!("Database migrations complete.");
|
2022-11-09 10:05:00 +00:00
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
pub async fn build_db_pool(settings: &Settings) -> Result<DbPool, LemmyError> {
|
|
|
|
|
build_db_pool_settings_opt(Some(settings)).await
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
pub async fn build_db_pool_for_tests() -> DbPool {
|
|
|
|
|
build_db_pool_settings_opt(None)
|
|
|
|
|
.await
|
|
|
|
|
.expect("db pool missing")
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
pub fn get_database_url(settings: Option<&Settings>) -> String {
|
|
|
|
|
// The env var should override anything in the settings config
|
|
|
|
|
match get_database_url_from_env() {
|
|
|
|
|
Ok(url) => url,
|
|
|
|
|
Err(e) => match settings {
|
|
|
|
|
Some(settings) => settings.get_database_url(),
|
2023-01-30 19:17:24 +00:00
|
|
|
|
None => panic!("Failed to read database URL from env var LEMMY_DATABASE_URL: {e}"),
|
2022-11-09 10:05:00 +00:00
|
|
|
|
},
|
|
|
|
|
}
|
2022-05-03 17:44:13 +00:00
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
pub fn naive_now() -> NaiveDateTime {
|
|
|
|
|
chrono::prelude::Utc::now().naive_utc()
|
|
|
|
|
}
|
|
|
|
|
|
2022-07-30 03:55:59 +00:00
|
|
|
|
pub fn post_to_comment_sort_type(sort: SortType) -> CommentSortType {
|
|
|
|
|
match sort {
|
|
|
|
|
SortType::Active | SortType::Hot => CommentSortType::Hot,
|
|
|
|
|
SortType::New | SortType::NewComments | SortType::MostComments => CommentSortType::New,
|
|
|
|
|
SortType::Old => CommentSortType::Old,
|
2023-06-20 14:05:43 +00:00
|
|
|
|
SortType::TopHour
|
|
|
|
|
| SortType::TopSixHour
|
|
|
|
|
| SortType::TopTwelveHour
|
|
|
|
|
| SortType::TopDay
|
2022-07-30 03:55:59 +00:00
|
|
|
|
| SortType::TopAll
|
|
|
|
|
| SortType::TopWeek
|
|
|
|
|
| SortType::TopYear
|
2023-06-26 19:03:35 +00:00
|
|
|
|
| SortType::TopMonth
|
|
|
|
|
| SortType::TopThreeMonths
|
|
|
|
|
| SortType::TopSixMonths
|
|
|
|
|
| SortType::TopNineMonths => CommentSortType::Top,
|
2022-07-30 03:55:59 +00:00
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
2022-05-03 17:44:13 +00:00
|
|
|
|
static EMAIL_REGEX: Lazy<Regex> = Lazy::new(|| {
|
|
|
|
|
Regex::new(r"^[a-zA-Z0-9.!#$%&’*+/=?^_`{|}~-]+@[a-zA-Z0-9-]+(?:\.[a-zA-Z0-9-]+)*$")
|
|
|
|
|
.expect("compile email regex")
|
|
|
|
|
});
|
|
|
|
|
|
|
|
|
|
pub mod functions {
|
2022-11-19 04:33:54 +00:00
|
|
|
|
use diesel::sql_types::{BigInt, Text, Timestamp};
|
2022-05-03 17:44:13 +00:00
|
|
|
|
|
|
|
|
|
sql_function! {
|
|
|
|
|
fn hot_rank(score: BigInt, time: Timestamp) -> Integer;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
sql_function!(fn lower(x: Text) -> Text);
|
|
|
|
|
}
|
|
|
|
|
|
2023-06-20 06:17:54 +00:00
|
|
|
|
pub const DELETED_REPLACEMENT_TEXT: &str = "*Permanently Deleted*";
|
|
|
|
|
|
2022-09-26 14:09:32 +00:00
|
|
|
|
impl ToSql<Text, Pg> for DbUrl {
|
|
|
|
|
fn to_sql(&self, out: &mut Output<Pg>) -> diesel::serialize::Result {
|
|
|
|
|
<std::string::String as ToSql<Text, Pg>>::to_sql(&self.0.to_string(), &mut out.reborrow())
|
2022-05-03 17:44:13 +00:00
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
impl<DB: Backend> FromSql<Text, DB> for DbUrl
|
|
|
|
|
where
|
|
|
|
|
String: FromSql<Text, DB>,
|
|
|
|
|
{
|
2023-06-07 10:54:40 +00:00
|
|
|
|
fn from_sql(value: DB::RawValue<'_>) -> diesel::deserialize::Result<Self> {
|
2022-09-26 14:09:32 +00:00
|
|
|
|
let str = String::from_sql(value)?;
|
2023-02-18 14:50:28 +00:00
|
|
|
|
Ok(DbUrl(Box::new(Url::parse(&str)?)))
|
2022-05-03 17:44:13 +00:00
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
impl<Kind> From<ObjectId<Kind>> for DbUrl
|
|
|
|
|
where
|
2023-03-21 15:03:05 +00:00
|
|
|
|
Kind: Object + Send + 'static,
|
|
|
|
|
for<'de2> <Kind as Object>::Kind: serde::Deserialize<'de2>,
|
2022-05-03 17:44:13 +00:00
|
|
|
|
{
|
|
|
|
|
fn from(id: ObjectId<Kind>) -> Self {
|
2023-02-18 14:50:28 +00:00
|
|
|
|
DbUrl(Box::new(id.into()))
|
2022-05-03 17:44:13 +00:00
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
#[cfg(test)]
|
|
|
|
|
mod tests {
|
|
|
|
|
use super::{fuzzy_search, *};
|
|
|
|
|
use crate::utils::is_email_regex;
|
|
|
|
|
|
|
|
|
|
#[test]
|
|
|
|
|
fn test_fuzzy_search() {
|
|
|
|
|
let test = "This %is% _a_ fuzzy search";
|
|
|
|
|
assert_eq!(
|
|
|
|
|
fuzzy_search(test),
|
|
|
|
|
"%This%\\%is\\%%\\_a\\_%fuzzy%search%".to_string()
|
|
|
|
|
);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
#[test]
|
|
|
|
|
fn test_email() {
|
|
|
|
|
assert!(is_email_regex("gush@gmail.com"));
|
|
|
|
|
assert!(!is_email_regex("nada_neutho"));
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
#[test]
|
|
|
|
|
fn test_diesel_option_overwrite() {
|
|
|
|
|
assert_eq!(diesel_option_overwrite(&None), None);
|
2022-11-19 04:33:54 +00:00
|
|
|
|
assert_eq!(diesel_option_overwrite(&Some(String::new())), Some(None));
|
2022-05-03 17:44:13 +00:00
|
|
|
|
assert_eq!(
|
|
|
|
|
diesel_option_overwrite(&Some("test".to_string())),
|
|
|
|
|
Some(Some("test".to_string()))
|
|
|
|
|
);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
#[test]
|
|
|
|
|
fn test_diesel_option_overwrite_to_url() {
|
|
|
|
|
assert!(matches!(diesel_option_overwrite_to_url(&None), Ok(None)));
|
|
|
|
|
assert!(matches!(
|
2022-11-19 04:33:54 +00:00
|
|
|
|
diesel_option_overwrite_to_url(&Some(String::new())),
|
2022-05-03 17:44:13 +00:00
|
|
|
|
Ok(Some(None))
|
|
|
|
|
));
|
2023-07-04 10:15:27 +00:00
|
|
|
|
assert!(diesel_option_overwrite_to_url(&Some("invalid_url".to_string())).is_err());
|
2022-05-03 17:44:13 +00:00
|
|
|
|
let example_url = "https://example.com";
|
|
|
|
|
assert!(matches!(
|
|
|
|
|
diesel_option_overwrite_to_url(&Some(example_url.to_string())),
|
|
|
|
|
Ok(Some(Some(url))) if url == Url::parse(example_url).unwrap().into()
|
|
|
|
|
));
|
|
|
|
|
}
|
|
|
|
|
}
|