de-global services
Signed-off-by: Jason Volk <jason@zemos.net>
This commit is contained in:
parent
7e50db4193
commit
2f85a5c1ac
36 changed files with 327 additions and 336 deletions
|
@ -2,7 +2,7 @@ use conduit::Result;
|
|||
use conduit_macros::implement;
|
||||
use ruma::events::room::message::RoomMessageEventContent;
|
||||
|
||||
use crate::{services, Command};
|
||||
use crate::Command;
|
||||
|
||||
/// Uses the iterator in `src/database/key_value/users.rs` to iterator over
|
||||
/// every user in our database (remote and local). Reports total count, any
|
||||
|
@ -10,7 +10,7 @@ use crate::{services, Command};
|
|||
#[implement(Command, params = "<'_>")]
|
||||
pub(super) async fn check_all_users(&self) -> Result<RoomMessageEventContent> {
|
||||
let timer = tokio::time::Instant::now();
|
||||
let results = services().users.db.iter();
|
||||
let results = self.services.users.db.iter();
|
||||
let query_time = timer.elapsed();
|
||||
|
||||
let users = results.collect::<Vec<_>>();
|
||||
|
|
|
@ -1,4 +1,4 @@
|
|||
use std::{panic::AssertUnwindSafe, time::Instant};
|
||||
use std::{panic::AssertUnwindSafe, sync::Arc, time::Instant};
|
||||
|
||||
use clap::{CommandFactory, Parser};
|
||||
use conduit::{error, trace, utils::string::common_prefix, Error, Result};
|
||||
|
@ -17,35 +17,25 @@ use service::{
|
|||
|
||||
use crate::{admin, admin::AdminCommand, Command};
|
||||
|
||||
struct Handler {
|
||||
services: &'static Services,
|
||||
}
|
||||
#[must_use]
|
||||
pub(super) fn complete(line: &str) -> String { complete_command(AdminCommand::command(), line) }
|
||||
|
||||
#[must_use]
|
||||
pub(super) fn complete(line: &str) -> String {
|
||||
Handler {
|
||||
services: service::services(),
|
||||
pub(super) fn handle(services: Arc<Services>, command: CommandInput) -> HandlerResult {
|
||||
Box::pin(handle_command(services, command))
|
||||
}
|
||||
.complete_command(AdminCommand::command(), line)
|
||||
}
|
||||
|
||||
#[must_use]
|
||||
pub(super) fn handle(command: CommandInput) -> HandlerResult { Box::pin(handle_command(command)) }
|
||||
|
||||
#[tracing::instrument(skip_all, name = "admin")]
|
||||
async fn handle_command(command: CommandInput) -> CommandResult {
|
||||
AssertUnwindSafe(Box::pin(process_command(&command)))
|
||||
async fn handle_command(services: Arc<Services>, command: CommandInput) -> CommandResult {
|
||||
AssertUnwindSafe(Box::pin(process_command(services, &command)))
|
||||
.catch_unwind()
|
||||
.await
|
||||
.map_err(Error::from_panic)
|
||||
.or_else(|error| handle_panic(&error, command))
|
||||
}
|
||||
|
||||
async fn process_command(command: &CommandInput) -> CommandOutput {
|
||||
Handler {
|
||||
services: service::services(),
|
||||
}
|
||||
.process(&command.command)
|
||||
async fn process_command(services: Arc<Services>, command: &CommandInput) -> CommandOutput {
|
||||
process(services, &command.command)
|
||||
.await
|
||||
.and_then(|content| reply(content, command.reply_id.clone()))
|
||||
}
|
||||
|
@ -68,15 +58,14 @@ fn reply(mut content: RoomMessageEventContent, reply_id: Option<OwnedEventId>) -
|
|||
Some(content)
|
||||
}
|
||||
|
||||
impl Handler {
|
||||
// Parse and process a message from the admin room
|
||||
async fn process(&self, msg: &str) -> CommandOutput {
|
||||
async fn process(services: Arc<Services>, msg: &str) -> CommandOutput {
|
||||
let mut lines = msg.lines().filter(|l| !l.trim().is_empty());
|
||||
let command = lines.next().expect("each string has at least one line");
|
||||
let (parsed, body) = match self.parse_command(command) {
|
||||
let (parsed, body) = match parse_command(command) {
|
||||
Ok(parsed) => parsed,
|
||||
Err(error) => {
|
||||
let server_name = self.services.globals.server_name();
|
||||
let server_name = services.globals.server_name();
|
||||
let message = error.replace("server.name", server_name.as_str());
|
||||
return Some(RoomMessageEventContent::notice_markdown(message));
|
||||
},
|
||||
|
@ -85,7 +74,7 @@ impl Handler {
|
|||
let timer = Instant::now();
|
||||
let body: Vec<&str> = body.iter().map(String::as_str).collect();
|
||||
let context = Command {
|
||||
services: self.services,
|
||||
services: &services,
|
||||
body: &body,
|
||||
};
|
||||
let result = Box::pin(admin::process(parsed, &context)).await;
|
||||
|
@ -100,14 +89,14 @@ impl Handler {
|
|||
}
|
||||
|
||||
// Parse chat messages from the admin room into an AdminCommand object
|
||||
fn parse_command(&self, command_line: &str) -> Result<(AdminCommand, Vec<String>), String> {
|
||||
let argv = self.parse_line(command_line);
|
||||
fn parse_command(command_line: &str) -> Result<(AdminCommand, Vec<String>), String> {
|
||||
let argv = parse_line(command_line);
|
||||
let com = AdminCommand::try_parse_from(&argv).map_err(|error| error.to_string())?;
|
||||
Ok((com, argv))
|
||||
}
|
||||
|
||||
fn complete_command(&self, mut cmd: clap::Command, line: &str) -> String {
|
||||
let argv = self.parse_line(line);
|
||||
fn complete_command(mut cmd: clap::Command, line: &str) -> String {
|
||||
let argv = parse_line(line);
|
||||
let mut ret = Vec::<String>::with_capacity(argv.len().saturating_add(1));
|
||||
|
||||
'token: for token in argv.into_iter().skip(1) {
|
||||
|
@ -150,7 +139,7 @@ impl Handler {
|
|||
}
|
||||
|
||||
// Parse chat messages from the admin room into an AdminCommand object
|
||||
fn parse_line(&self, command_line: &str) -> Vec<String> {
|
||||
fn parse_line(command_line: &str) -> Vec<String> {
|
||||
let mut argv = command_line
|
||||
.split_whitespace()
|
||||
.map(str::to_owned)
|
||||
|
@ -162,8 +151,7 @@ impl Handler {
|
|||
}
|
||||
|
||||
// First indice has to be "admin" but for console convenience we add it here
|
||||
let server_user = self.services.globals.server_user.as_str();
|
||||
if !argv.is_empty() && !argv[0].ends_with("admin") && !argv[0].starts_with(server_user) {
|
||||
if !argv.is_empty() && !argv[0].ends_with("admin") && !argv[0].starts_with('@') {
|
||||
argv.insert(0, "admin".to_owned());
|
||||
}
|
||||
|
||||
|
@ -193,4 +181,3 @@ impl Handler {
|
|||
trace!(?command_line, ?argv, "parse");
|
||||
argv
|
||||
}
|
||||
}
|
||||
|
|
|
@ -1,4 +1,4 @@
|
|||
#![recursion_limit = "168"]
|
||||
#![recursion_limit = "192"]
|
||||
#![allow(clippy::wildcard_imports)]
|
||||
#![allow(clippy::enum_glob_use)]
|
||||
|
||||
|
@ -24,7 +24,6 @@ extern crate conduit_service as service;
|
|||
|
||||
pub(crate) use conduit::Result;
|
||||
pub(crate) use conduit_macros::{admin_command, admin_command_dispatch};
|
||||
pub(crate) use service::services;
|
||||
|
||||
pub(crate) use crate::{
|
||||
command::Command,
|
||||
|
@ -38,26 +37,19 @@ conduit::mod_dtor! {}
|
|||
conduit::rustc_flags_capture! {}
|
||||
|
||||
/// Install the admin command handler
|
||||
pub async fn init() {
|
||||
_ = services()
|
||||
.admin
|
||||
pub async fn init(admin_service: &service::admin::Service) {
|
||||
_ = admin_service
|
||||
.complete
|
||||
.write()
|
||||
.expect("locked for writing")
|
||||
.insert(handler::complete);
|
||||
_ = services()
|
||||
.admin
|
||||
.handle
|
||||
.write()
|
||||
.await
|
||||
.insert(handler::handle);
|
||||
_ = admin_service.handle.write().await.insert(handler::handle);
|
||||
}
|
||||
|
||||
/// Uninstall the admin command handler
|
||||
pub async fn fini() {
|
||||
_ = services().admin.handle.write().await.take();
|
||||
_ = services()
|
||||
.admin
|
||||
pub async fn fini(admin_service: &service::admin::Service) {
|
||||
_ = admin_service.handle.write().await.take();
|
||||
_ = admin_service
|
||||
.complete
|
||||
.write()
|
||||
.expect("locked for writing")
|
||||
|
|
|
@ -370,7 +370,7 @@ pub(crate) async fn register_route(
|
|||
|
||||
if let Some(room_id_server_name) = room.server_name() {
|
||||
if let Err(e) = join_room_by_id_helper(
|
||||
services,
|
||||
&services,
|
||||
&user_id,
|
||||
room,
|
||||
Some("Automatically joining this room upon registration".to_owned()),
|
||||
|
@ -562,11 +562,11 @@ pub(crate) async fn deactivate_route(
|
|||
.rooms_joined(sender_user)
|
||||
.filter_map(Result::ok)
|
||||
.collect();
|
||||
super::update_displayname(services, sender_user.clone(), None, all_joined_rooms.clone()).await?;
|
||||
super::update_avatar_url(services, sender_user.clone(), None, None, all_joined_rooms).await?;
|
||||
super::update_displayname(&services, sender_user.clone(), None, all_joined_rooms.clone()).await?;
|
||||
super::update_avatar_url(&services, sender_user.clone(), None, None, all_joined_rooms).await?;
|
||||
|
||||
// Make the user leave all rooms before deactivation
|
||||
super::leave_all_rooms(services, sender_user).await;
|
||||
super::leave_all_rooms(&services, sender_user).await;
|
||||
|
||||
info!("User {sender_user} deactivated their account.");
|
||||
services
|
||||
|
|
|
@ -107,7 +107,7 @@ pub(crate) async fn get_alias_route(
|
|||
return Err(Error::BadRequest(ErrorKind::NotFound, "Room with alias not found."));
|
||||
};
|
||||
|
||||
let servers = room_available_servers(services, &room_id, &room_alias, &pre_servers);
|
||||
let servers = room_available_servers(&services, &room_id, &room_alias, &pre_servers);
|
||||
debug!(?room_alias, ?room_id, "available servers: {servers:?}");
|
||||
|
||||
Ok(get_alias::v3::Response::new(room_id, servers))
|
||||
|
|
|
@ -20,7 +20,7 @@ pub(crate) async fn set_global_account_data_route(
|
|||
State(services): State<crate::State>, body: Ruma<set_global_account_data::v3::Request>,
|
||||
) -> Result<set_global_account_data::v3::Response> {
|
||||
set_account_data(
|
||||
services,
|
||||
&services,
|
||||
None,
|
||||
&body.sender_user,
|
||||
&body.event_type.to_string(),
|
||||
|
@ -37,7 +37,7 @@ pub(crate) async fn set_room_account_data_route(
|
|||
State(services): State<crate::State>, body: Ruma<set_room_account_data::v3::Request>,
|
||||
) -> Result<set_room_account_data::v3::Response> {
|
||||
set_account_data(
|
||||
services,
|
||||
&services,
|
||||
Some(&body.room_id),
|
||||
&body.sender_user,
|
||||
&body.event_type.to_string(),
|
||||
|
|
|
@ -48,7 +48,7 @@ pub(crate) async fn get_public_rooms_filtered_route(
|
|||
}
|
||||
|
||||
let response = get_public_rooms_filtered_helper(
|
||||
services,
|
||||
&services,
|
||||
body.server.as_deref(),
|
||||
body.limit,
|
||||
body.since.as_deref(),
|
||||
|
@ -88,7 +88,7 @@ pub(crate) async fn get_public_rooms_route(
|
|||
}
|
||||
|
||||
let response = get_public_rooms_filtered_helper(
|
||||
services,
|
||||
&services,
|
||||
body.server.as_deref(),
|
||||
body.limit,
|
||||
body.since.as_deref(),
|
||||
|
@ -124,7 +124,7 @@ pub(crate) async fn set_room_visibility_route(
|
|||
return Err(Error::BadRequest(ErrorKind::NotFound, "Room not found"));
|
||||
}
|
||||
|
||||
if !user_can_publish_room(services, sender_user, &body.room_id)? {
|
||||
if !user_can_publish_room(&services, sender_user, &body.room_id)? {
|
||||
return Err(Error::BadRequest(
|
||||
ErrorKind::forbidden(),
|
||||
"User is not allowed to publish this room",
|
||||
|
|
|
@ -77,7 +77,7 @@ pub(crate) async fn get_keys_route(
|
|||
let sender_user = body.sender_user.as_ref().expect("user is authenticated");
|
||||
|
||||
get_keys_helper(
|
||||
services,
|
||||
&services,
|
||||
Some(sender_user),
|
||||
&body.device_keys,
|
||||
|u| u == sender_user,
|
||||
|
@ -92,7 +92,7 @@ pub(crate) async fn get_keys_route(
|
|||
pub(crate) async fn claim_keys_route(
|
||||
State(services): State<crate::State>, body: Ruma<claim_keys::v3::Request>,
|
||||
) -> Result<claim_keys::v3::Response> {
|
||||
claim_keys_helper(services, &body.one_time_keys).await
|
||||
claim_keys_helper(&services, &body.one_time_keys).await
|
||||
}
|
||||
|
||||
/// # `POST /_matrix/client/r0/keys/device_signing/upload`
|
||||
|
|
|
@ -76,12 +76,12 @@ pub(crate) async fn get_media_preview_route(
|
|||
let sender_user = body.sender_user.as_ref().expect("user is authenticated");
|
||||
|
||||
let url = &body.url;
|
||||
if !url_preview_allowed(services, url) {
|
||||
if !url_preview_allowed(&services, url) {
|
||||
warn!(%sender_user, "URL is not allowed to be previewed: {url}");
|
||||
return Err(Error::BadRequest(ErrorKind::forbidden(), "URL is not allowed to be previewed"));
|
||||
}
|
||||
|
||||
match get_url_preview(services, url).await {
|
||||
match get_url_preview(&services, url).await {
|
||||
Ok(preview) => {
|
||||
let res = serde_json::value::to_raw_value(&preview).map_err(|e| {
|
||||
error!(%sender_user, "Failed to convert UrlPreviewData into a serde json value: {e}");
|
||||
|
@ -221,7 +221,7 @@ pub(crate) async fn get_content_route(
|
|||
})
|
||||
} else if !services.globals.server_is_ours(&body.server_name) && body.allow_remote {
|
||||
let response = get_remote_content(
|
||||
services,
|
||||
&services,
|
||||
&mxc,
|
||||
&body.server_name,
|
||||
body.media_id.clone(),
|
||||
|
@ -311,7 +311,7 @@ pub(crate) async fn get_content_as_filename_route(
|
|||
})
|
||||
} else if !services.globals.server_is_ours(&body.server_name) && body.allow_remote {
|
||||
match get_remote_content(
|
||||
services,
|
||||
&services,
|
||||
&mxc,
|
||||
&body.server_name,
|
||||
body.media_id.clone(),
|
||||
|
|
|
@ -167,7 +167,7 @@ pub(crate) async fn join_room_by_id_route(
|
|||
let sender_user = body.sender_user.as_ref().expect("user is authenticated");
|
||||
|
||||
banned_room_check(
|
||||
services,
|
||||
&services,
|
||||
sender_user,
|
||||
Some(&body.room_id),
|
||||
body.room_id.server_name(),
|
||||
|
@ -202,7 +202,7 @@ pub(crate) async fn join_room_by_id_route(
|
|||
}
|
||||
|
||||
join_room_by_id_helper(
|
||||
services,
|
||||
&services,
|
||||
sender_user,
|
||||
&body.room_id,
|
||||
body.reason.clone(),
|
||||
|
@ -231,7 +231,7 @@ pub(crate) async fn join_room_by_id_or_alias_route(
|
|||
|
||||
let (servers, room_id) = match OwnedRoomId::try_from(body.room_id_or_alias) {
|
||||
Ok(room_id) => {
|
||||
banned_room_check(services, sender_user, Some(&room_id), room_id.server_name(), client).await?;
|
||||
banned_room_check(&services, sender_user, Some(&room_id), room_id.server_name(), client).await?;
|
||||
|
||||
let mut servers = body.server_name.clone();
|
||||
servers.extend(
|
||||
|
@ -270,7 +270,7 @@ pub(crate) async fn join_room_by_id_or_alias_route(
|
|||
.await?;
|
||||
let (room_id, mut pre_servers) = response;
|
||||
|
||||
banned_room_check(services, sender_user, Some(&room_id), Some(room_alias.server_name()), client).await?;
|
||||
banned_room_check(&services, sender_user, Some(&room_id), Some(room_alias.server_name()), client).await?;
|
||||
|
||||
let mut servers = body.server_name;
|
||||
if let Some(pre_servers) = &mut pre_servers {
|
||||
|
@ -303,7 +303,7 @@ pub(crate) async fn join_room_by_id_or_alias_route(
|
|||
};
|
||||
|
||||
let join_room_response = join_room_by_id_helper(
|
||||
services,
|
||||
&services,
|
||||
sender_user,
|
||||
&room_id,
|
||||
body.reason.clone(),
|
||||
|
@ -327,7 +327,7 @@ pub(crate) async fn leave_room_route(
|
|||
) -> Result<leave_room::v3::Response> {
|
||||
let sender_user = body.sender_user.as_ref().expect("user is authenticated");
|
||||
|
||||
leave_room(services, sender_user, &body.room_id, body.reason.clone()).await?;
|
||||
leave_room(&services, sender_user, &body.room_id, body.reason.clone()).await?;
|
||||
|
||||
Ok(leave_room::v3::Response::new())
|
||||
}
|
||||
|
@ -353,13 +353,13 @@ pub(crate) async fn invite_user_route(
|
|||
));
|
||||
}
|
||||
|
||||
banned_room_check(services, sender_user, Some(&body.room_id), body.room_id.server_name(), client).await?;
|
||||
banned_room_check(&services, sender_user, Some(&body.room_id), body.room_id.server_name(), client).await?;
|
||||
|
||||
if let invite_user::v3::InvitationRecipient::UserId {
|
||||
user_id,
|
||||
} = &body.recipient
|
||||
{
|
||||
invite_helper(services, sender_user, user_id, &body.room_id, body.reason.clone(), false).await?;
|
||||
invite_helper(&services, sender_user, user_id, &body.room_id, body.reason.clone(), false).await?;
|
||||
Ok(invite_user::v3::Response {})
|
||||
} else {
|
||||
Err(Error::BadRequest(ErrorKind::NotFound, "User not found."))
|
||||
|
|
|
@ -146,7 +146,7 @@ pub(crate) async fn get_message_events_route(
|
|||
.timeline
|
||||
.pdus_after(sender_user, &body.room_id, from)?
|
||||
.filter_map(Result::ok) // Filter out buggy events
|
||||
.filter(|(_, pdu)| { contains_url_filter(pdu, &body.filter) && visibility_filter(services, pdu, sender_user, &body.room_id)
|
||||
.filter(|(_, pdu)| { contains_url_filter(pdu, &body.filter) && visibility_filter(&services, pdu, sender_user, &body.room_id)
|
||||
|
||||
})
|
||||
.take_while(|&(k, _)| Some(k) != to) // Stop at `to`
|
||||
|
@ -193,7 +193,7 @@ pub(crate) async fn get_message_events_route(
|
|||
.timeline
|
||||
.pdus_until(sender_user, &body.room_id, from)?
|
||||
.filter_map(Result::ok) // Filter out buggy events
|
||||
.filter(|(_, pdu)| {contains_url_filter(pdu, &body.filter) && visibility_filter(services, pdu, sender_user, &body.room_id)})
|
||||
.filter(|(_, pdu)| {contains_url_filter(pdu, &body.filter) && visibility_filter(&services, pdu, sender_user, &body.room_id)})
|
||||
.take_while(|&(k, _)| Some(k) != to) // Stop at `to`
|
||||
.take(limit)
|
||||
.collect();
|
||||
|
|
|
@ -33,7 +33,7 @@ pub(crate) async fn set_displayname_route(
|
|||
.filter_map(Result::ok)
|
||||
.collect();
|
||||
|
||||
update_displayname(services, sender_user.clone(), body.displayname.clone(), all_joined_rooms).await?;
|
||||
update_displayname(&services, sender_user.clone(), body.displayname.clone(), all_joined_rooms).await?;
|
||||
|
||||
if services.globals.allow_local_presence() {
|
||||
// Presence update
|
||||
|
@ -118,7 +118,7 @@ pub(crate) async fn set_avatar_url_route(
|
|||
.collect();
|
||||
|
||||
update_avatar_url(
|
||||
services,
|
||||
&services,
|
||||
sender_user.clone(),
|
||||
body.avatar_url.clone(),
|
||||
body.blurhash.clone(),
|
||||
|
|
|
@ -40,7 +40,7 @@ pub(crate) async fn report_event_route(
|
|||
};
|
||||
|
||||
is_report_valid(
|
||||
services,
|
||||
&services,
|
||||
&pdu.event_id,
|
||||
&body.room_id,
|
||||
sender_user,
|
||||
|
|
|
@ -79,7 +79,7 @@ pub(crate) async fn create_room_route(
|
|||
}
|
||||
|
||||
let room_id: OwnedRoomId = if let Some(custom_room_id) = &body.room_id {
|
||||
custom_room_id_check(services, custom_room_id)?
|
||||
custom_room_id_check(&services, custom_room_id)?
|
||||
} else {
|
||||
RoomId::new(&services.globals.config.server_name)
|
||||
};
|
||||
|
@ -96,7 +96,7 @@ pub(crate) async fn create_room_route(
|
|||
let state_lock = services.rooms.state.mutex.lock(&room_id).await;
|
||||
|
||||
let alias: Option<OwnedRoomAliasId> = if let Some(alias) = &body.room_alias_name {
|
||||
Some(room_alias_check(services, alias, &body.appservice_info).await?)
|
||||
Some(room_alias_check(&services, alias, &body.appservice_info).await?)
|
||||
} else {
|
||||
None
|
||||
};
|
||||
|
@ -438,7 +438,7 @@ pub(crate) async fn create_room_route(
|
|||
// 8. Events implied by invite (and TODO: invite_3pid)
|
||||
drop(state_lock);
|
||||
for user_id in &body.invite {
|
||||
if let Err(e) = invite_helper(services, sender_user, user_id, &room_id, None, body.is_direct).await {
|
||||
if let Err(e) = invite_helper(&services, sender_user, user_id, &room_id, None, body.is_direct).await {
|
||||
warn!(%e, "Failed to send invite");
|
||||
}
|
||||
}
|
||||
|
|
|
@ -37,7 +37,7 @@ pub(crate) async fn send_state_event_for_key_route(
|
|||
|
||||
Ok(send_state_event::v3::Response {
|
||||
event_id: send_state_event_for_key_helper(
|
||||
services,
|
||||
&services,
|
||||
sender_user,
|
||||
&body.room_id,
|
||||
&body.event_type,
|
||||
|
|
|
@ -137,7 +137,7 @@ pub(crate) async fn sync_events_route(
|
|||
);
|
||||
|
||||
if services.globals.allow_local_presence() {
|
||||
process_presence_updates(services, &mut presence_updates, since, &sender_user).await?;
|
||||
process_presence_updates(&services, &mut presence_updates, since, &sender_user).await?;
|
||||
}
|
||||
|
||||
let all_joined_rooms = services
|
||||
|
@ -152,7 +152,7 @@ pub(crate) async fn sync_events_route(
|
|||
for room_id in all_joined_rooms {
|
||||
let room_id = room_id?;
|
||||
if let Ok(joined_room) = load_joined_room(
|
||||
services,
|
||||
&services,
|
||||
&sender_user,
|
||||
&sender_device,
|
||||
&room_id,
|
||||
|
@ -182,7 +182,7 @@ pub(crate) async fn sync_events_route(
|
|||
.collect();
|
||||
for result in all_left_rooms {
|
||||
handle_left_room(
|
||||
services,
|
||||
&services,
|
||||
since,
|
||||
&result?.0,
|
||||
&sender_user,
|
||||
|
@ -1214,7 +1214,7 @@ pub(crate) async fn sync_events_v4_route(
|
|||
match new_membership {
|
||||
MembershipState::Join => {
|
||||
// A new user joined an encrypted room
|
||||
if !share_encrypted_room(services, &sender_user, &user_id, room_id)? {
|
||||
if !share_encrypted_room(&services, &sender_user, &user_id, room_id)? {
|
||||
device_list_changes.insert(user_id);
|
||||
}
|
||||
},
|
||||
|
@ -1243,7 +1243,7 @@ pub(crate) async fn sync_events_v4_route(
|
|||
.filter(|user_id| {
|
||||
// Only send keys if the sender doesn't share an encrypted room with the target
|
||||
// already
|
||||
!share_encrypted_room(services, &sender_user, user_id, room_id).unwrap_or(false)
|
||||
!share_encrypted_room(&services, &sender_user, user_id, room_id).unwrap_or(false)
|
||||
}),
|
||||
);
|
||||
}
|
||||
|
@ -1407,7 +1407,8 @@ pub(crate) async fn sync_events_v4_route(
|
|||
for (room_id, (required_state_request, timeline_limit, roomsince)) in &todo_rooms {
|
||||
let roomsincecount = PduCount::Normal(*roomsince);
|
||||
|
||||
let (timeline_pdus, limited) = load_timeline(services, &sender_user, room_id, roomsincecount, *timeline_limit)?;
|
||||
let (timeline_pdus, limited) =
|
||||
load_timeline(&services, &sender_user, room_id, roomsincecount, *timeline_limit)?;
|
||||
|
||||
if roomsince != &0 && timeline_pdus.is_empty() {
|
||||
continue;
|
||||
|
|
|
@ -1,4 +1,4 @@
|
|||
#![recursion_limit = "160"]
|
||||
#![recursion_limit = "192"]
|
||||
|
||||
pub mod client;
|
||||
pub mod router;
|
||||
|
|
|
@ -4,6 +4,8 @@ mod handler;
|
|||
mod request;
|
||||
mod response;
|
||||
|
||||
use std::sync::Arc;
|
||||
|
||||
use axum::{
|
||||
response::IntoResponse,
|
||||
routing::{any, get, post},
|
||||
|
@ -16,7 +18,7 @@ use self::handler::RouterExt;
|
|||
pub(super) use self::{args::Args as Ruma, response::RumaResponse};
|
||||
use crate::{client, server};
|
||||
|
||||
pub type State = &'static service::Services;
|
||||
pub type State = Arc<service::Services>;
|
||||
|
||||
pub fn build(router: Router<State>, server: &Server) -> Router<State> {
|
||||
let config = &server.config;
|
||||
|
|
|
@ -7,7 +7,7 @@ use ruma::{api::IncomingRequest, CanonicalJsonValue, OwnedDeviceId, OwnedServerN
|
|||
use service::Services;
|
||||
|
||||
use super::{auth, auth::Auth, request, request::Request};
|
||||
use crate::service::appservice::RegistrationInfo;
|
||||
use crate::{service::appservice::RegistrationInfo, State};
|
||||
|
||||
/// Extractor for Ruma request structs
|
||||
pub(crate) struct Args<T> {
|
||||
|
@ -36,14 +36,13 @@ pub(crate) struct Args<T> {
|
|||
}
|
||||
|
||||
#[async_trait]
|
||||
impl<T, S> FromRequest<S, Body> for Args<T>
|
||||
impl<T> FromRequest<State, Body> for Args<T>
|
||||
where
|
||||
T: IncomingRequest,
|
||||
{
|
||||
type Rejection = Error;
|
||||
|
||||
async fn from_request(request: hyper::Request<Body>, _: &S) -> Result<Self, Self::Rejection> {
|
||||
let services = service::services(); // ???
|
||||
async fn from_request(request: hyper::Request<Body>, services: &State) -> Result<Self, Self::Rejection> {
|
||||
let mut request = request::from(services, request).await?;
|
||||
let mut json_body = serde_json::from_slice::<CanonicalJsonValue>(&request.body).ok();
|
||||
let auth = auth::auth(services, &mut request, &json_body, &T::METADATA).await?;
|
||||
|
|
|
@ -82,7 +82,7 @@ pub(crate) async fn create_join_event_template_route(
|
|||
.state_cache
|
||||
.is_left(&body.user_id, &body.room_id)
|
||||
.unwrap_or(true))
|
||||
&& user_can_perform_restricted_join(services, &body.user_id, &body.room_id, &room_version_id)?
|
||||
&& user_can_perform_restricted_join(&services, &body.user_id, &body.room_id, &room_version_id)?
|
||||
{
|
||||
let auth_user = services
|
||||
.rooms
|
||||
|
|
|
@ -26,7 +26,7 @@ pub(crate) async fn get_public_rooms_filtered_route(
|
|||
}
|
||||
|
||||
let response = crate::client::get_public_rooms_filtered_helper(
|
||||
services,
|
||||
&services,
|
||||
None,
|
||||
body.limit,
|
||||
body.since.as_deref(),
|
||||
|
@ -60,7 +60,7 @@ pub(crate) async fn get_public_rooms_route(
|
|||
}
|
||||
|
||||
let response = crate::client::get_public_rooms_filtered_helper(
|
||||
services,
|
||||
&services,
|
||||
None,
|
||||
body.limit,
|
||||
body.since.as_deref(),
|
||||
|
|
|
@ -62,8 +62,8 @@ pub(crate) async fn send_transaction_message_route(
|
|||
"Starting txn",
|
||||
);
|
||||
|
||||
let resolved_map = handle_pdus(services, &client, &body, origin, &txn_start_time).await?;
|
||||
handle_edus(services, &client, &body, origin).await?;
|
||||
let resolved_map = handle_pdus(&services, &client, &body, origin, &txn_start_time).await?;
|
||||
handle_edus(&services, &client, &body, origin).await?;
|
||||
|
||||
debug!(
|
||||
pdus = ?body.pdus.len(),
|
||||
|
|
|
@ -241,7 +241,7 @@ pub(crate) async fn create_join_event_v1_route(
|
|||
}
|
||||
}
|
||||
|
||||
let room_state = create_join_event(services, origin, &body.room_id, &body.pdu).await?;
|
||||
let room_state = create_join_event(&services, origin, &body.room_id, &body.pdu).await?;
|
||||
|
||||
Ok(create_join_event::v1::Response {
|
||||
room_state,
|
||||
|
@ -286,7 +286,7 @@ pub(crate) async fn create_join_event_v2_route(
|
|||
auth_chain,
|
||||
state,
|
||||
event,
|
||||
} = create_join_event(services, origin, &body.room_id, &body.pdu).await?;
|
||||
} = create_join_event(&services, origin, &body.room_id, &body.pdu).await?;
|
||||
let room_state = create_join_event::v2::RoomState {
|
||||
members_omitted: false,
|
||||
auth_chain,
|
||||
|
|
|
@ -28,7 +28,7 @@ pub(crate) async fn create_leave_event_v1_route(
|
|||
) -> Result<create_leave_event::v1::Response> {
|
||||
let origin = body.origin.as_ref().expect("server is authenticated");
|
||||
|
||||
create_leave_event(services, origin, &body.room_id, &body.pdu).await?;
|
||||
create_leave_event(&services, origin, &body.room_id, &body.pdu).await?;
|
||||
|
||||
Ok(create_leave_event::v1::Response::new())
|
||||
}
|
||||
|
@ -41,7 +41,7 @@ pub(crate) async fn create_leave_event_v2_route(
|
|||
) -> Result<create_leave_event::v2::Response> {
|
||||
let origin = body.origin.as_ref().expect("server is authenticated");
|
||||
|
||||
create_leave_event(services, origin, &body.room_id, &body.pdu).await?;
|
||||
create_leave_event(&services, origin, &body.room_id, &body.pdu).await?;
|
||||
|
||||
Ok(create_leave_event::v2::Response::new())
|
||||
}
|
||||
|
|
|
@ -84,7 +84,7 @@ pub(crate) async fn get_keys_route(
|
|||
}
|
||||
|
||||
let result = get_keys_helper(
|
||||
services,
|
||||
&services,
|
||||
None,
|
||||
&body.device_keys,
|
||||
|u| Some(u.server_name()) == body.origin.as_deref(),
|
||||
|
@ -116,7 +116,7 @@ pub(crate) async fn claim_keys_route(
|
|||
));
|
||||
}
|
||||
|
||||
let result = claim_keys_helper(services, &body.one_time_keys).await?;
|
||||
let result = claim_keys_helper(&services, &body.one_time_keys).await?;
|
||||
|
||||
Ok(claim_keys::v1::Response {
|
||||
one_time_keys: result.one_time_keys,
|
||||
|
|
|
@ -11,12 +11,12 @@ pub struct Database {
|
|||
|
||||
impl Database {
|
||||
/// Load an existing database or create a new one.
|
||||
pub async fn open(server: &Arc<Server>) -> Result<Self> {
|
||||
pub async fn open(server: &Arc<Server>) -> Result<Arc<Self>> {
|
||||
let db = Engine::open(server)?;
|
||||
Ok(Self {
|
||||
Ok(Arc::new(Self {
|
||||
db: db.clone(),
|
||||
map: maps::open(&db)?,
|
||||
})
|
||||
}))
|
||||
}
|
||||
|
||||
#[inline]
|
||||
|
|
|
@ -1,3 +1,5 @@
|
|||
#![recursion_limit = "192"]
|
||||
|
||||
pub(crate) mod clap;
|
||||
mod mods;
|
||||
mod restart;
|
||||
|
@ -57,17 +59,38 @@ fn main() -> Result<(), Error> {
|
|||
async fn async_main(server: &Arc<Server>) -> Result<(), Error> {
|
||||
extern crate conduit_router as router;
|
||||
|
||||
if let Err(error) = router::start(&server.server).await {
|
||||
match router::start(&server.server).await {
|
||||
Ok(services) => server.services.lock().await.insert(services),
|
||||
Err(error) => {
|
||||
error!("Critical error starting server: {error}");
|
||||
return Err(error);
|
||||
}
|
||||
},
|
||||
};
|
||||
|
||||
if let Err(error) = router::run(&server.server).await {
|
||||
if let Err(error) = router::run(
|
||||
server
|
||||
.services
|
||||
.lock()
|
||||
.await
|
||||
.as_ref()
|
||||
.expect("services initialized"),
|
||||
)
|
||||
.await
|
||||
{
|
||||
error!("Critical error running server: {error}");
|
||||
return Err(error);
|
||||
}
|
||||
|
||||
if let Err(error) = router::stop(&server.server).await {
|
||||
if let Err(error) = router::stop(
|
||||
server
|
||||
.services
|
||||
.lock()
|
||||
.await
|
||||
.take()
|
||||
.expect("services initialied"),
|
||||
)
|
||||
.await
|
||||
{
|
||||
error!("Critical error stopping server: {error}");
|
||||
return Err(error);
|
||||
}
|
||||
|
|
|
@ -1,7 +1,7 @@
|
|||
use std::sync::Arc;
|
||||
|
||||
use conduit::{config::Config, info, log::Log, utils::sys, Error, Result};
|
||||
use tokio::runtime;
|
||||
use tokio::{runtime, sync::Mutex};
|
||||
|
||||
use crate::{clap::Args, tracing::TracingFlameGuard};
|
||||
|
||||
|
@ -10,6 +10,8 @@ pub(crate) struct Server {
|
|||
/// Server runtime state; public portion
|
||||
pub(crate) server: Arc<conduit::Server>,
|
||||
|
||||
pub(crate) services: Mutex<Option<Arc<conduit_service::Services>>>,
|
||||
|
||||
_tracing_flame_guard: TracingFlameGuard,
|
||||
|
||||
#[cfg(feature = "sentry_telemetry")]
|
||||
|
@ -54,6 +56,8 @@ impl Server {
|
|||
},
|
||||
)),
|
||||
|
||||
services: None.into(),
|
||||
|
||||
_tracing_flame_guard: tracing_flame_guard,
|
||||
|
||||
#[cfg(feature = "sentry_telemetry")]
|
||||
|
|
|
@ -6,6 +6,7 @@ use axum::{
|
|||
};
|
||||
use axum_client_ip::SecureClientIpSource;
|
||||
use conduit::{error, Result, Server};
|
||||
use conduit_service::Services;
|
||||
use http::{
|
||||
header::{self, HeaderName},
|
||||
HeaderValue, Method, StatusCode,
|
||||
|
@ -34,7 +35,8 @@ const CONDUWUIT_CSP: &[&str] = &[
|
|||
|
||||
const CONDUWUIT_PERMISSIONS_POLICY: &[&str] = &["interest-cohort=()", "browsing-topics=()"];
|
||||
|
||||
pub(crate) fn build(server: &Arc<Server>) -> Result<Router> {
|
||||
pub(crate) fn build(services: &Arc<Services>) -> Result<Router> {
|
||||
let server = &services.server;
|
||||
let layers = ServiceBuilder::new();
|
||||
|
||||
#[cfg(feature = "sentry_telemetry")]
|
||||
|
@ -83,7 +85,7 @@ pub(crate) fn build(server: &Arc<Server>) -> Result<Router> {
|
|||
.layer(body_limit_layer(server))
|
||||
.layer(CatchPanicLayer::custom(catch_panic));
|
||||
|
||||
Ok(router::build(server).layer(layers))
|
||||
Ok(router::build(services).layer(layers))
|
||||
}
|
||||
|
||||
#[cfg(any(feature = "zstd_compression", feature = "gzip_compression", feature = "brotli_compression"))]
|
||||
|
@ -151,12 +153,14 @@ fn body_limit_layer(server: &Server) -> DefaultBodyLimit { DefaultBodyLimit::max
|
|||
#[allow(clippy::needless_pass_by_value)]
|
||||
#[tracing::instrument(skip_all, name = "panic")]
|
||||
fn catch_panic(err: Box<dyn Any + Send + 'static>) -> http::Response<http_body_util::Full<bytes::Bytes>> {
|
||||
//TODO: XXX
|
||||
/*
|
||||
conduit_service::services()
|
||||
.server
|
||||
.metrics
|
||||
.requests_panic
|
||||
.fetch_add(1, std::sync::atomic::Ordering::Release);
|
||||
|
||||
*/
|
||||
let details = if let Some(s) = err.downcast_ref::<String>() {
|
||||
s.clone()
|
||||
} else if let Some(s) = err.downcast_ref::<&str>() {
|
||||
|
|
|
@ -11,22 +11,23 @@ extern crate conduit_core as conduit;
|
|||
use std::{future::Future, pin::Pin, sync::Arc};
|
||||
|
||||
use conduit::{Result, Server};
|
||||
use conduit_service::Services;
|
||||
|
||||
conduit::mod_ctor! {}
|
||||
conduit::mod_dtor! {}
|
||||
conduit::rustc_flags_capture! {}
|
||||
|
||||
#[no_mangle]
|
||||
pub extern "Rust" fn start(server: &Arc<Server>) -> Pin<Box<dyn Future<Output = Result<()>> + Send>> {
|
||||
pub extern "Rust" fn start(server: &Arc<Server>) -> Pin<Box<dyn Future<Output = Result<Arc<Services>>> + Send>> {
|
||||
Box::pin(run::start(server.clone()))
|
||||
}
|
||||
|
||||
#[no_mangle]
|
||||
pub extern "Rust" fn stop(server: &Arc<Server>) -> Pin<Box<dyn Future<Output = Result<()>> + Send>> {
|
||||
Box::pin(run::stop(server.clone()))
|
||||
pub extern "Rust" fn stop(services: Arc<Services>) -> Pin<Box<dyn Future<Output = Result<()>> + Send>> {
|
||||
Box::pin(run::stop(services))
|
||||
}
|
||||
|
||||
#[no_mangle]
|
||||
pub extern "Rust" fn run(server: &Arc<Server>) -> Pin<Box<dyn Future<Output = Result<()>> + Send>> {
|
||||
Box::pin(run::run(server.clone()))
|
||||
pub extern "Rust" fn run(services: &Arc<Services>) -> Pin<Box<dyn Future<Output = Result<()>> + Send>> {
|
||||
Box::pin(run::run(services.clone()))
|
||||
}
|
||||
|
|
|
@ -1,20 +1,20 @@
|
|||
use std::sync::Arc;
|
||||
|
||||
use axum::{response::IntoResponse, routing::get, Router};
|
||||
use conduit::{Error, Server};
|
||||
use conduit::Error;
|
||||
use conduit_api::State;
|
||||
use conduit_service::Services;
|
||||
use http::{StatusCode, Uri};
|
||||
use ruma::api::client::error::ErrorKind;
|
||||
|
||||
extern crate conduit_api as api;
|
||||
extern crate conduit_service as service;
|
||||
pub(crate) fn build(services: &Arc<Services>) -> Router {
|
||||
let router = Router::<State>::new();
|
||||
let state = services.clone();
|
||||
|
||||
pub(crate) fn build(server: &Arc<Server>) -> Router {
|
||||
let router = Router::<api::State>::new();
|
||||
|
||||
api::router::build(router, server)
|
||||
conduit_api::router::build(router, &services.server)
|
||||
.route("/", get(it_works))
|
||||
.fallback(not_found)
|
||||
.with_state(service::services())
|
||||
.with_state(state)
|
||||
}
|
||||
|
||||
async fn not_found(_uri: Uri) -> impl IntoResponse {
|
||||
|
|
|
@ -1,28 +1,30 @@
|
|||
use std::{sync::Arc, time::Duration};
|
||||
extern crate conduit_admin as admin;
|
||||
extern crate conduit_core as conduit;
|
||||
extern crate conduit_service as service;
|
||||
|
||||
use std::{
|
||||
sync::{atomic::Ordering, Arc},
|
||||
time::Duration,
|
||||
};
|
||||
|
||||
use axum_server::Handle as ServerHandle;
|
||||
use conduit::{debug, debug_error, debug_info, error, info, Error, Result, Server};
|
||||
use service::Services;
|
||||
use tokio::{
|
||||
sync::broadcast::{self, Sender},
|
||||
task::JoinHandle,
|
||||
};
|
||||
|
||||
extern crate conduit_admin as admin;
|
||||
extern crate conduit_core as conduit;
|
||||
extern crate conduit_service as service;
|
||||
|
||||
use std::sync::atomic::Ordering;
|
||||
|
||||
use conduit::{debug, debug_info, error, info, Error, Result, Server};
|
||||
|
||||
use crate::serve;
|
||||
|
||||
/// Main loop base
|
||||
#[tracing::instrument(skip_all)]
|
||||
pub(crate) async fn run(server: Arc<Server>) -> Result<()> {
|
||||
pub(crate) async fn run(services: Arc<Services>) -> Result<()> {
|
||||
let server = &services.server;
|
||||
debug!("Start");
|
||||
|
||||
// Install the admin room callback here for now
|
||||
admin::init().await;
|
||||
admin::init(&services.admin).await;
|
||||
|
||||
// Setup shutdown/signal handling
|
||||
let handle = ServerHandle::new();
|
||||
|
@ -33,13 +35,13 @@ pub(crate) async fn run(server: Arc<Server>) -> Result<()> {
|
|||
|
||||
let mut listener = server
|
||||
.runtime()
|
||||
.spawn(serve::serve(server.clone(), handle.clone(), tx.subscribe()));
|
||||
.spawn(serve::serve(services.clone(), handle.clone(), tx.subscribe()));
|
||||
|
||||
// Focal point
|
||||
debug!("Running");
|
||||
let res = tokio::select! {
|
||||
res = &mut listener => res.map_err(Error::from).unwrap_or_else(Err),
|
||||
res = service::services().poll() => handle_services_poll(&server, res, listener).await,
|
||||
res = services.poll() => handle_services_poll(server, res, listener).await,
|
||||
};
|
||||
|
||||
// Join the signal handler before we leave.
|
||||
|
@ -47,7 +49,7 @@ pub(crate) async fn run(server: Arc<Server>) -> Result<()> {
|
|||
_ = sigs.await;
|
||||
|
||||
// Remove the admin room callback
|
||||
admin::fini().await;
|
||||
admin::fini(&services.admin).await;
|
||||
|
||||
debug_info!("Finish");
|
||||
res
|
||||
|
@ -55,26 +57,33 @@ pub(crate) async fn run(server: Arc<Server>) -> Result<()> {
|
|||
|
||||
/// Async initializations
|
||||
#[tracing::instrument(skip_all)]
|
||||
pub(crate) async fn start(server: Arc<Server>) -> Result<()> {
|
||||
pub(crate) async fn start(server: Arc<Server>) -> Result<Arc<Services>> {
|
||||
debug!("Starting...");
|
||||
|
||||
service::start(&server).await?;
|
||||
let services = Services::build(server).await?.start().await?;
|
||||
|
||||
#[cfg(feature = "systemd")]
|
||||
sd_notify::notify(true, &[sd_notify::NotifyState::Ready]).expect("failed to notify systemd of ready state");
|
||||
|
||||
debug!("Started");
|
||||
Ok(())
|
||||
Ok(services)
|
||||
}
|
||||
|
||||
/// Async destructions
|
||||
#[tracing::instrument(skip_all)]
|
||||
pub(crate) async fn stop(_server: Arc<Server>) -> Result<()> {
|
||||
pub(crate) async fn stop(services: Arc<Services>) -> Result<()> {
|
||||
debug!("Shutting down...");
|
||||
|
||||
// Wait for all completions before dropping or we'll lose them to the module
|
||||
// unload and explode.
|
||||
service::stop().await;
|
||||
services.stop().await;
|
||||
|
||||
if let Err(services) = Arc::try_unwrap(services) {
|
||||
debug_error!(
|
||||
"{} dangling references to Services after shutdown",
|
||||
Arc::strong_count(&services)
|
||||
);
|
||||
}
|
||||
|
||||
debug!("Cleaning up...");
|
||||
|
||||
|
|
|
@ -5,22 +5,26 @@ mod unix;
|
|||
use std::sync::Arc;
|
||||
|
||||
use axum_server::Handle as ServerHandle;
|
||||
use conduit::{Result, Server};
|
||||
use conduit::Result;
|
||||
use conduit_service::Services;
|
||||
use tokio::sync::broadcast;
|
||||
|
||||
use crate::layers;
|
||||
use super::layers;
|
||||
|
||||
/// Serve clients
|
||||
pub(super) async fn serve(server: Arc<Server>, handle: ServerHandle, shutdown: broadcast::Receiver<()>) -> Result<()> {
|
||||
pub(super) async fn serve(
|
||||
services: Arc<Services>, handle: ServerHandle, shutdown: broadcast::Receiver<()>,
|
||||
) -> Result<()> {
|
||||
let server = &services.server;
|
||||
let config = &server.config;
|
||||
let addrs = config.get_bind_addrs();
|
||||
let app = layers::build(&server)?;
|
||||
let app = layers::build(&services)?;
|
||||
|
||||
if cfg!(unix) && config.unix_socket_path.is_some() {
|
||||
unix::serve(&server, app, shutdown).await
|
||||
unix::serve(server, app, shutdown).await
|
||||
} else if config.tls.is_some() {
|
||||
tls::serve(&server, app, handle, addrs).await
|
||||
tls::serve(server, app, handle, addrs).await
|
||||
} else {
|
||||
plain::serve(&server, app, handle, addrs).await
|
||||
plain::serve(server, app, handle, addrs).await
|
||||
}
|
||||
}
|
||||
|
|
|
@ -9,7 +9,7 @@ use std::{
|
|||
};
|
||||
|
||||
use async_trait::async_trait;
|
||||
use conduit::{debug, error, error::default_log, pdu::PduBuilder, Error, PduEvent, Result, Server};
|
||||
use conduit::{debug, error, error::default_log, pdu::PduBuilder, Err, Error, PduEvent, Result, Server};
|
||||
pub use create::create_admin_room;
|
||||
use loole::{Receiver, Sender};
|
||||
use ruma::{
|
||||
|
@ -41,6 +41,7 @@ struct Services {
|
|||
timeline: Dep<rooms::timeline::Service>,
|
||||
state: Dep<rooms::state::Service>,
|
||||
state_cache: Dep<rooms::state_cache::Service>,
|
||||
services: StdRwLock<Option<Arc<crate::Services>>>,
|
||||
}
|
||||
|
||||
#[derive(Debug)]
|
||||
|
@ -50,7 +51,7 @@ pub struct CommandInput {
|
|||
}
|
||||
|
||||
pub type Completer = fn(&str) -> String;
|
||||
pub type Handler = fn(CommandInput) -> HandlerResult;
|
||||
pub type Handler = fn(Arc<crate::Services>, CommandInput) -> HandlerResult;
|
||||
pub type HandlerResult = Pin<Box<dyn Future<Output = CommandResult> + Send>>;
|
||||
pub type CommandResult = Result<CommandOutput, Error>;
|
||||
pub type CommandOutput = Option<RoomMessageEventContent>;
|
||||
|
@ -69,6 +70,7 @@ impl crate::Service for Service {
|
|||
timeline: args.depend::<rooms::timeline::Service>("rooms::timeline"),
|
||||
state: args.depend::<rooms::state::Service>("rooms::state"),
|
||||
state_cache: args.depend::<rooms::state_cache::Service>("rooms::state_cache"),
|
||||
services: None.into(),
|
||||
},
|
||||
sender,
|
||||
receiver: Mutex::new(receiver),
|
||||
|
@ -172,10 +174,14 @@ impl Service {
|
|||
}
|
||||
|
||||
async fn process_command(&self, command: CommandInput) -> CommandResult {
|
||||
let Some(services) = self.services.services.read().expect("locked").clone() else {
|
||||
return Err!("Services self-reference not initialized.");
|
||||
};
|
||||
|
||||
if let Some(handle) = self.handle.read().await.as_ref() {
|
||||
handle(command).await
|
||||
handle(services, command).await
|
||||
} else {
|
||||
Err(Error::Err("Admin module is not loaded.".into()))
|
||||
Err!("Admin module is not loaded.")
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -356,4 +362,10 @@ impl Service {
|
|||
#[cfg(feature = "console")]
|
||||
self.console.close().await;
|
||||
}
|
||||
|
||||
/// Sets the self-reference to crate::Services which will provide context to
|
||||
/// the admin commands.
|
||||
pub(super) fn set_services(&self, services: Option<Arc<crate::Services>>) {
|
||||
*self.services.services.write().expect("locked for writing") = services;
|
||||
}
|
||||
}
|
||||
|
|
|
@ -1,4 +1,4 @@
|
|||
#![recursion_limit = "160"]
|
||||
#![recursion_limit = "192"]
|
||||
#![allow(refining_impl_trait)]
|
||||
|
||||
mod manager;
|
||||
|
@ -26,11 +26,7 @@ pub mod users;
|
|||
extern crate conduit_core as conduit;
|
||||
extern crate conduit_database as database;
|
||||
|
||||
use std::sync::{Arc, RwLock};
|
||||
|
||||
pub use conduit::{pdu, PduBuilder, PduCount, PduEvent};
|
||||
use conduit::{Result, Server};
|
||||
use database::Database;
|
||||
pub(crate) use service::{Args, Dep, Service};
|
||||
|
||||
pub use crate::services::Services;
|
||||
|
@ -38,50 +34,3 @@ pub use crate::services::Services;
|
|||
conduit::mod_ctor! {}
|
||||
conduit::mod_dtor! {}
|
||||
conduit::rustc_flags_capture! {}
|
||||
|
||||
static SERVICES: RwLock<Option<&Services>> = RwLock::new(None);
|
||||
|
||||
pub async fn start(server: &Arc<Server>) -> Result<()> {
|
||||
let d = Arc::new(Database::open(server).await?);
|
||||
let s = Box::new(Services::build(server.clone(), d)?);
|
||||
_ = SERVICES.write().expect("write locked").insert(Box::leak(s));
|
||||
|
||||
services().start().await
|
||||
}
|
||||
|
||||
pub async fn stop() {
|
||||
services().stop().await;
|
||||
|
||||
// Deactivate services(). Any further use will panic the caller.
|
||||
let s = SERVICES
|
||||
.write()
|
||||
.expect("write locked")
|
||||
.take()
|
||||
.expect("services initialized");
|
||||
|
||||
let s: *mut Services = std::ptr::from_ref(s).cast_mut();
|
||||
//SAFETY: Services was instantiated in init() and leaked into the SERVICES
|
||||
// global perusing as 'static for the duration of service. Now we reclaim
|
||||
// it to drop it before unloading the module. If this is not done there wil
|
||||
// be multiple instances after module reload.
|
||||
let s = unsafe { Box::from_raw(s) };
|
||||
|
||||
// Drop it so we encounter any trouble before the infolog message
|
||||
drop(s);
|
||||
}
|
||||
|
||||
#[must_use]
|
||||
pub fn services() -> &'static Services {
|
||||
SERVICES
|
||||
.read()
|
||||
.expect("SERVICES locked for reading")
|
||||
.expect("SERVICES initialized with Services instance")
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn available() -> bool {
|
||||
SERVICES
|
||||
.read()
|
||||
.expect("SERVICES locked for reading")
|
||||
.is_some()
|
||||
}
|
||||
|
|
|
@ -44,7 +44,8 @@ pub struct Services {
|
|||
|
||||
impl Services {
|
||||
#[allow(clippy::cognitive_complexity)]
|
||||
pub fn build(server: Arc<Server>, db: Arc<Database>) -> Result<Self> {
|
||||
pub async fn build(server: Arc<Server>) -> Result<Arc<Self>> {
|
||||
let db = Database::open(&server).await?;
|
||||
let service: Arc<Map> = Arc::new(RwLock::new(BTreeMap::new()));
|
||||
macro_rules! build {
|
||||
($tyname:ty) => {{
|
||||
|
@ -58,7 +59,7 @@ impl Services {
|
|||
}};
|
||||
}
|
||||
|
||||
Ok(Self {
|
||||
Ok(Arc::new(Self {
|
||||
account_data: build!(account_data::Service),
|
||||
admin: build!(admin::Service),
|
||||
appservice: build!(appservice::Service),
|
||||
|
@ -102,12 +103,13 @@ impl Services {
|
|||
service,
|
||||
server,
|
||||
db,
|
||||
})
|
||||
}))
|
||||
}
|
||||
|
||||
pub(super) async fn start(&self) -> Result<()> {
|
||||
pub async fn start(self: &Arc<Self>) -> Result<Arc<Self>> {
|
||||
debug_info!("Starting services...");
|
||||
|
||||
self.admin.set_services(Some(Arc::clone(self)));
|
||||
globals::migrations::migrations(self).await?;
|
||||
self.manager
|
||||
.lock()
|
||||
|
@ -118,10 +120,10 @@ impl Services {
|
|||
.await?;
|
||||
|
||||
debug_info!("Services startup complete.");
|
||||
Ok(())
|
||||
Ok(Arc::clone(self))
|
||||
}
|
||||
|
||||
pub(super) async fn stop(&self) {
|
||||
pub async fn stop(&self) {
|
||||
info!("Shutting down services...");
|
||||
|
||||
self.interrupt();
|
||||
|
@ -129,6 +131,8 @@ impl Services {
|
|||
manager.stop().await;
|
||||
}
|
||||
|
||||
self.admin.set_services(None);
|
||||
|
||||
debug_info!("Services shutdown complete.");
|
||||
}
|
||||
|
||||
|
|
Loading…
Add table
Reference in a new issue