use std::{sync::Arc, time::Duration}; use axum_server::Handle as ServerHandle; use tokio::sync::broadcast::{self, Sender}; use tracing::{debug, error, info}; extern crate conduit_admin as admin; extern crate conduit_core as conduit; extern crate conduit_database as database; extern crate conduit_service as service; use std::sync::atomic::Ordering; use conduit::{debug_info, trace, Error, Result, Server}; use database::KeyValueDatabase; use service::{services, Services}; use crate::{layers, serve}; /// Main loop base #[tracing::instrument(skip_all)] #[allow(clippy::let_underscore_must_use)] // various of these are intended pub(crate) async fn run(server: Arc) -> Result<(), Error> { let app = layers::build(&server)?; // Install the admin room callback here for now _ = services().admin.handle.lock().await.insert(admin::handle); // Setup shutdown/signal handling let handle = ServerHandle::new(); _ = server .shutdown .lock() .expect("locked") .insert(handle.clone()); server.interrupt.store(false, Ordering::Release); let (tx, _) = broadcast::channel::<()>(1); let sigs = server.runtime().spawn(signal(server.clone(), tx.clone())); // Serve clients let res = serve::serve(&server, app, handle, tx.subscribe()).await; // Join the signal handler before we leave. sigs.abort(); _ = sigs.await; // Reset the axum handle instance; this should be reusable and might be // reload-survivable but better to be safe than sorry. _ = server.shutdown.lock().expect("locked").take(); // Remove the admin room callback _ = services().admin.handle.lock().await.take(); debug_info!("Finished"); res } /// Async initializations #[tracing::instrument(skip_all)] #[allow(clippy::let_underscore_must_use)] pub(crate) async fn start(server: Arc) -> Result<(), Error> { debug!("Starting..."); let d = Arc::new(KeyValueDatabase::load_or_create(&server).await?); let s = Box::new(Services::build(server, d.clone()).await?); _ = service::SERVICES .write() .expect("write locked") .insert(Box::leak(s)); services().start().await?; #[cfg(feature = "systemd")] sd_notify::notify(true, &[sd_notify::NotifyState::Ready]).expect("failed to notify systemd of ready state"); debug!("Started"); Ok(()) } /// Async destructions #[tracing::instrument(skip_all)] pub(crate) async fn stop(_server: Arc) -> Result<(), Error> { debug!("Shutting down..."); // Wait for all completions before dropping or we'll lose them to the module // unload and explode. services().shutdown().await; // Deactivate services(). Any further use will panic the caller. let s = service::SERVICES .write() .expect("write locked") .take() .unwrap(); let s: *mut Services = std::ptr::from_ref(s).cast_mut(); //SAFETY: Services was instantiated in start() and leaked into the SERVICES // global perusing as 'static for the duration of run_server(). Now we reclaim // it to drop it before unloading the module. If this is not done there will be // multiple instances after module reload. let s = unsafe { Box::from_raw(s) }; debug!("Cleaning up..."); // Drop it so we encounter any trouble before the infolog message drop(s); #[cfg(feature = "systemd")] sd_notify::notify(true, &[sd_notify::NotifyState::Stopping]).expect("failed to notify systemd of stopping state"); info!("Shutdown complete."); Ok(()) } #[tracing::instrument(skip_all)] async fn signal(server: Arc, tx: Sender<()>) { let sig: &'static str = server .signal .subscribe() .recv() .await .expect("channel error"); debug!("Received signal {}", sig); if sig == "Ctrl+C" { let reload = cfg!(unix) && cfg!(debug_assertions); server.reload.store(reload, Ordering::Release); } server.interrupt.store(true, Ordering::Release); services().globals.rotate.fire(); if let Err(e) = tx.send(()) { error!("failed sending shutdown transaction to channel: {e}"); } if let Some(handle) = server.shutdown.lock().expect("locked").as_ref() { let pending = server.requests_spawn_active.load(Ordering::Relaxed); if pending > 0 { let timeout = Duration::from_secs(36); trace!(pending, ?timeout, "Notifying for graceful shutdown"); handle.graceful_shutdown(Some(timeout)); } else { debug!(pending, "Notifying for immediate shutdown"); handle.shutdown(); } } }