mirror of
https://git.asonix.dog/asonix/background-jobs.git
synced 2024-11-25 05:21:00 +00:00
Introduce a manager to keep workers alive
This commit is contained in:
parent
7002366d67
commit
cff996d1f1
10 changed files with 388 additions and 68 deletions
|
@ -15,6 +15,7 @@ members = [
|
||||||
"jobs-core",
|
"jobs-core",
|
||||||
"jobs-sled",
|
"jobs-sled",
|
||||||
"examples/actix-example",
|
"examples/actix-example",
|
||||||
|
"examples/managed-example",
|
||||||
]
|
]
|
||||||
|
|
||||||
[features]
|
[features]
|
||||||
|
|
|
@ -38,7 +38,7 @@ async fn main() -> Result<(), Error> {
|
||||||
let arbiter = Arbiter::new();
|
let arbiter = Arbiter::new();
|
||||||
|
|
||||||
// Start the application server. This guards access to to the jobs store
|
// Start the application server. This guards access to to the jobs store
|
||||||
let queue_handle = create_server_in_arbiter(storage, arbiter.handle());
|
let queue_handle = create_server_in_arbiter(arbiter.handle(), storage);
|
||||||
|
|
||||||
// Configure and start our workers
|
// Configure and start our workers
|
||||||
WorkerConfig::new(move || MyState::new("My App"))
|
WorkerConfig::new(move || MyState::new("My App"))
|
||||||
|
|
1
examples/managed-example/.gitignore
vendored
Normal file
1
examples/managed-example/.gitignore
vendored
Normal file
|
@ -0,0 +1 @@
|
||||||
|
/my-sled-db
|
19
examples/managed-example/Cargo.toml
Normal file
19
examples/managed-example/Cargo.toml
Normal file
|
@ -0,0 +1,19 @@
|
||||||
|
[package]
|
||||||
|
name = "managed-example"
|
||||||
|
version = "0.1.0"
|
||||||
|
authors = ["asonix <asonix@asonix.dog>"]
|
||||||
|
edition = "2021"
|
||||||
|
|
||||||
|
# See more keys and their definitions at https://doc.rust-lang.org/cargo/reference/manifest.html
|
||||||
|
|
||||||
|
[dependencies]
|
||||||
|
actix-rt = "2.0.0"
|
||||||
|
anyhow = "1.0"
|
||||||
|
async-trait = "0.1.24"
|
||||||
|
background-jobs = { version = "0.11.0", path = "../..", features = ["error-logging"] }
|
||||||
|
background-jobs-sled-storage = { version = "0.10.0", path = "../../jobs-sled" }
|
||||||
|
chrono = "0.4"
|
||||||
|
tracing = "0.1"
|
||||||
|
tracing-subscriber = { version = "0.2", features = ["fmt"] }
|
||||||
|
serde = { version = "1.0", features = ["derive"] }
|
||||||
|
sled = "0.34"
|
143
examples/managed-example/src/main.rs
Normal file
143
examples/managed-example/src/main.rs
Normal file
|
@ -0,0 +1,143 @@
|
||||||
|
use actix_rt::Arbiter;
|
||||||
|
use anyhow::Error;
|
||||||
|
use background_jobs::{ActixJob as Job, Manager, MaxRetries, WorkerConfig};
|
||||||
|
use background_jobs_sled_storage::Storage;
|
||||||
|
use chrono::{Duration, Utc};
|
||||||
|
use std::future::{ready, Ready};
|
||||||
|
use tracing::info;
|
||||||
|
use tracing_subscriber::EnvFilter;
|
||||||
|
|
||||||
|
const DEFAULT_QUEUE: &str = "default";
|
||||||
|
|
||||||
|
#[derive(Clone, Debug)]
|
||||||
|
pub struct MyState {
|
||||||
|
pub app_name: String,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Clone, Debug, serde::Deserialize, serde::Serialize)]
|
||||||
|
pub struct MyJob {
|
||||||
|
some_usize: usize,
|
||||||
|
other_usize: usize,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Clone, Debug, serde::Deserialize, serde::Serialize)]
|
||||||
|
pub struct StopJob;
|
||||||
|
|
||||||
|
#[actix_rt::main]
|
||||||
|
async fn main() -> Result<(), Error> {
|
||||||
|
let env_filter = EnvFilter::try_from_default_env().unwrap_or_else(|_| EnvFilter::new("info"));
|
||||||
|
|
||||||
|
tracing_subscriber::fmt::fmt()
|
||||||
|
.with_env_filter(env_filter)
|
||||||
|
.init();
|
||||||
|
|
||||||
|
// Set up our Storage
|
||||||
|
let db = sled::Config::new().temporary(true).open()?;
|
||||||
|
let storage = Storage::new(db)?;
|
||||||
|
|
||||||
|
// Configure and start our workers
|
||||||
|
let worker_config = WorkerConfig::new(move || MyState::new("My App"))
|
||||||
|
.register::<StopJob>()
|
||||||
|
.register::<MyJob>()
|
||||||
|
.set_worker_count(DEFAULT_QUEUE, 16);
|
||||||
|
|
||||||
|
// Start the application server. This guards access to to the jobs store
|
||||||
|
let manager = Manager::new(storage, worker_config);
|
||||||
|
|
||||||
|
// Queue our jobs
|
||||||
|
manager.queue_handle().queue(MyJob::new(1, 2)).await?;
|
||||||
|
manager.queue_handle().queue(MyJob::new(3, 4)).await?;
|
||||||
|
manager.queue_handle().queue(MyJob::new(5, 6)).await?;
|
||||||
|
manager
|
||||||
|
.queue_handle()
|
||||||
|
.schedule(MyJob::new(7, 8), Utc::now() + Duration::seconds(2))
|
||||||
|
.await?;
|
||||||
|
|
||||||
|
// Block on Actix
|
||||||
|
actix_rt::signal::ctrl_c().await?;
|
||||||
|
|
||||||
|
// kill the current arbiter
|
||||||
|
manager.queue_handle().queue(StopJob).await?;
|
||||||
|
|
||||||
|
// Block on Actix
|
||||||
|
actix_rt::signal::ctrl_c().await?;
|
||||||
|
|
||||||
|
// See that the workers have respawned
|
||||||
|
manager.queue_handle().queue(MyJob::new(1, 2)).await?;
|
||||||
|
manager.queue_handle().queue(MyJob::new(3, 4)).await?;
|
||||||
|
manager.queue_handle().queue(MyJob::new(5, 6)).await?;
|
||||||
|
manager
|
||||||
|
.queue_handle()
|
||||||
|
.schedule(MyJob::new(7, 8), Utc::now() + Duration::seconds(2))
|
||||||
|
.await?;
|
||||||
|
|
||||||
|
// Block on Actix
|
||||||
|
actix_rt::signal::ctrl_c().await?;
|
||||||
|
|
||||||
|
drop(manager);
|
||||||
|
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
|
impl MyState {
|
||||||
|
pub fn new(app_name: &str) -> Self {
|
||||||
|
MyState {
|
||||||
|
app_name: app_name.to_owned(),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl MyJob {
|
||||||
|
pub fn new(some_usize: usize, other_usize: usize) -> Self {
|
||||||
|
MyJob {
|
||||||
|
some_usize,
|
||||||
|
other_usize,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[async_trait::async_trait]
|
||||||
|
impl Job for MyJob {
|
||||||
|
type State = MyState;
|
||||||
|
type Future = Ready<Result<(), Error>>;
|
||||||
|
|
||||||
|
// The name of the job. It is super important that each job has a unique name,
|
||||||
|
// because otherwise one job will overwrite another job when they're being
|
||||||
|
// registered.
|
||||||
|
const NAME: &'static str = "MyJob";
|
||||||
|
|
||||||
|
// The queue that this processor belongs to
|
||||||
|
//
|
||||||
|
// Workers have the option to subscribe to specific queues, so this is important to
|
||||||
|
// determine which worker will call the processor
|
||||||
|
//
|
||||||
|
// Jobs can optionally override the queue they're spawned on
|
||||||
|
const QUEUE: &'static str = DEFAULT_QUEUE;
|
||||||
|
|
||||||
|
// The number of times background-jobs should try to retry a job before giving up
|
||||||
|
//
|
||||||
|
// Jobs can optionally override this value
|
||||||
|
const MAX_RETRIES: MaxRetries = MaxRetries::Count(1);
|
||||||
|
|
||||||
|
fn run(self, state: MyState) -> Self::Future {
|
||||||
|
info!("{}: args, {:?}", state.app_name, self);
|
||||||
|
|
||||||
|
ready(Ok(()))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[async_trait::async_trait]
|
||||||
|
impl Job for StopJob {
|
||||||
|
type State = MyState;
|
||||||
|
type Future = Ready<Result<(), Error>>;
|
||||||
|
|
||||||
|
const NAME: &'static str = "StopJob";
|
||||||
|
const QUEUE: &'static str = DEFAULT_QUEUE;
|
||||||
|
const MAX_RETRIES: MaxRetries = MaxRetries::Count(1);
|
||||||
|
|
||||||
|
fn run(self, _: MyState) -> Self::Future {
|
||||||
|
Arbiter::current().stop();
|
||||||
|
|
||||||
|
ready(Ok(()))
|
||||||
|
}
|
||||||
|
}
|
|
@ -9,21 +9,18 @@ use tracing::error;
|
||||||
/// let server = create_server(storage);
|
/// let server = create_server(storage);
|
||||||
/// server.every(Duration::from_secs(60 * 30), MyJob::new());
|
/// server.every(Duration::from_secs(60 * 30), MyJob::new());
|
||||||
/// ```
|
/// ```
|
||||||
pub(crate) fn every<J>(spawner: &QueueHandle, duration: Duration, job: J)
|
pub(crate) async fn every<J>(spawner: QueueHandle, duration: Duration, job: J)
|
||||||
where
|
where
|
||||||
J: Job + Clone + Send,
|
J: Job + Clone + Send,
|
||||||
{
|
{
|
||||||
let spawner_clone = spawner.clone();
|
|
||||||
spawner.tokio_rt.spawn(async move {
|
|
||||||
let mut interval = interval_at(Instant::now(), duration);
|
let mut interval = interval_at(Instant::now(), duration);
|
||||||
|
|
||||||
loop {
|
loop {
|
||||||
interval.tick().await;
|
interval.tick().await;
|
||||||
|
|
||||||
let job = job.clone();
|
let job = job.clone();
|
||||||
if spawner_clone.queue::<J>(job).await.is_err() {
|
if spawner.queue::<J>(job).await.is_err() {
|
||||||
error!("Failed to queue job: {}", J::NAME);
|
error!("Failed to queue job: {}", J::NAME);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
});
|
|
||||||
}
|
}
|
||||||
|
|
|
@ -120,17 +120,136 @@ use actix_rt::{Arbiter, ArbiterHandle};
|
||||||
use anyhow::Error;
|
use anyhow::Error;
|
||||||
use background_jobs_core::{new_job, new_scheduled_job, Job, ProcessorMap, Stats, Storage};
|
use background_jobs_core::{new_job, new_scheduled_job, Job, ProcessorMap, Stats, Storage};
|
||||||
use chrono::{DateTime, Utc};
|
use chrono::{DateTime, Utc};
|
||||||
use std::{collections::BTreeMap, sync::Arc, time::Duration};
|
use std::{collections::BTreeMap, ops::Deref, sync::Arc, time::Duration};
|
||||||
|
|
||||||
mod every;
|
mod every;
|
||||||
mod server;
|
mod server;
|
||||||
mod storage;
|
mod storage;
|
||||||
mod worker;
|
mod worker;
|
||||||
|
|
||||||
use self::{every::every, server::Server, worker::local_worker};
|
use self::{every::every, server::Server, worker::LocalWorkerStarter};
|
||||||
|
|
||||||
pub use background_jobs_core::ActixJob;
|
pub use background_jobs_core::ActixJob;
|
||||||
|
|
||||||
|
/// Manager for worker threads
|
||||||
|
///
|
||||||
|
/// Manager attempts to restart workers as their arbiters die
|
||||||
|
pub struct Manager {
|
||||||
|
// the manager arbiter
|
||||||
|
_arbiter: ArbiterDropper,
|
||||||
|
|
||||||
|
// handle for queueing
|
||||||
|
queue_handle: QueueHandle,
|
||||||
|
}
|
||||||
|
|
||||||
|
struct ArbiterDropper {
|
||||||
|
arbiter: Option<Arbiter>,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl Manager {
|
||||||
|
/// Create a new manager to keep jobs alive
|
||||||
|
///
|
||||||
|
/// Manager works by startinng a new Arbiter to run jobs, and if that arbiter ever dies, it
|
||||||
|
/// spins up another one and spawns the workers again
|
||||||
|
pub fn new<S, State>(storage: S, worker_config: WorkerConfig<State>) -> Self
|
||||||
|
where
|
||||||
|
S: Storage + Sync + 'static,
|
||||||
|
State: Clone,
|
||||||
|
{
|
||||||
|
let arbiter = Arbiter::new();
|
||||||
|
let worker_arbiter = Arbiter::new();
|
||||||
|
let notifier = Arc::new(tokio::sync::Notify::new());
|
||||||
|
|
||||||
|
let queue_handle = create_server_managed(storage);
|
||||||
|
|
||||||
|
let drop_notifier = DropNotifier::new(Arc::clone(¬ifier));
|
||||||
|
let queue_handle_2 = queue_handle.clone();
|
||||||
|
arbiter.spawn(async move {
|
||||||
|
let queue_handle = queue_handle_2;
|
||||||
|
|
||||||
|
let mut drop_notifier = drop_notifier;
|
||||||
|
let mut arbiter = ArbiterDropper {
|
||||||
|
arbiter: Some(worker_arbiter),
|
||||||
|
};
|
||||||
|
|
||||||
|
loop {
|
||||||
|
queue_handle
|
||||||
|
.inner
|
||||||
|
.ticker(arbiter.handle(), drop_notifier.clone());
|
||||||
|
worker_config.start_managed(
|
||||||
|
&arbiter.handle(),
|
||||||
|
queue_handle.clone(),
|
||||||
|
&drop_notifier,
|
||||||
|
);
|
||||||
|
|
||||||
|
notifier.notified().await;
|
||||||
|
// drop_notifier needs to live at least as long as notifier.notified().await
|
||||||
|
// in order to ensure we get notified by ticker or a worker, and not ourselves
|
||||||
|
drop(drop_notifier);
|
||||||
|
|
||||||
|
// Assume arbiter is dead if we were notified
|
||||||
|
if arbiter.spawn(async {}) {
|
||||||
|
panic!("Arbiter should be dead by now");
|
||||||
|
}
|
||||||
|
|
||||||
|
arbiter = ArbiterDropper {
|
||||||
|
arbiter: Some(Arbiter::new()),
|
||||||
|
};
|
||||||
|
|
||||||
|
drop_notifier = DropNotifier::new(Arc::clone(¬ifier));
|
||||||
|
}
|
||||||
|
});
|
||||||
|
|
||||||
|
Manager {
|
||||||
|
_arbiter: ArbiterDropper {
|
||||||
|
arbiter: Some(arbiter),
|
||||||
|
},
|
||||||
|
queue_handle,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Retrieve the QueueHandle for the managed workers
|
||||||
|
pub fn queue_handle(&self) -> &QueueHandle {
|
||||||
|
&self.queue_handle
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl Deref for ArbiterDropper {
|
||||||
|
type Target = Arbiter;
|
||||||
|
|
||||||
|
fn deref(&self) -> &Self::Target {
|
||||||
|
self.arbiter.as_ref().unwrap()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl Drop for ArbiterDropper {
|
||||||
|
fn drop(&mut self) {
|
||||||
|
self.stop();
|
||||||
|
let _ = self.arbiter.take().unwrap().join();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Clone)]
|
||||||
|
struct DropNotifier {
|
||||||
|
inner: Arc<std::sync::Mutex<Option<Arc<tokio::sync::Notify>>>>,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl DropNotifier {
|
||||||
|
fn new(notify: Arc<tokio::sync::Notify>) -> Self {
|
||||||
|
DropNotifier {
|
||||||
|
inner: Arc::new(std::sync::Mutex::new(Some(notify))),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl Drop for DropNotifier {
|
||||||
|
fn drop(&mut self) {
|
||||||
|
if let Some(notifier) = self.inner.lock().unwrap().take() {
|
||||||
|
notifier.notify_one();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
/// Create a new Server
|
/// Create a new Server
|
||||||
///
|
///
|
||||||
/// In previous versions of this library, the server itself was run on it's own dedicated threads
|
/// In previous versions of this library, the server itself was run on it's own dedicated threads
|
||||||
|
@ -142,7 +261,7 @@ pub fn create_server<S>(storage: S) -> QueueHandle
|
||||||
where
|
where
|
||||||
S: Storage + Sync + 'static,
|
S: Storage + Sync + 'static,
|
||||||
{
|
{
|
||||||
create_server_in_arbiter(storage, Arbiter::current())
|
create_server_in_arbiter(Arbiter::current(), storage)
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Create a new Server
|
/// Create a new Server
|
||||||
|
@ -152,15 +271,26 @@ where
|
||||||
/// primitives, the Server has become an object that gets shared between client threads.
|
/// primitives, the Server has become an object that gets shared between client threads.
|
||||||
///
|
///
|
||||||
/// This method will panic if not called from an actix runtime
|
/// This method will panic if not called from an actix runtime
|
||||||
pub fn create_server_in_arbiter<S>(storage: S, arbiter: ArbiterHandle) -> QueueHandle
|
pub fn create_server_in_arbiter<S>(arbiter: ArbiterHandle, storage: S) -> QueueHandle
|
||||||
where
|
where
|
||||||
S: Storage + Sync + 'static,
|
S: Storage + Sync + 'static,
|
||||||
{
|
{
|
||||||
let tokio_rt = tokio::runtime::Handle::current();
|
let handle = create_server_managed(storage);
|
||||||
|
handle.inner.ticker(arbiter, ());
|
||||||
|
handle
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Create a new managed Server
|
||||||
|
///
|
||||||
|
/// In previous versions of this library, the server itself was run on it's own dedicated threads
|
||||||
|
/// and guarded access to jobs via messages. Since we now have futures-aware synchronization
|
||||||
|
/// primitives, the Server has become an object that gets shared between client threads.
|
||||||
|
pub fn create_server_managed<S>(storage: S) -> QueueHandle
|
||||||
|
where
|
||||||
|
S: Storage + Sync + 'static,
|
||||||
|
{
|
||||||
QueueHandle {
|
QueueHandle {
|
||||||
inner: Server::new(arbiter, storage),
|
inner: Server::new(storage),
|
||||||
tokio_rt,
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -228,14 +358,31 @@ where
|
||||||
|
|
||||||
/// Start the workers in the provided arbiter
|
/// Start the workers in the provided arbiter
|
||||||
pub fn start_in_arbiter(self, arbiter: &ArbiterHandle, queue_handle: QueueHandle) {
|
pub fn start_in_arbiter(self, arbiter: &ArbiterHandle, queue_handle: QueueHandle) {
|
||||||
for (key, count) in self.queues.into_iter() {
|
self.start_managed(arbiter, queue_handle, &())
|
||||||
for _ in 0..count {
|
}
|
||||||
let key = key.clone();
|
|
||||||
|
/// Start a workers in a managed way
|
||||||
|
pub fn start_managed<Extras: Clone + Send + 'static>(
|
||||||
|
&self,
|
||||||
|
arbiter: &ArbiterHandle,
|
||||||
|
queue_handle: QueueHandle,
|
||||||
|
extras: &Extras,
|
||||||
|
) {
|
||||||
|
for (key, count) in self.queues.iter() {
|
||||||
|
for _ in 0..*count {
|
||||||
|
let queue = key.clone();
|
||||||
let processors = self.processors.clone();
|
let processors = self.processors.clone();
|
||||||
let server = queue_handle.inner.clone();
|
let server = queue_handle.inner.clone();
|
||||||
|
|
||||||
|
let extras_2 = extras.clone();
|
||||||
|
|
||||||
arbiter.spawn_fn(move || {
|
arbiter.spawn_fn(move || {
|
||||||
actix_rt::spawn(local_worker(key, processors.cached(), server));
|
drop(LocalWorkerStarter::new(
|
||||||
|
queue,
|
||||||
|
processors.cached(),
|
||||||
|
server,
|
||||||
|
extras_2,
|
||||||
|
));
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -249,7 +396,6 @@ where
|
||||||
#[derive(Clone)]
|
#[derive(Clone)]
|
||||||
pub struct QueueHandle {
|
pub struct QueueHandle {
|
||||||
inner: Server,
|
inner: Server,
|
||||||
tokio_rt: tokio::runtime::Handle,
|
|
||||||
}
|
}
|
||||||
|
|
||||||
impl QueueHandle {
|
impl QueueHandle {
|
||||||
|
@ -266,17 +412,6 @@ impl QueueHandle {
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Queues a job for execution
|
|
||||||
///
|
|
||||||
/// This job will be sent to the server for storage, and will execute whenever a worker for the
|
|
||||||
/// job's queue is free to do so.
|
|
||||||
pub async fn blocking_queue<J>(&self, job: J) -> Result<(), Error>
|
|
||||||
where
|
|
||||||
J: Job,
|
|
||||||
{
|
|
||||||
self.tokio_rt.block_on(self.queue(job))
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Schedule a job for execution later
|
/// Schedule a job for execution later
|
||||||
///
|
///
|
||||||
/// This job will be sent to the server for storage, and will execute after the specified time
|
/// This job will be sent to the server for storage, and will execute after the specified time
|
||||||
|
@ -291,17 +426,6 @@ impl QueueHandle {
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Schedule a job for execution later
|
|
||||||
///
|
|
||||||
/// This job will be sent to the server for storage, and will execute after the specified time
|
|
||||||
/// and when a worker for the job's queue is free to do so.
|
|
||||||
pub fn blocking_schedule<J>(&self, job: J, after: DateTime<Utc>) -> Result<(), Error>
|
|
||||||
where
|
|
||||||
J: Job,
|
|
||||||
{
|
|
||||||
self.tokio_rt.block_on(self.schedule(job, after))
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Queues a job for recurring execution
|
/// Queues a job for recurring execution
|
||||||
///
|
///
|
||||||
/// This job will be added to it's queue on the server once every `Duration`. It will be
|
/// This job will be added to it's queue on the server once every `Duration`. It will be
|
||||||
|
@ -310,7 +434,7 @@ impl QueueHandle {
|
||||||
where
|
where
|
||||||
J: Job + Clone + Send + 'static,
|
J: Job + Clone + Send + 'static,
|
||||||
{
|
{
|
||||||
every(self, duration, job);
|
actix_rt::spawn(every(self.clone(), duration, job));
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Return an overview of the processor's statistics
|
/// Return an overview of the processor's statistics
|
||||||
|
|
|
@ -23,19 +23,24 @@ pub(crate) struct ServerCache {
|
||||||
cache: Arc<Mutex<HashMap<String, WorkerQueue>>>,
|
cache: Arc<Mutex<HashMap<String, WorkerQueue>>>,
|
||||||
}
|
}
|
||||||
|
|
||||||
struct Ticker {
|
pub(super) struct Ticker<Extras: Send + 'static> {
|
||||||
server: Server,
|
server: Server,
|
||||||
|
extras: Option<Extras>,
|
||||||
|
arbiter: ArbiterHandle,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl Drop for Ticker {
|
impl<Extras: Send + 'static> Drop for Ticker<Extras> {
|
||||||
fn drop(&mut self) {
|
fn drop(&mut self) {
|
||||||
let online = self.server.arbiter.spawn(async move {});
|
let online = self.arbiter.spawn(async move {});
|
||||||
|
|
||||||
|
let extras = self.extras.take().unwrap();
|
||||||
|
|
||||||
if online {
|
if online {
|
||||||
let server = self.server.clone();
|
let server = self.server.clone();
|
||||||
|
|
||||||
self.server.arbiter.spawn(async move {
|
let arbiter = self.arbiter.clone();
|
||||||
let _ticker = server.ticker();
|
let spawned = self.arbiter.spawn(async move {
|
||||||
|
let _ticker = server.ticker(arbiter, extras);
|
||||||
let mut interval = interval_at(Instant::now(), Duration::from_secs(1));
|
let mut interval = interval_at(Instant::now(), Duration::from_secs(1));
|
||||||
|
|
||||||
loop {
|
loop {
|
||||||
|
@ -45,10 +50,14 @@ impl Drop for Ticker {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
});
|
});
|
||||||
} else {
|
|
||||||
warn!("Not restarting ticker, arbiter is dead");
|
if spawned {
|
||||||
|
return;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
warn!("Not restarting ticker, arbiter is dead");
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/// The server Actor
|
/// The server Actor
|
||||||
|
@ -59,29 +68,30 @@ impl Drop for Ticker {
|
||||||
pub(crate) struct Server {
|
pub(crate) struct Server {
|
||||||
storage: Arc<dyn ActixStorage + Send + Sync>,
|
storage: Arc<dyn ActixStorage + Send + Sync>,
|
||||||
cache: ServerCache,
|
cache: ServerCache,
|
||||||
arbiter: ArbiterHandle,
|
|
||||||
}
|
}
|
||||||
|
|
||||||
impl Server {
|
impl Server {
|
||||||
fn ticker(&self) -> Ticker {
|
pub(super) fn ticker<Extras: Send + 'static>(
|
||||||
|
&self,
|
||||||
|
arbiter: ArbiterHandle,
|
||||||
|
extras: Extras,
|
||||||
|
) -> Ticker<Extras> {
|
||||||
Ticker {
|
Ticker {
|
||||||
server: self.clone(),
|
server: self.clone(),
|
||||||
|
extras: Some(extras),
|
||||||
|
arbiter,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Create a new Server from a compatible storage implementation
|
/// Create a new Server from a compatible storage implementation
|
||||||
pub(crate) fn new<S>(arbiter: ArbiterHandle, storage: S) -> Self
|
pub(crate) fn new<S>(storage: S) -> Self
|
||||||
where
|
where
|
||||||
S: Storage + Sync + 'static,
|
S: Storage + Sync + 'static,
|
||||||
{
|
{
|
||||||
let server = Server {
|
Server {
|
||||||
storage: Arc::new(StorageWrapper(storage)),
|
storage: Arc::new(StorageWrapper(storage)),
|
||||||
cache: ServerCache::new(),
|
cache: ServerCache::new(),
|
||||||
arbiter,
|
}
|
||||||
};
|
|
||||||
|
|
||||||
drop(server.ticker());
|
|
||||||
|
|
||||||
server
|
|
||||||
}
|
}
|
||||||
|
|
||||||
async fn check_db(&self) -> Result<(), Error> {
|
async fn check_db(&self) -> Result<(), Error> {
|
||||||
|
|
|
@ -56,24 +56,45 @@ impl Worker for LocalWorkerHandle {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
struct LocalWorkerStarter<State: Clone + 'static> {
|
pub(crate) struct LocalWorkerStarter<State: Clone + 'static, Extras: 'static> {
|
||||||
queue: String,
|
queue: String,
|
||||||
processors: CachedProcessorMap<State>,
|
processors: CachedProcessorMap<State>,
|
||||||
server: Server,
|
server: Server,
|
||||||
|
extras: Option<Extras>,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl<State: Clone + 'static> Drop for LocalWorkerStarter<State> {
|
impl<State: Clone + 'static, Extras: 'static> LocalWorkerStarter<State, Extras> {
|
||||||
|
pub(super) fn new(
|
||||||
|
queue: String,
|
||||||
|
processors: CachedProcessorMap<State>,
|
||||||
|
server: Server,
|
||||||
|
extras: Extras,
|
||||||
|
) -> Self {
|
||||||
|
LocalWorkerStarter {
|
||||||
|
queue,
|
||||||
|
processors,
|
||||||
|
server,
|
||||||
|
extras: Some(extras),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl<State: Clone + 'static, Extras: 'static> Drop for LocalWorkerStarter<State, Extras> {
|
||||||
fn drop(&mut self) {
|
fn drop(&mut self) {
|
||||||
let res = std::panic::catch_unwind(|| actix_rt::Arbiter::current().spawn(async move {}));
|
let res = std::panic::catch_unwind(|| actix_rt::Arbiter::current().spawn(async move {}));
|
||||||
|
|
||||||
|
let extras = self.extras.take().unwrap();
|
||||||
|
|
||||||
if let Ok(true) = res {
|
if let Ok(true) = res {
|
||||||
actix_rt::spawn(local_worker(
|
actix_rt::spawn(local_worker(
|
||||||
self.queue.clone(),
|
self.queue.clone(),
|
||||||
self.processors.clone(),
|
self.processors.clone(),
|
||||||
self.server.clone(),
|
self.server.clone(),
|
||||||
|
extras,
|
||||||
));
|
));
|
||||||
} else {
|
} else {
|
||||||
warn!("Not restarting worker, Arbiter is dead");
|
warn!("Not restarting worker, Arbiter is dead");
|
||||||
|
drop(extras);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -91,18 +112,22 @@ where
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
pub(crate) async fn local_worker<State>(
|
async fn local_worker<State, Extras>(
|
||||||
queue: String,
|
queue: String,
|
||||||
processors: CachedProcessorMap<State>,
|
processors: CachedProcessorMap<State>,
|
||||||
server: Server,
|
server: Server,
|
||||||
|
extras: Extras,
|
||||||
) where
|
) where
|
||||||
State: Clone + 'static,
|
State: Clone + 'static,
|
||||||
|
Extras: 'static,
|
||||||
{
|
{
|
||||||
let starter = LocalWorkerStarter {
|
let starter = LocalWorkerStarter {
|
||||||
queue: queue.clone(),
|
queue: queue.clone(),
|
||||||
processors: processors.clone(),
|
processors: processors.clone(),
|
||||||
server: server.clone(),
|
server: server.clone(),
|
||||||
|
extras: Some(extras),
|
||||||
};
|
};
|
||||||
|
|
||||||
let id = Uuid::new_v4();
|
let id = Uuid::new_v4();
|
||||||
let (tx, mut rx) = channel(16);
|
let (tx, mut rx) = channel(16);
|
||||||
|
|
||||||
|
|
|
@ -173,5 +173,5 @@ pub mod dev {
|
||||||
|
|
||||||
#[cfg(feature = "background-jobs-actix")]
|
#[cfg(feature = "background-jobs-actix")]
|
||||||
pub use background_jobs_actix::{
|
pub use background_jobs_actix::{
|
||||||
create_server, create_server_in_arbiter, ActixJob, QueueHandle, WorkerConfig,
|
create_server, create_server_in_arbiter, ActixJob, Manager, QueueHandle, WorkerConfig,
|
||||||
};
|
};
|
||||||
|
|
Loading…
Reference in a new issue