mirror of
https://git.asonix.dog/asonix/background-jobs.git
synced 2024-11-21 19:40:59 +00:00
Mark jobs staged, not running
Clear staged jobs on startup
This commit is contained in:
parent
9cff817ed6
commit
c8f1f6cd34
28 changed files with 186 additions and 65 deletions
10
Cargo.toml
10
Cargo.toml
|
@ -16,23 +16,23 @@ members = [
|
|||
]
|
||||
|
||||
[features]
|
||||
default = ["jobs-actix", "jobs-server-tokio", "jobs-server-tokio/tokio-zmq", "jobs-tokio"]
|
||||
default = ["background-jobs-actix", "background-jobs-server-tokio", "background-jobs-server-tokio/tokio-zmq", "background-jobs-tokio"]
|
||||
|
||||
[dependencies.jobs-actix]
|
||||
[dependencies.background-jobs-actix]
|
||||
version = "0.1"
|
||||
path = "jobs-actix"
|
||||
optional = true
|
||||
|
||||
[dependencies.jobs-core]
|
||||
[dependencies.background-jobs-core]
|
||||
version = "0.1"
|
||||
path = "jobs-core"
|
||||
|
||||
[dependencies.jobs-server-tokio]
|
||||
[dependencies.background-jobs-server-tokio]
|
||||
version = "0.1"
|
||||
path = "jobs-server-tokio"
|
||||
optional = true
|
||||
|
||||
[dependencies.jobs-tokio]
|
||||
[dependencies.background-jobs-tokio]
|
||||
version = "0.1"
|
||||
path = "jobs-tokio"
|
||||
optional = true
|
||||
|
|
5
TODO
5
TODO
|
@ -1,7 +1,2 @@
|
|||
1.
|
||||
Gracefull Shutdown
|
||||
|
||||
2.
|
||||
Don't mark pushed jobs as running, mark them as 'staged'
|
||||
Clear staged jobs that are 10 minutes old
|
||||
Send a Running notification from the worker to move a job from 'staged' to 'running'
|
||||
|
|
|
@ -14,8 +14,8 @@ log = "0.4"
|
|||
serde = "1.0"
|
||||
serde_derive = "1.0"
|
||||
|
||||
[dependencies.jobs]
|
||||
[dependencies.background-jobs]
|
||||
version = "0.1"
|
||||
path = "../.."
|
||||
default-features = false
|
||||
features = ["jobs-actix"]
|
||||
features = ["background-jobs-actix"]
|
||||
|
|
|
@ -3,9 +3,9 @@ extern crate log;
|
|||
#[macro_use]
|
||||
extern crate serde_derive;
|
||||
|
||||
use background_jobs::{Backoff, JobsBuilder, MaxRetries, Processor, QueueJob};
|
||||
use failure::Error;
|
||||
use futures::{future::IntoFuture, Future};
|
||||
use jobs::{Backoff, JobsBuilder, MaxRetries, Processor, QueueJob};
|
||||
|
||||
#[derive(Clone, Debug, Deserialize, Serialize)]
|
||||
struct MyJobArguments {
|
||||
|
@ -23,6 +23,10 @@ impl Processor for MyProcessor {
|
|||
"MyProcessor"
|
||||
}
|
||||
|
||||
fn queue() -> &'static str {
|
||||
"default"
|
||||
}
|
||||
|
||||
fn max_retries() -> MaxRetries {
|
||||
MaxRetries::Count(1)
|
||||
}
|
||||
|
|
|
@ -14,8 +14,8 @@ serde = "1.0"
|
|||
serde_derive = "1.0"
|
||||
tokio = "0.1"
|
||||
|
||||
[dependencies.jobs]
|
||||
[dependencies.background-jobs]
|
||||
version = "0.1"
|
||||
path = "../.."
|
||||
default-features = false
|
||||
features = ["jobs-server-tokio"]
|
||||
features = ["background-jobs-server-tokio"]
|
||||
|
|
|
@ -1,5 +1,5 @@
|
|||
use background_jobs::ServerConfig;
|
||||
use failure::Error;
|
||||
use jobs::ServerConfig;
|
||||
use server_jobs_example::queue_set;
|
||||
|
||||
fn main() -> Result<(), Error> {
|
||||
|
|
|
@ -1,5 +1,5 @@
|
|||
use background_jobs::{Processor, SpawnerConfig};
|
||||
use futures::{future::lazy, Future};
|
||||
use jobs::{Processor, SpawnerConfig};
|
||||
use server_jobs_example::{MyJobArguments, MyProcessor};
|
||||
|
||||
fn main() {
|
||||
|
|
|
@ -1,5 +1,5 @@
|
|||
use background_jobs::WorkerConfig;
|
||||
use failure::Error;
|
||||
use jobs::WorkerConfig;
|
||||
use server_jobs_example::{queue_map, MyProcessor};
|
||||
|
||||
fn main() -> Result<(), Error> {
|
||||
|
|
|
@ -5,9 +5,9 @@ extern crate serde_derive;
|
|||
|
||||
use std::collections::{BTreeMap, BTreeSet};
|
||||
|
||||
use background_jobs::{Backoff, MaxRetries, Processor};
|
||||
use failure::Error;
|
||||
use futures::{future::IntoFuture, Future};
|
||||
use jobs::{Backoff, MaxRetries, Processor};
|
||||
|
||||
pub fn queue_map() -> BTreeMap<String, usize> {
|
||||
let mut map = BTreeMap::new();
|
||||
|
|
|
@ -14,8 +14,8 @@ serde = "1.0"
|
|||
serde_derive = "1.0"
|
||||
tokio = "0.1"
|
||||
|
||||
[dependencies.jobs]
|
||||
[dependencies.background-jobs]
|
||||
version = "0.1"
|
||||
path = "../.."
|
||||
default-features = false
|
||||
features = ["jobs-tokio"]
|
||||
features = ["background-jobs-tokio"]
|
||||
|
|
|
@ -5,12 +5,12 @@ extern crate serde_derive;
|
|||
|
||||
use std::time::Duration;
|
||||
|
||||
use background_jobs::{Backoff, JobRunner, MaxRetries, Processor};
|
||||
use failure::Error;
|
||||
use futures::{
|
||||
future::{lazy, IntoFuture},
|
||||
Future,
|
||||
};
|
||||
use jobs::{Backoff, JobRunner, MaxRetries, Processor};
|
||||
|
||||
#[derive(Clone, Debug, Deserialize, Serialize)]
|
||||
struct MyJobArguments {
|
||||
|
@ -28,6 +28,10 @@ impl Processor for MyProcessor {
|
|||
"MyProcessor"
|
||||
}
|
||||
|
||||
fn queue() -> &'static str {
|
||||
"default"
|
||||
}
|
||||
|
||||
fn max_retries() -> MaxRetries {
|
||||
MaxRetries::Count(1)
|
||||
}
|
||||
|
|
|
@ -10,6 +10,6 @@ failure = "0.1"
|
|||
futures = "0.1"
|
||||
log = "0.4"
|
||||
|
||||
[dependencies.jobs-core]
|
||||
[dependencies.background-jobs-core]
|
||||
version = "0.1"
|
||||
path = "../jobs-core"
|
||||
|
|
|
@ -13,9 +13,9 @@ use actix::{
|
|||
fut::wrap_future, utils::IntervalFunc, Actor, ActorFuture, ActorStream, Addr, AsyncContext,
|
||||
Context, ContextFutureSpawner, Handler, Message, ResponseFuture, SyncArbiter, SyncContext,
|
||||
};
|
||||
use background_jobs_core::{JobInfo, Processor, Processors, Storage};
|
||||
use failure::Error;
|
||||
use futures::Future;
|
||||
use jobs_core::{JobInfo, Processor, Processors, Storage};
|
||||
|
||||
fn coerce<I, E, F>(res: Result<Result<I, E>, F>) -> Result<I, E>
|
||||
where
|
||||
|
@ -50,7 +50,7 @@ impl KvActor {
|
|||
}
|
||||
|
||||
pub fn dequeue_jobs(&self, limit: usize, queue: &str) -> Result<Vec<JobInfo>, Error> {
|
||||
let jobs = self.storage.dequeue_job(limit, queue)?;
|
||||
let jobs = self.storage.stage_jobs(limit, queue)?;
|
||||
|
||||
Ok(jobs)
|
||||
}
|
||||
|
|
|
@ -1,5 +1,5 @@
|
|||
[package]
|
||||
name = "jobs-core"
|
||||
name = "background-jobs-core"
|
||||
version = "0.1.0"
|
||||
authors = ["asonix <asonix@asonix.dog>"]
|
||||
edition = "2018"
|
||||
|
|
|
@ -122,6 +122,14 @@ impl JobInfo {
|
|||
self.queue == queue
|
||||
}
|
||||
|
||||
pub(crate) fn stage(&mut self) {
|
||||
self.status = JobStatus::Staged;
|
||||
}
|
||||
|
||||
pub fn run(&mut self) {
|
||||
self.status = JobStatus::Running;
|
||||
}
|
||||
|
||||
pub(crate) fn pending(&mut self) {
|
||||
self.status = JobStatus::Pending;
|
||||
}
|
||||
|
|
|
@ -32,6 +32,9 @@ pub enum JobStatus {
|
|||
/// Job should be queued
|
||||
Pending,
|
||||
|
||||
/// Job has been dequeued, but is not yet running
|
||||
Staged,
|
||||
|
||||
/// Job is running
|
||||
Running,
|
||||
|
||||
|
|
|
@ -14,6 +14,7 @@ use crate::{JobInfo, JobStatus};
|
|||
struct Buckets<'a> {
|
||||
queued: Bucket<'a, &'a [u8], ValueBuf<Json<usize>>>,
|
||||
running: Bucket<'a, &'a [u8], ValueBuf<Json<usize>>>,
|
||||
staged: Bucket<'a, &'a [u8], ValueBuf<Json<usize>>>,
|
||||
failed: Bucket<'a, &'a [u8], ValueBuf<Json<usize>>>,
|
||||
finished: Bucket<'a, &'a [u8], ValueBuf<Json<usize>>>,
|
||||
}
|
||||
|
@ -23,6 +24,7 @@ impl<'a> Buckets<'a> {
|
|||
let b = Buckets {
|
||||
queued: store.bucket(Some(Storage::job_queue()))?,
|
||||
running: store.bucket(Some(Storage::job_running()))?,
|
||||
staged: store.bucket(Some(Storage::job_staged()))?,
|
||||
failed: store.bucket(Some(Storage::job_failed()))?,
|
||||
finished: store.bucket(Some(Storage::job_finished()))?,
|
||||
};
|
||||
|
@ -87,6 +89,56 @@ impl Storage {
|
|||
Ok(new_id)
|
||||
}
|
||||
|
||||
pub fn requeue_staged_jobs(&self) -> Result<(), Error> {
|
||||
let store = self.store.write()?;
|
||||
let job_bucket =
|
||||
store.bucket::<&[u8], ValueBuf<Json<JobInfo>>>(Some(Storage::job_store()))?;
|
||||
|
||||
let lock_bucket =
|
||||
store.bucket::<&[u8], ValueBuf<Json<usize>>>(Some(Storage::job_lock()))?;
|
||||
|
||||
let buckets = Buckets::new(&store)?;
|
||||
|
||||
let mut write_txn = store.write_txn()?;
|
||||
let read_txn = store.read_txn()?;
|
||||
|
||||
self.with_lock::<_, (), _>(&lock_bucket, &mut write_txn, b"job-queue", |inner_txn| {
|
||||
let mut cursor = read_txn.read_cursor(&buckets.staged)?;
|
||||
match cursor.get(None, CursorOp::First) {
|
||||
Ok(_) => (),
|
||||
Err(e) => match e {
|
||||
Error::NotFound => {
|
||||
return Ok(());
|
||||
}
|
||||
e => {
|
||||
return Err(e);
|
||||
}
|
||||
},
|
||||
}
|
||||
|
||||
let initial_value = Ok(inner_txn) as Result<&mut Txn, Error>;
|
||||
|
||||
let _ = cursor.iter().fold(initial_value, |acc, (key, _)| {
|
||||
acc.and_then(|inner_txn| {
|
||||
let job = inner_txn.get(&job_bucket, &key)?.inner()?.to_serde();
|
||||
|
||||
let job_value = Json::to_value_buf(job)?;
|
||||
inner_txn.set(&job_bucket, key, job_value)?;
|
||||
self.queue_job(&buckets, inner_txn, key)?;
|
||||
|
||||
Ok(inner_txn)
|
||||
})
|
||||
})?;
|
||||
|
||||
Ok(())
|
||||
})?;
|
||||
|
||||
read_txn.commit()?;
|
||||
write_txn.commit()?;
|
||||
|
||||
Ok(())
|
||||
}
|
||||
|
||||
pub fn check_stalled_jobs(&self) -> Result<(), Error> {
|
||||
let store = self.store.write()?;
|
||||
let job_bucket =
|
||||
|
@ -146,7 +198,7 @@ impl Storage {
|
|||
Ok(())
|
||||
}
|
||||
|
||||
pub fn dequeue_job(&self, limit: usize, queue: &str) -> Result<Vec<JobInfo>, Error> {
|
||||
pub fn stage_jobs(&self, limit: usize, queue: &str) -> Result<Vec<JobInfo>, Error> {
|
||||
let store = self.store.write()?;
|
||||
|
||||
trace!("Got store");
|
||||
|
@ -194,10 +246,12 @@ impl Storage {
|
|||
let (_inner_txn, vec) = cursor.iter().fold(initial_value, |acc, (key, _)| {
|
||||
acc.and_then(|(inner_txn, mut jobs)| {
|
||||
if jobs.len() < limit {
|
||||
let job = inner_txn.get(&job_bucket, &key)?.inner()?.to_serde();
|
||||
let mut job = inner_txn.get(&job_bucket, &key)?.inner()?.to_serde();
|
||||
|
||||
job.stage();
|
||||
|
||||
if job.is_ready(now) && job.is_in_queue(queue) {
|
||||
self.run_job(&buckets, inner_txn, key)?;
|
||||
self.stage_job(&buckets, inner_txn, key)?;
|
||||
|
||||
jobs.push(job);
|
||||
}
|
||||
|
@ -263,6 +317,7 @@ impl Storage {
|
|||
match status {
|
||||
JobStatus::Pending => self.queue_job(&buckets, &mut txn, job_id.as_ref())?,
|
||||
JobStatus::Running => self.run_job(&buckets, &mut txn, job_id.as_ref())?,
|
||||
JobStatus::Staged => self.stage_job(&buckets, &mut txn, job_id.as_ref())?,
|
||||
JobStatus::Failed => self.fail_job(&buckets, &mut txn, job_id.as_ref())?,
|
||||
JobStatus::Finished => self.finish_job(&buckets, &mut txn, job_id.as_ref())?,
|
||||
}
|
||||
|
@ -350,6 +405,21 @@ impl Storage {
|
|||
Ok(queue_map)
|
||||
}
|
||||
|
||||
fn stage_job<'env>(
|
||||
&self,
|
||||
buckets: &'env Buckets<'env>,
|
||||
txn: &mut Txn<'env>,
|
||||
id: &[u8],
|
||||
) -> Result<(), Error> {
|
||||
self.add_job_to(&buckets.staged, txn, id)?;
|
||||
self.delete_job_from(&buckets.finished, txn, id)?;
|
||||
self.delete_job_from(&buckets.failed, txn, id)?;
|
||||
self.delete_job_from(&buckets.running, txn, id)?;
|
||||
self.delete_job_from(&buckets.queued, txn, id)?;
|
||||
|
||||
Ok(())
|
||||
}
|
||||
|
||||
fn queue_job<'env>(
|
||||
&self,
|
||||
buckets: &'env Buckets<'env>,
|
||||
|
@ -360,6 +430,7 @@ impl Storage {
|
|||
self.delete_job_from(&buckets.finished, txn, id)?;
|
||||
self.delete_job_from(&buckets.failed, txn, id)?;
|
||||
self.delete_job_from(&buckets.running, txn, id)?;
|
||||
self.delete_job_from(&buckets.staged, txn, id)?;
|
||||
|
||||
Ok(())
|
||||
}
|
||||
|
@ -373,6 +444,7 @@ impl Storage {
|
|||
self.add_job_to(&buckets.failed, txn, id)?;
|
||||
self.delete_job_from(&buckets.finished, txn, id)?;
|
||||
self.delete_job_from(&buckets.running, txn, id)?;
|
||||
self.delete_job_from(&buckets.staged, txn, id)?;
|
||||
self.delete_job_from(&buckets.queued, txn, id)?;
|
||||
|
||||
Ok(())
|
||||
|
@ -385,6 +457,7 @@ impl Storage {
|
|||
id: &[u8],
|
||||
) -> Result<(), Error> {
|
||||
self.add_job_to(&buckets.running, txn, id)?;
|
||||
self.delete_job_from(&buckets.staged, txn, id)?;
|
||||
self.delete_job_from(&buckets.finished, txn, id)?;
|
||||
self.delete_job_from(&buckets.failed, txn, id)?;
|
||||
self.delete_job_from(&buckets.queued, txn, id)?;
|
||||
|
@ -400,6 +473,7 @@ impl Storage {
|
|||
) -> Result<(), Error> {
|
||||
self.add_job_to(&buckets.finished, txn, id)?;
|
||||
self.delete_job_from(&buckets.running, txn, id)?;
|
||||
self.delete_job_from(&buckets.staged, txn, id)?;
|
||||
self.delete_job_from(&buckets.failed, txn, id)?;
|
||||
self.delete_job_from(&buckets.queued, txn, id)?;
|
||||
|
||||
|
@ -500,13 +574,14 @@ impl Storage {
|
|||
Ok(item)
|
||||
}
|
||||
|
||||
fn buckets() -> [&'static str; 8] {
|
||||
fn buckets() -> [&'static str; 9] {
|
||||
[
|
||||
Storage::id_store(),
|
||||
Storage::job_store(),
|
||||
Storage::job_queue(),
|
||||
Storage::job_failed(),
|
||||
Storage::job_running(),
|
||||
Storage::job_staged(),
|
||||
Storage::job_lock(),
|
||||
Storage::job_finished(),
|
||||
Storage::queue_port(),
|
||||
|
@ -533,6 +608,10 @@ impl Storage {
|
|||
"job-running"
|
||||
}
|
||||
|
||||
fn job_staged() -> &'static str {
|
||||
"job-staged"
|
||||
}
|
||||
|
||||
fn job_finished() -> &'static str {
|
||||
"job-finished"
|
||||
}
|
||||
|
|
|
@ -17,7 +17,7 @@ zmq = "0.8"
|
|||
[features]
|
||||
default = ["tokio-zmq"]
|
||||
|
||||
[dependencies.jobs-core]
|
||||
[dependencies.background-jobs-core]
|
||||
version = "0.1"
|
||||
path = "../jobs-core"
|
||||
|
||||
|
|
|
@ -4,9 +4,9 @@ use std::{
|
|||
sync::Arc,
|
||||
};
|
||||
|
||||
use background_jobs_core::Storage;
|
||||
use failure::{Error, Fail};
|
||||
use futures::{future::poll_fn, Future};
|
||||
use jobs_core::Storage;
|
||||
use log::{error, info};
|
||||
use tokio_threadpool::blocking;
|
||||
use zmq::Context;
|
||||
|
@ -47,6 +47,7 @@ impl Config {
|
|||
|
||||
blocking(move || {
|
||||
let storage = Arc::new(Storage::init(runner_id, db_path)?);
|
||||
storage.requeue_staged_jobs()?;
|
||||
storage.check_stalled_jobs()?;
|
||||
let port_map = storage.get_port_mapping(base_port, queues)?;
|
||||
|
||||
|
|
|
@ -1,10 +1,10 @@
|
|||
use std::{sync::Arc, time::Duration};
|
||||
|
||||
use background_jobs_core::{JobInfo, Storage};
|
||||
use failure::{Error, Fail};
|
||||
use futures::{future::poll_fn, Future, Stream};
|
||||
#[cfg(feature = "futures-zmq")]
|
||||
use futures_zmq::{prelude::*, Multipart, Pull};
|
||||
use jobs_core::{JobInfo, Storage};
|
||||
use log::{error, info, trace};
|
||||
use tokio::timer::Delay;
|
||||
use tokio_threadpool::blocking;
|
||||
|
|
|
@ -1,10 +1,10 @@
|
|||
use std::{sync::Arc, time::Duration};
|
||||
|
||||
use background_jobs_core::{JobInfo, Storage};
|
||||
use failure::Error;
|
||||
use futures::{future::poll_fn, stream::iter_ok, Future, Stream};
|
||||
#[cfg(feature = "futures-zmq")]
|
||||
use futures_zmq::{prelude::*, Multipart, Push};
|
||||
use jobs_core::{JobInfo, Storage};
|
||||
use log::{error, info};
|
||||
use tokio::timer::{Delay, Interval};
|
||||
use tokio_threadpool::blocking;
|
||||
|
@ -117,7 +117,7 @@ fn wrap_fetch_queue(storage: Arc<Storage>, queue: &str) -> Result<Vec<Multipart>
|
|||
}
|
||||
|
||||
fn fetch_queue(storage: Arc<Storage>, queue: &str) -> Result<Vec<JobInfo>, Error> {
|
||||
storage.dequeue_job(100, queue).map_err(Error::from)
|
||||
storage.stage_jobs(100, queue).map_err(Error::from)
|
||||
}
|
||||
|
||||
struct ResetPushConfig {
|
||||
|
|
|
@ -1,8 +1,8 @@
|
|||
use std::{sync::Arc, time::Duration};
|
||||
|
||||
use background_jobs_core::Storage;
|
||||
use failure::Error;
|
||||
use futures::{future::poll_fn, Future, Stream};
|
||||
use jobs_core::Storage;
|
||||
use log::{error, info};
|
||||
use tokio::timer::{Delay, Interval};
|
||||
use tokio_threadpool::blocking;
|
||||
|
|
|
@ -1,10 +1,10 @@
|
|||
use std::sync::Arc;
|
||||
|
||||
use background_jobs_core::JobInfo;
|
||||
use failure::Error;
|
||||
use futures::{future::IntoFuture, Future};
|
||||
#[cfg(feature = "futures-zmq")]
|
||||
use futures_zmq::{prelude::*, Push};
|
||||
use jobs_core::JobInfo;
|
||||
use log::{debug, trace};
|
||||
#[cfg(feature = "tokio-zmq")]
|
||||
use tokio_zmq::{prelude::*, Push};
|
||||
|
|
|
@ -1,13 +1,13 @@
|
|||
use std::{sync::Arc, time::Duration};
|
||||
|
||||
use background_jobs_core::{JobInfo, Processors};
|
||||
use failure::{Error, Fail};
|
||||
use futures::{
|
||||
future::{Either, IntoFuture},
|
||||
Future, Stream,
|
||||
sync::mpsc::{channel, Sender},
|
||||
Future, Sink, Stream,
|
||||
};
|
||||
#[cfg(feature = "futures-zmq")]
|
||||
use futures_zmq::{prelude::*, Multipart, Pull, Push};
|
||||
use jobs_core::{JobInfo, Processors};
|
||||
use log::{error, info};
|
||||
use tokio::timer::Delay;
|
||||
#[cfg(feature = "tokio-zmq")]
|
||||
|
@ -17,6 +17,7 @@ use zmq::{Context, Message};
|
|||
pub(crate) struct Worker {
|
||||
pull: Pull,
|
||||
push: Push,
|
||||
push2: Push,
|
||||
push_address: String,
|
||||
pull_address: String,
|
||||
queue: String,
|
||||
|
@ -49,6 +50,7 @@ impl Worker {
|
|||
|
||||
let Worker {
|
||||
push,
|
||||
push2,
|
||||
pull,
|
||||
push_address: _,
|
||||
pull_address: _,
|
||||
|
@ -57,11 +59,25 @@ impl Worker {
|
|||
context: _,
|
||||
} = self;
|
||||
|
||||
let (tx, rx) = channel(5);
|
||||
|
||||
tokio::spawn(
|
||||
rx.map_err(|_| RecvError)
|
||||
.from_err::<Error>()
|
||||
.and_then(serialize_request)
|
||||
.forward(push2.sink(1))
|
||||
.map(|_| ())
|
||||
.or_else(|_| Ok(())),
|
||||
);
|
||||
|
||||
let fut = pull
|
||||
.stream()
|
||||
.from_err::<Error>()
|
||||
.and_then(move |multipart| wrap_processing(multipart, &processors))
|
||||
.forward(push.sink(2))
|
||||
.and_then(parse_multipart)
|
||||
.and_then(move |job| report_running(job, tx.clone()))
|
||||
.and_then(move |job| process_job(job, &processors))
|
||||
.and_then(serialize_request)
|
||||
.forward(push.sink(1))
|
||||
.map(move |_| info!("worker for queue {} is shutting down", queue))
|
||||
.map_err(|e| {
|
||||
error!("Error processing job, {}", e);
|
||||
|
@ -105,14 +121,20 @@ impl ResetWorker {
|
|||
Push::builder(self.context.clone())
|
||||
.connect(&self.push_address)
|
||||
.build()
|
||||
.join(
|
||||
Push::builder(self.context.clone())
|
||||
.connect(&self.push_address)
|
||||
.build(),
|
||||
)
|
||||
.join(
|
||||
Pull::builder(self.context.clone())
|
||||
.connect(&self.pull_address)
|
||||
.build(),
|
||||
)
|
||||
.map(|(push, pull)| {
|
||||
.map(|((push, push2), pull)| {
|
||||
let config = Worker {
|
||||
push,
|
||||
push2,
|
||||
pull,
|
||||
push_address: self.push_address,
|
||||
pull_address: self.pull_address,
|
||||
|
@ -142,18 +164,15 @@ fn parse_multipart(mut multipart: Multipart) -> Result<JobInfo, Error> {
|
|||
Ok(parsed)
|
||||
}
|
||||
|
||||
fn wrap_processing(
|
||||
multipart: Multipart,
|
||||
processors: &Processors,
|
||||
) -> impl Future<Item = Multipart, Error = Error> {
|
||||
let msg = match parse_multipart(multipart) {
|
||||
Ok(msg) => msg,
|
||||
Err(e) => return Either::A(Err(e).into_future()),
|
||||
};
|
||||
fn report_running(
|
||||
mut job: JobInfo,
|
||||
push: Sender<JobInfo>,
|
||||
) -> impl Future<Item = JobInfo, Error = Error> {
|
||||
job.run();
|
||||
|
||||
let fut = process_job(msg, processors).and_then(serialize_request);
|
||||
|
||||
Either::B(fut)
|
||||
push.send(job.clone())
|
||||
.map(move |_| job)
|
||||
.map_err(|_| NotifyError.into())
|
||||
}
|
||||
|
||||
fn process_job(
|
||||
|
@ -173,3 +192,11 @@ struct ParseError;
|
|||
#[derive(Clone, Debug, Fail)]
|
||||
#[fail(display = "Error processing job")]
|
||||
struct ProcessError;
|
||||
|
||||
#[derive(Clone, Debug, Fail)]
|
||||
#[fail(display = "Error notifying running has started")]
|
||||
struct NotifyError;
|
||||
|
||||
#[derive(Clone, Debug, Fail)]
|
||||
#[fail(display = "Error receiving from mpsc")]
|
||||
struct RecvError;
|
||||
|
|
|
@ -1,8 +1,8 @@
|
|||
use std::{collections::BTreeMap, sync::Arc};
|
||||
|
||||
use background_jobs_core::{Processor, Processors};
|
||||
use failure::Fail;
|
||||
use futures::Future;
|
||||
use jobs_core::{Processor, Processors};
|
||||
use log::{error, info};
|
||||
use zmq::Context;
|
||||
|
||||
|
|
|
@ -10,6 +10,6 @@ log = "0.4"
|
|||
tokio = "0.1"
|
||||
tokio-threadpool = "0.1"
|
||||
|
||||
[dependencies.jobs-core]
|
||||
[dependencies.background-jobs-core]
|
||||
version = "0.1"
|
||||
path = "../jobs-core"
|
||||
|
|
|
@ -6,12 +6,12 @@ use std::{
|
|||
time::{Duration, Instant},
|
||||
};
|
||||
|
||||
use background_jobs_core::{JobInfo, Processor, Processors, Storage};
|
||||
use futures::{
|
||||
future::{poll_fn, Either, IntoFuture},
|
||||
sync::mpsc::{channel, Receiver, SendError, Sender},
|
||||
Future, Sink, Stream,
|
||||
};
|
||||
use jobs_core::{JobInfo, Processor, Processors, Storage};
|
||||
use tokio::timer::Interval;
|
||||
use tokio_threadpool::blocking;
|
||||
|
||||
|
@ -86,7 +86,7 @@ fn try_process_job(
|
|||
|
||||
blocking(move || {
|
||||
storage
|
||||
.dequeue_job(processor_count, &queue)
|
||||
.stage_jobs(processor_count, &queue)
|
||||
.map_err(|e| error!("Error dequeuing job, {}", e))
|
||||
})
|
||||
.map_err(|e| error!("Error blocking, {}", e))
|
||||
|
|
14
src/lib.rs
14
src/lib.rs
|
@ -1,12 +1,12 @@
|
|||
pub use jobs_core::{
|
||||
pub use background_jobs_core::{
|
||||
Backoff, JobError, JobInfo, JobStatus, MaxRetries, Processor, Processors, ShouldStop, Storage,
|
||||
};
|
||||
|
||||
#[cfg(feature = "jobs-tokio")]
|
||||
pub use jobs_tokio::{JobRunner, ProcessorHandle};
|
||||
#[cfg(feature = "background-jobs-tokio")]
|
||||
pub use background_jobs_tokio::{JobRunner, ProcessorHandle};
|
||||
|
||||
#[cfg(feature = "jobs-actix")]
|
||||
pub use jobs_actix::{JobsActor, JobsBuilder, QueueJob};
|
||||
#[cfg(feature = "background-jobs-actix")]
|
||||
pub use background_jobs_actix::{JobsActor, JobsBuilder, QueueJob};
|
||||
|
||||
#[cfg(feature = "jobs-server-tokio")]
|
||||
pub use jobs_server_tokio::{ServerConfig, SpawnerConfig, WorkerConfig};
|
||||
#[cfg(feature = "background-jobs-server-tokio")]
|
||||
pub use background_jobs_server_tokio::{ServerConfig, SpawnerConfig, WorkerConfig};
|
||||
|
|
Loading…
Reference in a new issue