This commit is contained in:
asonix 2020-03-22 12:52:43 -05:00
parent 1da4c7f64c
commit cc2d0fbf0d
9 changed files with 87 additions and 77 deletions

View file

@ -23,3 +23,4 @@ serde = { version = "1.0", features = ["derive"] }
serde_json = "1.0" serde_json = "1.0"
thiserror = "1.0" thiserror = "1.0"
tokio = { version = "0.2.13", features = ["sync"] } tokio = { version = "0.2.13", features = ["sync"] }
uuid = { version ="0.8.1", features = ["v4", "serde"] }

View file

@ -213,44 +213,32 @@ where
/// Start the workers in the current arbiter /// Start the workers in the current arbiter
pub fn start(self, queue_handle: QueueHandle) { pub fn start(self, queue_handle: QueueHandle) {
let processors = self.processors.clone(); for (key, count) in self.queues.into_iter() {
for _ in 0..count {
self.queues.into_iter().fold(0, |acc, (key, count)| {
(0..count).for_each(|i| {
local_worker( local_worker(
acc + i + 1000,
key.clone(), key.clone(),
processors.cached(), self.processors.cached(),
queue_handle.inner.clone(), queue_handle.inner.clone(),
); );
}); }
}
acc + count
});
} }
/// Start the workers in the provided arbiter /// Start the workers in the provided arbiter
pub fn start_in_arbiter(self, arbiter: &Arbiter, queue_handle: QueueHandle) { pub fn start_in_arbiter(self, arbiter: &Arbiter, queue_handle: QueueHandle) {
let processors = self.processors.clone(); for (key, count) in self.queues.into_iter() {
self.queues.into_iter().fold(0, |acc, (key, count)| { for _ in 0..count {
(0..count).for_each(|i| {
let processors = processors.clone();
let queue_handle = queue_handle.clone();
let key = key.clone(); let key = key.clone();
arbiter.exec_fn(move || { let processors = self.processors.clone();
local_worker( let server = queue_handle.inner.clone();
acc + i + 1000,
key.clone(),
processors.cached(),
queue_handle.inner.clone(),
);
});
});
acc + count arbiter.exec_fn(move || {
local_worker(key, processors.cached(), server);
}); });
} }
} }
}
}
/// A handle to the job server, used for queuing new jobs /// A handle to the job server, used for queuing new jobs
/// ///

View file

@ -1,11 +1,12 @@
use anyhow::Error; use anyhow::Error;
use background_jobs_core::{JobInfo, NewJobInfo, ReturnJobInfo, Stats, Storage}; use background_jobs_core::{JobInfo, NewJobInfo, ReturnJobInfo, Stats, Storage};
use uuid::Uuid;
#[async_trait::async_trait] #[async_trait::async_trait]
pub(crate) trait ActixStorage { pub(crate) trait ActixStorage {
async fn new_job(&self, job: NewJobInfo) -> Result<u64, Error>; async fn new_job(&self, job: NewJobInfo) -> Result<Uuid, Error>;
async fn request_job(&self, queue: &str, runner_id: u64) -> Result<Option<JobInfo>, Error>; async fn request_job(&self, queue: &str, runner_id: Uuid) -> Result<Option<JobInfo>, Error>;
async fn return_job(&self, ret: ReturnJobInfo) -> Result<(), Error>; async fn return_job(&self, ret: ReturnJobInfo) -> Result<(), Error>;
@ -23,11 +24,11 @@ where
S: Storage + Send + Sync, S: Storage + Send + Sync,
S::Error: Send + Sync + 'static, S::Error: Send + Sync + 'static,
{ {
async fn new_job(&self, job: NewJobInfo) -> Result<u64, Error> { async fn new_job(&self, job: NewJobInfo) -> Result<Uuid, Error> {
Ok(self.0.new_job(job).await?) Ok(self.0.new_job(job).await?)
} }
async fn request_job(&self, queue: &str, runner_id: u64) -> Result<Option<JobInfo>, Error> { async fn request_job(&self, queue: &str, runner_id: Uuid) -> Result<Option<JobInfo>, Error> {
Ok(self.0.request_job(queue, runner_id).await?) Ok(self.0.request_job(queue, runner_id).await?)
} }

View file

@ -2,12 +2,13 @@ use crate::Server;
use background_jobs_core::{CachedProcessorMap, JobInfo}; use background_jobs_core::{CachedProcessorMap, JobInfo};
use log::{debug, error, warn}; use log::{debug, error, warn};
use tokio::sync::mpsc::{channel, Sender}; use tokio::sync::mpsc::{channel, Sender};
use uuid::Uuid;
#[async_trait::async_trait] #[async_trait::async_trait]
pub trait Worker { pub trait Worker {
async fn process_job(&self, job: JobInfo) -> Result<(), JobInfo>; async fn process_job(&self, job: JobInfo) -> Result<(), JobInfo>;
fn id(&self) -> u64; fn id(&self) -> Uuid;
fn queue(&self) -> &str; fn queue(&self) -> &str;
} }
@ -15,7 +16,7 @@ pub trait Worker {
#[derive(Clone)] #[derive(Clone)]
pub(crate) struct LocalWorkerHandle { pub(crate) struct LocalWorkerHandle {
tx: Sender<JobInfo>, tx: Sender<JobInfo>,
id: u64, id: Uuid,
queue: String, queue: String,
} }
@ -31,7 +32,7 @@ impl Worker for LocalWorkerHandle {
} }
} }
fn id(&self) -> u64 { fn id(&self) -> Uuid {
self.id self.id
} }
@ -41,13 +42,14 @@ impl Worker for LocalWorkerHandle {
} }
pub(crate) fn local_worker<State>( pub(crate) fn local_worker<State>(
id: u64,
queue: String, queue: String,
processors: CachedProcessorMap<State>, processors: CachedProcessorMap<State>,
server: Server, server: Server,
) where ) where
State: Clone + 'static, State: Clone + 'static,
{ {
let id = Uuid::new_v4();
let (tx, mut rx) = channel(16); let (tx, mut rx) = channel(16);
let handle = LocalWorkerHandle { let handle = LocalWorkerHandle {

View file

@ -18,3 +18,4 @@ log = "0.4"
serde = { version = "1.0", features = ["derive"] } serde = { version = "1.0", features = ["derive"] }
serde_json = "1.0" serde_json = "1.0"
thiserror = "1.0" thiserror = "1.0"
uuid = { version = "0.8.1", features = ["serde", "v4"] }

View file

@ -2,30 +2,31 @@ use crate::{Backoff, JobResult, JobStatus, MaxRetries, ShouldStop};
use chrono::{offset::Utc, DateTime, Duration}; use chrono::{offset::Utc, DateTime, Duration};
use log::trace; use log::trace;
use serde_json::Value; use serde_json::Value;
use uuid::Uuid;
#[derive(Clone, Debug, PartialEq, serde::Deserialize, serde::Serialize)] #[derive(Clone, Debug, PartialEq, serde::Deserialize, serde::Serialize)]
/// Information about the sate of an attempted job /// Information about the sate of an attempted job
pub struct ReturnJobInfo { pub struct ReturnJobInfo {
pub(crate) id: u64, pub(crate) id: Uuid,
pub(crate) result: JobResult, pub(crate) result: JobResult,
} }
impl ReturnJobInfo { impl ReturnJobInfo {
pub(crate) fn fail(id: u64) -> Self { pub(crate) fn fail(id: Uuid) -> Self {
ReturnJobInfo { ReturnJobInfo {
id, id,
result: JobResult::Failure, result: JobResult::Failure,
} }
} }
pub(crate) fn pass(id: u64) -> Self { pub(crate) fn pass(id: Uuid) -> Self {
ReturnJobInfo { ReturnJobInfo {
id, id,
result: JobResult::Success, result: JobResult::Success,
} }
} }
pub(crate) fn missing_processor(id: u64) -> Self { pub(crate) fn missing_processor(id: Uuid) -> Self {
ReturnJobInfo { ReturnJobInfo {
id, id,
result: JobResult::MissingProcessor, result: JobResult::MissingProcessor,
@ -94,7 +95,7 @@ impl NewJobInfo {
self.next_queue.is_none() self.next_queue.is_none()
} }
pub(crate) fn with_id(self, id: u64) -> JobInfo { pub(crate) fn with_id(self, id: Uuid) -> JobInfo {
JobInfo { JobInfo {
id, id,
processor: self.processor, processor: self.processor,
@ -120,7 +121,7 @@ impl NewJobInfo {
/// new_job method. /// new_job method.
pub struct JobInfo { pub struct JobInfo {
/// ID of the job /// ID of the job
id: u64, id: Uuid,
/// Name of the processor that should handle this job /// Name of the processor that should handle this job
processor: String, processor: String,
@ -174,7 +175,7 @@ impl JobInfo {
} }
/// The ID of this job /// The ID of this job
pub fn id(&self) -> u64 { pub fn id(&self) -> Uuid {
self.id self.id
} }

View file

@ -1,8 +1,8 @@
use crate::{JobInfo, NewJobInfo, ReturnJobInfo, Stats};
use chrono::offset::Utc; use chrono::offset::Utc;
use log::info; use log::info;
use std::error::Error; use std::error::Error;
use uuid::Uuid;
use crate::{JobInfo, NewJobInfo, ReturnJobInfo, Stats};
/// Define a storage backend for jobs /// Define a storage backend for jobs
/// ///
@ -16,7 +16,7 @@ pub trait Storage: Clone + Send {
type Error: Error + Send + Sync; type Error: Error + Send + Sync;
/// This method generates unique IDs for jobs /// This method generates unique IDs for jobs
async fn generate_id(&self) -> Result<u64, Self::Error>; async fn generate_id(&self) -> Result<Uuid, Self::Error>;
/// This method should store the supplied job /// This method should store the supplied job
/// ///
@ -25,7 +25,7 @@ pub trait Storage: Clone + Send {
async fn save_job(&self, job: JobInfo) -> Result<(), Self::Error>; async fn save_job(&self, job: JobInfo) -> Result<(), Self::Error>;
/// This method should return the job with the given ID regardless of what state the job is in. /// This method should return the job with the given ID regardless of what state the job is in.
async fn fetch_job(&self, id: u64) -> Result<Option<JobInfo>, Self::Error>; async fn fetch_job(&self, id: Uuid) -> Result<Option<JobInfo>, Self::Error>;
/// This should fetch a job ready to be processed from the queue /// This should fetch a job ready to be processed from the queue
/// ///
@ -35,15 +35,15 @@ pub trait Storage: Clone + Send {
/// This method tells the storage mechanism to mark the given job as being in the provided /// This method tells the storage mechanism to mark the given job as being in the provided
/// queue /// queue
async fn queue_job(&self, queue: &str, id: u64) -> Result<(), Self::Error>; async fn queue_job(&self, queue: &str, id: Uuid) -> Result<(), Self::Error>;
/// This method tells the storage mechanism to mark a given job as running /// This method tells the storage mechanism to mark a given job as running
async fn run_job(&self, id: u64, runner_id: u64) -> Result<(), Self::Error>; async fn run_job(&self, id: Uuid, runner_id: Uuid) -> Result<(), Self::Error>;
/// This method tells the storage mechanism to remove the job /// This method tells the storage mechanism to remove the job
/// ///
/// This happens when a job has been completed or has failed too many times /// This happens when a job has been completed or has failed too many times
async fn delete_job(&self, id: u64) -> Result<(), Self::Error>; async fn delete_job(&self, id: Uuid) -> Result<(), Self::Error>;
/// This method returns the current statistics, or Stats::default() if none exists. /// This method returns the current statistics, or Stats::default() if none exists.
async fn get_stats(&self) -> Result<Stats, Self::Error>; async fn get_stats(&self) -> Result<Stats, Self::Error>;
@ -55,7 +55,7 @@ pub trait Storage: Clone + Send {
F: Fn(Stats) -> Stats + Send + 'static; F: Fn(Stats) -> Stats + Send + 'static;
/// Generate a new job based on the provided NewJobInfo /// Generate a new job based on the provided NewJobInfo
async fn new_job(&self, job: NewJobInfo) -> Result<u64, Self::Error> { async fn new_job(&self, job: NewJobInfo) -> Result<Uuid, Self::Error> {
let id = self.generate_id().await?; let id = self.generate_id().await?;
let job = job.with_id(id); let job = job.with_id(id);
@ -72,7 +72,7 @@ pub trait Storage: Clone + Send {
async fn request_job( async fn request_job(
&self, &self,
queue: &str, queue: &str,
runner_id: u64, runner_id: Uuid,
) -> Result<Option<JobInfo>, Self::Error> { ) -> Result<Option<JobInfo>, Self::Error> {
match self.fetch_job_from_queue(queue).await? { match self.fetch_job_from_queue(queue).await? {
Some(mut job) => { Some(mut job) => {
@ -138,6 +138,7 @@ pub mod memory_storage {
use chrono::Utc; use chrono::Utc;
use futures::lock::Mutex; use futures::lock::Mutex;
use std::{collections::HashMap, convert::Infallible, sync::Arc}; use std::{collections::HashMap, convert::Infallible, sync::Arc};
use uuid::Uuid;
#[derive(Clone)] #[derive(Clone)]
/// An In-Memory store for jobs /// An In-Memory store for jobs
@ -147,11 +148,10 @@ pub mod memory_storage {
#[derive(Clone)] #[derive(Clone)]
struct Inner { struct Inner {
count: u64, jobs: HashMap<Uuid, JobInfo>,
jobs: HashMap<u64, JobInfo>, queues: HashMap<Uuid, String>,
queues: HashMap<u64, String>, worker_ids: HashMap<Uuid, Uuid>,
worker_ids: HashMap<u64, u64>, worker_ids_inverse: HashMap<Uuid, Uuid>,
worker_ids_inverse: HashMap<u64, u64>,
stats: Stats, stats: Stats,
} }
@ -160,7 +160,6 @@ pub mod memory_storage {
pub fn new() -> Self { pub fn new() -> Self {
Storage { Storage {
inner: Arc::new(Mutex::new(Inner { inner: Arc::new(Mutex::new(Inner {
count: 0,
jobs: HashMap::new(), jobs: HashMap::new(),
queues: HashMap::new(), queues: HashMap::new(),
worker_ids: HashMap::new(), worker_ids: HashMap::new(),
@ -175,11 +174,15 @@ pub mod memory_storage {
impl super::Storage for Storage { impl super::Storage for Storage {
type Error = Infallible; type Error = Infallible;
async fn generate_id(&self) -> Result<u64, Self::Error> { async fn generate_id(&self) -> Result<Uuid, Self::Error> {
let mut inner = self.inner.lock().await; let uuid = loop {
let id = inner.count; let uuid = Uuid::new_v4();
inner.count = inner.count.wrapping_add(1); if !self.inner.lock().await.jobs.contains_key(&uuid) {
Ok(id) break uuid;
}
};
Ok(uuid)
} }
async fn save_job(&self, job: JobInfo) -> Result<(), Self::Error> { async fn save_job(&self, job: JobInfo) -> Result<(), Self::Error> {
@ -188,7 +191,7 @@ pub mod memory_storage {
Ok(()) Ok(())
} }
async fn fetch_job(&self, id: u64) -> Result<Option<JobInfo>, Self::Error> { async fn fetch_job(&self, id: Uuid) -> Result<Option<JobInfo>, Self::Error> {
let j = self.inner.lock().await.jobs.get(&id).map(|j| j.clone()); let j = self.inner.lock().await.jobs.get(&id).map(|j| j.clone());
Ok(j) Ok(j)
@ -221,12 +224,12 @@ pub mod memory_storage {
Ok(j) Ok(j)
} }
async fn queue_job(&self, queue: &str, id: u64) -> Result<(), Self::Error> { async fn queue_job(&self, queue: &str, id: Uuid) -> Result<(), Self::Error> {
self.inner.lock().await.queues.insert(id, queue.to_owned()); self.inner.lock().await.queues.insert(id, queue.to_owned());
Ok(()) Ok(())
} }
async fn run_job(&self, id: u64, worker_id: u64) -> Result<(), Self::Error> { async fn run_job(&self, id: Uuid, worker_id: Uuid) -> Result<(), Self::Error> {
let mut inner = self.inner.lock().await; let mut inner = self.inner.lock().await;
inner.worker_ids.insert(id, worker_id); inner.worker_ids.insert(id, worker_id);
@ -234,7 +237,7 @@ pub mod memory_storage {
Ok(()) Ok(())
} }
async fn delete_job(&self, id: u64) -> Result<(), Self::Error> { async fn delete_job(&self, id: Uuid) -> Result<(), Self::Error> {
let mut inner = self.inner.lock().await; let mut inner = self.inner.lock().await;
inner.jobs.remove(&id); inner.jobs.remove(&id);
inner.queues.remove(&id); inner.queues.remove(&id);

View file

@ -17,3 +17,4 @@ background-jobs-core = { version = "0.7", path = "../jobs-core" }
chrono = "0.4" chrono = "0.4"
sled-extensions = { version = "0.3.0-alpha.0", features = ["bincode", "cbor"], git = "https://git.asonix.dog/Aardwolf/sled-extensions" } sled-extensions = { version = "0.3.0-alpha.0", features = ["bincode", "cbor"], git = "https://git.asonix.dog/Aardwolf/sled-extensions" }
thiserror = "1.0" thiserror = "1.0"
uuid = { version = "0.8.1", features = ["v4", "serde"] }

View file

@ -17,6 +17,7 @@ use actix_threadpool::{run, BlockingError};
use background_jobs_core::{JobInfo, Stats, Storage}; use background_jobs_core::{JobInfo, Stats, Storage};
use chrono::offset::Utc; use chrono::offset::Utc;
use sled_extensions::{bincode::Tree, cbor, Db, DbExt}; use sled_extensions::{bincode::Tree, cbor, Db, DbExt};
use uuid::Uuid;
/// The error produced by sled storage calls /// The error produced by sled storage calls
#[derive(Debug, thiserror::Error)] #[derive(Debug, thiserror::Error)]
@ -37,11 +38,11 @@ pub type Result<T> = std::result::Result<T, Error>;
/// The Sled-backed storage implementation /// The Sled-backed storage implementation
pub struct SledStorage { pub struct SledStorage {
jobinfo: cbor::Tree<JobInfo>, jobinfo: cbor::Tree<JobInfo>,
running: Tree<u64>, running: Tree<Uuid>,
running_inverse: Tree<u64>, running_inverse: Tree<Uuid>,
queue: Tree<String>, queue: Tree<String>,
stats: Tree<Stats>, stats: Tree<Stats>,
lock: Tree<u64>, lock: Tree<Uuid>,
db: Db, db: Db,
} }
@ -49,10 +50,21 @@ pub struct SledStorage {
impl Storage for SledStorage { impl Storage for SledStorage {
type Error = Error; type Error = Error;
async fn generate_id(&self) -> Result<u64> { async fn generate_id(&self) -> Result<Uuid> {
let this = self.clone(); let this = self.clone();
Ok(run(move || Ok(this.db.generate_id()?) as sled_extensions::Result<u64>).await?) Ok(run(move || {
let uuid = loop {
let uuid = Uuid::new_v4();
if !this.jobinfo.contains_key(job_key(uuid))? {
break uuid;
}
};
Ok(uuid) as sled_extensions::Result<Uuid>
})
.await?)
} }
async fn save_job(&self, job: JobInfo) -> Result<()> { async fn save_job(&self, job: JobInfo) -> Result<()> {
@ -66,7 +78,7 @@ impl Storage for SledStorage {
.await?) .await?)
} }
async fn fetch_job(&self, id: u64) -> Result<Option<JobInfo>> { async fn fetch_job(&self, id: Uuid) -> Result<Option<JobInfo>> {
let this = self.clone(); let this = self.clone();
Ok(run(move || this.jobinfo.get(job_key(id))).await?) Ok(run(move || this.jobinfo.get(job_key(id))).await?)
@ -111,7 +123,7 @@ impl Storage for SledStorage {
.await?) .await?)
} }
async fn queue_job(&self, queue: &str, id: u64) -> Result<()> { async fn queue_job(&self, queue: &str, id: Uuid) -> Result<()> {
let this = self.clone(); let this = self.clone();
let queue = queue.to_owned(); let queue = queue.to_owned();
@ -125,7 +137,7 @@ impl Storage for SledStorage {
.await?) .await?)
} }
async fn run_job(&self, id: u64, runner_id: u64) -> Result<()> { async fn run_job(&self, id: Uuid, runner_id: Uuid) -> Result<()> {
let this = self.clone(); let this = self.clone();
Ok(run(move || { Ok(run(move || {
@ -139,7 +151,7 @@ impl Storage for SledStorage {
.await?) .await?)
} }
async fn delete_job(&self, id: u64) -> Result<()> { async fn delete_job(&self, id: Uuid) -> Result<()> {
let this = self.clone(); let this = self.clone();
Ok(run(move || { Ok(run(move || {
@ -204,7 +216,7 @@ impl SledStorage {
where where
F: Fn() -> sled_extensions::Result<T>, F: Fn() -> sled_extensions::Result<T>,
{ {
let id = self.db.generate_id()?; let id = Uuid::new_v4();
let mut prev; let mut prev;
while { while {
@ -224,11 +236,11 @@ impl SledStorage {
} }
} }
fn job_key(id: u64) -> String { fn job_key(id: Uuid) -> String {
format!("job-{}", id) format!("job-{}", id)
} }
fn runner_key(runner_id: u64) -> String { fn runner_key(runner_id: Uuid) -> String {
format!("runner-{}", runner_id) format!("runner-{}", runner_id)
} }