mirror of
https://git.asonix.dog/asonix/relay.git
synced 2024-11-21 17:11:05 +00:00
Measure bcrypt, change DashMap to RwLock<HashMap for collector
This commit is contained in:
parent
1a638f7f8d
commit
e9f312bed5
2 changed files with 28 additions and 135 deletions
156
src/collector.rs
156
src/collector.rs
|
@ -1,4 +1,3 @@
|
|||
use dashmap::DashMap;
|
||||
use indexmap::IndexMap;
|
||||
use metrics::{Key, Recorder, SetRecorderError};
|
||||
use metrics_util::{
|
||||
|
@ -8,7 +7,7 @@ use metrics_util::{
|
|||
use quanta::Clock;
|
||||
use std::{
|
||||
collections::{BTreeMap, HashMap},
|
||||
sync::{atomic::Ordering, Arc},
|
||||
sync::{atomic::Ordering, Arc, RwLock},
|
||||
time::Duration,
|
||||
};
|
||||
|
||||
|
@ -23,8 +22,8 @@ pub struct MemoryCollector {
|
|||
}
|
||||
|
||||
struct Inner {
|
||||
descriptions: DashMap<String, metrics::SharedString>,
|
||||
distributions: DashMap<String, IndexMap<Vec<(String, String)>, Summary>>,
|
||||
descriptions: RwLock<HashMap<String, metrics::SharedString>>,
|
||||
distributions: RwLock<HashMap<String, IndexMap<Vec<(String, String)>, Summary>>>,
|
||||
recency: Recency<Key>,
|
||||
registry: Registry<Key, GenerationalStorage<AtomicStorage>>,
|
||||
}
|
||||
|
@ -273,24 +272,24 @@ impl Inner {
|
|||
.recency
|
||||
.should_store_histogram(&key, gen, &self.registry)
|
||||
{
|
||||
let delete_by_name = if let Some(mut by_name) = self.distributions.get_mut(&name) {
|
||||
let mut d = self.distributions.write().unwrap();
|
||||
let delete_by_name = if let Some(by_name) = d.get_mut(&name) {
|
||||
by_name.remove(&labels);
|
||||
by_name.is_empty()
|
||||
} else {
|
||||
false
|
||||
};
|
||||
drop(d);
|
||||
|
||||
if delete_by_name {
|
||||
self.descriptions.remove(&name);
|
||||
self.descriptions.write().unwrap().remove(&name);
|
||||
}
|
||||
|
||||
continue;
|
||||
}
|
||||
|
||||
let mut outer_entry = self
|
||||
.distributions
|
||||
.entry(name.clone())
|
||||
.or_insert_with(IndexMap::new);
|
||||
let mut d = self.distributions.write().unwrap();
|
||||
let outer_entry = d.entry(name.clone()).or_insert_with(IndexMap::new);
|
||||
|
||||
let entry = outer_entry
|
||||
.entry(labels)
|
||||
|
@ -303,16 +302,16 @@ impl Inner {
|
|||
})
|
||||
}
|
||||
|
||||
self.distributions
|
||||
.iter()
|
||||
.map(|entry| {
|
||||
let d = self.distributions.read().unwrap().clone();
|
||||
let h = d
|
||||
.into_iter()
|
||||
.map(|(key, value)| {
|
||||
(
|
||||
entry.key().clone(),
|
||||
entry
|
||||
.value()
|
||||
.iter()
|
||||
key,
|
||||
value
|
||||
.into_iter()
|
||||
.map(|(labels, summary)| Histogram {
|
||||
labels: labels.iter().cloned().collect(),
|
||||
labels: labels.into_iter().collect(),
|
||||
value: [0.001, 0.01, 0.05, 0.1, 0.5, 0.9, 0.99, 1.0]
|
||||
.into_iter()
|
||||
.map(|q| (q, summary.quantile(q)))
|
||||
|
@ -321,7 +320,9 @@ impl Inner {
|
|||
.collect(),
|
||||
)
|
||||
})
|
||||
.collect()
|
||||
.collect();
|
||||
|
||||
h
|
||||
}
|
||||
|
||||
fn snapshot(&self) -> Snapshot {
|
||||
|
@ -362,10 +363,8 @@ impl MemoryCollector {
|
|||
key: &metrics::KeyName,
|
||||
description: metrics::SharedString,
|
||||
) {
|
||||
self.inner
|
||||
.descriptions
|
||||
.entry(key.as_str().to_owned())
|
||||
.or_insert(description);
|
||||
let mut d = self.inner.descriptions.write().unwrap();
|
||||
d.entry(key.as_str().to_owned()).or_insert(description);
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -415,114 +414,3 @@ impl Recorder for MemoryCollector {
|
|||
.get_or_create_histogram(key, |c| c.clone().into())
|
||||
}
|
||||
}
|
||||
|
||||
/*
|
||||
struct Bucket {
|
||||
begin: Instant,
|
||||
summary: Summary,
|
||||
}
|
||||
|
||||
pub(crate) struct RollingSummary {
|
||||
buckets: Vec<Bucket>,
|
||||
bucket_duration: Duration,
|
||||
expire_after: Duration,
|
||||
count: usize,
|
||||
}
|
||||
|
||||
impl Default for RollingSummary {
|
||||
fn default() -> Self {
|
||||
Self::new(
|
||||
Duration::from_secs(5 * MINUTES),
|
||||
Duration::from_secs(1 * DAYS),
|
||||
)
|
||||
}
|
||||
}
|
||||
|
||||
impl RollingSummary {
|
||||
fn new(bucket_duration: Duration, expire_after: Duration) -> Self {
|
||||
Self {
|
||||
buckets: Vec::new(),
|
||||
bucket_duration,
|
||||
expire_after,
|
||||
count: 0,
|
||||
}
|
||||
}
|
||||
|
||||
fn add(&mut self, value: f64, now: Instant) {
|
||||
self.count += 1;
|
||||
|
||||
// try adding to existing bucket
|
||||
for bucket in &mut self.buckets {
|
||||
let end = bucket.begin + self.bucket_duration;
|
||||
|
||||
if now >= end {
|
||||
break;
|
||||
}
|
||||
|
||||
if now >= bucket.begin {
|
||||
bucket.summary.add(value);
|
||||
return;
|
||||
}
|
||||
}
|
||||
|
||||
// if we're adding a new bucket, clean old buckets first
|
||||
if let Some(cutoff) = now.checked_sub(self.expire_after) {
|
||||
self.buckets.retain(|b| b.begin > cutoff);
|
||||
}
|
||||
|
||||
let mut summary = Summary::with_defaults();
|
||||
summary.add(value);
|
||||
|
||||
// if there's no buckets, make one and return
|
||||
if self.buckets.is_empty() {
|
||||
self.buckets.push(Bucket {
|
||||
summary,
|
||||
begin: now,
|
||||
});
|
||||
return;
|
||||
}
|
||||
|
||||
let mut begin = self.buckets[0].begin;
|
||||
|
||||
// there are buckets, but none can hold our value, see why
|
||||
if now < self.buckets[0].begin {
|
||||
// create an old bucket
|
||||
|
||||
while now < begin {
|
||||
begin -= self.bucket_duration;
|
||||
}
|
||||
|
||||
self.buckets.push(Bucket { begin, summary });
|
||||
self.buckets.sort_unstable_by(|a, b| b.begin.cmp(&a.begin));
|
||||
} else {
|
||||
// create a new bucket
|
||||
let mut end = self.buckets[0].begin + self.bucket_duration;
|
||||
|
||||
while now >= end {
|
||||
begin += self.bucket_duration;
|
||||
end += self.bucket_duration;
|
||||
}
|
||||
|
||||
self.buckets.insert(0, Bucket { begin, summary });
|
||||
}
|
||||
}
|
||||
|
||||
fn snapshot(&self, now: Instant) -> Summary {
|
||||
let cutoff = now.checked_sub(self.expire_after);
|
||||
let mut acc = Summary::with_defaults();
|
||||
|
||||
let summaries = self
|
||||
.buckets
|
||||
.iter()
|
||||
.filter(|b| cutoff.map(|c| b.begin > c).unwrap_or(true))
|
||||
.map(|b| &b.summary);
|
||||
|
||||
for item in summaries {
|
||||
acc.merge(item)
|
||||
.expect("All summaries are created with default settings");
|
||||
}
|
||||
|
||||
acc
|
||||
}
|
||||
}
|
||||
*/
|
||||
|
|
|
@ -11,7 +11,7 @@ use actix_web::{
|
|||
use bcrypt::{BcryptError, DEFAULT_COST};
|
||||
use futures_util::future::LocalBoxFuture;
|
||||
use http_signature_normalization_actix::prelude::InvalidHeaderValue;
|
||||
use std::{convert::Infallible, str::FromStr};
|
||||
use std::{convert::Infallible, str::FromStr, time::Instant};
|
||||
use tracing_error::SpanTrace;
|
||||
|
||||
use crate::db::Db;
|
||||
|
@ -178,10 +178,15 @@ impl FromRequest for Admin {
|
|||
type Future = LocalBoxFuture<'static, Result<Self, Self::Error>>;
|
||||
|
||||
fn from_request(req: &HttpRequest, _: &mut Payload) -> Self::Future {
|
||||
let now = Instant::now();
|
||||
let res = Self::prepare_verify(req);
|
||||
Box::pin(async move {
|
||||
let (db, c, t) = res?;
|
||||
Self::verify(c, t).await?;
|
||||
metrics::histogram!(
|
||||
"relay.admin.verify",
|
||||
now.elapsed().as_micros() as f64 / 1_000_000_f64
|
||||
);
|
||||
Ok(Admin { db })
|
||||
})
|
||||
}
|
||||
|
|
Loading…
Reference in a new issue