mirror of
https://github.com/denoland/deno.git
synced 2024-11-14 16:33:45 -05:00
228 lines
6.1 KiB
Rust
228 lines
6.1 KiB
Rust
// Copyright 2018-2022 the Deno authors. All rights reserved. MIT license.
|
|
|
|
use deno_core::parking_lot::Mutex;
|
|
use deno_core::serde::Deserialize;
|
|
use deno_core::serde::Serialize;
|
|
use deno_core::serde_json::json;
|
|
use std::cmp;
|
|
use std::collections::HashMap;
|
|
use std::collections::VecDeque;
|
|
use std::fmt;
|
|
use std::sync::Arc;
|
|
use std::time::Duration;
|
|
use std::time::Instant;
|
|
|
|
use super::logging::lsp_debug;
|
|
|
|
#[derive(Debug, Deserialize, Serialize, PartialEq, Eq)]
|
|
#[serde(rename_all = "camelCase")]
|
|
pub struct PerformanceAverage {
|
|
pub name: String,
|
|
pub count: u32,
|
|
pub average_duration: u32,
|
|
}
|
|
|
|
impl PartialOrd for PerformanceAverage {
|
|
fn partial_cmp(&self, other: &Self) -> Option<cmp::Ordering> {
|
|
Some(self.cmp(other))
|
|
}
|
|
}
|
|
|
|
impl Ord for PerformanceAverage {
|
|
fn cmp(&self, other: &Self) -> cmp::Ordering {
|
|
self.name.cmp(&other.name)
|
|
}
|
|
}
|
|
|
|
/// A structure which serves as a start of a measurement span.
|
|
#[derive(Debug)]
|
|
pub struct PerformanceMark {
|
|
name: String,
|
|
count: u32,
|
|
start: Instant,
|
|
}
|
|
|
|
/// A structure which holds the information about the measured span.
|
|
#[derive(Debug, Clone)]
|
|
pub struct PerformanceMeasure {
|
|
pub name: String,
|
|
pub count: u32,
|
|
pub duration: Duration,
|
|
}
|
|
|
|
impl fmt::Display for PerformanceMeasure {
|
|
fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result {
|
|
write!(f, "{} ({}ms)", self.name, self.duration.as_millis())
|
|
}
|
|
}
|
|
|
|
impl From<PerformanceMark> for PerformanceMeasure {
|
|
fn from(value: PerformanceMark) -> Self {
|
|
Self {
|
|
name: value.name,
|
|
count: value.count,
|
|
duration: value.start.elapsed(),
|
|
}
|
|
}
|
|
}
|
|
|
|
/// A simple structure for marking a start of something to measure the duration
|
|
/// of and measuring that duration. Each measurement is identified by a string
|
|
/// name and a counter is incremented each time a new measurement is marked.
|
|
///
|
|
/// The structure will limit the size of measurements to the most recent 1000,
|
|
/// and will roll off when that limit is reached.
|
|
#[derive(Debug)]
|
|
pub struct Performance {
|
|
counts: Mutex<HashMap<String, u32>>,
|
|
max_size: usize,
|
|
measures: Mutex<VecDeque<PerformanceMeasure>>,
|
|
}
|
|
|
|
impl Default for Performance {
|
|
fn default() -> Self {
|
|
Self {
|
|
counts: Default::default(),
|
|
max_size: 1_000,
|
|
measures: Default::default(),
|
|
}
|
|
}
|
|
}
|
|
|
|
impl Performance {
|
|
/// Return the count and average duration of a measurement identified by name.
|
|
#[cfg(test)]
|
|
pub fn average(&self, name: &str) -> Option<(usize, Duration)> {
|
|
let mut items = Vec::new();
|
|
for measure in self.measures.lock().iter() {
|
|
if measure.name == name {
|
|
items.push(measure.duration);
|
|
}
|
|
}
|
|
let len = items.len();
|
|
|
|
if len > 0 {
|
|
let average = items.into_iter().sum::<Duration>() / len as u32;
|
|
Some((len, average))
|
|
} else {
|
|
None
|
|
}
|
|
}
|
|
|
|
/// Return an iterator which provides the names, count, and average duration
|
|
/// of each measurement.
|
|
pub fn averages(&self) -> Vec<PerformanceAverage> {
|
|
let mut averages: HashMap<String, Vec<Duration>> = HashMap::new();
|
|
for measure in self.measures.lock().iter() {
|
|
averages
|
|
.entry(measure.name.clone())
|
|
.or_default()
|
|
.push(measure.duration);
|
|
}
|
|
averages
|
|
.into_iter()
|
|
.map(|(k, d)| {
|
|
let a = d.clone().into_iter().sum::<Duration>() / d.len() as u32;
|
|
PerformanceAverage {
|
|
name: k,
|
|
count: d.len() as u32,
|
|
average_duration: a.as_millis() as u32,
|
|
}
|
|
})
|
|
.collect()
|
|
}
|
|
|
|
/// Marks the start of a measurement which returns a performance mark
|
|
/// structure, which is then passed to `.measure()` to finalize the duration
|
|
/// and add it to the internal buffer.
|
|
pub fn mark<S: AsRef<str>, V: Serialize>(
|
|
&self,
|
|
name: S,
|
|
maybe_args: Option<V>,
|
|
) -> PerformanceMark {
|
|
let name = name.as_ref();
|
|
let mut counts = self.counts.lock();
|
|
let count = counts.entry(name.to_string()).or_insert(0);
|
|
*count += 1;
|
|
let msg = if let Some(args) = maybe_args {
|
|
json!({
|
|
"type": "mark",
|
|
"name": name,
|
|
"count": count,
|
|
"args": args,
|
|
})
|
|
} else {
|
|
json!({
|
|
"type": "mark",
|
|
"name": name,
|
|
})
|
|
};
|
|
lsp_debug!("{},", msg);
|
|
PerformanceMark {
|
|
name: name.to_string(),
|
|
count: *count,
|
|
start: Instant::now(),
|
|
}
|
|
}
|
|
|
|
/// A function which accepts a previously created performance mark which will
|
|
/// be used to finalize the duration of the span being measured, and add the
|
|
/// measurement to the internal buffer.
|
|
pub fn measure(&self, mark: PerformanceMark) -> Duration {
|
|
let measure = PerformanceMeasure::from(mark);
|
|
lsp_debug!(
|
|
"{},",
|
|
json!({
|
|
"type": "measure",
|
|
"name": measure.name,
|
|
"count": measure.count,
|
|
"duration": measure.duration.as_millis() as u32,
|
|
})
|
|
);
|
|
let duration = measure.duration;
|
|
let mut measures = self.measures.lock();
|
|
measures.push_front(measure);
|
|
while measures.len() > self.max_size {
|
|
measures.pop_back();
|
|
}
|
|
duration
|
|
}
|
|
|
|
pub fn to_vec(&self) -> Vec<PerformanceMeasure> {
|
|
let measures = self.measures.lock();
|
|
measures.iter().cloned().collect()
|
|
}
|
|
}
|
|
|
|
#[cfg(test)]
|
|
mod tests {
|
|
use super::*;
|
|
|
|
#[test]
|
|
fn test_average() {
|
|
let performance = Performance::default();
|
|
let mark1 = performance.mark("a", None::<()>);
|
|
let mark2 = performance.mark("a", None::<()>);
|
|
let mark3 = performance.mark("b", None::<()>);
|
|
performance.measure(mark2);
|
|
performance.measure(mark1);
|
|
performance.measure(mark3);
|
|
let (count, _) = performance.average("a").expect("should have had value");
|
|
assert_eq!(count, 2);
|
|
let (count, _) = performance.average("b").expect("should have had value");
|
|
assert_eq!(count, 1);
|
|
assert!(performance.average("c").is_none());
|
|
}
|
|
|
|
#[test]
|
|
fn test_averages() {
|
|
let performance = Performance::default();
|
|
let mark1 = performance.mark("a", None::<()>);
|
|
let mark2 = performance.mark("a", None::<()>);
|
|
performance.measure(mark2);
|
|
performance.measure(mark1);
|
|
let averages = performance.averages();
|
|
assert_eq!(averages.len(), 1);
|
|
assert_eq!(averages[0].count, 2);
|
|
}
|
|
}
|