// Copyright 2018-2023 the Deno authors. All rights reserved. MIT license. use deno_core::error::AnyError; use deno_core::serde_json; use deno_core::serde_json::Value; use std::collections::HashMap; use std::convert::From; use std::env; use std::fs; use std::path::Path; use std::path::PathBuf; use std::process::Command; use std::process::Stdio; use std::time::SystemTime; mod http; mod lsp; fn read_json(filename: &str) -> Result { let f = fs::File::open(filename)?; Ok(serde_json::from_reader(f)?) } fn write_json(filename: &str, value: &Value) -> Result<()> { let f = fs::File::create(filename)?; serde_json::to_writer(f, value)?; Ok(()) } /// The list of the tuples of the benchmark name, arguments and return code const EXEC_TIME_BENCHMARKS: &[(&str, &[&str], Option)] = &[ // we need to run the cold_* benchmarks before the _warm_ ones as they ensure // the cache is properly populated, instead of other tests possibly // invalidating that cache. ( "cold_hello", &["run", "--reload", "cli/tests/testdata/run/002_hello.ts"], None, ), ( "cold_relative_import", &[ "run", "--reload", "cli/tests/testdata/run/003_relative_import.ts", ], None, ), ( "hello", &["run", "cli/tests/testdata/run/002_hello.ts"], None, ), ( "relative_import", &["run", "cli/tests/testdata/run/003_relative_import.ts"], None, ), ( "error_001", &["run", "cli/tests/testdata/run/error_001.ts"], Some(1), ), ( "no_check_hello", &[ "run", "--reload", "--no-check", "cli/tests/testdata/run/002_hello.ts", ], None, ), ( "workers_startup", &[ "run", "--allow-read", "cli/tests/testdata/workers/bench_startup.ts", ], None, ), ( "workers_round_robin", &[ "run", "--allow-read", "cli/tests/testdata/workers/bench_round_robin.ts", ], None, ), ( "workers_large_message", &[ "run", "--allow-read", "cli/tests/testdata/workers/bench_large_message.ts", ], None, ), ( "text_decoder", &["run", "cli/tests/testdata/benches/text_decoder_perf.js"], None, ), ( "text_encoder", &["run", "cli/tests/testdata/benches/text_encoder_perf.js"], None, ), ( "text_encoder_into", &[ "run", "cli/tests/testdata/benches/text_encoder_into_perf.js", ], None, ), ( "response_string", &["run", "cli/tests/testdata/benches/response_string_perf.js"], None, ), ( "check", &[ "check", "--reload", "--unstable", "test_util/std/examples/chat/server_test.ts", ], None, ), ( "no_check", &[ "cache", "--reload", "--no-check", "--unstable", "test_util/std/examples/chat/server_test.ts", ], None, ), ( "bundle", &[ "bundle", "--unstable", "test_util/std/examples/chat/server_test.ts", ], None, ), ( "bundle_no_check", &[ "bundle", "--no-check", "--unstable", "test_util/std/examples/chat/server_test.ts", ], None, ), ]; const RESULT_KEYS: &[&str] = &["mean", "stddev", "user", "system", "min", "max"]; fn run_exec_time( deno_exe: &Path, target_dir: &Path, ) -> Result>> { let hyperfine_exe = test_util::prebuilt_tool_path("hyperfine"); let benchmark_file = target_dir.join("hyperfine_results.json"); let benchmark_file = benchmark_file.to_str().unwrap(); let mut command = [ hyperfine_exe.to_str().unwrap(), "--export-json", benchmark_file, "--warmup", "3", ] .iter() .map(|s| s.to_string()) .collect::>(); for (_, args, return_code) in EXEC_TIME_BENCHMARKS { let ret_code_test = if let Some(code) = return_code { // Bash test which asserts the return code value of the previous command // $? contains the return code of the previous command format!("; test $? -eq {code}") } else { "".to_string() }; command.push(format!( "{} {} {}", deno_exe.to_str().unwrap(), args.join(" "), ret_code_test )); } test_util::run( &command.iter().map(|s| s.as_ref()).collect::>(), None, None, None, true, ); let mut results = HashMap::>::new(); let hyperfine_results = read_json(benchmark_file)?; for ((name, _, _), data) in EXEC_TIME_BENCHMARKS.iter().zip( hyperfine_results .as_object() .unwrap() .get("results") .unwrap() .as_array() .unwrap(), ) { let data = data.as_object().unwrap().clone(); results.insert( name.to_string(), data .into_iter() .filter(|(key, _)| RESULT_KEYS.contains(&key.as_str())) .map(|(key, val)| (key, val.as_f64().unwrap())) .collect(), ); } Ok(results) } fn rlib_size(target_dir: &std::path::Path, prefix: &str) -> i64 { let mut size = 0; let mut seen = std::collections::HashSet::new(); for entry in std::fs::read_dir(target_dir.join("deps")).unwrap() { let entry = entry.unwrap(); let os_str = entry.file_name(); let name = os_str.to_str().unwrap(); if name.starts_with(prefix) && name.ends_with(".rlib") { let start = name.split('-').next().unwrap().to_string(); if seen.contains(&start) { println!("skip {name}"); } else { seen.insert(start); size += entry.metadata().unwrap().len(); println!("check size {name} {size}"); } } } assert!(size > 0); size as i64 } const BINARY_TARGET_FILES: &[&str] = &["CLI_SNAPSHOT.bin", "COMPILER_SNAPSHOT.bin"]; fn get_binary_sizes(target_dir: &Path) -> Result> { let mut sizes = HashMap::::new(); let mut mtimes = HashMap::::new(); sizes.insert( "deno".to_string(), test_util::deno_exe_path().metadata()?.len() as i64, ); // add up size for everything in target/release/deps/libswc* let swc_size = rlib_size(target_dir, "libswc"); println!("swc {swc_size} bytes"); sizes.insert("swc_rlib".to_string(), swc_size); let v8_size = rlib_size(target_dir, "libv8"); println!("v8 {v8_size} bytes"); sizes.insert("rusty_v8_rlib".to_string(), v8_size); // Because cargo's OUT_DIR is not predictable, search the build tree for // snapshot related files. for file in walkdir::WalkDir::new(target_dir) { let file = match file { Ok(file) => file, Err(_) => continue, }; let filename = file.file_name().to_str().unwrap().to_string(); if !BINARY_TARGET_FILES.contains(&filename.as_str()) { continue; } let meta = file.metadata()?; let file_mtime = meta.modified()?; // If multiple copies of a file are found, use the most recent one. if let Some(stored_mtime) = mtimes.get(&filename) { if *stored_mtime > file_mtime { continue; } } mtimes.insert(filename.clone(), file_mtime); sizes.insert(filename, meta.len() as i64); } Ok(sizes) } const BUNDLES: &[(&str, &str)] = &[ ("file_server", "./test_util/std/http/file_server.ts"), ("gist", "./test_util/std/examples/gist.ts"), ]; fn bundle_benchmark(deno_exe: &Path) -> Result> { let mut sizes = HashMap::::new(); for (name, url) in BUNDLES { let path = format!("{name}.bundle.js"); test_util::run( &[ deno_exe.to_str().unwrap(), "bundle", "--unstable", url, &path, ], None, None, None, true, ); let file = PathBuf::from(path); assert!(file.is_file()); sizes.insert(name.to_string(), file.metadata()?.len() as i64); let _ = fs::remove_file(file); } Ok(sizes) } fn run_max_mem_benchmark(deno_exe: &Path) -> Result> { let mut results = HashMap::::new(); for (name, args, return_code) in EXEC_TIME_BENCHMARKS { let proc = Command::new("time") .args(["-v", deno_exe.to_str().unwrap()]) .args(args.iter()) .stdout(Stdio::null()) .stderr(Stdio::piped()) .spawn()?; let proc_result = proc.wait_with_output()?; if let Some(code) = return_code { assert_eq!(proc_result.status.code().unwrap(), *code); } let out = String::from_utf8(proc_result.stderr)?; results.insert( name.to_string(), test_util::parse_max_mem(&out).unwrap() as i64, ); } Ok(results) } fn cargo_deps() -> usize { let cargo_lock = test_util::root_path().join("Cargo.lock"); let mut count = 0; let file = std::fs::File::open(cargo_lock).unwrap(); use std::io::BufRead; for line in std::io::BufReader::new(file).lines() { if line.unwrap().starts_with("[[package]]") { count += 1 } } println!("cargo_deps {count}"); assert!(count > 10); // Sanity check. count } // TODO(@littledivy): Remove this, denoland/benchmark_data is deprecated. #[derive(Default, serde::Serialize)] struct BenchResult { created_at: String, sha1: String, // TODO(ry) The "benchmark" benchmark should actually be called "exec_time". // When this is changed, the historical data in gh-pages branch needs to be // changed too. benchmark: HashMap>, binary_size: HashMap, bundle_size: HashMap, cargo_deps: usize, max_latency: HashMap, max_memory: HashMap, lsp_exec_time: HashMap, req_per_sec: HashMap, syscall_count: HashMap, thread_count: HashMap, } #[tokio::main] async fn main() -> Result<()> { let mut args = env::args(); let mut benchmarks = vec![ "bundle", "exec_time", "binary_size", "cargo_deps", "lsp", "http", "strace", "mem_usage", ]; let mut found_bench = false; let filter = args.nth(1); if let Some(filter) = filter { if filter != "--bench" { benchmarks.retain(|s| s == &filter); } else { found_bench = true; } } if !found_bench && !args.any(|s| s == "--bench") { return Ok(()); } println!("Starting Deno benchmark"); let target_dir = test_util::target_dir(); let deno_exe = test_util::deno_exe_path(); env::set_current_dir(test_util::root_path())?; let mut new_data = BenchResult { created_at: chrono::Utc::now() .to_rfc3339_opts(chrono::SecondsFormat::Secs, true), sha1: test_util::run_collect( &["git", "rev-parse", "HEAD"], None, None, None, true, ) .0 .trim() .to_string(), ..Default::default() }; if benchmarks.contains(&"bundle") { let bundle_size = bundle_benchmark(&deno_exe)?; new_data.bundle_size = bundle_size; } if benchmarks.contains(&"exec_time") { let exec_times = run_exec_time(&deno_exe, &target_dir)?; new_data.benchmark = exec_times; } if benchmarks.contains(&"binary_size") { let binary_sizes = get_binary_sizes(&target_dir)?; new_data.binary_size = binary_sizes; } if benchmarks.contains(&"cargo_deps") { let cargo_deps = cargo_deps(); new_data.cargo_deps = cargo_deps; } if benchmarks.contains(&"lsp") { let lsp_exec_times = lsp::benchmarks(&deno_exe); new_data.lsp_exec_time = lsp_exec_times; } if benchmarks.contains(&"http") && cfg!(not(target_os = "windows")) { let stats = http::benchmark(&target_dir)?; let req_per_sec = stats .iter() .map(|(name, result)| (name.clone(), result.requests as i64)) .collect(); new_data.req_per_sec = req_per_sec; let max_latency = stats .iter() .map(|(name, result)| (name.clone(), result.latency)) .collect(); new_data.max_latency = max_latency; } if cfg!(target_os = "linux") && benchmarks.contains(&"strace") { use std::io::Read; let mut thread_count = HashMap::::new(); let mut syscall_count = HashMap::::new(); for (name, args, expected_exit_code) in EXEC_TIME_BENCHMARKS { let mut file = secure_tempfile::NamedTempFile::new()?; let exit_status = Command::new("strace") .args([ "-c", "-f", "-o", file.path().to_str().unwrap(), deno_exe.to_str().unwrap(), ]) .args(args.iter()) .stdout(Stdio::null()) .env("LC_NUMERIC", "C") .spawn()? .wait()?; let expected_exit_code = expected_exit_code.unwrap_or(0); assert_eq!(exit_status.code(), Some(expected_exit_code)); let mut output = String::new(); file.as_file_mut().read_to_string(&mut output)?; let strace_result = test_util::parse_strace_output(&output); let clone = strace_result.get("clone").map(|d| d.calls).unwrap_or(0) + 1; let total = strace_result.get("total").unwrap().calls; thread_count.insert(name.to_string(), clone as i64); syscall_count.insert(name.to_string(), total as i64); } new_data.thread_count = thread_count; new_data.syscall_count = syscall_count; } if benchmarks.contains(&"mem_usage") { let max_memory = run_max_mem_benchmark(&deno_exe)?; new_data.max_memory = max_memory; } if let Some(filename) = target_dir.join("bench.json").to_str() { write_json(filename, &serde_json::to_value(&new_data)?)?; } else { eprintln!("Cannot write bench.json, path is invalid"); } Ok(()) } pub type Result = std::result::Result;