// Copyright 2018-2022 the Deno authors. All rights reserved. MIT license. use deno_core::error::AnyError; use deno_core::serde_json; use deno_core::serde_json::Value; use serde::Serialize; use std::collections::HashMap; use std::convert::From; use std::env; use std::fs; use std::path::Path; use std::path::PathBuf; use std::process::Command; use std::process::Stdio; use std::time::SystemTime; mod http; mod lsp; fn read_json(filename: &str) -> Result { let f = fs::File::open(filename)?; Ok(serde_json::from_reader(f)?) } fn write_json(filename: &str, value: &Value) -> Result<()> { let f = fs::File::create(filename)?; serde_json::to_writer(f, value)?; Ok(()) } /// The list of the tuples of the benchmark name, arguments and return code const EXEC_TIME_BENCHMARKS: &[(&str, &[&str], Option)] = &[ // we need to run the cold_* benchmarks before the _warm_ ones as they ensure // the cache is properly populated, instead of other tests possibly // invalidating that cache. ( "cold_hello", &["run", "--reload", "cli/tests/testdata/002_hello.ts"], None, ), ( "cold_relative_import", &[ "run", "--reload", "cli/tests/testdata/003_relative_import.ts", ], None, ), ("hello", &["run", "cli/tests/testdata/002_hello.ts"], None), ( "relative_import", &["run", "cli/tests/testdata/003_relative_import.ts"], None, ), ( "error_001", &["run", "cli/tests/testdata/error_001.ts"], Some(1), ), ( "no_check_hello", &[ "run", "--reload", "--no-check", "cli/tests/testdata/002_hello.ts", ], None, ), ( "workers_startup", &[ "run", "--allow-read", "cli/tests/testdata/workers/bench_startup.ts", ], None, ), ( "workers_round_robin", &[ "run", "--allow-read", "cli/tests/testdata/workers/bench_round_robin.ts", ], None, ), ( "workers_large_message", &[ "run", "--allow-read", "cli/tests/testdata/workers/bench_large_message.ts", ], None, ), ( "text_decoder", &["run", "cli/tests/testdata/text_decoder_perf.js"], None, ), ( "text_encoder", &["run", "cli/tests/testdata/text_encoder_perf.js"], None, ), ( "text_encoder_into", &["run", "cli/tests/testdata/text_encoder_into_perf.js"], None, ), ( "response_string", &["run", "cli/tests/testdata/response_string_perf.js"], None, ), ( "check", &[ "cache", "--reload", "test_util/std/examples/chat/server_test.ts", ], None, ), ( "no_check", &[ "cache", "--reload", "--no-check", "test_util/std/examples/chat/server_test.ts", ], None, ), ( "bundle", &["bundle", "test_util/std/examples/chat/server_test.ts"], None, ), ( "bundle_no_check", &[ "bundle", "--no-check", "test_util/std/examples/chat/server_test.ts", ], None, ), ]; const RESULT_KEYS: &[&str] = &["mean", "stddev", "user", "system", "min", "max"]; fn run_exec_time( deno_exe: &Path, target_dir: &Path, ) -> Result>> { let hyperfine_exe = test_util::prebuilt_tool_path("hyperfine"); let benchmark_file = target_dir.join("hyperfine_results.json"); let benchmark_file = benchmark_file.to_str().unwrap(); let mut command = [ hyperfine_exe.to_str().unwrap(), "--export-json", benchmark_file, "--warmup", "3", ] .iter() .map(|s| s.to_string()) .collect::>(); for (_, args, return_code) in EXEC_TIME_BENCHMARKS { let ret_code_test = if let Some(code) = return_code { // Bash test which asserts the return code value of the previous command // $? contains the return code of the previous command format!("; test $? -eq {}", code) } else { "".to_string() }; command.push(format!( "{} {} {}", deno_exe.to_str().unwrap(), args.join(" "), ret_code_test )); } test_util::run( &command.iter().map(|s| s.as_ref()).collect::>(), None, None, None, true, ); let mut results = HashMap::>::new(); let hyperfine_results = read_json(benchmark_file)?; for ((name, _, _), data) in EXEC_TIME_BENCHMARKS.iter().zip( hyperfine_results .as_object() .unwrap() .get("results") .unwrap() .as_array() .unwrap(), ) { let data = data.as_object().unwrap().clone(); results.insert( name.to_string(), data .into_iter() .filter(|(key, _)| RESULT_KEYS.contains(&key.as_str())) .map(|(key, val)| (key, val.as_f64().unwrap())) .collect(), ); } Ok(results) } fn rlib_size(target_dir: &std::path::Path, prefix: &str) -> u64 { let mut size = 0; let mut seen = std::collections::HashSet::new(); for entry in std::fs::read_dir(target_dir.join("deps")).unwrap() { let entry = entry.unwrap(); let os_str = entry.file_name(); let name = os_str.to_str().unwrap(); if name.starts_with(prefix) && name.ends_with(".rlib") { let start = name.split('-').next().unwrap().to_string(); if seen.contains(&start) { println!("skip {}", name); } else { seen.insert(start); size += entry.metadata().unwrap().len(); println!("check size {} {}", name, size); } } } assert!(size > 0); size } const BINARY_TARGET_FILES: &[&str] = &["CLI_SNAPSHOT.bin", "COMPILER_SNAPSHOT.bin"]; fn get_binary_sizes(target_dir: &Path) -> Result> { let mut sizes = HashMap::::new(); let mut mtimes = HashMap::::new(); sizes.insert( "deno".to_string(), test_util::deno_exe_path().metadata()?.len(), ); // add up size for everything in target/release/deps/libswc* let swc_size = rlib_size(target_dir, "libswc"); println!("swc {} bytes", swc_size); sizes.insert("swc_rlib".to_string(), swc_size); let v8_size = rlib_size(target_dir, "libv8"); println!("v8 {} bytes", v8_size); sizes.insert("rusty_v8_rlib".to_string(), v8_size); // Because cargo's OUT_DIR is not predictable, search the build tree for // snapshot related files. for file in walkdir::WalkDir::new(target_dir) { let file = match file { Ok(file) => file, Err(_) => continue, }; let filename = file.file_name().to_str().unwrap().to_string(); if !BINARY_TARGET_FILES.contains(&filename.as_str()) { continue; } let meta = file.metadata()?; let file_mtime = meta.modified()?; // If multiple copies of a file are found, use the most recent one. if let Some(stored_mtime) = mtimes.get(&filename) { if *stored_mtime > file_mtime { continue; } } mtimes.insert(filename.clone(), file_mtime); sizes.insert(filename, meta.len()); } Ok(sizes) } const BUNDLES: &[(&str, &str)] = &[ ("file_server", "./test_util/std/http/file_server.ts"), ("gist", "./test_util/std/examples/gist.ts"), ]; fn bundle_benchmark(deno_exe: &Path) -> Result> { let mut sizes = HashMap::::new(); for (name, url) in BUNDLES { let path = format!("{}.bundle.js", name); test_util::run( &[ deno_exe.to_str().unwrap(), "bundle", "--unstable", url, &path, ], None, None, None, true, ); let file = PathBuf::from(path); assert!(file.is_file()); sizes.insert(name.to_string(), file.metadata()?.len()); let _ = fs::remove_file(file); } Ok(sizes) } fn run_http(target_dir: &Path, new_data: &mut BenchResult) -> Result<()> { let stats = http::benchmark(target_dir)?; new_data.req_per_sec = stats .iter() .map(|(name, result)| (name.clone(), result.requests)) .collect(); new_data.max_latency = stats .iter() .map(|(name, result)| (name.clone(), result.latency)) .collect(); Ok(()) } fn run_strace_benchmarks( deno_exe: &Path, new_data: &mut BenchResult, ) -> Result<()> { use std::io::Read; let mut thread_count = HashMap::::new(); let mut syscall_count = HashMap::::new(); for (name, args, expected_exit_code) in EXEC_TIME_BENCHMARKS { let mut file = tempfile::NamedTempFile::new()?; let exit_status = Command::new("strace") .args(&[ "-c", "-f", "-o", file.path().to_str().unwrap(), deno_exe.to_str().unwrap(), ]) .args(args.iter()) .stdout(Stdio::null()) .spawn()? .wait()?; let expected_exit_code = expected_exit_code.unwrap_or(0); assert_eq!(exit_status.code(), Some(expected_exit_code)); let mut output = String::new(); file.as_file_mut().read_to_string(&mut output)?; let strace_result = test_util::parse_strace_output(&output); let clone = strace_result.get("clone").map(|d| d.calls).unwrap_or(0) + 1; let total = strace_result.get("total").unwrap().calls; thread_count.insert(name.to_string(), clone); syscall_count.insert(name.to_string(), total); } new_data.thread_count = thread_count; new_data.syscall_count = syscall_count; Ok(()) } fn run_max_mem_benchmark(deno_exe: &Path) -> Result> { let mut results = HashMap::::new(); for (name, args, return_code) in EXEC_TIME_BENCHMARKS { let proc = Command::new("time") .args(&["-v", deno_exe.to_str().unwrap()]) .args(args.iter()) .stdout(Stdio::null()) .stderr(Stdio::piped()) .spawn()?; let proc_result = proc.wait_with_output()?; if let Some(code) = return_code { assert_eq!(proc_result.status.code().unwrap(), *code); } let out = String::from_utf8(proc_result.stderr)?; results.insert(name.to_string(), test_util::parse_max_mem(&out).unwrap()); } Ok(results) } fn cargo_deps() -> usize { let cargo_lock = test_util::root_path().join("Cargo.lock"); let mut count = 0; let file = std::fs::File::open(cargo_lock).unwrap(); use std::io::BufRead; for line in std::io::BufReader::new(file).lines() { if line.unwrap().starts_with("[[package]]") { count += 1 } } println!("cargo_deps {}", count); assert!(count > 10); // Sanity check. count } #[derive(Default, Serialize)] struct BenchResult { created_at: String, sha1: String, // TODO(ry) The "benchmark" benchmark should actually be called "exec_time". // When this is changed, the historical data in gh-pages branch needs to be // changed too. benchmark: HashMap>, binary_size: HashMap, bundle_size: HashMap, cargo_deps: usize, max_latency: HashMap, max_memory: HashMap, lsp_exec_time: HashMap, req_per_sec: HashMap, syscall_count: HashMap, thread_count: HashMap, } /* TODO(SyrupThinker) Switch to the #[bench] attribute once it is stabilized. Before that the #[test] tests won't be run because we replace the harness with our own runner here. */ fn main() -> Result<()> { if !env::args().any(|s| s == "--bench") { return Ok(()); } println!("Starting Deno benchmark"); let target_dir = test_util::target_dir(); let deno_exe = test_util::deno_exe_path(); env::set_current_dir(&test_util::root_path())?; let mut new_data = BenchResult { created_at: chrono::Utc::now() .to_rfc3339_opts(chrono::SecondsFormat::Secs, true), sha1: test_util::run_collect( &["git", "rev-parse", "HEAD"], None, None, None, true, ) .0 .trim() .to_string(), benchmark: run_exec_time(&deno_exe, &target_dir)?, binary_size: get_binary_sizes(&target_dir)?, bundle_size: bundle_benchmark(&deno_exe)?, cargo_deps: cargo_deps(), lsp_exec_time: lsp::benchmarks(&deno_exe)?, ..Default::default() }; if cfg!(not(target_os = "windows")) { run_http(&target_dir, &mut new_data)?; } if cfg!(target_os = "linux") { run_strace_benchmarks(&deno_exe, &mut new_data)?; new_data.max_memory = run_max_mem_benchmark(&deno_exe)?; } println!("===== "); serde_json::to_writer_pretty(std::io::stdout(), &new_data)?; println!("\n===== "); if let Some(filename) = target_dir.join("bench.json").to_str() { write_json(filename, &serde_json::to_value(&new_data)?)?; } else { eprintln!("Cannot write bench.json, path is invalid"); } Ok(()) } pub(crate) type Result = std::result::Result;