// Copyright 2018-2020 the Deno authors. All rights reserved. MIT license. use crate::ast::parse; use crate::ast::Location; use crate::diagnostics::Diagnostics; use crate::disk_cache::DiskCache; use crate::file_fetcher::SourceFile; use crate::file_fetcher::SourceFileFetcher; use crate::flags::Flags; use crate::js; use crate::media_type::MediaType; use crate::module_graph::ModuleGraph; use crate::module_graph::ModuleGraphLoader; use crate::permissions::Permissions; use crate::program_state::ProgramState; use crate::tsc_config; use crate::version; use deno_core::error::generic_error; use deno_core::error::AnyError; use deno_core::error::JsError; use deno_core::json_op_sync; use deno_core::serde_json; use deno_core::serde_json::json; use deno_core::serde_json::Value; use deno_core::url::Url; use deno_core::JsRuntime; use deno_core::ModuleSpecifier; use deno_core::RuntimeOptions; use log::debug; use log::Level; use regex::Regex; use serde::Deserialize; use serde::Serialize; use serde::Serializer; use sourcemap::SourceMap; use std::collections::HashMap; use std::collections::HashSet; use std::fs; use std::io; use std::ops::Deref; use std::path::PathBuf; use std::str; use std::sync::Arc; use std::sync::Mutex; use swc_common::comments::Comment; use swc_common::comments::CommentKind; use swc_ecmascript::dep_graph; pub const AVAILABLE_LIBS: &[&str] = &[ "deno.ns", "deno.window", "deno.worker", "deno.shared_globals", "deno.unstable", "dom", "dom.iterable", "es5", "es6", "esnext", "es2020", "es2020.full", "es2019", "es2019.full", "es2018", "es2018.full", "es2017", "es2017.full", "es2016", "es2016.full", "es2015", "es2015.collection", "es2015.core", "es2015.generator", "es2015.iterable", "es2015.promise", "es2015.proxy", "es2015.reflect", "es2015.symbol", "es2015.symbol.wellknown", "es2016.array.include", "es2017.intl", "es2017.object", "es2017.sharedmemory", "es2017.string", "es2017.typedarrays", "es2018.asyncgenerator", "es2018.asynciterable", "es2018.intl", "es2018.promise", "es2018.regexp", "es2019.array", "es2019.object", "es2019.string", "es2019.symbol", "es2020.bigint", "es2020.promise", "es2020.string", "es2020.symbol.wellknown", "esnext.array", "esnext.asynciterable", "esnext.bigint", "esnext.intl", "esnext.promise", "esnext.string", "esnext.symbol", "esnext.weakref", "scripthost", "webworker", "webworker.importscripts", ]; #[derive(Debug, Clone)] pub struct CompiledModule { pub code: String, pub name: String, } lazy_static! { /// Matches the `@deno-types` pragma. static ref DENO_TYPES_RE: Regex = Regex::new(r#"(?i)^\s*@deno-types\s*=\s*(?:["']([^"']+)["']|(\S+))"#) .unwrap(); /// Matches a `/// ` comment reference. static ref TRIPLE_SLASH_REFERENCE_RE: Regex = Regex::new(r"(?i)^/\s*").unwrap(); /// Matches a path reference, which adds a dependency to a module static ref PATH_REFERENCE_RE: Regex = Regex::new(r#"(?i)\spath\s*=\s*["']([^"']*)["']"#).unwrap(); /// Matches a types reference, which for JavaScript files indicates the /// location of types to use when type checking a program that includes it as /// a dependency. static ref TYPES_REFERENCE_RE: Regex = Regex::new(r#"(?i)\stypes\s*=\s*["']([^"']*)["']"#).unwrap(); /// Matches a lib reference. static ref LIB_REFERENCE_RE: Regex = Regex::new(r#"(?i)\slib\s*=\s*["']([^"']*)["']"#).unwrap(); } fn warn_ignored_options( maybe_ignored_options: Option, ) { if let Some(ignored_options) = maybe_ignored_options { eprintln!("{}", ignored_options); } } #[derive(Clone)] pub enum TargetLib { Main, Worker, } /// Struct which represents the state of the compiler /// configuration where the first is canonical name for the configuration file, /// second is a vector of the bytes of the contents of the configuration file, /// third is bytes of the hash of contents. #[derive(Clone)] pub struct CompilerConfig { pub path: Option, pub options: Value, pub maybe_ignored_options: Option, pub hash: String, pub compile_js: bool, } impl CompilerConfig { /// Take the passed flag and resolve the file name relative to the cwd. pub fn load(maybe_config_path: Option) -> Result { if maybe_config_path.is_none() { return Ok(Self { path: Some(PathBuf::new()), options: json!({}), maybe_ignored_options: None, hash: "".to_string(), compile_js: false, }); } let raw_config_path = maybe_config_path.unwrap(); debug!("Compiler config file: {}", raw_config_path); let cwd = std::env::current_dir().unwrap(); let config_file = cwd.join(raw_config_path); // Convert the PathBuf to a canonicalized string. This is needed by the // compiler to properly deal with the configuration. let config_path = config_file.canonicalize().map_err(|_| { io::Error::new( io::ErrorKind::InvalidInput, format!( "Could not find the config file: {}", config_file.to_string_lossy() ), ) })?; // Load the contents of the configuration file debug!("Attempt to load config: {}", config_path.to_str().unwrap()); let config_bytes = fs::read(&config_file)?; let config_hash = crate::checksum::gen(&[&config_bytes]); let config_str = String::from_utf8(config_bytes)?; let (options, maybe_ignored_options) = if config_str.is_empty() { (json!({}), None) } else { tsc_config::parse_config(&config_str, &config_path)? }; // If `checkJs` is set to true in `compilerOptions` then we're gonna be compiling // JavaScript files as well let compile_js = options["checkJs"].as_bool().unwrap_or(false); Ok(Self { path: Some(config_path), options, maybe_ignored_options, hash: config_hash, compile_js, }) } } /// Information associated with compiled file in cache. /// version_hash is used to validate versions of the file /// and could be used to remove stale file in cache. #[derive(Deserialize, Serialize)] pub struct CompiledFileMetadata { pub version_hash: String, } impl CompiledFileMetadata { pub fn to_json_string(&self) -> Result { serde_json::to_string(self) } } /// Emit a SHA256 hash based on source code, deno version and TS config. /// Used to check if a recompilation for source code is needed. fn source_code_version_hash( source_code: &[u8], version: &str, config_hash: &[u8], ) -> String { crate::checksum::gen(&[source_code, version.as_bytes(), config_hash]) } fn maybe_log_stats(maybe_stats: Option>) { if let Some(stats) = maybe_stats { debug!("DEBUG - Compilation Statistics:"); for stat in stats { debug!("{}: {}", stat.key, stat.value); } } } pub struct TsCompilerInner { pub file_fetcher: SourceFileFetcher, pub flags: Flags, pub config: CompilerConfig, pub disk_cache: DiskCache, /// Set of all URLs that have been compiled. This prevents double /// compilation of module. pub compiled: Mutex>, /// This setting is controlled by `--reload` flag. Unless the flag /// is provided disk cache is used. pub use_disk_cache: bool, /// This setting is controlled by `compilerOptions.checkJs` pub compile_js: bool, } #[derive(Clone)] pub struct TsCompiler(Arc); impl Deref for TsCompiler { type Target = TsCompilerInner; fn deref(&self) -> &Self::Target { &self.0 } } #[derive(Deserialize, Debug)] #[serde(rename_all = "camelCase")] struct Stat { key: String, value: f64, } #[derive(Deserialize)] #[serde(rename_all = "camelCase")] struct EmittedSource { filename: String, contents: String, } #[derive(Deserialize)] #[serde(rename_all = "camelCase")] struct BundleResponse { diagnostics: Diagnostics, bundle_output: Option, stats: Option>, } // TODO(bartlomieju): possible deduplicate once TS refactor is stabilized #[derive(Deserialize)] #[serde(rename_all = "camelCase")] #[allow(unused)] struct RuntimeBundleResponse { diagnostics: Diagnostics, output: String, } #[derive(Deserialize)] #[serde(rename_all = "camelCase")] struct RuntimeCompileResponse { diagnostics: Diagnostics, emit_map: HashMap, } impl TsCompiler { pub fn new( file_fetcher: SourceFileFetcher, flags: Flags, disk_cache: DiskCache, ) -> Result { let config = CompilerConfig::load(flags.config_path.clone())?; let use_disk_cache = !flags.reload; Ok(TsCompiler(Arc::new(TsCompilerInner { file_fetcher, flags, disk_cache, compile_js: config.compile_js, config, compiled: Mutex::new(HashSet::new()), use_disk_cache, }))) } /// Mark given module URL as compiled to avoid multiple compilations of same /// module in single run. fn mark_compiled(&self, url: &Url) { let mut c = self.compiled.lock().unwrap(); c.insert(url.clone()); } /// For a given module, generate a single file JavaScript output that includes /// all the dependencies for that module. pub async fn bundle( &self, program_state: &Arc, module_specifier: ModuleSpecifier, ) -> Result { debug!( "Invoking the compiler to bundle. module_name: {}", module_specifier.to_string() ); let permissions = Permissions::allow_all(); let mut module_graph_loader = ModuleGraphLoader::new( self.file_fetcher.clone(), program_state.maybe_import_map.clone(), permissions.clone(), false, true, ); module_graph_loader .add_to_graph(&module_specifier, None) .await?; let module_graph = module_graph_loader.get_graph(); let module_graph_files = module_graph.values().collect::>(); // Check integrity of every file in module graph if let Some(ref lockfile) = program_state.lockfile { let mut g = lockfile.lock().unwrap(); for graph_file in &module_graph_files { let check_passed = g.check_or_insert(&graph_file.url, &graph_file.source_code); if !check_passed { eprintln!( "Subresource integrity check failed --lock={}\n{}", g.filename.display(), graph_file.url ); std::process::exit(10); } } } if let Some(ref lockfile) = program_state.lockfile { let g = lockfile.lock().unwrap(); g.write()?; } let module_graph_json = serde_json::to_value(module_graph).expect("Failed to serialize data"); let root_names = vec![module_specifier.to_string()]; let target = "main"; let performance = matches!(program_state.flags.log_level, Some(Level::Debug)); let unstable = self.flags.unstable; let mut lib = if target == "main" { vec!["deno.window"] } else { vec!["deno.worker"] }; if unstable { lib.push("deno.unstable"); } let mut compiler_options = json!({ "allowJs": true, "allowNonTsExtensions": true, "checkJs": false, "esModuleInterop": true, "inlineSourceMap": false, "jsx": "react", "lib": lib, "module": "system", "outFile": "deno:///bundle.js", // disabled until we have effective way to modify source maps "sourceMap": false, "strict": true, "removeComments": true, "target": "esnext", }); let compiler_config = self.config.clone(); tsc_config::json_merge(&mut compiler_options, &compiler_config.options); warn_ignored_options(compiler_config.maybe_ignored_options); let j = json!({ "type": CompilerRequestType::Bundle, "target": target, "rootNames": root_names, "performance": performance, "compilerOptions": compiler_options, "sourceFileMap": module_graph_json, }); let req_msg = j.to_string(); let json_str = execute_in_tsc(program_state.clone(), req_msg)?; let bundle_response: BundleResponse = serde_json::from_str(&json_str)?; maybe_log_stats(bundle_response.stats); if !bundle_response.diagnostics.0.is_empty() { return Err(generic_error(bundle_response.diagnostics.to_string())); } assert!(bundle_response.bundle_output.is_some()); let output = bundle_response.bundle_output.unwrap(); Ok(output) } fn cache_emitted_files( &self, emit_map: HashMap, ) -> std::io::Result<()> { for (emitted_name, source) in emit_map.iter() { let specifier = ModuleSpecifier::resolve_url(&source.filename) .expect("Should be a valid module specifier"); let source_file = self .file_fetcher .fetch_cached_source_file(&specifier, Permissions::allow_all()) .expect("Source file not found"); // NOTE: JavaScript files are only cached to disk if `checkJs` // option in on if source_file.media_type == MediaType::JavaScript && !self.compile_js { continue; } if emitted_name.ends_with(".map") { self.cache_source_map(&specifier, &source.contents)?; } else if emitted_name.ends_with(".js") { self.cache_compiled_file(&specifier, source_file, &source.contents)?; } else { panic!("Trying to cache unknown file type {}", emitted_name); } } Ok(()) } /// Save compiled JS file for given TS module to on-disk cache. /// /// Along compiled file a special metadata file is saved as well containing /// hash that can be validated to avoid unnecessary recompilation. fn cache_compiled_file( &self, module_specifier: &ModuleSpecifier, source_file: SourceFile, contents: &str, ) -> std::io::Result<()> { let js_key = self .disk_cache .get_cache_filename_with_extension(module_specifier.as_url(), "js"); self.disk_cache.set(&js_key, contents.as_bytes())?; self.mark_compiled(module_specifier.as_url()); let version_hash = source_code_version_hash( &source_file.source_code.as_bytes(), version::DENO, &self.config.hash.as_bytes(), ); let compiled_file_metadata = CompiledFileMetadata { version_hash }; let meta_key = self .disk_cache .get_cache_filename_with_extension(module_specifier.as_url(), "meta"); self.disk_cache.set( &meta_key, compiled_file_metadata.to_json_string()?.as_bytes(), ) } /// Save source map file for given TS module to on-disk cache. fn cache_source_map( &self, module_specifier: &ModuleSpecifier, contents: &str, ) -> std::io::Result<()> { let js_key = self .disk_cache .get_cache_filename_with_extension(module_specifier.as_url(), "js"); let js_path = self.disk_cache.location.join(js_key); let js_file_url = Url::from_file_path(js_path).expect("Bad file URL for file"); let source_map_key = self .disk_cache .get_cache_filename_with_extension(module_specifier.as_url(), "js.map"); let mut sm = SourceMap::from_slice(contents.as_bytes()) .expect("Invalid source map content"); sm.set_file(Some(&js_file_url.to_string())); sm.set_source(0, &module_specifier.to_string()); let mut output: Vec = vec![]; sm.to_writer(&mut output) .expect("Failed to write source map"); self.disk_cache.set(&source_map_key, &output) } } #[derive(Debug, Deserialize)] struct CreateHashArgs { data: String, } fn execute_in_tsc( program_state: Arc, req: String, ) -> Result { let mut js_runtime = JsRuntime::new(RuntimeOptions { startup_snapshot: Some(js::compiler_isolate_init()), ..Default::default() }); let debug_flag = program_state .flags .log_level .map_or(false, |l| l == log::Level::Debug); let response = Arc::new(Mutex::new(None)); { js_runtime.register_op( "op_fetch_asset", crate::op_fetch_asset::op_fetch_asset(HashMap::default()), ); let res = response.clone(); js_runtime.register_op( "op_compiler_respond", json_op_sync(move |_state, args, _bufs| { let mut response_slot = res.lock().unwrap(); let replaced_value = response_slot.replace(args.to_string()); assert!( replaced_value.is_none(), "op_compiler_respond found unexpected existing compiler output", ); Ok(json!({})) }), ); js_runtime.register_op( "op_create_hash", json_op_sync(move |_s, args, _bufs| { let v: CreateHashArgs = serde_json::from_value(args)?; let hash = crate::checksum::gen(&[v.data.as_bytes()]); Ok(json!({ "hash": hash })) }), ); } let bootstrap_script = format!( "globalThis.startup({{ debugFlag: {}, legacy: true }})", debug_flag ); js_runtime.execute("", &bootstrap_script)?; let script = format!("globalThis.tsCompilerOnMessage({{ data: {} }});", req); js_runtime.execute("", &script)?; let maybe_response = response.lock().unwrap().take(); assert!( maybe_response.is_some(), "Unexpected missing response from TS compiler" ); Ok(maybe_response.unwrap()) } async fn create_runtime_module_graph( program_state: &Arc, permissions: Permissions, root_name: &str, sources: &Option>, type_files: Vec, ) -> Result<(Vec, ModuleGraph), AnyError> { let mut root_names = vec![]; let mut module_graph_loader = ModuleGraphLoader::new( program_state.file_fetcher.clone(), None, permissions, false, false, ); if let Some(s_map) = sources { root_names.push(root_name.to_string()); module_graph_loader.build_local_graph(root_name, s_map)?; } else { let module_specifier = ModuleSpecifier::resolve_import(root_name, "")?; root_names.push(module_specifier.to_string()); module_graph_loader .add_to_graph(&module_specifier, None) .await?; } // download all additional files from TSconfig and add them to root_names for type_file in type_files { let type_specifier = ModuleSpecifier::resolve_url_or_path(&type_file)?; module_graph_loader .add_to_graph(&type_specifier, None) .await?; root_names.push(type_specifier.to_string()) } Ok((root_names, module_graph_loader.get_graph())) } fn extract_js_error(error: AnyError) -> AnyError { match error.downcast::() { Ok(js_error) => { let msg = format!("Error in TS compiler:\n{}", js_error); generic_error(msg) } Err(error) => error, } } /// This function is used by `Deno.compile()` API. pub async fn runtime_compile( program_state: &Arc, permissions: Permissions, root_name: &str, sources: &Option>, maybe_options: &Option, ) -> Result { let mut user_options = if let Some(options) = maybe_options { tsc_config::parse_raw_config(options)? } else { json!({}) }; // Intentionally calling "take()" to replace value with `null` - otherwise TSC will try to load that file // using `fileExists` API let type_files = if let Some(types) = user_options["types"].take().as_array() { types .iter() .map(|type_value| type_value.as_str().unwrap_or("").to_string()) .filter(|type_str| !type_str.is_empty()) .collect() } else { vec![] }; let unstable = program_state.flags.unstable; let mut lib = vec![]; if let Some(user_libs) = user_options["lib"].take().as_array() { let libs = user_libs .iter() .map(|type_value| type_value.as_str().unwrap_or("").to_string()) .filter(|type_str| !type_str.is_empty()) .collect::>(); lib.extend(libs); } else { lib.push("deno.window".to_string()); } if unstable { lib.push("deno.unstable".to_string()); } let mut compiler_options = json!({ "allowJs": false, "allowNonTsExtensions": true, "checkJs": false, "esModuleInterop": true, // TODO(lucacasonato): enable this by default in 1.5.0 "isolatedModules": unstable, "jsx": "react", "module": "esnext", "sourceMap": true, "strict": true, "removeComments": true, "target": "esnext", }); tsc_config::json_merge(&mut compiler_options, &user_options); tsc_config::json_merge(&mut compiler_options, &json!({ "lib": lib })); let (root_names, module_graph) = create_runtime_module_graph( &program_state, permissions.clone(), root_name, sources, type_files, ) .await?; let module_graph_json = serde_json::to_value(module_graph).expect("Failed to serialize data"); let req_msg = json!({ "type": CompilerRequestType::RuntimeCompile, "target": "runtime", "rootNames": root_names, "sourceFileMap": module_graph_json, "compilerOptions": compiler_options, }) .to_string(); let compiler = program_state.ts_compiler.clone(); let json_str = execute_in_tsc(program_state.clone(), req_msg).map_err(extract_js_error)?; let response: RuntimeCompileResponse = serde_json::from_str(&json_str)?; if response.diagnostics.0.is_empty() && sources.is_none() { compiler.cache_emitted_files(response.emit_map)?; } // We're returning `Ok()` instead of `Err()` because it's not runtime // error if there were diagnostics produced; we want to let user handle // diagnostics in the runtime. Ok(serde_json::from_str::(&json_str).unwrap()) } /// This function is used by `Deno.bundle()` API. pub async fn runtime_bundle( program_state: &Arc, permissions: Permissions, root_name: &str, sources: &Option>, maybe_options: &Option, ) -> Result { let mut user_options = if let Some(options) = maybe_options { tsc_config::parse_raw_config(options)? } else { json!({}) }; // Intentionally calling "take()" to replace value with `null` - otherwise TSC will try to load that file // using `fileExists` API let type_files = if let Some(types) = user_options["types"].take().as_array() { types .iter() .map(|type_value| type_value.as_str().unwrap_or("").to_string()) .filter(|type_str| !type_str.is_empty()) .collect() } else { vec![] }; let (root_names, module_graph) = create_runtime_module_graph( &program_state, permissions.clone(), root_name, sources, type_files, ) .await?; let module_graph_json = serde_json::to_value(module_graph).expect("Failed to serialize data"); let unstable = program_state.flags.unstable; let mut lib = vec![]; if let Some(user_libs) = user_options["lib"].take().as_array() { let libs = user_libs .iter() .map(|type_value| type_value.as_str().unwrap_or("").to_string()) .filter(|type_str| !type_str.is_empty()) .collect::>(); lib.extend(libs); } else { lib.push("deno.window".to_string()); } if unstable { lib.push("deno.unstable".to_string()); } let mut compiler_options = json!({ "allowJs": false, "allowNonTsExtensions": true, "checkJs": false, "esModuleInterop": true, "jsx": "react", "module": "esnext", "outDir": null, "sourceMap": true, "strict": true, "removeComments": true, "target": "esnext", }); let bundler_options = json!({ "allowJs": true, "inlineSourceMap": false, "module": "system", "outDir": null, "outFile": "deno:///bundle.js", // disabled until we have effective way to modify source maps "sourceMap": false, }); tsc_config::json_merge(&mut compiler_options, &user_options); tsc_config::json_merge(&mut compiler_options, &json!({ "lib": lib })); tsc_config::json_merge(&mut compiler_options, &bundler_options); let req_msg = json!({ "type": CompilerRequestType::RuntimeBundle, "target": "runtime", "rootNames": root_names, "sourceFileMap": module_graph_json, "compilerOptions": compiler_options, }) .to_string(); let json_str = execute_in_tsc(program_state.clone(), req_msg).map_err(extract_js_error)?; let _response: RuntimeBundleResponse = serde_json::from_str(&json_str)?; // We're returning `Ok()` instead of `Err()` because it's not runtime // error if there were diagnostics produced; we want to let user handle // diagnostics in the runtime. Ok(serde_json::from_str::(&json_str).unwrap()) } /// This function is used by `Deno.transpileOnly()` API. pub async fn runtime_transpile( program_state: Arc, sources: &HashMap, maybe_options: &Option, ) -> Result { let user_options = if let Some(options) = maybe_options { tsc_config::parse_raw_config(options)? } else { json!({}) }; let mut compiler_options = json!({ "esModuleInterop": true, "module": "esnext", "sourceMap": true, "scriptComments": true, "target": "esnext", }); tsc_config::json_merge(&mut compiler_options, &user_options); let req_msg = json!({ "type": CompilerRequestType::RuntimeTranspile, "sources": sources, "compilerOptions": compiler_options, }) .to_string(); let json_str = execute_in_tsc(program_state, req_msg).map_err(extract_js_error)?; let v = serde_json::from_str::(&json_str) .expect("Error decoding JSON string."); Ok(v) } #[derive(Clone, Debug, PartialEq)] pub struct ImportDesc { pub specifier: String, pub deno_types: Option, pub location: Location, } #[derive(Clone, Debug, PartialEq)] pub enum TsReferenceKind { Lib, Types, Path, } #[derive(Clone, Debug, PartialEq)] pub struct TsReferenceDesc { pub kind: TsReferenceKind, pub specifier: String, pub location: Location, } // TODO(bartlomieju): handle imports in ambient contexts/TS modules /// This function is a port of `ts.preProcessFile()` /// /// Additionally it captures `@deno-types` references directly /// preceeding `import .. from` and `export .. from` statements. pub fn pre_process_file( file_name: &str, media_type: MediaType, source_code: &str, analyze_dynamic_imports: bool, ) -> Result<(Vec, Vec), AnyError> { let specifier = ModuleSpecifier::resolve_url_or_path(file_name)?; let module = parse(&specifier, source_code, &media_type)?; let dependency_descriptors = module.analyze_dependencies(); // for each import check if there's relevant @deno-types directive let imports = dependency_descriptors .iter() .filter(|desc| desc.kind != dep_graph::DependencyKind::Require) .filter(|desc| { if analyze_dynamic_imports { return true; } !desc.is_dynamic }) .map(|desc| { let deno_types = get_deno_types(&desc.leading_comments); ImportDesc { specifier: desc.specifier.to_string(), deno_types, location: Location { filename: file_name.to_string(), col: desc.col, line: desc.line, }, } }) .collect(); // analyze comment from beginning of the file and find TS directives let comments = module.get_leading_comments(); let mut references = vec![]; for comment in comments { if comment.kind != CommentKind::Line { continue; } let text = comment.text.to_string(); if let Some((kind, specifier)) = parse_ts_reference(text.trim()) { let location = module.get_location(&comment.span); references.push(TsReferenceDesc { kind, specifier, location, }); } } Ok((imports, references)) } fn get_deno_types(comments: &[Comment]) -> Option { if comments.is_empty() { return None; } // @deno-types must directly prepend import statement - hence // checking last comment for span let last = comments.last().unwrap(); let comment = last.text.trim_start(); parse_deno_types(&comment) } fn parse_ts_reference(comment: &str) -> Option<(TsReferenceKind, String)> { if !TRIPLE_SLASH_REFERENCE_RE.is_match(comment) { return None; } let (kind, specifier) = if let Some(capture_groups) = PATH_REFERENCE_RE.captures(comment) { (TsReferenceKind::Path, capture_groups.get(1).unwrap()) } else if let Some(capture_groups) = TYPES_REFERENCE_RE.captures(comment) { (TsReferenceKind::Types, capture_groups.get(1).unwrap()) } else if let Some(capture_groups) = LIB_REFERENCE_RE.captures(comment) { (TsReferenceKind::Lib, capture_groups.get(1).unwrap()) } else { return None; }; Some((kind, specifier.as_str().to_string())) } fn parse_deno_types(comment: &str) -> Option { if let Some(capture_groups) = DENO_TYPES_RE.captures(comment) { if let Some(specifier) = capture_groups.get(1) { return Some(specifier.as_str().to_string()); } if let Some(specifier) = capture_groups.get(2) { return Some(specifier.as_str().to_string()); } } None } // Warning! The values in this enum are duplicated in js/compiler.ts // Update carefully! #[repr(i32)] #[derive(Clone, Copy, PartialEq, Debug)] pub enum CompilerRequestType { Bundle = 1, RuntimeCompile = 2, RuntimeBundle = 3, RuntimeTranspile = 4, } impl Serialize for CompilerRequestType { fn serialize(&self, serializer: S) -> Result where S: Serializer, { let value: i32 = match self { CompilerRequestType::Bundle => 1 as i32, CompilerRequestType::RuntimeCompile => 2 as i32, CompilerRequestType::RuntimeBundle => 3 as i32, CompilerRequestType::RuntimeTranspile => 4 as i32, }; Serialize::serialize(&value, serializer) } } #[cfg(test)] mod tests { use super::*; use crate::fs as deno_fs; use crate::program_state::ProgramState; use tempfile::TempDir; #[test] fn test_parse_deno_types() { assert_eq!( parse_deno_types("@deno-types=./a/b/c.d.ts"), Some("./a/b/c.d.ts".to_string()) ); assert_eq!( parse_deno_types("@deno-types=\"./a/b/c.d.ts\""), Some("./a/b/c.d.ts".to_string()) ); assert_eq!( parse_deno_types("@deno-types = https://dneo.land/x/some/package/a.d.ts"), Some("https://dneo.land/x/some/package/a.d.ts".to_string()) ); assert_eq!( parse_deno_types("@deno-types = ./a/b/c.d.ts"), Some("./a/b/c.d.ts".to_string()) ); assert!(parse_deno_types("asdf").is_none()); assert!(parse_deno_types("// deno-types = fooo").is_none()); assert_eq!( parse_deno_types("@deno-types=./a/b/c.d.ts some comment"), Some("./a/b/c.d.ts".to_string()) ); assert_eq!( parse_deno_types( "@deno-types=./a/b/c.d.ts // some comment after slashes" ), Some("./a/b/c.d.ts".to_string()) ); assert_eq!( parse_deno_types(r#"@deno-types="https://deno.land/x/foo/index.d.ts";"#), Some("https://deno.land/x/foo/index.d.ts".to_string()) ); } #[test] fn test_parse_ts_reference() { assert_eq!( parse_ts_reference(r#"/ "#), Some((TsReferenceKind::Lib, "deno.shared_globals".to_string())) ); assert_eq!( parse_ts_reference(r#"/ "#), Some((TsReferenceKind::Path, "./type/reference/dep.ts".to_string())) ); assert_eq!( parse_ts_reference(r#"/ "#), Some((TsReferenceKind::Types, "./type/reference.d.ts".to_string())) ); assert!(parse_ts_reference("asdf").is_none()); assert!( parse_ts_reference(r#"/ "#).is_none() ); assert!(parse_ts_reference(r#"/ "#).is_none()); } #[tokio::test] async fn test_bundle() { let p = std::path::PathBuf::from(env!("CARGO_MANIFEST_DIR")) .parent() .unwrap() .join("cli/tests/002_hello.ts"); use deno_core::ModuleSpecifier; let module_name = ModuleSpecifier::resolve_url_or_path(p.to_str().unwrap()).unwrap(); let mock_state = ProgramState::mock( vec![ String::from("deno"), p.to_string_lossy().into(), String::from("$deno$/bundle.js"), ], None, ); let result = mock_state .ts_compiler .bundle(&mock_state, module_name) .await; assert!(result.is_ok()); } #[test] fn test_source_code_version_hash() { assert_eq!( "0185b42de0686b4c93c314daaa8dee159f768a9e9a336c2a5e3d5b8ca6c4208c", source_code_version_hash(b"1+2", "0.4.0", b"{}") ); // Different source_code should result in different hash. assert_eq!( "e58631f1b6b6ce2b300b133ec2ad16a8a5ba6b7ecf812a8c06e59056638571ac", source_code_version_hash(b"1", "0.4.0", b"{}") ); // Different version should result in different hash. assert_eq!( "307e6200347a88dbbada453102deb91c12939c65494e987d2d8978f6609b5633", source_code_version_hash(b"1", "0.1.0", b"{}") ); // Different config should result in different hash. assert_eq!( "195eaf104a591d1d7f69fc169c60a41959c2b7a21373cd23a8f675f877ec385f", source_code_version_hash(b"1", "0.4.0", b"{\"compilerOptions\": {}}") ); } #[test] fn test_compile_js() { let temp_dir = TempDir::new().expect("tempdir fail"); let temp_dir_path = temp_dir.path(); let test_cases = vec![ // valid JSON (r#"{ "compilerOptions": { "checkJs": true } } "#, true), // JSON with comment ( r#"{ "compilerOptions": { // force .js file compilation by Deno "checkJs": true } }"#, true, ), // without content ("", false), ]; let path = temp_dir_path.join("tsconfig.json"); let path_str = path.to_str().unwrap().to_string(); for (json_str, expected) in test_cases { deno_fs::write_file(&path, json_str.as_bytes(), 0o666).unwrap(); let config = CompilerConfig::load(Some(path_str.clone())).unwrap(); assert_eq!(config.compile_js, expected); } } #[test] fn test_compiler_config_load() { let temp_dir = TempDir::new().expect("tempdir fail"); let temp_dir_path = temp_dir.path(); let path = temp_dir_path.join("doesnotexist.json"); let path_str = path.to_str().unwrap().to_string(); let res = CompilerConfig::load(Some(path_str)); assert!(res.is_err()); } }