1
0
Fork 0
mirror of https://github.com/denoland/deno.git synced 2024-11-21 15:04:11 -05:00

fix: better handling of npm resolution occurring on workers (#24094)

Closes https://github.com/denoland/deno/issues/24063
This commit is contained in:
David Sherret 2024-06-05 11:04:16 -04:00 committed by GitHub
parent 0544d60012
commit 7ed90a20d0
No known key found for this signature in database
GPG key ID: B5690EEEBB952194
69 changed files with 902 additions and 539 deletions

View file

@ -57,10 +57,10 @@
"ext/websocket/autobahn/reports" "ext/websocket/autobahn/reports"
], ],
"plugins": [ "plugins": [
"https://plugins.dprint.dev/typescript-0.90.5.wasm", "https://plugins.dprint.dev/typescript-0.91.1.wasm",
"https://plugins.dprint.dev/json-0.19.2.wasm", "https://plugins.dprint.dev/json-0.19.3.wasm",
"https://plugins.dprint.dev/markdown-0.17.0.wasm", "https://plugins.dprint.dev/markdown-0.17.1.wasm",
"https://plugins.dprint.dev/toml-0.6.1.wasm", "https://plugins.dprint.dev/toml-0.6.2.wasm",
"https://plugins.dprint.dev/exec-0.4.4.json@c207bf9b9a4ee1f0ecb75c594f774924baf62e8e53a2ce9d873816a408cecbf7" "https://plugins.dprint.dev/exec-0.4.4.json@c207bf9b9a4ee1f0ecb75c594f774924baf62e8e53a2ce9d873816a408cecbf7"
] ]
} }

40
Cargo.lock generated
View file

@ -1175,9 +1175,9 @@ dependencies = [
[[package]] [[package]]
name = "deno_ast" name = "deno_ast"
version = "0.38.2" version = "0.39.0"
source = "registry+https://github.com/rust-lang/crates.io-index" source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "584547d27786a734536fde7088f8429d355569c39410427be44695c300618408" checksum = "32edef567e3090862e865c75628f4d35ace80ca90e0fc5263a7d10fa307ae899"
dependencies = [ dependencies = [
"anyhow", "anyhow",
"base64 0.21.7", "base64 0.21.7",
@ -1385,9 +1385,9 @@ dependencies = [
[[package]] [[package]]
name = "deno_doc" name = "deno_doc"
version = "0.137.0" version = "0.139.0"
source = "registry+https://github.com/rust-lang/crates.io-index" source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "57f13d6254b2e05b014e8464647025fa28ef2f385c9c102744a27bd788eb3ebe" checksum = "c9cd9891748fbd9847c9aeed31635c4c1b5d9a949f6fdd80613b082bdd863518"
dependencies = [ dependencies = [
"ammonia", "ammonia",
"anyhow", "anyhow",
@ -1410,9 +1410,9 @@ dependencies = [
[[package]] [[package]]
name = "deno_emit" name = "deno_emit"
version = "0.41.0" version = "0.42.0"
source = "registry+https://github.com/rust-lang/crates.io-index" source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "ebe4b6c67f21a73901e962e92d51065f3c1bb42d2195bca8c2fef9f1808c4c2d" checksum = "25bc64f886c76647400ed8f807ba7dba82e0b52e57e5426a83094cfe22ee19c9"
dependencies = [ dependencies = [
"anyhow", "anyhow",
"base64 0.21.7", "base64 0.21.7",
@ -1478,9 +1478,9 @@ dependencies = [
[[package]] [[package]]
name = "deno_graph" name = "deno_graph"
version = "0.77.2" version = "0.78.0"
source = "registry+https://github.com/rust-lang/crates.io-index" source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "192d6f61d5418c928d29b2666b916df65a3b5677ce454fc6a4b4969983a02abe" checksum = "c4ccd2755a805983f96aeccd211c1f7585b6bfec77471f502c47227abe375682"
dependencies = [ dependencies = [
"anyhow", "anyhow",
"async-trait", "async-trait",
@ -1589,9 +1589,9 @@ dependencies = [
[[package]] [[package]]
name = "deno_lint" name = "deno_lint"
version = "0.59.1" version = "0.60.0"
source = "registry+https://github.com/rust-lang/crates.io-index" source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "0568595fd7f8ad76ddf66d70c1c5e3e680ad95c3d5abb44556e94d824643d6e2" checksum = "bf6a9540b371b123e3df4ab5fd59af0defc0d834e08ebfb3deacc41837963368"
dependencies = [ dependencies = [
"anyhow", "anyhow",
"deno_ast", "deno_ast",
@ -2269,9 +2269,9 @@ dependencies = [
[[package]] [[package]]
name = "dprint-plugin-json" name = "dprint-plugin-json"
version = "0.19.2" version = "0.19.3"
source = "registry+https://github.com/rust-lang/crates.io-index" source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "e292d0d74f3f51b1ff3e446c8809bcdd0b6079b49cf6c0d452c85927a2575246" checksum = "a19f4a9f2f548b2098b8ec597d7bb40af133b6e9a3187c1d3c4caa101b8c93c3"
dependencies = [ dependencies = [
"anyhow", "anyhow",
"dprint-core", "dprint-core",
@ -2296,9 +2296,9 @@ dependencies = [
[[package]] [[package]]
name = "dprint-plugin-markdown" name = "dprint-plugin-markdown"
version = "0.17.0" version = "0.17.1"
source = "registry+https://github.com/rust-lang/crates.io-index" source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "2b2642e4a5f3a2262bb9baef8739f90d99b73ca21bc65f46c320a7817fd65438" checksum = "da8df52eef864c2577ad3fb28c596935e2c0161eb09f6d5e239b10fecda2ec1c"
dependencies = [ dependencies = [
"anyhow", "anyhow",
"dprint-core", "dprint-core",
@ -2311,9 +2311,9 @@ dependencies = [
[[package]] [[package]]
name = "dprint-plugin-typescript" name = "dprint-plugin-typescript"
version = "0.90.5" version = "0.91.1"
source = "registry+https://github.com/rust-lang/crates.io-index" source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "d7c3c339020ebbbbbe5fc049350935ee2ea2ba5a3fc01f753588639a30404cda" checksum = "4170a1aea5c8d899e9fa96be972931b1f0beaf6f6ba2f3f40a48a13071b376ea"
dependencies = [ dependencies = [
"anyhow", "anyhow",
"deno_ast", "deno_ast",
@ -2539,9 +2539,9 @@ dependencies = [
[[package]] [[package]]
name = "eszip" name = "eszip"
version = "0.70.1" version = "0.71.0"
source = "registry+https://github.com/rust-lang/crates.io-index" source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "e5f9947a8dd5ba292461c84a5bf142497e2840c4165994c5c3b3ae4954d38fef" checksum = "3c3763e2d3e56ed5f770f9ab133aca20b1e7fa840f2408f79575ad96f942af2e"
dependencies = [ dependencies = [
"anyhow", "anyhow",
"base64 0.21.7", "base64 0.21.7",
@ -2652,9 +2652,9 @@ checksum = "c007b1ae3abe1cb6f85a16305acd418b7ca6343b953633fee2b76d8f108b830f"
[[package]] [[package]]
name = "file_test_runner" name = "file_test_runner"
version = "0.7.0" version = "0.7.2"
source = "registry+https://github.com/rust-lang/crates.io-index" source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "b8797fcdc5c6b8c06839900c30f5c59b3541ef2bec218579470ce7b1afc17ee9" checksum = "05b23dcc1b671771c6f59fdace6da685735c925f859733e8fd07fba6cae6462a"
dependencies = [ dependencies = [
"anyhow", "anyhow",
"crossbeam-channel", "crossbeam-channel",

View file

@ -43,7 +43,7 @@ license = "MIT"
repository = "https://github.com/denoland/deno" repository = "https://github.com/denoland/deno"
[workspace.dependencies] [workspace.dependencies]
deno_ast = { version = "=0.38.2", features = ["transpiling"] } deno_ast = { version = "=0.39.0", features = ["transpiling"] }
deno_core = { version = "0.284.0" } deno_core = { version = "0.284.0" }
deno_bench_util = { version = "0.148.0", path = "./bench_util" } deno_bench_util = { version = "0.148.0", path = "./bench_util" }

View file

@ -67,17 +67,17 @@ deno_ast = { workspace = true, features = ["bundler", "cjs", "codegen", "proposa
deno_cache_dir = { workspace = true } deno_cache_dir = { workspace = true }
deno_config = "=0.16.4" deno_config = "=0.16.4"
deno_core = { workspace = true, features = ["include_js_files_for_snapshotting"] } deno_core = { workspace = true, features = ["include_js_files_for_snapshotting"] }
deno_doc = { version = "=0.137.0", features = ["html", "syntect"] } deno_doc = { version = "=0.139.0", features = ["html", "syntect"] }
deno_emit = "=0.41.0" deno_emit = "=0.42.0"
deno_graph = { version = "=0.77.2", features = ["tokio_executor"] } deno_graph = { version = "=0.78.0", features = ["tokio_executor"] }
deno_lint = { version = "=0.59.1", features = ["docs"] } deno_lint = { version = "=0.60.0", features = ["docs"] }
deno_lockfile.workspace = true deno_lockfile.workspace = true
deno_npm = "=0.21.0" deno_npm = "=0.21.0"
deno_runtime = { workspace = true, features = ["include_js_files_for_snapshotting"] } deno_runtime = { workspace = true, features = ["include_js_files_for_snapshotting"] }
deno_semver = "=0.5.4" deno_semver = "=0.5.4"
deno_task_shell = "=0.16.1" deno_task_shell = "=0.16.1"
deno_terminal.workspace = true deno_terminal.workspace = true
eszip = "=0.70.1" eszip = "=0.71.0"
napi_sym.workspace = true napi_sym.workspace = true
async-trait.workspace = true async-trait.workspace = true
@ -96,10 +96,10 @@ dashmap = "5.5.3"
data-encoding.workspace = true data-encoding.workspace = true
dissimilar = "=1.0.4" dissimilar = "=1.0.4"
dotenvy = "0.15.7" dotenvy = "0.15.7"
dprint-plugin-json = "=0.19.2" dprint-plugin-json = "=0.19.3"
dprint-plugin-jupyter = "=0.1.3" dprint-plugin-jupyter = "=0.1.3"
dprint-plugin-markdown = "=0.17.0" dprint-plugin-markdown = "=0.17.1"
dprint-plugin-typescript = "=0.90.5" dprint-plugin-typescript = "=0.91.1"
env_logger = "=0.10.0" env_logger = "=0.10.0"
fancy-regex = "=0.10.0" fancy-regex = "=0.10.0"
faster-hex.workspace = true faster-hex.workspace = true

View file

@ -197,7 +197,7 @@ pub fn ts_config_to_transpile_and_emit_options(
}, },
deno_ast::EmitOptions { deno_ast::EmitOptions {
inline_sources: options.inline_sources, inline_sources: options.inline_sources,
keep_comments: true, remove_comments: false,
source_map, source_map,
source_map_file: None, source_map_file: None,
}, },

40
cli/cache/emit.rs vendored
View file

@ -45,7 +45,7 @@ impl EmitCache {
&self, &self,
specifier: &ModuleSpecifier, specifier: &ModuleSpecifier,
expected_source_hash: u64, expected_source_hash: u64,
) -> Option<String> { ) -> Option<Vec<u8>> {
let meta_filename = self.get_meta_filename(specifier)?; let meta_filename = self.get_meta_filename(specifier)?;
let emit_filename = self.get_emit_filename(specifier)?; let emit_filename = self.get_emit_filename(specifier)?;
@ -63,8 +63,7 @@ impl EmitCache {
} }
// everything looks good, return it // everything looks good, return it
let emit_text = String::from_utf8(emit_bytes).ok()?; Some(emit_bytes)
Some(emit_text)
} }
/// Gets the filepath which stores the emit. /// Gets the filepath which stores the emit.
@ -85,7 +84,7 @@ impl EmitCache {
&self, &self,
specifier: &ModuleSpecifier, specifier: &ModuleSpecifier,
source_hash: u64, source_hash: u64,
code: &str, code: &[u8],
) { ) {
if let Err(err) = self.set_emit_code_result(specifier, source_hash, code) { if let Err(err) = self.set_emit_code_result(specifier, source_hash, code) {
// should never error here, but if it ever does don't fail // should never error here, but if it ever does don't fail
@ -101,7 +100,7 @@ impl EmitCache {
&self, &self,
specifier: &ModuleSpecifier, specifier: &ModuleSpecifier,
source_hash: u64, source_hash: u64,
code: &str, code: &[u8],
) -> Result<(), AnyError> { ) -> Result<(), AnyError> {
let meta_filename = self let meta_filename = self
.get_meta_filename(specifier) .get_meta_filename(specifier)
@ -113,14 +112,14 @@ impl EmitCache {
// save the metadata // save the metadata
let metadata = EmitMetadata { let metadata = EmitMetadata {
source_hash, source_hash,
emit_hash: compute_emit_hash(code.as_bytes(), self.cli_version), emit_hash: compute_emit_hash(code, self.cli_version),
}; };
self self
.disk_cache .disk_cache
.set(&meta_filename, &serde_json::to_vec(&metadata)?)?; .set(&meta_filename, &serde_json::to_vec(&metadata)?)?;
// save the emit source // save the emit source
self.disk_cache.set(&emit_filename, code.as_bytes())?; self.disk_cache.set(&emit_filename, code)?;
Ok(()) Ok(())
} }
@ -163,6 +162,8 @@ mod test {
disk_cache: disk_cache.clone(), disk_cache: disk_cache.clone(),
cli_version: "1.0.0", cli_version: "1.0.0",
}; };
let to_string =
|bytes: Vec<u8>| -> String { String::from_utf8(bytes).unwrap() };
let specifier1 = let specifier1 =
ModuleSpecifier::from_file_path(temp_dir.path().join("file1.ts")) ModuleSpecifier::from_file_path(temp_dir.path().join("file1.ts"))
@ -173,16 +174,19 @@ mod test {
assert_eq!(cache.get_emit_code(&specifier1, 1), None); assert_eq!(cache.get_emit_code(&specifier1, 1), None);
let emit_code1 = "text1".to_string(); let emit_code1 = "text1".to_string();
let emit_code2 = "text2".to_string(); let emit_code2 = "text2".to_string();
cache.set_emit_code(&specifier1, 10, &emit_code1); cache.set_emit_code(&specifier1, 10, emit_code1.as_bytes());
cache.set_emit_code(&specifier2, 2, &emit_code2); cache.set_emit_code(&specifier2, 2, emit_code2.as_bytes());
// providing the incorrect source hash // providing the incorrect source hash
assert_eq!(cache.get_emit_code(&specifier1, 5), None); assert_eq!(cache.get_emit_code(&specifier1, 5), None);
// providing the correct source hash // providing the correct source hash
assert_eq!( assert_eq!(
cache.get_emit_code(&specifier1, 10), cache.get_emit_code(&specifier1, 10).map(to_string),
Some(emit_code1.clone()), Some(emit_code1.clone()),
); );
assert_eq!(cache.get_emit_code(&specifier2, 2), Some(emit_code2)); assert_eq!(
cache.get_emit_code(&specifier2, 2).map(to_string),
Some(emit_code2)
);
// try changing the cli version (should not load previous ones) // try changing the cli version (should not load previous ones)
let cache = EmitCache { let cache = EmitCache {
@ -190,19 +194,25 @@ mod test {
cli_version: "2.0.0", cli_version: "2.0.0",
}; };
assert_eq!(cache.get_emit_code(&specifier1, 10), None); assert_eq!(cache.get_emit_code(&specifier1, 10), None);
cache.set_emit_code(&specifier1, 5, &emit_code1); cache.set_emit_code(&specifier1, 5, emit_code1.as_bytes());
// recreating the cache should still load the data because the CLI version is the same // recreating the cache should still load the data because the CLI version is the same
let cache = EmitCache { let cache = EmitCache {
disk_cache, disk_cache,
cli_version: "2.0.0", cli_version: "2.0.0",
}; };
assert_eq!(cache.get_emit_code(&specifier1, 5), Some(emit_code1)); assert_eq!(
cache.get_emit_code(&specifier1, 5).map(to_string),
Some(emit_code1)
);
// adding when already exists should not cause issue // adding when already exists should not cause issue
let emit_code3 = "asdf".to_string(); let emit_code3 = "asdf".to_string();
cache.set_emit_code(&specifier1, 20, &emit_code3); cache.set_emit_code(&specifier1, 20, emit_code3.as_bytes());
assert_eq!(cache.get_emit_code(&specifier1, 5), None); assert_eq!(cache.get_emit_code(&specifier1, 5), None);
assert_eq!(cache.get_emit_code(&specifier1, 20), Some(emit_code3)); assert_eq!(
cache.get_emit_code(&specifier1, 20).map(to_string),
Some(emit_code3)
);
} }
} }

View file

@ -75,7 +75,7 @@ impl ParsedSourceCache {
) -> Result<ParsedSource, deno_ast::ParseDiagnostic> { ) -> Result<ParsedSource, deno_ast::ParseDiagnostic> {
if let Some(parsed_source) = self.remove_parsed_source(specifier) { if let Some(parsed_source) = self.remove_parsed_source(specifier) {
if parsed_source.media_type() == media_type if parsed_source.media_type() == media_type
&& parsed_source.text_info().text_str() == source.as_ref() && parsed_source.text().as_ref() == source.as_ref()
{ {
// note: message used tests // note: message used tests
log::debug!("Removed parsed source: {}", specifier); log::debug!("Removed parsed source: {}", specifier);

View file

@ -10,7 +10,7 @@ use deno_core::error::AnyError;
use deno_core::futures::stream::FuturesUnordered; use deno_core::futures::stream::FuturesUnordered;
use deno_core::futures::FutureExt; use deno_core::futures::FutureExt;
use deno_core::futures::StreamExt; use deno_core::futures::StreamExt;
use deno_core::ModuleCodeString; use deno_core::ModuleCodeBytes;
use deno_core::ModuleSpecifier; use deno_core::ModuleSpecifier;
use deno_graph::MediaType; use deno_graph::MediaType;
use deno_graph::Module; use deno_graph::Module;
@ -90,7 +90,7 @@ impl Emitter {
&self, &self,
specifier: &ModuleSpecifier, specifier: &ModuleSpecifier,
source: &str, source: &str,
) -> Option<String> { ) -> Option<Vec<u8>> {
let source_hash = self.get_source_hash(source); let source_hash = self.get_source_hash(source);
self.emit_cache.get_emit_code(specifier, source_hash) self.emit_cache.get_emit_code(specifier, source_hash)
} }
@ -100,7 +100,7 @@ impl Emitter {
specifier: &ModuleSpecifier, specifier: &ModuleSpecifier,
media_type: MediaType, media_type: MediaType,
source: &Arc<str>, source: &Arc<str>,
) -> Result<ModuleCodeString, AnyError> { ) -> Result<ModuleCodeBytes, AnyError> {
// Note: keep this in sync with the sync version below // Note: keep this in sync with the sync version below
let helper = EmitParsedSourceHelper(self); let helper = EmitParsedSourceHelper(self);
match helper.pre_emit_parsed_source(specifier, source) { match helper.pre_emit_parsed_source(specifier, source) {
@ -139,7 +139,7 @@ impl Emitter {
specifier: &ModuleSpecifier, specifier: &ModuleSpecifier,
media_type: MediaType, media_type: MediaType,
source: &Arc<str>, source: &Arc<str>,
) -> Result<ModuleCodeString, AnyError> { ) -> Result<ModuleCodeBytes, AnyError> {
// Note: keep this in sync with the async version above // Note: keep this in sync with the async version above
let helper = EmitParsedSourceHelper(self); let helper = EmitParsedSourceHelper(self);
match helper.pre_emit_parsed_source(specifier, source) { match helper.pre_emit_parsed_source(specifier, source) {
@ -172,16 +172,43 @@ impl Emitter {
ModuleSpecifier::to_file_path(specifier).unwrap(), ModuleSpecifier::to_file_path(specifier).unwrap(),
) )
.await?; .await?;
let source_arc: Arc<str> = source_code.into(); match media_type {
let parsed_source = self MediaType::TypeScript
.parsed_source_cache | MediaType::Mts
.remove_or_parse_module(specifier, source_arc, media_type)?; | MediaType::Cts
let mut options = self.transpile_and_emit_options.1.clone(); | MediaType::Jsx
options.source_map = SourceMapOption::None; | MediaType::Tsx => {
let transpiled_source = parsed_source let source_arc: Arc<str> = source_code.into();
.transpile(&self.transpile_and_emit_options.0, &options)? let parsed_source = self
.into_source(); .parsed_source_cache
Ok(transpiled_source.text) .remove_or_parse_module(specifier, source_arc, media_type)?;
// HMR doesn't work with embedded source maps for some reason, so set
// the option to not use them (though you should test this out because
// this statement is probably wrong)
let mut options = self.transpile_and_emit_options.1.clone();
options.source_map = SourceMapOption::None;
let transpiled_source = parsed_source
.transpile(&self.transpile_and_emit_options.0, &options)?
.into_source()
.into_string()?;
Ok(transpiled_source.text)
}
MediaType::JavaScript
| MediaType::Mjs
| MediaType::Cjs
| MediaType::Dts
| MediaType::Dmts
| MediaType::Dcts
| MediaType::Json
| MediaType::Wasm
| MediaType::TsBuildInfo
| MediaType::SourceMap
| MediaType::Unknown => {
// clear this specifier from the parsed source cache as it's now out of date
self.parsed_source_cache.free(specifier);
Ok(source_code)
}
}
} }
/// A hashing function that takes the source code and uses the global emit /// A hashing function that takes the source code and uses the global emit
@ -196,7 +223,7 @@ impl Emitter {
} }
enum PreEmitResult { enum PreEmitResult {
Cached(ModuleCodeString), Cached(ModuleCodeBytes),
NotCached { source_hash: u64 }, NotCached { source_hash: u64 },
} }
@ -214,7 +241,7 @@ impl<'a> EmitParsedSourceHelper<'a> {
if let Some(emit_code) = if let Some(emit_code) =
self.0.emit_cache.get_emit_code(specifier, source_hash) self.0.emit_cache.get_emit_code(specifier, source_hash)
{ {
PreEmitResult::Cached(emit_code.into()) PreEmitResult::Cached(emit_code.into_boxed_slice().into())
} else { } else {
PreEmitResult::NotCached { source_hash } PreEmitResult::NotCached { source_hash }
} }
@ -240,7 +267,7 @@ impl<'a> EmitParsedSourceHelper<'a> {
specifier: &ModuleSpecifier, specifier: &ModuleSpecifier,
transpile_result: TranspileResult, transpile_result: TranspileResult,
source_hash: u64, source_hash: u64,
) -> ModuleCodeString { ) -> ModuleCodeBytes {
let transpiled_source = match transpile_result { let transpiled_source = match transpile_result {
TranspileResult::Owned(source) => source, TranspileResult::Owned(source) => source,
TranspileResult::Cloned(source) => { TranspileResult::Cloned(source) => {
@ -252,8 +279,8 @@ impl<'a> EmitParsedSourceHelper<'a> {
self.0.emit_cache.set_emit_code( self.0.emit_cache.set_emit_code(
specifier, specifier,
source_hash, source_hash,
&transpiled_source.text, &transpiled_source.source,
); );
transpiled_source.text.into() transpiled_source.source.into_boxed_slice().into()
} }
} }

View file

@ -78,7 +78,7 @@ pub fn graph_valid(
let mut errors = graph let mut errors = graph
.walk( .walk(
roots, roots.iter(),
deno_graph::WalkOptions { deno_graph::WalkOptions {
check_js: options.check_js, check_js: options.check_js,
follow_type_only: options.follow_type_only, follow_type_only: options.follow_type_only,
@ -479,7 +479,7 @@ impl ModuleGraphBuilder {
}; };
let cli_resolver = &self.resolver; let cli_resolver = &self.resolver;
let graph_resolver = cli_resolver.as_graph_resolver(); let graph_resolver = cli_resolver.as_graph_resolver();
let graph_npm_resolver = cli_resolver.as_graph_npm_resolver(); let graph_npm_resolver = cli_resolver.create_graph_npm_resolver();
let maybe_file_watcher_reporter = self let maybe_file_watcher_reporter = self
.maybe_file_watcher_reporter .maybe_file_watcher_reporter
.as_ref() .as_ref()
@ -503,7 +503,7 @@ impl ModuleGraphBuilder {
executor: Default::default(), executor: Default::default(),
file_system: &DenoGraphFsAdapter(self.fs.as_ref()), file_system: &DenoGraphFsAdapter(self.fs.as_ref()),
jsr_url_provider: &CliJsrUrlProvider, jsr_url_provider: &CliJsrUrlProvider,
npm_resolver: Some(graph_npm_resolver), npm_resolver: Some(&graph_npm_resolver),
module_analyzer: &analyzer, module_analyzer: &analyzer,
reporter: maybe_file_watcher_reporter, reporter: maybe_file_watcher_reporter,
resolver: Some(graph_resolver), resolver: Some(graph_resolver),
@ -595,7 +595,6 @@ impl ModuleGraphBuilder {
if has_redirects_changed if has_redirects_changed
|| has_jsr_package_deps_changed || has_jsr_package_deps_changed
|| has_jsr_package_mappings_changed || has_jsr_package_mappings_changed
|| has_npm_packages_changed
{ {
if let Some(lockfile) = &self.lockfile { if let Some(lockfile) = &self.lockfile {
let mut lockfile = lockfile.lock(); let mut lockfile = lockfile.lock();
@ -624,10 +623,6 @@ impl ModuleGraphBuilder {
.add_package_deps(&name.to_string(), deps.map(|s| s.to_string())); .add_package_deps(&name.to_string(), deps.map(|s| s.to_string()));
} }
} }
// npm packages
if has_npm_packages_changed {
self.npm_resolver.as_managed().unwrap().lock(&mut lockfile);
}
} }
} }
@ -652,7 +647,7 @@ impl ModuleGraphBuilder {
let parser = self.parsed_source_cache.as_capturing_parser(); let parser = self.parsed_source_cache.as_capturing_parser();
let cli_resolver = &self.resolver; let cli_resolver = &self.resolver;
let graph_resolver = cli_resolver.as_graph_resolver(); let graph_resolver = cli_resolver.as_graph_resolver();
let graph_npm_resolver = cli_resolver.as_graph_npm_resolver(); let graph_npm_resolver = cli_resolver.create_graph_npm_resolver();
let workspace_members = if options.workspace_fast_check { let workspace_members = if options.workspace_fast_check {
Some(self.options.resolve_deno_graph_workspace_members()?) Some(self.options.resolve_deno_graph_workspace_members()?)
} else { } else {
@ -666,7 +661,7 @@ impl ModuleGraphBuilder {
fast_check_dts: false, fast_check_dts: false,
module_parser: Some(&parser), module_parser: Some(&parser),
resolver: Some(graph_resolver), resolver: Some(graph_resolver),
npm_resolver: Some(graph_npm_resolver), npm_resolver: Some(&graph_npm_resolver),
workspace_fast_check: if let Some(members) = &workspace_members { workspace_fast_check: if let Some(members) = &workspace_members {
deno_graph::WorkspaceFastCheckOption::Enabled(members) deno_graph::WorkspaceFastCheckOption::Enabled(members)
} else { } else {
@ -701,7 +696,10 @@ impl ModuleGraphBuilder {
/// so. Returns `Err(_)` if there is a known module graph or resolution /// so. Returns `Err(_)` if there is a known module graph or resolution
/// error statically reachable from `roots` and not a dynamic import. /// error statically reachable from `roots` and not a dynamic import.
pub fn graph_valid(&self, graph: &ModuleGraph) -> Result<(), AnyError> { pub fn graph_valid(&self, graph: &ModuleGraph) -> Result<(), AnyError> {
self.graph_roots_valid(graph, &graph.roots) self.graph_roots_valid(
graph,
&graph.roots.iter().cloned().collect::<Vec<_>>(),
)
} }
pub fn graph_roots_valid( pub fn graph_roots_valid(
@ -891,9 +889,8 @@ pub fn has_graph_root_local_dependent_changed(
root: &ModuleSpecifier, root: &ModuleSpecifier,
canonicalized_changed_paths: &HashSet<PathBuf>, canonicalized_changed_paths: &HashSet<PathBuf>,
) -> bool { ) -> bool {
let roots = vec![root.clone()];
let mut dependent_specifiers = graph.walk( let mut dependent_specifiers = graph.walk(
&roots, std::iter::once(root),
deno_graph::WalkOptions { deno_graph::WalkOptions {
follow_dynamic: true, follow_dynamic: true,
follow_type_only: true, follow_type_only: true,

View file

@ -724,8 +724,8 @@ impl CodeActionCollection {
&mut self, &mut self,
specifier: &ModuleSpecifier, specifier: &ModuleSpecifier,
diagnostic: &lsp::Diagnostic, diagnostic: &lsp::Diagnostic,
maybe_text_info: Option<SourceTextInfo>, maybe_text_info: Option<&SourceTextInfo>,
maybe_parsed_source: Option<deno_ast::ParsedSource>, maybe_parsed_source: Option<&deno_ast::ParsedSource>,
) -> Result<(), AnyError> { ) -> Result<(), AnyError> {
if let Some(data_quick_fixes) = diagnostic if let Some(data_quick_fixes) = diagnostic
.data .data
@ -774,8 +774,8 @@ impl CodeActionCollection {
&mut self, &mut self,
specifier: &ModuleSpecifier, specifier: &ModuleSpecifier,
diagnostic: &lsp::Diagnostic, diagnostic: &lsp::Diagnostic,
maybe_text_info: Option<SourceTextInfo>, maybe_text_info: Option<&SourceTextInfo>,
maybe_parsed_source: Option<deno_ast::ParsedSource>, maybe_parsed_source: Option<&deno_ast::ParsedSource>,
) -> Result<(), AnyError> { ) -> Result<(), AnyError> {
let code = diagnostic let code = diagnostic
.code .code
@ -830,7 +830,7 @@ impl CodeActionCollection {
.push(CodeActionKind::DenoLint(ignore_error_action)); .push(CodeActionKind::DenoLint(ignore_error_action));
// Disable a lint error for the entire file. // Disable a lint error for the entire file.
let maybe_ignore_comment = maybe_parsed_source.clone().and_then(|ps| { let maybe_ignore_comment = maybe_parsed_source.and_then(|ps| {
// Note: we can use ps.get_leading_comments() but it doesn't // Note: we can use ps.get_leading_comments() but it doesn't
// work when shebang is present at the top of the file. // work when shebang is present at the top of the file.
ps.comments().get_vec().iter().find_map(|c| { ps.comments().get_vec().iter().find_map(|c| {
@ -861,9 +861,8 @@ impl CodeActionCollection {
if let Some(ignore_comment) = maybe_ignore_comment { if let Some(ignore_comment) = maybe_ignore_comment {
new_text = format!(" {code}"); new_text = format!(" {code}");
// Get the end position of the comment. // Get the end position of the comment.
let line = maybe_parsed_source let line = maybe_text_info
.unwrap() .unwrap()
.text_info()
.line_and_column_index(ignore_comment.end()); .line_and_column_index(ignore_comment.end());
let position = lsp::Position { let position = lsp::Position {
line: line.line_index as u32, line: line.line_index as u32,

View file

@ -67,7 +67,7 @@ impl DenoTestCollector {
fn add_code_lenses<N: AsRef<str>>(&mut self, name: N, range: &SourceRange) { fn add_code_lenses<N: AsRef<str>>(&mut self, name: N, range: &SourceRange) {
let range = let range =
source_range_to_lsp_range(range, self.parsed_source.text_info()); source_range_to_lsp_range(range, self.parsed_source.text_info_lazy());
self.add_code_lens(&name, range, "\u{fe0e} Run Test", false); self.add_code_lens(&name, range, "\u{fe0e} Run Test", false);
self.add_code_lens(&name, range, "Debug", true); self.add_code_lens(&name, range, "Debug", true);
} }
@ -406,7 +406,7 @@ pub async fn resolve_code_lens(
pub fn collect_test( pub fn collect_test(
specifier: &ModuleSpecifier, specifier: &ModuleSpecifier,
parsed_source: ParsedSource, parsed_source: &ParsedSource,
) -> Result<Vec<lsp::CodeLens>, AnyError> { ) -> Result<Vec<lsp::CodeLens>, AnyError> {
let mut collector = let mut collector =
DenoTestCollector::new(specifier.clone(), parsed_source.clone()); DenoTestCollector::new(specifier.clone(), parsed_source.clone());
@ -537,7 +537,6 @@ pub fn collect_tsc(
#[cfg(test)] #[cfg(test)]
mod tests { mod tests {
use deno_ast::MediaType; use deno_ast::MediaType;
use deno_ast::SourceTextInfo;
use super::*; use super::*;
@ -562,7 +561,7 @@ mod tests {
"#; "#;
let parsed_module = deno_ast::parse_module(deno_ast::ParseParams { let parsed_module = deno_ast::parse_module(deno_ast::ParseParams {
specifier: specifier.clone(), specifier: specifier.clone(),
text_info: SourceTextInfo::new(source.into()), text: source.into(),
media_type: MediaType::TypeScript, media_type: MediaType::TypeScript,
capture_tokens: true, capture_tokens: true,
scope_analysis: true, scope_analysis: true,

View file

@ -164,7 +164,7 @@ pub async fn get_import_completions(
let document = documents.get(specifier)?; let document = documents.get(specifier)?;
let file_referrer = document.file_referrer(); let file_referrer = document.file_referrer();
let (text, _, range) = document.get_maybe_dependency(position)?; let (text, _, range) = document.get_maybe_dependency(position)?;
let range = to_narrow_lsp_range(&document.text_info(), &range); let range = to_narrow_lsp_range(document.text_info(), &range);
let resolved = resolver let resolved = resolver
.as_graph_resolver(file_referrer) .as_graph_resolver(file_referrer)
.resolve( .resolve(

View file

@ -854,7 +854,7 @@ fn generate_document_lint_diagnostics(
match document.maybe_parsed_source() { match document.maybe_parsed_source() {
Some(Ok(parsed_source)) => { Some(Ok(parsed_source)) => {
if let Ok(references) = if let Ok(references) =
analysis::get_lint_references(&parsed_source, lint_rules, lint_config) analysis::get_lint_references(parsed_source, lint_rules, lint_config)
{ {
references references
.into_iter() .into_iter()

View file

@ -155,7 +155,7 @@ impl AssetOrDocument {
pub fn text(&self) -> Arc<str> { pub fn text(&self) -> Arc<str> {
match self { match self {
AssetOrDocument::Asset(a) => a.text(), AssetOrDocument::Asset(a) => a.text(),
AssetOrDocument::Document(d) => d.text_info.text(), AssetOrDocument::Document(d) => d.text.clone(),
} }
} }
@ -191,7 +191,7 @@ impl AssetOrDocument {
pub fn maybe_parsed_source( pub fn maybe_parsed_source(
&self, &self,
) -> Option<Result<deno_ast::ParsedSource, deno_ast::ParseDiagnostic>> { ) -> Option<&Result<deno_ast::ParsedSource, deno_ast::ParseDiagnostic>> {
self.document().and_then(|d| d.maybe_parsed_source()) self.document().and_then(|d| d.maybe_parsed_source())
} }
@ -243,7 +243,7 @@ fn get_maybe_test_module_fut(
let handle = tokio::task::spawn_blocking(move || { let handle = tokio::task::spawn_blocking(move || {
let mut collector = TestCollector::new( let mut collector = TestCollector::new(
parsed_source.specifier().clone(), parsed_source.specifier().clone(),
parsed_source.text_info().clone(), parsed_source.text_info_lazy().clone(),
); );
parsed_source.module().visit_with(&mut collector); parsed_source.module().visit_with(&mut collector);
Arc::new(collector.take()) Arc::new(collector.take())
@ -283,7 +283,8 @@ pub struct Document {
open_data: Option<DocumentOpenData>, open_data: Option<DocumentOpenData>,
resolver: Arc<LspResolver>, resolver: Arc<LspResolver>,
specifier: ModuleSpecifier, specifier: ModuleSpecifier,
text_info: SourceTextInfo, text: Arc<str>,
text_info_cell: once_cell::sync::OnceCell<SourceTextInfo>,
} }
impl Document { impl Document {
@ -291,7 +292,7 @@ impl Document {
#[allow(clippy::too_many_arguments)] #[allow(clippy::too_many_arguments)]
fn new( fn new(
specifier: ModuleSpecifier, specifier: ModuleSpecifier,
content: Arc<str>, text: Arc<str>,
maybe_lsp_version: Option<i32>, maybe_lsp_version: Option<i32>,
maybe_language_id: Option<LanguageId>, maybe_language_id: Option<LanguageId>,
maybe_headers: Option<HashMap<String, String>>, maybe_headers: Option<HashMap<String, String>>,
@ -300,7 +301,6 @@ impl Document {
cache: &Arc<LspCache>, cache: &Arc<LspCache>,
file_referrer: Option<ModuleSpecifier>, file_referrer: Option<ModuleSpecifier>,
) -> Arc<Self> { ) -> Arc<Self> {
let text_info = SourceTextInfo::new(content);
let media_type = resolve_media_type( let media_type = resolve_media_type(
&specifier, &specifier,
maybe_headers.as_ref(), maybe_headers.as_ref(),
@ -311,7 +311,7 @@ impl Document {
if media_type_is_diagnosable(media_type) { if media_type_is_diagnosable(media_type) {
parse_and_analyze_module( parse_and_analyze_module(
specifier.clone(), specifier.clone(),
text_info.clone(), text.clone(),
maybe_headers.as_ref(), maybe_headers.as_ref(),
media_type, media_type,
file_referrer.as_ref(), file_referrer.as_ref(),
@ -328,7 +328,7 @@ impl Document {
let maybe_types_dependency = maybe_module let maybe_types_dependency = maybe_module
.as_ref() .as_ref()
.and_then(|m| Some(Arc::new(m.maybe_types_dependency.clone()?))); .and_then(|m| Some(Arc::new(m.maybe_types_dependency.clone()?)));
let line_index = Arc::new(LineIndex::new(text_info.text_str())); let line_index = Arc::new(LineIndex::new(text.as_ref()));
let maybe_test_module_fut = let maybe_test_module_fut =
get_maybe_test_module_fut(maybe_parsed_source.as_ref(), &config); get_maybe_test_module_fut(maybe_parsed_source.as_ref(), &config);
Arc::new(Self { Arc::new(Self {
@ -350,7 +350,8 @@ impl Document {
}), }),
resolver, resolver,
specifier, specifier,
text_info, text,
text_info_cell: once_cell::sync::OnceCell::new(),
}) })
} }
@ -370,11 +371,8 @@ impl Document {
let maybe_parsed_source; let maybe_parsed_source;
let maybe_test_module_fut; let maybe_test_module_fut;
if media_type != self.media_type { if media_type != self.media_type {
let parsed_source_result = parse_source( let parsed_source_result =
self.specifier.clone(), parse_source(self.specifier.clone(), self.text.clone(), media_type);
self.text_info.clone(),
media_type,
);
let maybe_module = analyze_module( let maybe_module = analyze_module(
self.specifier.clone(), self.specifier.clone(),
&parsed_source_result, &parsed_source_result,
@ -397,7 +395,7 @@ impl Document {
let graph_resolver = let graph_resolver =
resolver.as_graph_resolver(self.file_referrer.as_ref()); resolver.as_graph_resolver(self.file_referrer.as_ref());
let npm_resolver = let npm_resolver =
resolver.as_graph_npm_resolver(self.file_referrer.as_ref()); resolver.create_graph_npm_resolver(self.file_referrer.as_ref());
dependencies = Arc::new( dependencies = Arc::new(
self self
.dependencies .dependencies
@ -409,7 +407,7 @@ impl Document {
s, s,
&CliJsrUrlProvider, &CliJsrUrlProvider,
Some(graph_resolver), Some(graph_resolver),
Some(npm_resolver), Some(&npm_resolver),
), ),
) )
}) })
@ -419,10 +417,10 @@ impl Document {
Arc::new(d.with_new_resolver( Arc::new(d.with_new_resolver(
&CliJsrUrlProvider, &CliJsrUrlProvider,
Some(graph_resolver), Some(graph_resolver),
Some(npm_resolver), Some(&npm_resolver),
)) ))
}); });
maybe_parsed_source = self.maybe_parsed_source(); maybe_parsed_source = self.maybe_parsed_source().cloned();
maybe_test_module_fut = self maybe_test_module_fut = self
.maybe_test_module_fut .maybe_test_module_fut
.clone() .clone()
@ -450,7 +448,8 @@ impl Document {
}), }),
resolver, resolver,
specifier: self.specifier.clone(), specifier: self.specifier.clone(),
text_info: self.text_info.clone(), text: self.text.clone(),
text_info_cell: once_cell::sync::OnceCell::new(),
}) })
} }
@ -459,7 +458,7 @@ impl Document {
version: i32, version: i32,
changes: Vec<lsp::TextDocumentContentChangeEvent>, changes: Vec<lsp::TextDocumentContentChangeEvent>,
) -> Result<Arc<Self>, AnyError> { ) -> Result<Arc<Self>, AnyError> {
let mut content = self.text_info.text_str().to_string(); let mut content = self.text.to_string();
let mut line_index = self.line_index.clone(); let mut line_index = self.line_index.clone();
let mut index_valid = IndexValid::All; let mut index_valid = IndexValid::All;
for change in changes { for change in changes {
@ -475,7 +474,7 @@ impl Document {
index_valid = IndexValid::UpTo(0); index_valid = IndexValid::UpTo(0);
} }
} }
let text_info = SourceTextInfo::from_string(content); let text: Arc<str> = content.into();
let media_type = self.media_type; let media_type = self.media_type;
let (maybe_parsed_source, maybe_module) = if self let (maybe_parsed_source, maybe_module) = if self
.maybe_language_id .maybe_language_id
@ -485,7 +484,7 @@ impl Document {
{ {
parse_and_analyze_module( parse_and_analyze_module(
self.specifier.clone(), self.specifier.clone(),
text_info.clone(), text.clone(),
self.maybe_headers.as_ref(), self.maybe_headers.as_ref(),
media_type, media_type,
self.file_referrer.as_ref(), self.file_referrer.as_ref(),
@ -506,7 +505,7 @@ impl Document {
let line_index = if index_valid == IndexValid::All { let line_index = if index_valid == IndexValid::All {
line_index line_index
} else { } else {
Arc::new(LineIndex::new(text_info.text_str())) Arc::new(LineIndex::new(text.as_ref()))
}; };
let maybe_test_module_fut = let maybe_test_module_fut =
get_maybe_test_module_fut(maybe_parsed_source.as_ref(), &self.config); get_maybe_test_module_fut(maybe_parsed_source.as_ref(), &self.config);
@ -518,7 +517,8 @@ impl Document {
maybe_language_id: self.maybe_language_id, maybe_language_id: self.maybe_language_id,
dependencies, dependencies,
maybe_types_dependency, maybe_types_dependency,
text_info, text,
text_info_cell: once_cell::sync::OnceCell::new(),
line_index, line_index,
maybe_headers: self.maybe_headers.clone(), maybe_headers: self.maybe_headers.clone(),
maybe_navigation_tree: Mutex::new(None), maybe_navigation_tree: Mutex::new(None),
@ -542,7 +542,8 @@ impl Document {
maybe_language_id: self.maybe_language_id, maybe_language_id: self.maybe_language_id,
dependencies: self.dependencies.clone(), dependencies: self.dependencies.clone(),
maybe_types_dependency: self.maybe_types_dependency.clone(), maybe_types_dependency: self.maybe_types_dependency.clone(),
text_info: self.text_info.clone(), text: self.text.clone(),
text_info_cell: once_cell::sync::OnceCell::new(),
line_index: self.line_index.clone(), line_index: self.line_index.clone(),
maybe_headers: self.maybe_headers.clone(), maybe_headers: self.maybe_headers.clone(),
maybe_navigation_tree: Mutex::new( maybe_navigation_tree: Mutex::new(
@ -564,7 +565,8 @@ impl Document {
maybe_language_id: self.maybe_language_id, maybe_language_id: self.maybe_language_id,
dependencies: self.dependencies.clone(), dependencies: self.dependencies.clone(),
maybe_types_dependency: self.maybe_types_dependency.clone(), maybe_types_dependency: self.maybe_types_dependency.clone(),
text_info: self.text_info.clone(), text: self.text.clone(),
text_info_cell: once_cell::sync::OnceCell::new(),
line_index: self.line_index.clone(), line_index: self.line_index.clone(),
maybe_headers: self.maybe_headers.clone(), maybe_headers: self.maybe_headers.clone(),
maybe_navigation_tree: Mutex::new( maybe_navigation_tree: Mutex::new(
@ -585,12 +587,22 @@ impl Document {
self.file_referrer.as_ref() self.file_referrer.as_ref()
} }
pub fn content(&self) -> Arc<str> { pub fn content(&self) -> &Arc<str> {
self.text_info.text() &self.text
} }
pub fn text_info(&self) -> SourceTextInfo { pub fn text_info(&self) -> &SourceTextInfo {
self.text_info.clone() // try to get the text info from the parsed source and if
// not then create one in the cell
self
.maybe_parsed_source()
.and_then(|p| p.as_ref().ok())
.map(|p| p.text_info_lazy())
.unwrap_or_else(|| {
self
.text_info_cell
.get_or_init(|| SourceTextInfo::new(self.text.clone()))
})
} }
pub fn line_index(&self) -> Arc<LineIndex> { pub fn line_index(&self) -> Arc<LineIndex> {
@ -647,8 +659,8 @@ impl Document {
pub fn maybe_parsed_source( pub fn maybe_parsed_source(
&self, &self,
) -> Option<Result<deno_ast::ParsedSource, deno_ast::ParseDiagnostic>> { ) -> Option<&Result<deno_ast::ParsedSource, deno_ast::ParseDiagnostic>> {
self.open_data.as_ref()?.maybe_parsed_source.clone() self.open_data.as_ref()?.maybe_parsed_source.as_ref()
} }
pub async fn maybe_test_module(&self) -> Option<Arc<TestModule>> { pub async fn maybe_test_module(&self) -> Option<Arc<TestModule>> {
@ -1421,7 +1433,7 @@ impl<'a> OpenDocumentsGraphLoader<'a> {
if let Some(doc) = self.open_docs.get(specifier) { if let Some(doc) = self.open_docs.get(specifier) {
return Some( return Some(
future::ready(Ok(Some(deno_graph::source::LoadResponse::Module { future::ready(Ok(Some(deno_graph::source::LoadResponse::Module {
content: Arc::from(doc.content()), content: Arc::from(doc.content().clone()),
specifier: doc.specifier().clone(), specifier: doc.specifier().clone(),
maybe_headers: None, maybe_headers: None,
}))) })))
@ -1459,14 +1471,13 @@ impl<'a> deno_graph::source::Loader for OpenDocumentsGraphLoader<'a> {
fn parse_and_analyze_module( fn parse_and_analyze_module(
specifier: ModuleSpecifier, specifier: ModuleSpecifier,
text_info: SourceTextInfo, text: Arc<str>,
maybe_headers: Option<&HashMap<String, String>>, maybe_headers: Option<&HashMap<String, String>>,
media_type: MediaType, media_type: MediaType,
file_referrer: Option<&ModuleSpecifier>, file_referrer: Option<&ModuleSpecifier>,
resolver: &LspResolver, resolver: &LspResolver,
) -> (Option<ParsedSourceResult>, Option<ModuleResult>) { ) -> (Option<ParsedSourceResult>, Option<ModuleResult>) {
let parsed_source_result = let parsed_source_result = parse_source(specifier.clone(), text, media_type);
parse_source(specifier.clone(), text_info, media_type);
let module_result = analyze_module( let module_result = analyze_module(
specifier, specifier,
&parsed_source_result, &parsed_source_result,
@ -1479,12 +1490,12 @@ fn parse_and_analyze_module(
fn parse_source( fn parse_source(
specifier: ModuleSpecifier, specifier: ModuleSpecifier,
text_info: SourceTextInfo, text: Arc<str>,
media_type: MediaType, media_type: MediaType,
) -> ParsedSourceResult { ) -> ParsedSourceResult {
deno_ast::parse_module(deno_ast::ParseParams { deno_ast::parse_module(deno_ast::ParseParams {
specifier, specifier,
text_info, text,
media_type, media_type,
capture_tokens: true, capture_tokens: true,
scope_analysis: true, scope_analysis: true,
@ -1500,20 +1511,23 @@ fn analyze_module(
resolver: &LspResolver, resolver: &LspResolver,
) -> ModuleResult { ) -> ModuleResult {
match parsed_source_result { match parsed_source_result {
Ok(parsed_source) => Ok(deno_graph::parse_module_from_ast( Ok(parsed_source) => {
deno_graph::ParseModuleFromAstOptions { let npm_resolver = resolver.create_graph_npm_resolver(file_referrer);
graph_kind: deno_graph::GraphKind::TypesOnly, Ok(deno_graph::parse_module_from_ast(
specifier, deno_graph::ParseModuleFromAstOptions {
maybe_headers, graph_kind: deno_graph::GraphKind::TypesOnly,
parsed_source, specifier,
// use a null file system because there's no need to bother resolving maybe_headers,
// dynamic imports like import(`./dir/${something}`) in the LSP parsed_source,
file_system: &deno_graph::source::NullFileSystem, // use a null file system because there's no need to bother resolving
jsr_url_provider: &CliJsrUrlProvider, // dynamic imports like import(`./dir/${something}`) in the LSP
maybe_resolver: Some(resolver.as_graph_resolver(file_referrer)), file_system: &deno_graph::source::NullFileSystem,
maybe_npm_resolver: Some(resolver.as_graph_npm_resolver(file_referrer)), jsr_url_provider: &CliJsrUrlProvider,
}, maybe_resolver: Some(resolver.as_graph_resolver(file_referrer)),
)), maybe_npm_resolver: Some(&npm_resolver),
},
))
}
Err(err) => Err(deno_graph::ModuleGraphError::ModuleError( Err(err) => Err(deno_graph::ModuleGraphError::ModuleError(
deno_graph::ModuleError::ParseErr(specifier, err.clone()), deno_graph::ModuleError::ParseErr(specifier, err.clone()),
)), )),
@ -1602,7 +1616,7 @@ console.log(b);
) )
.unwrap(); .unwrap();
assert_eq!( assert_eq!(
&*documents.get(&specifier).unwrap().content(), documents.get(&specifier).unwrap().content().as_ref(),
r#"import * as b from "./b.ts"; r#"import * as b from "./b.ts";
console.log(b, "hello deno"); console.log(b, "hello deno");
"# "#

View file

@ -1317,7 +1317,7 @@ impl Inner {
move || { move || {
let format_result = match document.maybe_parsed_source() { let format_result = match document.maybe_parsed_source() {
Some(Ok(parsed_source)) => { Some(Ok(parsed_source)) => {
format_parsed_source(&parsed_source, &fmt_options) format_parsed_source(parsed_source, &fmt_options)
} }
Some(Err(err)) => Err(anyhow!("{:#}", err)), Some(Err(err)) => Err(anyhow!("{:#}", err)),
None => { None => {
@ -1330,12 +1330,12 @@ impl Inner {
.map(|ext| file_path.with_extension(ext)) .map(|ext| file_path.with_extension(ext))
.unwrap_or(file_path); .unwrap_or(file_path);
// it's not a js/ts file, so attempt to format its contents // it's not a js/ts file, so attempt to format its contents
format_file(&file_path, &document.content(), &fmt_options) format_file(&file_path, document.content(), &fmt_options)
} }
}; };
match format_result { match format_result {
Ok(Some(new_text)) => Some(text::get_edits( Ok(Some(new_text)) => Some(text::get_edits(
&document.content(), document.content(),
&new_text, &new_text,
document.line_index().as_ref(), document.line_index().as_ref(),
)), )),
@ -1605,7 +1605,9 @@ impl Inner {
&specifier, &specifier,
diagnostic, diagnostic,
asset_or_doc.document().map(|d| d.text_info()), asset_or_doc.document().map(|d| d.text_info()),
asset_or_doc.maybe_parsed_source().and_then(|r| r.ok()), asset_or_doc
.maybe_parsed_source()
.and_then(|r| r.as_ref().ok()),
) )
.map_err(|err| { .map_err(|err| {
error!("Unable to fix lint error: {:#}", err); error!("Unable to fix lint error: {:#}", err);

View file

@ -20,6 +20,7 @@ use crate::resolver::CliGraphResolver;
use crate::resolver::CliGraphResolverOptions; use crate::resolver::CliGraphResolverOptions;
use crate::resolver::CliNodeResolver; use crate::resolver::CliNodeResolver;
use crate::resolver::SloppyImportsResolver; use crate::resolver::SloppyImportsResolver;
use crate::resolver::WorkerCliNpmGraphResolver;
use crate::util::progress_bar::ProgressBar; use crate::util::progress_bar::ProgressBar;
use crate::util::progress_bar::ProgressBarStyle; use crate::util::progress_bar::ProgressBarStyle;
use dashmap::DashMap; use dashmap::DashMap;
@ -27,7 +28,6 @@ use deno_ast::MediaType;
use deno_cache_dir::HttpCache; use deno_cache_dir::HttpCache;
use deno_core::error::AnyError; use deno_core::error::AnyError;
use deno_core::url::Url; use deno_core::url::Url;
use deno_graph::source::NpmResolver;
use deno_graph::source::Resolver; use deno_graph::source::Resolver;
use deno_graph::GraphImport; use deno_graph::GraphImport;
use deno_graph::ModuleSpecifier; use deno_graph::ModuleSpecifier;
@ -105,6 +105,7 @@ impl LspResolver {
let redirect_resolver = Some(Arc::new(RedirectResolver::new( let redirect_resolver = Some(Arc::new(RedirectResolver::new(
cache.root_vendor_or_global(), cache.root_vendor_or_global(),
))); )));
let npm_graph_resolver = graph_resolver.create_graph_npm_resolver();
let graph_imports = config_data let graph_imports = config_data
.and_then(|d| d.config_file.as_ref()) .and_then(|d| d.config_file.as_ref())
.and_then(|cf| cf.to_maybe_imports().ok()) .and_then(|cf| cf.to_maybe_imports().ok())
@ -118,7 +119,7 @@ impl LspResolver {
imports, imports,
&CliJsrUrlProvider, &CliJsrUrlProvider,
Some(graph_resolver.as_ref()), Some(graph_resolver.as_ref()),
Some(graph_resolver.as_ref()), Some(&npm_graph_resolver),
); );
(referrer, graph_import) (referrer, graph_import)
}) })
@ -180,11 +181,11 @@ impl LspResolver {
self.graph_resolver.as_ref() self.graph_resolver.as_ref()
} }
pub fn as_graph_npm_resolver( pub fn create_graph_npm_resolver(
&self, &self,
_file_referrer: Option<&ModuleSpecifier>, _file_referrer: Option<&ModuleSpecifier>,
) -> &dyn NpmResolver { ) -> WorkerCliNpmGraphResolver {
self.graph_resolver.as_ref() self.graph_resolver.create_graph_npm_resolver()
} }
pub fn maybe_managed_npm_resolver( pub fn maybe_managed_npm_resolver(

View file

@ -641,14 +641,14 @@ pub mod tests {
let parsed_module = deno_ast::parse_module(deno_ast::ParseParams { let parsed_module = deno_ast::parse_module(deno_ast::ParseParams {
specifier: specifier.clone(), specifier: specifier.clone(),
text_info: deno_ast::SourceTextInfo::new(source.into()), text: source.into(),
media_type: deno_ast::MediaType::TypeScript, media_type: deno_ast::MediaType::TypeScript,
capture_tokens: true, capture_tokens: true,
scope_analysis: true, scope_analysis: true,
maybe_syntax: None, maybe_syntax: None,
}) })
.unwrap(); .unwrap();
let text_info = parsed_module.text_info().clone(); let text_info = parsed_module.text_info_lazy().clone();
let mut collector = TestCollector::new(specifier, text_info); let mut collector = TestCollector::new(specifier, text_info);
parsed_module.module().visit_with(&mut collector); parsed_module.module().visit_with(&mut collector);
collector.take() collector.take()

View file

@ -371,8 +371,7 @@ impl<TGraphContainer: ModuleGraphContainer>
// it to work with --inspect or --inspect-brk // it to work with --inspect or --inspect-brk
code_source.code code_source.code
} else { } else {
// reduce memory and throw away the source map // v8 is slower when source maps are present, so we strip them
// because we don't need it
code_without_source_map(code_source.code) code_without_source_map(code_source.code)
}; };
let module_type = match code_source.media_type { let module_type = match code_source.media_type {
@ -413,7 +412,7 @@ impl<TGraphContainer: ModuleGraphContainer>
Ok(ModuleSource::new_with_redirect( Ok(ModuleSource::new_with_redirect(
module_type, module_type,
ModuleSourceCode::String(code), code,
specifier, specifier,
&code_source.found_url, &code_source.found_url,
code_cache, code_cache,
@ -424,26 +423,6 @@ impl<TGraphContainer: ModuleGraphContainer>
&self, &self,
referrer: &str, referrer: &str,
) -> Result<ModuleSpecifier, AnyError> { ) -> Result<ModuleSpecifier, AnyError> {
// todo(https://github.com/denoland/deno_core/pull/741): use function from deno_core
fn specifier_has_uri_scheme(specifier: &str) -> bool {
let mut chars = specifier.chars();
let mut len = 0usize;
// The first character must be a letter.
match chars.next() {
Some(c) if c.is_ascii_alphabetic() => len += 1,
_ => return false,
}
// Second and following characters must be either a letter, number,
// plus sign, minus sign, or dot.
loop {
match chars.next() {
Some(c) if c.is_ascii_alphanumeric() || "+-.".contains(c) => len += 1,
Some(':') if len >= 2 => return true,
_ => return false,
}
}
}
let referrer = if referrer.is_empty() && self.shared.is_repl { let referrer = if referrer.is_empty() && self.shared.is_repl {
// FIXME(bartlomieju): this is a hacky way to provide compatibility with REPL // FIXME(bartlomieju): this is a hacky way to provide compatibility with REPL
// and `Deno.core.evalContext` API. Ideally we should always have a referrer filled // and `Deno.core.evalContext` API. Ideally we should always have a referrer filled
@ -452,7 +431,7 @@ impl<TGraphContainer: ModuleGraphContainer>
referrer referrer
}; };
if specifier_has_uri_scheme(referrer) { if deno_core::specifier_has_uri_scheme(referrer) {
deno_core::resolve_url(referrer).map_err(|e| e.into()) deno_core::resolve_url(referrer).map_err(|e| e.into())
} else if referrer == "." { } else if referrer == "." {
// main module, use the initial cwd // main module, use the initial cwd
@ -612,7 +591,7 @@ impl<TGraphContainer: ModuleGraphContainer>
self.parsed_source_cache.free(specifier); self.parsed_source_cache.free(specifier);
Ok(ModuleCodeStringSource { Ok(ModuleCodeStringSource {
code: transpile_result, code: ModuleSourceCode::Bytes(transpile_result),
found_url: specifier.clone(), found_url: specifier.clone(),
media_type, media_type,
}) })
@ -647,7 +626,7 @@ impl<TGraphContainer: ModuleGraphContainer>
self.parsed_source_cache.free(specifier); self.parsed_source_cache.free(specifier);
Ok(ModuleCodeStringSource { Ok(ModuleCodeStringSource {
code: transpile_result, code: ModuleSourceCode::Bytes(transpile_result),
found_url: specifier.clone(), found_url: specifier.clone(),
media_type, media_type,
}) })
@ -673,7 +652,7 @@ impl<TGraphContainer: ModuleGraphContainer>
specifier, specifier,
.. ..
})) => Ok(CodeOrDeferredEmit::Code(ModuleCodeStringSource { })) => Ok(CodeOrDeferredEmit::Code(ModuleCodeStringSource {
code: source.clone().into(), code: ModuleSourceCode::String(source.clone().into()),
found_url: specifier.clone(), found_url: specifier.clone(),
media_type: *media_type, media_type: *media_type,
})), })),
@ -712,7 +691,7 @@ impl<TGraphContainer: ModuleGraphContainer>
self.parsed_source_cache.free(specifier); self.parsed_source_cache.free(specifier);
Ok(CodeOrDeferredEmit::Code(ModuleCodeStringSource { Ok(CodeOrDeferredEmit::Code(ModuleCodeStringSource {
code, code: ModuleSourceCode::String(code),
found_url: specifier.clone(), found_url: specifier.clone(),
media_type: *media_type, media_type: *media_type,
})) }))
@ -892,7 +871,7 @@ impl<TGraphContainer: ModuleGraphContainer> SourceMapGetter
_ => return None, _ => return None,
} }
let source = self.0.load_prepared_module_sync(&specifier, None).ok()?; let source = self.0.load_prepared_module_sync(&specifier, None).ok()?;
source_map_from_code(&source.code) source_map_from_code(source.code.as_bytes())
} }
fn get_source_line( fn get_source_line(

View file

@ -85,7 +85,7 @@ impl CliCjsCodeAnalyzer {
move || -> Result<_, deno_ast::ParseDiagnostic> { move || -> Result<_, deno_ast::ParseDiagnostic> {
let parsed_source = deno_ast::parse_program(deno_ast::ParseParams { let parsed_source = deno_ast::parse_program(deno_ast::ParseParams {
specifier, specifier,
text_info: deno_ast::SourceTextInfo::new(source), text: source,
media_type, media_type,
capture_tokens: true, capture_tokens: true,
scope_analysis: false, scope_analysis: false,

View file

@ -28,6 +28,7 @@ use crate::util::fs::hard_link_dir_recursive;
mod registry_info; mod registry_info;
mod tarball; mod tarball;
mod tarball_extract; mod tarball_extract;
mod value_creator;
pub use registry_info::RegistryInfoDownloader; pub use registry_info::RegistryInfoDownloader;
pub use tarball::TarballCache; pub use tarball::TarballCache;

View file

@ -8,8 +8,7 @@ use deno_core::anyhow::bail;
use deno_core::anyhow::Context; use deno_core::anyhow::Context;
use deno_core::error::custom_error; use deno_core::error::custom_error;
use deno_core::error::AnyError; use deno_core::error::AnyError;
use deno_core::futures::future::BoxFuture; use deno_core::futures::future::LocalBoxFuture;
use deno_core::futures::future::Shared;
use deno_core::futures::FutureExt; use deno_core::futures::FutureExt;
use deno_core::parking_lot::Mutex; use deno_core::parking_lot::Mutex;
use deno_core::serde_json; use deno_core::serde_json;
@ -23,6 +22,7 @@ use crate::http_util::HttpClientProvider;
use crate::npm::common::maybe_auth_header_for_npm_registry; use crate::npm::common::maybe_auth_header_for_npm_registry;
use crate::util::progress_bar::ProgressBar; use crate::util::progress_bar::ProgressBar;
use super::value_creator::MultiRuntimeAsyncValueCreator;
use super::NpmCache; use super::NpmCache;
// todo(dsherret): create seams and unit test this // todo(dsherret): create seams and unit test this
@ -30,7 +30,7 @@ use super::NpmCache;
#[derive(Debug, Clone)] #[derive(Debug, Clone)]
enum MemoryCacheItem { enum MemoryCacheItem {
/// The cache item hasn't loaded yet. /// The cache item hasn't loaded yet.
PendingFuture(Shared<PendingRegistryLoadFuture>), Pending(Arc<MultiRuntimeAsyncValueCreator<FutureResult>>),
/// The item has loaded in the past and was stored in the file system cache. /// The item has loaded in the past and was stored in the file system cache.
/// There is no reason to request this package from the npm registry again /// There is no reason to request this package from the npm registry again
/// for the duration of execution. /// for the duration of execution.
@ -48,7 +48,7 @@ enum FutureResult {
} }
type PendingRegistryLoadFuture = type PendingRegistryLoadFuture =
BoxFuture<'static, Result<FutureResult, Arc<AnyError>>>; LocalBoxFuture<'static, Result<FutureResult, AnyError>>;
/// Downloads packuments from the npm registry. /// Downloads packuments from the npm registry.
/// ///
@ -79,7 +79,7 @@ impl RegistryInfoDownloader {
} }
pub async fn load_package_info( pub async fn load_package_info(
&self, self: &Arc<Self>,
name: &str, name: &str,
) -> Result<Option<Arc<NpmPackageInfo>>, AnyError> { ) -> Result<Option<Arc<NpmPackageInfo>>, AnyError> {
let registry_url = self.npmrc.get_registry_url(name); let registry_url = self.npmrc.get_registry_url(name);
@ -98,7 +98,7 @@ impl RegistryInfoDownloader {
} }
async fn load_package_info_inner( async fn load_package_info_inner(
&self, self: &Arc<Self>,
name: &str, name: &str,
registry_url: &Url, registry_url: &Url,
registry_config: &RegistryConfig, registry_config: &RegistryConfig,
@ -112,18 +112,20 @@ impl RegistryInfoDownloader {
)); ));
} }
let (created, cache_item) = { let cache_item = {
let mut mem_cache = self.memory_cache.lock(); let mut mem_cache = self.memory_cache.lock();
if let Some(cache_item) = mem_cache.get(name) { if let Some(cache_item) = mem_cache.get(name) {
(false, cache_item.clone()) cache_item.clone()
} else { } else {
let future = let future =
self.create_load_future(name, registry_url, registry_config); self.create_load_future(name, registry_url, registry_config);
let cache_item = MemoryCacheItem::PendingFuture(future); let value_creator = MultiRuntimeAsyncValueCreator::new(future);
let cache_item = MemoryCacheItem::Pending(Arc::new(value_creator));
mem_cache.insert(name.to_string(), cache_item.clone()); mem_cache.insert(name.to_string(), cache_item.clone());
(true, cache_item) cache_item
} }
}; };
match cache_item { match cache_item {
MemoryCacheItem::FsCached => { MemoryCacheItem::FsCached => {
// this struct previously loaded from the registry, so we can load it from the file system cache // this struct previously loaded from the registry, so we can load it from the file system cache
@ -135,40 +137,35 @@ impl RegistryInfoDownloader {
MemoryCacheItem::MemoryCached(maybe_info) => { MemoryCacheItem::MemoryCached(maybe_info) => {
maybe_info.clone().map_err(|e| anyhow!("{}", e)) maybe_info.clone().map_err(|e| anyhow!("{}", e))
} }
MemoryCacheItem::PendingFuture(future) => { MemoryCacheItem::Pending(value_creator) => {
if created { let downloader = self.clone();
match future.await { let future = value_creator.get(move || {
Ok(FutureResult::SavedFsCache(info)) => { downloader.create_load_future(name, registry_url, registry_config)
// return back the future and mark this package as having });
// been saved in the cache for next time it's requested match future.await {
*self.memory_cache.lock().get_mut(name).unwrap() = Ok(FutureResult::SavedFsCache(info)) => {
MemoryCacheItem::FsCached; // return back the future and mark this package as having
Ok(Some(info)) // been saved in the cache for next time it's requested
} *self.memory_cache.lock().get_mut(name).unwrap() =
Ok(FutureResult::ErroredFsCache(info)) => { MemoryCacheItem::FsCached;
// since saving to the fs cache failed, keep the package information in memory Ok(Some(info))
*self.memory_cache.lock().get_mut(name).unwrap() =
MemoryCacheItem::MemoryCached(Ok(Some(info.clone())));
Ok(Some(info))
}
Ok(FutureResult::PackageNotExists) => {
*self.memory_cache.lock().get_mut(name).unwrap() =
MemoryCacheItem::MemoryCached(Ok(None));
Ok(None)
}
Err(err) => {
let return_err = anyhow!("{}", err);
*self.memory_cache.lock().get_mut(name).unwrap() =
MemoryCacheItem::MemoryCached(Err(err));
Err(return_err)
}
} }
} else { Ok(FutureResult::ErroredFsCache(info)) => {
match future.await { // since saving to the fs cache failed, keep the package information in memory
Ok(FutureResult::SavedFsCache(info)) => Ok(Some(info)), *self.memory_cache.lock().get_mut(name).unwrap() =
Ok(FutureResult::ErroredFsCache(info)) => Ok(Some(info)), MemoryCacheItem::MemoryCached(Ok(Some(info.clone())));
Ok(FutureResult::PackageNotExists) => Ok(None), Ok(Some(info))
Err(err) => Err(anyhow!("{}", err)), }
Ok(FutureResult::PackageNotExists) => {
*self.memory_cache.lock().get_mut(name).unwrap() =
MemoryCacheItem::MemoryCached(Ok(None));
Ok(None)
}
Err(err) => {
let return_err = anyhow!("{}", err);
*self.memory_cache.lock().get_mut(name).unwrap() =
MemoryCacheItem::MemoryCached(Err(err));
Err(return_err)
} }
} }
} }
@ -203,23 +200,19 @@ impl RegistryInfoDownloader {
} }
fn create_load_future( fn create_load_future(
&self, self: &Arc<Self>,
name: &str, name: &str,
registry_url: &Url, registry_url: &Url,
registry_config: &RegistryConfig, registry_config: &RegistryConfig,
) -> Shared<PendingRegistryLoadFuture> { ) -> PendingRegistryLoadFuture {
let downloader = self.clone();
let package_url = self.get_package_url(name, registry_url); let package_url = self.get_package_url(name, registry_url);
let maybe_auth_header = maybe_auth_header_for_npm_registry(registry_config); let maybe_auth_header = maybe_auth_header_for_npm_registry(registry_config);
let guard = self.progress_bar.update(package_url.as_str()); let guard = self.progress_bar.update(package_url.as_str());
let cache = self.cache.clone();
let http_client_provider = self.http_client_provider.clone();
let name = name.to_string(); let name = name.to_string();
// force this future to be polled on the current runtime because it's not async move {
// safe to share `HttpClient`s across runtimes and because a restart of let maybe_bytes = downloader
// npm resolution might cause this package not to be resolved again .http_client_provider
// causing the future to never be polled
deno_core::unsync::spawn(async move {
let maybe_bytes = http_client_provider
.get_or_create()? .get_or_create()?
.download_with_progress(package_url, maybe_auth_header, &guard) .download_with_progress(package_url, maybe_auth_header, &guard)
.await?; .await?;
@ -228,7 +221,7 @@ impl RegistryInfoDownloader {
let future_result = deno_core::unsync::spawn_blocking( let future_result = deno_core::unsync::spawn_blocking(
move || -> Result<FutureResult, AnyError> { move || -> Result<FutureResult, AnyError> {
let package_info = serde_json::from_slice(&bytes)?; let package_info = serde_json::from_slice(&bytes)?;
match cache.save_package_info(&name, &package_info) { match downloader.cache.save_package_info(&name, &package_info) {
Ok(()) => { Ok(()) => {
Ok(FutureResult::SavedFsCache(Arc::new(package_info))) Ok(FutureResult::SavedFsCache(Arc::new(package_info)))
} }
@ -248,10 +241,8 @@ impl RegistryInfoDownloader {
} }
None => Ok(FutureResult::PackageNotExists), None => Ok(FutureResult::PackageNotExists),
} }
}) }
.map(|result| result.unwrap().map_err(Arc::new)) .boxed_local()
.boxed()
.shared()
} }
fn get_package_url(&self, name: &str, registry_url: &Url) -> Url { fn get_package_url(&self, name: &str, registry_url: &Url) -> Url {

View file

@ -8,8 +8,7 @@ use deno_core::anyhow::bail;
use deno_core::anyhow::Context; use deno_core::anyhow::Context;
use deno_core::error::custom_error; use deno_core::error::custom_error;
use deno_core::error::AnyError; use deno_core::error::AnyError;
use deno_core::futures::future::BoxFuture; use deno_core::futures::future::LocalBoxFuture;
use deno_core::futures::future::Shared;
use deno_core::futures::FutureExt; use deno_core::futures::FutureExt;
use deno_core::parking_lot::Mutex; use deno_core::parking_lot::Mutex;
use deno_npm::npm_rc::ResolvedNpmRc; use deno_npm::npm_rc::ResolvedNpmRc;
@ -24,6 +23,7 @@ use crate::util::progress_bar::ProgressBar;
use super::tarball_extract::verify_and_extract_tarball; use super::tarball_extract::verify_and_extract_tarball;
use super::tarball_extract::TarballExtractionMode; use super::tarball_extract::TarballExtractionMode;
use super::value_creator::MultiRuntimeAsyncValueCreator;
use super::NpmCache; use super::NpmCache;
// todo(dsherret): create seams and unit test this // todo(dsherret): create seams and unit test this
@ -31,7 +31,7 @@ use super::NpmCache;
#[derive(Debug, Clone)] #[derive(Debug, Clone)]
enum MemoryCacheItem { enum MemoryCacheItem {
/// The cache item hasn't finished yet. /// The cache item hasn't finished yet.
PendingFuture(Shared<BoxFuture<'static, Result<(), Arc<AnyError>>>>), Pending(Arc<MultiRuntimeAsyncValueCreator<()>>),
/// The result errored. /// The result errored.
Errored(Arc<AnyError>), Errored(Arc<AnyError>),
/// This package has already been cached. /// This package has already been cached.
@ -71,7 +71,7 @@ impl TarballCache {
} }
pub async fn ensure_package( pub async fn ensure_package(
&self, self: &Arc<Self>,
package: &PackageNv, package: &PackageNv,
dist: &NpmPackageVersionDistInfo, dist: &NpmPackageVersionDistInfo,
) -> Result<(), AnyError> { ) -> Result<(), AnyError> {
@ -82,69 +82,67 @@ impl TarballCache {
} }
async fn ensure_package_inner( async fn ensure_package_inner(
&self, self: &Arc<Self>,
package_nv: &PackageNv, package_nv: &PackageNv,
dist: &NpmPackageVersionDistInfo, dist: &NpmPackageVersionDistInfo,
) -> Result<(), AnyError> { ) -> Result<(), AnyError> {
let (created, cache_item) = { let cache_item = {
let mut mem_cache = self.memory_cache.lock(); let mut mem_cache = self.memory_cache.lock();
if let Some(cache_item) = mem_cache.get(package_nv) { if let Some(cache_item) = mem_cache.get(package_nv) {
(false, cache_item.clone()) cache_item.clone()
} else { } else {
let future = self.create_setup_future(package_nv.clone(), dist.clone()); let future = self.create_setup_future(package_nv.clone(), dist.clone());
let cache_item = MemoryCacheItem::PendingFuture(future); let value_creator = MultiRuntimeAsyncValueCreator::new(future);
let cache_item = MemoryCacheItem::Pending(Arc::new(value_creator));
mem_cache.insert(package_nv.clone(), cache_item.clone()); mem_cache.insert(package_nv.clone(), cache_item.clone());
(true, cache_item) cache_item
} }
}; };
match cache_item { match cache_item {
MemoryCacheItem::Cached => Ok(()), MemoryCacheItem::Cached => Ok(()),
MemoryCacheItem::Errored(err) => Err(anyhow!("{}", err)), MemoryCacheItem::Errored(err) => Err(anyhow!("{}", err)),
MemoryCacheItem::PendingFuture(future) => { MemoryCacheItem::Pending(creator) => {
if created { let tarball_cache = self.clone();
match future.await { let result = creator
Ok(_) => { .get(move || {
*self.memory_cache.lock().get_mut(package_nv).unwrap() = tarball_cache.create_setup_future(package_nv.clone(), dist.clone())
MemoryCacheItem::Cached; })
Ok(()) .await;
} match result {
Err(err) => { Ok(_) => {
let result_err = anyhow!("{}", err); *self.memory_cache.lock().get_mut(package_nv).unwrap() =
*self.memory_cache.lock().get_mut(package_nv).unwrap() = MemoryCacheItem::Cached;
MemoryCacheItem::Errored(err); Ok(())
Err(result_err) }
} Err(err) => {
let result_err = anyhow!("{}", err);
*self.memory_cache.lock().get_mut(package_nv).unwrap() =
MemoryCacheItem::Errored(err);
Err(result_err)
} }
} else {
future.await.map_err(|err| anyhow!("{}", err))
} }
} }
} }
} }
fn create_setup_future( fn create_setup_future(
&self, self: &Arc<Self>,
package_nv: PackageNv, package_nv: PackageNv,
dist: NpmPackageVersionDistInfo, dist: NpmPackageVersionDistInfo,
) -> Shared<BoxFuture<'static, Result<(), Arc<AnyError>>>> { ) -> LocalBoxFuture<'static, Result<(), AnyError>> {
let registry_url = self.npmrc.get_registry_url(&package_nv.name); let tarball_cache = self.clone();
let registry_config = async move {
self.npmrc.get_registry_config(&package_nv.name).clone(); let registry_url = tarball_cache.npmrc.get_registry_url(&package_nv.name);
let registry_config =
let cache = self.cache.clone(); tarball_cache.npmrc.get_registry_config(&package_nv.name).clone();
let fs = self.fs.clone(); let package_folder =
let progress_bar = self.progress_bar.clone(); tarball_cache.cache.package_folder_for_nv_and_url(&package_nv, registry_url);
let package_folder = let should_use_cache = tarball_cache.cache.should_use_cache_for_package(&package_nv);
cache.package_folder_for_nv_and_url(&package_nv, registry_url); let package_folder_exists = tarball_cache.fs.exists_sync(&package_folder);
let http_client_provider = self.http_client_provider.clone();
deno_core::unsync::spawn(async move {
let should_use_cache = cache.should_use_cache_for_package(&package_nv);
let package_folder_exists = fs.exists_sync(&package_folder);
if should_use_cache && package_folder_exists { if should_use_cache && package_folder_exists {
return Ok(()); return Ok(());
} else if cache.cache_setting() == &CacheSetting::Only { } else if tarball_cache.cache.cache_setting() == &CacheSetting::Only {
return Err(custom_error( return Err(custom_error(
"NotCached", "NotCached",
format!( format!(
@ -162,8 +160,9 @@ impl TarballCache {
let maybe_auth_header = let maybe_auth_header =
maybe_auth_header_for_npm_registry(&registry_config); maybe_auth_header_for_npm_registry(&registry_config);
let guard = progress_bar.update(&dist.tarball); let guard = tarball_cache.progress_bar.update(&dist.tarball);
let maybe_bytes = http_client_provider.get_or_create()? let maybe_bytes = tarball_cache.http_client_provider
.get_or_create()?
.download_with_progress(&dist.tarball, maybe_auth_header, &guard) .download_with_progress(&dist.tarball, maybe_auth_header, &guard)
.await?; .await?;
match maybe_bytes { match maybe_bytes {
@ -198,9 +197,6 @@ impl TarballCache {
bail!("Could not find npm package tarball at: {}", dist.tarball); bail!("Could not find npm package tarball at: {}", dist.tarball);
} }
} }
}) }.boxed_local()
.map(|result| result.unwrap().map_err(Arc::new))
.boxed()
.shared()
} }
} }

101
cli/npm/managed/cache/value_creator.rs vendored Normal file
View file

@ -0,0 +1,101 @@
// Copyright 2018-2024 the Deno authors. All rights reserved. MIT license.
use std::sync::Arc;
use deno_core::error::AnyError;
use deno_core::futures::future::BoxFuture;
use deno_core::futures::future::LocalBoxFuture;
use deno_core::futures::future::Shared;
use deno_core::futures::FutureExt;
use deno_core::parking_lot::Mutex;
use tokio::task::JoinError;
// todo(dsherret): unit test this
type FutureResult<TResult> = Result<TResult, Arc<AnyError>>;
type JoinResult<TResult> = Result<FutureResult<TResult>, Arc<JoinError>>;
#[derive(Debug)]
struct State<TResult> {
retry_index: usize,
future: Shared<BoxFuture<'static, JoinResult<TResult>>>,
}
/// Attempts to create a shared value asynchronously on one tokio runtime while
/// many runtimes are requesting the value.
///
/// This is only useful when the value needs to get created once across
/// many runtimes.
///
/// This handles the case where one tokio runtime goes down while another
/// one is still running.
#[derive(Debug)]
pub struct MultiRuntimeAsyncValueCreator<TResult: Send + Clone + 'static> {
state: Mutex<State<TResult>>,
}
impl<TResult: Send + Clone + 'static> MultiRuntimeAsyncValueCreator<TResult> {
pub fn new(
future: LocalBoxFuture<'static, Result<TResult, AnyError>>,
) -> Self {
Self {
state: Mutex::new(State {
retry_index: 0,
future: Self::create_shared_future(future),
}),
}
}
pub async fn get(
&self,
recreate_future: impl Fn() -> LocalBoxFuture<'static, Result<TResult, AnyError>>,
) -> Result<TResult, Arc<AnyError>> {
let (mut future, mut retry_index) = {
let state = self.state.lock();
(state.future.clone(), state.retry_index)
};
loop {
let result = future.await;
match result {
Ok(result) => return result,
Err(join_error) => {
if join_error.is_cancelled() {
let mut state = self.state.lock();
if state.retry_index == retry_index {
// we were the first one to retry, so create a new future
// that we'll run from the current runtime
state.retry_index += 1;
state.future = Self::create_shared_future(recreate_future());
}
retry_index = state.retry_index;
future = state.future.clone();
// just in case we're stuck in a loop
if retry_index > 1000 {
panic!("Something went wrong.") // should never happen
}
} else {
panic!("{}", join_error);
}
}
}
}
}
fn create_shared_future(
future: LocalBoxFuture<'static, Result<TResult, AnyError>>,
) -> Shared<BoxFuture<'static, JoinResult<TResult>>> {
deno_core::unsync::spawn(future)
.map(|result| match result {
Ok(Ok(value)) => Ok(Ok(value)),
Ok(Err(err)) => Ok(Err(Arc::new(err))),
Err(err) => Err(Arc::new(err)),
})
.boxed()
.shared()
}
}

View file

@ -105,7 +105,8 @@ impl PackageJsonDepsInstaller {
let (req, info) = result?; let (req, info) = result?;
let result = inner let result = inner
.npm_resolution .npm_resolution
.resolve_pkg_req_as_pending_with_info(req, &info); .resolve_pkg_req_as_pending_with_info(req, &info)
.await;
if let Err(err) = result { if let Err(err) = result {
if inner.npm_registry_api.mark_force_reload() { if inner.npm_registry_api.mark_force_reload() {
log::debug!("Failed to resolve package. Retrying. Error: {err:#}"); log::debug!("Failed to resolve package. Retrying. Error: {err:#}");

View file

@ -11,8 +11,9 @@ use deno_core::anyhow::Context;
use deno_core::error::AnyError; use deno_core::error::AnyError;
use deno_core::parking_lot::Mutex; use deno_core::parking_lot::Mutex;
use deno_core::serde_json; use deno_core::serde_json;
use deno_graph::NpmPackageReqResolution; use deno_graph::NpmPackageReqsResolution;
use deno_npm::npm_rc::ResolvedNpmRc; use deno_npm::npm_rc::ResolvedNpmRc;
use deno_npm::registry::NpmPackageInfo;
use deno_npm::registry::NpmRegistryApi; use deno_npm::registry::NpmRegistryApi;
use deno_npm::resolution::NpmResolutionSnapshot; use deno_npm::resolution::NpmResolutionSnapshot;
use deno_npm::resolution::PackageReqNotFoundError; use deno_npm::resolution::PackageReqNotFoundError;
@ -203,12 +204,12 @@ fn create_api(
) -> Arc<CliNpmRegistryApi> { ) -> Arc<CliNpmRegistryApi> {
Arc::new(CliNpmRegistryApi::new( Arc::new(CliNpmRegistryApi::new(
npm_cache.clone(), npm_cache.clone(),
RegistryInfoDownloader::new( Arc::new(RegistryInfoDownloader::new(
npm_cache, npm_cache,
options.http_client_provider.clone(), options.http_client_provider.clone(),
options.npmrc.clone(), options.npmrc.clone(),
options.text_only_progress_bar.clone(), options.text_only_progress_bar.clone(),
), )),
)) ))
} }
@ -387,11 +388,6 @@ impl ManagedCliNpmResolver {
self.resolution.add_package_reqs(packages).await?; self.resolution.add_package_reqs(packages).await?;
self.fs_resolver.cache_packages().await?; self.fs_resolver.cache_packages().await?;
// If there's a lock file, update it with all discovered npm packages
if let Some(lockfile) = &self.maybe_lockfile {
self.lock(&mut lockfile.lock());
}
Ok(()) Ok(())
} }
@ -418,10 +414,6 @@ impl ManagedCliNpmResolver {
.serialized_valid_snapshot_for_system(system_info) .serialized_valid_snapshot_for_system(system_info)
} }
pub fn lock(&self, lockfile: &mut Lockfile) {
self.resolution.lock(lockfile)
}
pub async fn inject_synthetic_types_node_package( pub async fn inject_synthetic_types_node_package(
&self, &self,
) -> Result<(), AnyError> { ) -> Result<(), AnyError> {
@ -442,25 +434,33 @@ impl ManagedCliNpmResolver {
self.fs_resolver.cache_packages().await self.fs_resolver.cache_packages().await
} }
/// Resolves a package requirement for deno graph. This should only be /// Resolves package requirements for deno graph.
/// called by deno_graph's NpmResolver or for resolving packages in pub async fn resolve_npm_for_deno_graph(
/// a package.json
pub fn resolve_npm_for_deno_graph(
&self, &self,
pkg_req: &PackageReq, reqs_with_pkg_infos: &[(&PackageReq, Arc<NpmPackageInfo>)],
) -> NpmPackageReqResolution { ) -> NpmPackageReqsResolution {
let result = self.resolution.resolve_pkg_req_as_pending(pkg_req); let results = self
match result { .resolution
Ok(nv) => NpmPackageReqResolution::Ok(nv), .resolve_pkg_reqs_as_pending_with_info(reqs_with_pkg_infos)
Err(err) => { .await;
if self.npm_api.mark_force_reload() {
log::debug!("Restarting npm specifier resolution to check for new registry information. Error: {:#}", err); let mut resolutions = Vec::with_capacity(results.len());
NpmPackageReqResolution::ReloadRegistryInfo(err.into()) for result in results {
} else { match result {
NpmPackageReqResolution::Err(err.into()) Ok(nv) => {
resolutions.push(Ok(nv));
}
Err(err) => {
if self.npm_api.mark_force_reload() {
log::debug!("Restarting npm specifier resolution to check for new registry information. Error: {:#}", err);
return NpmPackageReqsResolution::ReloadRegistryInfo;
} else {
resolutions.push(Err(Arc::new(err.into())));
}
} }
} }
} }
NpmPackageReqsResolution::Resolutions(resolutions)
} }
pub fn resolve_pkg_folder_from_deno_module( pub fn resolve_pkg_folder_from_deno_module(
@ -490,13 +490,12 @@ impl ManagedCliNpmResolver {
pub async fn cache_package_info( pub async fn cache_package_info(
&self, &self,
package_name: &str, package_name: &str,
) -> Result<(), AnyError> { ) -> Result<Arc<NpmPackageInfo>, AnyError> {
// this will internally cache the package information // this will internally cache the package information
self self
.npm_api .npm_api
.package_info(package_name) .package_info(package_name)
.await .await
.map(|_| ())
.map_err(|err| err.into()) .map_err(|err| err.into())
} }

View file

@ -21,14 +21,13 @@ use crate::util::sync::AtomicFlag;
use super::cache::NpmCache; use super::cache::NpmCache;
use super::cache::RegistryInfoDownloader; use super::cache::RegistryInfoDownloader;
// todo(dsherret): make this per worker
#[derive(Debug)] #[derive(Debug)]
pub struct CliNpmRegistryApi(Option<Arc<CliNpmRegistryApiInner>>); pub struct CliNpmRegistryApi(Option<Arc<CliNpmRegistryApiInner>>);
impl CliNpmRegistryApi { impl CliNpmRegistryApi {
pub fn new( pub fn new(
cache: Arc<NpmCache>, cache: Arc<NpmCache>,
registry_info_downloader: RegistryInfoDownloader, registry_info_downloader: Arc<RegistryInfoDownloader>,
) -> Self { ) -> Self {
Self(Some(Arc::new(CliNpmRegistryApiInner { Self(Some(Arc::new(CliNpmRegistryApiInner {
cache, cache,
@ -44,13 +43,6 @@ impl CliNpmRegistryApi {
self.inner().clear_memory_cache(); self.inner().clear_memory_cache();
} }
pub fn get_cached_package_info(
&self,
name: &str,
) -> Option<Arc<NpmPackageInfo>> {
self.inner().get_cached_package_info(name)
}
fn inner(&self) -> &Arc<CliNpmRegistryApiInner> { fn inner(&self) -> &Arc<CliNpmRegistryApiInner> {
// this panicking indicates a bug in the code where this // this panicking indicates a bug in the code where this
// wasn't initialized // wasn't initialized
@ -76,20 +68,7 @@ impl NpmRegistryApi for CliNpmRegistryApi {
} }
fn mark_force_reload(&self) -> bool { fn mark_force_reload(&self) -> bool {
// never force reload the registry information if reloading self.inner().mark_force_reload()
// is disabled or if we're already reloading
if matches!(
self.inner().cache.cache_setting(),
CacheSetting::Only | CacheSetting::ReloadAll
) {
return false;
}
if self.inner().force_reload_flag.raise() {
self.clear_memory_cache(); // clear the cache to force reloading
true
} else {
false
}
} }
} }
@ -108,7 +87,7 @@ struct CliNpmRegistryApiInner {
force_reload_flag: AtomicFlag, force_reload_flag: AtomicFlag,
mem_cache: Mutex<HashMap<String, CacheItem>>, mem_cache: Mutex<HashMap<String, CacheItem>>,
previously_reloaded_packages: Mutex<HashSet<String>>, previously_reloaded_packages: Mutex<HashSet<String>>,
registry_info_downloader: RegistryInfoDownloader, registry_info_downloader: Arc<RegistryInfoDownloader>,
} }
impl CliNpmRegistryApiInner { impl CliNpmRegistryApiInner {
@ -128,7 +107,7 @@ impl CliNpmRegistryApiInner {
let api = self.clone(); let api = self.clone();
let name = name.to_string(); let name = name.to_string();
async move { async move {
if (api.cache.cache_setting().should_use_for_npm_package(&name) && !api.force_reload()) if (api.cache.cache_setting().should_use_for_npm_package(&name) && !api.force_reload_flag.is_raised())
// if this has been previously reloaded, then try loading from the // if this has been previously reloaded, then try loading from the
// file system cache // file system cache
|| !api.previously_reloaded_packages.lock().insert(name.to_string()) || !api.previously_reloaded_packages.lock().insert(name.to_string())
@ -175,8 +154,21 @@ impl CliNpmRegistryApiInner {
} }
} }
fn force_reload(&self) -> bool { fn mark_force_reload(&self) -> bool {
self.force_reload_flag.is_raised() // never force reload the registry information if reloading
// is disabled or if we're already reloading
if matches!(
self.cache.cache_setting(),
CacheSetting::Only | CacheSetting::ReloadAll
) {
return false;
}
if self.force_reload_flag.raise() {
self.clear_memory_cache();
true
} else {
false
}
} }
async fn load_file_cached_package_info( async fn load_file_cached_package_info(
@ -205,16 +197,4 @@ impl CliNpmRegistryApiInner {
fn clear_memory_cache(&self) { fn clear_memory_cache(&self) {
self.mem_cache.lock().clear(); self.mem_cache.lock().clear();
} }
pub fn get_cached_package_info(
&self,
name: &str,
) -> Option<Arc<NpmPackageInfo>> {
let mem_cache = self.mem_cache.lock();
if let Some(CacheItem::Resolved(maybe_info)) = mem_cache.get(name) {
maybe_info.clone()
} else {
None
}
}
} }

View file

@ -223,34 +223,42 @@ impl NpmResolution {
/// Resolves a package requirement for deno graph. This should only be /// Resolves a package requirement for deno graph. This should only be
/// called by deno_graph's NpmResolver or for resolving packages in /// called by deno_graph's NpmResolver or for resolving packages in
/// a package.json /// a package.json
pub fn resolve_pkg_req_as_pending( pub async fn resolve_pkg_req_as_pending_with_info(
&self, &self,
pkg_req: &PackageReq, pkg_req: &PackageReq,
pkg_info: &NpmPackageInfo,
) -> Result<PackageNv, NpmPackageVersionResolutionError> { ) -> Result<PackageNv, NpmPackageVersionResolutionError> {
// we should always have this because it should have been cached before here debug_assert_eq!(pkg_req.name, pkg_info.name);
let package_info = self.api.get_cached_package_info(&pkg_req.name).unwrap(); let _permit = self.update_queue.acquire().await;
self.resolve_pkg_req_as_pending_with_info(pkg_req, &package_info)
}
/// Resolves a package requirement for deno graph. This should only be
/// called by deno_graph's NpmResolver or for resolving packages in
/// a package.json
pub fn resolve_pkg_req_as_pending_with_info(
&self,
pkg_req: &PackageReq,
package_info: &NpmPackageInfo,
) -> Result<PackageNv, NpmPackageVersionResolutionError> {
debug_assert_eq!(pkg_req.name, package_info.name);
let mut snapshot = self.snapshot.write(); let mut snapshot = self.snapshot.write();
let pending_resolver = get_npm_pending_resolver(&self.api); let pending_resolver = get_npm_pending_resolver(&self.api);
let nv = pending_resolver.resolve_package_req_as_pending( let nv = pending_resolver.resolve_package_req_as_pending(
&mut snapshot, &mut snapshot,
pkg_req, pkg_req,
package_info, pkg_info,
)?; )?;
Ok(nv) Ok(nv)
} }
pub async fn resolve_pkg_reqs_as_pending_with_info(
&self,
reqs_with_pkg_infos: &[(&PackageReq, Arc<NpmPackageInfo>)],
) -> Vec<Result<PackageNv, NpmPackageVersionResolutionError>> {
let _permit = self.update_queue.acquire().await;
let mut snapshot = self.snapshot.write();
let pending_resolver = get_npm_pending_resolver(&self.api);
let mut results = Vec::with_capacity(reqs_with_pkg_infos.len());
for (pkg_req, pkg_info) in reqs_with_pkg_infos {
debug_assert_eq!(pkg_req.name, pkg_info.name);
results.push(pending_resolver.resolve_package_req_as_pending(
&mut snapshot,
pkg_req,
pkg_info,
));
}
results
}
pub fn package_reqs(&self) -> HashMap<PackageReq, PackageNv> { pub fn package_reqs(&self) -> HashMap<PackageReq, PackageNv> {
self.snapshot.read().package_reqs().clone() self.snapshot.read().package_reqs().clone()
} }
@ -291,11 +299,6 @@ impl NpmResolution {
.read() .read()
.as_valid_serialized_for_system(system_info) .as_valid_serialized_for_system(system_info)
} }
pub fn lock(&self, lockfile: &mut Lockfile) {
let snapshot = self.snapshot.read();
populate_lockfile_from_snapshot(lockfile, &snapshot)
}
} }
async fn add_package_reqs_to_snapshot( async fn add_package_reqs_to_snapshot(
@ -370,6 +373,7 @@ fn populate_lockfile_from_snapshot(
lockfile: &mut Lockfile, lockfile: &mut Lockfile,
snapshot: &NpmResolutionSnapshot, snapshot: &NpmResolutionSnapshot,
) { ) {
assert!(!snapshot.has_pending());
for (package_req, nv) in snapshot.package_reqs() { for (package_req, nv) in snapshot.package_reqs() {
lockfile.insert_package_specifier( lockfile.insert_package_specifier(
format!("npm:{}", package_req), format!("npm:{}", package_req),

View file

@ -1,5 +1,6 @@
// Copyright 2018-2024 the Deno authors. All rights reserved. MIT license. // Copyright 2018-2024 the Deno authors. All rights reserved. MIT license.
use async_trait::async_trait;
use dashmap::DashMap; use dashmap::DashMap;
use dashmap::DashSet; use dashmap::DashSet;
use deno_ast::MediaType; use deno_ast::MediaType;
@ -9,15 +10,15 @@ use deno_core::error::AnyError;
use deno_core::futures::future; use deno_core::futures::future;
use deno_core::futures::future::LocalBoxFuture; use deno_core::futures::future::LocalBoxFuture;
use deno_core::futures::FutureExt; use deno_core::futures::FutureExt;
use deno_core::ModuleCodeString; use deno_core::ModuleSourceCode;
use deno_core::ModuleSpecifier; use deno_core::ModuleSpecifier;
use deno_graph::source::NpmPackageReqResolution;
use deno_graph::source::NpmResolver;
use deno_graph::source::ResolutionMode; use deno_graph::source::ResolutionMode;
use deno_graph::source::ResolveError; use deno_graph::source::ResolveError;
use deno_graph::source::Resolver; use deno_graph::source::Resolver;
use deno_graph::source::UnknownBuiltInNodeModuleError; use deno_graph::source::UnknownBuiltInNodeModuleError;
use deno_graph::source::DEFAULT_JSX_IMPORT_SOURCE_MODULE; use deno_graph::source::DEFAULT_JSX_IMPORT_SOURCE_MODULE;
use deno_graph::NpmPackageReqsResolution;
use deno_npm::registry::NpmPackageInfo;
use deno_runtime::deno_fs; use deno_runtime::deno_fs;
use deno_runtime::deno_fs::FileSystem; use deno_runtime::deno_fs::FileSystem;
use deno_runtime::deno_node::is_builtin_node_module; use deno_runtime::deno_node::is_builtin_node_module;
@ -34,6 +35,8 @@ use deno_semver::npm::NpmPackageReqReference;
use deno_semver::package::PackageReq; use deno_semver::package::PackageReq;
use import_map::ImportMap; use import_map::ImportMap;
use std::borrow::Cow; use std::borrow::Cow;
use std::cell::RefCell;
use std::collections::HashMap;
use std::path::Path; use std::path::Path;
use std::path::PathBuf; use std::path::PathBuf;
use std::rc::Rc; use std::rc::Rc;
@ -60,7 +63,7 @@ pub fn format_range_with_colors(range: &deno_graph::Range) -> String {
} }
pub struct ModuleCodeStringSource { pub struct ModuleCodeStringSource {
pub code: ModuleCodeString, pub code: ModuleSourceCode,
pub found_url: ModuleSpecifier, pub found_url: ModuleSpecifier,
pub media_type: MediaType, pub media_type: MediaType,
} }
@ -293,7 +296,7 @@ impl NpmModuleLoader {
let file_path = specifier.to_file_path().unwrap(); let file_path = specifier.to_file_path().unwrap();
let code = self let code = self
.fs .fs
.read_text_file_async(file_path.clone(), None) .read_file_async(file_path.clone(), None)
.await .await
.map_err(AnyError::from) .map_err(AnyError::from)
.with_context(|| { .with_context(|| {
@ -329,16 +332,20 @@ impl NpmModuleLoader {
let code = if self.cjs_resolutions.contains(specifier) { let code = if self.cjs_resolutions.contains(specifier) {
// translate cjs to esm if it's cjs and inject node globals // translate cjs to esm if it's cjs and inject node globals
self let code = String::from_utf8(code)?;
.node_code_translator ModuleSourceCode::String(
.translate_cjs_to_esm(specifier, Some(code), permissions) self
.await? .node_code_translator
.translate_cjs_to_esm(specifier, Some(code), permissions)
.await?
.into(),
)
} else { } else {
// esm and json code is untouched // esm and json code is untouched
code ModuleSourceCode::Bytes(code.into_boxed_slice().into())
}; };
Ok(ModuleCodeStringSource { Ok(ModuleCodeStringSource {
code: code.into(), code,
found_url: specifier.clone(), found_url: specifier.clone(),
media_type: MediaType::from_specifier(specifier), media_type: MediaType::from_specifier(specifier),
}) })
@ -498,8 +505,12 @@ impl CliGraphResolver {
self self
} }
pub fn as_graph_npm_resolver(&self) -> &dyn NpmResolver { pub fn create_graph_npm_resolver(&self) -> WorkerCliNpmGraphResolver {
self WorkerCliNpmGraphResolver {
npm_resolver: self.npm_resolver.as_ref(),
memory_cache: Default::default(),
bare_node_builtins_enabled: self.bare_node_builtins_enabled,
}
} }
pub fn found_package_json_dep(&self) -> bool { pub fn found_package_json_dep(&self) -> bool {
@ -782,7 +793,17 @@ fn resolve_package_json_dep(
Ok(None) Ok(None)
} }
impl NpmResolver for CliGraphResolver { #[derive(Debug)]
pub struct WorkerCliNpmGraphResolver<'a> {
npm_resolver: Option<&'a Arc<dyn CliNpmResolver>>,
// use a cache per worker so that another worker doesn't clear the
// cache of another worker
memory_cache: Rc<RefCell<HashMap<String, Arc<NpmPackageInfo>>>>,
bare_node_builtins_enabled: bool,
}
#[async_trait(?Send)]
impl<'a> deno_graph::source::NpmResolver for WorkerCliNpmGraphResolver<'a> {
fn resolve_builtin_node_module( fn resolve_builtin_node_module(
&self, &self,
specifier: &ModuleSpecifier, specifier: &ModuleSpecifier,
@ -818,17 +839,20 @@ impl NpmResolver for CliGraphResolver {
&self, &self,
package_name: &str, package_name: &str,
) -> LocalBoxFuture<'static, Result<(), AnyError>> { ) -> LocalBoxFuture<'static, Result<(), AnyError>> {
match &self.npm_resolver { match self.npm_resolver {
Some(npm_resolver) if npm_resolver.as_managed().is_some() => { Some(npm_resolver) if npm_resolver.as_managed().is_some() => {
let package_name = package_name.to_string();
let npm_resolver = npm_resolver.clone(); let npm_resolver = npm_resolver.clone();
let package_name = package_name.to_string();
let memory_cache = self.memory_cache.clone();
async move { async move {
if let Some(managed) = npm_resolver.as_managed() { if let Some(managed) = npm_resolver.as_managed() {
managed.cache_package_info(&package_name).await?; let package_info =
managed.cache_package_info(&package_name).await?;
memory_cache.borrow_mut().insert(package_name, package_info);
} }
Ok(()) Ok(())
} }
.boxed() .boxed_local()
} }
_ => { _ => {
// return it succeeded and error at the import site below // return it succeeded and error at the import site below
@ -837,19 +861,50 @@ impl NpmResolver for CliGraphResolver {
} }
} }
fn resolve_npm(&self, package_req: &PackageReq) -> NpmPackageReqResolution { async fn resolve_pkg_reqs(
&self,
package_reqs: &[&PackageReq],
) -> NpmPackageReqsResolution {
match &self.npm_resolver { match &self.npm_resolver {
Some(npm_resolver) => match npm_resolver.as_inner() { Some(npm_resolver) => {
InnerCliNpmResolverRef::Managed(npm_resolver) => { let npm_resolver = match npm_resolver.as_inner() {
npm_resolver.resolve_npm_for_deno_graph(package_req) InnerCliNpmResolverRef::Managed(npm_resolver) => npm_resolver,
// if we are using byonm, then this should never be called because
// we don't use deno_graph's npm resolution in this case
InnerCliNpmResolverRef::Byonm(_) => unreachable!(),
};
let reqs_with_package_infos = {
let memory_cache = self.memory_cache.borrow();
package_reqs
.iter()
.map(|package_req| {
let package_info = memory_cache
.get(&package_req.name)
.unwrap() // ok because deno_graph would have called load_and_cache_npm_package_info
.clone();
(*package_req, package_info)
})
.collect::<Vec<_>>()
};
let result = npm_resolver
.resolve_npm_for_deno_graph(&reqs_with_package_infos)
.await;
if matches!(result, NpmPackageReqsResolution::ReloadRegistryInfo) {
self.memory_cache.borrow_mut().clear();
} }
// if we are using byonm, then this should never be called because result
// we don't use deno_graph's npm resolution in this case }
InnerCliNpmResolverRef::Byonm(_) => unreachable!(), None => NpmPackageReqsResolution::Resolutions(
}, package_reqs
None => NpmPackageReqResolution::Err(anyhow!( .iter()
"npm specifiers were requested; but --no-npm is specified" .map(|_| {
)), Err(Arc::new(anyhow!(
"npm specifiers were requested; but --no-npm is specified"
)))
})
.collect(),
),
} }
} }

View file

@ -218,7 +218,7 @@ impl ModuleLoader for EmbeddedModuleLoader {
MediaType::Json => ModuleType::Json, MediaType::Json => ModuleType::Json,
_ => ModuleType::JavaScript, _ => ModuleType::JavaScript,
}, },
ModuleSourceCode::String(code_source.code), code_source.code,
&original_specifier, &original_specifier,
&code_source.found_url, &code_source.found_url,
None, None,

View file

@ -506,14 +506,14 @@ pub async fn run_benchmarks_with_watch(
let bench_modules_to_reload = if let Some(changed_paths) = changed_paths let bench_modules_to_reload = if let Some(changed_paths) = changed_paths
{ {
let changed_paths = changed_paths.into_iter().collect::<HashSet<_>>(); let changed_paths = changed_paths.into_iter().collect::<HashSet<_>>();
let mut result = Vec::new(); let mut result = IndexSet::with_capacity(bench_modules.len());
for bench_module_specifier in bench_modules { for bench_module_specifier in bench_modules {
if has_graph_root_local_dependent_changed( if has_graph_root_local_dependent_changed(
&graph, &graph,
bench_module_specifier, bench_module_specifier,
&changed_paths, &changed_paths,
) { ) {
result.push(bench_module_specifier.clone()); result.insert(bench_module_specifier.clone());
} }
} }
result result

View file

@ -25,7 +25,6 @@ use deno_core::serde_json;
use deno_core::sourcemap::SourceMap; use deno_core::sourcemap::SourceMap;
use deno_core::url::Url; use deno_core::url::Url;
use deno_core::LocalInspectorSession; use deno_core::LocalInspectorSession;
use deno_core::ModuleCodeString;
use regex::Regex; use regex::Regex;
use std::fs; use std::fs;
use std::fs::File; use std::fs::File;
@ -565,7 +564,7 @@ pub async fn cover_files(
| MediaType::Cjs | MediaType::Cjs
| MediaType::Mjs | MediaType::Mjs
| MediaType::Json => None, | MediaType::Json => None,
MediaType::Dts | MediaType::Dmts | MediaType::Dcts => Some(String::new()), MediaType::Dts | MediaType::Dmts | MediaType::Dcts => Some(Vec::new()),
MediaType::TypeScript MediaType::TypeScript
| MediaType::Jsx | MediaType::Jsx
| MediaType::Mts | MediaType::Mts
@ -586,11 +585,13 @@ pub async fn cover_files(
unreachable!() unreachable!()
} }
}; };
let runtime_code: ModuleCodeString = transpiled_code let runtime_code: String = match transpiled_code {
.map(|c| c.into()) Some(code) => String::from_utf8(code)
.unwrap_or_else(|| original_source.clone().into()); .with_context(|| format!("Failed decoding {}", file.specifier))?,
None => original_source.to_string(),
};
let source_map = source_map_from_code(&runtime_code); let source_map = source_map_from_code(runtime_code.as_bytes());
let coverage_report = generate_coverage_report( let coverage_report = generate_coverage_report(
&script_coverage, &script_coverage,
runtime_code.as_str().to_owned(), runtime_code.as_str().to_owned(),

View file

@ -350,7 +350,7 @@ fn check_diagnostics(diagnostics: &[DocDiagnostic]) -> Result<(), AnyError> {
for (_, diagnostics_by_col) in diagnostics_by_lc { for (_, diagnostics_by_col) in diagnostics_by_lc {
for (_, diagnostics) in diagnostics_by_col { for (_, diagnostics) in diagnostics_by_col {
for diagnostic in diagnostics { for diagnostic in diagnostics {
log::error!("{}", diagnostic.display()); log::error!("{}\n", diagnostic.display());
} }
} }
} }

View file

@ -211,7 +211,7 @@ fn format_markdown(
codeblock_config.line_width = line_width; codeblock_config.line_width = line_width;
dprint_plugin_typescript::format_text( dprint_plugin_typescript::format_text(
&fake_filename, &fake_filename,
text, text.to_string(),
&codeblock_config, &codeblock_config,
) )
} }
@ -255,7 +255,11 @@ pub fn format_file(
), ),
_ => { _ => {
let config = get_resolved_typescript_config(fmt_options); let config = get_resolved_typescript_config(fmt_options);
dprint_plugin_typescript::format_text(file_path, file_text, &config) dprint_plugin_typescript::format_text(
file_path,
file_text.to_string(),
&config,
)
} }
} }
} }

View file

@ -244,7 +244,7 @@ async fn lint_files(
incremental_cache.update_file( incremental_cache.update_file(
&file_path, &file_path,
// ensure the returned text is used here as it may have been modified via --fix // ensure the returned text is used here as it may have been modified via --fix
file_source.text_info().text_str(), file_source.text(),
) )
} }
} }
@ -396,7 +396,7 @@ fn lint_file_and_fix(
media_type, media_type,
linter, linter,
config.clone(), config.clone(),
source.text_info(), source.text_info_lazy(),
&diagnostics, &diagnostics,
)?; )?;
match change { match change {
@ -423,7 +423,7 @@ fn lint_file_and_fix(
if fix_iterations > 0 { if fix_iterations > 0 {
// everything looks good and the file still parses, so write it out // everything looks good and the file still parses, so write it out
fs::write(file_path, source.text_info().text_str()) fs::write(file_path, source.text().as_ref())
.context("Failed writing fix to file.")?; .context("Failed writing fix to file.")?;
} }
@ -668,7 +668,7 @@ impl LintReporter for PrettyLintReporter {
} }
} }
log::error!("{}", d.display()); log::error!("{}\n", d.display());
} }
fn visit_error(&mut self, file_path: &str, err: &AnyError) { fn visit_error(&mut self, file_path: &str, err: &AnyError) {

View file

@ -40,11 +40,7 @@ impl PublishDiagnosticsCollector {
diagnostics.sort_by_cached_key(|d| d.sorting_key()); diagnostics.sort_by_cached_key(|d| d.sorting_key());
for diagnostic in diagnostics { for diagnostic in diagnostics {
// todo(https://github.com/denoland/deno_ast/issues/245): use log crate here log::error!("{}", diagnostic.display());
#[allow(clippy::print_stderr)]
{
eprint!("{}", diagnostic.display());
}
if matches!(diagnostic.level(), DiagnosticLevel::Error) { if matches!(diagnostic.level(), DiagnosticLevel::Error) {
errors += 1; errors += 1;
} }
@ -287,7 +283,7 @@ impl Diagnostic for PublishDiagnostic {
Some(DiagnosticSnippet { Some(DiagnosticSnippet {
source: Cow::Borrowed(text_info), source: Cow::Borrowed(text_info),
highlight: DiagnosticSnippetHighlight { highlights: vec![DiagnosticSnippetHighlight {
style: DiagnosticSnippetHighlightStyle::Error, style: DiagnosticSnippetHighlightStyle::Error,
range: DiagnosticSourceRange { range: DiagnosticSourceRange {
start: DiagnosticSourcePos::LineAndCol { start: DiagnosticSourcePos::LineAndCol {
@ -300,7 +296,7 @@ impl Diagnostic for PublishDiagnostic {
}, },
}, },
description: Some("the specifier".into()), description: Some("the specifier".into()),
}, }],
}) })
} }
@ -314,14 +310,14 @@ impl Diagnostic for PublishDiagnostic {
.. ..
} => Some(DiagnosticSnippet { } => Some(DiagnosticSnippet {
source: Cow::Borrowed(text_info), source: Cow::Borrowed(text_info),
highlight: DiagnosticSnippetHighlight { highlights: vec![DiagnosticSnippetHighlight {
style: DiagnosticSnippetHighlightStyle::Warning, style: DiagnosticSnippetHighlightStyle::Warning,
range: DiagnosticSourceRange { range: DiagnosticSourceRange {
start: DiagnosticSourcePos::SourcePos(range.start), start: DiagnosticSourcePos::SourcePos(range.start),
end: DiagnosticSourcePos::SourcePos(range.end), end: DiagnosticSourcePos::SourcePos(range.end),
}, },
description: Some("the unanalyzable dynamic import".into()), description: Some("the unanalyzable dynamic import".into()),
}, }],
}), }),
}, },
InvalidPath { .. } => None, InvalidPath { .. } => None,
@ -343,14 +339,14 @@ impl Diagnostic for PublishDiagnostic {
range, text_info, .. range, text_info, ..
} => Some(DiagnosticSnippet { } => Some(DiagnosticSnippet {
source: Cow::Borrowed(text_info), source: Cow::Borrowed(text_info),
highlight: DiagnosticSnippetHighlight { highlights: vec![DiagnosticSnippetHighlight {
style: DiagnosticSnippetHighlightStyle::Error, style: DiagnosticSnippetHighlightStyle::Error,
range: DiagnosticSourceRange { range: DiagnosticSourceRange {
start: DiagnosticSourcePos::SourcePos(range.start), start: DiagnosticSourcePos::SourcePos(range.start),
end: DiagnosticSourcePos::SourcePos(range.end), end: DiagnosticSourcePos::SourcePos(range.end),
}, },
description: Some("the triple slash directive".into()), description: Some("the triple slash directive".into()),
}, }],
}), }),
} }
} }
@ -398,14 +394,14 @@ impl Diagnostic for PublishDiagnostic {
let end = replacement.line_end(0); let end = replacement.line_end(0);
Some(DiagnosticSnippet { Some(DiagnosticSnippet {
source: Cow::Owned(replacement), source: Cow::Owned(replacement),
highlight: DiagnosticSnippetHighlight { highlights: vec![DiagnosticSnippetHighlight {
style: DiagnosticSnippetHighlightStyle::Hint, style: DiagnosticSnippetHighlightStyle::Hint,
range: DiagnosticSourceRange { range: DiagnosticSourceRange {
start: DiagnosticSourcePos::SourcePos(start), start: DiagnosticSourcePos::SourcePos(start),
end: DiagnosticSourcePos::SourcePos(end), end: DiagnosticSourcePos::SourcePos(end),
}, },
description: Some("try this specifier".into()), description: Some("try this specifier".into()),
}, }],
}) })
} }
None => None, None => None,

View file

@ -130,7 +130,7 @@ impl GraphDiagnosticsCollector {
prefer_fast_check_graph: false, prefer_fast_check_graph: false,
follow_type_only: true, follow_type_only: true,
}; };
let mut iter = graph.walk(&graph.roots, options); let mut iter = graph.walk(graph.roots.iter(), options);
while let Some((specifier, entry)) = iter.next() { while let Some((specifier, entry)) = iter.next() {
if skip_specifiers.contains(specifier) { if skip_specifiers.contains(specifier) {
iter.skip_previous_dependencies(); iter.skip_previous_dependencies();
@ -196,7 +196,7 @@ fn check_for_banned_triple_slash_directives(
PublishDiagnostic::BannedTripleSlashDirectives { PublishDiagnostic::BannedTripleSlashDirectives {
specifier: parsed_source.specifier().clone(), specifier: parsed_source.specifier().clone(),
range: comment.range(), range: comment.range(),
text_info: parsed_source.text_info().clone(), text_info: parsed_source.text_info_lazy().clone(),
}, },
); );
} }

View file

@ -126,7 +126,7 @@ fn resolve_content_maybe_unfurling(
let text = String::from_utf8(data)?; let text = String::from_utf8(data)?;
deno_ast::parse_module(deno_ast::ParseParams { deno_ast::parse_module(deno_ast::ParseParams {
specifier: specifier.clone(), specifier: specifier.clone(),
text_info: deno_ast::SourceTextInfo::from_string(text), text: text.into(),
media_type, media_type,
capture_tokens: false, capture_tokens: false,
maybe_syntax: None, maybe_syntax: None,

View file

@ -121,16 +121,15 @@ impl<'a> SpecifierUnfurler<'a> {
fn try_unfurl_dynamic_dep( fn try_unfurl_dynamic_dep(
&self, &self,
module_url: &lsp_types::Url, module_url: &lsp_types::Url,
parsed_source: &ParsedSource, text_info: &SourceTextInfo,
dep: &deno_graph::DynamicDependencyDescriptor, dep: &deno_graph::DynamicDependencyDescriptor,
text_changes: &mut Vec<deno_ast::TextChange>, text_changes: &mut Vec<deno_ast::TextChange>,
) -> bool { ) -> bool {
match &dep.argument { match &dep.argument {
deno_graph::DynamicArgument::String(specifier) => { deno_graph::DynamicArgument::String(specifier) => {
let range = to_range(parsed_source, &dep.argument_range); let range = to_range(text_info, &dep.argument_range);
let maybe_relative_index = parsed_source.text_info().text_str() let maybe_relative_index =
[range.start..range.end] text_info.text_str()[range.start..range.end].find(specifier);
.find(specifier);
let Some(relative_index) = maybe_relative_index else { let Some(relative_index) = maybe_relative_index else {
return true; // always say it's analyzable for a string return true; // always say it's analyzable for a string
}; };
@ -159,9 +158,9 @@ impl<'a> SpecifierUnfurler<'a> {
let Some(unfurled) = unfurled else { let Some(unfurled) = unfurled else {
return true; // nothing to unfurl return true; // nothing to unfurl
}; };
let range = to_range(parsed_source, &dep.argument_range); let range = to_range(text_info, &dep.argument_range);
let maybe_relative_index = let maybe_relative_index =
parsed_source.text_info().text_str()[range.start..].find(specifier); text_info.text_str()[range.start..].find(specifier);
let Some(relative_index) = maybe_relative_index else { let Some(relative_index) = maybe_relative_index else {
return false; return false;
}; };
@ -192,6 +191,7 @@ impl<'a> SpecifierUnfurler<'a> {
diagnostic_reporter: &mut dyn FnMut(SpecifierUnfurlerDiagnostic), diagnostic_reporter: &mut dyn FnMut(SpecifierUnfurlerDiagnostic),
) -> String { ) -> String {
let mut text_changes = Vec::new(); let mut text_changes = Vec::new();
let text_info = parsed_source.text_info_lazy();
let module_info = ParserModuleAnalyzer::module_info(parsed_source); let module_info = ParserModuleAnalyzer::module_info(parsed_source);
let analyze_specifier = let analyze_specifier =
|specifier: &str, |specifier: &str,
@ -199,7 +199,7 @@ impl<'a> SpecifierUnfurler<'a> {
text_changes: &mut Vec<deno_ast::TextChange>| { text_changes: &mut Vec<deno_ast::TextChange>| {
if let Some(unfurled) = self.unfurl_specifier(url, specifier) { if let Some(unfurled) = self.unfurl_specifier(url, specifier) {
text_changes.push(deno_ast::TextChange { text_changes.push(deno_ast::TextChange {
range: to_range(parsed_source, range), range: to_range(text_info, range),
new_text: unfurled, new_text: unfurled,
}); });
} }
@ -214,27 +214,19 @@ impl<'a> SpecifierUnfurler<'a> {
); );
} }
DependencyDescriptor::Dynamic(dep) => { DependencyDescriptor::Dynamic(dep) => {
let success = self.try_unfurl_dynamic_dep( let success =
url, self.try_unfurl_dynamic_dep(url, text_info, dep, &mut text_changes);
parsed_source,
dep,
&mut text_changes,
);
if !success { if !success {
let start_pos = parsed_source let start_pos = text_info.line_start(dep.argument_range.start.line)
.text_info()
.line_start(dep.argument_range.start.line)
+ dep.argument_range.start.character; + dep.argument_range.start.character;
let end_pos = parsed_source let end_pos = text_info.line_start(dep.argument_range.end.line)
.text_info()
.line_start(dep.argument_range.end.line)
+ dep.argument_range.end.character; + dep.argument_range.end.character;
diagnostic_reporter( diagnostic_reporter(
SpecifierUnfurlerDiagnostic::UnanalyzableDynamicImport { SpecifierUnfurlerDiagnostic::UnanalyzableDynamicImport {
specifier: url.to_owned(), specifier: url.to_owned(),
range: SourceRange::new(start_pos, end_pos), range: SourceRange::new(start_pos, end_pos),
text_info: parsed_source.text_info().clone(), text_info: text_info.clone(),
}, },
); );
} }
@ -267,10 +259,8 @@ impl<'a> SpecifierUnfurler<'a> {
); );
} }
let rewritten_text = deno_ast::apply_text_changes( let rewritten_text =
parsed_source.text_info().text_str(), deno_ast::apply_text_changes(text_info.text_str(), text_changes);
text_changes,
);
rewritten_text rewritten_text
} }
} }
@ -295,13 +285,13 @@ fn relative_url(
} }
fn to_range( fn to_range(
parsed_source: &ParsedSource, text_info: &SourceTextInfo,
range: &deno_graph::PositionRange, range: &deno_graph::PositionRange,
) -> std::ops::Range<usize> { ) -> std::ops::Range<usize> {
let mut range = range let mut range = range
.as_source_range(parsed_source.text_info()) .as_source_range(text_info)
.as_byte_range(parsed_source.text_info().range().start); .as_byte_range(text_info.range().start);
let text = &parsed_source.text_info().text_str()[range.clone()]; let text = &text_info.text_str()[range.clone()];
if text.starts_with('"') || text.starts_with('\'') { if text.starts_with('"') || text.starts_with('\'') {
range.start += 1; range.start += 1;
} }
@ -338,7 +328,7 @@ mod tests {
capture_tokens: false, capture_tokens: false,
maybe_syntax: None, maybe_syntax: None,
scope_analysis: false, scope_analysis: false,
text_info: deno_ast::SourceTextInfo::new(source_code.into()), text: source_code.into(),
}) })
.unwrap() .unwrap()
} }

View file

@ -639,10 +639,11 @@ impl ReplSession {
source_map: deno_ast::SourceMapOption::None, source_map: deno_ast::SourceMapOption::None,
source_map_file: None, source_map_file: None,
inline_sources: false, inline_sources: false,
keep_comments: false, remove_comments: false,
}, },
)? )?
.into_source() .into_source()
.into_string()?
.text; .text;
let value = self let value = self
@ -817,7 +818,7 @@ fn parse_source_as(
let parsed = deno_ast::parse_module(deno_ast::ParseParams { let parsed = deno_ast::parse_module(deno_ast::ParseParams {
specifier, specifier,
text_info: deno_ast::SourceTextInfo::from_string(source), text: source.into(),
media_type, media_type,
capture_tokens: true, capture_tokens: true,
maybe_syntax: None, maybe_syntax: None,
@ -884,7 +885,7 @@ fn analyze_jsx_pragmas(
range: comment_source_to_position_range( range: comment_source_to_position_range(
c.start(), c.start(),
&m, &m,
parsed_source.text_info(), parsed_source.text_info_lazy(),
true, true,
), ),
}); });
@ -898,7 +899,7 @@ fn analyze_jsx_pragmas(
range: comment_source_to_position_range( range: comment_source_to_position_range(
c.start(), c.start(),
&m, &m,
parsed_source.text_info(), parsed_source.text_info_lazy(),
false, false,
), ),
}); });
@ -912,7 +913,7 @@ fn analyze_jsx_pragmas(
range: comment_source_to_position_range( range: comment_source_to_position_range(
c.start(), c.start(),
&m, &m,
parsed_source.text_info(), parsed_source.text_info_lazy(),
false, false,
), ),
}); });

View file

@ -1249,7 +1249,7 @@ fn extract_files_from_source_comments(
) -> Result<Vec<File>, AnyError> { ) -> Result<Vec<File>, AnyError> {
let parsed_source = deno_ast::parse_module(deno_ast::ParseParams { let parsed_source = deno_ast::parse_module(deno_ast::ParseParams {
specifier: specifier.clone(), specifier: specifier.clone(),
text_info: deno_ast::SourceTextInfo::new(source), text: source,
media_type, media_type,
capture_tokens: false, capture_tokens: false,
maybe_syntax: None, maybe_syntax: None,
@ -1273,7 +1273,7 @@ fn extract_files_from_source_comments(
specifier, specifier,
&comment.text, &comment.text,
media_type, media_type,
parsed_source.text_info().line_index(comment.start()), parsed_source.text_info_lazy().line_index(comment.start()),
blocks_regex, blocks_regex,
lines_regex, lines_regex,
) )
@ -1877,7 +1877,7 @@ pub async fn run_tests_with_watch(
let test_modules_to_reload = if let Some(changed_paths) = changed_paths let test_modules_to_reload = if let Some(changed_paths) = changed_paths
{ {
let mut result = Vec::new(); let mut result = IndexSet::with_capacity(test_modules.len());
let changed_paths = changed_paths.into_iter().collect::<HashSet<_>>(); let changed_paths = changed_paths.into_iter().collect::<HashSet<_>>();
for test_module_specifier in test_modules { for test_module_specifier in test_modules {
if has_graph_root_local_dependent_changed( if has_graph_root_local_dependent_changed(
@ -1885,7 +1885,7 @@ pub async fn run_tests_with_watch(
test_module_specifier, test_module_specifier,
&changed_paths, &changed_paths,
) { ) {
result.push(test_module_specifier.clone()); result.insert(test_module_specifier.clone());
} }
} }
result result

View file

@ -64,7 +64,6 @@ mod test {
use deno_ast::ModuleSpecifier; use deno_ast::ModuleSpecifier;
use deno_ast::ParseParams; use deno_ast::ParseParams;
use deno_ast::ParsedSource; use deno_ast::ParsedSource;
use deno_ast::SourceTextInfo;
use super::has_default_export; use super::has_default_export;
@ -107,7 +106,7 @@ mod test {
maybe_syntax: None, maybe_syntax: None,
media_type: MediaType::TypeScript, media_type: MediaType::TypeScript,
scope_analysis: false, scope_analysis: false,
text_info: SourceTextInfo::from_string(text.to_string()), text: text.into(),
}) })
.unwrap() .unwrap()
} }

View file

@ -118,7 +118,7 @@ pub async fn build<
graph_util::graph_valid( graph_util::graph_valid(
&graph, &graph,
&real_fs, &real_fs,
&graph.roots, &graph.roots.iter().cloned().collect::<Vec<_>>(),
graph_util::GraphValidOptions { graph_util::GraphValidOptions {
is_vendoring: true, is_vendoring: true,
check_js: true, check_js: true,

View file

@ -250,7 +250,7 @@ fn visit_modules(
let parsed_source = let parsed_source =
parsed_source_cache.get_parsed_source_from_js_module(module)?; parsed_source_cache.get_parsed_source_from_js_module(module)?;
let text_info = parsed_source.text_info().clone(); let text_info = parsed_source.text_info_lazy().clone();
for dep in module.dependencies.values() { for dep in module.dependencies.values() {
visit_resolution( visit_resolution(

View file

@ -1,44 +1,132 @@
// Copyright 2018-2024 the Deno authors. All rights reserved. MIT license. // Copyright 2018-2024 the Deno authors. All rights reserved. MIT license.
use std::ops::Range;
use base64::prelude::BASE64_STANDARD; use base64::prelude::BASE64_STANDARD;
use base64::Engine; use base64::Engine;
use deno_core::ModuleCodeString; use deno_core::ModuleSourceCode;
static SOURCE_MAP_PREFIX: &[u8] = static SOURCE_MAP_PREFIX: &[u8] =
b"//# sourceMappingURL=data:application/json;base64,"; b"//# sourceMappingURL=data:application/json;base64,";
pub fn source_map_from_code(code: &ModuleCodeString) -> Option<Vec<u8>> { pub fn source_map_from_code(code: &[u8]) -> Option<Vec<u8>> {
let bytes = code.as_bytes(); let range = find_source_map_range(code)?;
let last_line = bytes.rsplit(|u| *u == b'\n').next()?; let source_map_range = &code[range];
if last_line.starts_with(SOURCE_MAP_PREFIX) { let input = source_map_range.split_at(SOURCE_MAP_PREFIX.len()).1;
let input = last_line.split_at(SOURCE_MAP_PREFIX.len()).1; let decoded_map = BASE64_STANDARD.decode(input).ok()?;
let decoded_map = BASE64_STANDARD Some(decoded_map)
.decode(input) }
.expect("Unable to decode source map from emitted file.");
Some(decoded_map) /// Truncate the source code before the source map.
pub fn code_without_source_map(code: ModuleSourceCode) -> ModuleSourceCode {
use deno_core::ModuleCodeBytes;
match code {
ModuleSourceCode::String(mut code) => {
if let Some(range) = find_source_map_range(code.as_bytes()) {
code.truncate(range.start);
}
ModuleSourceCode::String(code)
}
ModuleSourceCode::Bytes(code) => {
if let Some(range) = find_source_map_range(code.as_bytes()) {
let source_map_index = range.start;
ModuleSourceCode::Bytes(match code {
ModuleCodeBytes::Static(bytes) => {
ModuleCodeBytes::Static(&bytes[..source_map_index])
}
ModuleCodeBytes::Boxed(bytes) => {
// todo(dsherret): should be possible without cloning
ModuleCodeBytes::Boxed(
bytes[..source_map_index].to_vec().into_boxed_slice(),
)
}
ModuleCodeBytes::Arc(bytes) => ModuleCodeBytes::Boxed(
bytes[..source_map_index].to_vec().into_boxed_slice(),
),
})
} else {
ModuleSourceCode::Bytes(code)
}
}
}
}
fn find_source_map_range(code: &[u8]) -> Option<Range<usize>> {
fn last_non_blank_line_range(code: &[u8]) -> Option<Range<usize>> {
let mut hit_non_whitespace = false;
let mut range_end = code.len();
for i in (0..code.len()).rev() {
match code[i] {
b' ' | b'\t' => {
if !hit_non_whitespace {
range_end = i;
}
}
b'\n' | b'\r' => {
if hit_non_whitespace {
return Some(i + 1..range_end);
}
range_end = i;
}
_ => {
hit_non_whitespace = true;
}
}
}
None
}
let range = last_non_blank_line_range(code)?;
if code[range.start..range.end].starts_with(SOURCE_MAP_PREFIX) {
Some(range)
} else { } else {
None None
} }
} }
/// Truncate the source code before the source map.
pub fn code_without_source_map(mut code: ModuleCodeString) -> ModuleCodeString {
let bytes = code.as_bytes();
for i in (0..bytes.len()).rev() {
if bytes[i] == b'\n' {
if bytes[i + 1..].starts_with(SOURCE_MAP_PREFIX) {
code.truncate(i + 1);
}
return code;
}
}
code
}
#[cfg(test)] #[cfg(test)]
mod tests { mod tests {
use std::sync::Arc;
use deno_core::ModuleCodeBytes;
use deno_core::ModuleCodeString;
use super::*; use super::*;
#[test]
fn test_source_map_from_code() {
let to_string =
|bytes: Vec<u8>| -> String { String::from_utf8(bytes).unwrap() };
assert_eq!(
source_map_from_code(
b"test\n//# sourceMappingURL=data:application/json;base64,dGVzdGluZ3Rlc3Rpbmc=",
).map(to_string),
Some("testingtesting".to_string())
);
assert_eq!(
source_map_from_code(
b"test\n//# sourceMappingURL=data:application/json;base64,dGVzdGluZ3Rlc3Rpbmc=\n \n",
).map(to_string),
Some("testingtesting".to_string())
);
assert_eq!(
source_map_from_code(
b"test\n//# sourceMappingURL=data:application/json;base64,dGVzdGluZ3Rlc3Rpbmc=\n test\n",
),
None
);
assert_eq!(
source_map_from_code(
b"\"use strict\";
throw new Error(\"Hello world!\");
//# sourceMappingURL=data:application/json;base64,{",
),
None
);
}
#[test] #[test]
fn test_source_without_source_map() { fn test_source_without_source_map() {
run_test("", ""); run_test("", "");
@ -62,14 +150,39 @@ mod tests {
"\n//# sourceMappingURL=data:application/json;base64,test", "\n//# sourceMappingURL=data:application/json;base64,test",
"\n", "\n",
); );
run_test(
"test\n//# sourceMappingURL=data:application/json;base64,test\n\n",
"test\n",
);
run_test(
"test\n//# sourceMappingURL=data:application/json;base64,test\n \n ",
"test\n",
);
fn run_test(input: &'static str, output: &'static str) { fn run_test(input: &'static str, output: &'static str) {
assert_eq!( let forms = [
code_without_source_map(ModuleCodeString::from_static(input)) ModuleSourceCode::String(ModuleCodeString::from_static(input)),
.as_str() ModuleSourceCode::String({
.to_owned(), let text: Arc<str> = input.into();
output text.into()
); }),
ModuleSourceCode::String({
let text: String = input.into();
text.into()
}),
ModuleSourceCode::Bytes(ModuleCodeBytes::Static(input.as_bytes())),
ModuleSourceCode::Bytes(ModuleCodeBytes::Boxed(
input.as_bytes().to_vec().into_boxed_slice(),
)),
ModuleSourceCode::Bytes(ModuleCodeBytes::Arc(
input.as_bytes().to_vec().into(),
)),
];
for form in forms {
let result = code_without_source_map(form);
let bytes = result.as_bytes();
assert_eq!(bytes, output.as_bytes());
}
} }
} }
} }

View file

@ -4,7 +4,6 @@
use deno_ast::MediaType; use deno_ast::MediaType;
use deno_ast::ParseParams; use deno_ast::ParseParams;
use deno_ast::SourceMapOption; use deno_ast::SourceMapOption;
use deno_ast::SourceTextInfo;
use deno_core::error::AnyError; use deno_core::error::AnyError;
use deno_core::extension; use deno_core::extension;
use deno_core::Extension; use deno_core::Extension;
@ -88,7 +87,7 @@ pub fn maybe_transpile_source(
let parsed = deno_ast::parse_module(ParseParams { let parsed = deno_ast::parse_module(ParseParams {
specifier: deno_core::url::Url::parse(&name).unwrap(), specifier: deno_core::url::Url::parse(&name).unwrap(),
text_info: SourceTextInfo::from_string(source.as_str().to_owned()), text: source.into(),
media_type, media_type,
capture_tokens: false, capture_tokens: false,
scope_analysis: false, scope_analysis: false,
@ -111,9 +110,9 @@ pub fn maybe_transpile_source(
)? )?
.into_source(); .into_source();
let maybe_source_map: Option<SourceMapData> = transpiled_source let maybe_source_map: Option<SourceMapData> =
.source_map transpiled_source.source_map.map(|sm| sm.into());
.map(|sm| sm.into_bytes().into()); let source_text = String::from_utf8(transpiled_source.source)?;
Ok((transpiled_source.text.into(), maybe_source_map)) Ok((source_text.into(), maybe_source_map))
} }

View file

@ -43,7 +43,7 @@ deno_lockfile.workspace = true
deno_terminal.workspace = true deno_terminal.workspace = true
deno_tls.workspace = true deno_tls.workspace = true
fastwebsockets = { workspace = true, features = ["upgrade", "unstable-split"] } fastwebsockets = { workspace = true, features = ["upgrade", "unstable-split"] }
file_test_runner = "0.7.0" file_test_runner = "0.7.2"
flaky_test = "=0.1.0" flaky_test = "=0.1.0"
http.workspace = true http.workspace = true
http-body-util.workspace = true http-body-util.workspace = true

View file

@ -1725,7 +1725,7 @@ globalThis.state = { i: 0 };
function bar() { function bar() {
globalThis.state.i = 0; globalThis.state.i = 0;
console.log("got request1", globalThis.state.i); console.error("got request1", globalThis.state.i);
} }
function handler(_req) { function handler(_req) {
@ -1738,8 +1738,14 @@ console.log("Listening...")
"#, "#,
); );
wait_contains("Failed to reload module", &mut stderr_lines).await; wait_contains("Replaced changed module", &mut stderr_lines).await;
wait_contains("File change detected", &mut stderr_lines).await; util::deno_cmd()
.current_dir(t.path())
.arg("eval")
.arg("await fetch('http://localhost:11111');")
.spawn()
.unwrap();
wait_contains("got request1", &mut stderr_lines).await;
check_alive_then_kill(child); check_alive_then_kill(child);
} }

View file

@ -76,9 +76,7 @@ Or if you want to run several tests at the same time:
### Top level properties ### Top level properties
- `base` - The base config to use for the test. Options: - `repeat` (number) - Number of times to repeat a test.
- `jsr` - Uses env vars for jsr.
- `npm` - Uses env vars for npm.
- `tempDir` (boolean) - Copy all the non-test files to a temporary directory and - `tempDir` (boolean) - Copy all the non-test files to a temporary directory and
execute the command in that temporary directory. execute the command in that temporary directory.
- By default, tests are executed with a current working directory of the test, - By default, tests are executed with a current working directory of the test,

View file

@ -3,6 +3,7 @@ error[no-slow-types]: missing explicit return type in the public API
| |
1 | export function add(a: number, b: number) { 1 | export function add(a: number, b: number) {
| ^^^ this function is missing an explicit return type | ^^^ this function is missing an explicit return type
|
= hint: add an explicit return type to the function = hint: add an explicit return type to the function
info: all functions in the public API must have an explicit return type info: all functions in the public API must have an explicit return type
@ -14,6 +15,7 @@ error[no-slow-types]: missing explicit return type in the public API
| |
1 | export function addB(a: number, b: number) { 1 | export function addB(a: number, b: number) {
| ^^^^ this function is missing an explicit return type | ^^^^ this function is missing an explicit return type
|
= hint: add an explicit return type to the function = hint: add an explicit return type to the function
info: all functions in the public API must have an explicit return type info: all functions in the public API must have an explicit return type
@ -25,6 +27,7 @@ error[no-slow-types]: missing explicit return type in the public API
| |
2 | export function addD(a: number, b: number) { 2 | export function addD(a: number, b: number) {
| ^^^^ this function is missing an explicit return type | ^^^^ this function is missing an explicit return type
|
= hint: add an explicit return type to the function = hint: add an explicit return type to the function
info: all functions in the public API must have an explicit return type info: all functions in the public API must have an explicit return type

View file

@ -3,6 +3,7 @@ error[no-slow-types]: missing explicit return type in the public API
| |
1 | export function add(a: number, b: number) { 1 | export function add(a: number, b: number) {
| ^^^ this function is missing an explicit return type | ^^^ this function is missing an explicit return type
|
= hint: add an explicit return type to the function = hint: add an explicit return type to the function
info: all functions in the public API must have an explicit return type info: all functions in the public API must have an explicit return type
@ -14,6 +15,7 @@ error[no-slow-types]: missing explicit return type in the public API
| |
1 | export function addB(a: number, b: number) { 1 | export function addB(a: number, b: number) {
| ^^^^ this function is missing an explicit return type | ^^^^ this function is missing an explicit return type
|
= hint: add an explicit return type to the function = hint: add an explicit return type to the function
info: all functions in the public API must have an explicit return type info: all functions in the public API must have an explicit return type
@ -25,6 +27,7 @@ error[no-slow-types]: missing explicit return type in the public API
| |
2 | export function addD(a: number, b: number) { 2 | export function addD(a: number, b: number) {
| ^^^^ this function is missing an explicit return type | ^^^^ this function is missing an explicit return type
|
= hint: add an explicit return type to the function = hint: add an explicit return type to the function
info: all functions in the public API must have an explicit return type info: all functions in the public API must have an explicit return type

View file

@ -3,6 +3,7 @@ error[no-slow-types]: missing explicit return type in the public API
| |
1 | export function add(a: number, b: number) { 1 | export function add(a: number, b: number) {
| ^^^ this function is missing an explicit return type | ^^^ this function is missing an explicit return type
|
= hint: add an explicit return type to the function = hint: add an explicit return type to the function
info: all functions in the public API must have an explicit return type info: all functions in the public API must have an explicit return type
@ -14,6 +15,7 @@ error[no-slow-types]: missing explicit return type in the public API
| |
1 | export function addB(a: number, b: number) { 1 | export function addB(a: number, b: number) {
| ^^^^ this function is missing an explicit return type | ^^^^ this function is missing an explicit return type
|
= hint: add an explicit return type to the function = hint: add an explicit return type to the function
info: all functions in the public API must have an explicit return type info: all functions in the public API must have an explicit return type
@ -25,6 +27,7 @@ error[no-slow-types]: missing explicit return type in the public API
| |
2 | export function addD(a: number, b: number) { 2 | export function addD(a: number, b: number) {
| ^^^^ this function is missing an explicit return type | ^^^^ this function is missing an explicit return type
|
= hint: add an explicit return type to the function = hint: add an explicit return type to the function
info: all functions in the public API must have an explicit return type info: all functions in the public API must have an explicit return type

View file

@ -3,6 +3,7 @@ error[no-slow-types]: missing explicit return type in the public API
| |
1 | export function addB(a: number, b: number) { 1 | export function addB(a: number, b: number) {
| ^^^^ this function is missing an explicit return type | ^^^^ this function is missing an explicit return type
|
= hint: add an explicit return type to the function = hint: add an explicit return type to the function
info: all functions in the public API must have an explicit return type info: all functions in the public API must have an explicit return type
@ -14,6 +15,7 @@ error[no-slow-types]: missing explicit return type in the public API
| |
2 | export function addD(a: number, b: number) { 2 | export function addD(a: number, b: number) {
| ^^^^ this function is missing an explicit return type | ^^^^ this function is missing an explicit return type
|
= hint: add an explicit return type to the function = hint: add an explicit return type to the function
info: all functions in the public API must have an explicit return type info: all functions in the public API must have an explicit return type
@ -25,6 +27,7 @@ error[no-slow-types]: missing explicit return type in the public API
| |
1 | export function add(a: number, b: number) { 1 | export function add(a: number, b: number) {
| ^^^ this function is missing an explicit return type | ^^^ this function is missing an explicit return type
|
= hint: add an explicit return type to the function = hint: add an explicit return type to the function
info: all functions in the public API must have an explicit return type info: all functions in the public API must have an explicit return type
@ -36,6 +39,7 @@ error[no-slow-types]: missing explicit return type in the public API
| |
2 | export function addC(a: number, b: number) { 2 | export function addC(a: number, b: number) {
| ^^^^ this function is missing an explicit return type | ^^^^ this function is missing an explicit return type
|
= hint: add an explicit return type to the function = hint: add an explicit return type to the function
info: all functions in the public API must have an explicit return type info: all functions in the public API must have an explicit return type

View file

@ -17,6 +17,7 @@ use file_test_runner::collection::CollectTestsError;
use file_test_runner::collection::CollectedCategoryOrTest; use file_test_runner::collection::CollectedCategoryOrTest;
use file_test_runner::collection::CollectedTest; use file_test_runner::collection::CollectedTest;
use file_test_runner::collection::CollectedTestCategory; use file_test_runner::collection::CollectedTestCategory;
use file_test_runner::SubTestResult;
use file_test_runner::TestResult; use file_test_runner::TestResult;
use serde::Deserialize; use serde::Deserialize;
use test_util::tests_path; use test_util::tests_path;
@ -109,6 +110,8 @@ struct MultiStepMetaData {
#[serde(default)] #[serde(default)]
pub envs: HashMap<String, String>, pub envs: HashMap<String, String>,
#[serde(default)] #[serde(default)]
pub repeat: Option<usize>,
#[serde(default)]
pub steps: Vec<StepMetaData>, pub steps: Vec<StepMetaData>,
} }
@ -119,6 +122,8 @@ struct SingleTestMetaData {
pub base: Option<String>, pub base: Option<String>,
#[serde(default)] #[serde(default)]
pub temp_dir: bool, pub temp_dir: bool,
#[serde(default)]
pub repeat: Option<usize>,
#[serde(flatten)] #[serde(flatten)]
pub step: StepMetaData, pub step: StepMetaData,
} }
@ -128,6 +133,7 @@ impl SingleTestMetaData {
MultiStepMetaData { MultiStepMetaData {
base: self.base, base: self.base,
temp_dir: self.temp_dir, temp_dir: self.temp_dir,
repeat: self.repeat,
envs: Default::default(), envs: Default::default(),
steps: vec![self.step], steps: vec![self.step],
} }
@ -213,8 +219,26 @@ fn run_test(test: &CollectedTest<serde_json::Value>) -> TestResult {
let cwd = PathRef::new(&test.path).parent(); let cwd = PathRef::new(&test.path).parent();
let metadata_value = test.data.clone(); let metadata_value = test.data.clone();
let diagnostic_logger = Rc::new(RefCell::new(Vec::<u8>::new())); let diagnostic_logger = Rc::new(RefCell::new(Vec::<u8>::new()));
let result = TestResult::from_maybe_panic(AssertUnwindSafe(|| { let result = TestResult::from_maybe_panic_or_result(AssertUnwindSafe(|| {
run_test_inner(metadata_value, &cwd, diagnostic_logger.clone()) let metadata = deserialize_value(metadata_value);
if let Some(repeat) = metadata.repeat {
TestResult::SubTests(
(0..repeat)
.map(|i| {
let diagnostic_logger = diagnostic_logger.clone();
SubTestResult {
name: format!("run {}", i + 1),
result: TestResult::from_maybe_panic(AssertUnwindSafe(|| {
run_test_inner(&metadata, &cwd, diagnostic_logger);
})),
}
})
.collect(),
)
} else {
run_test_inner(&metadata, &cwd, diagnostic_logger.clone());
TestResult::Passed
}
})); }));
match result { match result {
TestResult::Failed { TestResult::Failed {
@ -232,15 +256,13 @@ fn run_test(test: &CollectedTest<serde_json::Value>) -> TestResult {
} }
fn run_test_inner( fn run_test_inner(
metadata_value: serde_json::Value, metadata: &MultiStepMetaData,
cwd: &PathRef, cwd: &PathRef,
diagnostic_logger: Rc<RefCell<Vec<u8>>>, diagnostic_logger: Rc<RefCell<Vec<u8>>>,
) { ) {
let metadata = deserialize_value(metadata_value); let context = test_context_from_metadata(metadata, cwd, diagnostic_logger);
let context = test_context_from_metadata(&metadata, cwd, diagnostic_logger);
for step in metadata.steps.iter().filter(|s| should_run_step(s)) { for step in metadata.steps.iter().filter(|s| should_run_step(s)) {
let run_func = || run_step(step, &metadata, cwd, &context); let run_func = || run_step(step, metadata, cwd, &context);
if step.flaky { if step.flaky {
run_flaky(run_func); run_flaky(run_func);
} else { } else {

View file

@ -0,0 +1,5 @@
{
"tempDir": true,
"args": "run -A --quiet --lock=deno.lock main.ts",
"output": "main.out"
}

View file

@ -0,0 +1,5 @@
[UNORDERED_START]
1
2
3
[UNORDERED_END]

View file

@ -0,0 +1,9 @@
new Worker(new URL("./worker1.ts", import.meta.url), {
type: "module",
});
new Worker(new URL("./worker2.ts", import.meta.url), {
type: "module",
});
new Worker(new URL("./worker3.ts", import.meta.url), {
type: "module",
});

View file

@ -0,0 +1,9 @@
import "npm:chalk@4";
import "npm:react@18.2";
import "npm:preact@10.19";
import "npm:ajv";
import "npm:has";
import "npm:picocolors";
console.log(1);
self.close();

View file

@ -0,0 +1,6 @@
import "npm:@denotest/esm-basic";
import "npm:@denotest/add";
import "npm:@denotest/subtract";
console.log(2);
self.close();

View file

@ -0,0 +1,6 @@
import "npm:@denotest/subtract";
import "npm:@denotest/add";
import "npm:@denotest/esm-basic";
console.log(3);
self.close();

View file

@ -6,6 +6,7 @@ error[banned-triple-slash-directives]: triple slash directives that modify globa
| |
1 | /// <reference lib="deno.ns" /> 1 | /// <reference lib="deno.ns" />
| ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ the triple slash directive | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ the triple slash directive
|
= hint: remove the triple slash directive = hint: remove the triple slash directive
info: instead instruct the user of your package to specify these directives info: instead instruct the user of your package to specify these directives
@ -17,6 +18,7 @@ error[banned-triple-slash-directives]: triple slash directives that modify globa
| |
2 | /// <reference no-default-lib="true" /> 2 | /// <reference no-default-lib="true" />
| ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ the triple slash directive | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ the triple slash directive
|
= hint: remove the triple slash directive = hint: remove the triple slash directive
info: instead instruct the user of your package to specify these directives info: instead instruct the user of your package to specify these directives

View file

@ -5,6 +5,7 @@ error[missing-explicit-return-type]: missing explicit return type in the public
| |
2 | export function getRandom() { 2 | export function getRandom() {
| ^^^^^^^^^ this function is missing an explicit return type | ^^^^^^^^^ this function is missing an explicit return type
|
= hint: add an explicit return type to the function = hint: add an explicit return type to the function
info: all functions in the public API must have an explicit return type info: all functions in the public API must have an explicit return type

View file

@ -10,6 +10,7 @@ error[invalid-external-import]: invalid import to a non-JSR 'http' specifier
| |
1 | import "http://localhost:4545/welcome.ts"; 1 | import "http://localhost:4545/welcome.ts";
| ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ the specifier | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ the specifier
|
= hint: replace this import with one from jsr or npm, or vendor the dependency into your package = hint: replace this import with one from jsr or npm, or vendor the dependency into your package
info: the import was resolved to 'http://localhost:4545/welcome.ts' info: the import was resolved to 'http://localhost:4545/welcome.ts'
@ -22,6 +23,7 @@ error[invalid-external-import]: invalid import to a non-JSR 'http' specifier
| |
2 | import "$echo"; 2 | import "$echo";
| ^^^^^^^ the specifier | ^^^^^^^ the specifier
|
= hint: replace this import with one from jsr or npm, or vendor the dependency into your package = hint: replace this import with one from jsr or npm, or vendor the dependency into your package
info: the import was resolved to 'http://localhost:4545/echo.ts' info: the import was resolved to 'http://localhost:4545/echo.ts'

View file

@ -7,10 +7,12 @@ error[invalid-external-import]: invalid import to a non-JSR 'http' specifier
| |
1 | import "http://esm.sh/react-dom@18.2.0/server"; 1 | import "http://esm.sh/react-dom@18.2.0/server";
| ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ the specifier | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ the specifier
|
= hint: replace this import with one from jsr or npm, or vendor the dependency into your package = hint: replace this import with one from jsr or npm, or vendor the dependency into your package
| |
1 | "npm:react-dom@18.2.0" 1 | "npm:react-dom@18.2.0"
| ---------------------- try this specifier | ---------------------- try this specifier
|
info: the import was resolved to 'http://esm.sh/react-dom@18.2.0/server' info: the import was resolved to 'http://esm.sh/react-dom@18.2.0/server'
info: this specifier is not allowed to be imported on jsr info: this specifier is not allowed to be imported on jsr

View file

@ -6,6 +6,7 @@ error[missing-constraint]: specifier 'jsr:@denotest/add' is missing a version co
| |
1 | import { add } from "add"; 1 | import { add } from "add";
| ^^^^^ the specifier | ^^^^^ the specifier
|
= hint: specify a version constraint for the specifier in the import map = hint: specify a version constraint for the specifier in the import map
info: the specifier resolved to version 1.0.0 today, but will resolve to a different info: the specifier resolved to version 1.0.0 today, but will resolve to a different
@ -17,6 +18,7 @@ error[missing-constraint]: specifier 'npm:@denotest/esm-basic' is missing a vers
| |
2 | import * as basic from "basic"; 2 | import * as basic from "basic";
| ^^^^^^^ the specifier | ^^^^^^^ the specifier
|
= hint: specify a version constraint for the specifier in the import map = hint: specify a version constraint for the specifier in the import map
info: the specifier resolved to version 1.0.0 today, but will resolve to a different info: the specifier resolved to version 1.0.0 today, but will resolve to a different
@ -28,6 +30,7 @@ error[missing-constraint]: specifier 'jsr:@denotest/deps' is missing a version c
| |
3 | import * as deps from "jsr:@denotest/deps"; 3 | import * as deps from "jsr:@denotest/deps";
| ^^^^^^^^^^^^^^^^^^^^ the specifier | ^^^^^^^^^^^^^^^^^^^^ the specifier
|
= hint: specify a version constraint for the specifier = hint: specify a version constraint for the specifier
info: the specifier resolved to version 1.0.0 today, but will resolve to a different info: the specifier resolved to version 1.0.0 today, but will resolve to a different

View file

@ -7,6 +7,7 @@ error[invalid-external-import]: invalid import to a non-JSR 'https' specifier
| |
1 | export * from "https://deno.land/std/assert/assert.ts"; 1 | export * from "https://deno.land/std/assert/assert.ts";
| ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ the specifier | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ the specifier
|
= hint: replace this import with one from jsr or npm, or vendor the dependency into your package = hint: replace this import with one from jsr or npm, or vendor the dependency into your package
info: the import was resolved to 'https://deno.land/std/assert/assert.ts' info: the import was resolved to 'https://deno.land/std/assert/assert.ts'

View file

@ -6,6 +6,7 @@ warning[unanalyzable-dynamic-import]: unable to analyze dynamic import
| |
2 | await import("asd " + asd); 2 | await import("asd " + asd);
| ^^^^^^^^^^^^ the unanalyzable dynamic import | ^^^^^^^^^^^^ the unanalyzable dynamic import
|
info: after publishing this package, imports from the local import map / package.json do not work info: after publishing this package, imports from the local import map / package.json do not work
info: dynamic imports that can not be analyzed at publish time will not be rewritten automatically info: dynamic imports that can not be analyzed at publish time will not be rewritten automatically

View file

@ -69,6 +69,9 @@
"type": "string" "type": "string"
} }
}, },
"repeat": {
"type": "number"
},
"steps": { "steps": {
"type": "array", "type": "array",
"items": { "items": {
@ -112,6 +115,9 @@
"type": "string" "type": "string"
} }
}, },
"repeat": {
"type": "number"
},
"tests": { "tests": {
"type": "object", "type": "object",
"additionalProperties": { "additionalProperties": {

View file

@ -14,6 +14,7 @@ error[private-type-ref]: public type 'MyClass.prototype.prop' references private
| |
1 | interface MyInterface { 1 | interface MyInterface {
| - this is the referenced type | - this is the referenced type
|
info: to ensure documentation is complete all types that are exposed in the public API must be public info: to ensure documentation is complete all types that are exposed in the public API must be public

View file

@ -227,6 +227,12 @@ export async function downloadPrebuilt(toolName) {
); );
} }
spinner.text = `Successfully downloaded: ${toolName}`; spinner.text = `Successfully downloaded: ${toolName}`;
try {
// necessary on Windows it seems
await Deno.remove(toolPath);
} catch {
// ignore
}
await Deno.rename(tempFile, toolPath); await Deno.rename(tempFile, toolPath);
} catch (e) { } catch (e) {
spinner.fail(); spinner.fail();