2023-01-13 02:51:32 -05:00
|
|
|
// Copyright 2018-2023 the Deno authors. All rights reserved. MIT license.
|
|
|
|
|
2022-08-22 12:14:59 -04:00
|
|
|
use std::collections::HashMap;
|
|
|
|
use std::sync::Arc;
|
|
|
|
|
|
|
|
use deno_ast::MediaType;
|
|
|
|
use deno_ast::ModuleSpecifier;
|
|
|
|
use deno_ast::ParsedSource;
|
|
|
|
use deno_core::error::AnyError;
|
|
|
|
use deno_core::parking_lot::Mutex;
|
|
|
|
use deno_core::serde_json;
|
|
|
|
use deno_graph::CapturingModuleParser;
|
|
|
|
use deno_graph::DefaultModuleAnalyzer;
|
|
|
|
use deno_graph::ModuleInfo;
|
|
|
|
use deno_graph::ModuleParser;
|
|
|
|
use deno_runtime::deno_webstorage::rusqlite::params;
|
|
|
|
|
feat(core): initialize SQLite off-main-thread (#18401)
This gets SQLite off the flamegraph and reduces initialization time by
somewhere between 0.2ms and 0.5ms. In addition, I took the opportunity
to move all the cache management code to a single place and reduce
duplication. While the PR has a net gain of lines, much of that is just
being a bit more deliberate with how we're recovering from errors.
The existing caches had various policies for dealing with cache
corruption, so I've unified them and tried to isolate the decisions we
make for recovery in a single place (see `open_connection` in
`CacheDB`). The policy I chose was:
1. Retry twice to open on-disk caches
2. If that fails, try to delete the file and recreate it on-disk
3. If we fail to delete the file or re-create a new cache, use a
fallback strategy that can be chosen per-cache: InMemory (temporary
cache for the process run), BlackHole (ignore writes, return empty
reads), or Error (fail on every operation).
The caches all use the same general code now, and share the cache
failure recovery policy.
In addition, it cleans up a TODO in the `NodeAnalysisCache`.
2023-03-27 18:01:52 -04:00
|
|
|
use super::cache_db::CacheDB;
|
|
|
|
use super::cache_db::CacheDBConfiguration;
|
|
|
|
use super::cache_db::CacheFailure;
|
2022-08-22 12:14:59 -04:00
|
|
|
use super::FastInsecureHasher;
|
|
|
|
|
feat(core): initialize SQLite off-main-thread (#18401)
This gets SQLite off the flamegraph and reduces initialization time by
somewhere between 0.2ms and 0.5ms. In addition, I took the opportunity
to move all the cache management code to a single place and reduce
duplication. While the PR has a net gain of lines, much of that is just
being a bit more deliberate with how we're recovering from errors.
The existing caches had various policies for dealing with cache
corruption, so I've unified them and tried to isolate the decisions we
make for recovery in a single place (see `open_connection` in
`CacheDB`). The policy I chose was:
1. Retry twice to open on-disk caches
2. If that fails, try to delete the file and recreate it on-disk
3. If we fail to delete the file or re-create a new cache, use a
fallback strategy that can be chosen per-cache: InMemory (temporary
cache for the process run), BlackHole (ignore writes, return empty
reads), or Error (fail on every operation).
The caches all use the same general code now, and share the cache
failure recovery policy.
In addition, it cleans up a TODO in the `NodeAnalysisCache`.
2023-03-27 18:01:52 -04:00
|
|
|
const SELECT_MODULE_INFO: &str = "
|
|
|
|
SELECT
|
|
|
|
module_info
|
|
|
|
FROM
|
|
|
|
moduleinfocache
|
|
|
|
WHERE
|
|
|
|
specifier=?1
|
|
|
|
AND media_type=?2
|
|
|
|
AND source_hash=?3
|
|
|
|
LIMIT 1";
|
|
|
|
|
|
|
|
pub static PARSED_SOURCE_CACHE_DB: CacheDBConfiguration =
|
|
|
|
CacheDBConfiguration {
|
|
|
|
table_initializer: "CREATE TABLE IF NOT EXISTS moduleinfocache (
|
|
|
|
specifier TEXT PRIMARY KEY,
|
|
|
|
media_type TEXT NOT NULL,
|
|
|
|
source_hash TEXT NOT NULL,
|
|
|
|
module_info TEXT NOT NULL
|
|
|
|
);",
|
|
|
|
on_version_change: "DELETE FROM moduleinfocache;",
|
|
|
|
preheat_queries: &[SELECT_MODULE_INFO],
|
|
|
|
on_failure: CacheFailure::InMemory,
|
|
|
|
};
|
|
|
|
|
2022-08-22 12:14:59 -04:00
|
|
|
#[derive(Clone, Default)]
|
|
|
|
struct ParsedSourceCacheSources(
|
|
|
|
Arc<Mutex<HashMap<ModuleSpecifier, ParsedSource>>>,
|
|
|
|
);
|
|
|
|
|
|
|
|
/// It's ok that this is racy since in non-LSP situations
|
|
|
|
/// this will only ever store one form of a parsed source
|
|
|
|
/// and in LSP settings the concurrency will be enforced
|
|
|
|
/// at a higher level to ensure this will have the latest
|
|
|
|
/// parsed source.
|
|
|
|
impl deno_graph::ParsedSourceStore for ParsedSourceCacheSources {
|
|
|
|
fn set_parsed_source(
|
|
|
|
&self,
|
|
|
|
specifier: deno_graph::ModuleSpecifier,
|
|
|
|
parsed_source: ParsedSource,
|
|
|
|
) -> Option<ParsedSource> {
|
|
|
|
self.0.lock().insert(specifier, parsed_source)
|
|
|
|
}
|
|
|
|
|
|
|
|
fn get_parsed_source(
|
|
|
|
&self,
|
|
|
|
specifier: &deno_graph::ModuleSpecifier,
|
|
|
|
) -> Option<ParsedSource> {
|
|
|
|
self.0.lock().get(specifier).cloned()
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/// A cache of `ParsedSource`s, which may be used with `deno_graph`
|
|
|
|
/// for cached dependency analysis.
|
|
|
|
pub struct ParsedSourceCache {
|
feat(core): initialize SQLite off-main-thread (#18401)
This gets SQLite off the flamegraph and reduces initialization time by
somewhere between 0.2ms and 0.5ms. In addition, I took the opportunity
to move all the cache management code to a single place and reduce
duplication. While the PR has a net gain of lines, much of that is just
being a bit more deliberate with how we're recovering from errors.
The existing caches had various policies for dealing with cache
corruption, so I've unified them and tried to isolate the decisions we
make for recovery in a single place (see `open_connection` in
`CacheDB`). The policy I chose was:
1. Retry twice to open on-disk caches
2. If that fails, try to delete the file and recreate it on-disk
3. If we fail to delete the file or re-create a new cache, use a
fallback strategy that can be chosen per-cache: InMemory (temporary
cache for the process run), BlackHole (ignore writes, return empty
reads), or Error (fail on every operation).
The caches all use the same general code now, and share the cache
failure recovery policy.
In addition, it cleans up a TODO in the `NodeAnalysisCache`.
2023-03-27 18:01:52 -04:00
|
|
|
db: CacheDB,
|
2022-08-22 12:14:59 -04:00
|
|
|
sources: ParsedSourceCacheSources,
|
|
|
|
}
|
|
|
|
|
|
|
|
impl ParsedSourceCache {
|
feat(core): initialize SQLite off-main-thread (#18401)
This gets SQLite off the flamegraph and reduces initialization time by
somewhere between 0.2ms and 0.5ms. In addition, I took the opportunity
to move all the cache management code to a single place and reduce
duplication. While the PR has a net gain of lines, much of that is just
being a bit more deliberate with how we're recovering from errors.
The existing caches had various policies for dealing with cache
corruption, so I've unified them and tried to isolate the decisions we
make for recovery in a single place (see `open_connection` in
`CacheDB`). The policy I chose was:
1. Retry twice to open on-disk caches
2. If that fails, try to delete the file and recreate it on-disk
3. If we fail to delete the file or re-create a new cache, use a
fallback strategy that can be chosen per-cache: InMemory (temporary
cache for the process run), BlackHole (ignore writes, return empty
reads), or Error (fail on every operation).
The caches all use the same general code now, and share the cache
failure recovery policy.
In addition, it cleans up a TODO in the `NodeAnalysisCache`.
2023-03-27 18:01:52 -04:00
|
|
|
#[cfg(test)]
|
|
|
|
pub fn new_in_memory() -> Self {
|
|
|
|
Self {
|
|
|
|
db: CacheDB::in_memory(&PARSED_SOURCE_CACHE_DB, crate::version::deno()),
|
|
|
|
sources: Default::default(),
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
pub fn new(db: CacheDB) -> Self {
|
2022-08-22 12:14:59 -04:00
|
|
|
Self {
|
feat(core): initialize SQLite off-main-thread (#18401)
This gets SQLite off the flamegraph and reduces initialization time by
somewhere between 0.2ms and 0.5ms. In addition, I took the opportunity
to move all the cache management code to a single place and reduce
duplication. While the PR has a net gain of lines, much of that is just
being a bit more deliberate with how we're recovering from errors.
The existing caches had various policies for dealing with cache
corruption, so I've unified them and tried to isolate the decisions we
make for recovery in a single place (see `open_connection` in
`CacheDB`). The policy I chose was:
1. Retry twice to open on-disk caches
2. If that fails, try to delete the file and recreate it on-disk
3. If we fail to delete the file or re-create a new cache, use a
fallback strategy that can be chosen per-cache: InMemory (temporary
cache for the process run), BlackHole (ignore writes, return empty
reads), or Error (fail on every operation).
The caches all use the same general code now, and share the cache
failure recovery policy.
In addition, it cleans up a TODO in the `NodeAnalysisCache`.
2023-03-27 18:01:52 -04:00
|
|
|
db,
|
2023-01-10 10:28:10 -05:00
|
|
|
sources: Default::default(),
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-02-22 14:15:25 -05:00
|
|
|
pub fn get_parsed_source_from_esm_module(
|
2022-08-22 12:14:59 -04:00
|
|
|
&self,
|
2023-02-22 14:15:25 -05:00
|
|
|
module: &deno_graph::EsmModule,
|
|
|
|
) -> Result<ParsedSource, deno_ast::Diagnostic> {
|
|
|
|
self.get_or_parse_module(
|
|
|
|
&module.specifier,
|
|
|
|
module.source.clone(),
|
|
|
|
module.media_type,
|
|
|
|
)
|
2022-08-22 12:14:59 -04:00
|
|
|
}
|
|
|
|
|
|
|
|
/// Gets the matching `ParsedSource` from the cache
|
|
|
|
/// or parses a new one and stores that in the cache.
|
|
|
|
pub fn get_or_parse_module(
|
|
|
|
&self,
|
|
|
|
specifier: &deno_graph::ModuleSpecifier,
|
|
|
|
source: Arc<str>,
|
|
|
|
media_type: MediaType,
|
|
|
|
) -> deno_core::anyhow::Result<ParsedSource, deno_ast::Diagnostic> {
|
2022-08-29 14:24:10 -04:00
|
|
|
let parser = self.as_capturing_parser();
|
2022-08-22 12:14:59 -04:00
|
|
|
// this will conditionally parse because it's using a CapturingModuleParser
|
|
|
|
parser.parse_module(specifier, source, media_type)
|
|
|
|
}
|
|
|
|
|
|
|
|
/// Frees the parsed source from memory.
|
|
|
|
pub fn free(&self, specifier: &ModuleSpecifier) {
|
|
|
|
self.sources.0.lock().remove(specifier);
|
|
|
|
}
|
|
|
|
|
|
|
|
pub fn as_analyzer(&self) -> Box<dyn deno_graph::ModuleAnalyzer> {
|
feat(core): initialize SQLite off-main-thread (#18401)
This gets SQLite off the flamegraph and reduces initialization time by
somewhere between 0.2ms and 0.5ms. In addition, I took the opportunity
to move all the cache management code to a single place and reduce
duplication. While the PR has a net gain of lines, much of that is just
being a bit more deliberate with how we're recovering from errors.
The existing caches had various policies for dealing with cache
corruption, so I've unified them and tried to isolate the decisions we
make for recovery in a single place (see `open_connection` in
`CacheDB`). The policy I chose was:
1. Retry twice to open on-disk caches
2. If that fails, try to delete the file and recreate it on-disk
3. If we fail to delete the file or re-create a new cache, use a
fallback strategy that can be chosen per-cache: InMemory (temporary
cache for the process run), BlackHole (ignore writes, return empty
reads), or Error (fail on every operation).
The caches all use the same general code now, and share the cache
failure recovery policy.
In addition, it cleans up a TODO in the `NodeAnalysisCache`.
2023-03-27 18:01:52 -04:00
|
|
|
Box::new(ParsedSourceCacheModuleAnalyzer::new(
|
|
|
|
self.db.clone(),
|
2022-08-22 12:14:59 -04:00
|
|
|
self.sources.clone(),
|
feat(core): initialize SQLite off-main-thread (#18401)
This gets SQLite off the flamegraph and reduces initialization time by
somewhere between 0.2ms and 0.5ms. In addition, I took the opportunity
to move all the cache management code to a single place and reduce
duplication. While the PR has a net gain of lines, much of that is just
being a bit more deliberate with how we're recovering from errors.
The existing caches had various policies for dealing with cache
corruption, so I've unified them and tried to isolate the decisions we
make for recovery in a single place (see `open_connection` in
`CacheDB`). The policy I chose was:
1. Retry twice to open on-disk caches
2. If that fails, try to delete the file and recreate it on-disk
3. If we fail to delete the file or re-create a new cache, use a
fallback strategy that can be chosen per-cache: InMemory (temporary
cache for the process run), BlackHole (ignore writes, return empty
reads), or Error (fail on every operation).
The caches all use the same general code now, and share the cache
failure recovery policy.
In addition, it cleans up a TODO in the `NodeAnalysisCache`.
2023-03-27 18:01:52 -04:00
|
|
|
))
|
2022-08-22 12:14:59 -04:00
|
|
|
}
|
2022-08-29 14:24:10 -04:00
|
|
|
|
|
|
|
/// Creates a parser that will reuse a ParsedSource from the store
|
|
|
|
/// if it exists, or else parse.
|
|
|
|
pub fn as_capturing_parser(&self) -> CapturingModuleParser {
|
|
|
|
CapturingModuleParser::new(None, &self.sources)
|
|
|
|
}
|
2022-08-22 12:14:59 -04:00
|
|
|
}
|
|
|
|
|
|
|
|
struct ParsedSourceCacheModuleAnalyzer {
|
feat(core): initialize SQLite off-main-thread (#18401)
This gets SQLite off the flamegraph and reduces initialization time by
somewhere between 0.2ms and 0.5ms. In addition, I took the opportunity
to move all the cache management code to a single place and reduce
duplication. While the PR has a net gain of lines, much of that is just
being a bit more deliberate with how we're recovering from errors.
The existing caches had various policies for dealing with cache
corruption, so I've unified them and tried to isolate the decisions we
make for recovery in a single place (see `open_connection` in
`CacheDB`). The policy I chose was:
1. Retry twice to open on-disk caches
2. If that fails, try to delete the file and recreate it on-disk
3. If we fail to delete the file or re-create a new cache, use a
fallback strategy that can be chosen per-cache: InMemory (temporary
cache for the process run), BlackHole (ignore writes, return empty
reads), or Error (fail on every operation).
The caches all use the same general code now, and share the cache
failure recovery policy.
In addition, it cleans up a TODO in the `NodeAnalysisCache`.
2023-03-27 18:01:52 -04:00
|
|
|
conn: CacheDB,
|
2022-08-22 12:14:59 -04:00
|
|
|
sources: ParsedSourceCacheSources,
|
|
|
|
}
|
|
|
|
|
|
|
|
impl ParsedSourceCacheModuleAnalyzer {
|
feat(core): initialize SQLite off-main-thread (#18401)
This gets SQLite off the flamegraph and reduces initialization time by
somewhere between 0.2ms and 0.5ms. In addition, I took the opportunity
to move all the cache management code to a single place and reduce
duplication. While the PR has a net gain of lines, much of that is just
being a bit more deliberate with how we're recovering from errors.
The existing caches had various policies for dealing with cache
corruption, so I've unified them and tried to isolate the decisions we
make for recovery in a single place (see `open_connection` in
`CacheDB`). The policy I chose was:
1. Retry twice to open on-disk caches
2. If that fails, try to delete the file and recreate it on-disk
3. If we fail to delete the file or re-create a new cache, use a
fallback strategy that can be chosen per-cache: InMemory (temporary
cache for the process run), BlackHole (ignore writes, return empty
reads), or Error (fail on every operation).
The caches all use the same general code now, and share the cache
failure recovery policy.
In addition, it cleans up a TODO in the `NodeAnalysisCache`.
2023-03-27 18:01:52 -04:00
|
|
|
pub fn new(conn: CacheDB, sources: ParsedSourceCacheSources) -> Self {
|
|
|
|
Self { conn, sources }
|
2022-08-22 12:14:59 -04:00
|
|
|
}
|
|
|
|
|
|
|
|
pub fn get_module_info(
|
|
|
|
&self,
|
|
|
|
specifier: &ModuleSpecifier,
|
|
|
|
media_type: MediaType,
|
|
|
|
expected_source_hash: &str,
|
|
|
|
) -> Result<Option<ModuleInfo>, AnyError> {
|
feat(core): initialize SQLite off-main-thread (#18401)
This gets SQLite off the flamegraph and reduces initialization time by
somewhere between 0.2ms and 0.5ms. In addition, I took the opportunity
to move all the cache management code to a single place and reduce
duplication. While the PR has a net gain of lines, much of that is just
being a bit more deliberate with how we're recovering from errors.
The existing caches had various policies for dealing with cache
corruption, so I've unified them and tried to isolate the decisions we
make for recovery in a single place (see `open_connection` in
`CacheDB`). The policy I chose was:
1. Retry twice to open on-disk caches
2. If that fails, try to delete the file and recreate it on-disk
3. If we fail to delete the file or re-create a new cache, use a
fallback strategy that can be chosen per-cache: InMemory (temporary
cache for the process run), BlackHole (ignore writes, return empty
reads), or Error (fail on every operation).
The caches all use the same general code now, and share the cache
failure recovery policy.
In addition, it cleans up a TODO in the `NodeAnalysisCache`.
2023-03-27 18:01:52 -04:00
|
|
|
let query = SELECT_MODULE_INFO;
|
|
|
|
let res = self.conn.query_row(
|
|
|
|
query,
|
|
|
|
params![
|
|
|
|
&specifier.as_str(),
|
|
|
|
serialize_media_type(media_type),
|
|
|
|
&expected_source_hash,
|
|
|
|
],
|
|
|
|
|row| {
|
|
|
|
let module_info: String = row.get(0)?;
|
|
|
|
let module_info = serde_json::from_str(&module_info)?;
|
|
|
|
Ok(module_info)
|
|
|
|
},
|
|
|
|
)?;
|
|
|
|
Ok(res)
|
2022-08-22 12:14:59 -04:00
|
|
|
}
|
|
|
|
|
|
|
|
pub fn set_module_info(
|
|
|
|
&self,
|
|
|
|
specifier: &ModuleSpecifier,
|
|
|
|
media_type: MediaType,
|
|
|
|
source_hash: &str,
|
|
|
|
module_info: &ModuleInfo,
|
|
|
|
) -> Result<(), AnyError> {
|
|
|
|
let sql = "
|
|
|
|
INSERT OR REPLACE INTO
|
|
|
|
moduleinfocache (specifier, media_type, source_hash, module_info)
|
|
|
|
VALUES
|
|
|
|
(?1, ?2, ?3, ?4)";
|
feat(core): initialize SQLite off-main-thread (#18401)
This gets SQLite off the flamegraph and reduces initialization time by
somewhere between 0.2ms and 0.5ms. In addition, I took the opportunity
to move all the cache management code to a single place and reduce
duplication. While the PR has a net gain of lines, much of that is just
being a bit more deliberate with how we're recovering from errors.
The existing caches had various policies for dealing with cache
corruption, so I've unified them and tried to isolate the decisions we
make for recovery in a single place (see `open_connection` in
`CacheDB`). The policy I chose was:
1. Retry twice to open on-disk caches
2. If that fails, try to delete the file and recreate it on-disk
3. If we fail to delete the file or re-create a new cache, use a
fallback strategy that can be chosen per-cache: InMemory (temporary
cache for the process run), BlackHole (ignore writes, return empty
reads), or Error (fail on every operation).
The caches all use the same general code now, and share the cache
failure recovery policy.
In addition, it cleans up a TODO in the `NodeAnalysisCache`.
2023-03-27 18:01:52 -04:00
|
|
|
self.conn.execute(
|
|
|
|
sql,
|
|
|
|
params![
|
|
|
|
specifier.as_str(),
|
|
|
|
serialize_media_type(media_type),
|
|
|
|
&source_hash,
|
|
|
|
&serde_json::to_string(&module_info)?,
|
|
|
|
],
|
|
|
|
)?;
|
2022-08-22 12:14:59 -04:00
|
|
|
Ok(())
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-02-09 09:17:48 -05:00
|
|
|
// todo(dsherret): change this to be stored as an integer next time
|
|
|
|
// the cache version is bumped
|
|
|
|
fn serialize_media_type(media_type: MediaType) -> &'static str {
|
|
|
|
use MediaType::*;
|
|
|
|
match media_type {
|
|
|
|
JavaScript => "1",
|
|
|
|
Jsx => "2",
|
|
|
|
Mjs => "3",
|
|
|
|
Cjs => "4",
|
|
|
|
TypeScript => "5",
|
|
|
|
Mts => "6",
|
|
|
|
Cts => "7",
|
|
|
|
Dts => "8",
|
|
|
|
Dmts => "9",
|
|
|
|
Dcts => "10",
|
|
|
|
Tsx => "11",
|
|
|
|
Json => "12",
|
|
|
|
Wasm => "13",
|
|
|
|
TsBuildInfo => "14",
|
|
|
|
SourceMap => "15",
|
|
|
|
Unknown => "16",
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2022-08-22 12:14:59 -04:00
|
|
|
impl deno_graph::ModuleAnalyzer for ParsedSourceCacheModuleAnalyzer {
|
|
|
|
fn analyze(
|
|
|
|
&self,
|
|
|
|
specifier: &ModuleSpecifier,
|
|
|
|
source: Arc<str>,
|
|
|
|
media_type: MediaType,
|
|
|
|
) -> Result<ModuleInfo, deno_ast::Diagnostic> {
|
|
|
|
// attempt to load from the cache
|
|
|
|
let source_hash = compute_source_hash(source.as_bytes());
|
|
|
|
match self.get_module_info(specifier, media_type, &source_hash) {
|
|
|
|
Ok(Some(info)) => return Ok(info),
|
|
|
|
Ok(None) => {}
|
|
|
|
Err(err) => {
|
|
|
|
log::debug!(
|
|
|
|
"Error loading module cache info for {}. {:#}",
|
|
|
|
specifier,
|
|
|
|
err
|
|
|
|
);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// otherwise, get the module info from the parsed source cache
|
|
|
|
let parser = CapturingModuleParser::new(None, &self.sources);
|
|
|
|
let analyzer = DefaultModuleAnalyzer::new(&parser);
|
|
|
|
|
|
|
|
let module_info = analyzer.analyze(specifier, source, media_type)?;
|
|
|
|
|
|
|
|
// then attempt to cache it
|
|
|
|
if let Err(err) =
|
|
|
|
self.set_module_info(specifier, media_type, &source_hash, &module_info)
|
|
|
|
{
|
|
|
|
log::debug!(
|
|
|
|
"Error saving module cache info for {}. {:#}",
|
|
|
|
specifier,
|
|
|
|
err
|
|
|
|
);
|
|
|
|
}
|
|
|
|
|
|
|
|
Ok(module_info)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
fn compute_source_hash(bytes: &[u8]) -> String {
|
2023-07-10 17:45:09 -04:00
|
|
|
FastInsecureHasher::hash(bytes).to_string()
|
2022-08-22 12:14:59 -04:00
|
|
|
}
|
|
|
|
|
|
|
|
#[cfg(test)]
|
|
|
|
mod test {
|
|
|
|
use deno_graph::PositionRange;
|
|
|
|
use deno_graph::SpecifierWithRange;
|
|
|
|
|
|
|
|
use super::*;
|
|
|
|
|
|
|
|
#[test]
|
|
|
|
pub fn parsed_source_cache_module_analyzer_general_use() {
|
feat(core): initialize SQLite off-main-thread (#18401)
This gets SQLite off the flamegraph and reduces initialization time by
somewhere between 0.2ms and 0.5ms. In addition, I took the opportunity
to move all the cache management code to a single place and reduce
duplication. While the PR has a net gain of lines, much of that is just
being a bit more deliberate with how we're recovering from errors.
The existing caches had various policies for dealing with cache
corruption, so I've unified them and tried to isolate the decisions we
make for recovery in a single place (see `open_connection` in
`CacheDB`). The policy I chose was:
1. Retry twice to open on-disk caches
2. If that fails, try to delete the file and recreate it on-disk
3. If we fail to delete the file or re-create a new cache, use a
fallback strategy that can be chosen per-cache: InMemory (temporary
cache for the process run), BlackHole (ignore writes, return empty
reads), or Error (fail on every operation).
The caches all use the same general code now, and share the cache
failure recovery policy.
In addition, it cleans up a TODO in the `NodeAnalysisCache`.
2023-03-27 18:01:52 -04:00
|
|
|
let conn = CacheDB::in_memory(&PARSED_SOURCE_CACHE_DB, "1.0.0");
|
|
|
|
let cache = ParsedSourceCacheModuleAnalyzer::new(conn, Default::default());
|
2022-08-22 12:14:59 -04:00
|
|
|
let specifier1 =
|
|
|
|
ModuleSpecifier::parse("https://localhost/mod.ts").unwrap();
|
|
|
|
let specifier2 =
|
|
|
|
ModuleSpecifier::parse("https://localhost/mod2.ts").unwrap();
|
|
|
|
assert_eq!(
|
|
|
|
cache
|
|
|
|
.get_module_info(&specifier1, MediaType::JavaScript, "1")
|
|
|
|
.unwrap(),
|
|
|
|
None
|
|
|
|
);
|
|
|
|
|
|
|
|
let mut module_info = ModuleInfo::default();
|
|
|
|
module_info.jsdoc_imports.push(SpecifierWithRange {
|
|
|
|
range: PositionRange {
|
|
|
|
start: deno_graph::Position {
|
|
|
|
line: 0,
|
|
|
|
character: 3,
|
|
|
|
},
|
|
|
|
end: deno_graph::Position {
|
|
|
|
line: 1,
|
|
|
|
character: 2,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
text: "test".to_string(),
|
|
|
|
});
|
|
|
|
cache
|
|
|
|
.set_module_info(&specifier1, MediaType::JavaScript, "1", &module_info)
|
|
|
|
.unwrap();
|
|
|
|
assert_eq!(
|
|
|
|
cache
|
|
|
|
.get_module_info(&specifier1, MediaType::JavaScript, "1")
|
|
|
|
.unwrap(),
|
|
|
|
Some(module_info.clone())
|
|
|
|
);
|
|
|
|
assert_eq!(
|
|
|
|
cache
|
|
|
|
.get_module_info(&specifier2, MediaType::JavaScript, "1")
|
|
|
|
.unwrap(),
|
|
|
|
None,
|
|
|
|
);
|
|
|
|
// different media type
|
|
|
|
assert_eq!(
|
|
|
|
cache
|
|
|
|
.get_module_info(&specifier1, MediaType::TypeScript, "1")
|
|
|
|
.unwrap(),
|
|
|
|
None,
|
|
|
|
);
|
|
|
|
// different source hash
|
|
|
|
assert_eq!(
|
|
|
|
cache
|
|
|
|
.get_module_info(&specifier1, MediaType::JavaScript, "2")
|
|
|
|
.unwrap(),
|
|
|
|
None,
|
|
|
|
);
|
|
|
|
|
|
|
|
// try recreating with the same version
|
feat(core): initialize SQLite off-main-thread (#18401)
This gets SQLite off the flamegraph and reduces initialization time by
somewhere between 0.2ms and 0.5ms. In addition, I took the opportunity
to move all the cache management code to a single place and reduce
duplication. While the PR has a net gain of lines, much of that is just
being a bit more deliberate with how we're recovering from errors.
The existing caches had various policies for dealing with cache
corruption, so I've unified them and tried to isolate the decisions we
make for recovery in a single place (see `open_connection` in
`CacheDB`). The policy I chose was:
1. Retry twice to open on-disk caches
2. If that fails, try to delete the file and recreate it on-disk
3. If we fail to delete the file or re-create a new cache, use a
fallback strategy that can be chosen per-cache: InMemory (temporary
cache for the process run), BlackHole (ignore writes, return empty
reads), or Error (fail on every operation).
The caches all use the same general code now, and share the cache
failure recovery policy.
In addition, it cleans up a TODO in the `NodeAnalysisCache`.
2023-03-27 18:01:52 -04:00
|
|
|
let conn = cache.conn.recreate_with_version("1.0.0");
|
|
|
|
let cache = ParsedSourceCacheModuleAnalyzer::new(conn, Default::default());
|
2022-08-22 12:14:59 -04:00
|
|
|
|
|
|
|
// should get it
|
|
|
|
assert_eq!(
|
|
|
|
cache
|
|
|
|
.get_module_info(&specifier1, MediaType::JavaScript, "1")
|
|
|
|
.unwrap(),
|
|
|
|
Some(module_info)
|
|
|
|
);
|
|
|
|
|
|
|
|
// try recreating with a different version
|
feat(core): initialize SQLite off-main-thread (#18401)
This gets SQLite off the flamegraph and reduces initialization time by
somewhere between 0.2ms and 0.5ms. In addition, I took the opportunity
to move all the cache management code to a single place and reduce
duplication. While the PR has a net gain of lines, much of that is just
being a bit more deliberate with how we're recovering from errors.
The existing caches had various policies for dealing with cache
corruption, so I've unified them and tried to isolate the decisions we
make for recovery in a single place (see `open_connection` in
`CacheDB`). The policy I chose was:
1. Retry twice to open on-disk caches
2. If that fails, try to delete the file and recreate it on-disk
3. If we fail to delete the file or re-create a new cache, use a
fallback strategy that can be chosen per-cache: InMemory (temporary
cache for the process run), BlackHole (ignore writes, return empty
reads), or Error (fail on every operation).
The caches all use the same general code now, and share the cache
failure recovery policy.
In addition, it cleans up a TODO in the `NodeAnalysisCache`.
2023-03-27 18:01:52 -04:00
|
|
|
let conn = cache.conn.recreate_with_version("1.0.1");
|
|
|
|
let cache = ParsedSourceCacheModuleAnalyzer::new(conn, Default::default());
|
2022-08-22 12:14:59 -04:00
|
|
|
|
|
|
|
// should no longer exist
|
|
|
|
assert_eq!(
|
|
|
|
cache
|
|
|
|
.get_module_info(&specifier1, MediaType::JavaScript, "1")
|
|
|
|
.unwrap(),
|
|
|
|
None,
|
|
|
|
);
|
|
|
|
}
|
|
|
|
}
|