mirror of
https://github.com/denoland/deno.git
synced 2024-11-23 15:16:54 -05:00
94f040ac28
In https://github.com/denoland/deno/pull/23955 we changed the sqlite db journal mode to WAL. This causes issues when someone is running an old version of Deno using TRUNCATE and a new version because the two fight against each other.
191 lines
4.7 KiB
Rust
191 lines
4.7 KiB
Rust
// Copyright 2018-2024 the Deno authors. All rights reserved. MIT license.
|
|
|
|
use deno_core::error::AnyError;
|
|
use deno_core::serde_json;
|
|
use deno_runtime::deno_webstorage::rusqlite::params;
|
|
|
|
use crate::node::CliCjsAnalysis;
|
|
|
|
use super::cache_db::CacheDB;
|
|
use super::cache_db::CacheDBConfiguration;
|
|
use super::cache_db::CacheFailure;
|
|
use super::CacheDBHash;
|
|
|
|
pub static NODE_ANALYSIS_CACHE_DB: CacheDBConfiguration =
|
|
CacheDBConfiguration {
|
|
table_initializer: concat!(
|
|
"CREATE TABLE IF NOT EXISTS cjsanalysiscache (",
|
|
"specifier TEXT PRIMARY KEY,",
|
|
"source_hash INTEGER NOT NULL,",
|
|
"data TEXT NOT NULL",
|
|
");"
|
|
),
|
|
on_version_change: "DELETE FROM cjsanalysiscache;",
|
|
preheat_queries: &[],
|
|
on_failure: CacheFailure::InMemory,
|
|
};
|
|
|
|
#[derive(Clone)]
|
|
pub struct NodeAnalysisCache {
|
|
inner: NodeAnalysisCacheInner,
|
|
}
|
|
|
|
impl NodeAnalysisCache {
|
|
pub fn new(db: CacheDB) -> Self {
|
|
Self {
|
|
inner: NodeAnalysisCacheInner::new(db),
|
|
}
|
|
}
|
|
|
|
fn ensure_ok<T: Default>(res: Result<T, AnyError>) -> T {
|
|
match res {
|
|
Ok(x) => x,
|
|
Err(err) => {
|
|
// TODO(mmastrac): This behavior was inherited from before the refactoring but it probably makes sense to move it into the cache
|
|
// at some point.
|
|
// should never error here, but if it ever does don't fail
|
|
if cfg!(debug_assertions) {
|
|
panic!("Error using esm analysis: {err:#}");
|
|
} else {
|
|
log::debug!("Error using esm analysis: {:#}", err);
|
|
}
|
|
T::default()
|
|
}
|
|
}
|
|
}
|
|
|
|
pub fn get_cjs_analysis(
|
|
&self,
|
|
specifier: &str,
|
|
expected_source_hash: CacheDBHash,
|
|
) -> Option<CliCjsAnalysis> {
|
|
Self::ensure_ok(
|
|
self.inner.get_cjs_analysis(specifier, expected_source_hash),
|
|
)
|
|
}
|
|
|
|
pub fn set_cjs_analysis(
|
|
&self,
|
|
specifier: &str,
|
|
source_hash: CacheDBHash,
|
|
cjs_analysis: &CliCjsAnalysis,
|
|
) {
|
|
Self::ensure_ok(self.inner.set_cjs_analysis(
|
|
specifier,
|
|
source_hash,
|
|
cjs_analysis,
|
|
));
|
|
}
|
|
}
|
|
|
|
#[derive(Clone)]
|
|
struct NodeAnalysisCacheInner {
|
|
conn: CacheDB,
|
|
}
|
|
|
|
impl NodeAnalysisCacheInner {
|
|
pub fn new(conn: CacheDB) -> Self {
|
|
Self { conn }
|
|
}
|
|
|
|
pub fn get_cjs_analysis(
|
|
&self,
|
|
specifier: &str,
|
|
expected_source_hash: CacheDBHash,
|
|
) -> Result<Option<CliCjsAnalysis>, AnyError> {
|
|
let query = "
|
|
SELECT
|
|
data
|
|
FROM
|
|
cjsanalysiscache
|
|
WHERE
|
|
specifier=?1
|
|
AND source_hash=?2
|
|
LIMIT 1";
|
|
let res = self.conn.query_row(
|
|
query,
|
|
params![specifier, expected_source_hash],
|
|
|row| {
|
|
let analysis_info: String = row.get(0)?;
|
|
Ok(serde_json::from_str(&analysis_info)?)
|
|
},
|
|
)?;
|
|
Ok(res)
|
|
}
|
|
|
|
pub fn set_cjs_analysis(
|
|
&self,
|
|
specifier: &str,
|
|
source_hash: CacheDBHash,
|
|
cjs_analysis: &CliCjsAnalysis,
|
|
) -> Result<(), AnyError> {
|
|
let sql = "
|
|
INSERT OR REPLACE INTO
|
|
cjsanalysiscache (specifier, source_hash, data)
|
|
VALUES
|
|
(?1, ?2, ?3)";
|
|
self.conn.execute(
|
|
sql,
|
|
params![
|
|
specifier,
|
|
source_hash,
|
|
&serde_json::to_string(&cjs_analysis)?,
|
|
],
|
|
)?;
|
|
Ok(())
|
|
}
|
|
}
|
|
|
|
#[cfg(test)]
|
|
mod test {
|
|
use super::*;
|
|
|
|
#[test]
|
|
pub fn node_analysis_cache_general_use() {
|
|
let conn = CacheDB::in_memory(&NODE_ANALYSIS_CACHE_DB, "1.0.0");
|
|
let cache = NodeAnalysisCacheInner::new(conn);
|
|
|
|
assert!(cache
|
|
.get_cjs_analysis("file.js", CacheDBHash::new(2))
|
|
.unwrap()
|
|
.is_none());
|
|
let cjs_analysis = CliCjsAnalysis::Cjs {
|
|
exports: vec!["export1".to_string()],
|
|
reexports: vec!["re-export1".to_string()],
|
|
};
|
|
cache
|
|
.set_cjs_analysis("file.js", CacheDBHash::new(2), &cjs_analysis)
|
|
.unwrap();
|
|
assert!(cache
|
|
.get_cjs_analysis("file.js", CacheDBHash::new(3))
|
|
.unwrap()
|
|
.is_none()); // different hash
|
|
let actual_cjs_analysis = cache
|
|
.get_cjs_analysis("file.js", CacheDBHash::new(2))
|
|
.unwrap()
|
|
.unwrap();
|
|
assert_eq!(actual_cjs_analysis, cjs_analysis);
|
|
|
|
// adding when already exists should not cause issue
|
|
cache
|
|
.set_cjs_analysis("file.js", CacheDBHash::new(2), &cjs_analysis)
|
|
.unwrap();
|
|
|
|
// recreating with same cli version should still have it
|
|
let conn = cache.conn.recreate_with_version("1.0.0");
|
|
let cache = NodeAnalysisCacheInner::new(conn);
|
|
let actual_analysis = cache
|
|
.get_cjs_analysis("file.js", CacheDBHash::new(2))
|
|
.unwrap()
|
|
.unwrap();
|
|
assert_eq!(actual_analysis, cjs_analysis);
|
|
|
|
// now changing the cli version should clear it
|
|
let conn = cache.conn.recreate_with_version("2.0.0");
|
|
let cache = NodeAnalysisCacheInner::new(conn);
|
|
assert!(cache
|
|
.get_cjs_analysis("file.js", CacheDBHash::new(2))
|
|
.unwrap()
|
|
.is_none());
|
|
}
|
|
}
|