mirror of
https://github.com/denoland/deno.git
synced 2024-12-14 19:37:59 -05:00
7c3c13cecf
Fixes #26085. Adds a basic retry utility with some defaults, starts off with a 100ms wait, then 250ms, then 500ms I've applied the retry in the http client, reusing an existing function, so this also applies to retrying downloads of deno binaries in `upgrade` and `compile`. I can make a separate function that doesn't retry so this doesn't affect `upgrade` and `compile`, but it seemed desirable to have retries there too, so I left it in.
235 lines
7.7 KiB
Rust
235 lines
7.7 KiB
Rust
// Copyright 2018-2024 the Deno authors. All rights reserved. MIT license.
|
|
|
|
use std::collections::HashMap;
|
|
use std::sync::Arc;
|
|
|
|
use deno_core::anyhow::anyhow;
|
|
use deno_core::anyhow::bail;
|
|
use deno_core::anyhow::Context;
|
|
use deno_core::error::custom_error;
|
|
use deno_core::error::AnyError;
|
|
use deno_core::futures::future::LocalBoxFuture;
|
|
use deno_core::futures::FutureExt;
|
|
use deno_core::parking_lot::Mutex;
|
|
use deno_core::url::Url;
|
|
use deno_npm::npm_rc::ResolvedNpmRc;
|
|
use deno_npm::registry::NpmPackageVersionDistInfo;
|
|
use deno_runtime::deno_fs::FileSystem;
|
|
use deno_semver::package::PackageNv;
|
|
use http::StatusCode;
|
|
|
|
use crate::args::CacheSetting;
|
|
use crate::http_util::DownloadError;
|
|
use crate::http_util::HttpClientProvider;
|
|
use crate::npm::common::maybe_auth_header_for_npm_registry;
|
|
use crate::util::progress_bar::ProgressBar;
|
|
use crate::util::sync::MultiRuntimeAsyncValueCreator;
|
|
|
|
use super::tarball_extract::verify_and_extract_tarball;
|
|
use super::tarball_extract::TarballExtractionMode;
|
|
use super::NpmCache;
|
|
|
|
// todo(dsherret): create seams and unit test this
|
|
|
|
type LoadResult = Result<(), Arc<AnyError>>;
|
|
type LoadFuture = LocalBoxFuture<'static, LoadResult>;
|
|
|
|
#[derive(Debug, Clone)]
|
|
enum MemoryCacheItem {
|
|
/// The cache item hasn't finished yet.
|
|
Pending(Arc<MultiRuntimeAsyncValueCreator<LoadResult>>),
|
|
/// The result errored.
|
|
Errored(Arc<AnyError>),
|
|
/// This package has already been cached.
|
|
Cached,
|
|
}
|
|
|
|
/// Coordinates caching of tarballs being loaded from
|
|
/// the npm registry.
|
|
///
|
|
/// This is shared amongst all the workers.
|
|
#[derive(Debug)]
|
|
pub struct TarballCache {
|
|
cache: Arc<NpmCache>,
|
|
fs: Arc<dyn FileSystem>,
|
|
http_client_provider: Arc<HttpClientProvider>,
|
|
npmrc: Arc<ResolvedNpmRc>,
|
|
progress_bar: ProgressBar,
|
|
memory_cache: Mutex<HashMap<PackageNv, MemoryCacheItem>>,
|
|
}
|
|
|
|
impl TarballCache {
|
|
pub fn new(
|
|
cache: Arc<NpmCache>,
|
|
fs: Arc<dyn FileSystem>,
|
|
http_client_provider: Arc<HttpClientProvider>,
|
|
npmrc: Arc<ResolvedNpmRc>,
|
|
progress_bar: ProgressBar,
|
|
) -> Self {
|
|
Self {
|
|
cache,
|
|
fs,
|
|
http_client_provider,
|
|
npmrc,
|
|
progress_bar,
|
|
memory_cache: Default::default(),
|
|
}
|
|
}
|
|
|
|
pub async fn ensure_package(
|
|
self: &Arc<Self>,
|
|
package: &PackageNv,
|
|
dist: &NpmPackageVersionDistInfo,
|
|
) -> Result<(), AnyError> {
|
|
self
|
|
.ensure_package_inner(package, dist)
|
|
.await
|
|
.with_context(|| format!("Failed caching npm package '{}'.", package))
|
|
}
|
|
|
|
async fn ensure_package_inner(
|
|
self: &Arc<Self>,
|
|
package_nv: &PackageNv,
|
|
dist: &NpmPackageVersionDistInfo,
|
|
) -> Result<(), AnyError> {
|
|
let cache_item = {
|
|
let mut mem_cache = self.memory_cache.lock();
|
|
if let Some(cache_item) = mem_cache.get(package_nv) {
|
|
cache_item.clone()
|
|
} else {
|
|
let value_creator = MultiRuntimeAsyncValueCreator::new({
|
|
let tarball_cache = self.clone();
|
|
let package_nv = package_nv.clone();
|
|
let dist = dist.clone();
|
|
Box::new(move || {
|
|
tarball_cache.create_setup_future(package_nv.clone(), dist.clone())
|
|
})
|
|
});
|
|
let cache_item = MemoryCacheItem::Pending(Arc::new(value_creator));
|
|
mem_cache.insert(package_nv.clone(), cache_item.clone());
|
|
cache_item
|
|
}
|
|
};
|
|
|
|
match cache_item {
|
|
MemoryCacheItem::Cached => Ok(()),
|
|
MemoryCacheItem::Errored(err) => Err(anyhow!("{}", err)),
|
|
MemoryCacheItem::Pending(creator) => {
|
|
let result = creator.get().await;
|
|
match result {
|
|
Ok(_) => {
|
|
*self.memory_cache.lock().get_mut(package_nv).unwrap() =
|
|
MemoryCacheItem::Cached;
|
|
Ok(())
|
|
}
|
|
Err(err) => {
|
|
let result_err = anyhow!("{}", err);
|
|
*self.memory_cache.lock().get_mut(package_nv).unwrap() =
|
|
MemoryCacheItem::Errored(err);
|
|
Err(result_err)
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
fn create_setup_future(
|
|
self: &Arc<Self>,
|
|
package_nv: PackageNv,
|
|
dist: NpmPackageVersionDistInfo,
|
|
) -> LoadFuture {
|
|
let tarball_cache = self.clone();
|
|
async move {
|
|
let registry_url = tarball_cache.npmrc.get_registry_url(&package_nv.name);
|
|
let package_folder =
|
|
tarball_cache.cache.package_folder_for_nv_and_url(&package_nv, registry_url);
|
|
let should_use_cache = tarball_cache.cache.should_use_cache_for_package(&package_nv);
|
|
let package_folder_exists = tarball_cache.fs.exists_sync(&package_folder);
|
|
if should_use_cache && package_folder_exists {
|
|
return Ok(());
|
|
} else if tarball_cache.cache.cache_setting() == &CacheSetting::Only {
|
|
return Err(custom_error(
|
|
"NotCached",
|
|
format!(
|
|
"An npm specifier not found in cache: \"{}\", --cached-only is specified.",
|
|
&package_nv.name
|
|
)
|
|
)
|
|
);
|
|
}
|
|
|
|
if dist.tarball.is_empty() {
|
|
bail!("Tarball URL was empty.");
|
|
}
|
|
|
|
// IMPORTANT: npm registries may specify tarball URLs at different URLS than the
|
|
// registry, so we MUST get the auth for the tarball URL and not the registry URL.
|
|
let tarball_uri = Url::parse(&dist.tarball)?;
|
|
let maybe_registry_config =
|
|
tarball_cache.npmrc.tarball_config(&tarball_uri);
|
|
let maybe_auth_header = maybe_registry_config.and_then(|c| maybe_auth_header_for_npm_registry(c).ok()?);
|
|
|
|
let guard = tarball_cache.progress_bar.update(&dist.tarball);
|
|
let result = tarball_cache.http_client_provider
|
|
.get_or_create()?
|
|
.download_with_progress_and_retries(tarball_uri, maybe_auth_header, &guard)
|
|
.await;
|
|
let maybe_bytes = match result {
|
|
Ok(maybe_bytes) => maybe_bytes,
|
|
Err(DownloadError::BadResponse(err)) => {
|
|
if err.status_code == StatusCode::UNAUTHORIZED
|
|
&& maybe_registry_config.is_none()
|
|
&& tarball_cache.npmrc.get_registry_config(&package_nv.name).auth_token.is_some()
|
|
{
|
|
bail!(
|
|
concat!(
|
|
"No auth for tarball URI, but present for scoped registry.\n\n",
|
|
"Tarball URI: {}\n",
|
|
"Scope URI: {}\n\n",
|
|
"More info here: https://github.com/npm/cli/wiki/%22No-auth-for-URI,-but-auth-present-for-scoped-registry%22"
|
|
),
|
|
dist.tarball,
|
|
registry_url,
|
|
)
|
|
}
|
|
return Err(err.into())
|
|
},
|
|
Err(err) => return Err(err.into()),
|
|
};
|
|
match maybe_bytes {
|
|
Some(bytes) => {
|
|
let extraction_mode = if should_use_cache || !package_folder_exists {
|
|
TarballExtractionMode::SiblingTempDir
|
|
} else {
|
|
// The user ran with `--reload`, so overwrite the package instead of
|
|
// deleting it since the package might get corrupted if a user kills
|
|
// their deno process while it's deleting a package directory
|
|
//
|
|
// We can't rename this folder and delete it because the folder
|
|
// may be in use by another process or may now contain hardlinks,
|
|
// which will cause windows to throw an "AccessDenied" error when
|
|
// renaming. So we settle for overwriting.
|
|
TarballExtractionMode::Overwrite
|
|
};
|
|
let dist = dist.clone();
|
|
let package_nv = package_nv.clone();
|
|
deno_core::unsync::spawn_blocking(move || {
|
|
verify_and_extract_tarball(
|
|
&package_nv,
|
|
&bytes,
|
|
&dist,
|
|
&package_folder,
|
|
extraction_mode,
|
|
)
|
|
})
|
|
.await?
|
|
}
|
|
None => {
|
|
bail!("Could not find npm package tarball at: {}", dist.tarball);
|
|
}
|
|
}
|
|
}
|
|
.map(|r| r.map_err(Arc::new))
|
|
.boxed_local()
|
|
}
|
|
}
|