mirror of
https://github.com/denoland/rusty_v8.git
synced 2024-12-01 16:51:32 -05:00
1090 lines
32 KiB
Rust
1090 lines
32 KiB
Rust
// Copyright 2018-2019 the Deno authors. All rights reserved. MIT license.
|
|
use fslock::LockFile;
|
|
use miniz_oxide::inflate::stream::inflate;
|
|
use miniz_oxide::inflate::stream::InflateState;
|
|
use miniz_oxide::MZFlush;
|
|
use miniz_oxide::MZResult;
|
|
use miniz_oxide::MZStatus;
|
|
use miniz_oxide::StreamResult;
|
|
use std::collections::HashSet;
|
|
use std::env;
|
|
use std::fs;
|
|
use std::io;
|
|
use std::io::Read;
|
|
use std::io::Seek;
|
|
use std::io::Write;
|
|
use std::path::Path;
|
|
use std::path::PathBuf;
|
|
use std::process::Command;
|
|
use std::process::Stdio;
|
|
use which::which;
|
|
|
|
fn main() {
|
|
println!("cargo:rerun-if-changed=.gn");
|
|
println!("cargo:rerun-if-changed=BUILD.gn");
|
|
println!("cargo:rerun-if-changed=src/binding.cc");
|
|
|
|
// These are all the environment variables that we check. This is
|
|
// probably more than what is needed, but missing an important
|
|
// variable can lead to broken links when switching rusty_v8
|
|
// versions.
|
|
let envs = vec![
|
|
"CCACHE",
|
|
"CLANG_BASE_PATH",
|
|
"CXXSTDLIB",
|
|
"DENO_TRYBUILD",
|
|
"DOCS_RS",
|
|
"GN",
|
|
"GN_ARGS",
|
|
"HOST",
|
|
"NINJA",
|
|
"OUT_DIR",
|
|
"RUSTY_V8_ARCHIVE",
|
|
"RUSTY_V8_MIRROR",
|
|
"RUSTY_V8_SRC_BINDING_PATH",
|
|
"SCCACHE",
|
|
"V8_FORCE_DEBUG",
|
|
"V8_FROM_SOURCE",
|
|
"PYTHON",
|
|
"DISABLE_CLANG",
|
|
"EXTRA_GN_ARGS",
|
|
"PRINT_GN_ARGS",
|
|
"CARGO_ENCODED_RUSTFLAGS",
|
|
];
|
|
for env in envs {
|
|
println!("cargo:rerun-if-env-changed={}", env);
|
|
}
|
|
|
|
// Detect if trybuild tests are being compiled.
|
|
let is_trybuild = env::var_os("DENO_TRYBUILD").is_some();
|
|
|
|
// Don't build V8 if "cargo doc" is being run. This is to support docs.rs.
|
|
let is_cargo_doc = env::var_os("DOCS_RS").is_some();
|
|
|
|
// Don't build V8 if the rust language server (RLS) is running.
|
|
let is_rls = env::var_os("CARGO")
|
|
.map(PathBuf::from)
|
|
.as_ref()
|
|
.and_then(|p| p.file_stem())
|
|
.and_then(|f| f.to_str())
|
|
.map(|s| s.starts_with("rls"))
|
|
.unwrap_or(false);
|
|
|
|
// Early exit
|
|
if is_cargo_doc || is_rls {
|
|
print_prebuilt_src_binding_path();
|
|
return;
|
|
}
|
|
|
|
print_link_flags();
|
|
|
|
// Don't attempt rebuild but link
|
|
if is_trybuild {
|
|
println!(
|
|
"cargo:rustc-env=RUSTY_V8_SRC_BINDING_PATH={}",
|
|
env::var("RUSTY_V8_SRC_BINDING_PATH").unwrap()
|
|
);
|
|
return;
|
|
}
|
|
|
|
let is_asan = if let Some(rustflags) = env::var_os("CARGO_ENCODED_RUSTFLAGS")
|
|
{
|
|
let rustflags = rustflags.to_string_lossy();
|
|
rustflags.find("-Z sanitizer=address").is_some()
|
|
|| rustflags.find("-Zsanitizer=address").is_some()
|
|
} else {
|
|
false
|
|
};
|
|
|
|
// Cargo likes to run multiple build scripts at once sometimes.
|
|
// Nothing that follows is safe to run multiple times at once,
|
|
// because we store everything in a parent directory of OUT_DIR.
|
|
let _lockfile = acquire_lock();
|
|
|
|
// Build from source
|
|
if env_bool("V8_FROM_SOURCE") {
|
|
if is_asan && std::env::var_os("OPT_LEVEL").unwrap_or_default() == "0" {
|
|
panic!("v8 crate cannot be compiled with OPT_LEVEL=0 and ASAN.\nTry `[profile.dev.package.v8] opt-level = 1`.\nAborting before miscompilations cause issues.");
|
|
}
|
|
|
|
// cargo publish doesn't like pyc files.
|
|
env::set_var("PYTHONDONTWRITEBYTECODE", "1");
|
|
|
|
build_v8(is_asan);
|
|
build_binding();
|
|
|
|
return;
|
|
}
|
|
|
|
print_prebuilt_src_binding_path();
|
|
|
|
download_static_lib_binaries();
|
|
}
|
|
|
|
fn acquire_lock() -> LockFile {
|
|
let root = env::current_dir().unwrap();
|
|
let out_dir = env::var_os("OUT_DIR").unwrap();
|
|
let lockfilepath = root
|
|
.join(out_dir)
|
|
.parent()
|
|
.unwrap()
|
|
.parent()
|
|
.unwrap()
|
|
.join("v8.fslock");
|
|
let mut lockfile = LockFile::open(&lockfilepath)
|
|
.expect("Couldn't open lib download lockfile.");
|
|
lockfile.lock_with_pid().expect("Couldn't get lock");
|
|
println!("lockfile: {:?}", &lockfilepath);
|
|
lockfile
|
|
}
|
|
|
|
fn build_binding() {
|
|
let output = Command::new(python())
|
|
.arg("./tools/get_bindgen_args.py")
|
|
.arg("--gn-out")
|
|
.arg(build_dir().join("gn_out"))
|
|
.output()
|
|
.unwrap();
|
|
let args = String::from_utf8(output.stdout).unwrap();
|
|
let args = args.split('\0').collect::<Vec<_>>();
|
|
|
|
let bindings = bindgen::Builder::default()
|
|
.header("src/binding.hpp")
|
|
.parse_callbacks(Box::new(bindgen::CargoCallbacks::new()))
|
|
.clang_args(["-x", "c++", "-std=c++20", "-Iv8/include", "-I."])
|
|
.clang_args(args)
|
|
.generate_cstr(true)
|
|
.rustified_enum(".*UseCounterFeature")
|
|
.allowlist_item("v8__.*")
|
|
.allowlist_item("cppgc__.*")
|
|
.allowlist_item("RustObj")
|
|
.generate()
|
|
.expect("Unable to generate bindings");
|
|
|
|
let out_path = build_dir().join("gn_out").join("src_binding.rs");
|
|
println!(
|
|
"cargo:rustc-env=RUSTY_V8_SRC_BINDING_PATH={}",
|
|
out_path.display()
|
|
);
|
|
bindings
|
|
.write_to_file(out_path)
|
|
.expect("Couldn't write bindings!");
|
|
}
|
|
|
|
fn build_v8(is_asan: bool) {
|
|
env::set_var("DEPOT_TOOLS_WIN_TOOLCHAIN", "0");
|
|
|
|
if need_gn_ninja_download() {
|
|
download_ninja_gn_binaries();
|
|
}
|
|
// `#[cfg(...)]` attributes don't work as expected from build.rs -- they refer to the configuration
|
|
// of the host system which the build.rs script will be running on. In short, `cfg!(target_<os/arch>)`
|
|
// is actually the host os/arch instead of target os/arch while cross compiling. Instead, Environment variables
|
|
// are the officially approach to get the target os/arch in build.rs.
|
|
let target_os = env::var("CARGO_CFG_TARGET_OS").unwrap();
|
|
let target_arch = env::var("CARGO_CFG_TARGET_ARCH").unwrap();
|
|
// On windows, rustc cannot link with a V8 debug build.
|
|
let mut gn_args = if is_debug() && target_os != "windows" {
|
|
// Note: When building for Android aarch64-qemu, use release instead of debug.
|
|
vec!["is_debug=true".to_string()]
|
|
} else {
|
|
vec!["is_debug=false".to_string()]
|
|
};
|
|
if is_asan {
|
|
gn_args.push("is_asan=true".to_string());
|
|
}
|
|
if env::var("CARGO_FEATURE_USE_CUSTOM_LIBCXX").is_err() {
|
|
gn_args.push("use_custom_libcxx=false".to_string());
|
|
}
|
|
|
|
// Fix GN's host_cpu detection when using x86_64 bins on Apple Silicon
|
|
if cfg!(target_os = "macos") && cfg!(target_arch = "aarch64") {
|
|
gn_args.push("host_cpu=\"arm64\"".to_string())
|
|
}
|
|
|
|
if env::var_os("DISABLE_CLANG").is_some() {
|
|
gn_args.push("is_clang=false".into());
|
|
// -gline-tables-only is Clang-only
|
|
gn_args.push("line_tables_only=false".into());
|
|
} else if let Some(clang_base_path) = find_compatible_system_clang() {
|
|
println!("clang_base_path (system): {}", clang_base_path.display());
|
|
gn_args.push(format!("clang_base_path={:?}", clang_base_path));
|
|
gn_args.push("treat_warnings_as_errors=false".to_string());
|
|
} else {
|
|
println!("using Chromium's clang");
|
|
let clang_base_path = clang_download();
|
|
gn_args.push(format!("clang_base_path={:?}", clang_base_path));
|
|
|
|
if target_os == "android" && target_arch == "aarch64" {
|
|
gn_args.push("treat_warnings_as_errors=false".to_string());
|
|
}
|
|
}
|
|
|
|
if let Some(p) = env::var_os("SCCACHE") {
|
|
cc_wrapper(&mut gn_args, Path::new(&p));
|
|
} else if let Ok(p) = which("sccache") {
|
|
cc_wrapper(&mut gn_args, &p);
|
|
} else if let Some(p) = env::var_os("CCACHE") {
|
|
cc_wrapper(&mut gn_args, Path::new(&p));
|
|
} else if let Ok(p) = which("ccache") {
|
|
cc_wrapper(&mut gn_args, &p);
|
|
} else {
|
|
println!("cargo:warning=Not using sccache or ccache");
|
|
}
|
|
|
|
if let Ok(args) = env::var("GN_ARGS") {
|
|
for arg in args.split_whitespace() {
|
|
gn_args.push(arg.to_string());
|
|
}
|
|
}
|
|
// cross-compilation setup
|
|
if target_arch == "aarch64" {
|
|
gn_args.push(r#"target_cpu="arm64""#.to_string());
|
|
gn_args.push("use_sysroot=true".to_string());
|
|
maybe_install_sysroot("arm64");
|
|
maybe_install_sysroot("amd64");
|
|
}
|
|
if target_arch == "arm" {
|
|
gn_args.push(r#"target_cpu="arm""#.to_string());
|
|
gn_args.push(r#"v8_target_cpu="arm""#.to_string());
|
|
gn_args.push("use_sysroot=true".to_string());
|
|
maybe_install_sysroot("i386");
|
|
maybe_install_sysroot("arm");
|
|
}
|
|
|
|
let target_triple = env::var("TARGET").unwrap();
|
|
// check if the target triple describes a non-native environment
|
|
if target_triple != env::var("HOST").unwrap() && target_os == "android" {
|
|
let arch = if target_arch == "x86_64" {
|
|
"x64"
|
|
} else if target_arch == "aarch64" {
|
|
"arm64"
|
|
} else {
|
|
"unknown"
|
|
};
|
|
if target_arch == "x86_64" {
|
|
maybe_install_sysroot("amd64");
|
|
}
|
|
gn_args.push(format!(r#"v8_target_cpu="{}""#, arch).to_string());
|
|
gn_args.push(format!(r#"target_cpu="{}""#, arch).to_string());
|
|
gn_args.push(r#"target_os="android""#.to_string());
|
|
gn_args.push("treat_warnings_as_errors=false".to_string());
|
|
gn_args.push("use_sysroot=true".to_string());
|
|
|
|
// NDK 23 and above removes libgcc entirely.
|
|
// https://github.com/rust-lang/rust/pull/85806
|
|
if !Path::new("./third_party/android_ndk/toolchains/llvm/prebuilt/linux-x86_64/bin/aarch64-linux-android24-clang++").exists() {
|
|
assert!(Command::new("curl")
|
|
.arg("-L")
|
|
.arg("-o").arg("./third_party/android-ndk-r26c-linux.zip")
|
|
.arg("https://dl.google.com/android/repository/android-ndk-r26c-linux.zip")
|
|
.status()
|
|
.unwrap()
|
|
.success());
|
|
|
|
assert!(Command::new("unzip")
|
|
.arg("-d").arg("./third_party/")
|
|
.arg("-o")
|
|
.arg("-q")
|
|
.arg("./third_party/android-ndk-r26c-linux.zip")
|
|
.status()
|
|
.unwrap()
|
|
.success());
|
|
|
|
fs::rename("./third_party/android-ndk-r26c", "./third_party/android_ndk").unwrap();
|
|
fs::remove_file("./third_party/android-ndk-r26c-linux.zip").unwrap();
|
|
}
|
|
static CHROMIUM_URI: &str = "https://chromium.googlesource.com";
|
|
maybe_clone_repo(
|
|
"./third_party/android_platform",
|
|
&format!(
|
|
"{}/chromium/src/third_party/android_platform.git",
|
|
CHROMIUM_URI
|
|
),
|
|
);
|
|
maybe_clone_repo(
|
|
"./third_party/catapult",
|
|
&format!("{}/catapult.git", CHROMIUM_URI),
|
|
);
|
|
}
|
|
|
|
if target_triple.starts_with("i686-") {
|
|
gn_args.push(r#"target_cpu="x86""#.to_string());
|
|
}
|
|
|
|
let gn_out = maybe_gen(gn_args);
|
|
assert!(gn_out.exists());
|
|
assert!(gn_out.join("args.gn").exists());
|
|
if env_bool("PRINT_GN_ARGS") {
|
|
print_gn_args(&gn_out);
|
|
}
|
|
build("rusty_v8", None);
|
|
}
|
|
|
|
fn print_gn_args(gn_out_dir: &Path) {
|
|
assert!(Command::new(gn())
|
|
.arg(format!("--script-executable={}", python()))
|
|
.arg("args")
|
|
.arg(gn_out_dir)
|
|
.arg("--list")
|
|
.status()
|
|
.unwrap()
|
|
.success());
|
|
}
|
|
|
|
fn maybe_clone_repo(dest: &str, repo: &str) {
|
|
if !Path::new(&dest).exists() {
|
|
assert!(Command::new("git")
|
|
.arg("clone")
|
|
.arg("--depth=1")
|
|
.arg(repo)
|
|
.arg(dest)
|
|
.status()
|
|
.unwrap()
|
|
.success());
|
|
}
|
|
}
|
|
|
|
fn maybe_install_sysroot(arch: &str) {
|
|
let sysroot_path = format!("build/linux/debian_sid_{}-sysroot", arch);
|
|
if !PathBuf::from(sysroot_path).is_dir() {
|
|
assert!(Command::new(python())
|
|
.arg("./build/linux/sysroot_scripts/install-sysroot.py")
|
|
.arg(format!("--arch={}", arch))
|
|
.status()
|
|
.unwrap()
|
|
.success());
|
|
}
|
|
}
|
|
|
|
fn download_ninja_gn_binaries() {
|
|
let target_dir = build_dir().join("ninja_gn_binaries");
|
|
|
|
let gn = target_dir.join("gn").join("gn");
|
|
let ninja = target_dir.join("ninja").join("ninja");
|
|
#[cfg(windows)]
|
|
let gn = gn.with_extension("exe");
|
|
#[cfg(windows)]
|
|
let ninja = ninja.with_extension("exe");
|
|
|
|
if !gn.exists() || !ninja.exists() {
|
|
assert!(Command::new(python())
|
|
.arg("./tools/ninja_gn_binaries.py")
|
|
.arg("--dir")
|
|
.arg(&target_dir)
|
|
.status()
|
|
.unwrap()
|
|
.success());
|
|
}
|
|
assert!(gn.exists());
|
|
assert!(ninja.exists());
|
|
env::set_var("GN", gn);
|
|
if env::var("NINJA").is_err() {
|
|
env::set_var("NINJA", ninja);
|
|
}
|
|
}
|
|
|
|
fn prebuilt_profile() -> &'static str {
|
|
let target_os = env::var("CARGO_CFG_TARGET_OS").unwrap();
|
|
// Note: we always use the release build on windows.
|
|
if target_os == "windows" {
|
|
return "release";
|
|
}
|
|
// Use v8 in release mode unless $V8_FORCE_DEBUG=true
|
|
match env_bool("V8_FORCE_DEBUG") {
|
|
true => "debug",
|
|
_ => "release",
|
|
}
|
|
}
|
|
|
|
fn static_lib_name(suffix: &str) -> String {
|
|
let target_os = env::var("CARGO_CFG_TARGET_OS").unwrap();
|
|
if target_os == "windows" {
|
|
format!("rusty_v8{suffix}.lib")
|
|
} else {
|
|
format!("librusty_v8{suffix}.a")
|
|
}
|
|
}
|
|
|
|
fn static_lib_url() -> String {
|
|
if let Ok(custom_archive) = env::var("RUSTY_V8_ARCHIVE") {
|
|
return custom_archive;
|
|
}
|
|
let default_base = "https://github.com/denoland/rusty_v8/releases/download";
|
|
let base =
|
|
env::var("RUSTY_V8_MIRROR").unwrap_or_else(|_| default_base.into());
|
|
let version = env::var("CARGO_PKG_VERSION").unwrap();
|
|
let target = env::var("TARGET").unwrap();
|
|
let profile = prebuilt_profile();
|
|
format!(
|
|
"{}/v{}/{}.gz",
|
|
base,
|
|
version,
|
|
static_lib_name(&format!("_{}_{}", profile, target)),
|
|
)
|
|
}
|
|
|
|
fn static_lib_path() -> PathBuf {
|
|
static_lib_dir().join(static_lib_name(""))
|
|
}
|
|
|
|
fn static_checksum_path() -> PathBuf {
|
|
let mut t = static_lib_path();
|
|
t.set_extension("sum");
|
|
t
|
|
}
|
|
|
|
fn static_lib_dir() -> PathBuf {
|
|
build_dir().join("gn_out").join("obj")
|
|
}
|
|
|
|
fn build_dir() -> PathBuf {
|
|
let cwd = env::current_dir().unwrap();
|
|
|
|
// target/debug//build/rusty_v8-d9e5a424d4f96994/out/
|
|
let out_dir = env::var_os("OUT_DIR").expect(
|
|
"The 'OUT_DIR' environment is not set (it should be something like \
|
|
'target/debug/rusty_v8-{hash}').",
|
|
);
|
|
let out_dir_abs = cwd.join(out_dir);
|
|
|
|
// This would be target/debug or target/release
|
|
out_dir_abs
|
|
.parent()
|
|
.unwrap()
|
|
.parent()
|
|
.unwrap()
|
|
.parent()
|
|
.unwrap()
|
|
.to_path_buf()
|
|
}
|
|
|
|
fn replace_non_alphanumeric(url: &str) -> String {
|
|
url
|
|
.chars()
|
|
.map(|c| if c.is_ascii_alphanumeric() { c } else { '_' })
|
|
.collect()
|
|
}
|
|
|
|
fn download_file(url: String, filename: PathBuf) {
|
|
if !url.starts_with("http:") && !url.starts_with("https:") {
|
|
copy_archive(&url, &filename);
|
|
return;
|
|
}
|
|
|
|
// If there is a `.cargo/.rusty_v8/<escaped URL>` file, use that instead
|
|
// of downloading.
|
|
if let Ok(mut path) = home::cargo_home() {
|
|
path = path.join(".rusty_v8").join(replace_non_alphanumeric(&url));
|
|
println!("Looking for download in '{path:?}'");
|
|
if path.exists() {
|
|
copy_archive(&path.to_string_lossy(), &filename);
|
|
return;
|
|
}
|
|
}
|
|
|
|
// tmp file to download to so we don't clobber the existing one
|
|
let tmpfile = {
|
|
let mut t = filename.clone();
|
|
t.set_extension("tmp");
|
|
t
|
|
};
|
|
if tmpfile.exists() {
|
|
println!("Deleting old tmpfile {}", tmpfile.display());
|
|
std::fs::remove_file(&tmpfile).unwrap();
|
|
}
|
|
|
|
// Try downloading with python first. Python is a V8 build dependency,
|
|
// so this saves us from adding a Rust HTTP client dependency.
|
|
println!("Downloading (using Python) {}", url);
|
|
let status = Command::new(python())
|
|
.arg("./tools/download_file.py")
|
|
.arg("--url")
|
|
.arg(&url)
|
|
.arg("--filename")
|
|
.arg(&tmpfile)
|
|
.status();
|
|
|
|
// Python is only a required dependency for `V8_FROM_SOURCE` builds.
|
|
// If python is not available, try falling back to curl.
|
|
let status = match status {
|
|
Ok(status) if status.success() => status,
|
|
_ => {
|
|
println!("Python downloader failed, trying with curl.");
|
|
Command::new("curl")
|
|
.arg("-L")
|
|
.arg("-f")
|
|
.arg("-s")
|
|
.arg("-o")
|
|
.arg(&tmpfile)
|
|
.arg(&url)
|
|
.status()
|
|
.unwrap()
|
|
}
|
|
};
|
|
|
|
// Assert DL was successful
|
|
assert!(status.success());
|
|
assert!(tmpfile.exists());
|
|
|
|
// Write checksum (i.e url) & move file
|
|
std::fs::write(static_checksum_path(), url).unwrap();
|
|
copy_archive(&tmpfile.to_string_lossy(), &filename);
|
|
std::fs::remove_file(&tmpfile).unwrap();
|
|
|
|
assert!(filename.exists());
|
|
assert!(static_checksum_path().exists());
|
|
assert!(!tmpfile.exists());
|
|
}
|
|
|
|
fn download_static_lib_binaries() {
|
|
let url = static_lib_url();
|
|
println!("static lib URL: {}", url);
|
|
|
|
let dir = static_lib_dir();
|
|
std::fs::create_dir_all(&dir).unwrap();
|
|
println!("cargo:rustc-link-search={}", dir.display());
|
|
|
|
// Checksum (i.e: url) to avoid redownloads
|
|
match std::fs::read_to_string(static_checksum_path()) {
|
|
Ok(c) if c == static_lib_url() => return,
|
|
_ => {}
|
|
};
|
|
download_file(url, static_lib_path());
|
|
}
|
|
|
|
fn decompress_to_writer<R, W>(input: &mut R, output: &mut W) -> io::Result<()>
|
|
where
|
|
R: Read,
|
|
W: Write,
|
|
{
|
|
let mut inflate_state = InflateState::default();
|
|
let mut input_buffer = [0; 16 * 1024];
|
|
let mut output_buffer = [0; 16 * 1024];
|
|
let mut input_offset = 0;
|
|
|
|
// Skip the gzip header
|
|
gzip_header::read_gz_header(input).unwrap();
|
|
|
|
loop {
|
|
let bytes_read = input.read(&mut input_buffer[input_offset..])?;
|
|
let bytes_avail = input_offset + bytes_read;
|
|
|
|
let StreamResult {
|
|
bytes_consumed,
|
|
bytes_written,
|
|
status,
|
|
} = inflate(
|
|
&mut inflate_state,
|
|
&input_buffer[..bytes_avail],
|
|
&mut output_buffer,
|
|
MZFlush::None,
|
|
);
|
|
|
|
if status != MZResult::Ok(MZStatus::Ok)
|
|
&& status != MZResult::Ok(MZStatus::StreamEnd)
|
|
{
|
|
return Err(io::Error::new(
|
|
io::ErrorKind::Other,
|
|
format!("Decompression error {status:?}"),
|
|
));
|
|
}
|
|
|
|
output.write_all(&output_buffer[..bytes_written])?;
|
|
|
|
// Move remaining bytes to the beginning of the buffer
|
|
input_buffer.copy_within(bytes_consumed..bytes_avail, 0);
|
|
input_offset = bytes_avail - bytes_consumed;
|
|
|
|
if status == MZResult::Ok(MZStatus::StreamEnd) {
|
|
break; // End of decompression
|
|
}
|
|
}
|
|
|
|
Ok(())
|
|
}
|
|
|
|
/// Copy the V8 archive at `url` to `filename`.
|
|
///
|
|
/// This function doesn't use `std::fs::copy` because that would
|
|
/// preveserve the file attributes such as ownership and mode flags.
|
|
/// Instead, it copies the file contents to a new file.
|
|
/// This is necessary because the V8 archive could live inside a read-only
|
|
/// filesystem, and subsequent builds would fail to overwrite it.
|
|
fn copy_archive(url: &str, filename: &Path) {
|
|
println!("Copying {url} to {filename:?}");
|
|
let mut src = fs::File::open(url).unwrap();
|
|
let mut dst = fs::File::create(filename).unwrap();
|
|
|
|
// Allow both GZIP and non-GZIP downloads
|
|
let mut header = [0; 2];
|
|
src.read_exact(&mut header).unwrap();
|
|
src.seek(io::SeekFrom::Start(0)).unwrap();
|
|
if header == [0x1f, 0x8b] {
|
|
println!("Detected GZIP archive");
|
|
decompress_to_writer(&mut src, &mut dst).unwrap();
|
|
} else {
|
|
println!("Not a GZIP archive");
|
|
io::copy(&mut src, &mut dst).unwrap();
|
|
}
|
|
}
|
|
|
|
fn print_link_flags() {
|
|
println!("cargo:rustc-link-lib=static=rusty_v8");
|
|
let should_dyn_link_libcxx = env::var("CARGO_FEATURE_USE_CUSTOM_LIBCXX")
|
|
.is_err()
|
|
|| env::var("GN_ARGS").map_or(false, |gn_args| {
|
|
gn_args
|
|
.split_whitespace()
|
|
.any(|ba| ba == "use_custom_libcxx=false")
|
|
});
|
|
|
|
if should_dyn_link_libcxx {
|
|
// Based on https://github.com/alexcrichton/cc-rs/blob/fba7feded71ee4f63cfe885673ead6d7b4f2f454/src/lib.rs#L2462
|
|
if let Ok(stdlib) = env::var("CXXSTDLIB") {
|
|
if !stdlib.is_empty() {
|
|
println!("cargo:rustc-link-lib=dylib={}", stdlib);
|
|
}
|
|
} else {
|
|
let target = env::var("TARGET").unwrap();
|
|
if target.contains("msvc") {
|
|
// nothing to link to
|
|
} else if target.contains("apple")
|
|
|| target.contains("freebsd")
|
|
|| target.contains("openbsd")
|
|
{
|
|
println!("cargo:rustc-link-lib=dylib=c++");
|
|
} else if target.contains("android") {
|
|
println!("cargo:rustc-link-lib=dylib=c++_shared");
|
|
} else {
|
|
println!("cargo:rustc-link-lib=dylib=stdc++");
|
|
}
|
|
}
|
|
}
|
|
let target_os = env::var("CARGO_CFG_TARGET_OS").unwrap();
|
|
let target_env = env::var("CARGO_CFG_TARGET_ENV").unwrap();
|
|
|
|
if target_os == "windows" {
|
|
println!("cargo:rustc-link-lib=dylib=winmm");
|
|
println!("cargo:rustc-link-lib=dylib=dbghelp");
|
|
}
|
|
|
|
if target_env == "msvc" {
|
|
// On Windows, including libcpmt[d]/msvcprt[d] explicitly links the C++
|
|
// standard library, which libc++ needs for exception_ptr internals.
|
|
if env::var("CARGO_FEATURE_CRT_STATIC").is_ok() {
|
|
println!("cargo:rustc-link-lib=libcpmt");
|
|
} else {
|
|
println!("cargo:rustc-link-lib=dylib=msvcprt");
|
|
}
|
|
}
|
|
}
|
|
|
|
fn print_prebuilt_src_binding_path() {
|
|
if let Ok(binding) = env::var("RUSTY_V8_SRC_BINDING_PATH") {
|
|
println!("cargo:rustc-env=RUSTY_V8_SRC_BINDING_PATH={}", binding);
|
|
return;
|
|
}
|
|
let target = env::var("TARGET").unwrap();
|
|
let profile = prebuilt_profile();
|
|
let src_binding_path = get_dirs()
|
|
.root
|
|
.join("gen")
|
|
.join(format!("src_binding_{}_{}.rs", profile, target));
|
|
println!(
|
|
"cargo:rustc-env=RUSTY_V8_SRC_BINDING_PATH={}",
|
|
src_binding_path.display()
|
|
);
|
|
}
|
|
|
|
// Chromium depot_tools contains helpers
|
|
// which delegate to the "relevant" `buildtools`
|
|
// directory when invoked, so they don't count.
|
|
fn not_in_depot_tools(p: PathBuf) -> bool {
|
|
!p.as_path().to_str().unwrap().contains("depot_tools")
|
|
}
|
|
|
|
fn need_gn_ninja_download() -> bool {
|
|
let has_ninja = which("ninja").map_or(false, not_in_depot_tools)
|
|
|| env::var_os("NINJA").is_some();
|
|
let has_gn = which("gn").map_or(false, not_in_depot_tools)
|
|
|| env::var_os("GN").is_some();
|
|
|
|
!has_ninja || !has_gn
|
|
}
|
|
|
|
// Chromiums gn arg clang_base_path is currently compatible with:
|
|
// * Apples clang and clang from homebrew's llvm@x packages
|
|
// * the official binaries from releases.llvm.org
|
|
// * unversioned (Linux) packages of clang (if recent enough)
|
|
// but unfortunately it doesn't work with version-suffixed packages commonly
|
|
// found in Linux packet managers
|
|
fn is_compatible_clang_version(clang_path: &Path) -> bool {
|
|
if let Ok(o) = Command::new(clang_path).arg("--version").output() {
|
|
let _output = String::from_utf8(o.stdout).unwrap();
|
|
// TODO check version output to make sure it's supported.
|
|
const _MIN_APPLE_CLANG_VER: f32 = 11.0;
|
|
const _MIN_LLVM_CLANG_VER: f32 = 8.0;
|
|
return true;
|
|
}
|
|
false
|
|
}
|
|
|
|
fn find_compatible_system_clang() -> Option<PathBuf> {
|
|
if let Ok(p) = env::var("CLANG_BASE_PATH") {
|
|
let base_path = Path::new(&p);
|
|
let clang_path = base_path.join("bin").join("clang");
|
|
if is_compatible_clang_version(&clang_path) {
|
|
return Some(base_path.to_path_buf());
|
|
}
|
|
}
|
|
|
|
None
|
|
}
|
|
|
|
// Download chromium's clang into OUT_DIR because Cargo will not allow us to
|
|
// modify the source directory.
|
|
fn clang_download() -> PathBuf {
|
|
let clang_base_path = build_dir().join("clang");
|
|
println!("clang_base_path (downloaded) {}", clang_base_path.display());
|
|
assert!(Command::new(python())
|
|
.arg("./tools/clang/scripts/update.py")
|
|
.arg("--output-dir")
|
|
.arg(&clang_base_path)
|
|
.status()
|
|
.unwrap()
|
|
.success());
|
|
assert!(clang_base_path.exists());
|
|
clang_base_path
|
|
}
|
|
|
|
fn cc_wrapper(gn_args: &mut Vec<String>, sccache_path: &Path) {
|
|
gn_args.push(format!("cc_wrapper={:?}", sccache_path));
|
|
}
|
|
|
|
struct Dirs {
|
|
pub out: PathBuf,
|
|
pub root: PathBuf,
|
|
}
|
|
|
|
fn get_dirs() -> Dirs {
|
|
// The OUT_DIR is going to be a crate-specific directory like
|
|
// "target/debug/build/cargo_gn_example-eee5160084460b2c"
|
|
// But we want to share the GN build amongst all crates
|
|
// and return the path "target/debug". So to find it, we walk up three
|
|
// directories.
|
|
// TODO(ry) This is quite brittle - if Cargo changes the directory structure
|
|
// this could break.
|
|
let out = env::var("OUT_DIR").map(PathBuf::from).unwrap();
|
|
let out = out
|
|
.parent()
|
|
.unwrap()
|
|
.parent()
|
|
.unwrap()
|
|
.parent()
|
|
.unwrap()
|
|
.to_owned();
|
|
|
|
let root = env::var("CARGO_MANIFEST_DIR").map(PathBuf::from).unwrap();
|
|
let mut dirs = Dirs { out, root };
|
|
maybe_symlink_root_dir(&mut dirs);
|
|
dirs
|
|
}
|
|
|
|
#[cfg(not(target_os = "windows"))]
|
|
fn maybe_symlink_root_dir(_: &mut Dirs) {}
|
|
|
|
#[cfg(target_os = "windows")]
|
|
fn maybe_symlink_root_dir(dirs: &mut Dirs) {
|
|
// GN produces invalid paths if the source (a.k.a. root) directory is on a
|
|
// different drive than the output. If this is the case we'll create a
|
|
// symlink called 'gn_root' in the out directory, next to 'gn_out', so it
|
|
// appears as if they're both on the same drive.
|
|
use std::fs::remove_dir_all;
|
|
use std::fs::remove_file;
|
|
use std::os::windows::fs::symlink_dir;
|
|
|
|
let get_prefix = |p: &Path| {
|
|
p.components()
|
|
.find_map(|c| match c {
|
|
std::path::Component::Prefix(p) => Some(p),
|
|
_ => None,
|
|
})
|
|
.map(|p| p.as_os_str().to_owned())
|
|
};
|
|
|
|
let Dirs { out, root } = dirs;
|
|
if get_prefix(out) != get_prefix(root) {
|
|
let symlink = &*out.join("gn_root");
|
|
let target = &*root.canonicalize().unwrap();
|
|
|
|
println!("Creating symlink {:?} to {:?}", &symlink, &root);
|
|
|
|
let mut retries = 0;
|
|
loop {
|
|
match symlink.canonicalize() {
|
|
Ok(existing) if existing == target => break,
|
|
Ok(_) => remove_dir_all(symlink).expect("remove_dir_all failed"),
|
|
Err(err) => {
|
|
println!("symlink.canonicalize failed: {:?}", err);
|
|
// we're having very strange issues on GHA when the cache
|
|
// is restored, so trying this out temporarily
|
|
if let Err(err) = remove_dir_all(symlink) {
|
|
eprintln!("remove_dir_all failed: {:?}", err);
|
|
if let Err(err) = remove_file(symlink) {
|
|
eprintln!("remove_file failed: {:?}", err);
|
|
}
|
|
}
|
|
match symlink_dir(target, symlink) {
|
|
Ok(_) => break,
|
|
Err(err) => {
|
|
println!("symlink_dir failed: {:?}", err);
|
|
retries += 1;
|
|
std::thread::sleep(std::time::Duration::from_millis(
|
|
50 * retries,
|
|
));
|
|
if retries > 4 {
|
|
panic!("Failed to create symlink");
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
dirs.root = symlink.to_path_buf();
|
|
}
|
|
}
|
|
|
|
pub fn is_debug() -> bool {
|
|
// Cargo sets PROFILE to either "debug" or "release", which conveniently
|
|
// matches the build modes we support.
|
|
let m = env::var("PROFILE").unwrap();
|
|
if m == "release" {
|
|
false
|
|
} else if m == "debug" {
|
|
true
|
|
} else {
|
|
panic!("unhandled PROFILE value {}", m)
|
|
}
|
|
}
|
|
|
|
fn gn() -> String {
|
|
env::var("GN").unwrap_or_else(|_| "gn".to_owned())
|
|
}
|
|
|
|
/*
|
|
* Get the system's python binary - specified via the PYTHON environment
|
|
* variable or defaulting to `python3`.
|
|
*/
|
|
fn python() -> String {
|
|
env::var("PYTHON").unwrap_or_else(|_| "python3".to_owned())
|
|
}
|
|
|
|
type NinjaEnv = Vec<(String, String)>;
|
|
|
|
fn ninja(gn_out_dir: &Path, maybe_env: Option<NinjaEnv>) -> Command {
|
|
let cmd_string = env::var("NINJA").unwrap_or_else(|_| "ninja".to_owned());
|
|
let mut cmd = Command::new(&cmd_string);
|
|
cmd.arg("-C");
|
|
cmd.arg(gn_out_dir);
|
|
if !cmd_string.ends_with("autoninja") {
|
|
if let Ok(jobs) = env::var("NUM_JOBS") {
|
|
cmd.arg("-j");
|
|
cmd.arg(jobs);
|
|
}
|
|
}
|
|
if let Some(env) = maybe_env {
|
|
for item in env {
|
|
cmd.env(item.0, item.1);
|
|
}
|
|
}
|
|
cmd
|
|
}
|
|
|
|
pub type GnArgs = Vec<String>;
|
|
|
|
pub fn maybe_gen(gn_args: GnArgs) -> PathBuf {
|
|
let dirs = get_dirs();
|
|
let gn_out_dir = dirs.out.join("gn_out");
|
|
|
|
if !gn_out_dir.exists() || !gn_out_dir.join("build.ninja").exists() {
|
|
let args = if let Ok(extra_args) = env::var("EXTRA_GN_ARGS") {
|
|
format!("{} {}", gn_args.join(" "), extra_args)
|
|
} else {
|
|
gn_args.join(" ")
|
|
};
|
|
|
|
let path = env::current_dir().unwrap();
|
|
println!("The current directory is {}", path.display());
|
|
println!(
|
|
"gn gen --root={} {}",
|
|
dirs.root.display(),
|
|
gn_out_dir.display()
|
|
);
|
|
assert!(Command::new(gn())
|
|
.arg(format!("--root={}", dirs.root.display()))
|
|
.arg(format!("--script-executable={}", python()))
|
|
.arg("gen")
|
|
.arg(&gn_out_dir)
|
|
.arg("--ide=json")
|
|
.arg("--args=".to_owned() + &args)
|
|
.stdout(Stdio::inherit())
|
|
.stderr(Stdio::inherit())
|
|
.envs(env::vars())
|
|
.status()
|
|
.expect("Could not run `gn`")
|
|
.success());
|
|
}
|
|
gn_out_dir
|
|
}
|
|
|
|
pub fn build(target: &str, maybe_env: Option<NinjaEnv>) {
|
|
let gn_out_dir = get_dirs().out.join("gn_out");
|
|
|
|
rerun_if_changed(&gn_out_dir, maybe_env.clone(), target);
|
|
|
|
// This helps Rust source files locate the snapshot, source map etc.
|
|
println!("cargo:rustc-env=GN_OUT_DIR={}", gn_out_dir.display());
|
|
|
|
assert!(ninja(&gn_out_dir, maybe_env)
|
|
.arg(target)
|
|
.status()
|
|
.unwrap()
|
|
.success());
|
|
|
|
// TODO This is not sufficent. We need to use "gn desc" to query the target
|
|
// and figure out what else we need to add to the link.
|
|
println!(
|
|
"cargo:rustc-link-search=native={}/obj/",
|
|
gn_out_dir.display()
|
|
);
|
|
}
|
|
|
|
/// build.rs does not get re-run unless we tell cargo about what files we
|
|
/// depend on. This outputs a bunch of rerun-if-changed lines to stdout.
|
|
fn rerun_if_changed(out_dir: &Path, maybe_env: Option<NinjaEnv>, target: &str) {
|
|
let deps = ninja_get_deps(out_dir, maybe_env, target);
|
|
for d in deps {
|
|
if let Ok(p) = out_dir.join(d).canonicalize() {
|
|
println!("cargo:rerun-if-changed={}", p.display());
|
|
}
|
|
}
|
|
}
|
|
|
|
fn ninja_get_deps(
|
|
out_dir: &Path,
|
|
maybe_env: Option<NinjaEnv>,
|
|
target: &str,
|
|
) -> HashSet<String> {
|
|
let mut cmd = ninja(out_dir, maybe_env.clone());
|
|
cmd.arg("-t");
|
|
cmd.arg("graph");
|
|
cmd.arg(target);
|
|
let output = cmd.output().expect("ninja -t graph failed");
|
|
let stdout = String::from_utf8(output.stdout).unwrap();
|
|
let graph_files = parse_ninja_graph(&stdout);
|
|
|
|
let mut cmd = ninja(out_dir, maybe_env);
|
|
cmd.arg(target);
|
|
cmd.arg("-t");
|
|
cmd.arg("deps");
|
|
let output = cmd.output().expect("ninja -t deps failed");
|
|
let stdout = String::from_utf8(output.stdout).unwrap();
|
|
let deps_files = parse_ninja_deps(&stdout);
|
|
|
|
graph_files.union(&deps_files).map(String::from).collect()
|
|
}
|
|
|
|
pub fn parse_ninja_deps(s: &str) -> HashSet<String> {
|
|
let mut out = HashSet::new();
|
|
for line in s.lines() {
|
|
if line.starts_with(" ") {
|
|
let filename = line.trim().to_string();
|
|
out.insert(filename);
|
|
}
|
|
}
|
|
out
|
|
}
|
|
|
|
/// A parser for the output of "ninja -t graph". It returns all of the input
|
|
/// files.
|
|
pub fn parse_ninja_graph(s: &str) -> HashSet<String> {
|
|
let mut out = HashSet::new();
|
|
// This is extremely hacky and likely to break.
|
|
for line in s.lines() {
|
|
if line.starts_with('\"')
|
|
&& line.contains("label=")
|
|
&& !line.contains("shape=")
|
|
&& !line.contains(" -> ")
|
|
{
|
|
let filename = line.split('\"').nth(3).unwrap();
|
|
if !filename.starts_with("..") {
|
|
continue;
|
|
}
|
|
out.insert(filename.to_string());
|
|
}
|
|
}
|
|
out
|
|
}
|
|
|
|
fn env_bool(key: &str) -> bool {
|
|
matches!(
|
|
env::var(key).unwrap_or_default().as_str(),
|
|
"true" | "1" | "yes"
|
|
)
|
|
}
|
|
|
|
#[cfg(test)]
|
|
mod test {
|
|
use super::*;
|
|
|
|
const MOCK_GRAPH: &str = r#"
|
|
digraph ninja {
|
|
rankdir="LR"
|
|
node [fontsize=10, shape=box, height=0.25]
|
|
edge [fontsize=10]
|
|
"0x7fc3c040c210" [label="default"]
|
|
"0x7fc3c040a7f0" -> "0x7fc3c040c210" [label=" phony"]
|
|
"0x7fc3c040a7f0" [label="obj/default.stamp"]
|
|
"0x7fc3c040a790" [label="stamp", shape=ellipse]
|
|
"0x7fc3c040a790" -> "0x7fc3c040a7f0"
|
|
"0x7fc3c040a6c0" -> "0x7fc3c040a790" [arrowhead=none]
|
|
"0x7fc3c040a8a0" -> "0x7fc3c040a790" [arrowhead=none]
|
|
"0x7fc3c040a920" -> "0x7fc3c040a790" [arrowhead=none]
|
|
"0x7fc3c040a6c0" [label="obj/count_bytes.stamp"]
|
|
"0x7fc3c040a4d0" -> "0x7fc3c040a6c0" [label=" stamp"]
|
|
"0x7fc3c040a4d0" [label="gen/output.txt"]
|
|
"0x7fc3c040a400" [label="___count_bytes___build_toolchain_mac_clang_x64__rule", shape=ellipse]
|
|
"0x7fc3c040a400" -> "0x7fc3c040a4d0"
|
|
"0x7fc3c040a580" -> "0x7fc3c040a400" [arrowhead=none]
|
|
"0x7fc3c040a620" -> "0x7fc3c040a400" [arrowhead=none]
|
|
"0x7fc3c040a580" [label="../../../example/src/count_bytes.py"]
|
|
"0x7fc3c040a620" [label="../../../example/src/input.txt"]
|
|
"0x7fc3c040a8a0" [label="foo"]
|
|
"0x7fc3c040b5e0" [label="link", shape=ellipse]
|
|
"0x7fc3c040b5e0" -> "0x7fc3c040a8a0"
|
|
"0x7fc3c040b5e0" -> "0x7fc3c040b6d0"
|
|
"0x7fc3c040b5e0" -> "0x7fc3c040b780"
|
|
"0x7fc3c040b5e0" -> "0x7fc3c040b820"
|
|
"0x7fc3c040b020" -> "0x7fc3c040b5e0" [arrowhead=none]
|
|
"0x7fc3c040a920" -> "0x7fc3c040b5e0" [arrowhead=none]
|
|
"0x7fc3c040b020" [label="obj/foo/foo.o"]
|
|
"0x7fc3c040b0d0" -> "0x7fc3c040b020" [label=" cxx"]
|
|
"0x7fc3c040b0d0" [label="../../../example/src/foo.cc"]
|
|
"0x7fc3c040a920" [label="obj/libhello.a"]
|
|
"0x7fc3c040be00" -> "0x7fc3c040a920" [label=" alink"]
|
|
"0x7fc3c040be00" [label="obj/hello/hello.o"]
|
|
"0x7fc3c040beb0" -> "0x7fc3c040be00" [label=" cxx"]
|
|
"0x7fc3c040beb0" [label="../../../example/src/hello.cc"]
|
|
}
|
|
"#;
|
|
|
|
#[test]
|
|
fn test_parse_ninja_graph() {
|
|
let files = parse_ninja_graph(MOCK_GRAPH);
|
|
assert!(files.contains("../../../example/src/input.txt"));
|
|
assert!(files.contains("../../../example/src/count_bytes.py"));
|
|
assert!(!files.contains("obj/hello/hello.o"));
|
|
}
|
|
}
|