| //! Logic to invoke `cargo` for building build-dependencies (build scripts and proc-macros) as well as |
| //! executing the build scripts to fetch required dependency information (`OUT_DIR` env var, extra |
| //! cfg flags, etc). |
| //! |
| //! In essence this just invokes `cargo` with the appropriate output format which we consume, |
| //! but if enabled we will also use `RUSTC_WRAPPER` to only compile the build scripts and |
| //! proc-macros and skip everything else. |
| |
| use std::{cell::RefCell, io, mem, process::Command}; |
| |
| use base_db::Env; |
| use cargo_metadata::{Message, camino::Utf8Path}; |
| use cfg::CfgAtom; |
| use itertools::Itertools; |
| use la_arena::ArenaMap; |
| use paths::{AbsPath, AbsPathBuf, Utf8PathBuf}; |
| use rustc_hash::{FxHashMap, FxHashSet}; |
| use serde::Deserialize as _; |
| use toolchain::Tool; |
| |
| use crate::{ |
| CargoConfig, CargoFeatures, CargoWorkspace, InvocationStrategy, ManifestPath, Package, Sysroot, |
| TargetKind, utf8_stdout, |
| }; |
| |
| /// Output of the build script and proc-macro building steps for a workspace. |
| #[derive(Debug, Default, Clone, PartialEq, Eq)] |
| pub struct WorkspaceBuildScripts { |
| outputs: ArenaMap<Package, BuildScriptOutput>, |
| error: Option<String>, |
| } |
| |
| /// Output of the build script and proc-macro building step for a concrete package. |
| #[derive(Debug, Clone, Default, PartialEq, Eq)] |
| pub(crate) struct BuildScriptOutput { |
| /// List of config flags defined by this package's build script. |
| pub(crate) cfgs: Vec<CfgAtom>, |
| /// List of cargo-related environment variables with their value. |
| /// |
| /// If the package has a build script which defines environment variables, |
| /// they can also be found here. |
| pub(crate) envs: Env, |
| /// Directory where a build script might place its output. |
| pub(crate) out_dir: Option<AbsPathBuf>, |
| /// Path to the proc-macro library file if this package exposes proc-macros. |
| pub(crate) proc_macro_dylib_path: Option<AbsPathBuf>, |
| } |
| |
| impl BuildScriptOutput { |
| fn is_empty(&self) -> bool { |
| self.cfgs.is_empty() |
| && self.envs.is_empty() |
| && self.out_dir.is_none() |
| && self.proc_macro_dylib_path.is_none() |
| } |
| } |
| |
| impl WorkspaceBuildScripts { |
| /// Runs the build scripts for the given workspace |
| pub(crate) fn run_for_workspace( |
| config: &CargoConfig, |
| workspace: &CargoWorkspace, |
| progress: &dyn Fn(String), |
| sysroot: &Sysroot, |
| ) -> io::Result<WorkspaceBuildScripts> { |
| let current_dir = workspace.workspace_root(); |
| |
| let allowed_features = workspace.workspace_features(); |
| let cmd = Self::build_command( |
| config, |
| &allowed_features, |
| workspace.manifest_path(), |
| current_dir, |
| sysroot, |
| )?; |
| Self::run_per_ws(cmd, workspace, progress) |
| } |
| |
| /// Runs the build scripts by invoking the configured command *once*. |
| /// This populates the outputs for all passed in workspaces. |
| pub(crate) fn run_once( |
| config: &CargoConfig, |
| workspaces: &[&CargoWorkspace], |
| progress: &dyn Fn(String), |
| working_directory: &AbsPathBuf, |
| ) -> io::Result<Vec<WorkspaceBuildScripts>> { |
| assert_eq!(config.invocation_strategy, InvocationStrategy::Once); |
| |
| let cmd = Self::build_command( |
| config, |
| &Default::default(), |
| // This is not gonna be used anyways, so just construct a dummy here |
| &ManifestPath::try_from(working_directory.clone()).unwrap(), |
| working_directory, |
| &Sysroot::empty(), |
| )?; |
| // NB: Cargo.toml could have been modified between `cargo metadata` and |
| // `cargo check`. We shouldn't assume that package ids we see here are |
| // exactly those from `config`. |
| let mut by_id = FxHashMap::default(); |
| // some workspaces might depend on the same crates, so we need to duplicate the outputs |
| // to those collisions |
| let mut collisions = Vec::new(); |
| let mut res: Vec<_> = workspaces |
| .iter() |
| .enumerate() |
| .map(|(idx, workspace)| { |
| let mut res = WorkspaceBuildScripts::default(); |
| for package in workspace.packages() { |
| res.outputs.insert(package, BuildScriptOutput::default()); |
| if by_id.contains_key(&workspace[package].id) { |
| collisions.push((&workspace[package].id, idx, package)); |
| } else { |
| by_id.insert(workspace[package].id.clone(), (package, idx)); |
| } |
| } |
| res |
| }) |
| .collect(); |
| |
| let errors = Self::run_command( |
| cmd, |
| |package, cb| { |
| if let Some(&(package, workspace)) = by_id.get(package) { |
| cb(&workspaces[workspace][package].name, &mut res[workspace].outputs[package]); |
| } |
| }, |
| progress, |
| )?; |
| res.iter_mut().for_each(|it| it.error.clone_from(&errors)); |
| collisions.into_iter().for_each(|(id, workspace, package)| { |
| if let Some(&(p, w)) = by_id.get(id) { |
| res[workspace].outputs[package] = res[w].outputs[p].clone(); |
| } |
| }); |
| |
| if tracing::enabled!(tracing::Level::INFO) { |
| for (idx, workspace) in workspaces.iter().enumerate() { |
| for package in workspace.packages() { |
| let package_build_data = &mut res[idx].outputs[package]; |
| if !package_build_data.is_empty() { |
| tracing::info!( |
| "{}: {package_build_data:?}", |
| workspace[package].manifest.parent(), |
| ); |
| } |
| } |
| } |
| } |
| |
| Ok(res) |
| } |
| |
| pub fn error(&self) -> Option<&str> { |
| self.error.as_deref() |
| } |
| |
| pub(crate) fn get_output(&self, idx: Package) -> Option<&BuildScriptOutput> { |
| self.outputs.get(idx) |
| } |
| |
| /// Assembles build script outputs for the rustc crates via `--print target-libdir`. |
| pub(crate) fn rustc_crates( |
| rustc: &CargoWorkspace, |
| current_dir: &AbsPath, |
| extra_env: &FxHashMap<String, String>, |
| sysroot: &Sysroot, |
| ) -> Self { |
| let mut bs = WorkspaceBuildScripts::default(); |
| for p in rustc.packages() { |
| bs.outputs.insert(p, BuildScriptOutput::default()); |
| } |
| let res = (|| { |
| let target_libdir = (|| { |
| let mut cargo_config = sysroot.tool(Tool::Cargo, current_dir); |
| cargo_config.envs(extra_env); |
| cargo_config |
| .args(["rustc", "-Z", "unstable-options", "--print", "target-libdir"]) |
| .env("RUSTC_BOOTSTRAP", "1"); |
| if let Ok(it) = utf8_stdout(&mut cargo_config) { |
| return Ok(it); |
| } |
| let mut cmd = sysroot.tool(Tool::Rustc, current_dir); |
| cmd.envs(extra_env); |
| cmd.args(["--print", "target-libdir"]); |
| utf8_stdout(&mut cmd) |
| })()?; |
| |
| let target_libdir = AbsPathBuf::try_from(Utf8PathBuf::from(target_libdir)) |
| .map_err(|_| anyhow::format_err!("target-libdir was not an absolute path"))?; |
| tracing::info!("Loading rustc proc-macro paths from {target_libdir}"); |
| |
| let proc_macro_dylibs: Vec<(String, AbsPathBuf)> = std::fs::read_dir(target_libdir)? |
| .filter_map(|entry| { |
| let dir_entry = entry.ok()?; |
| if dir_entry.file_type().ok()?.is_file() { |
| let path = dir_entry.path(); |
| let extension = path.extension()?; |
| if extension == std::env::consts::DLL_EXTENSION { |
| let name = path.file_stem()?.to_str()?.split_once('-')?.0.to_owned(); |
| let path = AbsPathBuf::try_from(Utf8PathBuf::from_path_buf(path).ok()?) |
| .ok()?; |
| return Some((name, path)); |
| } |
| } |
| None |
| }) |
| .collect(); |
| for p in rustc.packages() { |
| let package = &rustc[p]; |
| if package |
| .targets |
| .iter() |
| .any(|&it| matches!(rustc[it].kind, TargetKind::Lib { is_proc_macro: true })) |
| { |
| if let Some((_, path)) = proc_macro_dylibs |
| .iter() |
| .find(|(name, _)| *name.trim_start_matches("lib") == package.name) |
| { |
| bs.outputs[p].proc_macro_dylib_path = Some(path.clone()); |
| } |
| } |
| } |
| |
| if tracing::enabled!(tracing::Level::INFO) { |
| for package in rustc.packages() { |
| let package_build_data = &bs.outputs[package]; |
| if !package_build_data.is_empty() { |
| tracing::info!( |
| "{}: {package_build_data:?}", |
| rustc[package].manifest.parent(), |
| ); |
| } |
| } |
| } |
| Ok(()) |
| })(); |
| if let Err::<_, anyhow::Error>(e) = res { |
| bs.error = Some(e.to_string()); |
| } |
| bs |
| } |
| |
| fn run_per_ws( |
| cmd: Command, |
| workspace: &CargoWorkspace, |
| progress: &dyn Fn(String), |
| ) -> io::Result<WorkspaceBuildScripts> { |
| let mut res = WorkspaceBuildScripts::default(); |
| let outputs = &mut res.outputs; |
| // NB: Cargo.toml could have been modified between `cargo metadata` and |
| // `cargo check`. We shouldn't assume that package ids we see here are |
| // exactly those from `config`. |
| let mut by_id: FxHashMap<String, Package> = FxHashMap::default(); |
| for package in workspace.packages() { |
| outputs.insert(package, BuildScriptOutput::default()); |
| by_id.insert(workspace[package].id.clone(), package); |
| } |
| |
| res.error = Self::run_command( |
| cmd, |
| |package, cb| { |
| if let Some(&package) = by_id.get(package) { |
| cb(&workspace[package].name, &mut outputs[package]); |
| } |
| }, |
| progress, |
| )?; |
| |
| if tracing::enabled!(tracing::Level::INFO) { |
| for package in workspace.packages() { |
| let package_build_data = &outputs[package]; |
| if !package_build_data.is_empty() { |
| tracing::info!( |
| "{}: {package_build_data:?}", |
| workspace[package].manifest.parent(), |
| ); |
| } |
| } |
| } |
| |
| Ok(res) |
| } |
| |
| fn run_command( |
| cmd: Command, |
| // ideally this would be something like: |
| // with_output_for: impl FnMut(&str, dyn FnOnce(&mut BuildScriptOutput)), |
| // but owned trait objects aren't a thing |
| mut with_output_for: impl FnMut(&str, &mut dyn FnMut(&str, &mut BuildScriptOutput)), |
| progress: &dyn Fn(String), |
| ) -> io::Result<Option<String>> { |
| let errors = RefCell::new(String::new()); |
| let push_err = |err: &str| { |
| let mut e = errors.borrow_mut(); |
| e.push_str(err); |
| e.push('\n'); |
| }; |
| |
| tracing::info!("Running build scripts: {:?}", cmd); |
| let output = stdx::process::spawn_with_streaming_output( |
| cmd, |
| &mut |line| { |
| // Copy-pasted from existing cargo_metadata. It seems like we |
| // should be using serde_stacker here? |
| let mut deserializer = serde_json::Deserializer::from_str(line); |
| deserializer.disable_recursion_limit(); |
| let message = Message::deserialize(&mut deserializer) |
| .unwrap_or_else(|_| Message::TextLine(line.to_owned())); |
| |
| match message { |
| Message::BuildScriptExecuted(mut message) => { |
| with_output_for(&message.package_id.repr, &mut |name, data| { |
| progress(format!("running build-script: {name}")); |
| let cfgs = { |
| let mut acc = Vec::new(); |
| for cfg in &message.cfgs { |
| match crate::parse_cfg(cfg) { |
| Ok(it) => acc.push(it), |
| Err(err) => { |
| push_err(&format!( |
| "invalid cfg from cargo-metadata: {err}" |
| )); |
| return; |
| } |
| }; |
| } |
| acc |
| }; |
| data.envs.extend(message.env.drain(..)); |
| // cargo_metadata crate returns default (empty) path for |
| // older cargos, which is not absolute, so work around that. |
| let out_dir = mem::take(&mut message.out_dir); |
| if !out_dir.as_str().is_empty() { |
| let out_dir = AbsPathBuf::assert(out_dir); |
| // inject_cargo_env(package, package_build_data); |
| data.envs.insert("OUT_DIR", out_dir.as_str()); |
| data.out_dir = Some(out_dir); |
| data.cfgs = cfgs; |
| } |
| }); |
| } |
| Message::CompilerArtifact(message) => { |
| with_output_for(&message.package_id.repr, &mut |name, data| { |
| progress(format!("building proc-macros: {name}")); |
| if message.target.kind.contains(&cargo_metadata::TargetKind::ProcMacro) |
| { |
| // Skip rmeta file |
| if let Some(filename) = |
| message.filenames.iter().find(|file| is_dylib(file)) |
| { |
| let filename = AbsPath::assert(filename); |
| data.proc_macro_dylib_path = Some(filename.to_owned()); |
| } |
| } |
| }); |
| } |
| Message::CompilerMessage(message) => { |
| progress(message.target.name); |
| |
| if let Some(diag) = message.message.rendered.as_deref() { |
| push_err(diag); |
| } |
| } |
| Message::BuildFinished(_) => {} |
| Message::TextLine(_) => {} |
| _ => {} |
| } |
| }, |
| &mut |line| { |
| push_err(line); |
| }, |
| )?; |
| |
| let errors = if !output.status.success() { |
| let errors = errors.into_inner(); |
| Some(if errors.is_empty() { "cargo check failed".to_owned() } else { errors }) |
| } else { |
| None |
| }; |
| Ok(errors) |
| } |
| |
| fn build_command( |
| config: &CargoConfig, |
| allowed_features: &FxHashSet<String>, |
| manifest_path: &ManifestPath, |
| current_dir: &AbsPath, |
| sysroot: &Sysroot, |
| ) -> io::Result<Command> { |
| let mut cmd = match config.run_build_script_command.as_deref() { |
| Some([program, args @ ..]) => { |
| let mut cmd = toolchain::command(program, current_dir); |
| cmd.args(args); |
| cmd |
| } |
| _ => { |
| let mut cmd = sysroot.tool(Tool::Cargo, current_dir); |
| |
| cmd.args(["check", "--quiet", "--workspace", "--message-format=json"]); |
| cmd.args(&config.extra_args); |
| |
| cmd.arg("--manifest-path"); |
| cmd.arg(manifest_path); |
| |
| if let Some(target_dir) = &config.target_dir { |
| cmd.arg("--target-dir").arg(target_dir); |
| } |
| |
| // --all-targets includes tests, benches and examples in addition to the |
| // default lib and bins. This is an independent concept from the --target |
| // flag below. |
| if config.all_targets { |
| cmd.arg("--all-targets"); |
| } |
| |
| if let Some(target) = &config.target { |
| cmd.args(["--target", target]); |
| } |
| |
| match &config.features { |
| CargoFeatures::All => { |
| cmd.arg("--all-features"); |
| } |
| CargoFeatures::Selected { features, no_default_features } => { |
| if *no_default_features { |
| cmd.arg("--no-default-features"); |
| } |
| if !features.is_empty() { |
| cmd.arg("--features"); |
| cmd.arg( |
| features |
| .iter() |
| .filter(|&feat| allowed_features.contains(feat)) |
| .join(","), |
| ); |
| } |
| } |
| } |
| |
| if manifest_path.is_rust_manifest() { |
| cmd.arg("-Zscript"); |
| } |
| |
| cmd.arg("--keep-going"); |
| |
| cmd |
| } |
| }; |
| |
| cmd.envs(&config.extra_env); |
| if config.wrap_rustc_in_build_scripts { |
| // Setup RUSTC_WRAPPER to point to `rust-analyzer` binary itself. We use |
| // that to compile only proc macros and build scripts during the initial |
| // `cargo check`. |
| let myself = std::env::current_exe()?; |
| cmd.env("RUSTC_WRAPPER", myself); |
| cmd.env("RA_RUSTC_WRAPPER", "1"); |
| } |
| |
| Ok(cmd) |
| } |
| } |
| |
| // FIXME: Find a better way to know if it is a dylib. |
| fn is_dylib(path: &Utf8Path) -> bool { |
| match path.extension().map(|e| e.to_owned().to_lowercase()) { |
| None => false, |
| Some(ext) => matches!(ext.as_str(), "dll" | "dylib" | "so"), |
| } |
| } |