| // Copyright 2012-2015 The Rust Project Developers. See the COPYRIGHT |
| // file at the top-level directory of this distribution and at |
| // http://rust-lang.org/COPYRIGHT. |
| // |
| // Licensed under the Apache License, Version 2.0 <LICENSE-APACHE or |
| // http://www.apache.org/licenses/LICENSE-2.0> or the MIT license |
| // <LICENSE-MIT or http://opensource.org/licenses/MIT>, at your |
| // option. This file may not be copied, modified, or distributed |
| // except according to those terms. |
| |
| use rustc::dep_graph::DepGraph; |
| use rustc::hir::{self, map as hir_map}; |
| use rustc::hir::lowering::lower_crate; |
| use rustc::ich::Fingerprint; |
| use rustc_data_structures::stable_hasher::StableHasher; |
| use rustc_mir as mir; |
| use rustc::session::{CompileResult, CrateDisambiguator, Session}; |
| use rustc::session::CompileIncomplete; |
| use rustc::session::config::{self, Input, OutputFilenames, OutputType}; |
| use rustc::session::search_paths::PathKind; |
| use rustc::lint; |
| use rustc::middle::{self, reachable, resolve_lifetime, stability}; |
| use rustc::middle::cstore::CrateStoreDyn; |
| use rustc::middle::privacy::AccessLevels; |
| use rustc::ty::{self, AllArenas, Resolutions, TyCtxt}; |
| use rustc::traits; |
| use rustc::util::common::{install_panic_hook, time, ErrorReported}; |
| use rustc_allocator as allocator; |
| use rustc_borrowck as borrowck; |
| use rustc_incremental; |
| use rustc_resolve::{MakeGlobMap, Resolver, ResolverArenas}; |
| use rustc_metadata::creader::CrateLoader; |
| use rustc_metadata::cstore::{self, CStore}; |
| use rustc_traits; |
| use rustc_codegen_utils::codegen_backend::CodegenBackend; |
| use rustc_typeck as typeck; |
| use rustc_privacy; |
| use rustc_plugin::registry::Registry; |
| use rustc_plugin as plugin; |
| use rustc_passes::{self, ast_validation, hir_stats, loops, rvalue_promotion}; |
| use super::Compilation; |
| |
| use serialize::json; |
| |
| use std::any::Any; |
| use std::env; |
| use std::ffi::{OsStr, OsString}; |
| use std::fs; |
| use std::io::{self, Write}; |
| use std::iter; |
| use std::path::{Path, PathBuf}; |
| use rustc_data_structures::sync::{self, Lrc, Lock}; |
| use std::sync::mpsc; |
| use syntax::{self, ast, attr, diagnostics, visit}; |
| use syntax::early_buffered_lints::BufferedEarlyLint; |
| use syntax::ext::base::ExtCtxt; |
| use syntax::fold::Folder; |
| use syntax::parse::{self, PResult}; |
| use syntax::util::node_count::NodeCounter; |
| use syntax_pos::FileName; |
| use syntax_ext; |
| |
| use derive_registrar; |
| use pretty::ReplaceBodyWithLoop; |
| |
| use profile; |
| |
| #[cfg(not(parallel_queries))] |
| pub fn spawn_thread_pool<F: FnOnce(config::Options) -> R + sync::Send, R: sync::Send>( |
| opts: config::Options, |
| f: F |
| ) -> R { |
| ty::tls::GCX_PTR.set(&Lock::new(0), || { |
| f(opts) |
| }) |
| } |
| |
| #[cfg(parallel_queries)] |
| pub fn spawn_thread_pool<F: FnOnce(config::Options) -> R + sync::Send, R: sync::Send>( |
| opts: config::Options, |
| f: F |
| ) -> R { |
| use syntax; |
| use syntax_pos; |
| use rayon::{ThreadPoolBuilder, ThreadPool}; |
| |
| let gcx_ptr = &Lock::new(0); |
| |
| let config = ThreadPoolBuilder::new() |
| .num_threads(Session::query_threads_from_opts(&opts)) |
| .deadlock_handler(|| unsafe { ty::query::handle_deadlock() }) |
| .stack_size(16 * 1024 * 1024); |
| |
| let with_pool = move |pool: &ThreadPool| { |
| pool.install(move || f(opts)) |
| }; |
| |
| syntax::GLOBALS.with(|syntax_globals| { |
| syntax_pos::GLOBALS.with(|syntax_pos_globals| { |
| // The main handler run for each Rayon worker thread and sets up |
| // the thread local rustc uses. syntax_globals and syntax_pos_globals are |
| // captured and set on the new threads. ty::tls::with_thread_locals sets up |
| // thread local callbacks from libsyntax |
| let main_handler = move |worker: &mut dyn FnMut()| { |
| syntax::GLOBALS.set(syntax_globals, || { |
| syntax_pos::GLOBALS.set(syntax_pos_globals, || { |
| ty::tls::with_thread_locals(|| { |
| ty::tls::GCX_PTR.set(gcx_ptr, || { |
| worker() |
| }) |
| }) |
| }) |
| }) |
| }; |
| |
| ThreadPool::scoped_pool(config, main_handler, with_pool).unwrap() |
| }) |
| }) |
| } |
| |
| pub fn compile_input( |
| codegen_backend: Box<dyn CodegenBackend>, |
| sess: &Session, |
| cstore: &CStore, |
| input_path: &Option<PathBuf>, |
| input: &Input, |
| outdir: &Option<PathBuf>, |
| output: &Option<PathBuf>, |
| addl_plugins: Option<Vec<String>>, |
| control: &CompileController, |
| ) -> CompileResult { |
| macro_rules! controller_entry_point { |
| ($point: ident, $tsess: expr, $make_state: expr, $phase_result: expr) => {{ |
| let state = &mut $make_state; |
| let phase_result: &CompileResult = &$phase_result; |
| if phase_result.is_ok() || control.$point.run_callback_on_error { |
| (control.$point.callback)(state); |
| } |
| |
| if control.$point.stop == Compilation::Stop { |
| // FIXME: shouldn't this return Err(CompileIncomplete::Stopped) |
| // if there are no errors? |
| return $tsess.compile_status(); |
| } |
| }} |
| } |
| |
| if sess.profile_queries() { |
| profile::begin(sess); |
| } |
| |
| // We need nested scopes here, because the intermediate results can keep |
| // large chunks of memory alive and we want to free them as soon as |
| // possible to keep the peak memory usage low |
| let (outputs, ongoing_codegen, dep_graph) = { |
| let krate = match phase_1_parse_input(control, sess, input) { |
| Ok(krate) => krate, |
| Err(mut parse_error) => { |
| parse_error.emit(); |
| return Err(CompileIncomplete::Errored(ErrorReported)); |
| } |
| }; |
| |
| let (krate, registry) = { |
| let mut compile_state = |
| CompileState::state_after_parse(input, sess, outdir, output, krate, &cstore); |
| controller_entry_point!(after_parse, sess, compile_state, Ok(())); |
| |
| (compile_state.krate.unwrap(), compile_state.registry) |
| }; |
| |
| let outputs = build_output_filenames(input, outdir, output, &krate.attrs, sess); |
| let crate_name = |
| ::rustc_codegen_utils::link::find_crate_name(Some(sess), &krate.attrs, input); |
| install_panic_hook(); |
| |
| let ExpansionResult { |
| expanded_crate, |
| defs, |
| analysis, |
| resolutions, |
| mut hir_forest, |
| } = { |
| phase_2_configure_and_expand( |
| sess, |
| &cstore, |
| krate, |
| registry, |
| &crate_name, |
| addl_plugins, |
| control.make_glob_map, |
| |expanded_crate| { |
| let mut state = CompileState::state_after_expand( |
| input, |
| sess, |
| outdir, |
| output, |
| &cstore, |
| expanded_crate, |
| &crate_name, |
| ); |
| controller_entry_point!(after_expand, sess, state, Ok(())); |
| Ok(()) |
| }, |
| )? |
| }; |
| |
| let output_paths = generated_output_paths(sess, &outputs, output.is_some(), &crate_name); |
| |
| // Ensure the source file isn't accidentally overwritten during compilation. |
| if let Some(ref input_path) = *input_path { |
| if sess.opts.will_create_output_file() { |
| if output_contains_path(&output_paths, input_path) { |
| sess.err(&format!( |
| "the input file \"{}\" would be overwritten by the generated \ |
| executable", |
| input_path.display() |
| )); |
| return Err(CompileIncomplete::Stopped); |
| } |
| if let Some(dir_path) = output_conflicts_with_dir(&output_paths) { |
| sess.err(&format!( |
| "the generated executable for the input file \"{}\" conflicts with the \ |
| existing directory \"{}\"", |
| input_path.display(), |
| dir_path.display() |
| )); |
| return Err(CompileIncomplete::Stopped); |
| } |
| } |
| } |
| |
| write_out_deps(sess, &outputs, &output_paths); |
| if sess.opts.output_types.contains_key(&OutputType::DepInfo) |
| && sess.opts.output_types.len() == 1 |
| { |
| return Ok(()); |
| } |
| |
| if let &Some(ref dir) = outdir { |
| if fs::create_dir_all(dir).is_err() { |
| sess.err("failed to find or create the directory specified by --out-dir"); |
| return Err(CompileIncomplete::Stopped); |
| } |
| } |
| |
| let arenas = AllArenas::new(); |
| |
| // Construct the HIR map |
| let hir_map = time(sess, "indexing hir", || { |
| hir_map::map_crate(sess, cstore, &mut hir_forest, &defs) |
| }); |
| |
| { |
| hir_map.dep_graph.assert_ignored(); |
| controller_entry_point!( |
| after_hir_lowering, |
| sess, |
| CompileState::state_after_hir_lowering( |
| input, |
| sess, |
| outdir, |
| output, |
| &arenas, |
| &cstore, |
| &hir_map, |
| &analysis, |
| &resolutions, |
| &expanded_crate, |
| &hir_map.krate(), |
| &outputs, |
| &crate_name |
| ), |
| Ok(()) |
| ); |
| } |
| |
| let opt_crate = if control.keep_ast { |
| Some(&expanded_crate) |
| } else { |
| drop(expanded_crate); |
| None |
| }; |
| |
| phase_3_run_analysis_passes( |
| &*codegen_backend, |
| control, |
| sess, |
| cstore, |
| hir_map, |
| analysis, |
| resolutions, |
| &arenas, |
| &crate_name, |
| &outputs, |
| |tcx, analysis, rx, result| { |
| { |
| // Eventually, we will want to track plugins. |
| tcx.dep_graph.with_ignore(|| { |
| let mut state = CompileState::state_after_analysis( |
| input, |
| sess, |
| outdir, |
| output, |
| opt_crate, |
| tcx.hir.krate(), |
| &analysis, |
| tcx, |
| &crate_name, |
| ); |
| (control.after_analysis.callback)(&mut state); |
| }); |
| |
| if control.after_analysis.stop == Compilation::Stop { |
| return result.and_then(|_| Err(CompileIncomplete::Stopped)); |
| } |
| } |
| |
| result?; |
| |
| if log_enabled!(::log::Level::Info) { |
| println!("Pre-codegen"); |
| tcx.print_debug_stats(); |
| } |
| |
| let ongoing_codegen = phase_4_codegen(&*codegen_backend, tcx, rx); |
| |
| if log_enabled!(::log::Level::Info) { |
| println!("Post-codegen"); |
| tcx.print_debug_stats(); |
| } |
| |
| if tcx.sess.opts.output_types.contains_key(&OutputType::Mir) { |
| if let Err(e) = mir::transform::dump_mir::emit_mir(tcx, &outputs) { |
| sess.err(&format!("could not emit MIR: {}", e)); |
| sess.abort_if_errors(); |
| } |
| } |
| |
| Ok((outputs.clone(), ongoing_codegen, tcx.dep_graph.clone())) |
| }, |
| )?? |
| }; |
| |
| if sess.opts.debugging_opts.print_type_sizes { |
| sess.code_stats.borrow().print_type_sizes(); |
| } |
| |
| codegen_backend.join_codegen_and_link(ongoing_codegen, sess, &dep_graph, &outputs)?; |
| |
| if sess.opts.debugging_opts.perf_stats { |
| sess.print_perf_stats(); |
| } |
| |
| controller_entry_point!( |
| compilation_done, |
| sess, |
| CompileState::state_when_compilation_done(input, sess, outdir, output), |
| Ok(()) |
| ); |
| |
| Ok(()) |
| } |
| |
| pub fn source_name(input: &Input) -> FileName { |
| match *input { |
| Input::File(ref ifile) => ifile.clone().into(), |
| Input::Str { ref name, .. } => name.clone(), |
| } |
| } |
| |
| /// CompileController is used to customize compilation, it allows compilation to |
| /// be stopped and/or to call arbitrary code at various points in compilation. |
| /// It also allows for various flags to be set to influence what information gets |
| /// collected during compilation. |
| /// |
| /// This is a somewhat higher level controller than a Session - the Session |
| /// controls what happens in each phase, whereas the CompileController controls |
| /// whether a phase is run at all and whether other code (from outside the |
| /// compiler) is run between phases. |
| /// |
| /// Note that if compilation is set to stop and a callback is provided for a |
| /// given entry point, the callback is called before compilation is stopped. |
| /// |
| /// Expect more entry points to be added in the future. |
| pub struct CompileController<'a> { |
| pub after_parse: PhaseController<'a>, |
| pub after_expand: PhaseController<'a>, |
| pub after_hir_lowering: PhaseController<'a>, |
| pub after_analysis: PhaseController<'a>, |
| pub compilation_done: PhaseController<'a>, |
| |
| // FIXME we probably want to group the below options together and offer a |
| // better API, rather than this ad-hoc approach. |
| pub make_glob_map: MakeGlobMap, |
| // Whether the compiler should keep the ast beyond parsing. |
| pub keep_ast: bool, |
| // -Zcontinue-parse-after-error |
| pub continue_parse_after_error: bool, |
| |
| /// Allows overriding default rustc query providers, |
| /// after `default_provide` has installed them. |
| pub provide: Box<dyn Fn(&mut ty::query::Providers) + 'a>, |
| /// Same as `provide`, but only for non-local crates, |
| /// applied after `default_provide_extern`. |
| pub provide_extern: Box<dyn Fn(&mut ty::query::Providers) + 'a>, |
| } |
| |
| impl<'a> CompileController<'a> { |
| pub fn basic() -> CompileController<'a> { |
| CompileController { |
| after_parse: PhaseController::basic(), |
| after_expand: PhaseController::basic(), |
| after_hir_lowering: PhaseController::basic(), |
| after_analysis: PhaseController::basic(), |
| compilation_done: PhaseController::basic(), |
| make_glob_map: MakeGlobMap::No, |
| keep_ast: false, |
| continue_parse_after_error: false, |
| provide: box |_| {}, |
| provide_extern: box |_| {}, |
| } |
| } |
| } |
| |
| /// This implementation makes it easier to create a custom driver when you only want to hook |
| /// into callbacks from `CompileController`. |
| /// |
| /// # Example |
| /// |
| /// ```no_run |
| /// # extern crate rustc_driver; |
| /// # use rustc_driver::driver::CompileController; |
| /// let mut controller = CompileController::basic(); |
| /// controller.after_analysis.callback = Box::new(move |_state| {}); |
| /// rustc_driver::run_compiler(&[], Box::new(controller), None, None); |
| /// ``` |
| impl<'a> ::CompilerCalls<'a> for CompileController<'a> { |
| fn early_callback( |
| &mut self, |
| matches: &::getopts::Matches, |
| sopts: &config::Options, |
| cfg: &ast::CrateConfig, |
| descriptions: &::errors::registry::Registry, |
| output: ::ErrorOutputType, |
| ) -> Compilation { |
| ::RustcDefaultCalls.early_callback( |
| matches, |
| sopts, |
| cfg, |
| descriptions, |
| output, |
| ) |
| } |
| fn no_input( |
| &mut self, |
| matches: &::getopts::Matches, |
| sopts: &config::Options, |
| cfg: &ast::CrateConfig, |
| odir: &Option<PathBuf>, |
| ofile: &Option<PathBuf>, |
| descriptions: &::errors::registry::Registry, |
| ) -> Option<(Input, Option<PathBuf>)> { |
| ::RustcDefaultCalls.no_input( |
| matches, |
| sopts, |
| cfg, |
| odir, |
| ofile, |
| descriptions, |
| ) |
| } |
| fn late_callback( |
| &mut self, |
| codegen_backend: &dyn (::CodegenBackend), |
| matches: &::getopts::Matches, |
| sess: &Session, |
| cstore: &dyn (::CrateStore), |
| input: &Input, |
| odir: &Option<PathBuf>, |
| ofile: &Option<PathBuf>, |
| ) -> Compilation { |
| ::RustcDefaultCalls |
| .late_callback(codegen_backend, matches, sess, cstore, input, odir, ofile) |
| } |
| fn build_controller( |
| self: Box<Self>, |
| _: &Session, |
| _: &::getopts::Matches |
| ) -> CompileController<'a> { |
| *self |
| } |
| } |
| |
| pub struct PhaseController<'a> { |
| pub stop: Compilation, |
| // If true then the compiler will try to run the callback even if the phase |
| // ends with an error. Note that this is not always possible. |
| pub run_callback_on_error: bool, |
| pub callback: Box<dyn Fn(&mut CompileState) + 'a>, |
| } |
| |
| impl<'a> PhaseController<'a> { |
| pub fn basic() -> PhaseController<'a> { |
| PhaseController { |
| stop: Compilation::Continue, |
| run_callback_on_error: false, |
| callback: box |_| {}, |
| } |
| } |
| } |
| |
| /// State that is passed to a callback. What state is available depends on when |
| /// during compilation the callback is made. See the various constructor methods |
| /// (`state_*`) in the impl to see which data is provided for any given entry point. |
| pub struct CompileState<'a, 'tcx: 'a> { |
| pub input: &'a Input, |
| pub session: &'tcx Session, |
| pub krate: Option<ast::Crate>, |
| pub registry: Option<Registry<'a>>, |
| pub cstore: Option<&'tcx CStore>, |
| pub crate_name: Option<&'a str>, |
| pub output_filenames: Option<&'a OutputFilenames>, |
| pub out_dir: Option<&'a Path>, |
| pub out_file: Option<&'a Path>, |
| pub arenas: Option<&'tcx AllArenas<'tcx>>, |
| pub expanded_crate: Option<&'a ast::Crate>, |
| pub hir_crate: Option<&'a hir::Crate>, |
| pub hir_map: Option<&'a hir_map::Map<'tcx>>, |
| pub resolutions: Option<&'a Resolutions>, |
| pub analysis: Option<&'a ty::CrateAnalysis>, |
| pub tcx: Option<TyCtxt<'a, 'tcx, 'tcx>>, |
| } |
| |
| impl<'a, 'tcx> CompileState<'a, 'tcx> { |
| fn empty(input: &'a Input, session: &'tcx Session, out_dir: &'a Option<PathBuf>) -> Self { |
| CompileState { |
| input, |
| session, |
| out_dir: out_dir.as_ref().map(|s| &**s), |
| out_file: None, |
| arenas: None, |
| krate: None, |
| registry: None, |
| cstore: None, |
| crate_name: None, |
| output_filenames: None, |
| expanded_crate: None, |
| hir_crate: None, |
| hir_map: None, |
| resolutions: None, |
| analysis: None, |
| tcx: None, |
| } |
| } |
| |
| fn state_after_parse( |
| input: &'a Input, |
| session: &'tcx Session, |
| out_dir: &'a Option<PathBuf>, |
| out_file: &'a Option<PathBuf>, |
| krate: ast::Crate, |
| cstore: &'tcx CStore, |
| ) -> Self { |
| CompileState { |
| // Initialize the registry before moving `krate` |
| registry: Some(Registry::new(&session, krate.span)), |
| krate: Some(krate), |
| cstore: Some(cstore), |
| out_file: out_file.as_ref().map(|s| &**s), |
| ..CompileState::empty(input, session, out_dir) |
| } |
| } |
| |
| fn state_after_expand( |
| input: &'a Input, |
| session: &'tcx Session, |
| out_dir: &'a Option<PathBuf>, |
| out_file: &'a Option<PathBuf>, |
| cstore: &'tcx CStore, |
| expanded_crate: &'a ast::Crate, |
| crate_name: &'a str, |
| ) -> Self { |
| CompileState { |
| crate_name: Some(crate_name), |
| cstore: Some(cstore), |
| expanded_crate: Some(expanded_crate), |
| out_file: out_file.as_ref().map(|s| &**s), |
| ..CompileState::empty(input, session, out_dir) |
| } |
| } |
| |
| fn state_after_hir_lowering( |
| input: &'a Input, |
| session: &'tcx Session, |
| out_dir: &'a Option<PathBuf>, |
| out_file: &'a Option<PathBuf>, |
| arenas: &'tcx AllArenas<'tcx>, |
| cstore: &'tcx CStore, |
| hir_map: &'a hir_map::Map<'tcx>, |
| analysis: &'a ty::CrateAnalysis, |
| resolutions: &'a Resolutions, |
| krate: &'a ast::Crate, |
| hir_crate: &'a hir::Crate, |
| output_filenames: &'a OutputFilenames, |
| crate_name: &'a str, |
| ) -> Self { |
| CompileState { |
| crate_name: Some(crate_name), |
| arenas: Some(arenas), |
| cstore: Some(cstore), |
| hir_map: Some(hir_map), |
| analysis: Some(analysis), |
| resolutions: Some(resolutions), |
| expanded_crate: Some(krate), |
| hir_crate: Some(hir_crate), |
| output_filenames: Some(output_filenames), |
| out_file: out_file.as_ref().map(|s| &**s), |
| ..CompileState::empty(input, session, out_dir) |
| } |
| } |
| |
| fn state_after_analysis( |
| input: &'a Input, |
| session: &'tcx Session, |
| out_dir: &'a Option<PathBuf>, |
| out_file: &'a Option<PathBuf>, |
| krate: Option<&'a ast::Crate>, |
| hir_crate: &'a hir::Crate, |
| analysis: &'a ty::CrateAnalysis, |
| tcx: TyCtxt<'a, 'tcx, 'tcx>, |
| crate_name: &'a str, |
| ) -> Self { |
| CompileState { |
| analysis: Some(analysis), |
| tcx: Some(tcx), |
| expanded_crate: krate, |
| hir_crate: Some(hir_crate), |
| crate_name: Some(crate_name), |
| out_file: out_file.as_ref().map(|s| &**s), |
| ..CompileState::empty(input, session, out_dir) |
| } |
| } |
| |
| fn state_when_compilation_done( |
| input: &'a Input, |
| session: &'tcx Session, |
| out_dir: &'a Option<PathBuf>, |
| out_file: &'a Option<PathBuf>, |
| ) -> Self { |
| CompileState { |
| out_file: out_file.as_ref().map(|s| &**s), |
| ..CompileState::empty(input, session, out_dir) |
| } |
| } |
| } |
| |
| pub fn phase_1_parse_input<'a>( |
| control: &CompileController, |
| sess: &'a Session, |
| input: &Input, |
| ) -> PResult<'a, ast::Crate> { |
| sess.diagnostic() |
| .set_continue_after_error(control.continue_parse_after_error); |
| |
| if sess.profile_queries() { |
| profile::begin(sess); |
| } |
| |
| let krate = time(sess, "parsing", || match *input { |
| Input::File(ref file) => parse::parse_crate_from_file(file, &sess.parse_sess), |
| Input::Str { |
| ref input, |
| ref name, |
| } => parse::parse_crate_from_source_str(name.clone(), input.clone(), &sess.parse_sess), |
| })?; |
| |
| sess.diagnostic().set_continue_after_error(true); |
| |
| if sess.opts.debugging_opts.ast_json_noexpand { |
| println!("{}", json::as_json(&krate)); |
| } |
| |
| if sess.opts.debugging_opts.input_stats { |
| println!( |
| "Lines of code: {}", |
| sess.codemap().count_lines() |
| ); |
| println!("Pre-expansion node count: {}", count_nodes(&krate)); |
| } |
| |
| if let Some(ref s) = sess.opts.debugging_opts.show_span { |
| syntax::show_span::run(sess.diagnostic(), s, &krate); |
| } |
| |
| if sess.opts.debugging_opts.hir_stats { |
| hir_stats::print_ast_stats(&krate, "PRE EXPANSION AST STATS"); |
| } |
| |
| Ok(krate) |
| } |
| |
| fn count_nodes(krate: &ast::Crate) -> usize { |
| let mut counter = NodeCounter::new(); |
| visit::walk_crate(&mut counter, krate); |
| counter.count |
| } |
| |
| // For continuing compilation after a parsed crate has been |
| // modified |
| |
| pub struct ExpansionResult { |
| pub expanded_crate: ast::Crate, |
| pub defs: hir_map::Definitions, |
| pub analysis: ty::CrateAnalysis, |
| pub resolutions: Resolutions, |
| pub hir_forest: hir_map::Forest, |
| } |
| |
| pub struct InnerExpansionResult<'a> { |
| pub expanded_crate: ast::Crate, |
| pub resolver: Resolver<'a>, |
| pub hir_forest: hir_map::Forest, |
| } |
| |
| /// Run the "early phases" of the compiler: initial `cfg` processing, |
| /// loading compiler plugins (including those from `addl_plugins`), |
| /// syntax expansion, secondary `cfg` expansion, synthesis of a test |
| /// harness if one is to be provided, injection of a dependency on the |
| /// standard library and prelude, and name resolution. |
| /// |
| /// Returns `None` if we're aborting after handling -W help. |
| pub fn phase_2_configure_and_expand<F>( |
| sess: &Session, |
| cstore: &CStore, |
| krate: ast::Crate, |
| registry: Option<Registry>, |
| crate_name: &str, |
| addl_plugins: Option<Vec<String>>, |
| make_glob_map: MakeGlobMap, |
| after_expand: F, |
| ) -> Result<ExpansionResult, CompileIncomplete> |
| where |
| F: FnOnce(&ast::Crate) -> CompileResult, |
| { |
| // Currently, we ignore the name resolution data structures for the purposes of dependency |
| // tracking. Instead we will run name resolution and include its output in the hash of each |
| // item, much like we do for macro expansion. In other words, the hash reflects not just |
| // its contents but the results of name resolution on those contents. Hopefully we'll push |
| // this back at some point. |
| let mut crate_loader = CrateLoader::new(sess, &cstore, &crate_name); |
| let resolver_arenas = Resolver::arenas(); |
| let result = phase_2_configure_and_expand_inner( |
| sess, |
| cstore, |
| krate, |
| registry, |
| crate_name, |
| addl_plugins, |
| make_glob_map, |
| &resolver_arenas, |
| &mut crate_loader, |
| after_expand, |
| ); |
| match result { |
| Ok(InnerExpansionResult { |
| expanded_crate, |
| resolver, |
| hir_forest, |
| }) => Ok(ExpansionResult { |
| expanded_crate, |
| defs: resolver.definitions, |
| hir_forest, |
| resolutions: Resolutions { |
| freevars: resolver.freevars, |
| export_map: resolver.export_map, |
| trait_map: resolver.trait_map, |
| maybe_unused_trait_imports: resolver.maybe_unused_trait_imports, |
| maybe_unused_extern_crates: resolver.maybe_unused_extern_crates, |
| }, |
| |
| analysis: ty::CrateAnalysis { |
| access_levels: Lrc::new(AccessLevels::default()), |
| name: crate_name.to_string(), |
| glob_map: if resolver.make_glob_map { |
| Some(resolver.glob_map) |
| } else { |
| None |
| }, |
| }, |
| }), |
| Err(x) => Err(x), |
| } |
| } |
| |
| /// Same as phase_2_configure_and_expand, but doesn't let you keep the resolver |
| /// around |
| pub fn phase_2_configure_and_expand_inner<'a, F>( |
| sess: &'a Session, |
| cstore: &'a CStore, |
| mut krate: ast::Crate, |
| registry: Option<Registry>, |
| crate_name: &str, |
| addl_plugins: Option<Vec<String>>, |
| make_glob_map: MakeGlobMap, |
| resolver_arenas: &'a ResolverArenas<'a>, |
| crate_loader: &'a mut CrateLoader, |
| after_expand: F, |
| ) -> Result<InnerExpansionResult<'a>, CompileIncomplete> |
| where |
| F: FnOnce(&ast::Crate) -> CompileResult, |
| { |
| krate = time(sess, "attributes injection", || { |
| syntax::attr::inject(krate, &sess.parse_sess, &sess.opts.debugging_opts.crate_attr) |
| }); |
| |
| let (mut krate, features) = syntax::config::features( |
| krate, |
| &sess.parse_sess, |
| sess.opts.test, |
| sess.edition(), |
| ); |
| // these need to be set "early" so that expansion sees `quote` if enabled. |
| sess.init_features(features); |
| |
| let crate_types = collect_crate_types(sess, &krate.attrs); |
| sess.crate_types.set(crate_types); |
| |
| let disambiguator = compute_crate_disambiguator(sess); |
| sess.crate_disambiguator.set(disambiguator); |
| rustc_incremental::prepare_session_directory(sess, &crate_name, disambiguator); |
| |
| if sess.opts.incremental.is_some() { |
| time(sess, "garbage collect incremental cache directory", || { |
| if let Err(e) = rustc_incremental::garbage_collect_session_directories(sess) { |
| warn!( |
| "Error while trying to garbage collect incremental \ |
| compilation cache directory: {}", |
| e |
| ); |
| } |
| }); |
| } |
| |
| // If necessary, compute the dependency graph (in the background). |
| let future_dep_graph = if sess.opts.build_dep_graph() { |
| Some(rustc_incremental::load_dep_graph(sess)) |
| } else { |
| None |
| }; |
| |
| time(sess, "recursion limit", || { |
| middle::recursion_limit::update_limits(sess, &krate); |
| }); |
| |
| krate = time(sess, "crate injection", || { |
| let alt_std_name = sess.opts.alt_std_name.as_ref().map(|s| &**s); |
| syntax::std_inject::maybe_inject_crates_ref(krate, alt_std_name) |
| }); |
| |
| let mut addl_plugins = Some(addl_plugins); |
| let registrars = time(sess, "plugin loading", || { |
| plugin::load::load_plugins( |
| sess, |
| &cstore, |
| &krate, |
| crate_name, |
| addl_plugins.take().unwrap(), |
| ) |
| }); |
| |
| let mut registry = registry.unwrap_or(Registry::new(sess, krate.span)); |
| |
| time(sess, "plugin registration", || { |
| if sess.features_untracked().rustc_diagnostic_macros { |
| registry.register_macro( |
| "__diagnostic_used", |
| diagnostics::plugin::expand_diagnostic_used, |
| ); |
| registry.register_macro( |
| "__register_diagnostic", |
| diagnostics::plugin::expand_register_diagnostic, |
| ); |
| registry.register_macro( |
| "__build_diagnostic_array", |
| diagnostics::plugin::expand_build_diagnostic_array, |
| ); |
| } |
| |
| for registrar in registrars { |
| registry.args_hidden = Some(registrar.args); |
| (registrar.fun)(&mut registry); |
| } |
| }); |
| |
| let whitelisted_legacy_custom_derives = registry.take_whitelisted_custom_derives(); |
| let Registry { |
| syntax_exts, |
| early_lint_passes, |
| late_lint_passes, |
| lint_groups, |
| llvm_passes, |
| attributes, |
| .. |
| } = registry; |
| |
| sess.track_errors(|| { |
| let mut ls = sess.lint_store.borrow_mut(); |
| for pass in early_lint_passes { |
| ls.register_early_pass(Some(sess), true, pass); |
| } |
| for pass in late_lint_passes { |
| ls.register_late_pass(Some(sess), true, pass); |
| } |
| |
| for (name, to) in lint_groups { |
| ls.register_group(Some(sess), true, name, to); |
| } |
| |
| *sess.plugin_llvm_passes.borrow_mut() = llvm_passes; |
| *sess.plugin_attributes.borrow_mut() = attributes.clone(); |
| })?; |
| |
| // Lint plugins are registered; now we can process command line flags. |
| if sess.opts.describe_lints { |
| super::describe_lints(&sess, &sess.lint_store.borrow(), true); |
| return Err(CompileIncomplete::Stopped); |
| } |
| |
| time(sess, "pre ast expansion lint checks", || { |
| lint::check_ast_crate(sess, &krate, true) |
| }); |
| |
| let mut resolver = Resolver::new( |
| sess, |
| cstore, |
| &krate, |
| crate_name, |
| make_glob_map, |
| crate_loader, |
| &resolver_arenas, |
| ); |
| resolver.whitelisted_legacy_custom_derives = whitelisted_legacy_custom_derives; |
| syntax_ext::register_builtins(&mut resolver, syntax_exts, sess.features_untracked().quote); |
| |
| // Expand all macros |
| krate = time(sess, "expansion", || { |
| // Windows dlls do not have rpaths, so they don't know how to find their |
| // dependencies. It's up to us to tell the system where to find all the |
| // dependent dlls. Note that this uses cfg!(windows) as opposed to |
| // targ_cfg because syntax extensions are always loaded for the host |
| // compiler, not for the target. |
| // |
| // This is somewhat of an inherently racy operation, however, as |
| // multiple threads calling this function could possibly continue |
| // extending PATH far beyond what it should. To solve this for now we |
| // just don't add any new elements to PATH which are already there |
| // within PATH. This is basically a targeted fix at #17360 for rustdoc |
| // which runs rustc in parallel but has been seen (#33844) to cause |
| // problems with PATH becoming too long. |
| let mut old_path = OsString::new(); |
| if cfg!(windows) { |
| old_path = env::var_os("PATH").unwrap_or(old_path); |
| let mut new_path = sess.host_filesearch(PathKind::All).get_dylib_search_paths(); |
| for path in env::split_paths(&old_path) { |
| if !new_path.contains(&path) { |
| new_path.push(path); |
| } |
| } |
| env::set_var( |
| "PATH", |
| &env::join_paths( |
| new_path |
| .iter() |
| .filter(|p| env::join_paths(iter::once(p)).is_ok()), |
| ).unwrap(), |
| ); |
| } |
| |
| // Create the config for macro expansion |
| let features = sess.features_untracked(); |
| let cfg = syntax::ext::expand::ExpansionConfig { |
| features: Some(&features), |
| recursion_limit: *sess.recursion_limit.get(), |
| trace_mac: sess.opts.debugging_opts.trace_macros, |
| should_test: sess.opts.test, |
| ..syntax::ext::expand::ExpansionConfig::default(crate_name.to_string()) |
| }; |
| |
| let mut ecx = ExtCtxt::new(&sess.parse_sess, cfg, &mut resolver); |
| let err_count = ecx.parse_sess.span_diagnostic.err_count(); |
| |
| // Expand macros now! |
| let krate = time(sess, "expand crate", || { |
| ecx.monotonic_expander().expand_crate(krate) |
| }); |
| |
| // The rest is error reporting |
| |
| time(sess, "check unused macros", || { |
| ecx.check_unused_macros(); |
| }); |
| |
| let mut missing_fragment_specifiers: Vec<_> = ecx.parse_sess |
| .missing_fragment_specifiers |
| .borrow() |
| .iter() |
| .cloned() |
| .collect(); |
| missing_fragment_specifiers.sort(); |
| for span in missing_fragment_specifiers { |
| let lint = lint::builtin::MISSING_FRAGMENT_SPECIFIER; |
| let msg = "missing fragment specifier"; |
| sess.buffer_lint(lint, ast::CRATE_NODE_ID, span, msg); |
| } |
| if ecx.parse_sess.span_diagnostic.err_count() - ecx.resolve_err_count > err_count { |
| ecx.parse_sess.span_diagnostic.abort_if_errors(); |
| } |
| if cfg!(windows) { |
| env::set_var("PATH", &old_path); |
| } |
| krate |
| }); |
| |
| krate = time(sess, "maybe building test harness", || { |
| syntax::test::modify_for_testing( |
| &sess.parse_sess, |
| &mut resolver, |
| sess.opts.test, |
| krate, |
| sess.diagnostic(), |
| &sess.features_untracked(), |
| ) |
| }); |
| |
| // If we're actually rustdoc then there's no need to actually compile |
| // anything, so switch everything to just looping |
| if sess.opts.actually_rustdoc { |
| krate = ReplaceBodyWithLoop::new(sess).fold_crate(krate); |
| } |
| |
| // If we're in rustdoc we're always compiling as an rlib, but that'll trip a |
| // bunch of checks in the `modify` function below. For now just skip this |
| // step entirely if we're rustdoc as it's not too useful anyway. |
| if !sess.opts.actually_rustdoc { |
| krate = time(sess, "maybe creating a macro crate", || { |
| let crate_types = sess.crate_types.borrow(); |
| let num_crate_types = crate_types.len(); |
| let is_proc_macro_crate = crate_types.contains(&config::CrateTypeProcMacro); |
| let is_test_crate = sess.opts.test; |
| syntax_ext::proc_macro_registrar::modify( |
| &sess.parse_sess, |
| &mut resolver, |
| krate, |
| is_proc_macro_crate, |
| is_test_crate, |
| num_crate_types, |
| sess.diagnostic(), |
| ) |
| }); |
| } |
| |
| // Expand global allocators, which are treated as an in-tree proc macro |
| krate = time(sess, "creating allocators", || { |
| allocator::expand::modify( |
| &sess.parse_sess, |
| &mut resolver, |
| krate, |
| crate_name.to_string(), |
| sess.diagnostic(), |
| ) |
| }); |
| |
| // Add all buffered lints from the `ParseSess` to the `Session`. |
| sess.parse_sess.buffered_lints.with_lock(|buffered_lints| { |
| info!("{} parse sess buffered_lints", buffered_lints.len()); |
| for BufferedEarlyLint{id, span, msg, lint_id} in buffered_lints.drain(..) { |
| let lint = lint::Lint::from_parser_lint_id(lint_id); |
| sess.buffer_lint(lint, id, span, &msg); |
| } |
| }); |
| |
| // Done with macro expansion! |
| |
| after_expand(&krate)?; |
| |
| if sess.opts.debugging_opts.input_stats { |
| println!("Post-expansion node count: {}", count_nodes(&krate)); |
| } |
| |
| if sess.opts.debugging_opts.hir_stats { |
| hir_stats::print_ast_stats(&krate, "POST EXPANSION AST STATS"); |
| } |
| |
| if sess.opts.debugging_opts.ast_json { |
| println!("{}", json::as_json(&krate)); |
| } |
| |
| time(sess, "AST validation", || { |
| ast_validation::check_crate(sess, &krate) |
| }); |
| |
| time(sess, "name resolution", || -> CompileResult { |
| resolver.resolve_crate(&krate); |
| Ok(()) |
| })?; |
| |
| // Needs to go *after* expansion to be able to check the results of macro expansion. |
| time(sess, "complete gated feature checking", || { |
| sess.track_errors(|| { |
| syntax::feature_gate::check_crate( |
| &krate, |
| &sess.parse_sess, |
| &sess.features_untracked(), |
| &attributes, |
| sess.opts.unstable_features, |
| ); |
| }) |
| })?; |
| |
| // Unresolved macros might be due to mistyped `#[macro_use]`, |
| // so abort after checking for unknown attributes. (#49074) |
| if resolver.found_unresolved_macro { |
| sess.parse_sess.span_diagnostic.abort_if_errors(); |
| } |
| |
| // Lower ast -> hir. |
| // First, we need to collect the dep_graph. |
| let dep_graph = match future_dep_graph { |
| None => DepGraph::new_disabled(), |
| Some(future) => { |
| let (prev_graph, prev_work_products) = |
| time(sess, "blocked while dep-graph loading finishes", || { |
| future |
| .open() |
| .unwrap_or_else(|e| rustc_incremental::LoadResult::Error { |
| message: format!("could not decode incremental cache: {:?}", e), |
| }) |
| .open(sess) |
| }); |
| DepGraph::new(prev_graph, prev_work_products) |
| } |
| }; |
| let hir_forest = time(sess, "lowering ast -> hir", || { |
| let hir_crate = lower_crate(sess, cstore, &dep_graph, &krate, &mut resolver); |
| |
| if sess.opts.debugging_opts.hir_stats { |
| hir_stats::print_hir_stats(&hir_crate); |
| } |
| |
| hir_map::Forest::new(hir_crate, &dep_graph) |
| }); |
| |
| time(sess, "early lint checks", || { |
| lint::check_ast_crate(sess, &krate, false) |
| }); |
| |
| // Discard hygiene data, which isn't required after lowering to HIR. |
| if !sess.opts.debugging_opts.keep_hygiene_data { |
| syntax::ext::hygiene::clear_markings(); |
| } |
| |
| Ok(InnerExpansionResult { |
| expanded_crate: krate, |
| resolver, |
| hir_forest, |
| }) |
| } |
| |
| pub fn default_provide(providers: &mut ty::query::Providers) { |
| hir::provide(providers); |
| borrowck::provide(providers); |
| mir::provide(providers); |
| reachable::provide(providers); |
| resolve_lifetime::provide(providers); |
| rustc_privacy::provide(providers); |
| typeck::provide(providers); |
| ty::provide(providers); |
| traits::provide(providers); |
| reachable::provide(providers); |
| rustc_passes::provide(providers); |
| rustc_traits::provide(providers); |
| middle::region::provide(providers); |
| cstore::provide(providers); |
| lint::provide(providers); |
| } |
| |
| pub fn default_provide_extern(providers: &mut ty::query::Providers) { |
| cstore::provide_extern(providers); |
| } |
| |
| /// Run the resolution, typechecking, region checking and other |
| /// miscellaneous analysis passes on the crate. Return various |
| /// structures carrying the results of the analysis. |
| pub fn phase_3_run_analysis_passes<'tcx, F, R>( |
| codegen_backend: &dyn CodegenBackend, |
| control: &CompileController, |
| sess: &'tcx Session, |
| cstore: &'tcx CrateStoreDyn, |
| hir_map: hir_map::Map<'tcx>, |
| mut analysis: ty::CrateAnalysis, |
| resolutions: Resolutions, |
| arenas: &'tcx AllArenas<'tcx>, |
| name: &str, |
| output_filenames: &OutputFilenames, |
| f: F, |
| ) -> Result<R, CompileIncomplete> |
| where |
| F: for<'a> FnOnce( |
| TyCtxt<'a, 'tcx, 'tcx>, |
| ty::CrateAnalysis, |
| mpsc::Receiver<Box<dyn Any + Send>>, |
| CompileResult, |
| ) -> R, |
| { |
| let query_result_on_disk_cache = time(sess, "load query result cache", || { |
| rustc_incremental::load_query_result_cache(sess) |
| }); |
| |
| time(sess, "looking for entry point", || { |
| middle::entry::find_entry_point(sess, &hir_map, name) |
| }); |
| |
| sess.plugin_registrar_fn |
| .set(time(sess, "looking for plugin registrar", || { |
| plugin::build::find_plugin_registrar(sess.diagnostic(), &hir_map) |
| })); |
| sess.derive_registrar_fn |
| .set(derive_registrar::find(&hir_map)); |
| |
| time(sess, "loop checking", || loops::check_crate(sess, &hir_map)); |
| |
| let mut local_providers = ty::query::Providers::default(); |
| default_provide(&mut local_providers); |
| codegen_backend.provide(&mut local_providers); |
| (control.provide)(&mut local_providers); |
| |
| let mut extern_providers = local_providers; |
| default_provide_extern(&mut extern_providers); |
| codegen_backend.provide_extern(&mut extern_providers); |
| (control.provide_extern)(&mut extern_providers); |
| |
| let (tx, rx) = mpsc::channel(); |
| |
| TyCtxt::create_and_enter( |
| sess, |
| cstore, |
| local_providers, |
| extern_providers, |
| arenas, |
| resolutions, |
| hir_map, |
| query_result_on_disk_cache, |
| name, |
| tx, |
| output_filenames, |
| |tcx| { |
| // Do some initialization of the DepGraph that can only be done with the |
| // tcx available. |
| rustc_incremental::dep_graph_tcx_init(tcx); |
| |
| time(sess, "attribute checking", || { |
| hir::check_attr::check_crate(tcx) |
| }); |
| |
| time(sess, "stability checking", || { |
| stability::check_unstable_api_usage(tcx) |
| }); |
| |
| // passes are timed inside typeck |
| match typeck::check_crate(tcx) { |
| Ok(x) => x, |
| Err(x) => { |
| f(tcx, analysis, rx, Err(x)); |
| return Err(x); |
| } |
| } |
| |
| time(sess, "rvalue promotion", || { |
| rvalue_promotion::check_crate(tcx) |
| }); |
| |
| analysis.access_levels = |
| time(sess, "privacy checking", || rustc_privacy::check_crate(tcx)); |
| |
| time(sess, "intrinsic checking", || { |
| middle::intrinsicck::check_crate(tcx) |
| }); |
| |
| time(sess, "match checking", || mir::matchck_crate(tcx)); |
| |
| // this must run before MIR dump, because |
| // "not all control paths return a value" is reported here. |
| // |
| // maybe move the check to a MIR pass? |
| time(sess, "liveness checking", || { |
| middle::liveness::check_crate(tcx) |
| }); |
| |
| time(sess, "borrow checking", || { |
| if tcx.use_ast_borrowck() { |
| borrowck::check_crate(tcx); |
| } |
| }); |
| |
| time(sess, |
| "MIR borrow checking", |
| || tcx.par_body_owners(|def_id| { tcx.mir_borrowck(def_id); })); |
| |
| time(sess, "dumping chalk-like clauses", || { |
| rustc_traits::lowering::dump_program_clauses(tcx); |
| }); |
| |
| time(sess, "MIR effect checking", || { |
| for def_id in tcx.body_owners() { |
| mir::transform::check_unsafety::check_unsafety(tcx, def_id) |
| } |
| }); |
| // Avoid overwhelming user with errors if type checking failed. |
| // I'm not sure how helpful this is, to be honest, but it avoids |
| // a |
| // lot of annoying errors in the compile-fail tests (basically, |
| // lint warnings and so on -- kindck used to do this abort, but |
| // kindck is gone now). -nmatsakis |
| if sess.err_count() > 0 { |
| return Ok(f(tcx, analysis, rx, sess.compile_status())); |
| } |
| |
| time(sess, "death checking", || middle::dead::check_crate(tcx)); |
| |
| time(sess, "unused lib feature checking", || { |
| stability::check_unused_or_stable_features(tcx) |
| }); |
| |
| time(sess, "lint checking", || lint::check_crate(tcx)); |
| |
| return Ok(f(tcx, analysis, rx, tcx.sess.compile_status())); |
| }, |
| ) |
| } |
| |
| /// Run the codegen backend, after which the AST and analysis can |
| /// be discarded. |
| pub fn phase_4_codegen<'a, 'tcx>( |
| codegen_backend: &dyn CodegenBackend, |
| tcx: TyCtxt<'a, 'tcx, 'tcx>, |
| rx: mpsc::Receiver<Box<dyn Any + Send>>, |
| ) -> Box<dyn Any> { |
| time(tcx.sess, "resolving dependency formats", || { |
| ::rustc::middle::dependency_format::calculate(tcx) |
| }); |
| |
| let codegen = time(tcx.sess, "codegen", move || codegen_backend.codegen_crate(tcx, rx)); |
| if tcx.sess.profile_queries() { |
| profile::dump(&tcx.sess, "profile_queries".to_string()) |
| } |
| |
| codegen |
| } |
| |
| fn escape_dep_filename(filename: &FileName) -> String { |
| // Apparently clang and gcc *only* escape spaces: |
| // http://llvm.org/klaus/clang/commit/9d50634cfc268ecc9a7250226dd5ca0e945240d4 |
| filename.to_string().replace(" ", "\\ ") |
| } |
| |
| // Returns all the paths that correspond to generated files. |
| fn generated_output_paths( |
| sess: &Session, |
| outputs: &OutputFilenames, |
| exact_name: bool, |
| crate_name: &str, |
| ) -> Vec<PathBuf> { |
| let mut out_filenames = Vec::new(); |
| for output_type in sess.opts.output_types.keys() { |
| let file = outputs.path(*output_type); |
| match *output_type { |
| // If the filename has been overridden using `-o`, it will not be modified |
| // by appending `.rlib`, `.exe`, etc., so we can skip this transformation. |
| OutputType::Exe if !exact_name => for crate_type in sess.crate_types.borrow().iter() { |
| let p = ::rustc_codegen_utils::link::filename_for_input( |
| sess, |
| *crate_type, |
| crate_name, |
| outputs, |
| ); |
| out_filenames.push(p); |
| }, |
| OutputType::DepInfo if sess.opts.debugging_opts.dep_info_omit_d_target => { |
| // Don't add the dep-info output when omitting it from dep-info targets |
| } |
| _ => { |
| out_filenames.push(file); |
| } |
| } |
| } |
| out_filenames |
| } |
| |
| // Runs `f` on every output file path and returns the first non-None result, or None if `f` |
| // returns None for every file path. |
| fn check_output<F, T>(output_paths: &[PathBuf], f: F) -> Option<T> |
| where |
| F: Fn(&PathBuf) -> Option<T>, |
| { |
| for output_path in output_paths { |
| if let Some(result) = f(output_path) { |
| return Some(result); |
| } |
| } |
| None |
| } |
| |
| pub fn output_contains_path(output_paths: &[PathBuf], input_path: &PathBuf) -> bool { |
| let input_path = input_path.canonicalize().ok(); |
| if input_path.is_none() { |
| return false; |
| } |
| let check = |output_path: &PathBuf| { |
| if output_path.canonicalize().ok() == input_path { |
| Some(()) |
| } else { |
| None |
| } |
| }; |
| check_output(output_paths, check).is_some() |
| } |
| |
| pub fn output_conflicts_with_dir(output_paths: &[PathBuf]) -> Option<PathBuf> { |
| let check = |output_path: &PathBuf| { |
| if output_path.is_dir() { |
| Some(output_path.clone()) |
| } else { |
| None |
| } |
| }; |
| check_output(output_paths, check) |
| } |
| |
| fn write_out_deps(sess: &Session, outputs: &OutputFilenames, out_filenames: &[PathBuf]) { |
| // Write out dependency rules to the dep-info file if requested |
| if !sess.opts.output_types.contains_key(&OutputType::DepInfo) { |
| return; |
| } |
| let deps_filename = outputs.path(OutputType::DepInfo); |
| |
| let result = (|| -> io::Result<()> { |
| // Build a list of files used to compile the output and |
| // write Makefile-compatible dependency rules |
| let files: Vec<String> = sess.codemap() |
| .files() |
| .iter() |
| .filter(|fmap| fmap.is_real_file()) |
| .filter(|fmap| !fmap.is_imported()) |
| .map(|fmap| escape_dep_filename(&fmap.name)) |
| .collect(); |
| let mut file = fs::File::create(&deps_filename)?; |
| for path in out_filenames { |
| write!(file, "{}: {}\n\n", path.display(), files.join(" "))?; |
| } |
| |
| // Emit a fake target for each input file to the compilation. This |
| // prevents `make` from spitting out an error if a file is later |
| // deleted. For more info see #28735 |
| for path in files { |
| writeln!(file, "{}:", path)?; |
| } |
| Ok(()) |
| })(); |
| |
| match result { |
| Ok(()) => {} |
| Err(e) => { |
| sess.fatal(&format!( |
| "error writing dependencies to `{}`: {}", |
| deps_filename.display(), |
| e |
| )); |
| } |
| } |
| } |
| |
| pub fn collect_crate_types(session: &Session, attrs: &[ast::Attribute]) -> Vec<config::CrateType> { |
| // Unconditionally collect crate types from attributes to make them used |
| let attr_types: Vec<config::CrateType> = attrs |
| .iter() |
| .filter_map(|a| { |
| if a.check_name("crate_type") { |
| match a.value_str() { |
| Some(ref n) if *n == "rlib" => Some(config::CrateTypeRlib), |
| Some(ref n) if *n == "dylib" => Some(config::CrateTypeDylib), |
| Some(ref n) if *n == "cdylib" => Some(config::CrateTypeCdylib), |
| Some(ref n) if *n == "lib" => Some(config::default_lib_output()), |
| Some(ref n) if *n == "staticlib" => Some(config::CrateTypeStaticlib), |
| Some(ref n) if *n == "proc-macro" => Some(config::CrateTypeProcMacro), |
| Some(ref n) if *n == "bin" => Some(config::CrateTypeExecutable), |
| Some(_) => { |
| session.buffer_lint( |
| lint::builtin::UNKNOWN_CRATE_TYPES, |
| ast::CRATE_NODE_ID, |
| a.span, |
| "invalid `crate_type` value", |
| ); |
| None |
| } |
| _ => { |
| session |
| .struct_span_err(a.span, "`crate_type` requires a value") |
| .note("for example: `#![crate_type=\"lib\"]`") |
| .emit(); |
| None |
| } |
| } |
| } else { |
| None |
| } |
| }) |
| .collect(); |
| |
| // If we're generating a test executable, then ignore all other output |
| // styles at all other locations |
| if session.opts.test { |
| return vec![config::CrateTypeExecutable]; |
| } |
| |
| // Only check command line flags if present. If no types are specified by |
| // command line, then reuse the empty `base` Vec to hold the types that |
| // will be found in crate attributes. |
| let mut base = session.opts.crate_types.clone(); |
| if base.is_empty() { |
| base.extend(attr_types); |
| if base.is_empty() { |
| base.push(::rustc_codegen_utils::link::default_output_for_target( |
| session, |
| )); |
| } |
| base.sort(); |
| base.dedup(); |
| } |
| |
| base.into_iter() |
| .filter(|crate_type| { |
| let res = !::rustc_codegen_utils::link::invalid_output_for_target(session, *crate_type); |
| |
| if !res { |
| session.warn(&format!( |
| "dropping unsupported crate type `{}` for target `{}`", |
| *crate_type, session.opts.target_triple |
| )); |
| } |
| |
| res |
| }) |
| .collect() |
| } |
| |
| pub fn compute_crate_disambiguator(session: &Session) -> CrateDisambiguator { |
| use std::hash::Hasher; |
| |
| // The crate_disambiguator is a 128 bit hash. The disambiguator is fed |
| // into various other hashes quite a bit (symbol hashes, incr. comp. hashes, |
| // debuginfo type IDs, etc), so we don't want it to be too wide. 128 bits |
| // should still be safe enough to avoid collisions in practice. |
| let mut hasher = StableHasher::<Fingerprint>::new(); |
| |
| let mut metadata = session.opts.cg.metadata.clone(); |
| // We don't want the crate_disambiguator to dependent on the order |
| // -C metadata arguments, so sort them: |
| metadata.sort(); |
| // Every distinct -C metadata value is only incorporated once: |
| metadata.dedup(); |
| |
| hasher.write(b"metadata"); |
| for s in &metadata { |
| // Also incorporate the length of a metadata string, so that we generate |
| // different values for `-Cmetadata=ab -Cmetadata=c` and |
| // `-Cmetadata=a -Cmetadata=bc` |
| hasher.write_usize(s.len()); |
| hasher.write(s.as_bytes()); |
| } |
| |
| // Also incorporate crate type, so that we don't get symbol conflicts when |
| // linking against a library of the same name, if this is an executable. |
| let is_exe = session |
| .crate_types |
| .borrow() |
| .contains(&config::CrateTypeExecutable); |
| hasher.write(if is_exe { b"exe" } else { b"lib" }); |
| |
| CrateDisambiguator::from(hasher.finish()) |
| } |
| |
| pub fn build_output_filenames( |
| input: &Input, |
| odir: &Option<PathBuf>, |
| ofile: &Option<PathBuf>, |
| attrs: &[ast::Attribute], |
| sess: &Session, |
| ) -> OutputFilenames { |
| match *ofile { |
| None => { |
| // "-" as input file will cause the parser to read from stdin so we |
| // have to make up a name |
| // We want to toss everything after the final '.' |
| let dirpath = match *odir { |
| Some(ref d) => d.clone(), |
| None => PathBuf::new(), |
| }; |
| |
| // If a crate name is present, we use it as the link name |
| let stem = sess.opts |
| .crate_name |
| .clone() |
| .or_else(|| attr::find_crate_name(attrs).map(|n| n.to_string())) |
| .unwrap_or(input.filestem()); |
| |
| OutputFilenames { |
| out_directory: dirpath, |
| out_filestem: stem, |
| single_output_file: None, |
| extra: sess.opts.cg.extra_filename.clone(), |
| outputs: sess.opts.output_types.clone(), |
| } |
| } |
| |
| Some(ref out_file) => { |
| let unnamed_output_types = sess.opts |
| .output_types |
| .values() |
| .filter(|a| a.is_none()) |
| .count(); |
| let ofile = if unnamed_output_types > 1 { |
| sess.warn( |
| "due to multiple output types requested, the explicitly specified \ |
| output file name will be adapted for each output type", |
| ); |
| None |
| } else { |
| Some(out_file.clone()) |
| }; |
| if *odir != None { |
| sess.warn("ignoring --out-dir flag due to -o flag"); |
| } |
| if !sess.opts.cg.extra_filename.is_empty() { |
| sess.warn("ignoring -C extra-filename flag due to -o flag"); |
| } |
| |
| let cur_dir = Path::new(""); |
| |
| OutputFilenames { |
| out_directory: out_file.parent().unwrap_or(cur_dir).to_path_buf(), |
| out_filestem: out_file |
| .file_stem() |
| .unwrap_or(OsStr::new("")) |
| .to_str() |
| .unwrap() |
| .to_string(), |
| single_output_file: ofile, |
| extra: sess.opts.cg.extra_filename.clone(), |
| outputs: sess.opts.output_types.clone(), |
| } |
| } |
| } |
| } |