| use crate::abi::FnAbi; |
| use crate::attributes; |
| use crate::debuginfo; |
| use crate::llvm; |
| use crate::llvm_util; |
| use crate::value::Value; |
| use rustc::dep_graph::DepGraphSafe; |
| |
| use crate::type_::Type; |
| use rustc_codegen_ssa::traits::*; |
| |
| use crate::callee::get_fn; |
| use rustc::bug; |
| use rustc::mir::mono::CodegenUnit; |
| use rustc::session::config::{self, CFGuard, DebugInfo}; |
| use rustc::session::Session; |
| use rustc::ty::layout::{ |
| FnAbiExt, HasParamEnv, LayoutError, LayoutOf, PointeeInfo, Size, TyLayout, VariantIdx, |
| }; |
| use rustc::ty::{self, Instance, Ty, TyCtxt}; |
| use rustc_codegen_ssa::base::wants_msvc_seh; |
| use rustc_data_structures::base_n; |
| use rustc_data_structures::const_cstr; |
| use rustc_data_structures::fx::FxHashMap; |
| use rustc_data_structures::small_c_str::SmallCStr; |
| use rustc_hir::Unsafety; |
| use rustc_target::spec::{HasTargetSpec, Target}; |
| |
| use crate::abi::Abi; |
| use rustc_span::source_map::{Span, DUMMY_SP}; |
| use rustc_span::symbol::Symbol; |
| use std::cell::{Cell, RefCell}; |
| use std::ffi::CStr; |
| use std::iter; |
| use std::str; |
| use std::sync::Arc; |
| |
| /// There is one `CodegenCx` per compilation unit. Each one has its own LLVM |
| /// `llvm::Context` so that several compilation units may be optimized in parallel. |
| /// All other LLVM data structures in the `CodegenCx` are tied to that `llvm::Context`. |
| pub struct CodegenCx<'ll, 'tcx> { |
| pub tcx: TyCtxt<'tcx>, |
| pub check_overflow: bool, |
| pub use_dll_storage_attrs: bool, |
| pub tls_model: llvm::ThreadLocalMode, |
| |
| pub llmod: &'ll llvm::Module, |
| pub llcx: &'ll llvm::Context, |
| pub codegen_unit: Arc<CodegenUnit<'tcx>>, |
| |
| /// Cache instances of monomorphic and polymorphic items |
| pub instances: RefCell<FxHashMap<Instance<'tcx>, &'ll Value>>, |
| /// Cache generated vtables |
| pub vtables: |
| RefCell<FxHashMap<(Ty<'tcx>, Option<ty::PolyExistentialTraitRef<'tcx>>), &'ll Value>>, |
| /// Cache of constant strings, |
| pub const_cstr_cache: RefCell<FxHashMap<Symbol, &'ll Value>>, |
| |
| /// Reverse-direction for const ptrs cast from globals. |
| /// Key is a Value holding a *T, |
| /// Val is a Value holding a *[T]. |
| /// |
| /// Needed because LLVM loses pointer->pointee association |
| /// when we ptrcast, and we have to ptrcast during codegen |
| /// of a [T] const because we form a slice, a (*T,usize) pair, not |
| /// a pointer to an LLVM array type. Similar for trait objects. |
| pub const_unsized: RefCell<FxHashMap<&'ll Value, &'ll Value>>, |
| |
| /// Cache of emitted const globals (value -> global) |
| pub const_globals: RefCell<FxHashMap<&'ll Value, &'ll Value>>, |
| |
| /// List of globals for static variables which need to be passed to the |
| /// LLVM function ReplaceAllUsesWith (RAUW) when codegen is complete. |
| /// (We have to make sure we don't invalidate any Values referring |
| /// to constants.) |
| pub statics_to_rauw: RefCell<Vec<(&'ll Value, &'ll Value)>>, |
| |
| /// Statics that will be placed in the llvm.used variable |
| /// See <http://llvm.org/docs/LangRef.html#the-llvm-used-global-variable> for details |
| pub used_statics: RefCell<Vec<&'ll Value>>, |
| |
| pub lltypes: RefCell<FxHashMap<(Ty<'tcx>, Option<VariantIdx>), &'ll Type>>, |
| pub scalar_lltypes: RefCell<FxHashMap<Ty<'tcx>, &'ll Type>>, |
| pub pointee_infos: RefCell<FxHashMap<(Ty<'tcx>, Size), Option<PointeeInfo>>>, |
| pub isize_ty: &'ll Type, |
| |
| pub dbg_cx: Option<debuginfo::CrateDebugContext<'ll, 'tcx>>, |
| |
| eh_personality: Cell<Option<&'ll Value>>, |
| eh_unwind_resume: Cell<Option<&'ll Value>>, |
| pub rust_try_fn: Cell<Option<&'ll Value>>, |
| |
| intrinsics: RefCell<FxHashMap<&'static str, &'ll Value>>, |
| |
| /// A counter that is used for generating local symbol names |
| local_gen_sym_counter: Cell<usize>, |
| } |
| |
| impl<'ll, 'tcx> DepGraphSafe for CodegenCx<'ll, 'tcx> {} |
| |
| pub fn get_reloc_model(sess: &Session) -> llvm::RelocMode { |
| let reloc_model_arg = match sess.opts.cg.relocation_model { |
| Some(ref s) => &s[..], |
| None => &sess.target.target.options.relocation_model[..], |
| }; |
| |
| match crate::back::write::RELOC_MODEL_ARGS.iter().find(|&&arg| arg.0 == reloc_model_arg) { |
| Some(x) => x.1, |
| _ => { |
| sess.err(&format!("{:?} is not a valid relocation mode", reloc_model_arg)); |
| sess.abort_if_errors(); |
| bug!(); |
| } |
| } |
| } |
| |
| fn get_tls_model(sess: &Session) -> llvm::ThreadLocalMode { |
| let tls_model_arg = match sess.opts.debugging_opts.tls_model { |
| Some(ref s) => &s[..], |
| None => &sess.target.target.options.tls_model[..], |
| }; |
| |
| match crate::back::write::TLS_MODEL_ARGS.iter().find(|&&arg| arg.0 == tls_model_arg) { |
| Some(x) => x.1, |
| _ => { |
| sess.err(&format!("{:?} is not a valid TLS model", tls_model_arg)); |
| sess.abort_if_errors(); |
| bug!(); |
| } |
| } |
| } |
| |
| fn is_any_library(sess: &Session) -> bool { |
| sess.crate_types.borrow().iter().any(|ty| *ty != config::CrateType::Executable) |
| } |
| |
| pub fn is_pie_binary(sess: &Session) -> bool { |
| !is_any_library(sess) && get_reloc_model(sess) == llvm::RelocMode::PIC |
| } |
| |
| fn strip_function_ptr_alignment(data_layout: String) -> String { |
| // FIXME: Make this more general. |
| data_layout.replace("-Fi8-", "-") |
| } |
| |
| fn strip_x86_address_spaces(data_layout: String) -> String { |
| data_layout.replace("-p270:32:32-p271:32:32-p272:64:64-", "-") |
| } |
| |
| pub unsafe fn create_module( |
| tcx: TyCtxt<'_>, |
| llcx: &'ll llvm::Context, |
| mod_name: &str, |
| ) -> &'ll llvm::Module { |
| let sess = tcx.sess; |
| let mod_name = SmallCStr::new(mod_name); |
| let llmod = llvm::LLVMModuleCreateWithNameInContext(mod_name.as_ptr(), llcx); |
| |
| let mut target_data_layout = sess.target.target.data_layout.clone(); |
| if llvm_util::get_major_version() < 9 { |
| target_data_layout = strip_function_ptr_alignment(target_data_layout); |
| } |
| if llvm_util::get_major_version() < 10 { |
| if sess.target.target.arch == "x86" || sess.target.target.arch == "x86_64" { |
| target_data_layout = strip_x86_address_spaces(target_data_layout); |
| } |
| } |
| |
| // Ensure the data-layout values hardcoded remain the defaults. |
| if sess.target.target.options.is_builtin { |
| let tm = crate::back::write::create_informational_target_machine(&tcx.sess, false); |
| llvm::LLVMRustSetDataLayoutFromTargetMachine(llmod, tm); |
| llvm::LLVMRustDisposeTargetMachine(tm); |
| |
| let llvm_data_layout = llvm::LLVMGetDataLayout(llmod); |
| let llvm_data_layout = str::from_utf8(CStr::from_ptr(llvm_data_layout).to_bytes()) |
| .ok() |
| .expect("got a non-UTF8 data-layout from LLVM"); |
| |
| // Unfortunately LLVM target specs change over time, and right now we |
| // don't have proper support to work with any more than one |
| // `data_layout` than the one that is in the rust-lang/rust repo. If |
| // this compiler is configured against a custom LLVM, we may have a |
| // differing data layout, even though we should update our own to use |
| // that one. |
| // |
| // As an interim hack, if CFG_LLVM_ROOT is not an empty string then we |
| // disable this check entirely as we may be configured with something |
| // that has a different target layout. |
| // |
| // Unsure if this will actually cause breakage when rustc is configured |
| // as such. |
| // |
| // FIXME(#34960) |
| let cfg_llvm_root = option_env!("CFG_LLVM_ROOT").unwrap_or(""); |
| let custom_llvm_used = cfg_llvm_root.trim() != ""; |
| |
| if !custom_llvm_used && target_data_layout != llvm_data_layout { |
| bug!( |
| "data-layout for builtin `{}` target, `{}`, \ |
| differs from LLVM default, `{}`", |
| sess.target.target.llvm_target, |
| target_data_layout, |
| llvm_data_layout |
| ); |
| } |
| } |
| |
| let data_layout = SmallCStr::new(&target_data_layout); |
| llvm::LLVMSetDataLayout(llmod, data_layout.as_ptr()); |
| |
| let llvm_target = SmallCStr::new(&sess.target.target.llvm_target); |
| llvm::LLVMRustSetNormalizedTarget(llmod, llvm_target.as_ptr()); |
| |
| if get_reloc_model(sess) == llvm::RelocMode::PIC { |
| llvm::LLVMRustSetModulePICLevel(llmod); |
| } |
| |
| if is_pie_binary(sess) { |
| llvm::LLVMRustSetModulePIELevel(llmod); |
| } |
| |
| // If skipping the PLT is enabled, we need to add some module metadata |
| // to ensure intrinsic calls don't use it. |
| if !sess.needs_plt() { |
| let avoid_plt = "RtLibUseGOT\0".as_ptr().cast(); |
| llvm::LLVMRustAddModuleFlag(llmod, avoid_plt, 1); |
| } |
| |
| // Set module flags to enable Windows Control Flow Guard (/guard:cf) metadata |
| // only (`cfguard=1`) or metadata and checks (`cfguard=2`). |
| match sess.opts.debugging_opts.control_flow_guard { |
| CFGuard::Disabled => {} |
| CFGuard::NoChecks => { |
| llvm::LLVMRustAddModuleFlag(llmod, "cfguard\0".as_ptr() as *const _, 1) |
| } |
| CFGuard::Checks => llvm::LLVMRustAddModuleFlag(llmod, "cfguard\0".as_ptr() as *const _, 2), |
| } |
| |
| llmod |
| } |
| |
| impl<'ll, 'tcx> CodegenCx<'ll, 'tcx> { |
| crate fn new( |
| tcx: TyCtxt<'tcx>, |
| codegen_unit: Arc<CodegenUnit<'tcx>>, |
| llvm_module: &'ll crate::ModuleLlvm, |
| ) -> Self { |
| // An interesting part of Windows which MSVC forces our hand on (and |
| // apparently MinGW didn't) is the usage of `dllimport` and `dllexport` |
| // attributes in LLVM IR as well as native dependencies (in C these |
| // correspond to `__declspec(dllimport)`). |
| // |
| // Whenever a dynamic library is built by MSVC it must have its public |
| // interface specified by functions tagged with `dllexport` or otherwise |
| // they're not available to be linked against. This poses a few problems |
| // for the compiler, some of which are somewhat fundamental, but we use |
| // the `use_dll_storage_attrs` variable below to attach the `dllexport` |
| // attribute to all LLVM functions that are exported e.g., they're |
| // already tagged with external linkage). This is suboptimal for a few |
| // reasons: |
| // |
| // * If an object file will never be included in a dynamic library, |
| // there's no need to attach the dllexport attribute. Most object |
| // files in Rust are not destined to become part of a dll as binaries |
| // are statically linked by default. |
| // * If the compiler is emitting both an rlib and a dylib, the same |
| // source object file is currently used but with MSVC this may be less |
| // feasible. The compiler may be able to get around this, but it may |
| // involve some invasive changes to deal with this. |
| // |
| // The flipside of this situation is that whenever you link to a dll and |
| // you import a function from it, the import should be tagged with |
| // `dllimport`. At this time, however, the compiler does not emit |
| // `dllimport` for any declarations other than constants (where it is |
| // required), which is again suboptimal for even more reasons! |
| // |
| // * Calling a function imported from another dll without using |
| // `dllimport` causes the linker/compiler to have extra overhead (one |
| // `jmp` instruction on x86) when calling the function. |
| // * The same object file may be used in different circumstances, so a |
| // function may be imported from a dll if the object is linked into a |
| // dll, but it may be just linked against if linked into an rlib. |
| // * The compiler has no knowledge about whether native functions should |
| // be tagged dllimport or not. |
| // |
| // For now the compiler takes the perf hit (I do not have any numbers to |
| // this effect) by marking very little as `dllimport` and praying the |
| // linker will take care of everything. Fixing this problem will likely |
| // require adding a few attributes to Rust itself (feature gated at the |
| // start) and then strongly recommending static linkage on MSVC! |
| let use_dll_storage_attrs = tcx.sess.target.target.options.is_like_msvc; |
| |
| let check_overflow = tcx.sess.overflow_checks(); |
| |
| let tls_model = get_tls_model(&tcx.sess); |
| |
| let (llcx, llmod) = (&*llvm_module.llcx, llvm_module.llmod()); |
| |
| let dbg_cx = if tcx.sess.opts.debuginfo != DebugInfo::None { |
| let dctx = debuginfo::CrateDebugContext::new(llmod); |
| debuginfo::metadata::compile_unit_metadata(tcx, &codegen_unit.name().as_str(), &dctx); |
| Some(dctx) |
| } else { |
| None |
| }; |
| |
| let isize_ty = Type::ix_llcx(llcx, tcx.data_layout.pointer_size.bits()); |
| |
| CodegenCx { |
| tcx, |
| check_overflow, |
| use_dll_storage_attrs, |
| tls_model, |
| llmod, |
| llcx, |
| codegen_unit, |
| instances: Default::default(), |
| vtables: Default::default(), |
| const_cstr_cache: Default::default(), |
| const_unsized: Default::default(), |
| const_globals: Default::default(), |
| statics_to_rauw: RefCell::new(Vec::new()), |
| used_statics: RefCell::new(Vec::new()), |
| lltypes: Default::default(), |
| scalar_lltypes: Default::default(), |
| pointee_infos: Default::default(), |
| isize_ty, |
| dbg_cx, |
| eh_personality: Cell::new(None), |
| eh_unwind_resume: Cell::new(None), |
| rust_try_fn: Cell::new(None), |
| intrinsics: Default::default(), |
| local_gen_sym_counter: Cell::new(0), |
| } |
| } |
| |
| crate fn statics_to_rauw(&self) -> &RefCell<Vec<(&'ll Value, &'ll Value)>> { |
| &self.statics_to_rauw |
| } |
| } |
| |
| impl MiscMethods<'tcx> for CodegenCx<'ll, 'tcx> { |
| fn vtables( |
| &self, |
| ) -> &RefCell<FxHashMap<(Ty<'tcx>, Option<ty::PolyExistentialTraitRef<'tcx>>), &'ll Value>> |
| { |
| &self.vtables |
| } |
| |
| fn get_fn(&self, instance: Instance<'tcx>) -> &'ll Value { |
| get_fn(self, instance) |
| } |
| |
| fn get_fn_addr(&self, instance: Instance<'tcx>) -> &'ll Value { |
| get_fn(self, instance) |
| } |
| |
| fn eh_personality(&self) -> &'ll Value { |
| // The exception handling personality function. |
| // |
| // If our compilation unit has the `eh_personality` lang item somewhere |
| // within it, then we just need to codegen that. Otherwise, we're |
| // building an rlib which will depend on some upstream implementation of |
| // this function, so we just codegen a generic reference to it. We don't |
| // specify any of the types for the function, we just make it a symbol |
| // that LLVM can later use. |
| // |
| // Note that MSVC is a little special here in that we don't use the |
| // `eh_personality` lang item at all. Currently LLVM has support for |
| // both Dwarf and SEH unwind mechanisms for MSVC targets and uses the |
| // *name of the personality function* to decide what kind of unwind side |
| // tables/landing pads to emit. It looks like Dwarf is used by default, |
| // injecting a dependency on the `_Unwind_Resume` symbol for resuming |
| // an "exception", but for MSVC we want to force SEH. This means that we |
| // can't actually have the personality function be our standard |
| // `rust_eh_personality` function, but rather we wired it up to the |
| // CRT's custom personality function, which forces LLVM to consider |
| // landing pads as "landing pads for SEH". |
| if let Some(llpersonality) = self.eh_personality.get() { |
| return llpersonality; |
| } |
| let tcx = self.tcx; |
| let llfn = match tcx.lang_items().eh_personality() { |
| Some(def_id) if !wants_msvc_seh(self.sess()) => self.get_fn_addr( |
| ty::Instance::resolve( |
| tcx, |
| ty::ParamEnv::reveal_all(), |
| def_id, |
| tcx.intern_substs(&[]), |
| ) |
| .unwrap(), |
| ), |
| _ => { |
| let name = if wants_msvc_seh(self.sess()) { |
| "__CxxFrameHandler3" |
| } else { |
| "rust_eh_personality" |
| }; |
| let fty = self.type_variadic_func(&[], self.type_i32()); |
| self.declare_cfn(name, fty) |
| } |
| }; |
| attributes::apply_target_cpu_attr(self, llfn); |
| self.eh_personality.set(Some(llfn)); |
| llfn |
| } |
| |
| // Returns a Value of the "eh_unwind_resume" lang item if one is defined, |
| // otherwise declares it as an external function. |
| fn eh_unwind_resume(&self) -> &'ll Value { |
| let unwresume = &self.eh_unwind_resume; |
| if let Some(llfn) = unwresume.get() { |
| return llfn; |
| } |
| |
| let tcx = self.tcx; |
| assert!(self.sess().target.target.options.custom_unwind_resume); |
| if let Some(def_id) = tcx.lang_items().eh_unwind_resume() { |
| let llfn = self.get_fn_addr( |
| ty::Instance::resolve( |
| tcx, |
| ty::ParamEnv::reveal_all(), |
| def_id, |
| tcx.intern_substs(&[]), |
| ) |
| .unwrap(), |
| ); |
| unwresume.set(Some(llfn)); |
| return llfn; |
| } |
| |
| let sig = ty::Binder::bind(tcx.mk_fn_sig( |
| iter::once(tcx.mk_mut_ptr(tcx.types.u8)), |
| tcx.types.never, |
| false, |
| Unsafety::Unsafe, |
| Abi::C, |
| )); |
| |
| let fn_abi = FnAbi::of_fn_ptr(self, sig, &[]); |
| let llfn = self.declare_fn("rust_eh_unwind_resume", &fn_abi); |
| attributes::apply_target_cpu_attr(self, llfn); |
| unwresume.set(Some(llfn)); |
| llfn |
| } |
| |
| fn sess(&self) -> &Session { |
| &self.tcx.sess |
| } |
| |
| fn check_overflow(&self) -> bool { |
| self.check_overflow |
| } |
| |
| fn codegen_unit(&self) -> &Arc<CodegenUnit<'tcx>> { |
| &self.codegen_unit |
| } |
| |
| fn used_statics(&self) -> &RefCell<Vec<&'ll Value>> { |
| &self.used_statics |
| } |
| |
| fn set_frame_pointer_elimination(&self, llfn: &'ll Value) { |
| attributes::set_frame_pointer_elimination(self, llfn) |
| } |
| |
| fn apply_target_cpu_attr(&self, llfn: &'ll Value) { |
| attributes::apply_target_cpu_attr(self, llfn) |
| } |
| |
| fn create_used_variable(&self) { |
| let name = const_cstr!("llvm.used"); |
| let section = const_cstr!("llvm.metadata"); |
| let array = |
| self.const_array(&self.type_ptr_to(self.type_i8()), &*self.used_statics.borrow()); |
| |
| unsafe { |
| let g = llvm::LLVMAddGlobal(self.llmod, self.val_ty(array), name.as_ptr()); |
| llvm::LLVMSetInitializer(g, array); |
| llvm::LLVMRustSetLinkage(g, llvm::Linkage::AppendingLinkage); |
| llvm::LLVMSetSection(g, section.as_ptr()); |
| } |
| } |
| } |
| |
| impl CodegenCx<'b, 'tcx> { |
| crate fn get_intrinsic(&self, key: &str) -> &'b Value { |
| if let Some(v) = self.intrinsics.borrow().get(key).cloned() { |
| return v; |
| } |
| |
| self.declare_intrinsic(key).unwrap_or_else(|| bug!("unknown intrinsic '{}'", key)) |
| } |
| |
| fn insert_intrinsic( |
| &self, |
| name: &'static str, |
| args: Option<&[&'b llvm::Type]>, |
| ret: &'b llvm::Type, |
| ) -> &'b llvm::Value { |
| let fn_ty = if let Some(args) = args { |
| self.type_func(args, ret) |
| } else { |
| self.type_variadic_func(&[], ret) |
| }; |
| let f = self.declare_cfn(name, fn_ty); |
| llvm::SetUnnamedAddr(f, false); |
| self.intrinsics.borrow_mut().insert(name, f); |
| f |
| } |
| |
| fn declare_intrinsic(&self, key: &str) -> Option<&'b Value> { |
| macro_rules! ifn { |
| ($name:expr, fn() -> $ret:expr) => ( |
| if key == $name { |
| return Some(self.insert_intrinsic($name, Some(&[]), $ret)); |
| } |
| ); |
| ($name:expr, fn(...) -> $ret:expr) => ( |
| if key == $name { |
| return Some(self.insert_intrinsic($name, None, $ret)); |
| } |
| ); |
| ($name:expr, fn($($arg:expr),*) -> $ret:expr) => ( |
| if key == $name { |
| return Some(self.insert_intrinsic($name, Some(&[$($arg),*]), $ret)); |
| } |
| ); |
| } |
| macro_rules! mk_struct { |
| ($($field_ty:expr),*) => (self.type_struct( &[$($field_ty),*], false)) |
| } |
| |
| let i8p = self.type_i8p(); |
| let void = self.type_void(); |
| let i1 = self.type_i1(); |
| let t_i8 = self.type_i8(); |
| let t_i16 = self.type_i16(); |
| let t_i32 = self.type_i32(); |
| let t_i64 = self.type_i64(); |
| let t_i128 = self.type_i128(); |
| let t_f32 = self.type_f32(); |
| let t_f64 = self.type_f64(); |
| |
| macro_rules! vector_types { |
| ($id_out:ident: $elem_ty:ident, $len:expr) => { |
| let $id_out = self.type_vector($elem_ty, $len); |
| }; |
| ($($id_out:ident: $elem_ty:ident, $len:expr;)*) => { |
| $(vector_types!($id_out: $elem_ty, $len);)* |
| } |
| } |
| vector_types! { |
| t_v2f32: t_f32, 2; |
| t_v4f32: t_f32, 4; |
| t_v8f32: t_f32, 8; |
| t_v16f32: t_f32, 16; |
| |
| t_v2f64: t_f64, 2; |
| t_v4f64: t_f64, 4; |
| t_v8f64: t_f64, 8; |
| } |
| |
| ifn!("llvm.trap", fn() -> void); |
| ifn!("llvm.debugtrap", fn() -> void); |
| ifn!("llvm.frameaddress", fn(t_i32) -> i8p); |
| ifn!("llvm.sideeffect", fn() -> void); |
| |
| ifn!("llvm.powi.f32", fn(t_f32, t_i32) -> t_f32); |
| ifn!("llvm.powi.v2f32", fn(t_v2f32, t_i32) -> t_v2f32); |
| ifn!("llvm.powi.v4f32", fn(t_v4f32, t_i32) -> t_v4f32); |
| ifn!("llvm.powi.v8f32", fn(t_v8f32, t_i32) -> t_v8f32); |
| ifn!("llvm.powi.v16f32", fn(t_v16f32, t_i32) -> t_v16f32); |
| ifn!("llvm.powi.f64", fn(t_f64, t_i32) -> t_f64); |
| ifn!("llvm.powi.v2f64", fn(t_v2f64, t_i32) -> t_v2f64); |
| ifn!("llvm.powi.v4f64", fn(t_v4f64, t_i32) -> t_v4f64); |
| ifn!("llvm.powi.v8f64", fn(t_v8f64, t_i32) -> t_v8f64); |
| |
| ifn!("llvm.pow.f32", fn(t_f32, t_f32) -> t_f32); |
| ifn!("llvm.pow.v2f32", fn(t_v2f32, t_v2f32) -> t_v2f32); |
| ifn!("llvm.pow.v4f32", fn(t_v4f32, t_v4f32) -> t_v4f32); |
| ifn!("llvm.pow.v8f32", fn(t_v8f32, t_v8f32) -> t_v8f32); |
| ifn!("llvm.pow.v16f32", fn(t_v16f32, t_v16f32) -> t_v16f32); |
| ifn!("llvm.pow.f64", fn(t_f64, t_f64) -> t_f64); |
| ifn!("llvm.pow.v2f64", fn(t_v2f64, t_v2f64) -> t_v2f64); |
| ifn!("llvm.pow.v4f64", fn(t_v4f64, t_v4f64) -> t_v4f64); |
| ifn!("llvm.pow.v8f64", fn(t_v8f64, t_v8f64) -> t_v8f64); |
| |
| ifn!("llvm.sqrt.f32", fn(t_f32) -> t_f32); |
| ifn!("llvm.sqrt.v2f32", fn(t_v2f32) -> t_v2f32); |
| ifn!("llvm.sqrt.v4f32", fn(t_v4f32) -> t_v4f32); |
| ifn!("llvm.sqrt.v8f32", fn(t_v8f32) -> t_v8f32); |
| ifn!("llvm.sqrt.v16f32", fn(t_v16f32) -> t_v16f32); |
| ifn!("llvm.sqrt.f64", fn(t_f64) -> t_f64); |
| ifn!("llvm.sqrt.v2f64", fn(t_v2f64) -> t_v2f64); |
| ifn!("llvm.sqrt.v4f64", fn(t_v4f64) -> t_v4f64); |
| ifn!("llvm.sqrt.v8f64", fn(t_v8f64) -> t_v8f64); |
| |
| ifn!("llvm.sin.f32", fn(t_f32) -> t_f32); |
| ifn!("llvm.sin.v2f32", fn(t_v2f32) -> t_v2f32); |
| ifn!("llvm.sin.v4f32", fn(t_v4f32) -> t_v4f32); |
| ifn!("llvm.sin.v8f32", fn(t_v8f32) -> t_v8f32); |
| ifn!("llvm.sin.v16f32", fn(t_v16f32) -> t_v16f32); |
| ifn!("llvm.sin.f64", fn(t_f64) -> t_f64); |
| ifn!("llvm.sin.v2f64", fn(t_v2f64) -> t_v2f64); |
| ifn!("llvm.sin.v4f64", fn(t_v4f64) -> t_v4f64); |
| ifn!("llvm.sin.v8f64", fn(t_v8f64) -> t_v8f64); |
| |
| ifn!("llvm.cos.f32", fn(t_f32) -> t_f32); |
| ifn!("llvm.cos.v2f32", fn(t_v2f32) -> t_v2f32); |
| ifn!("llvm.cos.v4f32", fn(t_v4f32) -> t_v4f32); |
| ifn!("llvm.cos.v8f32", fn(t_v8f32) -> t_v8f32); |
| ifn!("llvm.cos.v16f32", fn(t_v16f32) -> t_v16f32); |
| ifn!("llvm.cos.f64", fn(t_f64) -> t_f64); |
| ifn!("llvm.cos.v2f64", fn(t_v2f64) -> t_v2f64); |
| ifn!("llvm.cos.v4f64", fn(t_v4f64) -> t_v4f64); |
| ifn!("llvm.cos.v8f64", fn(t_v8f64) -> t_v8f64); |
| |
| ifn!("llvm.exp.f32", fn(t_f32) -> t_f32); |
| ifn!("llvm.exp.v2f32", fn(t_v2f32) -> t_v2f32); |
| ifn!("llvm.exp.v4f32", fn(t_v4f32) -> t_v4f32); |
| ifn!("llvm.exp.v8f32", fn(t_v8f32) -> t_v8f32); |
| ifn!("llvm.exp.v16f32", fn(t_v16f32) -> t_v16f32); |
| ifn!("llvm.exp.f64", fn(t_f64) -> t_f64); |
| ifn!("llvm.exp.v2f64", fn(t_v2f64) -> t_v2f64); |
| ifn!("llvm.exp.v4f64", fn(t_v4f64) -> t_v4f64); |
| ifn!("llvm.exp.v8f64", fn(t_v8f64) -> t_v8f64); |
| |
| ifn!("llvm.exp2.f32", fn(t_f32) -> t_f32); |
| ifn!("llvm.exp2.v2f32", fn(t_v2f32) -> t_v2f32); |
| ifn!("llvm.exp2.v4f32", fn(t_v4f32) -> t_v4f32); |
| ifn!("llvm.exp2.v8f32", fn(t_v8f32) -> t_v8f32); |
| ifn!("llvm.exp2.v16f32", fn(t_v16f32) -> t_v16f32); |
| ifn!("llvm.exp2.f64", fn(t_f64) -> t_f64); |
| ifn!("llvm.exp2.v2f64", fn(t_v2f64) -> t_v2f64); |
| ifn!("llvm.exp2.v4f64", fn(t_v4f64) -> t_v4f64); |
| ifn!("llvm.exp2.v8f64", fn(t_v8f64) -> t_v8f64); |
| |
| ifn!("llvm.log.f32", fn(t_f32) -> t_f32); |
| ifn!("llvm.log.v2f32", fn(t_v2f32) -> t_v2f32); |
| ifn!("llvm.log.v4f32", fn(t_v4f32) -> t_v4f32); |
| ifn!("llvm.log.v8f32", fn(t_v8f32) -> t_v8f32); |
| ifn!("llvm.log.v16f32", fn(t_v16f32) -> t_v16f32); |
| ifn!("llvm.log.f64", fn(t_f64) -> t_f64); |
| ifn!("llvm.log.v2f64", fn(t_v2f64) -> t_v2f64); |
| ifn!("llvm.log.v4f64", fn(t_v4f64) -> t_v4f64); |
| ifn!("llvm.log.v8f64", fn(t_v8f64) -> t_v8f64); |
| |
| ifn!("llvm.log10.f32", fn(t_f32) -> t_f32); |
| ifn!("llvm.log10.v2f32", fn(t_v2f32) -> t_v2f32); |
| ifn!("llvm.log10.v4f32", fn(t_v4f32) -> t_v4f32); |
| ifn!("llvm.log10.v8f32", fn(t_v8f32) -> t_v8f32); |
| ifn!("llvm.log10.v16f32", fn(t_v16f32) -> t_v16f32); |
| ifn!("llvm.log10.f64", fn(t_f64) -> t_f64); |
| ifn!("llvm.log10.v2f64", fn(t_v2f64) -> t_v2f64); |
| ifn!("llvm.log10.v4f64", fn(t_v4f64) -> t_v4f64); |
| ifn!("llvm.log10.v8f64", fn(t_v8f64) -> t_v8f64); |
| |
| ifn!("llvm.log2.f32", fn(t_f32) -> t_f32); |
| ifn!("llvm.log2.v2f32", fn(t_v2f32) -> t_v2f32); |
| ifn!("llvm.log2.v4f32", fn(t_v4f32) -> t_v4f32); |
| ifn!("llvm.log2.v8f32", fn(t_v8f32) -> t_v8f32); |
| ifn!("llvm.log2.v16f32", fn(t_v16f32) -> t_v16f32); |
| ifn!("llvm.log2.f64", fn(t_f64) -> t_f64); |
| ifn!("llvm.log2.v2f64", fn(t_v2f64) -> t_v2f64); |
| ifn!("llvm.log2.v4f64", fn(t_v4f64) -> t_v4f64); |
| ifn!("llvm.log2.v8f64", fn(t_v8f64) -> t_v8f64); |
| |
| ifn!("llvm.fma.f32", fn(t_f32, t_f32, t_f32) -> t_f32); |
| ifn!("llvm.fma.v2f32", fn(t_v2f32, t_v2f32, t_v2f32) -> t_v2f32); |
| ifn!("llvm.fma.v4f32", fn(t_v4f32, t_v4f32, t_v4f32) -> t_v4f32); |
| ifn!("llvm.fma.v8f32", fn(t_v8f32, t_v8f32, t_v8f32) -> t_v8f32); |
| ifn!("llvm.fma.v16f32", fn(t_v16f32, t_v16f32, t_v16f32) -> t_v16f32); |
| ifn!("llvm.fma.f64", fn(t_f64, t_f64, t_f64) -> t_f64); |
| ifn!("llvm.fma.v2f64", fn(t_v2f64, t_v2f64, t_v2f64) -> t_v2f64); |
| ifn!("llvm.fma.v4f64", fn(t_v4f64, t_v4f64, t_v4f64) -> t_v4f64); |
| ifn!("llvm.fma.v8f64", fn(t_v8f64, t_v8f64, t_v8f64) -> t_v8f64); |
| |
| ifn!("llvm.fabs.f32", fn(t_f32) -> t_f32); |
| ifn!("llvm.fabs.v2f32", fn(t_v2f32) -> t_v2f32); |
| ifn!("llvm.fabs.v4f32", fn(t_v4f32) -> t_v4f32); |
| ifn!("llvm.fabs.v8f32", fn(t_v8f32) -> t_v8f32); |
| ifn!("llvm.fabs.v16f32", fn(t_v16f32) -> t_v16f32); |
| ifn!("llvm.fabs.f64", fn(t_f64) -> t_f64); |
| ifn!("llvm.fabs.v2f64", fn(t_v2f64) -> t_v2f64); |
| ifn!("llvm.fabs.v4f64", fn(t_v4f64) -> t_v4f64); |
| ifn!("llvm.fabs.v8f64", fn(t_v8f64) -> t_v8f64); |
| |
| ifn!("llvm.minnum.f32", fn(t_f32, t_f32) -> t_f32); |
| ifn!("llvm.minnum.f64", fn(t_f64, t_f64) -> t_f64); |
| ifn!("llvm.maxnum.f32", fn(t_f32, t_f32) -> t_f32); |
| ifn!("llvm.maxnum.f64", fn(t_f64, t_f64) -> t_f64); |
| |
| ifn!("llvm.floor.f32", fn(t_f32) -> t_f32); |
| ifn!("llvm.floor.v2f32", fn(t_v2f32) -> t_v2f32); |
| ifn!("llvm.floor.v4f32", fn(t_v4f32) -> t_v4f32); |
| ifn!("llvm.floor.v8f32", fn(t_v8f32) -> t_v8f32); |
| ifn!("llvm.floor.v16f32", fn(t_v16f32) -> t_v16f32); |
| ifn!("llvm.floor.f64", fn(t_f64) -> t_f64); |
| ifn!("llvm.floor.v2f64", fn(t_v2f64) -> t_v2f64); |
| ifn!("llvm.floor.v4f64", fn(t_v4f64) -> t_v4f64); |
| ifn!("llvm.floor.v8f64", fn(t_v8f64) -> t_v8f64); |
| |
| ifn!("llvm.ceil.f32", fn(t_f32) -> t_f32); |
| ifn!("llvm.ceil.v2f32", fn(t_v2f32) -> t_v2f32); |
| ifn!("llvm.ceil.v4f32", fn(t_v4f32) -> t_v4f32); |
| ifn!("llvm.ceil.v8f32", fn(t_v8f32) -> t_v8f32); |
| ifn!("llvm.ceil.v16f32", fn(t_v16f32) -> t_v16f32); |
| ifn!("llvm.ceil.f64", fn(t_f64) -> t_f64); |
| ifn!("llvm.ceil.v2f64", fn(t_v2f64) -> t_v2f64); |
| ifn!("llvm.ceil.v4f64", fn(t_v4f64) -> t_v4f64); |
| ifn!("llvm.ceil.v8f64", fn(t_v8f64) -> t_v8f64); |
| |
| ifn!("llvm.trunc.f32", fn(t_f32) -> t_f32); |
| ifn!("llvm.trunc.f64", fn(t_f64) -> t_f64); |
| |
| ifn!("llvm.copysign.f32", fn(t_f32, t_f32) -> t_f32); |
| ifn!("llvm.copysign.f64", fn(t_f64, t_f64) -> t_f64); |
| ifn!("llvm.round.f32", fn(t_f32) -> t_f32); |
| ifn!("llvm.round.f64", fn(t_f64) -> t_f64); |
| |
| ifn!("llvm.rint.f32", fn(t_f32) -> t_f32); |
| ifn!("llvm.rint.f64", fn(t_f64) -> t_f64); |
| ifn!("llvm.nearbyint.f32", fn(t_f32) -> t_f32); |
| ifn!("llvm.nearbyint.f64", fn(t_f64) -> t_f64); |
| |
| ifn!("llvm.ctpop.i8", fn(t_i8) -> t_i8); |
| ifn!("llvm.ctpop.i16", fn(t_i16) -> t_i16); |
| ifn!("llvm.ctpop.i32", fn(t_i32) -> t_i32); |
| ifn!("llvm.ctpop.i64", fn(t_i64) -> t_i64); |
| ifn!("llvm.ctpop.i128", fn(t_i128) -> t_i128); |
| |
| ifn!("llvm.ctlz.i8", fn(t_i8, i1) -> t_i8); |
| ifn!("llvm.ctlz.i16", fn(t_i16, i1) -> t_i16); |
| ifn!("llvm.ctlz.i32", fn(t_i32, i1) -> t_i32); |
| ifn!("llvm.ctlz.i64", fn(t_i64, i1) -> t_i64); |
| ifn!("llvm.ctlz.i128", fn(t_i128, i1) -> t_i128); |
| |
| ifn!("llvm.cttz.i8", fn(t_i8, i1) -> t_i8); |
| ifn!("llvm.cttz.i16", fn(t_i16, i1) -> t_i16); |
| ifn!("llvm.cttz.i32", fn(t_i32, i1) -> t_i32); |
| ifn!("llvm.cttz.i64", fn(t_i64, i1) -> t_i64); |
| ifn!("llvm.cttz.i128", fn(t_i128, i1) -> t_i128); |
| |
| ifn!("llvm.bswap.i16", fn(t_i16) -> t_i16); |
| ifn!("llvm.bswap.i32", fn(t_i32) -> t_i32); |
| ifn!("llvm.bswap.i64", fn(t_i64) -> t_i64); |
| ifn!("llvm.bswap.i128", fn(t_i128) -> t_i128); |
| |
| ifn!("llvm.bitreverse.i8", fn(t_i8) -> t_i8); |
| ifn!("llvm.bitreverse.i16", fn(t_i16) -> t_i16); |
| ifn!("llvm.bitreverse.i32", fn(t_i32) -> t_i32); |
| ifn!("llvm.bitreverse.i64", fn(t_i64) -> t_i64); |
| ifn!("llvm.bitreverse.i128", fn(t_i128) -> t_i128); |
| |
| ifn!("llvm.fshl.i8", fn(t_i8, t_i8, t_i8) -> t_i8); |
| ifn!("llvm.fshl.i16", fn(t_i16, t_i16, t_i16) -> t_i16); |
| ifn!("llvm.fshl.i32", fn(t_i32, t_i32, t_i32) -> t_i32); |
| ifn!("llvm.fshl.i64", fn(t_i64, t_i64, t_i64) -> t_i64); |
| ifn!("llvm.fshl.i128", fn(t_i128, t_i128, t_i128) -> t_i128); |
| |
| ifn!("llvm.fshr.i8", fn(t_i8, t_i8, t_i8) -> t_i8); |
| ifn!("llvm.fshr.i16", fn(t_i16, t_i16, t_i16) -> t_i16); |
| ifn!("llvm.fshr.i32", fn(t_i32, t_i32, t_i32) -> t_i32); |
| ifn!("llvm.fshr.i64", fn(t_i64, t_i64, t_i64) -> t_i64); |
| ifn!("llvm.fshr.i128", fn(t_i128, t_i128, t_i128) -> t_i128); |
| |
| ifn!("llvm.sadd.with.overflow.i8", fn(t_i8, t_i8) -> mk_struct! {t_i8, i1}); |
| ifn!("llvm.sadd.with.overflow.i16", fn(t_i16, t_i16) -> mk_struct! {t_i16, i1}); |
| ifn!("llvm.sadd.with.overflow.i32", fn(t_i32, t_i32) -> mk_struct! {t_i32, i1}); |
| ifn!("llvm.sadd.with.overflow.i64", fn(t_i64, t_i64) -> mk_struct! {t_i64, i1}); |
| ifn!("llvm.sadd.with.overflow.i128", fn(t_i128, t_i128) -> mk_struct! {t_i128, i1}); |
| |
| ifn!("llvm.uadd.with.overflow.i8", fn(t_i8, t_i8) -> mk_struct! {t_i8, i1}); |
| ifn!("llvm.uadd.with.overflow.i16", fn(t_i16, t_i16) -> mk_struct! {t_i16, i1}); |
| ifn!("llvm.uadd.with.overflow.i32", fn(t_i32, t_i32) -> mk_struct! {t_i32, i1}); |
| ifn!("llvm.uadd.with.overflow.i64", fn(t_i64, t_i64) -> mk_struct! {t_i64, i1}); |
| ifn!("llvm.uadd.with.overflow.i128", fn(t_i128, t_i128) -> mk_struct! {t_i128, i1}); |
| |
| ifn!("llvm.ssub.with.overflow.i8", fn(t_i8, t_i8) -> mk_struct! {t_i8, i1}); |
| ifn!("llvm.ssub.with.overflow.i16", fn(t_i16, t_i16) -> mk_struct! {t_i16, i1}); |
| ifn!("llvm.ssub.with.overflow.i32", fn(t_i32, t_i32) -> mk_struct! {t_i32, i1}); |
| ifn!("llvm.ssub.with.overflow.i64", fn(t_i64, t_i64) -> mk_struct! {t_i64, i1}); |
| ifn!("llvm.ssub.with.overflow.i128", fn(t_i128, t_i128) -> mk_struct! {t_i128, i1}); |
| |
| ifn!("llvm.usub.with.overflow.i8", fn(t_i8, t_i8) -> mk_struct! {t_i8, i1}); |
| ifn!("llvm.usub.with.overflow.i16", fn(t_i16, t_i16) -> mk_struct! {t_i16, i1}); |
| ifn!("llvm.usub.with.overflow.i32", fn(t_i32, t_i32) -> mk_struct! {t_i32, i1}); |
| ifn!("llvm.usub.with.overflow.i64", fn(t_i64, t_i64) -> mk_struct! {t_i64, i1}); |
| ifn!("llvm.usub.with.overflow.i128", fn(t_i128, t_i128) -> mk_struct! {t_i128, i1}); |
| |
| ifn!("llvm.smul.with.overflow.i8", fn(t_i8, t_i8) -> mk_struct! {t_i8, i1}); |
| ifn!("llvm.smul.with.overflow.i16", fn(t_i16, t_i16) -> mk_struct! {t_i16, i1}); |
| ifn!("llvm.smul.with.overflow.i32", fn(t_i32, t_i32) -> mk_struct! {t_i32, i1}); |
| ifn!("llvm.smul.with.overflow.i64", fn(t_i64, t_i64) -> mk_struct! {t_i64, i1}); |
| ifn!("llvm.smul.with.overflow.i128", fn(t_i128, t_i128) -> mk_struct! {t_i128, i1}); |
| |
| ifn!("llvm.umul.with.overflow.i8", fn(t_i8, t_i8) -> mk_struct! {t_i8, i1}); |
| ifn!("llvm.umul.with.overflow.i16", fn(t_i16, t_i16) -> mk_struct! {t_i16, i1}); |
| ifn!("llvm.umul.with.overflow.i32", fn(t_i32, t_i32) -> mk_struct! {t_i32, i1}); |
| ifn!("llvm.umul.with.overflow.i64", fn(t_i64, t_i64) -> mk_struct! {t_i64, i1}); |
| ifn!("llvm.umul.with.overflow.i128", fn(t_i128, t_i128) -> mk_struct! {t_i128, i1}); |
| |
| ifn!("llvm.sadd.sat.i8", fn(t_i8, t_i8) -> t_i8); |
| ifn!("llvm.sadd.sat.i16", fn(t_i16, t_i16) -> t_i16); |
| ifn!("llvm.sadd.sat.i32", fn(t_i32, t_i32) -> t_i32); |
| ifn!("llvm.sadd.sat.i64", fn(t_i64, t_i64) -> t_i64); |
| ifn!("llvm.sadd.sat.i128", fn(t_i128, t_i128) -> t_i128); |
| |
| ifn!("llvm.uadd.sat.i8", fn(t_i8, t_i8) -> t_i8); |
| ifn!("llvm.uadd.sat.i16", fn(t_i16, t_i16) -> t_i16); |
| ifn!("llvm.uadd.sat.i32", fn(t_i32, t_i32) -> t_i32); |
| ifn!("llvm.uadd.sat.i64", fn(t_i64, t_i64) -> t_i64); |
| ifn!("llvm.uadd.sat.i128", fn(t_i128, t_i128) -> t_i128); |
| |
| ifn!("llvm.ssub.sat.i8", fn(t_i8, t_i8) -> t_i8); |
| ifn!("llvm.ssub.sat.i16", fn(t_i16, t_i16) -> t_i16); |
| ifn!("llvm.ssub.sat.i32", fn(t_i32, t_i32) -> t_i32); |
| ifn!("llvm.ssub.sat.i64", fn(t_i64, t_i64) -> t_i64); |
| ifn!("llvm.ssub.sat.i128", fn(t_i128, t_i128) -> t_i128); |
| |
| ifn!("llvm.usub.sat.i8", fn(t_i8, t_i8) -> t_i8); |
| ifn!("llvm.usub.sat.i16", fn(t_i16, t_i16) -> t_i16); |
| ifn!("llvm.usub.sat.i32", fn(t_i32, t_i32) -> t_i32); |
| ifn!("llvm.usub.sat.i64", fn(t_i64, t_i64) -> t_i64); |
| ifn!("llvm.usub.sat.i128", fn(t_i128, t_i128) -> t_i128); |
| |
| ifn!("llvm.lifetime.start.p0i8", fn(t_i64, i8p) -> void); |
| ifn!("llvm.lifetime.end.p0i8", fn(t_i64, i8p) -> void); |
| |
| ifn!("llvm.expect.i1", fn(i1, i1) -> i1); |
| ifn!("llvm.eh.typeid.for", fn(i8p) -> t_i32); |
| ifn!("llvm.localescape", fn(...) -> void); |
| ifn!("llvm.localrecover", fn(i8p, i8p, t_i32) -> i8p); |
| ifn!("llvm.x86.seh.recoverfp", fn(i8p, i8p) -> i8p); |
| |
| ifn!("llvm.assume", fn(i1) -> void); |
| ifn!("llvm.prefetch", fn(i8p, t_i32, t_i32, t_i32) -> void); |
| |
| // variadic intrinsics |
| ifn!("llvm.va_start", fn(i8p) -> void); |
| ifn!("llvm.va_end", fn(i8p) -> void); |
| ifn!("llvm.va_copy", fn(i8p, i8p) -> void); |
| |
| if self.sess().opts.debuginfo != DebugInfo::None { |
| ifn!("llvm.dbg.declare", fn(self.type_metadata(), self.type_metadata()) -> void); |
| ifn!("llvm.dbg.value", fn(self.type_metadata(), t_i64, self.type_metadata()) -> void); |
| } |
| return None; |
| } |
| } |
| |
| impl<'b, 'tcx> CodegenCx<'b, 'tcx> { |
| /// Generates a new symbol name with the given prefix. This symbol name must |
| /// only be used for definitions with `internal` or `private` linkage. |
| pub fn generate_local_symbol_name(&self, prefix: &str) -> String { |
| let idx = self.local_gen_sym_counter.get(); |
| self.local_gen_sym_counter.set(idx + 1); |
| // Include a '.' character, so there can be no accidental conflicts with |
| // user defined names |
| let mut name = String::with_capacity(prefix.len() + 6); |
| name.push_str(prefix); |
| name.push_str("."); |
| base_n::push_str(idx as u128, base_n::ALPHANUMERIC_ONLY, &mut name); |
| name |
| } |
| } |
| |
| impl ty::layout::HasDataLayout for CodegenCx<'ll, 'tcx> { |
| fn data_layout(&self) -> &ty::layout::TargetDataLayout { |
| &self.tcx.data_layout |
| } |
| } |
| |
| impl HasTargetSpec for CodegenCx<'ll, 'tcx> { |
| fn target_spec(&self) -> &Target { |
| &self.tcx.sess.target.target |
| } |
| } |
| |
| impl ty::layout::HasTyCtxt<'tcx> for CodegenCx<'ll, 'tcx> { |
| fn tcx(&self) -> TyCtxt<'tcx> { |
| self.tcx |
| } |
| } |
| |
| impl LayoutOf for CodegenCx<'ll, 'tcx> { |
| type Ty = Ty<'tcx>; |
| type TyLayout = TyLayout<'tcx>; |
| |
| fn layout_of(&self, ty: Ty<'tcx>) -> Self::TyLayout { |
| self.spanned_layout_of(ty, DUMMY_SP) |
| } |
| |
| fn spanned_layout_of(&self, ty: Ty<'tcx>, span: Span) -> Self::TyLayout { |
| self.tcx.layout_of(ty::ParamEnv::reveal_all().and(ty)).unwrap_or_else(|e| { |
| if let LayoutError::SizeOverflow(_) = e { |
| self.sess().span_fatal(span, &e.to_string()) |
| } else { |
| bug!("failed to get layout for `{}`: {}", ty, e) |
| } |
| }) |
| } |
| } |
| |
| impl<'tcx, 'll> HasParamEnv<'tcx> for CodegenCx<'ll, 'tcx> { |
| fn param_env(&self) -> ty::ParamEnv<'tcx> { |
| ty::ParamEnv::reveal_all() |
| } |
| } |