-------------------------------------------------------------------------------- I1 cache: 65536 B, 64 B, 4-way associative D1 cache: 32768 B, 64 B, 8-way associative LL cache: 67108864 B, 64 B, 64-way associative Command: /usr/home/liquid/.rustup/toolchains/w-profiling/bin/rustc --crate-name ctrlc --edition=2018 src/lib.rs --error-format=json --json=diagnostic-rendered-ansi,artifacts,future-incompat --crate-type lib --emit=dep-info,metadata,link -C opt-level=3 -C embed-bitcode=no -C metadata=6760245488357057 -C extra-filename=-6760245488357057 --out-dir /usr/home/liquid/tmp/.tmpNHkXLx/target/release/deps -L dependency=/usr/home/liquid/tmp/.tmpNHkXLx/target/release/deps --extern nix=/usr/home/liquid/tmp/.tmpNHkXLx/target/release/deps/libnix-a70a0fdadbfbb35a.rmeta -Adeprecated -Aunknown-lints -Zincremental-verify-ich Data file: results/cgout-w-profiling-ctrlc-3.2.1-Opt-Full Events recorded: Ir Events shown: Ir Event sort order: Ir Thresholds: 0.1 Include dirs: User annotated: Auto-annotation: on -------------------------------------------------------------------------------- Ir -------------------------------------------------------------------------------- 473,243,703 (100.0%) PROGRAM TOTALS -------------------------------------------------------------------------------- Ir file:function -------------------------------------------------------------------------------- 18,245,159 ( 3.86%) ./malloc/malloc.c:_int_malloc 17,678,813 ( 3.74%) ./malloc/malloc.c:_int_free 16,665,784 ( 3.52%) ???:llvm::FPPassManager::runOnFunction(llvm::Function&) 11,273,832 ( 2.38%) ./malloc/malloc.c:malloc 10,309,110 ( 2.18%) ./elf/dl-lookup.c:_dl_lookup_symbol_x 7,069,559 ( 1.49%) ???:llvm::PassRegistry::enumerateWith(llvm::PassRegistrationListener*) 5,576,057 ( 1.18%) ./malloc/malloc.c:free 5,242,299 ( 1.11%) ???:llvm::AnalysisManager::getResultImpl(llvm::AnalysisKey*, llvm::Function&) 4,845,521 ( 1.02%) ./elf/dl-lookup.c:do_lookup_x 4,740,874 ( 1.00%) ???:llvm::AnalysisManager::invalidate(llvm::Function&, llvm::PreservedAnalyses const&) 4,643,660 ( 0.98%) ./string/../sysdeps/x86_64/multiarch/memmove-vec-unaligned-erms.S:__memcpy_avx_unaligned_erms 4,533,666 ( 0.96%) ./string/../sysdeps/x86_64/multiarch/memset-vec-unaligned-erms.S:__memset_avx2_erms 4,461,060 ( 0.94%) ???:llvm::X86TargetLowering::X86TargetLowering(llvm::X86TargetMachine const&, llvm::X86Subtarget const&) 4,141,271 ( 0.88%) ./elf/../sysdeps/x86_64/dl-machine.h:_dl_relocate_object 3,865,192 ( 0.82%) ???:llvm::BitstreamCursor::readRecord(unsigned int, llvm::SmallVectorImpl&, llvm::StringRef*) 3,816,200 ( 0.81%) ???:llvm::TargetLoweringBase::computeRegisterProperties(llvm::TargetRegisterInfo const*) 3,416,924 ( 0.72%) ./string/../sysdeps/x86_64/multiarch/memcmp-avx2-movbe.S:__memcmp_avx2_movbe 3,121,634 ( 0.66%) ???:SetImpliedBits(llvm::FeatureBitset&, llvm::FeatureBitset const&, llvm::ArrayRef) 3,097,129 ( 0.65%) /usr/home/liquid/rust/worktree-benchmarking/compiler/rustc_serialize/src/opaque.rs:>::decode 2,917,136 ( 0.62%) ???:llvm::PMTopLevelManager::setLastUser(llvm::ArrayRef, llvm::Pass*) 2,866,770 ( 0.61%) ./malloc/malloc.c:malloc_consolidate 2,779,507 ( 0.59%) /usr/home/liquid/rust/worktree-benchmarking/library/alloc/src/vec/mod.rs:>::decode 2,706,753 ( 0.57%) ???:llvm::AttributeList::addAttributes(llvm::LLVMContext&, unsigned int, llvm::AttrBuilder const&) const 2,682,640 ( 0.57%) ???:llvm::InstCombinerImpl::run() 2,597,817 ( 0.55%) /usr/home/liquid/rust/worktree-benchmarking/compiler/rustc_span/src/lib.rs:>::decode 2,523,502 ( 0.53%) ???:llvm::SelectionDAG::Combine(llvm::CombineLevel, llvm::AAResults*, llvm::CodeGenOpt::Level) 2,486,055 ( 0.53%) ???:llvm::PMTopLevelManager::AUFoldingSetNode::Profile(llvm::FoldingSetNodeID&, llvm::AnalysisUsage const&) 2,344,865 ( 0.50%) ???:llvm::PMDataManager::verifyPreservedAnalysis(llvm::Pass*) 2,329,280 ( 0.49%) ???:(anonymous namespace)::GetCFGOnlyPasses::passEnumerate(llvm::PassInfo const*) [clone .llvm.764396836974782617] 2,139,676 ( 0.45%) ???:llvm::PMDataManager::removeNotPreservedAnalysis(llvm::Pass*) 2,139,033 ( 0.45%) ./elf/do-rel.h:_dl_relocate_object 2,104,137 ( 0.44%) ./malloc/malloc.c:unlink_chunk.constprop.0 1,940,554 ( 0.41%) ???:combineInstructionsOverFunction(llvm::Function&, llvm::InstCombineWorklist&, llvm::AAResults*, llvm::AssumptionCache&, llvm::TargetLibraryInfo&, llvm::TargetTransformInfo&, llvm::DominatorTree&, llvm::OptimizationRemarkEmitter&, llvm::BlockFrequencyInfo*, llvm::ProfileSummaryInfo*, unsigned int, llvm::LoopInfo*) 1,918,938 ( 0.41%) ./string/../sysdeps/x86_64/multiarch/strlen-avx2.S:__strlen_avx2 1,726,965 ( 0.36%) ???:llvm::DomTreeBuilder::SemiNCAInfo >::CalculateFromScratch(llvm::DominatorTreeBase&, llvm::DomTreeBuilder::SemiNCAInfo >::BatchUpdateInfo*) 1,708,461 ( 0.36%) /usr/home/liquid/rust/worktree-benchmarking/compiler/rustc_metadata/src/rmeta/decoder.rs:>::decode 1,707,634 ( 0.36%) ???:llvm::coro::declaresIntrinsics(llvm::Module const&, std::initializer_list) 1,659,077 ( 0.35%) ???:runCVP(llvm::Module&) [clone .llvm.11785992503873176614] 1,628,000 ( 0.34%) /tmp/gcc-build/x86_64-unknown-linux-gnu/libstdc++-v3/libsupc++/../../../../gcc-5.5.0/libstdc++-v3/libsupc++/new_op.cc:operator new(unsigned long) 1,574,214 ( 0.33%) ./string/../sysdeps/x86_64/multiarch/strcmp-avx2.S:__strncmp_avx2 1,526,578 ( 0.32%) ???:llvm::TargetLibraryInfoImpl::getLibFunc(llvm::Function const&, llvm::LibFunc&) const 1,472,624 ( 0.31%) /usr/home/liquid/rust/worktree-benchmarking/compiler/rustc_serialize/src/leb128.rs:::read_u32 1,422,785 ( 0.30%) ???:llvm::LiveVariables::runOnBlock(llvm::MachineBasicBlock*, unsigned int) 1,353,562 ( 0.29%) ???:llvm::detail::PassModel>, llvm::PreservedAnalyses, llvm::AnalysisManager>::run(llvm::Function&, llvm::AnalysisManager&) 1,346,436 ( 0.28%) ???:llvm::X86_MC::initLLVMToSEHAndCVRegMapping(llvm::MCRegisterInfo*) 1,299,748 ( 0.27%) /usr/home/liquid/rust/worktree-benchmarking/compiler/rustc_serialize/src/leb128.rs:::read_u32 1,292,709 ( 0.27%) ???:llvm::AnalysisUsage::pushUnique(llvm::SmallVectorImpl&, void const*) 1,203,527 ( 0.25%) /usr/home/liquid/rust/worktree-benchmarking/library/core/src/slice/iter/macros.rs:, >::decode::{closure#0}>, ::imported_source_files::{closure#3}::{closure#0}> as core::iter::traits::iterator::Iterator>::fold::<(), core::iter::traits::iterator::Iterator::for_each::call as alloc::vec::spec_extend::SpecExtend, >::decode::{closure#0}>, ::imported_source_files::{closure#3}::{closure#0}>>>::spec_extend::{closure#0}>::{closure#0}> 1,198,801 ( 0.25%) /usr/home/liquid/rust/worktree-benchmarking/library/core/src/slice/iter/macros.rs:::new_imported_source_file 1,195,135 ( 0.25%) /usr/home/liquid/rust/worktree-benchmarking/compiler/rustc_span/src/lib.rs:::new_imported_source_file 1,177,842 ( 0.25%) ???:llvm::FoldingSetBase::FindNodeOrInsertPos(llvm::FoldingSetNodeID const&, void*&, llvm::FoldingSetBase::FoldingSetInfo const&) 1,177,702 ( 0.25%) ???:llvm::AttributeList::get(llvm::LLVMContext&, llvm::ArrayRef) 1,173,467 ( 0.25%) /usr/home/liquid/.cargo/registry/src/github.com-1ecc6299db9ec823/rustc-hash-1.1.0/src/lib.rs:<&str as core::hash::Hash>::hash:: 1,147,214 ( 0.24%) /usr/home/liquid/rust/worktree-benchmarking/library/core/src/slice/mod.rs:>::decode 1,145,749 ( 0.24%) ???:llvm::PMTopLevelManager::findAnalysisUsage(llvm::Pass*) 1,034,604 ( 0.22%) ???:llvm::StringMapImpl::LookupBucketFor(llvm::StringRef) 1,027,360 ( 0.22%) ???:llvm::LegacyLegalizerInfo::computeTables() 989,060 ( 0.21%) ./string/../sysdeps/x86_64/multiarch/memset-vec-unaligned-erms.S:__memset_avx2_unaligned_erms 988,224 ( 0.21%) ???:llvm::MD5::final(llvm::MD5::MD5Result&) 942,634 ( 0.20%) ???:llvm::removeUnreachableBlocks(llvm::Function&, llvm::DomTreeUpdater*, llvm::MemorySSAUpdater*) 919,510 ( 0.19%) ???:bool llvm::DenseMapBase*, llvm::DenseMapInfo<(anonymous namespace)::SimpleValue>, llvm::detail::DenseMapPair<(anonymous namespace)::SimpleValue, llvm::ScopedHashTableVal<(anonymous namespace)::SimpleValue, llvm::Value*>*> >, (anonymous namespace)::SimpleValue, llvm::ScopedHashTableVal<(anonymous namespace)::SimpleValue, llvm::Value*>*, llvm::DenseMapInfo<(anonymous namespace)::SimpleValue>, llvm::detail::DenseMapPair<(anonymous namespace)::SimpleValue, llvm::ScopedHashTableVal<(anonymous namespace)::SimpleValue, llvm::Value*>*> >::LookupBucketFor<(anonymous namespace)::SimpleValue>((anonymous namespace)::SimpleValue const&, llvm::detail::DenseMapPair<(anonymous namespace)::SimpleValue, llvm::ScopedHashTableVal<(anonymous namespace)::SimpleValue, llvm::Value*>*> const*&) const 917,163 ( 0.19%) ???:??? 910,501 ( 0.19%) ???:llvm::AttributeList::addAttribute(llvm::LLVMContext&, unsigned int, llvm::Attribute::AttrKind) const 901,273 ( 0.19%) ???:llvm::ValueHandleBase::AddToUseList() 879,912 ( 0.19%) /usr/home/liquid/rust/worktree-benchmarking/compiler/rustc_metadata/src/rmeta/decoder.rs:, >::decode::{closure#0}>, ::imported_source_files::{closure#3}::{closure#0}> as core::iter::traits::iterator::Iterator>::fold::<(), core::iter::traits::iterator::Iterator::for_each::call as alloc::vec::spec_extend::SpecExtend, >::decode::{closure#0}>, ::imported_source_files::{closure#3}::{closure#0}>>>::spec_extend::{closure#0}>::{closure#0}> 866,356 ( 0.18%) ./elf/dl-lookup.c:check_match 858,505 ( 0.18%) ./malloc/malloc.c:realloc 832,356 ( 0.18%) ???:llvm::DenseMapBase, llvm::detail::DenseMapPair >, llvm::MCRegister, int, llvm::DenseMapInfo, llvm::detail::DenseMapPair >::moveFromOldBuckets(llvm::detail::DenseMapPair*, llvm::detail::DenseMapPair*) 797,836 ( 0.17%) /usr/home/liquid/rust/worktree-benchmarking/compiler/rustc_span/src/lib.rs:, >::decode::{closure#0}>, ::imported_source_files::{closure#3}::{closure#0}> as core::iter::traits::iterator::Iterator>::fold::<(), core::iter::traits::iterator::Iterator::for_each::call as alloc::vec::spec_extend::SpecExtend, >::decode::{closure#0}>, ::imported_source_files::{closure#3}::{closure#0}>>>::spec_extend::{closure#0}>::{closure#0}> 796,059 ( 0.17%) /usr/home/liquid/.cargo/registry/src/github.com-1ecc6299db9ec823/hashbrown-0.12.0/src/raw/mod.rs:)>>::remove_entry::>::{closure#0}> 763,661 ( 0.16%) /usr/home/liquid/rust/worktree-benchmarking/compiler/rustc_serialize/src/leb128.rs:::read_usize 763,344 ( 0.16%) ???:llvm::DomTreeBuilder::SemiNCAInfo >::CalculateFromScratch(llvm::DominatorTreeBase&, llvm::DomTreeBuilder::SemiNCAInfo >::BatchUpdateInfo*) 751,540 ( 0.16%) ???:(anonymous namespace)::MachineCopyPropagation::runOnMachineFunction(llvm::MachineFunction&) 745,044 ( 0.16%) ???:llvm::AttributeSetNode::get(llvm::LLVMContext&, llvm::AttrBuilder const&) 734,983 ( 0.16%) ???:llvm::LiveVariables::HandleRegMask(llvm::MachineOperand const&) 709,733 ( 0.15%) ???:llvm::SHA1::hashBlock() 707,233 ( 0.15%) ???:llvm::DataLayout::getAlignment(llvm::Type*, bool) const 705,073 ( 0.15%) ???:llvm::PMTopLevelManager::findAnalysisPass(void const*) 700,861 ( 0.15%) /usr/home/liquid/.cargo/registry/src/github.com-1ecc6299db9ec823/hashbrown-0.12.0/src/raw/mod.rs:, (), core::hash::BuildHasherDefault>>::from_hash::>::{closure#0}> 698,572 ( 0.15%) ???:llvm::SelectionDAGISel::SelectCodeCommon(llvm::SDNode*, unsigned char const*, unsigned int) 681,873 ( 0.14%) /usr/home/liquid/rust/worktree-benchmarking/compiler/rustc_serialize/src/leb128.rs:::read_usize 660,036 ( 0.14%) ???:llvm::BitstreamCursor::ReadAbbrevRecord() 659,054 ( 0.14%) /usr/home/liquid/.cargo/registry/src/github.com-1ecc6299db9ec823/hashbrown-0.12.0/src/raw/mod.rs:>::insert::>::{closure#0}> 655,233 ( 0.14%) ???:llvm::InstCombinerImpl::visitCallInst(llvm::CallInst&) 617,673 ( 0.13%) ???:computeKnownBits(llvm::Value const*, llvm::KnownBits&, unsigned int, (anonymous namespace)::Query const&) [clone .llvm.15619146473165121143] 612,521 ( 0.13%) ???:llvm::InlineFunction(llvm::CallBase&, llvm::InlineFunctionInfo&, llvm::AAResults*, bool, llvm::Function*) 610,385 ( 0.13%) ./malloc/malloc.c:_int_realloc 603,793 ( 0.13%) ???:llvm::DomTreeBuilder::SemiNCAInfo >::FindRoots(llvm::DominatorTreeBase const&, llvm::DomTreeBuilder::SemiNCAInfo >::BatchUpdateInfo*) 595,685 ( 0.13%) ???:(anonymous namespace)::EarlyCSE::run() [clone .llvm.7062997131228810369] 592,403 ( 0.13%) ???:llvm::X86TargetMachine::getTargetTransformInfo(llvm::Function const&) 592,378 ( 0.13%) ???:llvm::SimplifyInstruction(llvm::Instruction*, llvm::SimplifyQuery const&, llvm::OptimizationRemarkEmitter*) 586,087 ( 0.12%) ???:(anonymous namespace)::BitcodeReader::parseModule(unsigned long, bool, llvm::function_ref, std::allocator > > (llvm::StringRef)>) 580,462 ( 0.12%) ???:llvm::simplifyCFG(llvm::BasicBlock*, llvm::TargetTransformInfo const&, llvm::DomTreeUpdater*, llvm::SimplifyCFGOptions const&, llvm::ArrayRef) 569,834 ( 0.12%) ???:(anonymous namespace)::DeadMachineInstructionElim::eliminateDeadMI(llvm::MachineFunction&) 569,012 ( 0.12%) /usr/home/liquid/rust/worktree-benchmarking/compiler/rustc_span/src/lib.rs:>::decode 565,533 ( 0.12%) ???:(anonymous namespace)::LazyValueInfoImpl::solve() [clone .llvm.4316243980339171764] 560,140 ( 0.12%) ???:llvm::MVT::getVectorMinNumElements() const 555,468 ( 0.12%) /usr/home/liquid/rust/worktree-benchmarking/compiler/rustc_span/src/symbol.rs:::intern 553,813 ( 0.12%) ???:char const** std::__find_if >(char const**, char const**, __gnu_cxx::__ops::_Iter_equals_val, std::random_access_iterator_tag) 550,995 ( 0.12%) ./string/../sysdeps/x86_64/multiarch/memmove-vec-unaligned-erms.S:__memcpy_sse2_unaligned_erms 535,757 ( 0.11%) ???:llvm::FoldingSetNodeID::AddPointer(void const*) 534,388 ( 0.11%) ???:getFeatures(llvm::StringRef, llvm::StringRef, llvm::StringRef, llvm::ArrayRef, llvm::ArrayRef) 532,036 ( 0.11%) ???:(anonymous namespace)::Verifier::verifyFunctionAttrs(llvm::FunctionType*, llvm::AttributeList, llvm::Value const*, bool) 531,469 ( 0.11%) ./string/../sysdeps/x86_64/strcmp.S:strcmp 529,140 ( 0.11%) ???:SimplifyICmpInst(unsigned int, llvm::Value*, llvm::Value*, llvm::SimplifyQuery const&, unsigned int) [clone .llvm.1619516508949622737] 528,592 ( 0.11%) ???:llvm::FindFunctionBackedges(llvm::Function const&, llvm::SmallVectorImpl >&) 516,557 ( 0.11%) /usr/home/liquid/rust/worktree-benchmarking/compiler/rustc_resolve/src/build_reduced_graph.rs:::build_reduced_graph_for_external_crate_res 496,440 ( 0.10%) /usr/home/liquid/.cargo/registry/src/github.com-1ecc6299db9ec823/hashbrown-0.12.0/src/raw/mod.rs:::intern 494,951 ( 0.10%) ???:llvm::LivePhysRegs::stepBackward(llvm::MachineInstr const&) 494,205 ( 0.10%) ./nptl/pthread_rwlock_common.c:pthread_rwlock_unlock 486,330 ( 0.10%) ???:isKnownNonZero(llvm::Value const*, llvm::APInt const&, unsigned int, (anonymous namespace)::Query const&) [clone .llvm.15619146473165121143] 485,527 ( 0.10%) /usr/home/liquid/rust/worktree-benchmarking/compiler/rustc_resolve/src/imports.rs:::try_define 478,676 ( 0.10%) ???:initialize(llvm::TargetLibraryInfoImpl&, llvm::Triple const&, llvm::ArrayRef) [clone .llvm.8179168538081115032] 477,069 ( 0.10%) ???:void llvm::BitstreamWriter::EmitRecord >(unsigned int, llvm::SmallVector const&, unsigned int) 474,202 ( 0.10%) ???:llvm::MachineFunctionPass::runOnFunction(llvm::Function&) -------------------------------------------------------------------------------- -- Auto-annotated source: /usr/home/liquid/rust/worktree-benchmarking/library/alloc/src/vec/mod.rs -------------------------------------------------------------------------------- Ir -- line 417 ---------------------------------------- . /// # #![allow(unused_mut)] . /// let mut vec: Vec = Vec::new(); . /// ``` . #[inline] . #[rustc_const_stable(feature = "const_vec_new", since = "1.39.0")] . #[stable(feature = "rust1", since = "1.0.0")] . #[must_use] . pub const fn new() -> Self { 208,716 ( 0.04%) Vec { buf: RawVec::NEW, len: 0 } 39 ( 0.00%) } . . /// Constructs a new, empty `Vec` with the specified capacity. . /// . /// The vector will be able to hold exactly `capacity` elements without . /// reallocating. If `capacity` is 0, the vector will not allocate. . /// . /// It is important to note that although the returned vector has the . /// *capacity* specified, the vector will have a zero *length*. For an -- line 434 ---------------------------------------- -- line 601 ---------------------------------------- . /// vec.push(11); . /// assert_eq!(vec.len(), 11); . /// assert!(vec.capacity() >= 11); . /// ``` . #[cfg(not(no_global_oom_handling))] . #[inline] . #[unstable(feature = "allocator_api", issue = "32838")] . pub fn with_capacity_in(capacity: usize, alloc: A) -> Self { 110,846 ( 0.02%) Vec { buf: RawVec::with_capacity_in(capacity, alloc), len: 0 } . } . . /// Creates a `Vec` directly from the raw components of another vector. . /// . /// # Safety . /// . /// This is highly unsafe, due to the number of invariants that aren't . /// checked: -- line 617 ---------------------------------------- -- line 677 ---------------------------------------- . /// // Put everything back together into a Vec . /// let rebuilt = Vec::from_raw_parts_in(p, len, cap, alloc.clone()); . /// assert_eq!(rebuilt, [4, 5, 6]); . /// } . /// ``` . #[inline] . #[unstable(feature = "allocator_api", issue = "32838")] . pub unsafe fn from_raw_parts_in(ptr: *mut T, length: usize, capacity: usize, alloc: A) -> Self { 3,677 ( 0.00%) unsafe { Vec { buf: RawVec::from_raw_parts_in(ptr, capacity, alloc), len: length } } . } . . /// Decomposes a `Vec` into its raw components. . /// . /// Returns the raw pointer to the underlying data, the length of . /// the vector (in elements), and the allocated capacity of the . /// data (in elements). These are the same arguments in the same . /// order as the arguments to [`from_raw_parts`]. -- line 693 ---------------------------------------- -- line 778 ---------------------------------------- . /// . /// ``` . /// let vec: Vec = Vec::with_capacity(10); . /// assert_eq!(vec.capacity(), 10); . /// ``` . #[inline] . #[stable(feature = "rust1", since = "1.0.0")] . pub fn capacity(&self) -> usize { 28,931 ( 0.01%) self.buf.capacity() . } . . /// Reserves capacity for at least `additional` more elements to be inserted . /// in the given `Vec`. The collection may reserve more space to avoid . /// frequent reallocations. After calling `reserve`, capacity will be . /// greater than or equal to `self.len() + additional`. Does nothing if . /// capacity is already sufficient. . /// -- line 794 ---------------------------------------- -- line 801 ---------------------------------------- . /// ``` . /// let mut vec = vec![1]; . /// vec.reserve(10); . /// assert!(vec.capacity() >= 11); . /// ``` . #[cfg(not(no_global_oom_handling))] . #[stable(feature = "rust1", since = "1.0.0")] . pub fn reserve(&mut self, additional: usize) { 17,281 ( 0.00%) self.buf.reserve(self.len, additional); . } . . /// Reserves the minimum capacity for exactly `additional` more elements to . /// be inserted in the given `Vec`. After calling `reserve_exact`, . /// capacity will be greater than or equal to `self.len() + additional`. . /// Does nothing if the capacity is already sufficient. . /// . /// Note that the allocator may give the collection more space than it -- line 817 ---------------------------------------- -- line 829 ---------------------------------------- . /// ``` . /// let mut vec = vec![1]; . /// vec.reserve_exact(10); . /// assert!(vec.capacity() >= 11); . /// ``` . #[cfg(not(no_global_oom_handling))] . #[stable(feature = "rust1", since = "1.0.0")] . pub fn reserve_exact(&mut self, additional: usize) { 1,104 ( 0.00%) self.buf.reserve_exact(self.len, additional); . } . . /// Tries to reserve capacity for at least `additional` more elements to be inserted . /// in the given `Vec`. The collection may reserve more space to avoid . /// frequent reallocations. After calling `try_reserve`, capacity will be . /// greater than or equal to `self.len() + additional`. Does nothing if . /// capacity is already sufficient. . /// -- line 845 ---------------------------------------- -- line 930 ---------------------------------------- . /// assert!(vec.capacity() >= 3); . /// ``` . #[cfg(not(no_global_oom_handling))] . #[stable(feature = "rust1", since = "1.0.0")] . pub fn shrink_to_fit(&mut self) { . // The capacity is never less than the length, and there's nothing to do when . // they are equal, so we can avoid the panic case in `RawVec::shrink_to_fit` . // by only calling it with a greater capacity. 1,707 ( 0.00%) if self.capacity() > self.len { 652 ( 0.00%) self.buf.shrink_to_fit(self.len); . } . } . . /// Shrinks the capacity of the vector with a lower bound. . /// . /// The capacity will remain at least as large as both the length . /// and the supplied value. . /// -- line 947 ---------------------------------------- -- line 990 ---------------------------------------- . /// let slice = vec.into_boxed_slice(); . /// assert_eq!(slice.into_vec().capacity(), 3); . /// ``` . #[cfg(not(no_global_oom_handling))] . #[stable(feature = "rust1", since = "1.0.0")] . pub fn into_boxed_slice(mut self) -> Box<[T], A> { . unsafe { . self.shrink_to_fit(); 306 ( 0.00%) let me = ManuallyDrop::new(self); . let buf = ptr::read(&me.buf); . let len = me.len(); . buf.into_box(len).assume_init() . } . } . . /// Shortens the vector, keeping the first `len` elements and dropping . /// the rest. -- line 1006 ---------------------------------------- -- line 1040 ---------------------------------------- . /// let mut vec = vec![1, 2, 3]; . /// vec.truncate(0); . /// assert_eq!(vec, []); . /// ``` . /// . /// [`clear`]: Vec::clear . /// [`drain`]: Vec::drain . #[stable(feature = "rust1", since = "1.0.0")] 9,207 ( 0.00%) pub fn truncate(&mut self, len: usize) { . // This is safe because: . // . // * the slice passed to `drop_in_place` is valid; the `len > self.len` . // case avoids creating an invalid slice, and . // * the `len` of the vector is shrunk before calling `drop_in_place`, . // such that no value will be dropped twice in case `drop_in_place` . // were to panic once (if it panics twice, the program aborts). . unsafe { . // Note: It's intentional that this is `>` and not `>=`. . // Changing it to `>=` has negative performance . // implications in some cases. See #78884 for more. 7,749 ( 0.00%) if len > self.len { . return; . } . let remaining_len = self.len - len; . let s = ptr::slice_from_raw_parts_mut(self.as_mut_ptr().add(len), remaining_len); 9,022 ( 0.00%) self.len = len; 254 ( 0.00%) ptr::drop_in_place(s); . } 11,026 ( 0.00%) } . . /// Extracts a slice containing the entire vector. . /// . /// Equivalent to `&s[..]`. . /// . /// # Examples . /// . /// ``` -- line 1076 ---------------------------------------- -- line 1126 ---------------------------------------- . /// ``` . /// . /// [`as_mut_ptr`]: Vec::as_mut_ptr . #[stable(feature = "vec_as_ptr", since = "1.37.0")] . #[inline] . pub fn as_ptr(&self) -> *const T { . // We shadow the slice method of the same name to avoid going through . // `deref`, which creates an intermediate reference. 237,878 ( 0.05%) let ptr = self.buf.ptr(); . unsafe { . assume(!ptr.is_null()); . } . ptr . } . . /// Returns an unsafe mutable pointer to the vector's buffer. . /// -- line 1142 ---------------------------------------- -- line 1162 ---------------------------------------- . /// } . /// assert_eq!(&*x, &[0, 1, 2, 3]); . /// ``` . #[stable(feature = "vec_as_ptr", since = "1.37.0")] . #[inline] . pub fn as_mut_ptr(&mut self) -> *mut T { . // We shadow the slice method of the same name to avoid going through . // `deref_mut`, which creates an intermediate reference. 580,181 ( 0.12%) let ptr = self.buf.ptr(); . unsafe { . assume(!ptr.is_null()); . } . ptr . } . . /// Returns a reference to the underlying allocator. . #[unstable(feature = "allocator_api", issue = "32838")] -- line 1178 ---------------------------------------- -- line 1259 ---------------------------------------- . /// . /// Normally, here, one would use [`clear`] instead to correctly drop . /// the contents and thus not leak memory. . #[inline] . #[stable(feature = "rust1", since = "1.0.0")] . pub unsafe fn set_len(&mut self, new_len: usize) { . debug_assert!(new_len <= self.capacity()); . 48,210 ( 0.01%) self.len = new_len; 666 ( 0.00%) } . . /// Removes an element from the vector and returns it. . /// . /// The removed element is replaced by the last element of the vector. . /// . /// This does not preserve ordering, but is *O*(1). . /// If you need to preserve the element order, use [`remove`] instead. . /// -- line 1276 ---------------------------------------- -- line 1329 ---------------------------------------- . /// let mut vec = vec![1, 2, 3]; . /// vec.insert(1, 4); . /// assert_eq!(vec, [1, 4, 2, 3]); . /// vec.insert(4, 5); . /// assert_eq!(vec, [1, 4, 2, 3, 5]); . /// ``` . #[cfg(not(no_global_oom_handling))] . #[stable(feature = "rust1", since = "1.0.0")] 406 ( 0.00%) pub fn insert(&mut self, index: usize, element: T) { . #[cold] . #[inline(never)] . fn assert_failed(index: usize, len: usize) -> ! { . panic!("insertion index (is {}) should be <= len (is {})", index, len); . } . 69 ( 0.00%) let len = self.len(); 190 ( 0.00%) if index > len { . assert_failed(index, len); . } . . // space for the new element 278 ( 0.00%) if len == self.buf.capacity() { . self.reserve(1); . } . . unsafe { . // infallible . // The spot to put the new value . { . let p = self.as_mut_ptr().add(index); . // Shift everything over to make space. (Duplicating the . // `index`th element into two consecutive places.) 384 ( 0.00%) ptr::copy(p, p.offset(1), len - index); . // Write it in, overwriting the first copy of the `index`th . // element. . ptr::write(p, element); . } 278 ( 0.00%) self.set_len(len + 1); . } 348 ( 0.00%) } . . /// Removes and returns the element at position `index` within the vector, . /// shifting all elements after it to the left. . /// . /// Note: Because this shifts over the remaining elements, it has a . /// worst-case performance of *O*(*n*). If you don't need the order of elements . /// to be preserved, use [`swap_remove`] instead. If you'd like to remove . /// elements from the beginning of the `Vec`, consider using -- line 1376 ---------------------------------------- -- line 1395 ---------------------------------------- . pub fn remove(&mut self, index: usize) -> T { . #[cold] . #[inline(never)] . #[track_caller] . fn assert_failed(index: usize, len: usize) -> ! { . panic!("removal index (is {}) should be < len (is {})", index, len); . } . 46 ( 0.00%) let len = self.len(); 175 ( 0.00%) if index >= len { . assert_failed(index, len); . } . unsafe { . // infallible . let ret; . { . // the place we are taking from. . let ptr = self.as_mut_ptr().add(index); . // copy it out, unsafely having a copy of the value on . // the stack and in the vector at the same time. 144 ( 0.00%) ret = ptr::read(ptr); . . // Shift everything down to fill in that spot. 122 ( 0.00%) ptr::copy(ptr.offset(1), ptr, len - index - 1); . } 46 ( 0.00%) self.set_len(len - 1); . ret . } . } . . /// Retains only the elements specified by the predicate. . /// . /// In other words, remove all elements `e` such that `f(&e)` returns `false`. . /// This method operates in place, visiting each element exactly once in the -- line 1428 ---------------------------------------- -- line 1442 ---------------------------------------- . /// ``` . /// let mut vec = vec![1, 2, 3, 4, 5]; . /// let keep = [false, true, true, false, true]; . /// let mut iter = keep.iter(); . /// vec.retain(|_| *iter.next().unwrap()); . /// assert_eq!(vec, [2, 3, 5]); . /// ``` . #[stable(feature = "rust1", since = "1.0.0")] 16,494 ( 0.00%) pub fn retain(&mut self, mut f: F) . where . F: FnMut(&T) -> bool, . { 577 ( 0.00%) self.retain_mut(|elem| f(elem)); 16,176 ( 0.00%) } . . /// Retains only the elements specified by the predicate, passing a mutable reference to it. . /// . /// In other words, remove all elements `e` such that `f(&mut e)` returns `false`. . /// This method operates in place, visiting each element exactly once in the . /// original order, and preserves the order of the retained elements. . /// . /// # Examples -- line 1463 ---------------------------------------- -- line 1474 ---------------------------------------- . /// }); . /// assert_eq!(vec, [2, 3, 4]); . /// ``` . #[unstable(feature = "vec_retain_mut", issue = "90829")] . pub fn retain_mut(&mut self, mut f: F) . where . F: FnMut(&mut T) -> bool, . { 2,080 ( 0.00%) let original_len = self.len(); . // Avoid double drop if the drop guard is not executed, . // since we may make some holes during the process. . unsafe { self.set_len(0) }; . . // Vec: [Kept, Kept, Hole, Hole, Hole, Hole, Unchecked, Unchecked] . // |<- processed len ->| ^- next to check . // |<- deleted cnt ->| . // |<- original_len ->| -- line 1490 ---------------------------------------- -- line 1499 ---------------------------------------- . v: &'a mut Vec, . processed_len: usize, . deleted_cnt: usize, . original_len: usize, . } . . impl Drop for BackshiftOnDrop<'_, T, A> { . fn drop(&mut self) { 1,946 ( 0.00%) if self.deleted_cnt > 0 { . // SAFETY: Trailing unchecked items must be valid since we never touch them. . unsafe { . ptr::copy( . self.v.as_ptr().add(self.processed_len), 1,182 ( 0.00%) self.v.as_mut_ptr().add(self.processed_len - self.deleted_cnt), . self.original_len - self.processed_len, . ); . } . } . // SAFETY: After filling holes, all items are in contiguous memory. . unsafe { 2 ( 0.00%) self.v.set_len(self.original_len - self.deleted_cnt); . } . } . } . . let mut g = BackshiftOnDrop { v: self, processed_len: 0, deleted_cnt: 0, original_len }; . . fn process_loop( . original_len: usize, . f: &mut F, . g: &mut BackshiftOnDrop<'_, T, A>, . ) where . F: FnMut(&mut T) -> bool, . { 25,299 ( 0.01%) while g.processed_len != original_len { . // SAFETY: Unchecked element must be valid. . let cur = unsafe { &mut *g.v.as_mut_ptr().add(g.processed_len) }; 5,685 ( 0.00%) if !f(cur) { . // Advance early to avoid double drop if `drop_in_place` panicked. 154 ( 0.00%) g.processed_len += 1; 20 ( 0.00%) g.deleted_cnt += 1; . // SAFETY: We never touch this element again after dropped. 8 ( 0.00%) unsafe { ptr::drop_in_place(cur) }; . // We already advanced the counter. . if DELETED { . continue; . } else { . break; . } . } . if DELETED { . // SAFETY: `deleted_cnt` > 0, so the hole slot must not overlap with current element. . // We use copy for move, and never touch this element again. . unsafe { 699 ( 0.00%) let hole_slot = g.v.as_mut_ptr().add(g.processed_len - g.deleted_cnt); . ptr::copy_nonoverlapping(cur, hole_slot, 1); . } . } 623 ( 0.00%) g.processed_len += 1; . } . } . . // Stage 1: Nothing was deleted. . process_loop::(original_len, &mut f, &mut g); . . // Stage 2: Some elements were deleted. . process_loop::(original_len, &mut f, &mut g); -- line 1565 ---------------------------------------- -- line 1606 ---------------------------------------- . /// ``` . /// let mut vec = vec!["foo", "bar", "Bar", "baz", "bar"]; . /// . /// vec.dedup_by(|a, b| a.eq_ignore_ascii_case(b)); . /// . /// assert_eq!(vec, ["foo", "bar", "baz", "bar"]); . /// ``` . #[stable(feature = "dedup_by", since = "1.16.0")] 1,225 ( 0.00%) pub fn dedup_by(&mut self, mut same_bucket: F) . where . F: FnMut(&mut T, &mut T) -> bool, . { 222 ( 0.00%) let len = self.len(); 546 ( 0.00%) if len <= 1 { . return; . } . . /* INVARIANT: vec.len() > read >= write > write-1 >= 0 */ . struct FillGapOnDrop<'a, T, A: core::alloc::Allocator> { . /* Offset of the element we want to check if it is duplicate */ . read: usize, . -- line 1627 ---------------------------------------- -- line 1670 ---------------------------------------- . let ptr = gap.vec.as_mut_ptr(); . . /* Drop items while going through Vec, it should be more efficient than . * doing slice partition_dedup + truncate */ . . /* SAFETY: Because of the invariant, read_ptr, prev_ptr and write_ptr . * are always in-bounds and read_ptr never aliases prev_ptr */ . unsafe { 875 ( 0.00%) while gap.read < len { . let read_ptr = ptr.add(gap.read); . let prev_ptr = ptr.add(gap.write.wrapping_sub(1)); . 46 ( 0.00%) if same_bucket(&mut *read_ptr, &mut *prev_ptr) { . // Increase `gap.read` now since the drop may panic. . gap.read += 1; . /* We have found duplicate, drop it in-place */ . ptr::drop_in_place(read_ptr); . } else { . let write_ptr = ptr.add(gap.write); . . /* Because `read_ptr` can be equal to `write_ptr`, we either . * have to use `copy` or conditional `copy_nonoverlapping`. . * Looks like the first option is faster. */ . ptr::copy(read_ptr, write_ptr, 1); . . /* We have filled that place, so go further */ 616 ( 0.00%) gap.write += 1; . gap.read += 1; . } . } . . /* Technically we could let `gap` clean up with its Drop, but . * when `same_bucket` is guaranteed to not panic, this bloats a little . * the codegen, so we just do it manually */ . gap.vec.set_len(gap.write); . mem::forget(gap); . } 1,400 ( 0.00%) } . . /// Appends an element to the back of a collection. . /// . /// # Panics . /// . /// Panics if the new capacity exceeds `isize::MAX` bytes. . /// . /// # Examples -- line 1715 ---------------------------------------- -- line 1717 ---------------------------------------- . /// ``` . /// let mut vec = vec![1, 2]; . /// vec.push(3); . /// assert_eq!(vec, [1, 2, 3]); . /// ``` . #[cfg(not(no_global_oom_handling))] . #[inline] . #[stable(feature = "rust1", since = "1.0.0")] 50 ( 0.00%) pub fn push(&mut self, value: T) { . // This will panic or abort if we would allocate > isize::MAX bytes . // or if the length increment would overflow for zero-sized types. 1,488,712 ( 0.31%) if self.len == self.buf.capacity() { 40,802 ( 0.01%) self.buf.reserve_for_push(self.len); . } . unsafe { 16,793 ( 0.00%) let end = self.as_mut_ptr().add(self.len); . ptr::write(end, value); 1,591,180 ( 0.34%) self.len += 1; . } 40 ( 0.00%) } . . /// Removes the last element from a vector and returns it, or [`None`] if it . /// is empty. . /// . /// If you'd like to pop the first element, consider using . /// [`VecDeque::pop_front`] instead. . /// . /// [`VecDeque::pop_front`]: crate::collections::VecDeque::pop_front -- line 1744 ---------------------------------------- -- line 1748 ---------------------------------------- . /// ``` . /// let mut vec = vec![1, 2, 3]; . /// assert_eq!(vec.pop(), Some(3)); . /// assert_eq!(vec, [1, 2]); . /// ``` . #[inline] . #[stable(feature = "rust1", since = "1.0.0")] . pub fn pop(&mut self) -> Option { 57,106 ( 0.01%) if self.len == 0 { 15 ( 0.00%) None . } else { . unsafe { 37,634 ( 0.01%) self.len -= 1; 7,758 ( 0.00%) Some(ptr::read(self.as_ptr().add(self.len()))) . } . } . } . . /// Moves all the elements of `other` into `Self`, leaving `other` empty. . /// . /// # Panics . /// -- line 1769 ---------------------------------------- -- line 1776 ---------------------------------------- . /// let mut vec2 = vec![4, 5, 6]; . /// vec.append(&mut vec2); . /// assert_eq!(vec, [1, 2, 3, 4, 5, 6]); . /// assert_eq!(vec2, []); . /// ``` . #[cfg(not(no_global_oom_handling))] . #[inline] . #[stable(feature = "append", since = "1.4.0")] 28 ( 0.00%) pub fn append(&mut self, other: &mut Self) { . unsafe { . self.append_elements(other.as_slice() as _); . other.set_len(0); . } 24 ( 0.00%) } . . /// Appends elements to `Self` from other buffer. . #[cfg(not(no_global_oom_handling))] . #[inline] . unsafe fn append_elements(&mut self, other: *const [T]) { . let count = unsafe { (*other).len() }; . self.reserve(count); 2,762 ( 0.00%) let len = self.len(); . unsafe { ptr::copy_nonoverlapping(other as *const T, self.as_mut_ptr().add(len), count) }; 23,939 ( 0.01%) self.len += count; . } . . /// Creates a draining iterator that removes the specified range in the vector . /// and yields the removed items. . /// . /// When the iterator **is** dropped, all elements in the range are removed . /// from the vector, even if the iterator was not fully consumed. If the . /// iterator **is not** dropped (with [`mem::forget`] for example), it is -- line 1807 ---------------------------------------- -- line 1834 ---------------------------------------- . // When the Drain is first created, it shortens the length of . // the source vector to make sure no uninitialized or moved-from elements . // are accessible at all if the Drain's destructor never gets to run. . // . // Drain will ptr::read out the values to remove. . // When finished, remaining tail of the vec is copied back to cover . // the hole, and the vector length is restored to the new length. . // 1,000 ( 0.00%) let len = self.len(); . let Range { start, end } = slice::range(range, ..len); . . unsafe { . // set self.vec length's to start, to be safe in case Drain is leaked . self.set_len(start); . // Use the borrow in the IterMut to indicate borrowing behavior of the . // whole Drain iterator (like &mut T). 143 ( 0.00%) let range_slice = slice::from_raw_parts_mut(self.as_mut_ptr().add(start), end - start); 305 ( 0.00%) Drain { . tail_start: end, 128 ( 0.00%) tail_len: len - end, . iter: range_slice.iter(), . vec: NonNull::from(self), . } . } . } . . /// Clears the vector, removing all values. . /// -- line 1861 ---------------------------------------- -- line 1869 ---------------------------------------- . /// . /// v.clear(); . /// . /// assert!(v.is_empty()); . /// ``` . #[inline] . #[stable(feature = "rust1", since = "1.0.0")] . pub fn clear(&mut self) { 15,635 ( 0.00%) self.truncate(0) . } . . /// Returns the number of elements in the vector, also referred to . /// as its 'length'. . /// . /// # Examples . /// . /// ``` -- line 1885 ---------------------------------------- -- line 1900 ---------------------------------------- . /// let mut v = Vec::new(); . /// assert!(v.is_empty()); . /// . /// v.push(1); . /// assert!(!v.is_empty()); . /// ``` . #[stable(feature = "rust1", since = "1.0.0")] . pub fn is_empty(&self) -> bool { 20,625 ( 0.00%) self.len() == 0 . } . . /// Splits the collection into two at the given index. . /// . /// Returns a newly allocated vector containing the elements in the range . /// `[at, len)`. After the call, the original vector will be left containing . /// the elements `[0, at)` with its previous capacity unchanged. . /// -- line 1916 ---------------------------------------- -- line 1935 ---------------------------------------- . A: Clone, . { . #[cold] . #[inline(never)] . fn assert_failed(at: usize, len: usize) -> ! { . panic!("`at` split index (is {}) should be <= len (is {})", at, len); . } . 60 ( 0.00%) if at > self.len() { . assert_failed(at, self.len()); . } . 48 ( 0.00%) if at == 0 { . // the new vector can take over the original buffer and avoid the copy . return mem::replace( . self, . Vec::with_capacity_in(self.capacity(), self.allocator().clone()), . ); . } . . let other_len = self.len - at; -- line 1955 ---------------------------------------- -- line 1988 ---------------------------------------- . /// . /// let mut vec = vec![]; . /// let mut p = 1; . /// vec.resize_with(4, || { p *= 2; p }); . /// assert_eq!(vec, [2, 4, 8, 16]); . /// ``` . #[cfg(not(no_global_oom_handling))] . #[stable(feature = "vec_resize_with", since = "1.33.0")] 1,163 ( 0.00%) pub fn resize_with(&mut self, new_len: usize, f: F) . where . F: FnMut() -> T, . { 155 ( 0.00%) let len = self.len(); 310 ( 0.00%) if new_len > len { 1,180 ( 0.00%) self.extend_with(new_len - len, ExtendFunc(f)); . } else { . self.truncate(new_len); . } 1,008 ( 0.00%) } . . /// Consumes and leaks the `Vec`, returning a mutable reference to the contents, . /// `&'a mut [T]`. Note that the type `T` must outlive the chosen lifetime . /// `'a`. If the type has only static references, or none at all, then this . /// may be chosen to be `'static`. . /// . /// As of Rust 1.57, this method does not reallocate or shrink the `Vec`, . /// so the leaked allocation may include unused capacity that is not part -- line 2014 ---------------------------------------- -- line 2070 ---------------------------------------- . #[stable(feature = "vec_spare_capacity", since = "1.60.0")] . #[inline] . pub fn spare_capacity_mut(&mut self) -> &mut [MaybeUninit] { . // Note: . // This method is not implemented in terms of `split_at_spare_mut`, . // to prevent invalidation of pointers to the buffer. . unsafe { . slice::from_raw_parts_mut( 2 ( 0.00%) self.as_mut_ptr().add(self.len) as *mut MaybeUninit, 38 ( 0.00%) self.buf.capacity() - self.len, . ) . } . } . . /// Returns vector content as a slice of `T`, along with the remaining spare . /// capacity of the vector as a slice of `MaybeUninit`. . /// . /// The returned spare capacity slice can be used to fill the vector with data -- line 2087 ---------------------------------------- -- line 2189 ---------------------------------------- . /// assert_eq!(vec, ["hello", "world", "world"]); . /// . /// let mut vec = vec![1, 2, 3, 4]; . /// vec.resize(2, 0); . /// assert_eq!(vec, [1, 2]); . /// ``` . #[cfg(not(no_global_oom_handling))] . #[stable(feature = "vec_resize", since = "1.5.0")] 1,272 ( 0.00%) pub fn resize(&mut self, new_len: usize, value: T) { 128 ( 0.00%) let len = self.len(); . 1,000 ( 0.00%) if new_len > len { 527 ( 0.00%) self.extend_with(new_len - len, ExtendElement(value)) . } else { . self.truncate(new_len); . } 304 ( 0.00%) } . . /// Clones and appends all elements in a slice to the `Vec`. . /// . /// Iterates over the slice `other`, clones each element, and then appends . /// it to this `Vec`. The `other` slice is traversed in-order. . /// . /// Note that this function is same as [`extend`] except that it is . /// specialized to work with slices instead. If and when Rust gets -- line 2213 ---------------------------------------- -- line 2271 ---------------------------------------- . trait ExtendWith { . fn next(&mut self) -> T; . fn last(self) -> T; . } . . struct ExtendElement(T); . impl ExtendWith for ExtendElement { . fn next(&mut self) -> T { 32 ( 0.00%) self.0.clone() . } . fn last(self) -> T { . self.0 . } . } . . struct ExtendFunc(F); . impl T> ExtendWith for ExtendFunc { -- line 2287 ---------------------------------------- -- line 2291 ---------------------------------------- . fn last(mut self) -> T { . (self.0)() . } . } . . impl Vec { . #[cfg(not(no_global_oom_handling))] . /// Extend the vector by `n` values, using the given generator. 13,967 ( 0.00%) fn extend_with>(&mut self, n: usize, mut value: E) { . self.reserve(n); . . unsafe { 428 ( 0.00%) let mut ptr = self.as_mut_ptr().add(self.len()); . // Use SetLenOnDrop to work around bug where compiler . // might not realize the store through `ptr` through self.set_len() . // don't alias. . let mut local_len = SetLenOnDrop::new(&mut self.len); . . // Write all elements except the last one . for _ in 1..n { . ptr::write(ptr, value.next()); . ptr = ptr.offset(1); . // Increment the length in every step in case next() panics . local_len.increment_len(1); . } . 18,834 ( 0.00%) if n > 0 { . // We can write the last element directly without cloning needlessly . ptr::write(ptr, value.last()); . local_len.increment_len(1); . } . . // len set by scope guard . } 11,123 ( 0.00%) } . } . . impl Vec { . /// Removes consecutive repeated elements in the vector according to the . /// [`PartialEq`] trait implementation. . /// . /// If the vector is sorted, this removes all duplicates. . /// -- line 2333 ---------------------------------------- -- line 2338 ---------------------------------------- . /// . /// vec.dedup(); . /// . /// assert_eq!(vec, [1, 2, 3, 2]); . /// ``` . #[stable(feature = "rust1", since = "1.0.0")] . #[inline] . pub fn dedup(&mut self) { 176 ( 0.00%) self.dedup_by(|a, b| a == b) . } . } . . //////////////////////////////////////////////////////////////////////////////// . // Internal methods and functions . //////////////////////////////////////////////////////////////////////////////// . . #[doc(hidden)] . #[cfg(not(no_global_oom_handling))] . #[stable(feature = "rust1", since = "1.0.0")] 7,665 ( 0.00%) pub fn from_elem(elem: T, n: usize) -> Vec { 21,162 ( 0.00%) ::from_elem(elem, n, Global) 7,685 ( 0.00%) } . . #[doc(hidden)] . #[cfg(not(no_global_oom_handling))] . #[unstable(feature = "allocator_api", issue = "32838")] . pub fn from_elem_in(elem: T, n: usize, alloc: A) -> Vec { . ::from_elem(elem, n, alloc) . } . -- line 2367 ---------------------------------------- -- line 2424 ---------------------------------------- . // Common trait implementations for Vec . //////////////////////////////////////////////////////////////////////////////// . . #[stable(feature = "rust1", since = "1.0.0")] . impl ops::Deref for Vec { . type Target = [T]; . . fn deref(&self) -> &[T] { 344,875 ( 0.07%) unsafe { slice::from_raw_parts(self.as_ptr(), self.len) } 7 ( 0.00%) } . } . . #[stable(feature = "rust1", since = "1.0.0")] . impl ops::DerefMut for Vec { . fn deref_mut(&mut self) -> &mut [T] { 70,564 ( 0.01%) unsafe { slice::from_raw_parts_mut(self.as_mut_ptr(), self.len) } . } . } . . #[cfg(not(no_global_oom_handling))] . trait SpecCloneFrom { . fn clone_from(this: &mut Self, other: &Self); . } . -- line 2447 ---------------------------------------- -- line 2468 ---------------------------------------- . this.extend_from_slice(other); . } . } . . #[cfg(not(no_global_oom_handling))] . #[stable(feature = "rust1", since = "1.0.0")] . impl Clone for Vec { . #[cfg(not(test))] 9,211 ( 0.00%) fn clone(&self) -> Self { . let alloc = self.allocator().clone(); 2 ( 0.00%) <[T]>::to_vec_in(&**self, alloc) 11,397 ( 0.00%) } . . // HACK(japaric): with cfg(test) the inherent `[T]::to_vec` method, which is . // required for this method definition, is not available. Instead use the . // `slice::to_vec` function which is only available with cfg(test) . // NB see the slice::hack module in slice.rs for more information . #[cfg(test)] . fn clone(&self) -> Self { . let alloc = self.allocator().clone(); -- line 2487 ---------------------------------------- -- line 2518 ---------------------------------------- . message = "vector indices are of type `usize` or ranges of `usize`", . label = "vector indices are of type `usize` or ranges of `usize`" . )] . impl, A: Allocator> Index for Vec { . type Output = I::Output; . . #[inline] . fn index(&self, index: I) -> &Self::Output { 6,579 ( 0.00%) Index::index(&**self, index) . } . } . . #[stable(feature = "rust1", since = "1.0.0")] . #[rustc_on_unimplemented( . message = "vector indices are of type `usize` or ranges of `usize`", . label = "vector indices are of type `usize` or ranges of `usize`" . )] -- line 2534 ---------------------------------------- -- line 2539 ---------------------------------------- . } . } . . #[cfg(not(no_global_oom_handling))] . #[stable(feature = "rust1", since = "1.0.0")] . impl FromIterator for Vec { . #[inline] . fn from_iter>(iter: I) -> Vec { 108,406 ( 0.02%) >::from_iter(iter.into_iter()) . } . } . . #[stable(feature = "rust1", since = "1.0.0")] . impl IntoIterator for Vec { . type Item = T; . type IntoIter = IntoIter; . -- line 2555 ---------------------------------------- -- line 2564 ---------------------------------------- . /// for s in v.into_iter() { . /// // s has type String, not &String . /// println!("{}", s); . /// } . /// ``` . #[inline] . fn into_iter(self) -> IntoIter { . unsafe { 32,963 ( 0.01%) let mut me = ManuallyDrop::new(self); . let alloc = ptr::read(me.allocator()); . let begin = me.as_mut_ptr(); . let end = if mem::size_of::() == 0 { . arith_offset(begin as *const i8, me.len() as isize) as *const T . } else { . begin.add(me.len()) as *const T . }; . let cap = me.buf.capacity(); 13,725 ( 0.00%) IntoIter { . buf: NonNull::new_unchecked(begin), . phantom: PhantomData, . cap, . alloc, . ptr: begin, . end, . } . } -- line 2589 ---------------------------------------- -- line 2591 ---------------------------------------- . } . . #[stable(feature = "rust1", since = "1.0.0")] . impl<'a, T, A: Allocator> IntoIterator for &'a Vec { . type Item = &'a T; . type IntoIter = slice::Iter<'a, T>; . . fn into_iter(self) -> slice::Iter<'a, T> { 5 ( 0.00%) self.iter() . } . } . . #[stable(feature = "rust1", since = "1.0.0")] . impl<'a, T, A: Allocator> IntoIterator for &'a mut Vec { . type Item = &'a mut T; . type IntoIter = slice::IterMut<'a, T>; . . fn into_iter(self) -> slice::IterMut<'a, T> { 1,268 ( 0.00%) self.iter_mut() . } . } . . #[cfg(not(no_global_oom_handling))] . #[stable(feature = "rust1", since = "1.0.0")] . impl Extend for Vec { . #[inline] 80 ( 0.00%) fn extend>(&mut self, iter: I) { 32,215 ( 0.01%) >::spec_extend(self, iter.into_iter()) 80 ( 0.00%) } . . #[inline] . fn extend_one(&mut self, item: T) { 16 ( 0.00%) self.push(item); . } . . #[inline] . fn extend_reserve(&mut self, additional: usize) { . self.reserve(additional); . } . } . -- line 2631 ---------------------------------------- -- line 2636 ---------------------------------------- . fn extend_desugared>(&mut self, mut iterator: I) { . // This is the case for a general iterator. . // . // This function should be the moral equivalent of: . // . // for item in iterator { . // self.push(item); . // } 17,342 ( 0.00%) while let Some(element) = iterator.next() { 976 ( 0.00%) let len = self.len(); 21,198 ( 0.00%) if len == self.capacity() { 175 ( 0.00%) let (lower, _) = iterator.size_hint(); . self.reserve(lower.saturating_add(1)); . } . unsafe { . ptr::write(self.as_mut_ptr().add(len), element); . // Since next() executes user code which can panic we have to bump the length . // after each step. . // NB can't overflow since we would have had to alloc the address space 19,152 ( 0.00%) self.set_len(len + 1); . } . } 78 ( 0.00%) } . . /// Creates a splicing iterator that replaces the specified range in the vector . /// with the given `replace_with` iterator and yields the removed items. . /// `replace_with` does not need to be the same length as `range`. . /// . /// `range` is removed even if the iterator is not consumed until the end. . /// . /// It is unspecified how many elements are removed from the vector -- line 2666 ---------------------------------------- -- line 2693 ---------------------------------------- . #[cfg(not(no_global_oom_handling))] . #[inline] . #[stable(feature = "vec_splice", since = "1.21.0")] . pub fn splice(&mut self, range: R, replace_with: I) -> Splice<'_, I::IntoIter, A> . where . R: RangeBounds, . I: IntoIterator, . { 165 ( 0.00%) Splice { drain: self.drain(range), replace_with: replace_with.into_iter() } . } . . /// Creates an iterator which uses a closure to determine if an element should be removed. . /// . /// If the closure returns true, then the element is removed and yielded. . /// If the closure returns false, the element will remain in the vector and will not be yielded . /// by the iterator. . /// -- line 2709 ---------------------------------------- -- line 2745 ---------------------------------------- . /// assert_eq!(evens, vec![2, 4, 6, 8, 14]); . /// assert_eq!(odds, vec![1, 3, 5, 9, 11, 13, 15]); . /// ``` . #[unstable(feature = "drain_filter", reason = "recently added", issue = "43244")] . pub fn drain_filter(&mut self, filter: F) -> DrainFilter<'_, T, F, A> . where . F: FnMut(&mut T) -> bool, . { 406 ( 0.00%) let old_len = self.len(); . . // Guard against us getting leaked (leak amplification) . unsafe { . self.set_len(0); . } . 200 ( 0.00%) DrainFilter { vec: self, idx: 0, del: 0, old_len, pred: filter, panic_flag: false } . } . } . . /// Extend implementation that copies elements out of references before pushing them onto the Vec. . /// . /// This implementation is specialized for slice iterators, where it uses [`copy_from_slice`] to . /// append the entire slice at once. . /// -- line 2768 ---------------------------------------- -- line 2803 ---------------------------------------- . #[inline] . fn cmp(&self, other: &Self) -> Ordering { . Ord::cmp(&**self, &**other) . } . } . . #[stable(feature = "rust1", since = "1.0.0")] . unsafe impl<#[may_dangle] T, A: Allocator> Drop for Vec { 54,834 ( 0.01%) fn drop(&mut self) { . unsafe { . // use drop for [T] . // use a raw slice to refer to the elements of the vector as weakest necessary type; . // could avoid questions of validity in certain cases 26,670 ( 0.01%) ptr::drop_in_place(ptr::slice_from_raw_parts_mut(self.as_mut_ptr(), self.len)) . } . // RawVec handles deallocation 64,130 ( 0.01%) } . } . . #[stable(feature = "rust1", since = "1.0.0")] . #[rustc_const_unstable(feature = "const_default_impls", issue = "87864")] . impl const Default for Vec { . /// Creates an empty `Vec`. . fn default() -> Vec { . Vec::new() -- line 2827 ---------------------------------------- -- line 2976 ---------------------------------------- . /// newly-allocated buffer with exactly the right capacity. . /// . /// # Examples . /// . /// ``` . /// assert_eq!(Box::from(vec![1, 2, 3]), vec![1, 2, 3].into_boxed_slice()); . /// ``` . fn from(v: Vec) -> Self { 4 ( 0.00%) v.into_boxed_slice() . } . } . . #[cfg(not(no_global_oom_handling))] . #[stable(feature = "rust1", since = "1.0.0")] . impl From<&str> for Vec { . /// Allocate a `Vec` and fill it with a UTF-8 string. . /// -- line 2992 ---------------------------------------- 122,819 ( 0.03%) -------------------------------------------------------------------------------- -- Auto-annotated source: /usr/home/liquid/.cargo/registry/src/github.com-1ecc6299db9ec823/hashbrown-0.12.0/src/raw/mod.rs -------------------------------------------------------------------------------- Ir -- line 111 ---------------------------------------- . const EMPTY: u8 = 0b1111_1111; . . /// Control byte value for a deleted bucket. . const DELETED: u8 = 0b1000_0000; . . /// Checks whether a control byte represents a full bucket (top bit is clear). . #[inline] . fn is_full(ctrl: u8) -> bool { 332,031 ( 0.07%) ctrl & 0x80 == 0 . } . . /// Checks whether a control byte represents a special value (top bit is set). . #[inline] . fn is_special(ctrl: u8) -> bool { . ctrl & 0x80 != 0 . } . . /// Checks whether a special control value is EMPTY (just check 1 bit). . #[inline] . fn special_is_empty(ctrl: u8) -> bool { . debug_assert!(is_special(ctrl)); 27,520 ( 0.01%) ctrl & 0x01 != 0 . } . . /// Primary hash function, used to select the initial bucket to probe from. . #[inline] . #[allow(clippy::cast_possible_truncation)] . fn h1(hash: u64) -> usize { . // On 32-bit platforms we simply ignore the higher hash bits. . hash as usize -- line 140 ---------------------------------------- -- line 143 ---------------------------------------- . /// Secondary hash function, saved in the low 7 bits of the control byte. . #[inline] . #[allow(clippy::cast_possible_truncation)] . fn h2(hash: u64) -> u8 { . // Grab the top 7 bits of the hash. While the hash is normally a full 64-bit . // value, some hash functions (such as FxHash) produce a usize result . // instead, which means that the top 32 bits are 0 on 32-bit platforms. . let hash_len = usize::min(mem::size_of::(), mem::size_of::()); 2,368,133 ( 0.50%) let top7 = hash >> (hash_len * 8 - 7); . (top7 & 0x7f) as u8 // truncation . } . . /// Probe sequence based on triangular numbers, which is guaranteed (since our . /// table size is a power of two) to visit every group of elements exactly once. . /// . /// A triangular probe has us jump by 1 more group every time. So first we . /// jump by 1 group (meaning we just continue our linear scan), then 2 groups -- line 159 ---------------------------------------- -- line 170 ---------------------------------------- . #[inline] . fn move_next(&mut self, bucket_mask: usize) { . // We should have found an empty bucket by now and ended the probe. . debug_assert!( . self.stride <= bucket_mask, . "Went past end of probe sequence" . ); . 37,738 ( 0.01%) self.stride += Group::WIDTH; 37,738 ( 0.01%) self.pos += self.stride; 30,784 ( 0.01%) self.pos &= bucket_mask; . } . } . . /// Returns the number of buckets needed to hold the given number of items, . /// taking the maximum load factor into account. . /// . /// Returns `None` if an overflow occurs. . // Workaround for emscripten bug emscripten-core/emscripten-fastcomp#258 . #[cfg_attr(target_os = "emscripten", inline(never))] . #[cfg_attr(not(target_os = "emscripten"), inline)] . fn capacity_to_buckets(cap: usize) -> Option { . debug_assert_ne!(cap, 0); . . // For small tables we require at least 1 empty bucket so that lookups are . // guaranteed to terminate if an element doesn't exist in the table. 11,398 ( 0.00%) if cap < 8 { . // We don't bother with a table size of 2 buckets since that can only . // hold a single element. Instead we skip directly to a 4 bucket table . // which can hold 3 elements. 22,635 ( 0.00%) return Some(if cap < 4 { 4 } else { 8 }); . } . . // Otherwise require 1/8 buckets to be empty (87.5% load) . // . // Be careful when modifying this, calculate_layout relies on the . // overflow check here. 7,032 ( 0.00%) let adjusted_cap = cap.checked_mul(8)? / 7; . . // Any overflows will have been caught by the checked_mul. Also, any . // rounding errors from the division above will be cleaned up by . // next_power_of_two (which can't overflow because of the previous division). . Some(adjusted_cap.next_power_of_two()) . } . . /// Returns the maximum effective capacity for the given bucket mask, taking . /// the maximum load factor into account. . #[inline] . fn bucket_mask_to_capacity(bucket_mask: usize) -> usize { 31,873 ( 0.01%) if bucket_mask < 8 { . // For tables with 1/2/4/8 buckets, we always reserve one empty slot. . // Keep in mind that the bucket mask is one less than the bucket count. . bucket_mask . } else { . // For larger tables we reserve 12.5% of the slots as empty. 9,594 ( 0.00%) ((bucket_mask + 1) / 8) * 7 . } . } . . /// Helper which allows the max calculation for ctrl_align to be statically computed for each T . /// while keeping the rest of `calculate_layout_for` independent of `T` . #[derive(Copy, Clone)] . struct TableLayout { . size: usize, -- line 233 ---------------------------------------- -- line 246 ---------------------------------------- . . #[inline] . fn calculate_layout_for(self, buckets: usize) -> Option<(Layout, usize)> { . debug_assert!(buckets.is_power_of_two()); . . let TableLayout { size, ctrl_align } = self; . // Manual layout calculation since Layout methods are not yet stable. . let ctrl_offset = 20,764 ( 0.00%) size.checked_mul(buckets)?.checked_add(ctrl_align - 1)? & !(ctrl_align - 1); 29,660 ( 0.01%) let len = ctrl_offset.checked_add(buckets + Group::WIDTH)?; . . Some(( . unsafe { Layout::from_size_align_unchecked(len, ctrl_align) }, . ctrl_offset, . )) . } . } . -- line 263 ---------------------------------------- -- line 337 ---------------------------------------- . } . } . #[cfg_attr(feature = "inline-more", inline)] . pub unsafe fn drop(&self) { . self.as_ptr().drop_in_place(); . } . #[inline] . pub unsafe fn read(&self) -> T { 14 ( 0.00%) self.as_ptr().read() . } . #[inline] . pub unsafe fn write(&self, val: T) { . self.as_ptr().write(val); . } . #[inline] . pub unsafe fn as_ref<'a>(&self) -> &'a T { . &*self.as_ptr() -- line 353 ---------------------------------------- -- line 422 ---------------------------------------- . /// Creates a new empty hash table without allocating any memory, using the . /// given allocator. . /// . /// In effect this returns a table with exactly 1 bucket. However we can . /// leave the data pointer dangling since that bucket is never written to . /// due to our load factor forcing us to always have at least 1 free bucket. . #[inline] . pub fn new_in(alloc: A) -> Self { 21 ( 0.00%) Self { . table: RawTableInner::new_in(alloc), . marker: PhantomData, . } . } . . /// Allocates a new hash table with the given number of buckets. . /// . /// The control bytes are left uninitialized. -- line 438 ---------------------------------------- -- line 440 ---------------------------------------- . unsafe fn new_uninitialized( . alloc: A, . buckets: usize, . fallibility: Fallibility, . ) -> Result { . debug_assert!(buckets.is_power_of_two()); . . Ok(Self { 20 ( 0.00%) table: RawTableInner::new_uninitialized( . alloc, . TableLayout::new::(), . buckets, . fallibility, . )?, . marker: PhantomData, . }) . } -- line 456 ---------------------------------------- -- line 458 ---------------------------------------- . /// Attempts to allocate a new hash table with at least enough capacity . /// for inserting the given number of elements without reallocating. . fn fallible_with_capacity( . alloc: A, . capacity: usize, . fallibility: Fallibility, . ) -> Result { . Ok(Self { 1,806 ( 0.00%) table: RawTableInner::fallible_with_capacity( . alloc, . TableLayout::new::(), . capacity, . fallibility, . )?, . marker: PhantomData, . }) . } -- line 474 ---------------------------------------- -- line 527 ---------------------------------------- . debug_assert_ne!(self.table.bucket_mask, 0); . debug_assert!(index < self.buckets()); . Bucket::from_base_index(self.data_end(), index) . } . . /// Erases an element from the table without dropping it. . #[cfg_attr(feature = "inline-more", inline)] . #[deprecated(since = "0.8.1", note = "use erase or remove instead")] 891 ( 0.00%) pub unsafe fn erase_no_drop(&mut self, item: &Bucket) { 891 ( 0.00%) let index = self.bucket_index(item); . self.table.erase(index); 1,782 ( 0.00%) } . . /// Erases an element from the table, dropping it in place. . #[cfg_attr(feature = "inline-more", inline)] . #[allow(clippy::needless_pass_by_value)] . #[allow(deprecated)] . pub unsafe fn erase(&mut self, item: Bucket) { . // Erase the element from the table first since drop might panic. 298 ( 0.00%) self.erase_no_drop(&item); . item.drop(); . } . . /// Finds and erases an element from the table, dropping it in place. . /// Returns true if an element was found. . #[cfg(feature = "raw")] . #[cfg_attr(feature = "inline-more", inline)] . pub fn erase_entry(&mut self, hash: u64, eq: impl FnMut(&T) -> bool) -> bool { -- line 554 ---------------------------------------- -- line 563 ---------------------------------------- . } . } . . /// Removes an element from the table, returning it. . #[cfg_attr(feature = "inline-more", inline)] . #[allow(clippy::needless_pass_by_value)] . #[allow(deprecated)] . pub unsafe fn remove(&mut self, item: Bucket) -> T { 1,484 ( 0.00%) self.erase_no_drop(&item); 4 ( 0.00%) item.read() . } . . /// Finds and removes an element from the table, returning it. . #[cfg_attr(feature = "inline-more", inline)] 88,557 ( 0.02%) pub fn remove_entry(&mut self, hash: u64, eq: impl FnMut(&T) -> bool) -> Option { . // Avoid `Option::map` because it bloats LLVM IR. 168 ( 0.00%) match self.find(hash, eq) { 277 ( 0.00%) Some(bucket) => Some(unsafe { self.remove(bucket) }), 5,833 ( 0.00%) None => None, . } 131,383 ( 0.03%) } . . /// Marks all table buckets as empty without dropping their contents. . #[cfg_attr(feature = "inline-more", inline)] . pub fn clear_no_drop(&mut self) { . self.table.clear_no_drop(); . } . . /// Removes all elements from the table without freeing the backing memory. . #[cfg_attr(feature = "inline-more", inline)] . pub fn clear(&mut self) { . // Ensure that the table is reset even if one of the drops panic . let mut self_ = guard(self, |self_| self_.clear_no_drop()); . unsafe { 1 ( 0.00%) self_.drop_elements(); . } . } . 7 ( 0.00%) unsafe fn drop_elements(&mut self) { 1,474 ( 0.00%) if mem::needs_drop::() && !self.is_empty() { . for item in self.iter() { . item.drop(); . } . } 8 ( 0.00%) } . . /// Shrinks the table to fit `max(self.len(), min_size)` elements. . #[cfg_attr(feature = "inline-more", inline)] . pub fn shrink_to(&mut self, min_size: usize, hasher: impl Fn(&T) -> u64) { . // Calculate the minimal number of elements that we need to reserve . // space for. . let min_size = usize::max(self.table.items, min_size); . if min_size == 0 { -- line 615 ---------------------------------------- -- line 642 ---------------------------------------- . } . } . } . . /// Ensures that at least `additional` items can be inserted into the table . /// without reallocation. . #[cfg_attr(feature = "inline-more", inline)] . pub fn reserve(&mut self, additional: usize, hasher: impl Fn(&T) -> u64) { 59,465 ( 0.01%) if additional > self.table.growth_left { . // Avoid `Result::unwrap_or_else` because it bloats LLVM IR. 18,744 ( 0.00%) if self . .reserve_rehash(additional, hasher, Fallibility::Infallible) . .is_err() . { . unsafe { hint::unreachable_unchecked() } . } . } . } . -- line 660 ---------------------------------------- -- line 671 ---------------------------------------- . } else { . Ok(()) . } . } . . /// Out-of-line slow path for `reserve` and `try_reserve`. . #[cold] . #[inline(never)] 45,430 ( 0.01%) fn reserve_rehash( . &mut self, . additional: usize, . hasher: impl Fn(&T) -> u64, . fallibility: Fallibility, . ) -> Result<(), TryReserveError> { . unsafe { . self.table.reserve_rehash_inner( . additional, -- line 687 ---------------------------------------- -- line 690 ---------------------------------------- . TableLayout::new::(), . if mem::needs_drop::() { . Some(mem::transmute(ptr::drop_in_place:: as unsafe fn(*mut T))) . } else { . None . }, . ) . } 30,544 ( 0.01%) } . . /// Allocates a new table of a different size and moves the contents of the . /// current table into it. . fn resize( . &mut self, . capacity: usize, . hasher: impl Fn(&T) -> u64, . fallibility: Fallibility, -- line 706 ---------------------------------------- -- line 714 ---------------------------------------- . ) . } . } . . /// Inserts a new element into the table, and returns its raw bucket. . /// . /// This does not check if the given element already exists in the table. . #[cfg_attr(feature = "inline-more", inline)] 549,714 ( 0.12%) pub fn insert(&mut self, hash: u64, value: T, hasher: impl Fn(&T) -> u64) -> Bucket { . unsafe { . let mut index = self.table.find_insert_slot(hash); . . // We can avoid growing the table once we have reached our load . // factor if we are replacing a tombstone. This works since the . // number of EMPTY slots does not change in this case. 484 ( 0.00%) let old_ctrl = *self.table.ctrl(index); 312,313 ( 0.07%) if unlikely(self.table.growth_left == 0 && special_is_empty(old_ctrl)) { . self.reserve(1, hasher); . index = self.table.find_insert_slot(hash); . } . . self.table.record_item_insert_at(index, old_ctrl, hash); . . let bucket = self.bucket(index); 4 ( 0.00%) bucket.write(value); . bucket . } 403,153 ( 0.09%) } . . /// Attempts to insert a new element without growing the table and return its raw bucket. . /// . /// Returns an `Err` containing the given element if inserting it would require growing the . /// table. . /// . /// This does not check if the given element already exists in the table. . #[cfg(feature = "raw")] -- line 749 ---------------------------------------- -- line 760 ---------------------------------------- . } . } . } . . /// Inserts a new element into the table, and returns a mutable reference to it. . /// . /// This does not check if the given element already exists in the table. . #[cfg_attr(feature = "inline-more", inline)] 95,392 ( 0.02%) pub fn insert_entry(&mut self, hash: u64, value: T, hasher: impl Fn(&T) -> u64) -> &mut T { 97 ( 0.00%) unsafe { self.insert(hash, value, hasher).as_mut() } 71,544 ( 0.02%) } . . /// Inserts a new element into the table, without growing the table. . /// . /// There must be enough space in the table to insert the new element. . /// . /// This does not check if the given element already exists in the table. . #[cfg_attr(feature = "inline-more", inline)] . #[cfg(any(feature = "raw", feature = "rustc-internal-api"))] 66 ( 0.00%) pub unsafe fn insert_no_grow(&mut self, hash: u64, value: T) -> Bucket { 83,009 ( 0.02%) let (index, old_ctrl) = self.table.prepare_insert_slot(hash); 560 ( 0.00%) let bucket = self.table.bucket(index); . . // If we are replacing a DELETED entry then we don't need to update . // the load counter. 168,905 ( 0.04%) self.table.growth_left -= special_is_empty(old_ctrl) as usize; . . bucket.write(value); 132,739 ( 0.03%) self.table.items += 1; . bucket 127 ( 0.00%) } . . /// Temporary removes a bucket, applying the given function to the removed . /// element and optionally put back the returned value in the same bucket. . /// . /// Returns `true` if the bucket still contains an element . /// . /// This does not check if the given bucket is actually occupied. . #[cfg_attr(feature = "inline-more", inline)] -- line 798 ---------------------------------------- -- line 813 ---------------------------------------- . true . } else { . false . } . } . . /// Searches for an element in the table. . #[inline] 3,069 ( 0.00%) pub fn find(&self, hash: u64, mut eq: impl FnMut(&T) -> bool) -> Option> { 1,803 ( 0.00%) let result = self.table.find_inner(hash, &mut |index| unsafe { 2,419 ( 0.00%) eq(self.bucket(index).as_ref()) 43 ( 0.00%) }); . . // Avoid `Option::map` because it bloats LLVM IR. . match result { 178 ( 0.00%) Some(index) => Some(unsafe { self.bucket(index) }), . None => None, . } 3,488 ( 0.00%) } . . /// Gets a reference to an element in the table. . #[inline] . pub fn get(&self, hash: u64, eq: impl FnMut(&T) -> bool) -> Option<&T> { . // Avoid `Option::map` because it bloats LLVM IR. 1,357 ( 0.00%) match self.find(hash, eq) { . Some(bucket) => Some(unsafe { bucket.as_ref() }), . None => None, . } . } . . /// Gets a mutable reference to an element in the table. . #[inline] 245 ( 0.00%) pub fn get_mut(&mut self, hash: u64, eq: impl FnMut(&T) -> bool) -> Option<&mut T> { . // Avoid `Option::map` because it bloats LLVM IR. 1,378 ( 0.00%) match self.find(hash, eq) { . Some(bucket) => Some(unsafe { bucket.as_mut() }), . None => None, . } 280 ( 0.00%) } . . /// Attempts to get mutable references to `N` entries in the table at once. . /// . /// Returns an array of length `N` with the results of each query. . /// . /// At most one mutable reference will be returned to any entry. `None` will be returned if any . /// of the hashes are duplicates. `None` will be returned if the hash is not found. . /// -- line 859 ---------------------------------------- -- line 920 ---------------------------------------- . #[inline] . pub fn len(&self) -> usize { . self.table.items . } . . /// Returns `true` if the table contains no elements. . #[inline] . pub fn is_empty(&self) -> bool { 70,292 ( 0.01%) self.len() == 0 . } . . /// Returns the number of buckets in the table. . #[inline] . pub fn buckets(&self) -> usize { . self.table.bucket_mask + 1 . } . -- line 936 ---------------------------------------- -- line 938 ---------------------------------------- . /// the caller to ensure that the `RawTable` outlives the `RawIter`. . /// Because we cannot make the `next` method unsafe on the `RawIter` . /// struct, we have to make the `iter` method unsafe. . #[inline] . pub unsafe fn iter(&self) -> RawIter { . let data = Bucket::from_base_index(self.data_end(), 0); . RawIter { . iter: RawIterRange::new(self.table.ctrl.as_ptr(), data, self.table.buckets()), 6,759 ( 0.00%) items: self.table.items, . } . } . . /// Returns an iterator over occupied buckets that could match a given hash. . /// . /// `RawTable` only stores 7 bits of the hash value, so this iterator may . /// return items that have a hash value different than the one provided. You . /// should always validate the returned values before using them. -- line 954 ---------------------------------------- -- line 995 ---------------------------------------- . /// Iteration starts at the provided iterator's current location. . /// . /// It is up to the caller to ensure that the iterator is valid for this . /// `RawTable` and covers all items that remain in the table. . pub unsafe fn into_iter_from(self, iter: RawIter) -> RawIntoIter { . debug_assert_eq!(iter.len(), self.len()); . . let alloc = self.table.alloc.clone(); 760 ( 0.00%) let allocation = self.into_allocation(); 570 ( 0.00%) RawIntoIter { 950 ( 0.00%) iter, . allocation, . marker: PhantomData, . alloc, . } . } . . /// Converts the table into a raw allocation. The contents of the table . /// should be dropped using a `RawIter` before freeing the allocation. . #[cfg_attr(feature = "inline-more", inline)] . pub(crate) fn into_allocation(self) -> Option<(NonNull, Layout)> { 244 ( 0.00%) let alloc = if self.table.is_empty_singleton() { . None . } else { . // Avoid `Option::unwrap_or_else` because it bloats LLVM IR. . let (layout, ctrl_offset) = match calculate_layout::(self.table.buckets()) { . Some(lco) => lco, . None => unsafe { hint::unreachable_unchecked() }, . }; . Some(( 104 ( 0.00%) unsafe { NonNull::new_unchecked(self.table.ctrl.as_ptr().sub(ctrl_offset)) }, . layout, . )) . }; . mem::forget(self); . alloc . } . } . -- line 1033 ---------------------------------------- -- line 1042 ---------------------------------------- . T: Sync, . A: Sync, . { . } . . impl RawTableInner { . #[inline] . const fn new_in(alloc: A) -> Self { 54,086 ( 0.01%) Self { . // Be careful to cast the entire slice to a raw pointer. . ctrl: unsafe { NonNull::new_unchecked(Group::static_empty() as *const _ as *mut u8) }, . bucket_mask: 0, . items: 0, . growth_left: 0, . alloc, . } . } . } . . impl RawTableInner { . #[cfg_attr(feature = "inline-more", inline)] 29,066 ( 0.01%) unsafe fn new_uninitialized( . alloc: A, . table_layout: TableLayout, . buckets: usize, . fallibility: Fallibility, . ) -> Result { . debug_assert!(buckets.is_power_of_two()); . . // Avoid `Option::ok_or_else` because it bloats LLVM IR. -- line 1071 ---------------------------------------- -- line 1078 ---------------------------------------- . // exceed `isize::MAX`. We can skip this check on 64-bit systems since . // such allocations will never succeed anyways. . // . // This mirrors what Vec does in the standard library. . if mem::size_of::() < 8 && layout.size() > isize::MAX as usize { . return Err(fallibility.capacity_overflow()); . } . 10,118 ( 0.00%) let ptr: NonNull = match do_alloc(&alloc, layout) { . Ok(block) => block.cast(), . Err(_) => return Err(fallibility.alloc_err(layout)), . }; . . let ctrl = NonNull::new_unchecked(ptr.as_ptr().add(ctrl_offset)); 17,574 ( 0.00%) Ok(Self { . ctrl, 8,632 ( 0.00%) bucket_mask: buckets - 1, . items: 0, . growth_left: bucket_mask_to_capacity(buckets - 1), . alloc, . }) 21,330 ( 0.00%) } . . #[inline] 6,422 ( 0.00%) fn fallible_with_capacity( . alloc: A, . table_layout: TableLayout, . capacity: usize, . fallibility: Fallibility, . ) -> Result { 1,616 ( 0.00%) if capacity == 0 { 1,662 ( 0.00%) Ok(Self::new_in(alloc)) . } else { . unsafe { . let buckets = . capacity_to_buckets(capacity).ok_or_else(|| fallibility.capacity_overflow())?; . 13,425 ( 0.00%) let result = Self::new_uninitialized(alloc, table_layout, buckets, fallibility)?; . result.ctrl(0).write_bytes(EMPTY, result.num_ctrl_bytes()); . 2,837 ( 0.00%) Ok(result) . } . } 6,422 ( 0.00%) } . . /// Searches for an empty or deleted bucket which is suitable for inserting . /// a new element and sets the hash for that slot. . /// . /// There must be at least 1 empty bucket in the table. . #[inline] 27,459 ( 0.01%) unsafe fn prepare_insert_slot(&self, hash: u64) -> (usize, u8) { . let index = self.find_insert_slot(hash); 27,459 ( 0.01%) let old_ctrl = *self.ctrl(index); . self.set_ctrl_h2(index, hash); . (index, old_ctrl) 54,918 ( 0.01%) } . . /// Searches for an empty or deleted bucket which is suitable for inserting . /// a new element. . /// . /// There must be at least 1 empty bucket in the table. . #[inline] . fn find_insert_slot(&self, hash: u64) -> usize { . let mut probe_seq = self.probe_seq(hash); . loop { . unsafe { . let group = Group::load(self.ctrl(probe_seq.pos)); 218,122 ( 0.05%) if let Some(bit) = group.match_empty_or_deleted().lowest_set_bit() { 635,099 ( 0.13%) let result = (probe_seq.pos + bit) & self.bucket_mask; . . // In tables smaller than the group width, trailing control . // bytes outside the range of the table are filled with . // EMPTY entries. These will unfortunately trigger a . // match, but once masked may point to a full bucket that . // is already occupied. We detect this situation here and . // perform a second scan starting at the beginning of the . // table. This second scan is guaranteed to find an empty . // slot (due to the load factor) before hitting the trailing . // control bytes (containing EMPTY). 289,763 ( 0.06%) if unlikely(is_full(*self.ctrl(result))) { . debug_assert!(self.bucket_mask < Group::WIDTH); . debug_assert_ne!(probe_seq.pos, 0); . return Group::load_aligned(self.ctrl(0)) . .match_empty_or_deleted() . .lowest_set_bit_nonzero(); . } . . return result; -- line 1165 ---------------------------------------- -- line 1171 ---------------------------------------- . . /// Searches for an element in the table. This uses dynamic dispatch to reduce the amount of . /// code generated, but it is eliminated by LLVM optimizations. . #[inline] . fn find_inner(&self, hash: u64, eq: &mut dyn FnMut(usize) -> bool) -> Option { . let h2_hash = h2(hash); . let mut probe_seq = self.probe_seq(hash); . 5,060 ( 0.00%) loop { . let group = unsafe { Group::load(self.ctrl(probe_seq.pos)) }; . 336,072 ( 0.07%) for bit in group.match_byte(h2_hash) { 476,438 ( 0.10%) let index = (probe_seq.pos + bit) & self.bucket_mask; . 356,328 ( 0.08%) if likely(eq(index)) { . return Some(index); . } . } . 168,382 ( 0.04%) if likely(group.match_empty().any_bit_set()) { . return None; . } . . probe_seq.move_next(self.bucket_mask); . } . } . . #[allow(clippy::mut_mut)] -- line 1198 ---------------------------------------- -- line 1225 ---------------------------------------- . Bucket::from_base_index(self.data_end(), index) . } . . #[inline] . unsafe fn bucket_ptr(&self, index: usize, size_of: usize) -> *mut u8 { . debug_assert_ne!(self.bucket_mask, 0); . debug_assert!(index < self.buckets()); . let base: *mut u8 = self.data_end().as_ptr(); 813,248 ( 0.17%) base.sub((index + 1) * size_of) . } . . #[inline] . unsafe fn data_end(&self) -> NonNull { . NonNull::new_unchecked(self.ctrl.as_ptr().cast()) . } . . /// Returns an iterator-like object for a probe sequence on the table. . /// . /// This iterator never terminates, but is guaranteed to visit each bucket . /// group exactly once. The loop using `probe_seq` must terminate upon . /// reaching a group containing an empty bucket. . #[inline] . fn probe_seq(&self, hash: u64) -> ProbeSeq { . ProbeSeq { 2,562,539 ( 0.54%) pos: h1(hash) & self.bucket_mask, . stride: 0, . } . } . . /// Returns the index of a bucket for which a value must be inserted if there is enough rooom . /// in the table, otherwise returns error . #[cfg(feature = "raw")] . #[inline] -- line 1257 ---------------------------------------- -- line 1263 ---------------------------------------- . } else { . self.record_item_insert_at(index, old_ctrl, hash); . Ok(index) . } . } . . #[inline] . unsafe fn record_item_insert_at(&mut self, index: usize, old_ctrl: u8, hash: u64) { 390,399 ( 0.08%) self.growth_left -= special_is_empty(old_ctrl) as usize; . self.set_ctrl_h2(index, hash); 312,308 ( 0.07%) self.items += 1; . } . . #[inline] . fn is_in_same_group(&self, i: usize, new_i: usize, hash: u64) -> bool { . let probe_seq_pos = self.probe_seq(hash).pos; . let probe_index = . |pos: usize| (pos.wrapping_sub(probe_seq_pos) & self.bucket_mask) / Group::WIDTH; . probe_index(i) == probe_index(new_i) -- line 1281 ---------------------------------------- -- line 1312 ---------------------------------------- . // replicate the buckets at the end of the trailing group. For example . // with 2 buckets and a group size of 4, the control bytes will look . // like this: . // . // Real | Replicated . // --------------------------------------------- . // | [A] | [B] | [EMPTY] | [EMPTY] | [A] | [B] | . // --------------------------------------------- 621,973 ( 0.13%) let index2 = ((index.wrapping_sub(Group::WIDTH)) & self.bucket_mask) + Group::WIDTH; . 207,253 ( 0.04%) *self.ctrl(index) = ctrl; 207,259 ( 0.04%) *self.ctrl(index2) = ctrl; . } . . /// Returns a pointer to a control byte. . #[inline] . unsafe fn ctrl(&self, index: usize) -> *mut u8 { . debug_assert!(index < self.num_ctrl_bytes()); . self.ctrl.as_ptr().add(index) . } . . #[inline] . fn buckets(&self) -> usize { 24,647 ( 0.01%) self.bucket_mask + 1 . } . . #[inline] . fn num_ctrl_bytes(&self) -> usize { 16,754 ( 0.00%) self.bucket_mask + 1 + Group::WIDTH . } . . #[inline] . fn is_empty_singleton(&self) -> bool { 118,578 ( 0.03%) self.bucket_mask == 0 . } . . #[allow(clippy::mut_mut)] . #[inline] . unsafe fn prepare_resize( . &self, . table_layout: TableLayout, . capacity: usize, . fallibility: Fallibility, . ) -> Result, TryReserveError> { . debug_assert!(self.items <= capacity); . . // Allocate and initialize the new table. 1,059 ( 0.00%) let mut new_table = RawTableInner::fallible_with_capacity( . self.alloc.clone(), . table_layout, . capacity, . fallibility, . )?; 12,650 ( 0.00%) new_table.growth_left -= self.items; . new_table.items = self.items; . . // The hash function may panic, in which case we simply free the new . // table without dropping any elements that may have been copied into . // it. . // . // This guard is also used to free the old table on success, see . // the comment at the bottom of this function. . Ok(guard(new_table, move |self_| { 5,579 ( 0.00%) if !self_.is_empty_singleton() { . self_.free_buckets(table_layout); . } . })) . } . . /// Reserves or rehashes to make room for `additional` more elements. . /// . /// This uses dynamic dispatch to reduce the amount of -- line 1383 ---------------------------------------- -- line 1388 ---------------------------------------- . &mut self, . additional: usize, . hasher: &dyn Fn(&mut Self, usize) -> u64, . fallibility: Fallibility, . layout: TableLayout, . drop: Option, . ) -> Result<(), TryReserveError> { . // Avoid `Option::ok_or_else` because it bloats LLVM IR. 11,158 ( 0.00%) let new_items = match self.items.checked_add(additional) { . Some(new_items) => new_items, . None => return Err(fallibility.capacity_overflow()), . }; 11,158 ( 0.00%) let full_capacity = bucket_mask_to_capacity(self.bucket_mask); 23,261 ( 0.00%) if new_items <= full_capacity / 2 { . // Rehash in-place without re-allocating if we have plenty of spare . // capacity that is locked up due to DELETED entries. . self.rehash_in_place(hasher, layout.size, drop); . Ok(()) . } else { . // Otherwise, conservatively resize to at least the next size up . // to avoid churning deletes into frequent rehashes. . self.resize_inner( 5,579 ( 0.00%) usize::max(new_items, full_capacity + 1), . hasher, . fallibility, . layout, . ) . } . } . . /// Allocates a new table of a different size and moves the contents of the -- line 1418 ---------------------------------------- -- line 1424 ---------------------------------------- . #[inline(always)] . unsafe fn resize_inner( . &mut self, . capacity: usize, . hasher: &dyn Fn(&mut Self, usize) -> u64, . fallibility: Fallibility, . layout: TableLayout, . ) -> Result<(), TryReserveError> { 2,621 ( 0.00%) let mut new_table = self.prepare_resize(layout, capacity, fallibility)?; . . // Copy all elements to the new table. . for i in 0..self.buckets() { 120,345 ( 0.03%) if !is_full(*self.ctrl(i)) { . continue; . } . . // This may panic. . let hash = hasher(self, i); . . // We can use a simpler version of insert() here since: . // - there are no DELETED entries. -- line 1444 ---------------------------------------- -- line 1454 ---------------------------------------- . } . . // We successfully copied all elements without panicking. Now replace . // self with the new table. The old table will have its memory freed but . // the items will not be dropped (since they have been moved into the . // new table). . mem::swap(self, &mut new_table); . 5,579 ( 0.00%) Ok(()) . } . . /// Rehashes the contents of the table in place (i.e. without changing the . /// allocation). . /// . /// If `hasher` panics then some the table's contents may be lost. . /// . /// This uses dynamic dispatch to reduce the amount of -- line 1470 ---------------------------------------- -- line 1554 ---------------------------------------- . #[inline] . unsafe fn free_buckets(&mut self, table_layout: TableLayout) { . // Avoid `Option::unwrap_or_else` because it bloats LLVM IR. . let (layout, ctrl_offset) = match table_layout.calculate_layout_for(self.buckets()) { . Some(lco) => lco, . None => hint::unreachable_unchecked(), . }; . self.alloc.deallocate( 3,705 ( 0.00%) NonNull::new_unchecked(self.ctrl.as_ptr().sub(ctrl_offset)), . layout, . ); . } . . /// Marks all table buckets as empty without dropping their contents. . #[inline] . fn clear_no_drop(&mut self) { 453 ( 0.00%) if !self.is_empty_singleton() { . unsafe { . self.ctrl(0).write_bytes(EMPTY, self.num_ctrl_bytes()); . } . } 570 ( 0.00%) self.items = 0; 453 ( 0.00%) self.growth_left = bucket_mask_to_capacity(self.bucket_mask); . } . . #[inline] . unsafe fn erase(&mut self, index: usize) { . debug_assert!(is_full(*self.ctrl(index))); 22,230 ( 0.00%) let index_before = index.wrapping_sub(Group::WIDTH) & self.bucket_mask; . let empty_before = Group::load(self.ctrl(index_before)).match_empty(); . let empty_after = Group::load(self.ctrl(index)).match_empty(); . . // If we are inside a continuous block of Group::WIDTH full or deleted . // cells then a probe window may have seen a full block when trying to . // insert. We therefore need to keep that block non-empty so that . // lookups will continue searching to the next probe window. . // . // Note that in this context `leading_zeros` refers to the bytes at the . // end of a group, while `trailing_zeros` refers to the bytes at the . // beginning of a group. 88,920 ( 0.02%) let ctrl = if empty_before.leading_zeros() + empty_after.trailing_zeros() >= Group::WIDTH { . DELETED . } else { 111,150 ( 0.02%) self.growth_left += 1; . EMPTY . }; . self.set_ctrl(index, ctrl); 88,920 ( 0.02%) self.items -= 1; . } . } . . impl Clone for RawTable { 72 ( 0.00%) fn clone(&self) -> Self { 11 ( 0.00%) if self.table.is_empty_singleton() { . Self::new_in(self.table.alloc.clone()) . } else { . unsafe { . let mut new_table = ManuallyDrop::new( . // Avoid `Result::ok_or_else` because it bloats LLVM IR. . match Self::new_uninitialized( . self.table.alloc.clone(), . self.table.buckets(), -- line 1615 ---------------------------------------- -- line 1624 ---------------------------------------- . // We need to free the memory allocated for the new table. . new_table.free_buckets(); . }); . . // Return the newly created table. . ManuallyDrop::into_inner(new_table) . } . } 81 ( 0.00%) } . . fn clone_from(&mut self, source: &Self) { . if source.table.is_empty_singleton() { . *self = Self::new_in(self.table.alloc.clone()); . } else { . unsafe { . // First, drop all our elements without clearing the control bytes. . self.drop_elements(); -- line 1640 ---------------------------------------- -- line 1687 ---------------------------------------- . .table . .ctrl(0) . .copy_to_nonoverlapping(self.table.ctrl(0), self.table.num_ctrl_bytes()); . source . .data_start() . .copy_to_nonoverlapping(self.data_start(), self.table.buckets()); . . self.table.items = source.table.items; 8 ( 0.00%) self.table.growth_left = source.table.growth_left; . } . } . . impl RawTable { . /// Common code for clone and clone_from. Assumes `self.buckets() == source.buckets()`. . #[cfg_attr(feature = "inline-more", inline)] . unsafe fn clone_from_impl(&mut self, source: &Self, mut on_panic: impl FnMut(&mut Self)) { . // Copy the control bytes unchanged. We do this in a single pass -- line 1703 ---------------------------------------- -- line 1790 ---------------------------------------- . fn default() -> Self { . Self::new_in(Default::default()) . } . } . . #[cfg(feature = "nightly")] . unsafe impl<#[may_dangle] T, A: Allocator + Clone> Drop for RawTable { . #[cfg_attr(feature = "inline-more", inline)] 66,249 ( 0.01%) fn drop(&mut self) { 52,635 ( 0.01%) if !self.table.is_empty_singleton() { . unsafe { . self.drop_elements(); . self.free_buckets(); . } . } 69,744 ( 0.01%) } . } . #[cfg(not(feature = "nightly"))] . impl Drop for RawTable { . #[cfg_attr(feature = "inline-more", inline)] . fn drop(&mut self) { . if !self.table.is_empty_singleton() { . unsafe { . self.drop_elements(); -- line 1813 ---------------------------------------- -- line 1817 ---------------------------------------- . } . } . . impl IntoIterator for RawTable { . type Item = T; . type IntoIter = RawIntoIter; . . #[cfg_attr(feature = "inline-more", inline)] 760 ( 0.00%) fn into_iter(self) -> RawIntoIter { . unsafe { . let iter = self.iter(); . self.into_iter_from(iter) . } 950 ( 0.00%) } . } . . /// Iterator over a sub-range of a table. Unlike `RawIter` this iterator does . /// not track an item count. . pub(crate) struct RawIterRange { . // Mask of full buckets in the current group. Bits are cleared from this . // mask as each element is processed. . current_group: BitMask, -- line 1838 ---------------------------------------- -- line 1934 ---------------------------------------- . . impl Iterator for RawIterRange { . type Item = Bucket; . . #[cfg_attr(feature = "inline-more", inline)] . fn next(&mut self) -> Option> { . unsafe { . loop { 17,480 ( 0.00%) if let Some(index) = self.current_group.lowest_set_bit() { 2,076 ( 0.00%) self.current_group = self.current_group.remove_lowest_bit(); 3,135 ( 0.00%) return Some(self.data.next_n(index)); . } . 16,338 ( 0.00%) if self.next_ctrl >= self.end { . return None; . } . . // We might read past self.end up to the next group boundary, . // but this is fine because it only occurs on tables smaller . // than the group size where the trailing control bytes are all . // EMPTY. On larger tables self.end is guaranteed to be aligned . // to the group size (since tables are power-of-two sized). 534 ( 0.00%) self.current_group = Group::load_aligned(self.next_ctrl).match_full(); 244 ( 0.00%) self.data = self.data.next_n(Group::WIDTH); 1,220 ( 0.00%) self.next_ctrl = self.next_ctrl.add(Group::WIDTH); . } . } . } . . #[inline] . fn size_hint(&self) -> (usize, Option) { . // We don't have an item count, so just guess based on the range size. . ( -- line 1966 ---------------------------------------- -- line 2102 ---------------------------------------- . } . } else { . // We must have already iterated past the removed item. . } . } . } . . unsafe fn drop_elements(&mut self) { 325 ( 0.00%) if mem::needs_drop::() && self.len() != 0 { . for item in self { . item.drop(); . } . } . } . } . . impl Clone for RawIter { -- line 2118 ---------------------------------------- -- line 2124 ---------------------------------------- . } . } . } . . impl Iterator for RawIter { . type Item = Bucket; . . #[cfg_attr(feature = "inline-more", inline)] 3,054 ( 0.00%) fn next(&mut self) -> Option> { 4,988 ( 0.00%) if let Some(b) = self.iter.next() { 18,924 ( 0.00%) self.items -= 1; . Some(b) . } else { . // We don't check against items == 0 here to allow the . // compiler to optimize away the item count entirely if the . // iterator length is never queried. . debug_assert_eq!(self.items, 0); . None . } 6,108 ( 0.00%) } . . #[inline] . fn size_hint(&self) -> (usize, Option) { . (self.items, Some(self.items)) . } . } . . impl ExactSizeIterator for RawIter {} -- line 2151 ---------------------------------------- -- line 2177 ---------------------------------------- . T: Sync, . A: Sync, . { . } . . #[cfg(feature = "nightly")] . unsafe impl<#[may_dangle] T, A: Allocator + Clone> Drop for RawIntoIter { . #[cfg_attr(feature = "inline-more", inline)] 636 ( 0.00%) fn drop(&mut self) { . unsafe { . // Drop all remaining elements . self.iter.drop_elements(); . . // Free the table 1,115 ( 0.00%) if let Some((ptr, layout)) = self.allocation { . self.alloc.deallocate(ptr, layout); . } . } 198 ( 0.00%) } . } . #[cfg(not(feature = "nightly"))] . impl Drop for RawIntoIter { . #[cfg_attr(feature = "inline-more", inline)] . fn drop(&mut self) { . unsafe { . // Drop all remaining elements . self.iter.drop_elements(); -- line 2203 ---------------------------------------- -- line 2209 ---------------------------------------- . } . } . } . . impl Iterator for RawIntoIter { . type Item = T; . . #[cfg_attr(feature = "inline-more", inline)] 137 ( 0.00%) fn next(&mut self) -> Option { 21 ( 0.00%) unsafe { Some(self.iter.next()?.read()) } 374 ( 0.00%) } . . #[inline] . fn size_hint(&self) -> (usize, Option) { 3 ( 0.00%) self.iter.size_hint() . } . } . . impl ExactSizeIterator for RawIntoIter {} . impl FusedIterator for RawIntoIter {} . . /// Iterator which consumes elements without freeing the table storage. . pub struct RawDrain<'a, T, A: Allocator + Clone = Global> { -- line 2231 ---------------------------------------- -- line 2259 ---------------------------------------- . where . T: Sync, . A: Sync, . { . } . . impl Drop for RawDrain<'_, T, A> { . #[cfg_attr(feature = "inline-more", inline)] 16 ( 0.00%) fn drop(&mut self) { . unsafe { . // Drop all remaining elements. Note that this may panic. . self.iter.drop_elements(); . . // Reset the contents of the table now that all elements have been . // dropped. . self.table.clear_no_drop(); . . // Move the now empty table back to its original location. 2 ( 0.00%) self.orig_table . .as_ptr() . .copy_from_nonoverlapping(&*self.table, 1); . } 16 ( 0.00%) } . } . . impl Iterator for RawDrain<'_, T, A> { . type Item = T; . . #[cfg_attr(feature = "inline-more", inline)] . fn next(&mut self) -> Option { . unsafe { -- line 2289 ---------------------------------------- 1,402,062 ( 0.30%) -------------------------------------------------------------------------------- -- Auto-annotated source: /usr/home/liquid/rust/worktree-benchmarking/compiler/rustc_serialize/src/leb128.rs -------------------------------------------------------------------------------- Ir -- line 17 ---------------------------------------- . #[inline] . pub fn $fn_name( . out: &mut [::std::mem::MaybeUninit; max_leb128_len!($int_ty)], . mut value: $int_ty, . ) -> &[u8] { . let mut i = 0; . . loop { 29,682 ( 0.01%) if value < 0x80 { . unsafe { 14,311 ( 0.00%) *out.get_unchecked_mut(i).as_mut_ptr() = value as u8; . } . 20,416 ( 0.00%) i += 1; . break; . } else { . unsafe { 12,052 ( 0.00%) *out.get_unchecked_mut(i).as_mut_ptr() = ((value & 0x7f) | 0x80) as u8; . } . 5,226 ( 0.00%) value >>= 7; . i += 1; . } . } . . unsafe { ::std::mem::MaybeUninit::slice_assume_init_ref(&out.get_unchecked(..i)) } . } . }; . } -- line 45 ---------------------------------------- -- line 53 ---------------------------------------- . macro_rules! impl_read_unsigned_leb128 { . ($fn_name:ident, $int_ty:ty) => { . #[inline] . pub fn $fn_name(slice: &[u8], position: &mut usize) -> $int_ty { . // The first iteration of this loop is unpeeled. This is a . // performance win because this code is hot and integer values less . // than 128 are very common, typically occurring 50-80% or more of . // the time, even for u64 and u128. 1,726,816 ( 0.36%) let byte = slice[*position]; 863,408 ( 0.18%) *position += 1; 863,408 ( 0.18%) if (byte & 0x80) == 0 { 632 ( 0.00%) return byte as $int_ty; . } 1,292 ( 0.00%) let mut result = (byte & 0x7F) as $int_ty; . let mut shift = 7; . loop { 578,814 ( 0.12%) let byte = slice[*position]; 378,057 ( 0.08%) *position += 1; 385,876 ( 0.08%) if (byte & 0x80) == 0 { 313,849 ( 0.07%) result |= (byte as $int_ty) << shift; . return result; . } else { . result |= ((byte & 0x7F) as $int_ty) << shift; . } . shift += 7; . } . } . }; -- line 80 ---------------------------------------- 617,325 ( 0.13%) -------------------------------------------------------------------------------- -- Auto-annotated source: /usr/home/liquid/rust/worktree-benchmarking/compiler/rustc_span/src/lib.rs -------------------------------------------------------------------------------- Ir -- line 84 ---------------------------------------- . pub struct SessionGlobals { . symbol_interner: symbol::Interner, . span_interner: Lock, . hygiene_data: Lock, . source_map: Lock>>, . } . . impl SessionGlobals { 12 ( 0.00%) pub fn new(edition: Edition) -> SessionGlobals { 50 ( 0.00%) SessionGlobals { 2 ( 0.00%) symbol_interner: symbol::Interner::fresh(), . span_interner: Lock::new(span_encoding::SpanInterner::default()), 6 ( 0.00%) hygiene_data: Lock::new(hygiene::HygieneData::new(edition)), . source_map: Lock::new(None), . } 10 ( 0.00%) } . } . . #[inline] . pub fn create_session_globals_then(edition: Edition, f: impl FnOnce() -> R) -> R { 1 ( 0.00%) assert!( . !SESSION_GLOBALS.is_set(), . "SESSION_GLOBALS should never be overwritten! \ . Use another thread if you need another SessionGlobals" . ); 3 ( 0.00%) let session_globals = SessionGlobals::new(edition); 7 ( 0.00%) SESSION_GLOBALS.set(&session_globals, f) 1 ( 0.00%) } . . #[inline] . pub fn set_session_globals_then(session_globals: &SessionGlobals, f: impl FnOnce() -> R) -> R { . assert!( . !SESSION_GLOBALS.is_set(), . "SESSION_GLOBALS should never be overwritten! \ . Use another thread if you need another SessionGlobals" . ); -- line 119 ---------------------------------------- -- line 120 ---------------------------------------- . SESSION_GLOBALS.set(session_globals, f) . } . . #[inline] . pub fn create_default_session_if_not_set_then(f: F) -> R . where . F: FnOnce(&SessionGlobals) -> R, . { 4 ( 0.00%) create_session_if_not_set_then(edition::DEFAULT_EDITION, f) . } . . #[inline] . pub fn create_session_if_not_set_then(edition: Edition, f: F) -> R . where . F: FnOnce(&SessionGlobals) -> R, . { 1 ( 0.00%) if !SESSION_GLOBALS.is_set() { 3 ( 0.00%) let session_globals = SessionGlobals::new(edition); 11 ( 0.00%) SESSION_GLOBALS.set(&session_globals, || SESSION_GLOBALS.with(f)) 1 ( 0.00%) } else { . SESSION_GLOBALS.with(f) . } . } . . #[inline] . pub fn with_session_globals(f: F) -> R . where . F: FnOnce(&SessionGlobals) -> R, . { 75,614 ( 0.02%) SESSION_GLOBALS.with(f) . } . . #[inline] . pub fn create_default_session_globals_then(f: impl FnOnce() -> R) -> R { . create_session_globals_then(edition::DEFAULT_EDITION, f) . } . . // If this ever becomes non thread-local, `decode_syntax_context` . // and `decode_expn_id` will need to be updated to handle concurrent . // deserialization. . scoped_tls::scoped_thread_local!(static SESSION_GLOBALS: SessionGlobals); . . // FIXME: We should use this enum or something like it to get rid of the . // use of magic `/rust/1.x/...` paths across the board. 31 ( 0.00%) #[derive(Debug, Eq, PartialEq, Clone, Ord, PartialOrd)] 1,250 ( 0.00%) #[derive(Decodable)] . pub enum RealFileName { 2 ( 0.00%) LocalPath(PathBuf), . /// For remapped paths (namely paths into libstd that have been mapped . /// to the appropriate spot on the local host's file system, and local file . /// system paths that have been remapped with `FilePathMapping`), . Remapped { . /// `local_path` is the (host-dependent) local path to the file. This is . /// None if the file was imported from another crate . local_path: Option, . /// `virtual_name` is the stable path rustc will store internally within -- line 175 ---------------------------------------- -- line 179 ---------------------------------------- . } . . impl Hash for RealFileName { . fn hash(&self, state: &mut H) { . // To prevent #70924 from happening again we should only hash the . // remapped (virtualized) path if that exists. This is because . // virtualized paths to sysroot crates (/rust/$hash or /rust/$version) . // remain stable even if the corresponding local_path changes 3,226 ( 0.00%) self.remapped_path_if_available().hash(state) . } . } . . // This is functionally identical to #[derive(Encodable)], with the exception of . // an added assert statement . impl Encodable for RealFileName { . fn encode(&self, encoder: &mut S) -> Result<(), S::Error> { 10 ( 0.00%) encoder.emit_enum(|encoder| match *self { . RealFileName::LocalPath(ref local_path) => { 40 ( 0.00%) encoder.emit_enum_variant("LocalPath", 0, 1, |encoder| { . encoder.emit_enum_variant_arg(true, |encoder| local_path.encode(encoder))?; . Ok(()) . }) . } . . RealFileName::Remapped { ref local_path, ref virtual_name } => encoder . .emit_enum_variant("Remapped", 1, 2, |encoder| { . // For privacy and build reproducibility, we must not embed host-dependant path in artifacts -- line 205 ---------------------------------------- -- line 224 ---------------------------------------- . p.as_ref().map(PathBuf::as_path) . } . } . } . . /// Returns the path suitable for reading from the file system on the local host, . /// if this information exists. . /// Avoid embedding this in build artifacts; see `remapped_path_if_available()` for that. 2 ( 0.00%) pub fn into_local_path(self) -> Option { 2 ( 0.00%) match self { . RealFileName::LocalPath(p) => Some(p), . RealFileName::Remapped { local_path: p, virtual_name: _ } => p, . } 3 ( 0.00%) } . . /// Returns the path suitable for embedding into build artifacts. This would still . /// be a local path if it has not been remapped. A remapped path will not correspond . /// to a valid file system path: see `local_path_if_available()` for something that . /// is more likely to return paths into the local host file system. . pub fn remapped_path_if_available(&self) -> &Path { 2,568 ( 0.00%) match self { . RealFileName::LocalPath(p) . | RealFileName::Remapped { local_path: _, virtual_name: p } => &p, . } 6 ( 0.00%) } . . /// Returns the path suitable for reading from the file system on the local host, . /// if this information exists. Otherwise returns the remapped name. . /// Avoid embedding this in build artifacts; see `remapped_path_if_available()` for that. . pub fn local_path_if_available(&self) -> &Path { 8 ( 0.00%) match self { . RealFileName::LocalPath(path) . | RealFileName::Remapped { local_path: None, virtual_name: path } . | RealFileName::Remapped { local_path: Some(path), virtual_name: _ } => path, . } . } . . pub fn to_string_lossy(&self, display_pref: FileNameDisplayPreference) -> Cow<'_, str> { 10 ( 0.00%) match display_pref { . FileNameDisplayPreference::Local => self.local_path_if_available().to_string_lossy(), . FileNameDisplayPreference::Remapped => { . self.remapped_path_if_available().to_string_lossy() . } . } . } . } . . /// Differentiates between real files and common virtual files. 6,630 ( 0.00%) #[derive(Debug, Eq, PartialEq, Clone, Ord, PartialOrd, Hash)] 7,610 ( 0.00%) #[derive(Decodable, Encodable)] . pub enum FileName { . Real(RealFileName), . /// Call to `quote!`. . QuoteExpansion(u64), . /// Command line. . Anon(u64), . /// Hack in `src/librustc_ast/parse.rs`. . // FIXME(jseyfried) -- line 281 ---------------------------------------- -- line 288 ---------------------------------------- . /// Custom sources for explicit parser calls from plugins and drivers. . Custom(String), . DocTest(PathBuf, isize), . /// Post-substitution inline assembly from LLVM. . InlineAsm(u64), . } . . impl From for FileName { 35 ( 0.00%) fn from(p: PathBuf) -> Self { 10 ( 0.00%) assert!(!p.to_string_lossy().ends_with('>')); 30 ( 0.00%) FileName::Real(RealFileName::LocalPath(p)) 35 ( 0.00%) } . } . 24 ( 0.00%) #[derive(Clone, Copy, Eq, PartialEq, Hash, Debug)] . pub enum FileNameDisplayPreference { . Remapped, . Local, . } . . pub struct FileNameDisplay<'a> { . inner: &'a FileName, . display_pref: FileNameDisplayPreference, . } . . impl fmt::Display for FileNameDisplay<'_> { 32 ( 0.00%) fn fmt(&self, fmt: &mut std::fmt::Formatter<'_>) -> std::fmt::Result { . use FileName::*; 48 ( 0.00%) match *self.inner { . Real(ref name) => { 48 ( 0.00%) write!(fmt, "{}", name.to_string_lossy(self.display_pref)) . } . QuoteExpansion(_) => write!(fmt, ""), . MacroExpansion(_) => write!(fmt, ""), . Anon(_) => write!(fmt, ""), . ProcMacroSourceCode(_) => write!(fmt, ""), . CfgSpec(_) => write!(fmt, ""), . CliCrateAttr(_) => write!(fmt, ""), . Custom(ref s) => write!(fmt, "<{}>", s), . DocTest(ref path, _) => write!(fmt, "{}", path.display()), . InlineAsm(_) => write!(fmt, ""), . } 40 ( 0.00%) } . } . . impl FileNameDisplay<'_> { 8 ( 0.00%) pub fn to_string_lossy(&self) -> Cow<'_, str> { 6 ( 0.00%) match self.inner { 2 ( 0.00%) FileName::Real(ref inner) => inner.to_string_lossy(self.display_pref), . _ => Cow::from(format!("{}", self)), . } 8 ( 0.00%) } . } . . impl FileName { . pub fn is_real(&self) -> bool { . use FileName::*; 1,268 ( 0.00%) match *self { . Real(_) => true, . Anon(_) . | MacroExpansion(_) . | ProcMacroSourceCode(_) . | CfgSpec(_) . | CliCrateAttr(_) . | Custom(_) . | QuoteExpansion(_) . | DocTest(_, _) . | InlineAsm(_) => false, . } . } . 2 ( 0.00%) pub fn prefer_remapped(&self) -> FileNameDisplay<'_> { . FileNameDisplay { inner: self, display_pref: FileNameDisplayPreference::Remapped } 4 ( 0.00%) } . . // This may include transient local filesystem information. . // Must not be embedded in build outputs. 5 ( 0.00%) pub fn prefer_local(&self) -> FileNameDisplay<'_> { . FileNameDisplay { inner: self, display_pref: FileNameDisplayPreference::Local } 10 ( 0.00%) } . . pub fn display(&self, display_pref: FileNameDisplayPreference) -> FileNameDisplay<'_> { . FileNameDisplay { inner: self, display_pref } . } . . pub fn macro_expansion_source_code(src: &str) -> FileName { . let mut hasher = StableHasher::new(); . src.hash(&mut hasher); -- line 375 ---------------------------------------- -- line 423 ---------------------------------------- . /// that the length of the span is equal to `span.hi - span.lo`; there may be space in the . /// [`BytePos`] range between files. . /// . /// `SpanData` is public because `Span` uses a thread-local interner and can't be . /// sent to other threads, but some pieces of performance infra run in a separate thread. . /// Using `Span` is generally preferred. . #[derive(Clone, Copy, Hash, PartialEq, Eq)] . pub struct SpanData { 9 ( 0.00%) pub lo: BytePos, 9 ( 0.00%) pub hi: BytePos, . /// Information about where the macro came from, if this piece of . /// code was created by a macro expansion. 27 ( 0.00%) pub ctxt: SyntaxContext, 9 ( 0.00%) pub parent: Option, . } . . // Order spans by position in the file. . impl Ord for SpanData { . fn cmp(&self, other: &Self) -> Ordering { . let SpanData { . lo: s_lo, . hi: s_hi, -- line 444 ---------------------------------------- -- line 485 ---------------------------------------- . } . #[inline] . pub fn with_parent(&self, parent: Option) -> Span { . Span::new(self.lo, self.hi, self.ctxt, parent) . } . /// Returns `true` if this is a dummy span with any hygienic context. . #[inline] . pub fn is_dummy(self) -> bool { 9,158 ( 0.00%) self.lo.0 == 0 && self.hi.0 == 0 . } . /// Returns `true` if `self` fully encloses `other`. . pub fn contains(self, other: Self) -> bool { 60 ( 0.00%) self.lo <= other.lo && other.hi <= self.hi . } . } . . // The interner is pointed to by a thread local value which is only set on the main thread . // with parallelization is disabled. So we don't allow `Span` to transfer between threads . // to avoid panics and other errors, even though it would be memory safe to do so. . #[cfg(not(parallel_compiler))] . impl !Send for Span {} . #[cfg(not(parallel_compiler))] . impl !Sync for Span {} . . impl PartialOrd for Span { 8,128 ( 0.00%) fn partial_cmp(&self, rhs: &Self) -> Option { 6,096 ( 0.00%) PartialOrd::partial_cmp(&self.data(), &rhs.data()) 8,128 ( 0.00%) } . } . impl Ord for Span { . fn cmp(&self, rhs: &Self) -> Ordering { . Ord::cmp(&self.data(), &rhs.data()) . } . } . . /// A collection of `Span`s. . /// . /// Spans have two orthogonal attributes: . /// . /// - They can be *primary spans*. In this case they are the locus of . /// the error, and would be rendered with `^^^`. . /// - They can have a *label*. In this case, the label is written next . /// to the mark in the snippet when we render. 36 ( 0.00%) #[derive(Clone, Debug, Hash, PartialEq, Eq, Encodable, Decodable)] . pub struct MultiSpan { . primary_spans: Vec, 8 ( 0.00%) span_labels: Vec<(Span, String)>, . } . . impl Span { . #[inline] . pub fn lo(self) -> BytePos { . self.data().lo . } . #[inline] 2,205 ( 0.00%) pub fn with_lo(self, lo: BytePos) -> Span { . self.data().with_lo(lo) 1,470 ( 0.00%) } . #[inline] 136 ( 0.00%) pub fn hi(self) -> BytePos { . self.data().hi 136 ( 0.00%) } . #[inline] 603 ( 0.00%) pub fn with_hi(self, hi: BytePos) -> Span { . self.data().with_hi(hi) 402 ( 0.00%) } . #[inline] . pub fn ctxt(self) -> SyntaxContext { . self.data_untracked().ctxt . } . #[inline] 220 ( 0.00%) pub fn with_ctxt(self, ctxt: SyntaxContext) -> Span { . self.data_untracked().with_ctxt(ctxt) 132 ( 0.00%) } . #[inline] . pub fn parent(self) -> Option { . self.data().parent . } . #[inline] . pub fn with_parent(self, ctxt: Option) -> Span { . self.data().with_parent(ctxt) . } -- line 566 ---------------------------------------- -- line 600 ---------------------------------------- . /// Returns a new span representing an empty span at the beginning of this span. . #[inline] . pub fn shrink_to_lo(self) -> Span { . let span = self.data_untracked(); . span.with_hi(span.lo) . } . /// Returns a new span representing an empty span at the end of this span. . #[inline] 6 ( 0.00%) pub fn shrink_to_hi(self) -> Span { . let span = self.data_untracked(); . span.with_lo(span.hi) 4 ( 0.00%) } . . #[inline] . /// Returns `true` if `hi == lo`. . pub fn is_empty(self) -> bool { . let span = self.data_untracked(); . span.hi == span.lo . } . . /// Returns `self` if `self` is not the dummy span, and `other` otherwise. . pub fn substitute_dummy(self, other: Span) -> Span { . if self.is_dummy() { other } else { self } . } . . /// Returns `true` if `self` fully encloses `other`. 180 ( 0.00%) pub fn contains(self, other: Span) -> bool { . let span = self.data(); . let other = other.data(); . span.contains(other) 120 ( 0.00%) } . . /// Returns `true` if `self` touches `other`. . pub fn overlaps(self, other: Span) -> bool { . let span = self.data(); . let other = other.data(); . span.lo < other.hi && other.lo < span.hi . } . -- line 638 ---------------------------------------- -- line 663 ---------------------------------------- . /// The `Span` for the tokens in the previous macro expansion from which `self` was generated, . /// if any. . pub fn parent_callsite(self) -> Option { . let expn_data = self.ctxt().outer_expn_data(); . if !expn_data.is_root() { Some(expn_data.call_site) } else { None } . } . . /// Walk down the expansion ancestors to find a span that's contained within `outer`. 170 ( 0.00%) pub fn find_ancestor_inside(mut self, outer: Span) -> Option { 85 ( 0.00%) while !outer.contains(self) { . self = self.parent_callsite()?; . } . Some(self) 187 ( 0.00%) } . . /// Edition of the crate from which this span came. 843 ( 0.00%) pub fn edition(self) -> edition::Edition { . self.ctxt().edition() 562 ( 0.00%) } . . #[inline] . pub fn rust_2015(self) -> bool { 222 ( 0.00%) self.edition() == edition::Edition::Edition2015 . } . . #[inline] . pub fn rust_2018(self) -> bool { 64 ( 0.00%) self.edition() >= edition::Edition::Edition2018 . } . . #[inline] . pub fn rust_2021(self) -> bool { 85 ( 0.00%) self.edition() >= edition::Edition::Edition2021 . } . . /// Returns the source callee. . /// . /// Returns `None` if the supplied span has no expansion trace, . /// else returns the `ExpnData` for the macro definition . /// corresponding to the source callsite. . pub fn source_callee(self) -> Option { -- line 703 ---------------------------------------- -- line 707 ---------------------------------------- . } . let expn_data = self.ctxt().outer_expn_data(); . if !expn_data.is_root() { Some(source_callee(expn_data)) } else { None } . } . . /// Checks if a span is "internal" to a macro in which `#[unstable]` . /// items can be used (that is, a macro marked with . /// `#[allow_internal_unstable]`). 40 ( 0.00%) pub fn allows_unstable(self, feature: Symbol) -> bool { 8 ( 0.00%) self.ctxt() . .outer_expn_data() . .allow_internal_unstable . .map_or(false, |features| features.iter().any(|&f| f == feature)) 32 ( 0.00%) } . . /// Checks if this span arises from a compiler desugaring of kind `kind`. 679 ( 0.00%) pub fn is_desugaring(self, kind: DesugaringKind) -> bool { 582 ( 0.00%) match self.ctxt().outer_expn_data().kind { . ExpnKind::Desugaring(k) => k == kind, . _ => false, . } 485 ( 0.00%) } . . /// Returns the compiler desugaring that created this span, or `None` . /// if this span is not from a desugaring. . pub fn desugaring_kind(self) -> Option { . match self.ctxt().outer_expn_data().kind { . ExpnKind::Desugaring(k) => Some(k), . _ => None, . } . } . . /// Checks if a span is "internal" to a macro in which `unsafe` . /// can be used without triggering the `unsafe_code` lint. . // (that is, a macro marked with `#[allow_internal_unsafe]`). 20 ( 0.00%) pub fn allows_unsafe(self) -> bool { 5 ( 0.00%) self.ctxt().outer_expn_data().allow_internal_unsafe 20 ( 0.00%) } . 4 ( 0.00%) pub fn macro_backtrace(mut self) -> impl Iterator { . let mut prev_span = DUMMY_SP; . std::iter::from_fn(move || { . loop { 8 ( 0.00%) let expn_data = self.ctxt().outer_expn_data(); 5 ( 0.00%) if expn_data.is_root() { 1 ( 0.00%) return None; . } . . let is_recursive = expn_data.call_site.source_equal(prev_span); . . prev_span = self; . self = expn_data.call_site; . . // Don't print recursive invocations. . if !is_recursive { . return Some(expn_data); . } . } . }) 8 ( 0.00%) } . . /// Returns a `Span` that would enclose both `self` and `end`. . /// . /// ```text . /// ____ ___ . /// self lorem ipsum end . /// ^^^^^^^^^^^^^^^^^^^^ . /// ``` 16,467 ( 0.00%) pub fn to(self, end: Span) -> Span { . let span_data = self.data(); . let end_data = end.data(); . // FIXME(jseyfried): `self.ctxt` should always equal `end.ctxt` here (cf. issue #23480). . // Return the macro span on its own to avoid weird diagnostic output. It is preferable to . // have an incomplete span than a completely nonsensical one. 2,966 ( 0.00%) if span_data.ctxt != end_data.ctxt { 28 ( 0.00%) if span_data.ctxt == SyntaxContext::root() { . return end; 27 ( 0.00%) } else if end_data.ctxt == SyntaxContext::root() { . return self; . } . // Both spans fall within a macro. . // FIXME(estebank): check if it is the *same* macro. . } . Span::new( . cmp::min(span_data.lo, end_data.lo), . cmp::max(span_data.hi, end_data.hi), . if span_data.ctxt == SyntaxContext::root() { end_data.ctxt } else { span_data.ctxt }, 10,283 ( 0.00%) if span_data.parent == end_data.parent { span_data.parent } else { None }, . ) 13,473 ( 0.00%) } . . /// Returns a `Span` between the end of `self` to the beginning of `end`. . /// . /// ```text . /// ____ ___ . /// self lorem ipsum end . /// ^^^^^^^^^^^^^ . /// ``` 44 ( 0.00%) pub fn between(self, end: Span) -> Span { . let span = self.data(); . let end = end.data(); . Span::new( . span.hi, . end.lo, . if end.ctxt == SyntaxContext::root() { end.ctxt } else { span.ctxt }, 28 ( 0.00%) if span.parent == end.parent { span.parent } else { None }, . ) 32 ( 0.00%) } . . /// Returns a `Span` from the beginning of `self` until the beginning of `end`. . /// . /// ```text . /// ____ ___ . /// self lorem ipsum end . /// ^^^^^^^^^^^^^^^^^ . /// ``` 66 ( 0.00%) pub fn until(self, end: Span) -> Span { . // Most of this function's body is copied from `to`. . // We can't just do `self.to(end.shrink_to_lo())`, . // because to also does some magic where it uses min/max so . // it can handle overlapping spans. Some advanced mis-use of . // `until` with different ctxts makes this visible. . let span_data = self.data(); . let end_data = end.data(); . // FIXME(jseyfried): `self.ctxt` should always equal `end.ctxt` here (cf. issue #23480). . // Return the macro span on its own to avoid weird diagnostic output. It is preferable to . // have an incomplete span than a completely nonsensical one. 12 ( 0.00%) if span_data.ctxt != end_data.ctxt { . if span_data.ctxt == SyntaxContext::root() { . return end; . } else if end_data.ctxt == SyntaxContext::root() { . return self; . } . // Both spans fall within a macro. . // FIXME(estebank): check if it is the *same* macro. . } . Span::new( . span_data.lo, . end_data.lo, . if end_data.ctxt == SyntaxContext::root() { end_data.ctxt } else { span_data.ctxt }, 42 ( 0.00%) if span_data.parent == end_data.parent { span_data.parent } else { None }, . ) 54 ( 0.00%) } . 11 ( 0.00%) pub fn from_inner(self, inner: InnerSpan) -> Span { . let span = self.data(); . Span::new( . span.lo + BytePos::from_usize(inner.start), . span.lo + BytePos::from_usize(inner.end), . span.ctxt, . span.parent, . ) 7 ( 0.00%) } . . /// Equivalent of `Span::def_site` from the proc macro API, . /// except that the location is taken from the `self` span. . pub fn with_def_site_ctxt(self, expn_id: ExpnId) -> Span { 76 ( 0.00%) self.with_ctxt_from_mark(expn_id, Transparency::Opaque) . } . . /// Equivalent of `Span::call_site` from the proc macro API, . /// except that the location is taken from the `self` span. . pub fn with_call_site_ctxt(self, expn_id: ExpnId) -> Span { 2 ( 0.00%) self.with_ctxt_from_mark(expn_id, Transparency::Transparent) . } . . /// Equivalent of `Span::mixed_site` from the proc macro API, . /// except that the location is taken from the `self` span. . pub fn with_mixed_site_ctxt(self, expn_id: ExpnId) -> Span { . self.with_ctxt_from_mark(expn_id, Transparency::SemiTransparent) . } . . /// Produces a span with the same location as `self` and context produced by a macro with the . /// given ID and transparency, assuming that macro was defined directly and not produced by . /// some other macro (which is the case for built-in and procedural macros). 468 ( 0.00%) pub fn with_ctxt_from_mark(self, expn_id: ExpnId, transparency: Transparency) -> Span { . self.with_ctxt(SyntaxContext::root().apply_mark(expn_id, transparency)) 273 ( 0.00%) } . . #[inline] . pub fn apply_mark(self, expn_id: ExpnId, transparency: Transparency) -> Span { . let span = self.data(); 99 ( 0.00%) span.with_ctxt(span.ctxt.apply_mark(expn_id, transparency)) . } . . #[inline] . pub fn remove_mark(&mut self) -> ExpnId { . let mut span = self.data(); . let mark = span.ctxt.remove_mark(); . *self = Span::new(span.lo, span.hi, span.ctxt, span.parent); . mark -- line 897 ---------------------------------------- -- line 901 ---------------------------------------- . pub fn adjust(&mut self, expn_id: ExpnId) -> Option { . let mut span = self.data(); . let mark = span.ctxt.adjust(expn_id); . *self = Span::new(span.lo, span.hi, span.ctxt, span.parent); . mark . } . . #[inline] 2,655 ( 0.00%) pub fn normalize_to_macros_2_0_and_adjust(&mut self, expn_id: ExpnId) -> Option { 1,770 ( 0.00%) let mut span = self.data(); 1,463 ( 0.00%) let mark = span.ctxt.normalize_to_macros_2_0_and_adjust(expn_id); 2,655 ( 0.00%) *self = Span::new(span.lo, span.hi, span.ctxt, span.parent); . mark 2,938 ( 0.00%) } . . #[inline] . pub fn glob_adjust(&mut self, expn_id: ExpnId, glob_span: Span) -> Option> { 26 ( 0.00%) let mut span = self.data(); 39 ( 0.00%) let mark = span.ctxt.glob_adjust(expn_id, glob_span); 104 ( 0.00%) *self = Span::new(span.lo, span.hi, span.ctxt, span.parent); . mark . } . . #[inline] 18 ( 0.00%) pub fn reverse_glob_adjust( . &mut self, . expn_id: ExpnId, . glob_span: Span, . ) -> Option> { 32 ( 0.00%) let mut span = self.data(); 50 ( 0.00%) let mark = span.ctxt.reverse_glob_adjust(expn_id, glob_span); 98 ( 0.00%) *self = Span::new(span.lo, span.hi, span.ctxt, span.parent); . mark 20 ( 0.00%) } . . #[inline] 2,104 ( 0.00%) pub fn normalize_to_macros_2_0(self) -> Span { . let span = self.data(); . span.with_ctxt(span.ctxt.normalize_to_macros_2_0()) 1,841 ( 0.00%) } . . #[inline] . pub fn normalize_to_macro_rules(self) -> Span { . let span = self.data(); . span.with_ctxt(span.ctxt.normalize_to_macro_rules()) . } . } . -- line 948 ---------------------------------------- -- line 958 ---------------------------------------- . . /// What label should we attach to this span (if any)? . pub label: Option, . } . . impl Default for Span { . fn default() -> Self { . DUMMY_SP 2 ( 0.00%) } . } . . impl Encodable for Span { . default fn encode(&self, s: &mut E) -> Result<(), E::Error> { . let span = self.data(); . s.emit_struct(false, |s| { . s.emit_struct_field("lo", true, |s| span.lo.encode(s))?; . s.emit_struct_field("hi", false, |s| span.hi.encode(s)) -- line 974 ---------------------------------------- -- line 990 ---------------------------------------- . /// any spans that are debug-printed during the closure's execution. . /// . /// Normally, the global `TyCtxt` is used to retrieve the `SourceMap` . /// (see `rustc_interface::callbacks::span_debug1`). However, some parts . /// of the compiler (e.g. `rustc_parse`) may debug-print `Span`s before . /// a `TyCtxt` is available. In this case, we fall back to . /// the `SourceMap` provided to this function. If that is not available, . /// we fall back to printing the raw `Span` field values. 9 ( 0.00%) pub fn with_source_map T>(source_map: Lrc, f: F) -> T { . with_session_globals(|session_globals| { 2 ( 0.00%) *session_globals.source_map.borrow_mut() = Some(source_map); . }); . struct ClearSourceMap; . impl Drop for ClearSourceMap { . fn drop(&mut self) { . with_session_globals(|session_globals| { 1 ( 0.00%) session_globals.source_map.borrow_mut().take(); . }); . } . } . . let _guard = ClearSourceMap; 4 ( 0.00%) f() 8 ( 0.00%) } . . pub fn debug_with_source_map( . span: Span, . f: &mut fmt::Formatter<'_>, . source_map: &SourceMap, . ) -> fmt::Result { . write!(f, "{} ({:?})", source_map.span_to_diagnostic_string(span), span.ctxt()) . } -- line 1021 ---------------------------------------- -- line 1044 ---------------------------------------- . fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { . (*SPAN_DEBUG)(Span::new(self.lo, self.hi, self.ctxt, self.parent), f) . } . } . . impl MultiSpan { . #[inline] . pub fn new() -> MultiSpan { 2 ( 0.00%) MultiSpan { primary_spans: vec![], span_labels: vec![] } . } . . pub fn from_span(primary_span: Span) -> MultiSpan { 215 ( 0.00%) MultiSpan { primary_spans: vec![primary_span], span_labels: vec![] } . } . . pub fn from_spans(mut vec: Vec) -> MultiSpan { . vec.sort(); 20 ( 0.00%) MultiSpan { primary_spans: vec, span_labels: vec![] } . } . 6 ( 0.00%) pub fn push_span_label(&mut self, span: Span, label: String) { 2 ( 0.00%) self.span_labels.push((span, label)); 5 ( 0.00%) } . . /// Selects the first primary span (if any). . pub fn primary_span(&self) -> Option { . self.primary_spans.first().cloned() 27 ( 0.00%) } . . /// Returns all primary spans. . pub fn primary_spans(&self) -> &[Span] { . &self.primary_spans 13 ( 0.00%) } . . /// Returns `true` if any of the primary spans are displayable. 18 ( 0.00%) pub fn has_primary_spans(&self) -> bool { 3 ( 0.00%) self.primary_spans.iter().any(|sp| !sp.is_dummy()) 21 ( 0.00%) } . . /// Returns `true` if this contains only a dummy primary span with any hygienic context. . pub fn is_dummy(&self) -> bool { . let mut is_dummy = true; . for span in &self.primary_spans { . if !span.is_dummy() { . is_dummy = false; . } -- line 1089 ---------------------------------------- -- line 1110 ---------------------------------------- . replacements_occurred . } . . /// Returns the strings to highlight. We always ensure that there . /// is an entry for each of the primary spans -- for each primary . /// span `P`, if there is at least one label with span `P`, we return . /// those labels (marked as primary). But otherwise we return . /// `SpanLabel` instances with empty labels. 84 ( 0.00%) pub fn span_labels(&self) -> Vec { 15 ( 0.00%) let is_primary = |span| self.primary_spans.contains(&span); . . let mut span_labels = self . .span_labels . .iter() 9 ( 0.00%) .map(|&(span, ref label)| SpanLabel { . span, . is_primary: is_primary(span), 6 ( 0.00%) label: Some(label.clone()), . }) . .collect::>(); . 4 ( 0.00%) for &span in &self.primary_spans { . if !span_labels.iter().any(|sl| sl.span == span) { . span_labels.push(SpanLabel { span, is_primary: true, label: None }); . } . } . . span_labels 84 ( 0.00%) } . . /// Returns `true` if any of the span labels is displayable. 12 ( 0.00%) pub fn has_span_labels(&self) -> bool { . self.span_labels.iter().any(|(sp, _)| !sp.is_dummy()) 14 ( 0.00%) } . } . . impl From for MultiSpan { 215 ( 0.00%) fn from(span: Span) -> MultiSpan { . MultiSpan::from_span(span) 215 ( 0.00%) } . } . . impl From> for MultiSpan { 20 ( 0.00%) fn from(spans: Vec) -> MultiSpan { 16 ( 0.00%) MultiSpan::from_spans(spans) 20 ( 0.00%) } . } . . /// Identifies an offset of a multi-byte character in a `SourceFile`. . #[derive(Copy, Clone, Encodable, Decodable, Eq, PartialEq, Debug)] . pub struct MultiByteChar { . /// The absolute offset of the character in the `SourceMap`. . pub pos: BytePos, . /// The number of bytes, `>= 2`. . pub bytes: u8, . } . . /// Identifies an offset of a non-narrow character in a `SourceFile`. 308 ( 0.00%) #[derive(Copy, Clone, Encodable, Decodable, Eq, PartialEq, Debug)] . pub enum NonNarrowChar { . /// Represents a zero-width character. . ZeroWidth(BytePos), . /// Represents a wide (full-width) character. . Wide(BytePos), . /// Represents a tab character, represented visually with a width of 4 characters. . Tab(BytePos), . } -- line 1176 ---------------------------------------- -- line 1201 ---------------------------------------- . } . } . } . . impl Add for NonNarrowChar { . type Output = Self; . . fn add(self, rhs: BytePos) -> Self { 532 ( 0.00%) match self { . NonNarrowChar::ZeroWidth(pos) => NonNarrowChar::ZeroWidth(pos + rhs), . NonNarrowChar::Wide(pos) => NonNarrowChar::Wide(pos + rhs), . NonNarrowChar::Tab(pos) => NonNarrowChar::Tab(pos + rhs), . } . } . } . . impl Sub for NonNarrowChar { . type Output = Self; . 154 ( 0.00%) fn sub(self, rhs: BytePos) -> Self { 532 ( 0.00%) match self { . NonNarrowChar::ZeroWidth(pos) => NonNarrowChar::ZeroWidth(pos - rhs), . NonNarrowChar::Wide(pos) => NonNarrowChar::Wide(pos - rhs), . NonNarrowChar::Tab(pos) => NonNarrowChar::Tab(pos - rhs), . } 462 ( 0.00%) } . } . . /// Identifies an offset of a character that was normalized away from `SourceFile`. . #[derive(Copy, Clone, Encodable, Decodable, Eq, PartialEq, Debug)] . pub struct NormalizedPos { . /// The absolute offset of the character in the `SourceMap`. . pub pos: BytePos, . /// The difference between original and normalized string at position. . pub diff: u32, . } . 15 ( 0.00%) #[derive(PartialEq, Eq, Clone, Debug)] . pub enum ExternalSource { . /// No external source has to be loaded, since the `SourceFile` represents a local crate. . Unneeded, . Foreign { . kind: ExternalSourceKind, . /// This SourceFile's byte-offset within the source_map of its original crate. . original_start_pos: BytePos, . /// The end of this SourceFile within the source_map of its original crate. -- line 1246 ---------------------------------------- -- line 1257 ---------------------------------------- . AbsentOk, . /// A failed attempt has been made to load the external source. . AbsentErr, . Unneeded, . } . . impl ExternalSource { . pub fn get_source(&self) -> Option<&Lrc> { 8 ( 0.00%) match self { . ExternalSource::Foreign { kind: ExternalSourceKind::Present(ref src), .. } => Some(src), . _ => None, . } . } . } . . #[derive(Debug)] . pub struct OffsetOverflowError; . 1,268 ( 0.00%) #[derive(Copy, Clone, Debug, PartialEq, Eq, PartialOrd, Ord, Hash, Encodable, Decodable)] . pub enum SourceFileHashAlgorithm { . Md5, . Sha1, . Sha256, . } . . impl FromStr for SourceFileHashAlgorithm { . type Err = (); -- line 1283 ---------------------------------------- -- line 1290 ---------------------------------------- . _ => Err(()), . } . } . } . . rustc_data_structures::impl_stable_hash_via_hash!(SourceFileHashAlgorithm); . . /// The hash of the on-disk source file used for debug info. 30 ( 0.00%) #[derive(Copy, Clone, PartialEq, Eq, Debug)] 2,536 ( 0.00%) #[derive(HashStable_Generic, Encodable, Decodable)] . pub struct SourceFileHash { . pub kind: SourceFileHashAlgorithm, . value: [u8; 32], . } . . impl SourceFileHash { . pub fn new(kind: SourceFileHashAlgorithm, src: &str) -> SourceFileHash { . let mut hash = SourceFileHash { kind, value: Default::default() }; . let len = hash.hash_len(); . let value = &mut hash.value[..len]; . let data = src.as_bytes(); . match kind { . SourceFileHashAlgorithm::Md5 => { 15 ( 0.00%) value.copy_from_slice(&Md5::digest(data)); . } . SourceFileHashAlgorithm::Sha1 => { . value.copy_from_slice(&Sha1::digest(data)); . } . SourceFileHashAlgorithm::Sha256 => { . value.copy_from_slice(&Sha256::digest(data)); . } . } -- line 1321 ---------------------------------------- -- line 1329 ---------------------------------------- . . /// The bytes of the hash. . pub fn hash_bytes(&self) -> &[u8] { . let len = self.hash_len(); . &self.value[..len] . } . . fn hash_len(&self) -> usize { 15 ( 0.00%) match self.kind { . SourceFileHashAlgorithm::Md5 => 16, . SourceFileHashAlgorithm::Sha1 => 20, . SourceFileHashAlgorithm::Sha256 => 32, . } . } . } . . /// A single source in the [`SourceMap`]. 290 ( 0.00%) #[derive(Clone)] . pub struct SourceFile { . /// The name of the file that the source came from. Source that doesn't . /// originate from files has names between angle brackets by convention . /// (e.g., ``). . pub name: FileName, . /// The complete source code. 5 ( 0.00%) pub src: Option>, . /// The source code's hash. . pub src_hash: SourceFileHash, . /// The external source code (used for external crates, which will have a `None` . /// value as `self.src`. . pub external_src: Lock, . /// The start position of this source in the `SourceMap`. . pub start_pos: BytePos, . /// The end position of this source in the `SourceMap`. -- line 1361 ---------------------------------------- -- line 1364 ---------------------------------------- . pub lines: Vec, . /// Locations of multi-byte characters in the source code. . pub multibyte_chars: Vec, . /// Width of characters that are not narrow in the source code. . pub non_narrow_chars: Vec, . /// Locations of characters removed during normalization. . pub normalized_pos: Vec, . /// A hash of the filename, used for speeding up hashing in incremental compilation. 5 ( 0.00%) pub name_hash: u128, . /// Indicates which crate this `SourceFile` was imported from. 5 ( 0.00%) pub cnum: CrateNum, . } . . impl Encodable for SourceFile { . fn encode(&self, s: &mut S) -> Result<(), S::Error> { . s.emit_struct(false, |s| { . s.emit_struct_field("name", true, |s| self.name.encode(s))?; . s.emit_struct_field("src_hash", false, |s| self.src_hash.encode(s))?; . s.emit_struct_field("start_pos", false, |s| self.start_pos.encode(s))?; . s.emit_struct_field("end_pos", false, |s| self.end_pos.encode(s))?; . s.emit_struct_field("lines", false, |s| { . let lines = &self.lines[..]; . // Store the length. . s.emit_u32(lines.len() as u32)?; . 20 ( 0.00%) if !lines.is_empty() { . // In order to preserve some space, we exploit the fact that . // the lines list is sorted and individual lines are . // probably not that long. Because of that we can store lines . // as a difference list, using as little space as possible . // for the differences. . let max_line_length = if lines.len() == 1 { . 0 . } else { -- line 1397 ---------------------------------------- -- line 1399 ---------------------------------------- . .array_windows() . .map(|&[fst, snd]| snd - fst) . .map(|bp| bp.to_usize()) . .max() . .unwrap() . }; . . let bytes_per_diff: u8 = match max_line_length { 10 ( 0.00%) 0..=0xFF => 1, . 0x100..=0xFFFF => 2, . _ => 4, . }; . . // Encode the number of bytes used per diff. . bytes_per_diff.encode(s)?; . . // Encode the first element. . lines[0].encode(s)?; . . let diff_iter = lines.array_windows().map(|&[fst, snd]| snd - fst); . 10 ( 0.00%) match bytes_per_diff { . 1 => { . for diff in diff_iter { . (diff.0 as u8).encode(s)? . } . } . 2 => { . for diff in diff_iter { . (diff.0 as u16).encode(s)? -- line 1428 ---------------------------------------- -- line 1436 ---------------------------------------- . _ => unreachable!(), . } . } . . Ok(()) . })?; . s.emit_struct_field("multibyte_chars", false, |s| self.multibyte_chars.encode(s))?; . s.emit_struct_field("non_narrow_chars", false, |s| self.non_narrow_chars.encode(s))?; 15 ( 0.00%) s.emit_struct_field("name_hash", false, |s| self.name_hash.encode(s))?; . s.emit_struct_field("normalized_pos", false, |s| self.normalized_pos.encode(s))?; 55 ( 0.00%) s.emit_struct_field("cnum", false, |s| self.cnum.encode(s)) . }) . } . } . . impl Decodable for SourceFile { 5,706 ( 0.00%) fn decode(d: &mut D) -> SourceFile { . d.read_struct(|d| { . let name: FileName = d.read_struct_field("name", |d| Decodable::decode(d)); . let src_hash: SourceFileHash = . d.read_struct_field("src_hash", |d| Decodable::decode(d)); . let start_pos: BytePos = d.read_struct_field("start_pos", |d| Decodable::decode(d)); . let end_pos: BytePos = d.read_struct_field("end_pos", |d| Decodable::decode(d)); . let lines: Vec = d.read_struct_field("lines", |d| { . let num_lines: u32 = Decodable::decode(d); 634 ( 0.00%) let mut lines = Vec::with_capacity(num_lines as usize); . . if num_lines > 0 { . // Read the number of bytes used per diff. . let bytes_per_diff: u8 = Decodable::decode(d); . . // Read the first element. . let mut line_start: BytePos = Decodable::decode(d); . lines.push(line_start); . . for _ in 1..num_lines { 952,186 ( 0.20%) let diff = match bytes_per_diff { . 1 => d.read_u8() as u32, . 2 => d.read_u16() as u32, . 4 => d.read_u32(), . _ => unreachable!(), . }; . . line_start = line_start + BytePos(diff); . -- line 1480 ---------------------------------------- -- line 1483 ---------------------------------------- . } . . lines . }); . let multibyte_chars: Vec = . d.read_struct_field("multibyte_chars", |d| Decodable::decode(d)); . let non_narrow_chars: Vec = . d.read_struct_field("non_narrow_chars", |d| Decodable::decode(d)); 1,268 ( 0.00%) let name_hash: u128 = d.read_struct_field("name_hash", |d| Decodable::decode(d)); . let normalized_pos: Vec = . d.read_struct_field("normalized_pos", |d| Decodable::decode(d)); . let cnum: CrateNum = d.read_struct_field("cnum", |d| Decodable::decode(d)); 6,974 ( 0.00%) SourceFile { 5,072 ( 0.00%) name, . start_pos, . end_pos, . src: None, 2,536 ( 0.00%) src_hash, . // Unused - the metadata decoder will construct . // a new SourceFile, filling in `external_src` properly . external_src: Lock::new(ExternalSource::Unneeded), 2,536 ( 0.00%) lines, 2,536 ( 0.00%) multibyte_chars, 2,536 ( 0.00%) non_narrow_chars, 2,536 ( 0.00%) normalized_pos, . name_hash, . cnum, . } . }) 5,706 ( 0.00%) } . } . . impl fmt::Debug for SourceFile { . fn fmt(&self, fmt: &mut fmt::Formatter<'_>) -> fmt::Result { . write!(fmt, "SourceFile({:?})", self.name) . } . } . . impl SourceFile { 55 ( 0.00%) pub fn new( . name: FileName, . mut src: String, . start_pos: BytePos, . hash_kind: SourceFileHashAlgorithm, . ) -> Self { . // Compute the file hash before any normalization. . let src_hash = SourceFileHash::new(hash_kind, &src); 15 ( 0.00%) let normalized_pos = normalize_src(&mut src, start_pos); . . let name_hash = { . let mut hasher: StableHasher = StableHasher::new(); 10 ( 0.00%) name.hash(&mut hasher); . hasher.finish::() . }; 10 ( 0.00%) let end_pos = start_pos.to_usize() + src.len(); 15 ( 0.00%) assert!(end_pos <= u32::MAX as usize); . 60 ( 0.00%) let (lines, multibyte_chars, non_narrow_chars) = 10 ( 0.00%) analyze_source_file::analyze_source_file(&src, start_pos); . 85 ( 0.00%) SourceFile { . name, . src: Some(Lrc::new(src)), 55 ( 0.00%) src_hash, . external_src: Lock::new(ExternalSource::Unneeded), . start_pos, . end_pos: Pos::from_usize(end_pos), 20 ( 0.00%) lines, 20 ( 0.00%) multibyte_chars, 20 ( 0.00%) non_narrow_chars, 20 ( 0.00%) normalized_pos, . name_hash, . cnum: LOCAL_CRATE, . } 45 ( 0.00%) } . . /// Returns the `BytePos` of the beginning of the current line. . pub fn line_begin_pos(&self, pos: BytePos) -> BytePos { . let line_index = self.lookup_line(pos).unwrap(); . self.lines[line_index] . } . . /// Add externally loaded source. . /// If the hash of the input doesn't match or no input is supplied via None, . /// it is interpreted as an error and the corresponding enum variant is set. . /// The return value signifies whether some kind of source is present. 133 ( 0.00%) pub fn add_external_src(&self, get_src: F) -> bool . where . F: FnOnce() -> Option, . { 19 ( 0.00%) if matches!( 38 ( 0.00%) *self.external_src.borrow(), . ExternalSource::Foreign { kind: ExternalSourceKind::AbsentOk, .. } . ) { . let src = get_src(); . let mut external_src = self.external_src.borrow_mut(); . // Check that no-one else have provided the source while we were getting it . if let ExternalSource::Foreign { . kind: src_kind @ ExternalSourceKind::AbsentOk, .. . } = &mut *external_src -- line 1582 ---------------------------------------- -- line 1592 ---------------------------------------- . *src_kind = ExternalSourceKind::AbsentErr; . } . . false . } else { . self.src.is_some() || external_src.get_source().is_some() . } . } else { 19 ( 0.00%) self.src.is_some() || self.external_src.borrow().get_source().is_some() . } 171 ( 0.00%) } . . /// Gets a line from the list of pre-computed line-beginnings. . /// The line number here is 0-based. 35 ( 0.00%) pub fn get_line(&self, line_number: usize) -> Option> { 35 ( 0.00%) fn get_until_newline(src: &str, begin: usize) -> &str { . // We can't use `lines.get(line_number+1)` because we might . // be parsing when we call this function and thus the current . // line is the last one we have line info for. . let slice = &src[begin..]; . match slice.find('\n') { . Some(e) => &slice[..e], . None => slice, . } 40 ( 0.00%) } . . let begin = { 5 ( 0.00%) let line = self.lines.get(line_number)?; 5 ( 0.00%) let begin: BytePos = *line - self.start_pos; . begin.to_usize() . }; . 15 ( 0.00%) if let Some(ref src) = self.src { 25 ( 0.00%) Some(Cow::from(get_until_newline(src, begin))) . } else if let Some(src) = self.external_src.borrow().get_source() { . Some(Cow::Owned(String::from(get_until_newline(src, begin)))) . } else { . None . } 40 ( 0.00%) } . . pub fn is_real_file(&self) -> bool { . self.name.is_real() 634 ( 0.00%) } . . pub fn is_imported(&self) -> bool { . self.src.is_none() 1,605 ( 0.00%) } . . pub fn count_lines(&self) -> usize { . self.lines.len() . } . . /// Finds the line containing the given position. The return value is the . /// index into the `lines` array of this `SourceFile`, not the 1-based line . /// number. If the source_file is empty or the position is located before the . /// first line, `None` is returned. . pub fn lookup_line(&self, pos: BytePos) -> Option { 447 ( 0.00%) match self.lines.binary_search(&pos) { . Ok(idx) => Some(idx), . Err(0) => None, . Err(idx) => Some(idx - 1), . } . } . . pub fn line_bounds(&self, line_index: usize) -> Range { 812 ( 0.00%) if self.is_empty() { . return self.start_pos..self.end_pos; . } . 559 ( 0.00%) assert!(line_index < self.lines.len()); 777 ( 0.00%) if line_index == (self.lines.len() - 1) { 36 ( 0.00%) self.lines[line_index]..self.end_pos . } else { 490 ( 0.00%) self.lines[line_index]..self.lines[line_index + 1] . } . } . . /// Returns whether or not the file contains the given `SourceMap` byte . /// position. The position one past the end of the file is considered to be . /// contained by the file. This implies that files for which `is_empty` . /// returns true still contain one byte position according to this function. . #[inline] -- line 1674 ---------------------------------------- -- line 1678 ---------------------------------------- . . #[inline] . pub fn is_empty(&self) -> bool { . self.start_pos == self.end_pos . } . . /// Calculates the original byte position relative to the start of the file . /// based on the given byte position. 4 ( 0.00%) pub fn original_relative_byte_pos(&self, pos: BytePos) -> BytePos { . // Diff before any records is 0. Otherwise use the previously recorded . // diff as that applies to the following characters until a new diff . // is recorded. . let diff = match self.normalized_pos.binary_search_by(|np| np.pos.cmp(&pos)) { . Ok(i) => self.normalized_pos[i].diff, . Err(i) if i == 0 => 0, . Err(i) => self.normalized_pos[i - 1].diff, . }; . 16 ( 0.00%) BytePos::from_u32(pos.0 - self.start_pos.0 + diff) 12 ( 0.00%) } . . /// Converts an absolute `BytePos` to a `CharPos` relative to the `SourceFile`. 32 ( 0.00%) pub fn bytepos_to_file_charpos(&self, bpos: BytePos) -> CharPos { . // The number of extra bytes due to multibyte chars in the `SourceFile`. . let mut total_extra_bytes = 0; . 96 ( 0.00%) for mbc in self.multibyte_chars.iter() { . debug!("{}-byte char at {:?}", mbc.bytes, mbc.pos); 28 ( 0.00%) if mbc.pos < bpos { . // Every character is at least one byte, so we only . // count the actual extra bytes. 72 ( 0.00%) total_extra_bytes += mbc.bytes as u32 - 1; . // We should never see a byte position in the middle of a . // character. 60 ( 0.00%) assert!(bpos.to_u32() >= mbc.pos.to_u32() + mbc.bytes as u32); . } else { . break; . } . } . 192 ( 0.00%) assert!(self.start_pos.to_u32() + total_extra_bytes <= bpos.to_u32()); 160 ( 0.00%) CharPos(bpos.to_usize() - self.start_pos.to_usize() - total_extra_bytes as usize) 64 ( 0.00%) } . . /// Looks up the file's (1-based) line number and (0-based `CharPos`) column offset, for a . /// given `BytePos`. 112 ( 0.00%) pub fn lookup_file_pos(&self, pos: BytePos) -> (usize, CharPos) { 32 ( 0.00%) let chpos = self.bytepos_to_file_charpos(pos); 15 ( 0.00%) match self.lookup_line(pos) { . Some(a) => { . let line = a + 1; // Line numbers start at 1 16 ( 0.00%) let linebpos = self.lines[a]; 32 ( 0.00%) let linechpos = self.bytepos_to_file_charpos(linebpos); . let col = chpos - linechpos; . debug!("byte pos {:?} is on the line at byte pos {:?}", pos, linebpos); . debug!("char pos {:?} is on the line at char pos {:?}", chpos, linechpos); . debug!("byte is on line: {}", line); 16 ( 0.00%) assert!(chpos >= linechpos); . (line, col) . } . None => (0, chpos), . } 128 ( 0.00%) } . . /// Looks up the file's (1-based) line number, (0-based `CharPos`) column offset, and (0-based) . /// column offset when displayed, for a given `BytePos`. 128 ( 0.00%) pub fn lookup_file_pos_with_col_display(&self, pos: BytePos) -> (usize, CharPos, usize) { 48 ( 0.00%) let (line, col_or_chpos) = self.lookup_file_pos(pos); 32 ( 0.00%) if line > 0 { . let col = col_or_chpos; 16 ( 0.00%) let linebpos = self.lines[line - 1]; . let col_display = { . let start_width_idx = self . .non_narrow_chars . .binary_search_by_key(&linebpos, |x| x.pos()) . .unwrap_or_else(|x| x); . let end_width_idx = self . .non_narrow_chars . .binary_search_by_key(&pos, |x| x.pos()) . .unwrap_or_else(|x| x); . let special_chars = end_width_idx - start_width_idx; . let non_narrow: usize = self.non_narrow_chars[start_width_idx..end_width_idx] . .iter() . .map(|x| x.width()) . .sum(); 80 ( 0.00%) col.0 - special_chars + non_narrow . }; . (line, col, col_display) . } else { . let chpos = col_or_chpos; . let col_display = { . let end_width_idx = self . .non_narrow_chars . .binary_search_by_key(&pos, |x| x.pos()) . .unwrap_or_else(|x| x); . let non_narrow: usize = . self.non_narrow_chars[0..end_width_idx].iter().map(|x| x.width()).sum(); . chpos.0 - end_width_idx + non_narrow . }; . (0, chpos, col_display) . } 112 ( 0.00%) } . } . . /// Normalizes the source code and records the normalizations. 50 ( 0.00%) fn normalize_src(src: &mut String, start_pos: BytePos) -> Vec { . let mut normalized_pos = vec![]; . remove_bom(src, &mut normalized_pos); . normalize_newlines(src, &mut normalized_pos); . . // Offset all the positions by start_pos to match the final file positions. . for np in &mut normalized_pos { . np.pos.0 += start_pos.0; . } . . normalized_pos 45 ( 0.00%) } . . /// Removes UTF-8 BOM, if any. . fn remove_bom(src: &mut String, normalized_pos: &mut Vec) { 5 ( 0.00%) if src.starts_with('\u{feff}') { . src.drain(..3); . normalized_pos.push(NormalizedPos { pos: BytePos(0), diff: 3 }); . } . } . . /// Replaces `\r\n` with `\n` in-place in `src`. . /// . /// Returns error if there's a lone `\r` in the string. . fn normalize_newlines(src: &mut String, normalized_pos: &mut Vec) { 5 ( 0.00%) if !src.as_bytes().contains(&b'\r') { . return; . } . . // We replace `\r\n` with `\n` in-place, which doesn't break utf-8 encoding. . // While we *can* call `as_mut_vec` and do surgery on the live string . // directly, let's rather steal the contents of `src`. This makes the code . // safe even if a panic occurs. . -- line 1816 ---------------------------------------- -- line 1877 ---------------------------------------- . ( . $( . $(#[$attr:meta])* . $vis:vis struct $ident:ident($inner_vis:vis $inner_ty:ty); . )* . ) => { . $( . $(#[$attr])* 2,738 ( 0.00%) $vis struct $ident($inner_vis $inner_ty); . . impl Pos for $ident { . #[inline(always)] . fn from_usize(n: usize) -> $ident { 261 ( 0.00%) $ident(n as $inner_ty) . } . . #[inline(always)] . fn to_usize(&self) -> usize { 5,176 ( 0.00%) self.0 as usize . } . . #[inline(always)] . fn from_u32(n: u32) -> $ident { . $ident(n as $inner_ty) . } . . #[inline(always)] -- line 1903 ---------------------------------------- -- line 1906 ---------------------------------------- . } . } . . impl Add for $ident { . type Output = $ident; . . #[inline(always)] . fn add(self, rhs: $ident) -> $ident { 1,766,417 ( 0.37%) $ident(self.0 + rhs.0) . } . } . . impl Sub for $ident { . type Output = $ident; . . #[inline(always)] . fn sub(self, rhs: $ident) -> $ident { 943,795 ( 0.20%) $ident(self.0 - rhs.0) . } . } . )* . }; . } . . impl_pos! { . /// A byte offset. -- line 1931 ---------------------------------------- -- line 1946 ---------------------------------------- . impl Encodable for BytePos { . fn encode(&self, s: &mut S) -> Result<(), S::Error> { . s.emit_u32(self.0) . } . } . . impl Decodable for BytePos { . fn decode(d: &mut D) -> BytePos { 176,076 ( 0.04%) BytePos(d.read_u32()) . } . } . . // _____________________________________________________________________________ . // Loc, SourceFileAndLine, SourceFileAndBytePos . // . . /// A source code location used for error reporting. 4 ( 0.00%) #[derive(Debug, Clone)] . pub struct Loc { . /// Information about the original source. . pub file: Lrc, . /// The (1-based) line number. 1 ( 0.00%) pub line: usize, . /// The (0-based) column offset. . pub col: CharPos, . /// The (0-based) column offset when displayed. 1 ( 0.00%) pub col_display: usize, . } . . // Used to be structural records. . #[derive(Debug)] . pub struct SourceFileAndLine { . pub sf: Lrc, . /// Index of line, starting from 0. . pub line: usize, -- line 1980 ---------------------------------------- -- line 2042 ---------------------------------------- . /// Range inside of a `Span` used for diagnostics when we only have access to relative positions. . #[derive(Copy, Clone, PartialEq, Eq, Debug)] . pub struct InnerSpan { . pub start: usize, . pub end: usize, . } . . impl InnerSpan { 6 ( 0.00%) pub fn new(start: usize, end: usize) -> InnerSpan { . InnerSpan { start, end } 3 ( 0.00%) } . } . . /// Requirements for a `StableHashingContext` to be used in this crate. . /// . /// This is a hack to allow using the [`HashStable_Generic`] derive macro . /// instead of implementing everything in rustc_middle. . pub trait HashStableContext { . fn def_path_hash(&self, def_id: DefId) -> DefPathHash; -- line 2060 ---------------------------------------- -- line 2079 ---------------------------------------- . /// offsets into the `SourceMap`). Instead, we hash the (file name, line, column) . /// triple, which stays the same even if the containing `SourceFile` has moved . /// within the `SourceMap`. . /// . /// Also note that we are hashing byte offsets for the column, not unicode . /// codepoint offsets. For the purpose of the hash that's sufficient. . /// Also, hashing filenames is expensive so we avoid doing it twice when the . /// span starts and ends in the same file, which is almost always the case. 14,259 ( 0.00%) fn hash_stable(&self, ctx: &mut CTX, hasher: &mut StableHasher) { . const TAG_VALID_SPAN: u8 = 0; . const TAG_INVALID_SPAN: u8 = 1; . const TAG_RELATIVE_SPAN: u8 = 2; . 2,668 ( 0.00%) if !ctx.hash_spans() { . return; . } . 11,591 ( 0.00%) let span = self.data_untracked(); 4,218 ( 0.00%) span.ctxt.hash_stable(ctx, hasher); 6,111 ( 0.00%) span.parent.hash_stable(ctx, hasher); . 4,119 ( 0.00%) if span.is_dummy() { . Hash::hash(&TAG_INVALID_SPAN, hasher); . return; . } . 3,984 ( 0.00%) if let Some(parent) = span.parent { . let def_span = ctx.def_span(parent).data_untracked(); . if def_span.contains(span) { . // This span is enclosed in a definition: only hash the relative position. . Hash::hash(&TAG_RELATIVE_SPAN, hasher); . (span.lo - def_span.lo).to_u32().hash_stable(ctx, hasher); . (span.hi - def_span.lo).to_u32().hash_stable(ctx, hasher); . return; . } . } . . // If this is not an empty or invalid span, we want to hash the last . // position that belongs to it, as opposed to hashing the first . // position past it. 7,968 ( 0.00%) let (file, line_lo, col_lo, line_hi, col_hi) = match ctx.span_data_to_lines_and_cols(&span) . { 3,984 ( 0.00%) Some(pos) => pos, . None => { . Hash::hash(&TAG_INVALID_SPAN, hasher); . return; . } . }; . . Hash::hash(&TAG_VALID_SPAN, hasher); . // We truncate the stable ID hash and line and column numbers. The chances -- line 2129 ---------------------------------------- -- line 2134 ---------------------------------------- . // hash only the length, for example, then two otherwise equal spans with . // different end locations will have the same hash. This can cause a problem . // during incremental compilation wherein a previous result for a query that . // depends on the end location of a span will be incorrectly reused when the . // end location of the span it depends on has changed (see issue #74890). A . // similar analysis applies if some query depends specifically on the length . // of the span, but we only hash the end location. So hash both. . 1,992 ( 0.00%) let col_lo_trunc = (col_lo.0 as u64) & 0xFF; . let line_lo_trunc = ((line_lo as u64) & 0xFF_FF_FF) << 8; . let col_hi_trunc = (col_hi.0 as u64) & 0xFF << 32; . let line_hi_trunc = ((line_hi as u64) & 0xFF_FF_FF) << 40; . let col_line = col_lo_trunc | line_lo_trunc | col_hi_trunc | line_hi_trunc; 1,992 ( 0.00%) let len = (span.hi - span.lo).0; . Hash::hash(&col_line, hasher); . Hash::hash(&len, hasher); 16,296 ( 0.00%) } . } 1,347,678 ( 0.28%) -------------------------------------------------------------------------------- -- Auto-annotated source: /usr/home/liquid/.cargo/registry/src/github.com-1ecc6299db9ec823/rustc-hash-1.1.0/src/lib.rs -------------------------------------------------------------------------------- Ir -- line 71 ---------------------------------------- . fn default() -> FxHasher { . FxHasher { hash: 0 } . } . } . . impl FxHasher { . #[inline] . fn add_to_hash(&mut self, i: usize) { 191,812 ( 0.04%) self.hash = self.hash.rotate_left(5).bitxor(i).wrapping_mul(K); . } . } . . impl Hasher for FxHasher { . #[inline] 2,308 ( 0.00%) fn write(&mut self, mut bytes: &[u8]) { . #[cfg(target_pointer_width = "32")] . let read_usize = |bytes: &[u8]| u32::from_ne_bytes(bytes[..4].try_into().unwrap()); . #[cfg(target_pointer_width = "64")] . let read_usize = |bytes: &[u8]| u64::from_ne_bytes(bytes[..8].try_into().unwrap()); . 58,569 ( 0.01%) let mut hash = FxHasher { hash: self.hash }; . assert!(size_of::() <= 8); 777,293 ( 0.16%) while bytes.len() >= size_of::() { . hash.add_to_hash(read_usize(bytes) as usize); . bytes = &bytes[size_of::()..]; . } 120,150 ( 0.03%) if (size_of::() > 4) && (bytes.len() >= 4) { . hash.add_to_hash(u32::from_ne_bytes(bytes[..4].try_into().unwrap()) as usize); . bytes = &bytes[4..]; . } 120,150 ( 0.03%) if (size_of::() > 2) && bytes.len() >= 2 { . hash.add_to_hash(u16::from_ne_bytes(bytes[..2].try_into().unwrap()) as usize); . bytes = &bytes[2..]; . } 133,662 ( 0.03%) if (size_of::() > 1) && bytes.len() >= 1 { 29,368 ( 0.01%) hash.add_to_hash(bytes[0] as usize); . } 3,760 ( 0.00%) self.hash = hash.hash; 2,308 ( 0.00%) } . . #[inline] . fn write_u8(&mut self, i: u8) { 189 ( 0.00%) self.add_to_hash(i as usize); . } . . #[inline] . fn write_u16(&mut self, i: u16) { 20 ( 0.00%) self.add_to_hash(i as usize); . } . . #[inline] . fn write_u32(&mut self, i: u32) { 108,287 ( 0.02%) self.add_to_hash(i as usize); . } . . #[cfg(target_pointer_width = "32")] . #[inline] . fn write_u64(&mut self, i: u64) { . self.add_to_hash(i as usize); . self.add_to_hash((i >> 32) as usize); . } -- line 131 ---------------------------------------- 673 ( 0.00%) -------------------------------------------------------------------------------- -- Auto-annotated source: /usr/home/liquid/rust/worktree-benchmarking/library/core/src/slice/mod.rs -------------------------------------------------------------------------------- Ir -- line 141 ---------------------------------------- . /// ``` . /// let a = [1, 2, 3]; . /// assert!(!a.is_empty()); . /// ``` . #[stable(feature = "rust1", since = "1.0.0")] . #[rustc_const_stable(feature = "const_slice_is_empty", since = "1.39.0")] . #[inline] . pub const fn is_empty(&self) -> bool { 47,024 ( 0.01%) self.len() == 0 . } . . /// Returns the first element of the slice, or `None` if it is empty. . /// . /// # Examples . /// . /// ``` . /// let v = [10, 40, 30]; -- line 157 ---------------------------------------- -- line 159 ---------------------------------------- . /// . /// let w: &[i32] = &[]; . /// assert_eq!(None, w.first()); . /// ``` . #[stable(feature = "rust1", since = "1.0.0")] . #[rustc_const_stable(feature = "const_slice_first_last_not_mut", since = "1.56.0")] . #[inline] . pub const fn first(&self) -> Option<&T> { 196 ( 0.00%) if let [first, ..] = self { Some(first) } else { None } . } . . /// Returns a mutable pointer to the first element of the slice, or `None` if it is empty. . /// . /// # Examples . /// . /// ``` . /// let x = &mut [0, 1, 2]; -- line 175 ---------------------------------------- -- line 178 ---------------------------------------- . /// *first = 5; . /// } . /// assert_eq!(x, &[5, 1, 2]); . /// ``` . #[stable(feature = "rust1", since = "1.0.0")] . #[rustc_const_unstable(feature = "const_slice_first_last", issue = "83570")] . #[inline] . pub const fn first_mut(&mut self) -> Option<&mut T> { 92 ( 0.00%) if let [first, ..] = self { Some(first) } else { None } . } . . /// Returns the first and all the rest of the elements of the slice, or `None` if it is empty. . /// . /// # Examples . /// . /// ``` . /// let x = &[0, 1, 2]; -- line 194 ---------------------------------------- -- line 197 ---------------------------------------- . /// assert_eq!(first, &0); . /// assert_eq!(elements, &[1, 2]); . /// } . /// ``` . #[stable(feature = "slice_splits", since = "1.5.0")] . #[rustc_const_stable(feature = "const_slice_first_last_not_mut", since = "1.56.0")] . #[inline] . pub const fn split_first(&self) -> Option<(&T, &[T])> { 4 ( 0.00%) if let [first, tail @ ..] = self { Some((first, tail)) } else { None } . } . . /// Returns the first and all the rest of the elements of the slice, or `None` if it is empty. . /// . /// # Examples . /// . /// ``` . /// let x = &mut [0, 1, 2]; -- line 213 ---------------------------------------- -- line 237 ---------------------------------------- . /// assert_eq!(last, &2); . /// assert_eq!(elements, &[0, 1]); . /// } . /// ``` . #[stable(feature = "slice_splits", since = "1.5.0")] . #[rustc_const_stable(feature = "const_slice_first_last_not_mut", since = "1.56.0")] . #[inline] . pub const fn split_last(&self) -> Option<(&T, &[T])> { 734 ( 0.00%) if let [init @ .., last] = self { Some((last, init)) } else { None } . } . . /// Returns the last and all the rest of the elements of the slice, or `None` if it is empty. . /// . /// # Examples . /// . /// ``` . /// let x = &mut [0, 1, 2]; -- line 253 ---------------------------------------- -- line 276 ---------------------------------------- . /// . /// let w: &[i32] = &[]; . /// assert_eq!(None, w.last()); . /// ``` . #[stable(feature = "rust1", since = "1.0.0")] . #[rustc_const_stable(feature = "const_slice_first_last_not_mut", since = "1.56.0")] . #[inline] . pub const fn last(&self) -> Option<&T> { 17,318 ( 0.00%) if let [.., last] = self { Some(last) } else { None } . } . . /// Returns a mutable pointer to the last item in the slice. . /// . /// # Examples . /// . /// ``` . /// let x = &mut [0, 1, 2]; -- line 292 ---------------------------------------- -- line 295 ---------------------------------------- . /// *last = 10; . /// } . /// assert_eq!(x, &[0, 1, 10]); . /// ``` . #[stable(feature = "rust1", since = "1.0.0")] . #[rustc_const_unstable(feature = "const_slice_first_last", issue = "83570")] . #[inline] . pub const fn last_mut(&mut self) -> Option<&mut T> { 25,759 ( 0.01%) if let [.., last] = self { Some(last) } else { None } . } . . /// Returns a reference to an element or subslice depending on the type of . /// index. . /// . /// - If given a position, returns a reference to the element at that . /// position or `None` if out of bounds. . /// - If given a range, returns the subslice corresponding to that range, -- line 311 ---------------------------------------- -- line 448 ---------------------------------------- . /// } . /// ``` . /// . /// [`as_mut_ptr`]: slice::as_mut_ptr . #[stable(feature = "rust1", since = "1.0.0")] . #[rustc_const_stable(feature = "const_slice_as_ptr", since = "1.32.0")] . #[inline] . pub const fn as_ptr(&self) -> *const T { 42,623 ( 0.01%) self as *const [T] as *const T . } . . /// Returns an unsafe mutable pointer to the slice's buffer. . /// . /// The caller must ensure that the slice outlives the pointer this . /// function returns, or else it will end up pointing to garbage. . /// . /// Modifying the container referenced by this slice may cause its buffer -- line 464 ---------------------------------------- -- line 476 ---------------------------------------- . /// } . /// } . /// assert_eq!(x, &[3, 4, 6]); . /// ``` . #[stable(feature = "rust1", since = "1.0.0")] . #[rustc_const_unstable(feature = "const_ptr_offset", issue = "71499")] . #[inline] . pub const fn as_mut_ptr(&mut self) -> *mut T { 5 ( 0.00%) self as *mut [T] as *mut T . } . . /// Returns the two raw pointers spanning the slice. . /// . /// The returned range is half-open, which means that the end pointer . /// points *one past* the last element of the slice. This way, an empty . /// slice is represented by two equal pointers, and the difference between . /// the two pointers represents the size of the slice. -- line 492 ---------------------------------------- -- line 582 ---------------------------------------- . /// v.swap(2, 4); . /// assert!(v == ["a", "b", "e", "d", "c"]); . /// ``` . #[stable(feature = "rust1", since = "1.0.0")] . #[rustc_const_unstable(feature = "const_swap", issue = "83163")] . #[inline] . #[track_caller] . pub const fn swap(&mut self, a: usize, b: usize) { 812 ( 0.00%) let _ = &self[a]; 1,919 ( 0.00%) let _ = &self[b]; . . // SAFETY: we just checked that both `a` and `b` are in bounds . unsafe { self.swap_unchecked(a, b) } . } . . /// Swaps two elements in the slice, without doing bounds checking. . /// . /// For a safe alternative see [`swap`]. -- line 599 ---------------------------------------- -- line 677 ---------------------------------------- . . // Because this function is first compiled in isolation, . // this check tells LLVM that the indexing below is . // in-bounds. Then after inlining -- once the actual . // lengths of the slices are known -- it's removed. . let (a, b) = (&mut a[..n], &mut b[..n]); . . for i in 0..n { 555 ( 0.00%) mem::swap(&mut a[i], &mut b[n - 1 - i]); . } . } . } . . /// Returns an iterator over the slice. . /// . /// # Examples . /// -- line 693 ---------------------------------------- -- line 1499 ---------------------------------------- . /// assert_eq!(left, [1, 2, 3, 4, 5, 6]); . /// assert_eq!(right, []); . /// } . /// ``` . #[stable(feature = "rust1", since = "1.0.0")] . #[inline] . #[track_caller] . pub fn split_at(&self, mid: usize) -> (&[T], &[T]) { 294 ( 0.00%) assert!(mid <= self.len()); . // SAFETY: `[ptr; mid]` and `[mid; len]` are inside `self`, which . // fulfills the requirements of `from_raw_parts_mut`. . unsafe { self.split_at_unchecked(mid) } . } . . /// Divides one mutable slice into two at an index. . /// . /// The first will contain all indices from `[0, mid)` (excluding -- line 1515 ---------------------------------------- -- line 1530 ---------------------------------------- . /// left[1] = 2; . /// right[1] = 4; . /// assert_eq!(v, [1, 2, 3, 4, 5, 6]); . /// ``` . #[stable(feature = "rust1", since = "1.0.0")] . #[inline] . #[track_caller] . pub fn split_at_mut(&mut self, mid: usize) -> (&mut [T], &mut [T]) { 2,246 ( 0.00%) assert!(mid <= self.len()); . // SAFETY: `[ptr; mid]` and `[mid; len]` are inside `self`, which . // fulfills the requirements of `from_raw_parts_mut`. . unsafe { self.split_at_mut_unchecked(mid) } . } . . /// Divides one slice into two at an index, without doing bounds checking. . /// . /// The first will contain all indices from `[0, mid)` (excluding -- line 1546 ---------------------------------------- -- line 1628 ---------------------------------------- . pub unsafe fn split_at_mut_unchecked(&mut self, mid: usize) -> (&mut [T], &mut [T]) { . let len = self.len(); . let ptr = self.as_mut_ptr(); . . // SAFETY: Caller has to check that `0 <= mid <= self.len()`. . // . // `[ptr; mid]` and `[mid; len]` are not overlapping, so returning a mutable reference . // is fine. 603 ( 0.00%) unsafe { (from_raw_parts_mut(ptr, mid), from_raw_parts_mut(ptr.add(mid), len - mid)) } . } . . /// Divides one slice into an array and a remainder slice at an index. . /// . /// The array will contain all indices from `[0, N)` (excluding . /// the index `N` itself) and the slice will contain all . /// indices from `[N, len)` (excluding the index `len` itself). . /// -- line 1644 ---------------------------------------- -- line 2113 ---------------------------------------- . /// assert!(!v.iter().any(|e| e == "hi")); . /// ``` . #[stable(feature = "rust1", since = "1.0.0")] . #[inline] . pub fn contains(&self, x: &T) -> bool . where . T: PartialEq, . { 32 ( 0.00%) cmp::SliceContains::slice_contains(x, self) . } . . /// Returns `true` if `needle` is a prefix of the slice. . /// . /// # Examples . /// . /// ``` . /// let v = [10, 40, 30]; -- line 2129 ---------------------------------------- -- line 2142 ---------------------------------------- . /// assert!(v.starts_with(&[])); . /// ``` . #[stable(feature = "rust1", since = "1.0.0")] . pub fn starts_with(&self, needle: &[T]) -> bool . where . T: PartialEq, . { . let n = needle.len(); 7,501 ( 0.00%) self.len() >= n && needle == &self[..n] . } . . /// Returns `true` if `needle` is a suffix of the slice. . /// . /// # Examples . /// . /// ``` . /// let v = [10, 40, 30]; -- line 2158 ---------------------------------------- -- line 2171 ---------------------------------------- . /// assert!(v.ends_with(&[])); . /// ``` . #[stable(feature = "rust1", since = "1.0.0")] . pub fn ends_with(&self, needle: &[T]) -> bool . where . T: PartialEq, . { . let (m, n) = (self.len(), needle.len()); 4,446 ( 0.00%) m >= n && needle == &self[m - n..] . } . . /// Returns a subslice with the prefix removed. . /// . /// If the slice starts with `prefix`, returns the subslice after the prefix, wrapped in `Some`. . /// If `prefix` is empty, simply returns the original slice. . /// . /// If the slice does not start with `prefix`, returns `None`. -- line 2187 ---------------------------------------- -- line 2293 ---------------------------------------- . /// s.insert(idx, num); . /// assert_eq!(s, [0, 1, 1, 1, 1, 2, 3, 5, 8, 13, 21, 34, 42, 55]); . /// ``` . #[stable(feature = "rust1", since = "1.0.0")] . pub fn binary_search(&self, x: &T) -> Result . where . T: Ord, . { 3 ( 0.00%) self.binary_search_by(|p| p.cmp(x)) . } . . /// Binary searches this sorted slice with a comparator function. . /// . /// The comparator function should implement an order consistent . /// with the sort order of the underlying slice, returning an . /// order code that indicates whether its argument is `Less`, . /// `Equal` or `Greater` the desired target. -- line 2309 ---------------------------------------- -- line 2345 ---------------------------------------- . #[inline] . pub fn binary_search_by<'a, F>(&'a self, mut f: F) -> Result . where . F: FnMut(&'a T) -> Ordering, . { . let mut size = self.len(); . let mut left = 0; . let mut right = size; 222,438 ( 0.05%) while left < right { 312,399 ( 0.07%) let mid = left + size / 2; . . // SAFETY: the call is made safe by the following invariants: . // - `mid >= 0` . // - `mid < size`: `mid` is limited by `[left; right)` bound. 180,764 ( 0.04%) let cmp = f(unsafe { self.get_unchecked(mid) }); . . // The reason why we use if/else control flow rather than match . // is because match reorders comparison operations, which is perf sensitive. . // This is x86 asm for u8: https://rust.godbolt.org/z/8Y8Pra. 101,418 ( 0.02%) if cmp == Less { 83,827 ( 0.02%) left = mid + 1; 62,655 ( 0.01%) } else if cmp == Greater { . right = mid; . } else { . // SAFETY: same as the `get_unchecked` above . unsafe { crate::intrinsics::assume(mid < self.len()) }; . return Ok(mid); . } . 279,354 ( 0.06%) size = right - left; . } . Err(left) . } . . /// Binary searches this sorted slice with a key extraction function. . /// . /// Assumes that the slice is sorted by the key, for instance with . /// [`sort_by_key`] using the same key extraction function. -- line 2382 ---------------------------------------- -- line 3203 ---------------------------------------- . #[track_caller] . fn len_mismatch_fail(dst_len: usize, src_len: usize) -> ! { . panic!( . "source slice length ({}) does not match destination slice length ({})", . src_len, dst_len, . ); . } . 6,028 ( 0.00%) if self.len() != src.len() { . len_mismatch_fail(self.len(), src.len()); . } . . // SAFETY: `self` is valid for `self.len()` elements by definition, and `src` was . // checked to have the same length. The slices cannot overlap because . // mutable references are exclusive. . unsafe { . ptr::copy_nonoverlapping(src.as_ptr(), self.as_mut_ptr(), self.len()); -- line 3219 ---------------------------------------- -- line 3382 ---------------------------------------- . } . let gcd: usize = gcd(mem::size_of::(), mem::size_of::()); . let ts: usize = mem::size_of::() / gcd; . let us: usize = mem::size_of::() / gcd; . . // Armed with this knowledge, we can find how many `U`s we can fit! . let us_len = self.len() / ts * us; . // And how many `T`s will be in the trailing slice! 984 ( 0.00%) let ts_len = self.len() % ts; . (us_len, ts_len) . } . . /// Transmute the slice to a slice of another type, ensuring alignment of the types is . /// maintained. . /// . /// This method splits the slice into three distinct slices: prefix, correctly aligned middle . /// slice of a new type, and the suffix slice. The method may make the middle slice the greatest -- line 3398 ---------------------------------------- -- line 3429 ---------------------------------------- . return (self, &[], &[]); . } . . // First, find at what point do we split between the first and 2nd slice. Easy with . // ptr.align_offset. . let ptr = self.as_ptr(); . // SAFETY: See the `align_to_mut` method for the detailed safety comment. . let offset = unsafe { crate::ptr::align_offset(ptr, mem::align_of::()) }; 984 ( 0.00%) if offset > self.len() { . (self, &[], &[]) . } else { . let (left, rest) = self.split_at(offset); . let (us_len, ts_len) = rest.align_to_offsets::(); . // SAFETY: now `rest` is definitely aligned, so `from_raw_parts` below is okay, . // since the caller guarantees that we can transmute `T` to `U` safely. . unsafe { . ( . left, . from_raw_parts(rest.as_ptr() as *const U, us_len), 984 ( 0.00%) from_raw_parts(rest.as_ptr().add(rest.len() - ts_len), ts_len), . ) . } . } . } . . /// Transmute the slice to a slice of another type, ensuring alignment of the types is . /// maintained. . /// -- line 3456 ---------------------------------------- 133,523 ( 0.03%) -------------------------------------------------------------------------------- -- Auto-annotated source: /usr/home/liquid/rust/worktree-benchmarking/compiler/rustc_serialize/src/opaque.rs -------------------------------------------------------------------------------- Ir -- line 14 ---------------------------------------- . . pub type EncodeResult = Result<(), !>; . . pub struct Encoder { . pub data: Vec, . } . . impl Encoder { 1 ( 0.00%) pub fn new(data: Vec) -> Encoder { 4 ( 0.00%) Encoder { data } 1 ( 0.00%) } . 1 ( 0.00%) pub fn into_inner(self) -> Vec { 4 ( 0.00%) self.data 1 ( 0.00%) } . . #[inline] . pub fn position(&self) -> usize { 1,196 ( 0.00%) self.data.len() . } . } . . macro_rules! write_leb128 { . ($enc:expr, $value:expr, $int_ty:ty, $fun:ident) => {{ . const MAX_ENCODED_LEN: usize = max_leb128_len!($int_ty); . let old_len = $enc.data.len(); . -- line 40 ---------------------------------------- -- line 67 ---------------------------------------- . . #[inline] . fn emit_unit(&mut self) -> EncodeResult { . Ok(()) . } . . #[inline] . fn emit_usize(&mut self, v: usize) -> EncodeResult { 44,577 ( 0.01%) write_leb128!(self, v, usize, write_usize_leb128) . } . . #[inline] . fn emit_u128(&mut self, v: u128) -> EncodeResult { 455 ( 0.00%) write_leb128!(self, v, u128, write_u128_leb128) . } . . #[inline] . fn emit_u64(&mut self, v: u64) -> EncodeResult { 511 ( 0.00%) write_leb128!(self, v, u64, write_u64_leb128) . } . . #[inline] . fn emit_u32(&mut self, v: u32) -> EncodeResult { 46,617 ( 0.01%) write_leb128!(self, v, u32, write_u32_leb128) . } . . #[inline] . fn emit_u16(&mut self, v: u16) -> EncodeResult { . self.data.extend_from_slice(&v.to_le_bytes()); . Ok(()) . } . -- line 98 ---------------------------------------- -- line 131 ---------------------------------------- . #[inline] . fn emit_i8(&mut self, v: i8) -> EncodeResult { . let as_u8: u8 = unsafe { std::mem::transmute(v) }; . self.emit_u8(as_u8) . } . . #[inline] . fn emit_bool(&mut self, v: bool) -> EncodeResult { 409 ( 0.00%) self.emit_u8(if v { 1 } else { 0 }) . } . . #[inline] . fn emit_f64(&mut self, v: f64) -> EncodeResult { . let as_u64: u64 = v.to_bits(); . self.emit_u64(as_u64) . } . -- line 147 ---------------------------------------- -- line 549 ---------------------------------------- . #[inline] . pub fn advance(&mut self, bytes: usize) { . self.position += bytes; . } . . #[inline] . pub fn read_raw_bytes(&mut self, bytes: usize) -> &'a [u8] { . let start = self.position; 24 ( 0.00%) self.position += bytes; 24 ( 0.00%) &self.data[start..self.position] . } . } . . macro_rules! read_leb128 { . ($dec:expr, $fun:ident) => {{ leb128::$fun($dec.data, &mut $dec.position) }}; . } . . impl<'a> serialize::Decoder for Decoder<'a> { . #[inline] . fn read_unit(&mut self) -> () { . () . } . . #[inline] . fn read_u128(&mut self) -> u128 { 962 ( 0.00%) read_leb128!(self, read_u128_leb128) . } . . #[inline] . fn read_u64(&mut self) -> u64 { 3,120 ( 0.00%) read_leb128!(self, read_u64_leb128) . } . . #[inline] 47,748 ( 0.01%) fn read_u32(&mut self) -> u32 { 208,026 ( 0.04%) read_leb128!(self, read_u32_leb128) 95,496 ( 0.02%) } . . #[inline] . fn read_u16(&mut self) -> u16 { 561,421 ( 0.12%) let bytes = [self.data[self.position], self.data[self.position + 1]]; 80,203 ( 0.02%) let value = u16::from_le_bytes(bytes); 160,416 ( 0.03%) self.position += 2; . value . } . . #[inline] . fn read_u8(&mut self) -> u8 { 2,198,166 ( 0.46%) let value = self.data[self.position]; 435,814 ( 0.09%) self.position += 1; . value . } . . #[inline] 219,592 ( 0.05%) fn read_usize(&mut self) -> usize { 219,596 ( 0.05%) read_leb128!(self, read_usize_leb128) 439,184 ( 0.09%) } . . #[inline] . fn read_i128(&mut self) -> i128 { . read_leb128!(self, read_i128_leb128) . } . . #[inline] . fn read_i64(&mut self) -> i64 { -- line 613 ---------------------------------------- -- line 637 ---------------------------------------- . #[inline] . fn read_isize(&mut self) -> isize { . read_leb128!(self, read_isize_leb128) . } . . #[inline] . fn read_bool(&mut self) -> bool { . let value = self.read_u8(); 14,557 ( 0.00%) value != 0 . } . . #[inline] . fn read_f64(&mut self) -> f64 { . let bits = self.read_u64(); . f64::from_bits(bits) . } . -- line 653 ---------------------------------------- -- line 660 ---------------------------------------- . #[inline] . fn read_char(&mut self) -> char { . let bits = self.read_u32(); . std::char::from_u32(bits).unwrap() . } . . #[inline] . fn read_str(&mut self) -> Cow<'_, str> { 39,116 ( 0.01%) let len = self.read_usize(); 218,388 ( 0.05%) let sentinel = self.data[self.position + len]; 54,600 ( 0.01%) assert!(sentinel == STR_SENTINEL); . let s = unsafe { . std::str::from_utf8_unchecked(&self.data[self.position..self.position + len]) . }; 136,500 ( 0.03%) self.position += len + 1; 35,211 ( 0.01%) Cow::Borrowed(s) . } . . #[inline] . fn read_raw_bytes_into(&mut self, s: &mut [u8]) { 824 ( 0.00%) let start = self.position; 3,968 ( 0.00%) self.position += s.len(); 824 ( 0.00%) s.copy_from_slice(&self.data[start..self.position]); . } . } . . // Specializations for contiguous byte sequences follow. The default implementations for slices . // encode and decode each element individually. This isn't necessary for `u8` slices when using . // opaque encoders and decoders, because each `u8` is unchanged by encoding and decoding. . // Therefore, we can use more efficient implementations that process the entire sequence at once. . -- line 690 ---------------------------------------- 544,012 ( 0.11%) -------------------------------------------------------------------------------- -- Auto-annotated source: /usr/home/liquid/rust/worktree-benchmarking/compiler/rustc_resolve/src/build_reduced_graph.rs -------------------------------------------------------------------------------- Ir -- line 36 ---------------------------------------- . . use std::cell::Cell; . use std::ptr; . use tracing::debug; . . type Res = def::Res; . . impl<'a> ToNameBinding<'a> for (Module<'a>, ty::Visibility, Span, LocalExpnId) { 8 ( 0.00%) fn to_name_binding(self, arenas: &'a ResolverArenas<'a>) -> &'a NameBinding<'a> { . arenas.alloc_name_binding(NameBinding { 1 ( 0.00%) kind: NameBindingKind::Module(self.0), . ambiguity: None, 3 ( 0.00%) vis: self.1, 1 ( 0.00%) span: self.2, 1 ( 0.00%) expansion: self.3, . }) 8 ( 0.00%) } . } . . impl<'a> ToNameBinding<'a> for (Res, ty::Visibility, Span, LocalExpnId) { . fn to_name_binding(self, arenas: &'a ResolverArenas<'a>) -> &'a NameBinding<'a> { 37,800 ( 0.01%) arenas.alloc_name_binding(NameBinding { 25,200 ( 0.01%) kind: NameBindingKind::Res(self.0, false), . ambiguity: None, . vis: self.1, . span: self.2, . expansion: self.3, . }) . } . } . -- line 66 ---------------------------------------- -- line 76 ---------------------------------------- . expansion: self.3, . }) . } . } . . impl<'a> Resolver<'a> { . /// Defines `name` in namespace `ns` of module `parent` to be `def` if it is not yet defined; . /// otherwise, reports an error. 81,817 ( 0.02%) crate fn define(&mut self, parent: Module<'a>, ident: Ident, ns: Namespace, def: T) . where . T: ToNameBinding<'a>, . { 48,753 ( 0.01%) let binding = def.to_name_binding(self.arenas); . let key = self.new_key(ident, ns); 75,152 ( 0.02%) if let Err(old_binding) = self.try_define(parent, key, binding) { . self.report_conflict(parent, ident, ns, old_binding, &binding); . } 54,544 ( 0.01%) } . . /// Walks up the tree of definitions starting at `def_id`, . /// stopping at the first encountered module. . /// Parent block modules for arbitrary def-ids are not recorded for the local crate, . /// and are not preserved in metadata for foreign crates, so block modules are never . /// returned by this function. . /// . /// For the local crate ignoring block modules may be incorrect, so use this method with care. -- line 101 ---------------------------------------- -- line 103 ---------------------------------------- . /// For foreign crates block modules can be ignored without introducing observable differences, . /// moreover they has to be ignored right now because they are not kept in metadata. . /// Foreign parent modules are used for resolving names used by foreign macros with def-site . /// hygiene, therefore block module ignorability relies on macros with def-site hygiene and . /// block module parents being unreachable from other crates. . /// Reachable macros with block module parents exist due to `#[macro_export] macro_rules!`, . /// but they cannot use def-site hygiene, so the assumption holds . /// (). 6,227 ( 0.00%) pub fn get_nearest_non_block_module(&mut self, mut def_id: DefId) -> Module<'a> { . loop { 2,839 ( 0.00%) match self.get_module(def_id) { . Some(module) => return module, . None => { . def_id.index = . self.def_key(def_id).parent.expect("non-root `DefId` without parent") . } . } . } 3,832 ( 0.00%) } . 1 ( 0.00%) pub fn expect_module(&mut self, def_id: DefId) -> Module<'a> { 1,340 ( 0.00%) self.get_module(def_id).expect("argument `DefId` is not a module") 2 ( 0.00%) } . . /// If `def_id` refers to a module (in resolver's sense, i.e. a module item, crate root, enum, . /// or trait), then this function returns that module's resolver representation, otherwise it . /// returns `None`. 11,600 ( 0.00%) crate fn get_module(&mut self, def_id: DefId) -> Option> { . if let module @ Some(..) = self.module_map.get(&def_id) { . return module.copied(); . } . 488 ( 0.00%) if !def_id.is_local() { 2,928 ( 0.00%) let def_kind = self.cstore().def_kind(def_id); 976 ( 0.00%) match def_kind { . DefKind::Mod | DefKind::Enum | DefKind::Trait => { 2,484 ( 0.00%) let def_key = self.cstore().def_key(def_id); . let parent = def_key.parent.map(|index| { 2,025 ( 0.00%) self.get_nearest_non_block_module(DefId { index, krate: def_id.krate }) . }); 414 ( 0.00%) let name = if def_id.index == CRATE_DEF_INDEX { 36 ( 0.00%) self.cstore().crate_name(def_id.krate) . } else { 1,215 ( 0.00%) def_key.disambiguated_data.data.get_opt_name().expect("module without name") . }; . 3,312 ( 0.00%) Some(self.new_module( . parent, . ModuleKind::Def(def_kind, def_id, name), 2,898 ( 0.00%) self.cstore().module_expansion_untracked(def_id, &self.session), 2,484 ( 0.00%) self.cstore().get_span_untracked(def_id, &self.session), . // FIXME: Account for `#[no_implicit_prelude]` attributes. 1,215 ( 0.00%) parent.map_or(false, |module| module.no_implicit_prelude), . )) . } . _ => None, . } . } else { . None . } 9,280 ( 0.00%) } . 693 ( 0.00%) crate fn expn_def_scope(&mut self, expn_id: ExpnId) -> Module<'a> { 231 ( 0.00%) match expn_id.expn_data().macro_def_id { 204 ( 0.00%) Some(def_id) => self.macro_def_scope(def_id), 9 ( 0.00%) None => expn_id . .as_local() . .and_then(|expn_id| self.ast_transform_scopes.get(&expn_id)) 9 ( 0.00%) .unwrap_or(&self.graph_root), . } 462 ( 0.00%) } . 74 ( 0.00%) crate fn macro_def_scope(&mut self, def_id: DefId) -> Module<'a> { 74 ( 0.00%) if let Some(id) = def_id.as_local() { . self.local_macro_def_scopes[&id] . } else { 148 ( 0.00%) self.get_nearest_non_block_module(def_id) . } . } . . crate fn get_macro(&mut self, res: Res) -> Option> { 82 ( 0.00%) match res { 28 ( 0.00%) Res::Def(DefKind::Macro(..), def_id) => Some(self.get_macro_by_def_id(def_id)), . Res::NonMacroAttr(_) => Some(self.non_macro_attr.clone()), . _ => None, . } . } . 88 ( 0.00%) crate fn get_macro_by_def_id(&mut self, def_id: DefId) -> Lrc { . if let Some(ext) = self.macro_map.get(&def_id) { 6 ( 0.00%) return ext.clone(); . } . 45 ( 0.00%) let ext = Lrc::new(match self.cstore().load_macro_untracked(def_id, &self.session) { 55 ( 0.00%) LoadedMacro::MacroDef(item, edition) => self.compile_macro(&item, edition), . LoadedMacro::ProcMacro(ext) => ext, . }); . . self.macro_map.insert(def_id, ext.clone()); . ext 77 ( 0.00%) } . . crate fn build_reduced_graph( . &mut self, . fragment: &AstFragment, . parent_scope: ParentScope<'a>, . ) -> MacroRulesScopeRef<'a> { 28 ( 0.00%) collect_definitions(self, fragment, parent_scope.expansion); 49 ( 0.00%) let mut visitor = BuildReducedGraphVisitor { r: self, parent_scope }; 14 ( 0.00%) fragment.visit_with(&mut visitor); 7 ( 0.00%) visitor.parent_scope.macro_rules . } . . crate fn build_reduced_graph_external(&mut self, module: Module<'a>) { 54,776 ( 0.01%) for child in self.cstore().module_children_untracked(module.def_id(), self.session) { . let parent_scope = ParentScope::module(module, self); 67,900 ( 0.01%) BuildReducedGraphVisitor { r: self, parent_scope } 54,320 ( 0.01%) .build_reduced_graph_for_external_crate_res(child); . } . } . } . . struct BuildReducedGraphVisitor<'a, 'b> { . r: &'b mut Resolver<'a>, . parent_scope: ParentScope<'a>, . } . . impl<'a> AsMut> for BuildReducedGraphVisitor<'a, '_> { . fn as_mut(&mut self) -> &mut Resolver<'a> { 1 ( 0.00%) self.r . } . } . . impl<'a, 'b> BuildReducedGraphVisitor<'a, 'b> { . fn resolve_visibility(&mut self, vis: &ast::Visibility) -> ty::Visibility { 223 ( 0.00%) self.resolve_visibility_speculative(vis, false).unwrap_or_else(|err| { . self.r.report_vis_error(err); . ty::Visibility::Public . }) . } . 540 ( 0.00%) fn resolve_visibility_speculative<'ast>( . &mut self, . vis: &'ast ast::Visibility, . speculative: bool, . ) -> Result> { 54 ( 0.00%) let parent_scope = &self.parent_scope; 270 ( 0.00%) match vis.kind { 22 ( 0.00%) ast::VisibilityKind::Public => Ok(ty::Visibility::Public), . ast::VisibilityKind::Crate(..) => { . Ok(ty::Visibility::Restricted(DefId::local(CRATE_DEF_INDEX))) . } . ast::VisibilityKind::Inherited => { 501 ( 0.00%) Ok(match self.parent_scope.module.kind { . // Any inherited visibility resolved directly inside an enum or trait . // (i.e. variants, fields, and trait items) inherits from the visibility . // of the enum or trait. 18 ( 0.00%) ModuleKind::Def(DefKind::Enum | DefKind::Trait, def_id, _) => { 45 ( 0.00%) self.r.visibilities[&def_id.expect_local()] . } . // Otherwise, the visibility is restricted to the nearest parent `mod` item. 68 ( 0.00%) _ => ty::Visibility::Restricted(self.parent_scope.module.nearest_parent_mod()), . }) . } . ast::VisibilityKind::Restricted { ref path, id, .. } => { . // For visibilities we are not ready to provide correct implementation of "uniform . // paths" right now, so on 2018 edition we only allow module-relative paths for now. . // On 2015 edition visibilities are resolved as crate-relative by default, . // so we are prepending a root segment if necessary. . let ident = path.segments.get(0).expect("empty path in visibility").ident; -- line 272 ---------------------------------------- -- line 326 ---------------------------------------- . } . PathResult::Failed { span, label, suggestion, .. } => { . Err(VisResolutionError::FailedToResolve(span, label, suggestion)) . } . PathResult::Indeterminate => Err(VisResolutionError::Indeterminate(path.span)), . } . } . } 432 ( 0.00%) } . . fn insert_field_names_local(&mut self, def_id: DefId, vdata: &ast::VariantData) { 12 ( 0.00%) let field_names = vdata . .fields() . .iter() 6 ( 0.00%) .map(|field| respan(field.span, field.ident.map_or(kw::Empty, |ident| ident.name))) . .collect(); . self.insert_field_names(def_id, field_names); . } . . fn insert_field_names(&mut self, def_id: DefId, field_names: Vec>) { . self.r.field_names.insert(def_id, field_names); . } . . fn block_needs_anonymous_module(&mut self, block: &Block) -> bool { . // If any statements are items, we need to create an anonymous module . block . .stmts . .iter() 82 ( 0.00%) .any(|statement| matches!(statement.kind, StmtKind::Item(_) | StmtKind::MacCall(_))) . } . . // Add an import to the current module. 221 ( 0.00%) fn add_import( . &mut self, . module_path: Vec, . kind: ImportKind<'a>, . span: Span, . id: NodeId, . item: &ast::Item, . root_span: Span, . root_id: NodeId, . vis: ty::Visibility, . ) { 51 ( 0.00%) let current_module = self.parent_scope.module; 476 ( 0.00%) let import = self.r.arenas.alloc_import(Import { . kind, . parent_scope: self.parent_scope, 68 ( 0.00%) module_path, . imported_module: Cell::new(None), . span, . id, 34 ( 0.00%) use_span: item.span, 51 ( 0.00%) use_span_with_attributes: item.span_with_attributes(), . has_attributes: !item.attrs.is_empty(), . root_span, . root_id, . vis: Cell::new(vis), . used: Cell::new(false), . }); . . self.r.indeterminate_imports.push(import); 91 ( 0.00%) match import.kind { . // Don't add unresolved underscore imports to modules . ImportKind::Single { target: Ident { name: kw::Underscore, .. }, .. } => {} 98 ( 0.00%) ImportKind::Single { target, type_ns_only, .. } => { 392 ( 0.00%) self.r.per_ns(|this, ns| { 126 ( 0.00%) if !type_ns_only || ns == TypeNS { 126 ( 0.00%) let key = this.new_key(target, ns); 336 ( 0.00%) let mut resolution = this.resolution(current_module, key).borrow_mut(); 42 ( 0.00%) resolution.add_single_import(import); . } 210 ( 0.00%) }); . } . // We don't add prelude imports to the globs since they only affect lexical scopes, . // which are not relevant to import resolution. . ImportKind::Glob { is_prelude: true, .. } => {} . ImportKind::Glob { .. } => current_module.globs.borrow_mut().push(import), . _ => unreachable!(), . } 136 ( 0.00%) } . 162 ( 0.00%) fn build_reduced_graph_for_use_tree( . &mut self, . // This particular use tree . use_tree: &ast::UseTree, . id: NodeId, . parent_prefix: &[Segment], . nested: bool, . // The whole `use` item . item: &Item, -- line 415 ---------------------------------------- -- line 426 ---------------------------------------- . .cloned() . .chain(use_tree.prefix.segments.iter().map(|seg| seg.into())) . .peekable(); . . // On 2015 edition imports are resolved as crate-relative by default, . // so prefixes are prepended with crate root segment if necessary. . // The root is prepended lazily, when the first non-empty prefix or terminating glob . // appears, so imports in braced groups can have roots prepended independently. 18 ( 0.00%) let is_glob = matches!(use_tree.kind, ast::UseTreeKind::Glob); 16 ( 0.00%) let crate_root = match prefix_iter.peek() { 212 ( 0.00%) Some(seg) if !seg.ident.is_path_segment_keyword() && seg.ident.span.rust_2015() => { . Some(seg.ident.span.ctxt()) . } . None if is_glob && use_tree.span.rust_2015() => Some(use_tree.span.ctxt()), . _ => None, . } . .map(|ctxt| { . Segment::from_ident(Ident::new( . kw::PathRoot, . use_tree.prefix.span.shrink_to_lo().with_ctxt(ctxt), . )) . }); . 90 ( 0.00%) let prefix = crate_root.into_iter().chain(prefix_iter).collect::>(); . debug!("build_reduced_graph_for_use_tree: prefix={:?}", prefix); . . let empty_for_self = |prefix: &[Segment]| { 2 ( 0.00%) prefix.is_empty() || prefix.len() == 1 && prefix[0].ident.name == kw::PathRoot . }; 66 ( 0.00%) match use_tree.kind { 70 ( 0.00%) ast::UseTreeKind::Simple(rename, id1, id2) => { 14 ( 0.00%) let mut ident = use_tree.ident(); 56 ( 0.00%) let mut module_path = prefix; . let mut source = module_path.pop().unwrap(); . let mut type_ns_only = false; . 28 ( 0.00%) if nested { . // Correctly handle `self` 2 ( 0.00%) if source.ident.name == kw::SelfLower { . type_ns_only = true; . . if empty_for_self(&module_path) { . self.r.report_error( . use_tree.span, . ResolutionError::SelfImportOnlyInImportListWithNonEmptyPrefix, . ); . return; -- line 472 ---------------------------------------- -- line 475 ---------------------------------------- . // Replace `use foo::{ self };` with `use foo;` . source = module_path.pop().unwrap(); . if rename.is_none() { . ident = source.ident; . } . } . } else { . // Disallow `self` 12 ( 0.00%) if source.ident.name == kw::SelfLower { . let parent = module_path.last(); . . let span = match parent { . // only `::self` from `use foo::self as bar` . Some(seg) => seg.ident.span.shrink_to_hi().to(source.ident.span), . None => source.ident.span, . }; . let span_with_rename = match rename { -- line 491 ---------------------------------------- -- line 506 ---------------------------------------- . source = parent; . if rename.is_none() { . ident = source.ident; . } . } . } . . // Disallow `use $crate;` 12 ( 0.00%) if source.ident.name == kw::DollarCrate && module_path.is_empty() { . let crate_root = self.r.resolve_crate_root(source.ident); . let crate_name = match crate_root.kind { . ModuleKind::Def(.., name) => name, . ModuleKind::Block(..) => unreachable!(), . }; . // HACK(eddyb) unclear how good this is, but keeping `$crate` . // in `source` breaks `src/test/ui/imports/import-crate-var.rs`, . // while the current crate doesn't have a valid `crate_name`. -- line 522 ---------------------------------------- -- line 535 ---------------------------------------- . . self.r . .session . .struct_span_err(item.span, "`$crate` may not be imported") . .emit(); . } . } . 26 ( 0.00%) if ident.name == kw::Crate { . self.r.session.span_err( . ident.span, . "crate root imports need to be explicitly named: \ . `use crate as name;`", . ); . } . . let kind = ImportKind::Single { 70 ( 0.00%) source: source.ident, . target: ident, . source_bindings: PerNS { . type_ns: Cell::new(Err(Determinacy::Undetermined)), . value_ns: Cell::new(Err(Determinacy::Undetermined)), . macro_ns: Cell::new(Err(Determinacy::Undetermined)), . }, . target_bindings: PerNS { . type_ns: Cell::new(None), -- line 560 ---------------------------------------- -- line 562 ---------------------------------------- . macro_ns: Cell::new(None), . }, . type_ns_only, . nested, . additional_ids: (id1, id2), . }; . . self.add_import( 56 ( 0.00%) module_path, 336 ( 0.00%) kind, 28 ( 0.00%) use_tree.span, . id, . item, . root_span, 56 ( 0.00%) item.id, . vis, . ); . } . ast::UseTreeKind::Glob => { . let kind = ImportKind::Glob { 15 ( 0.00%) is_prelude: self.r.session.contains_name(&item.attrs, sym::prelude_import), . max_vis: Cell::new(ty::Visibility::Invisible), . }; 45 ( 0.00%) self.add_import(prefix, kind, use_tree.span, id, item, root_span, item.id, vis); . } . ast::UseTreeKind::Nested(ref items) => { . // Ensure there is at most one `self` in the list . let self_spans = items . .iter() . .filter_map(|&(ref use_tree, _)| { 7 ( 0.00%) if let ast::UseTreeKind::Simple(..) = use_tree.kind { 6 ( 0.00%) if use_tree.ident().name == kw::SelfLower { . return Some(use_tree.span); . } . } . . None . }) . .collect::>(); 3 ( 0.00%) if self_spans.len() > 1 { . let mut e = self.r.into_struct_error( . self_spans[0], . ResolutionError::SelfImportCanOnlyAppearOnceInTheList, . ); . . for other_span in self_spans.iter().skip(1) { . e.span_label(*other_span, "another `self` import appears here"); . } . . e.emit(); . } . 4 ( 0.00%) for &(ref tree, id) in items { 30 ( 0.00%) self.build_reduced_graph_for_use_tree( . // This particular use tree . tree, id, &prefix, true, // The whole `use` item . item, vis, root_span, . ); . } . . // Empty groups `a::b::{}` are turned into synthetic `self` imports . // `a::b::c::{self as _}`, so that their prefixes are correctly . // resolved and checked for privacy/stability/etc. 1 ( 0.00%) if items.is_empty() && !empty_for_self(&prefix) { . let new_span = prefix[prefix.len() - 1].ident.span; . let tree = ast::UseTree { . prefix: ast::Path::from_ident(Ident::new(kw::SelfLower, new_span)), . kind: ast::UseTreeKind::Simple( . Some(Ident::new(kw::Underscore, new_span)), . ast::DUMMY_NODE_ID, . ast::DUMMY_NODE_ID, . ), -- line 633 ---------------------------------------- -- line 642 ---------------------------------------- . // The whole `use` item . item, . ty::Visibility::Invisible, . root_span, . ); . } . } . } 144 ( 0.00%) } . . /// Constructs the reduced graph for one item. . fn build_reduced_graph_for_item(&mut self, item: &'b Item) { . let parent_scope = &self.parent_scope; . let parent = parent_scope.module; 76 ( 0.00%) let expansion = parent_scope.expansion; 114 ( 0.00%) let ident = item.ident; 76 ( 0.00%) let sp = item.span; 76 ( 0.00%) let vis = self.resolve_visibility(&item.vis); 114 ( 0.00%) let local_def_id = self.r.local_def_id(item.id); . let def_id = local_def_id.to_def_id(); . . self.r.visibilities.insert(local_def_id, vis); . 208 ( 0.00%) match item.kind { 16 ( 0.00%) ItemKind::Use(ref use_tree) => { 176 ( 0.00%) self.build_reduced_graph_for_use_tree( . // This particular use tree . use_tree, 16 ( 0.00%) item.id, . &[], . false, . // The whole `use` item . item, . vis, . use_tree.span, . ); . } . 1 ( 0.00%) ItemKind::ExternCrate(orig_name) => { . self.build_reduced_graph_for_extern_crate( . orig_name, . item, . local_def_id, . vis, . parent, . ); . } . . ItemKind::Mod(..) => { 32 ( 0.00%) let module = self.r.new_module( . Some(parent), . ModuleKind::Def(DefKind::Mod, def_id, ident.name), . expansion.to_expn_id(), 8 ( 0.00%) item.span, 8 ( 0.00%) parent.no_implicit_prelude 28 ( 0.00%) || self.r.session.contains_name(&item.attrs, sym::no_implicit_prelude), . ); 76 ( 0.00%) self.r.define(parent, ident, TypeNS, (module, vis, sp, expansion)); . . // Descend into the module. 8 ( 0.00%) self.parent_scope.module = module; . } . . // These items live in the value namespace. . ItemKind::Static(..) => { . let res = Res::Def(DefKind::Static, def_id); 8 ( 0.00%) self.r.define(parent, ident, ValueNS, (res, vis, sp, expansion)); . } . ItemKind::Const(..) => { . let res = Res::Def(DefKind::Const, def_id); . self.r.define(parent, ident, ValueNS, (res, vis, sp, expansion)); . } . ItemKind::Fn(..) => { . let res = Res::Def(DefKind::Fn, def_id); 105 ( 0.00%) self.r.define(parent, ident, ValueNS, (res, vis, sp, expansion)); . . // Functions introducing procedural macros reserve a slot . // in the macro namespace as well (see #52225). 20 ( 0.00%) self.define_macro(item); . } . . // These items live in the type namespace. . ItemKind::TyAlias(..) => { . let res = Res::Def(DefKind::TyAlias, def_id); 8 ( 0.00%) self.r.define(parent, ident, TypeNS, (res, vis, sp, expansion)); . } . . ItemKind::Enum(_, _) => { 16 ( 0.00%) let module = self.r.new_module( . Some(parent), . ModuleKind::Def(DefKind::Enum, def_id, ident.name), . expansion.to_expn_id(), . item.span, 2 ( 0.00%) parent.no_implicit_prelude, . ); . self.r.define(parent, ident, TypeNS, (module, vis, sp, expansion)); . self.parent_scope.module = module; . } . . ItemKind::TraitAlias(..) => { . let res = Res::Def(DefKind::TraitAlias, def_id); . self.r.define(parent, ident, TypeNS, (res, vis, sp, expansion)); -- line 743 ---------------------------------------- -- line 825 ---------------------------------------- . fn build_reduced_graph_for_extern_crate( . &mut self, . orig_name: Option, . item: &Item, . local_def_id: LocalDefId, . vis: ty::Visibility, . parent: Module<'a>, . ) { 4 ( 0.00%) let ident = item.ident; 2 ( 0.00%) let sp = item.span; 2 ( 0.00%) let parent_scope = self.parent_scope; . let expansion = parent_scope.expansion; . 7 ( 0.00%) let (used, module, binding) = if orig_name.is_none() && ident.name == kw::SelfLower { . self.r . .session . .struct_span_err(item.span, "`extern crate self;` requires renaming") . .span_suggestion( . item.span, . "rename the `self` crate to be able to import it", . "extern crate self as name;".into(), . Applicability::HasPlaceholders, . ) . .emit(); . return; . } else if orig_name == Some(kw::SelfLower) { . Some(self.r.graph_root) . } else { 6 ( 0.00%) self.r.crate_loader.process_extern_crate(item, &self.r.definitions, local_def_id).map( 1 ( 0.00%) |crate_id| { 1 ( 0.00%) self.r.extern_crate_map.insert(local_def_id, crate_id); 6 ( 0.00%) self.r.expect_module(crate_id.as_def_id()) . }, . ) . } . .map(|module| { . let used = self.process_macro_use_imports(item, module); . let binding = 11 ( 0.00%) (module, ty::Visibility::Public, sp, expansion).to_name_binding(self.r.arenas); . (used, Some(ModuleOrUniformRoot::Module(module)), binding) . }) 2 ( 0.00%) .unwrap_or((true, None, self.r.dummy_binding)); 31 ( 0.00%) let import = self.r.arenas.alloc_import(Import { . kind: ImportKind::ExternCrate { source: orig_name, target: ident }, 1 ( 0.00%) root_id: item.id, . id: item.id, 6 ( 0.00%) parent_scope: self.parent_scope, . imported_module: Cell::new(module), . has_attributes: !item.attrs.is_empty(), 2 ( 0.00%) use_span_with_attributes: item.span_with_attributes(), 1 ( 0.00%) use_span: item.span, . root_span: item.span, . span: item.span, . module_path: Vec::new(), . vis: Cell::new(vis), . used: Cell::new(used), . }); 1 ( 0.00%) self.r.potentially_unused_imports.push(import); 5 ( 0.00%) let imported_binding = self.r.import(binding, import); 2 ( 0.00%) if ptr::eq(parent, self.r.graph_root) { 10 ( 0.00%) if let Some(entry) = self.r.extern_prelude.get(&ident.normalize_to_macros_2_0()) { . if expansion != LocalExpnId::ROOT . && orig_name.is_some() . && entry.extern_crate_item.is_none() . { . let msg = "macro-expanded `extern crate` items cannot \ . shadow names passed with `--extern`"; . self.r.session.span_err(item.span, msg); . } . } 6 ( 0.00%) let entry = self.r.extern_prelude.entry(ident.normalize_to_macros_2_0()).or_insert( . ExternPreludeEntry { extern_crate_item: None, introduced_by_item: true }, . ); 1 ( 0.00%) entry.extern_crate_item = Some(imported_binding); 1 ( 0.00%) if orig_name.is_some() { . entry.introduced_by_item = true; . } . } 6 ( 0.00%) self.r.define(parent, ident, TypeNS, imported_binding); . } . . /// Constructs the reduced graph for one foreign item. . fn build_reduced_graph_for_foreign_item(&mut self, item: &ForeignItem) { . let local_def_id = self.r.local_def_id(item.id); . let def_id = local_def_id.to_def_id(); . let (def_kind, ns) = match item.kind { . ForeignItemKind::Fn(..) => (DefKind::Fn, ValueNS), -- line 911 ---------------------------------------- -- line 918 ---------------------------------------- . let vis = self.resolve_visibility(&item.vis); . let res = Res::Def(def_kind, def_id); . self.r.define(parent, item.ident, ns, (res, vis, item.span, expansion)); . self.r.visibilities.insert(local_def_id, vis); . } . . fn build_reduced_graph_for_block(&mut self, block: &Block) { . let parent = self.parent_scope.module; 29 ( 0.00%) let expansion = self.parent_scope.expansion; . if self.block_needs_anonymous_module(block) { 6 ( 0.00%) let module = self.r.new_module( . Some(parent), 4 ( 0.00%) ModuleKind::Block(block.id), . expansion.to_expn_id(), . block.span, 2 ( 0.00%) parent.no_implicit_prelude, . ); . self.r.block_map.insert(block.id, module); 2 ( 0.00%) self.parent_scope.module = module; // Descend into the block. . } . } . . /// Builds the reduced graph for a single item in an external crate. 54,320 ( 0.01%) fn build_reduced_graph_for_external_crate_res(&mut self, child: ModChild) { 6,790 ( 0.00%) let parent = self.parent_scope.module; 40,740 ( 0.01%) let ModChild { ident, res, vis, span } = child; 33,950 ( 0.01%) let res = res.expect_non_local(); 13,580 ( 0.00%) let expansion = self.parent_scope.expansion; . // Record primary definitions. 33,950 ( 0.01%) match res { . Res::Def(DefKind::Mod | DefKind::Enum | DefKind::Trait, def_id) => { 511 ( 0.00%) let module = self.r.expect_module(def_id); 7,154 ( 0.00%) self.r.define(parent, ident, TypeNS, (module, vis, span, expansion)); . } . Res::Def( . DefKind::Struct . | DefKind::Union . | DefKind::Variant . | DefKind::TyAlias . | DefKind::ForeignTy . | DefKind::OpaqueTy . | DefKind::TraitAlias . | DefKind::AssocTy, . _, . ) . | Res::PrimTy(..) 20,682 ( 0.00%) | Res::ToolMod => self.r.define(parent, ident, TypeNS, (res, vis, span, expansion)), . Res::Def( . DefKind::Fn . | DefKind::AssocFn . | DefKind::Static . | DefKind::Const . | DefKind::AssocConst . | DefKind::Ctor(..), . _, 155,469 ( 0.03%) ) => self.r.define(parent, ident, ValueNS, (res, vis, span, expansion)), . Res::Def(DefKind::Macro(..), _) | Res::NonMacroAttr(..) => { 4,256 ( 0.00%) self.r.define(parent, ident, MacroNS, (res, vis, span, expansion)) . } . Res::Def( . DefKind::TyParam . | DefKind::ConstParam . | DefKind::ExternCrate . | DefKind::Use . | DefKind::ForeignMod . | DefKind::AnonConst -- line 983 ---------------------------------------- -- line 992 ---------------------------------------- . ) . | Res::Local(..) . | Res::SelfTy(..) . | Res::SelfCtor(..) . | Res::Err => bug!("unexpected resolution: {:?}", res), . } . // Record some extra data for better diagnostics. . let cstore = self.r.cstore(); 65,143 ( 0.01%) match res { 993 ( 0.00%) Res::Def(DefKind::Struct, def_id) => { . let field_names = 1,986 ( 0.00%) cstore.struct_field_names_untracked(def_id, self.r.session).collect(); 2,317 ( 0.00%) let ctor = cstore.ctor_def_id_and_kind_untracked(def_id); 662 ( 0.00%) if let Some((ctor_def_id, ctor_kind)) = ctor { . let ctor_res = Res::Def(DefKind::Ctor(CtorOf::Struct, ctor_kind), ctor_def_id); 120 ( 0.00%) let ctor_vis = cstore.visibility_untracked(ctor_def_id); . let field_visibilities = 96 ( 0.00%) cstore.struct_field_visibilities_untracked(def_id).collect(); . self.r . .struct_constructors . .insert(def_id, (ctor_res, ctor_vis, field_visibilities)); . } . self.insert_field_names(def_id, field_names); . } 6 ( 0.00%) Res::Def(DefKind::Union, def_id) => { . let field_names = 14 ( 0.00%) cstore.struct_field_names_untracked(def_id, self.r.session).collect(); . self.insert_field_names(def_id, field_names); . } 234 ( 0.00%) Res::Def(DefKind::AssocFn, def_id) => { 702 ( 0.00%) if cstore.fn_has_self_parameter_untracked(def_id) { . self.r.has_self.insert(def_id); . } . } . _ => {} . } 53,400 ( 0.01%) } . 408 ( 0.00%) fn add_macro_use_binding( . &mut self, . name: Symbol, . binding: &'a NameBinding<'a>, . span: Span, . allow_shadowing: bool, . ) { 204 ( 0.00%) if self.r.macro_use_prelude.insert(name, binding).is_some() && !allow_shadowing { . let msg = format!("`{}` is already in scope", name); . let note = . "macro-expanded `#[macro_use]`s may not shadow existing macros (see RFC 1560)"; . self.r.session.struct_span_err(span, &msg).note(note).emit(); . } 306 ( 0.00%) } . . /// Returns `true` if we should consider the underlying `extern crate` to be used. . fn process_macro_use_imports(&mut self, item: &Item, module: Module<'a>) -> bool { . let mut import_all = None; . let mut single_imports = Vec::new(); . for attr in &item.attrs { 1 ( 0.00%) if attr.has_name(sym::macro_use) { 2 ( 0.00%) if self.parent_scope.module.parent.is_some() { . struct_span_err!( . self.r.session, . item.span, . E0468, . "an `extern crate` loading macros must be at the crate root" . ) . .emit(); . } 4 ( 0.00%) if let ItemKind::ExternCrate(Some(orig_name)) = item.kind { . if orig_name == kw::SelfLower { . self.r . .session . .struct_span_err( . attr.span, . "`#[macro_use]` is not supported on `extern crate self`", . ) . .emit(); . } . } . let ill_formed = 3 ( 0.00%) |span| struct_span_err!(self.r.session, span, E0466, "bad macro import").emit(); 5 ( 0.00%) match attr.meta() { 22 ( 0.00%) Some(meta) => match meta.kind { . MetaItemKind::Word => { 2 ( 0.00%) import_all = Some(meta.span); . break; . } . MetaItemKind::List(nested_metas) => { . for nested_meta in nested_metas { . match nested_meta.ident() { . Some(ident) if nested_meta.is_word() => { . single_imports.push(ident) . } . _ => ill_formed(nested_meta.span()), . } . } . } . MetaItemKind::NameValue(..) => ill_formed(meta.span), 2 ( 0.00%) }, . None => ill_formed(attr.span), . } . } . } . . let macro_use_import = |this: &Self, span| { 20 ( 0.00%) this.r.arenas.alloc_import(Import { . kind: ImportKind::MacroUse, 1 ( 0.00%) root_id: item.id, . id: item.id, 6 ( 0.00%) parent_scope: this.parent_scope, . imported_module: Cell::new(Some(ModuleOrUniformRoot::Module(module))), 1 ( 0.00%) use_span_with_attributes: item.span_with_attributes(), . has_attributes: !item.attrs.is_empty(), 1 ( 0.00%) use_span: item.span, . root_span: span, . span, . module_path: Vec::new(), . vis: Cell::new(ty::Visibility::Restricted(DefId::local(CRATE_DEF_INDEX))), . used: Cell::new(false), . }) . }; . 4 ( 0.00%) let allow_shadowing = self.parent_scope.expansion == LocalExpnId::ROOT; 3 ( 0.00%) if let Some(span) = import_all { . let import = macro_use_import(self, span); 1 ( 0.00%) self.r.potentially_unused_imports.push(import); 9 ( 0.00%) module.for_each_child(self, |this, ident, ns, binding| { 195 ( 0.00%) if ns == MacroNS { 153 ( 0.00%) let imported_binding = this.r.import(binding, import); 408 ( 0.00%) this.add_macro_use_binding(ident.name, imported_binding, span, allow_shadowing); . } . }); . } else { . for ident in single_imports.iter().cloned() { . let result = self.r.resolve_ident_in_module( . ModuleOrUniformRoot::Module(module), . ident, . MacroNS, -- line 1129 ---------------------------------------- -- line 1146 ---------------------------------------- . .emit(); . } . } . } . import_all.is_some() || !single_imports.is_empty() . } . . /// Returns `true` if this attribute list contains `macro_use`. 25 ( 0.00%) fn contains_macro_use(&mut self, attrs: &[ast::Attribute]) -> bool { . for attr in attrs { 13 ( 0.00%) if attr.has_name(sym::macro_escape) { . let msg = "`#[macro_escape]` is a deprecated synonym for `#[macro_use]`"; . let mut err = self.r.session.struct_span_warn(attr.span, msg); . if let ast::AttrStyle::Inner = attr.style { . err.help("try an outer attribute: `#[macro_use]`").emit(); . } else { . err.emit(); . } . } else if !attr.has_name(sym::macro_use) { . continue; . } . 5 ( 0.00%) if !attr.is_word() { . self.r.session.span_err(attr.span, "arguments to `macro_use` are not allowed here"); . } . return true; . } . . false 35 ( 0.00%) } . . fn visit_invoc(&mut self, id: NodeId) -> LocalExpnId { 17 ( 0.00%) let invoc_id = id.placeholder_to_expn_id(); 4 ( 0.00%) let old_parent_scope = self.r.invocation_parent_scopes.insert(invoc_id, self.parent_scope); 5 ( 0.00%) assert!(old_parent_scope.is_none(), "invocation data is reset for an invocation"); . invoc_id . } . . /// Visit invocation in context in which it can emit a named item (possibly `macro_rules`) . /// directly into its parent scope's module. 24 ( 0.00%) fn visit_invoc_in_module(&mut self, id: NodeId) -> MacroRulesScopeRef<'a> { . let invoc_id = self.visit_invoc(id); 4 ( 0.00%) self.parent_scope.module.unexpanded_invocations.borrow_mut().insert(invoc_id); 12 ( 0.00%) self.r.arenas.alloc_macro_rules_scope(MacroRulesScope::Invocation(invoc_id)) 24 ( 0.00%) } . . fn proc_macro_stub(&self, item: &ast::Item) -> Option<(MacroKind, Ident, Span)> { 30 ( 0.00%) if self.r.session.contains_name(&item.attrs, sym::proc_macro) { . return Some((MacroKind::Bang, item.ident, item.span)); 25 ( 0.00%) } else if self.r.session.contains_name(&item.attrs, sym::proc_macro_attribute) { . return Some((MacroKind::Attr, item.ident, item.span)); 25 ( 0.00%) } else if let Some(attr) = self.r.session.find_by_name(&item.attrs, sym::proc_macro_derive) . { . if let Some(nested_meta) = attr.meta_item_list().and_then(|list| list.get(0).cloned()) { . if let Some(ident) = nested_meta.ident() { . return Some((MacroKind::Derive, ident, ident.span)); . } . } . } . None -- line 1205 ---------------------------------------- -- line 1209 ---------------------------------------- . // Macro uses will remove items from this set, and the remaining . // items will be reported as `unused_macros`. . fn insert_unused_macro(&mut self, ident: Ident, def_id: LocalDefId, node_id: NodeId) { . if !ident.as_str().starts_with('_') { . self.r.unused_macros.insert(def_id, (node_id, ident)); . } . } . 45 ( 0.00%) fn define_macro(&mut self, item: &ast::Item) -> MacroRulesScopeRef<'a> { 20 ( 0.00%) let parent_scope = self.parent_scope; . let expansion = parent_scope.expansion; 20 ( 0.00%) let def_id = self.r.local_def_id(item.id); 20 ( 0.00%) let (ext, ident, span, macro_rules) = match &item.kind { . ItemKind::MacroDef(def) => { . let ext = Lrc::new(self.r.compile_macro(item, self.r.session.edition())); . (ext, item.ident, item.span, def.macro_rules) . } . ItemKind::Fn(..) => match self.proc_macro_stub(item) { . Some((macro_kind, ident, span)) => { . self.r.proc_macro_stubs.insert(def_id); . (self.r.dummy_ext(macro_kind), ident, span, false) -- line 1229 ---------------------------------------- -- line 1276 ---------------------------------------- . }; . if vis != ty::Visibility::Public { . self.insert_unused_macro(ident, def_id, item.id); . } . self.r.define(module, ident, MacroNS, (res, vis, span, expansion)); . self.r.visibilities.insert(def_id, vis); . self.parent_scope.macro_rules . } 40 ( 0.00%) } . } . . macro_rules! method { . ($visit:ident: $ty:ty, $invoc:path, $walk:ident) => { 18 ( 0.00%) fn $visit(&mut self, node: &'b $ty) { 1,091 ( 0.00%) if let $invoc(..) = node.kind { 2 ( 0.00%) self.visit_invoc(node.id); . } else { 910 ( 0.00%) visit::$walk(self, node); . } . } . }; . } . . impl<'a, 'b> Visitor<'b> for BuildReducedGraphVisitor<'a, 'b> { . method!(visit_expr: ast::Expr, ast::ExprKind::MacCall, walk_expr); . method!(visit_pat: ast::Pat, ast::PatKind::MacCall, walk_pat); . method!(visit_ty: ast::Ty, ast::TyKind::MacCall, walk_ty); . 378 ( 0.00%) fn visit_item(&mut self, item: &'b Item) { 42 ( 0.00%) let orig_module_scope = self.parent_scope.module; 244 ( 0.00%) self.parent_scope.macro_rules = match item.kind { . ItemKind::MacroDef(..) => { . let macro_rules_scope = self.define_macro(item); . visit::walk_item(self, item); . macro_rules_scope . } . ItemKind::MacCall(..) => { 12 ( 0.00%) let macro_rules_scope = self.visit_invoc_in_module(item.id); . visit::walk_item(self, item); . macro_rules_scope . } . _ => { 76 ( 0.00%) let orig_macro_rules_scope = self.parent_scope.macro_rules; . self.build_reduced_graph_for_item(item); 114 ( 0.00%) visit::walk_item(self, item); 76 ( 0.00%) match item.kind { 24 ( 0.00%) ItemKind::Mod(..) if self.contains_macro_use(&item.attrs) => { . self.parent_scope.macro_rules . } . _ => orig_macro_rules_scope, . } . } . }; 42 ( 0.00%) self.parent_scope.module = orig_module_scope; 336 ( 0.00%) } . . fn visit_stmt(&mut self, stmt: &'b ast::Stmt) { 104 ( 0.00%) if let ast::StmtKind::MacCall(..) = stmt.kind { . self.parent_scope.macro_rules = self.visit_invoc_in_module(stmt.id); . } else { 156 ( 0.00%) visit::walk_stmt(self, stmt); . } . } . . fn visit_foreign_item(&mut self, foreign_item: &'b ForeignItem) { . if let ForeignItemKind::MacCall(_) = foreign_item.kind { . self.visit_invoc_in_module(foreign_item.id); . return; . } . . self.build_reduced_graph_for_foreign_item(foreign_item); . visit::walk_foreign_item(self, foreign_item); . } . 261 ( 0.00%) fn visit_block(&mut self, block: &'b Block) { 29 ( 0.00%) let orig_current_module = self.parent_scope.module; 58 ( 0.00%) let orig_current_macro_rules_scope = self.parent_scope.macro_rules; . self.build_reduced_graph_for_block(block); . visit::walk_block(self, block); 58 ( 0.00%) self.parent_scope.module = orig_current_module; 29 ( 0.00%) self.parent_scope.macro_rules = orig_current_macro_rules_scope; 232 ( 0.00%) } . 70 ( 0.00%) fn visit_assoc_item(&mut self, item: &'b AssocItem, ctxt: AssocCtxt) { 14 ( 0.00%) if let AssocItemKind::MacCall(_) = item.kind { . match ctxt { . AssocCtxt::Trait => { . self.visit_invoc_in_module(item.id); . } . AssocCtxt::Impl => { . self.visit_invoc(item.id); . } . } . return; . } . 14 ( 0.00%) let vis = self.resolve_visibility(&item.vis); 21 ( 0.00%) let local_def_id = self.r.local_def_id(item.id); . let def_id = local_def_id.to_def_id(); . 42 ( 0.00%) if !(ctxt == AssocCtxt::Impl . && matches!(item.vis.kind, ast::VisibilityKind::Inherited) 7 ( 0.00%) && self . .r . .trait_impl_items 7 ( 0.00%) .contains(&ty::DefIdTree::parent(&*self.r, def_id).unwrap().expect_local())) . { . // Trait impl item visibility is inherited from its trait when not specified . // explicitly. In that case we cannot determine it here in early resolve, . // so we leave a hole in the visibility table to be filled later. . self.r.visibilities.insert(local_def_id, vis); . } . 14 ( 0.00%) if ctxt == AssocCtxt::Trait { . let (def_kind, ns) = match item.kind { . AssocItemKind::Const(..) => (DefKind::AssocConst, ValueNS), . AssocItemKind::Fn(box Fn { ref sig, .. }) => { . if sig.decl.has_self() { . self.r.has_self.insert(def_id); . } . (DefKind::AssocFn, ValueNS) . } -- line 1397 ---------------------------------------- -- line 1400 ---------------------------------------- . }; . . let parent = self.parent_scope.module; . let expansion = self.parent_scope.expansion; . let res = Res::Def(def_kind, def_id); . self.r.define(parent, item.ident, ns, (res, vis, item.span, expansion)); . } . 28 ( 0.00%) visit::walk_assoc_item(self, item, ctxt); 56 ( 0.00%) } . 612 ( 0.00%) fn visit_attribute(&mut self, attr: &'b ast::Attribute) { 464 ( 0.00%) if !attr.is_doc_comment() && attr::is_builtin_attr(attr) { 14 ( 0.00%) self.r . .builtin_attrs 168 ( 0.00%) .push((attr.get_normal_item().path.segments[0].ident, self.parent_scope)); . } 714 ( 0.00%) visit::walk_attribute(self, attr); . } . . fn visit_arm(&mut self, arm: &'b ast::Arm) { 80 ( 0.00%) if arm.is_placeholder { . self.visit_invoc(arm.id); . } else { 38 ( 0.00%) visit::walk_arm(self, arm); . } . } . . fn visit_expr_field(&mut self, f: &'b ast::ExprField) { . if f.is_placeholder { . self.visit_invoc(f.id); . } else { . visit::walk_expr_field(self, f); -- line 1432 ---------------------------------------- -- line 1437 ---------------------------------------- . if fp.is_placeholder { . self.visit_invoc(fp.id); . } else { . visit::walk_pat_field(self, fp); . } . } . . fn visit_generic_param(&mut self, param: &'b ast::GenericParam) { 7 ( 0.00%) if param.is_placeholder { . self.visit_invoc(param.id); . } else { 3 ( 0.00%) visit::walk_generic_param(self, param); . } . } . . fn visit_param(&mut self, p: &'b ast::Param) { 83 ( 0.00%) if p.is_placeholder { . self.visit_invoc(p.id); . } else { 42 ( 0.00%) visit::walk_param(self, p); . } . } . 24 ( 0.00%) fn visit_field_def(&mut self, sf: &'b ast::FieldDef) { 6 ( 0.00%) if sf.is_placeholder { . self.visit_invoc(sf.id); . } else { 6 ( 0.00%) let vis = self.resolve_visibility(&sf.vis); 9 ( 0.00%) self.r.visibilities.insert(self.r.local_def_id(sf.id), vis); 27 ( 0.00%) visit::walk_field_def(self, sf); . } . } . . // Constructs the reduced graph for one variant. Variants exist in the . // type and value namespaces. 54 ( 0.00%) fn visit_variant(&mut self, variant: &'b ast::Variant) { 12 ( 0.00%) if variant.is_placeholder { . self.visit_invoc_in_module(variant.id); . return; . } . 6 ( 0.00%) let parent = self.parent_scope.module; 12 ( 0.00%) let expn_id = self.parent_scope.expansion; 18 ( 0.00%) let ident = variant.ident; . . // Define a name in the type namespace. 12 ( 0.00%) let def_id = self.r.local_def_id(variant.id); . let res = Res::Def(DefKind::Variant, def_id.to_def_id()); 12 ( 0.00%) let vis = self.resolve_visibility(&variant.vis); 126 ( 0.00%) self.r.define(parent, ident, TypeNS, (res, vis, variant.span, expn_id)); . self.r.visibilities.insert(def_id, vis); . . // If the variant is marked as non_exhaustive then lower the visibility to within the crate. 36 ( 0.00%) let ctor_vis = if vis == ty::Visibility::Public 24 ( 0.00%) && self.r.session.contains_name(&variant.attrs, sym::non_exhaustive) . { . ty::Visibility::Restricted(DefId::local(CRATE_DEF_INDEX)) . } else { . vis . }; . . // Define a constructor name in the value namespace. . // Braced variants, unlike structs, generate unusable names in . // value namespace, they are reserved for possible future use. . // It's ok to use the variant's id as a ctor id since an . // error will be reported on any use of such resolution anyway. 18 ( 0.00%) let ctor_node_id = variant.data.ctor_id().unwrap_or(variant.id); . let ctor_def_id = self.r.local_def_id(ctor_node_id); 12 ( 0.00%) let ctor_kind = CtorKind::from_ast(&variant.data); . let ctor_res = Res::Def(DefKind::Ctor(CtorOf::Variant, ctor_kind), ctor_def_id.to_def_id()); 114 ( 0.00%) self.r.define(parent, ident, ValueNS, (ctor_res, ctor_vis, variant.span, expn_id)); 6 ( 0.00%) if ctor_def_id != def_id { . self.r.visibilities.insert(ctor_def_id, ctor_vis); . } . // Record field names for error reporting. . self.insert_field_names_local(ctor_def_id.to_def_id(), &variant.data); . 60 ( 0.00%) visit::walk_variant(self, variant); . } . . fn visit_crate(&mut self, krate: &'b ast::Crate) { 2 ( 0.00%) if krate.is_placeholder { . self.visit_invoc_in_module(krate.id); . } else { . visit::walk_crate(self, krate); 11 ( 0.00%) self.contains_macro_use(&krate.attrs); . } . } . } 23,006 ( 0.00%) -------------------------------------------------------------------------------- -- Auto-annotated source: /usr/home/liquid/rust/worktree-benchmarking/compiler/rustc_span/src/symbol.rs -------------------------------------------------------------------------------- Ir -- line 15 ---------------------------------------- . use std::str; . . use crate::{with_session_globals, Edition, Span, DUMMY_SP}; . . #[cfg(test)] . mod tests; . . // The proc macro code for this is in `compiler/rustc_macros/src/symbols.rs`. 8 ( 0.00%) symbols! { . // After modifying this list adjust `is_special`, `is_used_keyword`/`is_unused_keyword`, . // this should be rarely necessary though if the keywords are kept in alphabetic order. . Keywords { . // Special reserved identifiers used internally for elided lifetimes, . // unnamed method parameters, crate root module, error recovery etc. . Empty: "", . PathRoot: "{{root}}", . DollarCrate: "$crate", -- line 31 ---------------------------------------- -- line 1498 ---------------------------------------- . write_str, . writeln_macro, . x87_reg, . xer, . xmm_reg, . ymm_reg, . zmm_reg, . } 10 ( 0.00%) } . 3,862 ( 0.00%) #[derive(Copy, Clone, Eq, HashStable_Generic, Encodable, Decodable)] . pub struct Ident { . pub name: Symbol, . pub span: Span, . } . . impl Ident { . #[inline] . /// Constructs a new identifier from a symbol and a span. . pub const fn new(name: Symbol, span: Span) -> Ident { . Ident { name, span } 50,213 ( 0.01%) } . . /// Constructs a new identifier with a dummy span. . #[inline] . pub const fn with_dummy_span(name: Symbol) -> Ident { . Ident::new(name, DUMMY_SP) . } . . #[inline] . pub fn empty() -> Ident { . Ident::with_dummy_span(kw::Empty) . } . . /// Maps a string to an identifier with a dummy span. 3 ( 0.00%) pub fn from_str(string: &str) -> Ident { 3 ( 0.00%) Ident::with_dummy_span(Symbol::intern(string)) 12 ( 0.00%) } . . /// Maps a string and a span to an identifier. 14 ( 0.00%) pub fn from_str_and_span(string: &str, span: Span) -> Ident { 7 ( 0.00%) Ident::new(Symbol::intern(string), span) 14 ( 0.00%) } . . /// Replaces `lo` and `hi` with those from `span`, but keep hygiene context. 1,208 ( 0.00%) pub fn with_span_pos(self, span: Span) -> Ident { . Ident::new(self.name, span.with_ctxt(self.span.ctxt())) 755 ( 0.00%) } . . pub fn without_first_quote(self) -> Ident { . Ident::new(Symbol::intern(self.as_str().trim_start_matches('\'')), self.span) . } . . /// "Normalize" ident for use in comparisons using "item hygiene". . /// Identifiers with same string value become same if they came from the same macro 2.0 macro . /// (e.g., `macro` item, but not `macro_rules` item) and stay different if they came from . /// different macro 2.0 macros. . /// Technically, this operation strips all non-opaque marks from ident's syntactic context. 70,065 ( 0.01%) pub fn normalize_to_macros_2_0(self) -> Ident { . Ident::new(self.name, self.span.normalize_to_macros_2_0()) 54,495 ( 0.01%) } . . /// "Normalize" ident for use in comparisons using "local variable hygiene". . /// Identifiers with same string value become same if they came from the same non-transparent . /// macro (e.g., `macro` or `macro_rules!` items) and stay different if they came from different . /// non-transparent macros. . /// Technically, this operation strips all transparent marks from ident's syntactic context. 639 ( 0.00%) pub fn normalize_to_macro_rules(self) -> Ident { . Ident::new(self.name, self.span.normalize_to_macro_rules()) 497 ( 0.00%) } . . /// Access the underlying string. This is a slowish operation because it . /// requires locking the symbol interner. . /// . /// Note that the lifetime of the return value is a lie. See . /// `Symbol::as_str()` for details. . pub fn as_str(&self) -> &str { 17 ( 0.00%) self.name.as_str() . } . } . . impl PartialEq for Ident { 6,315 ( 0.00%) fn eq(&self, rhs: &Self) -> bool { 8,506 ( 0.00%) self.name == rhs.name && self.span.ctxt() == rhs.span.ctxt() 8,420 ( 0.00%) } . } . . impl Hash for Ident { . fn hash(&self, state: &mut H) { 3 ( 0.00%) self.name.hash(state); 680 ( 0.00%) self.span.ctxt().hash(state); . } . } . . impl fmt::Debug for Ident { . fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { . fmt::Display::fmt(self, f)?; . fmt::Debug::fmt(&self.span.ctxt(), f) . } . } . . /// This implementation is supposed to be used in error messages, so it's expected to be identical . /// to printing the original identifier token written in source code (`token_to_string`), . /// except that AST identifiers don't keep the rawness flag, so we have to guess it. . impl fmt::Display for Ident { 4 ( 0.00%) fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { 5 ( 0.00%) fmt::Display::fmt(&IdentPrinter::new(self.name, self.is_raw_guess(), None), f) 4 ( 0.00%) } . } . . /// This is the most general way to print identifiers. . /// AST pretty-printer is used as a fallback for turning AST structures into token streams for . /// proc macros. Additionally, proc macros may stringify their input and expect it survive the . /// stringification (especially true for proc macro derives written between Rust 1.15 and 1.30). . /// So we need to somehow pretty-print `$crate` in a way preserving at least some of its . /// hygiene data, most importantly name of the crate it refers to. -- line 1613 ---------------------------------------- -- line 1622 ---------------------------------------- . is_raw: bool, . /// Span used for retrieving the crate name to which `$crate` refers to, . /// if this field is `None` then the `$crate` conversion doesn't happen. . convert_dollar_crate: Option, . } . . impl IdentPrinter { . /// The most general `IdentPrinter` constructor. Do not use this. 23 ( 0.00%) pub fn new(symbol: Symbol, is_raw: bool, convert_dollar_crate: Option) -> IdentPrinter { 133 ( 0.00%) IdentPrinter { symbol, is_raw, convert_dollar_crate } 23 ( 0.00%) } . . /// This implementation is supposed to be used when printing identifiers . /// as a part of pretty-printing for larger AST pieces. . /// Do not use this either. 18 ( 0.00%) pub fn for_ast_ident(ident: Ident, is_raw: bool) -> IdentPrinter { 18 ( 0.00%) IdentPrinter::new(ident.name, is_raw, Some(ident.span)) 9 ( 0.00%) } . } . . impl fmt::Display for IdentPrinter { 165 ( 0.00%) fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { 66 ( 0.00%) if self.is_raw { . f.write_str("r#")?; 33 ( 0.00%) } else if self.symbol == kw::DollarCrate { . if let Some(span) = self.convert_dollar_crate { . let converted = span.ctxt().dollar_crate_name(); . if !converted.is_path_segment_keyword() { . f.write_str("::")?; . } . return fmt::Display::fmt(&converted, f); . } . } -- line 1654 ---------------------------------------- -- line 1656 ---------------------------------------- . } . } . . /// An newtype around `Ident` that calls [Ident::normalize_to_macro_rules] on . /// construction. . // FIXME(matthewj, petrochenkov) Use this more often, add a similar . // `ModernIdent` struct and use that as well. . #[derive(Copy, Clone, Eq, PartialEq, Hash)] 104 ( 0.00%) pub struct MacroRulesNormalizedIdent(Ident); . . impl MacroRulesNormalizedIdent { . pub fn new(ident: Ident) -> Self { 40 ( 0.00%) Self(ident.normalize_to_macro_rules()) . } . } . . impl fmt::Debug for MacroRulesNormalizedIdent { . fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { . fmt::Debug::fmt(&self.0, f) . } . } -- line 1676 ---------------------------------------- -- line 1686 ---------------------------------------- . /// Internally, a `Symbol` is implemented as an index, and all operations . /// (including hashing, equality, and ordering) operate on that index. The use . /// of `rustc_index::newtype_index!` means that `Option` only takes up 4 bytes, . /// because `rustc_index::newtype_index!` reserves the last 256 values for tagging purposes. . /// . /// Note that `Symbol` cannot directly be a `rustc_index::newtype_index!` because it . /// implements `fmt::Debug`, `Encodable`, and `Decodable` in special ways. . #[derive(Clone, Copy, PartialEq, Eq, PartialOrd, Ord, Hash)] 5,285 ( 0.00%) pub struct Symbol(SymbolIndex); . . rustc_index::newtype_index! { . struct SymbolIndex { .. } . } . . impl Symbol { . const fn new(n: u32) -> Self { . Symbol(SymbolIndex::from_u32(n)) 20 ( 0.00%) } . . /// Maps a string to its interned representation. 244,791 ( 0.05%) pub fn intern(string: &str) -> Self { . with_session_globals(|session_globals| session_globals.symbol_interner.intern(string)) 244,791 ( 0.05%) } . . /// Access the underlying string. This is a slowish operation because it . /// requires locking the symbol interner. . /// . /// Note that the lifetime of the return value is a lie. It's not the same . /// as `&self`, but actually tied to the lifetime of the underlying . /// interner. Interners are long-lived, and there are very few of them, and . /// this function is typically used for short-lived things, so in practice . /// it works out ok. 11,049 ( 0.00%) pub fn as_str(&self) -> &str { 3,683 ( 0.00%) with_session_globals(|session_globals| unsafe { . std::mem::transmute::<&str, &str>(session_globals.symbol_interner.get(*self)) . }) 14,732 ( 0.00%) } . . pub fn as_u32(self) -> u32 { . self.0.as_u32() . } . . pub fn is_empty(self) -> bool { . self == kw::Empty . } -- line 1730 ---------------------------------------- -- line 1740 ---------------------------------------- . . impl fmt::Debug for Symbol { . fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { . fmt::Debug::fmt(self.as_str(), f) . } . } . . impl fmt::Display for Symbol { 498 ( 0.00%) fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { 1,791 ( 0.00%) fmt::Display::fmt(self.as_str(), f) . } . } . . impl Encodable for Symbol { . fn encode(&self, s: &mut S) -> Result<(), S::Error> { 3,734 ( 0.00%) s.emit_str(self.as_str()) . } . } . . impl Decodable for Symbol { . #[inline] 21,142 ( 0.00%) fn decode(d: &mut D) -> Symbol { 67,847 ( 0.01%) Symbol::intern(&d.read_str()) . } . } . . impl HashStable for Symbol { . #[inline] 60 ( 0.00%) fn hash_stable(&self, hcx: &mut CTX, hasher: &mut StableHasher) { 1,904 ( 0.00%) self.as_str().hash_stable(hcx, hasher); 54 ( 0.00%) } . } . . impl ToStableHashKey for Symbol { . type KeyType = String; . #[inline] . fn to_stable_hash_key(&self, _: &CTX) -> String { . self.as_str().to_string() . } -- line 1778 ---------------------------------------- -- line 1785 ---------------------------------------- . // . // The `FxHashMap`+`Vec` pair could be replaced by `FxIndexSet`, but #75278 . // found that to regress performance up to 2% in some cases. This might be . // revisited after further improvements to `indexmap`. . // . // This type is private to prevent accidentally constructing more than one . // `Interner` on the same thread, which makes it easy to mixup `Symbol`s . // between `Interner`s. 2 ( 0.00%) #[derive(Default)] . struct InternerInner { . arena: DroplessArena, . names: FxHashMap<&'static str, Symbol>, . strings: Vec<&'static str>, . } . . impl Interner { . fn prefill(init: &[&'static str]) -> Self { 30 ( 0.00%) Interner(Lock::new(InternerInner { . strings: init.into(), . names: init.iter().copied().zip((0..).map(Symbol::new)).collect(), . ..Default::default() . })) . } . . #[inline] . fn intern(&self, string: &str) -> Symbol { . let mut inner = self.0.lock(); 31,430 ( 0.01%) if let Some(&name) = inner.names.get(string) { . return name; . } . 11,484 ( 0.00%) let name = Symbol::new(inner.strings.len() as u32); . . // SAFETY: we convert from `&str` to `&[u8]`, clone it into the arena, . // and immediately convert the clone back to `&[u8], all because there . // is no `inner.arena.alloc_str()` method. This is clearly safe. . let string: &str = . unsafe { str::from_utf8_unchecked(inner.arena.alloc_slice(string.as_bytes())) }; . . // SAFETY: we can extend the arena allocation to `'static` because we -- line 1824 ---------------------------------------- -- line 1831 ---------------------------------------- . // #91445 for details. . inner.names.insert(string, name); . name . } . . // Get the symbol as a string. `Symbol::as_str()` should be used in . // preference to this function. . fn get(&self, symbol: Symbol) -> &str { 11,049 ( 0.00%) self.0.lock().strings[symbol.0.as_usize()] . } . } . . // This module has a very short name because it's used a lot. . /// This module contains all the defined keyword `Symbol`s. . /// . /// Given that `kw` is imported, use them like `kw::keyword_name`. . /// For example `kw::Loop` or `kw::Break`. -- line 1847 ---------------------------------------- -- line 1863 ---------------------------------------- . . // Used from a macro in `librustc_feature/accepted.rs` . pub use super::kw::MacroRules as macro_rules; . . /// Get the symbol for an integer. . /// . /// The first few non-negative integers each have a static symbol and therefore . /// are fast. 32 ( 0.00%) pub fn integer + Copy + ToString>(n: N) -> Symbol { 14 ( 0.00%) if let Result::Ok(idx) = n.try_into() { 12 ( 0.00%) if idx < 10 { 40 ( 0.00%) return Symbol::new(super::SYMBOL_DIGITS_BASE + idx as u32); . } . } . Symbol::intern(&n.to_string()) . } . } . . impl Symbol { . fn is_special(self) -> bool { . self <= kw::Underscore . } . . fn is_used_keyword_always(self) -> bool { 69 ( 0.00%) self >= kw::As && self <= kw::While . } . . fn is_used_keyword_conditional(self, edition: impl FnOnce() -> Edition) -> bool { 2,403 ( 0.00%) (self >= kw::Async && self <= kw::Dyn) && edition() >= Edition::Edition2018 . } . . fn is_unused_keyword_always(self) -> bool { 69 ( 0.00%) self >= kw::Abstract && self <= kw::Yield . } . . fn is_unused_keyword_conditional(self, edition: impl FnOnce() -> Edition) -> bool { 23 ( 0.00%) self == kw::Try && edition() >= Edition::Edition2018 . } . . pub fn is_reserved(self, edition: impl Copy + FnOnce() -> Edition) -> bool { 4,165 ( 0.00%) self.is_special() . || self.is_used_keyword_always() . || self.is_unused_keyword_always() . || self.is_used_keyword_conditional(edition) . || self.is_unused_keyword_conditional(edition) . } . . /// A keyword or reserved identifier that can be used as a path segment. . pub fn is_path_segment_keyword(self) -> bool { 10,262 ( 0.00%) self == kw::Super . || self == kw::SelfLower . || self == kw::SelfUpper . || self == kw::Crate . || self == kw::PathRoot . || self == kw::DollarCrate . } . . /// Returns `true` if the symbol is `true` or `false`. . pub fn is_bool_lit(self) -> bool { 113 ( 0.00%) self == kw::True || self == kw::False 113 ( 0.00%) } . . /// Returns `true` if this symbol can be a raw identifier. . pub fn can_be_raw(self) -> bool { 264 ( 0.00%) self != kw::Empty && self != kw::Underscore && !self.is_path_segment_keyword() . } . } . . impl Ident { . // Returns `true` for reserved identifiers used internally for elided lifetimes, . // unnamed method parameters, crate root module, error recovery etc. . pub fn is_special(self) -> bool { . self.name.is_special() 23 ( 0.00%) } . . /// Returns `true` if the token is a keyword used in the language. 23 ( 0.00%) pub fn is_used_keyword(self) -> bool { . // Note: `span.edition()` is relatively expensive, don't call it unless necessary. 23 ( 0.00%) self.name.is_used_keyword_always() . || self.name.is_used_keyword_conditional(|| self.span.edition()) 46 ( 0.00%) } . . /// Returns `true` if the token is a keyword reserved for possible future use. 23 ( 0.00%) pub fn is_unused_keyword(self) -> bool { . // Note: `span.edition()` is relatively expensive, don't call it unless necessary. 23 ( 0.00%) self.name.is_unused_keyword_always() . || self.name.is_unused_keyword_conditional(|| self.span.edition()) 46 ( 0.00%) } . . /// Returns `true` if the token is either a special identifier or a keyword. 701 ( 0.00%) pub fn is_reserved(self) -> bool { . // Note: `span.edition()` is relatively expensive, don't call it unless necessary. . self.name.is_reserved(|| self.span.edition()) 1,402 ( 0.00%) } . . /// A keyword or reserved identifier that can be used as a path segment. . pub fn is_path_segment_keyword(self) -> bool { . self.name.is_path_segment_keyword() 1,466 ( 0.00%) } . . /// We see this identifier in a normal identifier position, like variable name or a type. . /// How was it written originally? Did it use the raw form? Let's try to guess. 131 ( 0.00%) pub fn is_raw_guess(self) -> bool { . self.name.can_be_raw() && self.is_reserved() 262 ( 0.00%) } . } 44,264 ( 0.01%) -------------------------------------------------------------------------------- -- Auto-annotated source: /usr/home/liquid/rust/worktree-benchmarking/compiler/rustc_metadata/src/rmeta/decoder.rs -------------------------------------------------------------------------------- Ir -- line 176 ---------------------------------------- . } . fn sess(self) -> Option<&'tcx Session> { . None . } . fn tcx(self) -> Option> { . None . } . 9,573 ( 0.00%) fn decoder(self, pos: usize) -> DecodeContext<'a, 'tcx> { 3,191 ( 0.00%) let tcx = self.tcx(); 332,074 ( 0.07%) DecodeContext { 36,045 ( 0.01%) opaque: opaque::Decoder::new(self.blob(), pos), 3,191 ( 0.00%) cdata: self.cdata(), . blob: self.blob(), 11,462 ( 0.00%) sess: self.sess().or(tcx.map(|tcx| tcx.sess)), . tcx, . last_source_file_index: 0, . lazy_state: LazyState::NoNode, . alloc_decoding_session: self . .cdata() 30,119 ( 0.01%) .map(|cdata| cdata.cdata.alloc_decoding_state.new_decoding_session()), . } 9,573 ( 0.00%) } . } . . impl<'a, 'tcx> Metadata<'a, 'tcx> for &'a MetadataBlob { . #[inline] . fn blob(self) -> &'a MetadataBlob { . self . } . } -- line 206 ---------------------------------------- -- line 255 ---------------------------------------- . } . #[inline] . fn tcx(self) -> Option> { . Some(self.1) . } . } . . impl<'a, 'tcx, T: Decodable>> Lazy { 33,922 ( 0.01%) fn decode>(self, metadata: M) -> T { 29,982 ( 0.01%) let mut dcx = metadata.decoder(self.position.get()); 95,853 ( 0.02%) dcx.lazy_state = LazyState::NodeStart(self.position); 35,569 ( 0.01%) T::decode(&mut dcx) 43,545 ( 0.01%) } . } . . impl<'a: 'x, 'tcx: 'x, 'x, T: Decodable>> Lazy<[T]> { . fn decode>( . self, . metadata: M, . ) -> impl ExactSizeIterator + Captures<'a> + Captures<'tcx> + 'x { 56 ( 0.00%) let mut dcx = metadata.decoder(self.position.get()); 16 ( 0.00%) dcx.lazy_state = LazyState::NodeStart(self.position); 26,523 ( 0.01%) (0..self.meta).map(move |_| T::decode(&mut dcx)) . } . } . . impl<'a, 'tcx> DecodeContext<'a, 'tcx> { . #[inline] . fn tcx(&self) -> TyCtxt<'tcx> { . debug_assert!(self.tcx.is_some(), "missing TyCtxt in DecodeContext"); . self.tcx.unwrap() -- line 285 ---------------------------------------- -- line 292 ---------------------------------------- . . #[inline] . pub fn cdata(&self) -> CrateMetadataRef<'a> { . debug_assert!(self.cdata.is_some(), "missing CrateMetadata in DecodeContext"); . self.cdata.unwrap() . } . . fn map_encoded_cnum_to_current(&self, cnum: CrateNum) -> CrateNum { 117,199 ( 0.02%) if cnum == LOCAL_CRATE { self.cdata().cnum } else { self.cdata().cnum_map[cnum] } . } . 36,005 ( 0.01%) fn read_lazy_with_meta(&mut self, meta: T::Meta) -> Lazy { . let distance = self.read_usize(); 34,311 ( 0.01%) let position = match self.lazy_state { . LazyState::NoNode => bug!("read_lazy_with_meta: outside of a metadata node"), 5,502 ( 0.00%) LazyState::NodeStart(start) => { . let start = start.get(); 11,004 ( 0.00%) assert!(distance <= start); . start - distance . } 10,683 ( 0.00%) LazyState::Previous(last_pos) => last_pos.get() + distance, . }; 18,126 ( 0.00%) self.lazy_state = LazyState::Previous(NonZeroUsize::new(position).unwrap()); . Lazy::from_position_and_meta(NonZeroUsize::new(position).unwrap(), meta) 36,005 ( 0.01%) } . . #[inline] . pub fn read_raw_bytes(&mut self, len: usize) -> &'a [u8] { . self.opaque.read_raw_bytes(len) . } . } . . impl<'a, 'tcx> TyDecoder<'tcx> for DecodeContext<'a, 'tcx> { -- line 324 ---------------------------------------- -- line 326 ---------------------------------------- . . #[inline] . fn tcx(&self) -> TyCtxt<'tcx> { . self.tcx.expect("missing TyCtxt in DecodeContext") . } . . #[inline] . fn peek_byte(&self) -> u8 { 61,345 ( 0.01%) self.opaque.data[self.opaque.position()] . } . . #[inline] . fn position(&self) -> usize { . self.opaque.position() . } . 51,912 ( 0.01%) fn cached_ty_for_shorthand(&mut self, shorthand: usize, or_insert_with: F) -> Ty<'tcx> . where . F: FnOnce(&mut Self) -> Ty<'tcx>, . { 7,416 ( 0.00%) let tcx = self.tcx(); . 7,416 ( 0.00%) let key = ty::CReaderCacheKey { cnum: Some(self.cdata().cnum), pos: shorthand }; . 20,052 ( 0.00%) if let Some(&ty) = tcx.ty_rcache.borrow().get(&key) { . return ty; . } . 2,403 ( 0.00%) let ty = or_insert_with(self); . tcx.ty_rcache.borrow_mut().insert(key, ty); . ty 66,744 ( 0.01%) } . . fn with_position(&mut self, pos: usize, f: F) -> R . where . F: FnOnce(&mut Self) -> R, . { . let new_opaque = opaque::Decoder::new(self.opaque.data, pos); . let old_opaque = mem::replace(&mut self.opaque, new_opaque); . let old_state = mem::replace(&mut self.lazy_state, LazyState::NoNode); . let r = f(self); 12,440 ( 0.00%) self.opaque = old_opaque; 2,488 ( 0.00%) self.lazy_state = old_state; . r . } . . fn decode_alloc_id(&mut self) -> rustc_middle::mir::interpret::AllocId { . if let Some(alloc_decoding_session) = self.alloc_decoding_session { . alloc_decoding_session.decode_alloc_id(self) . } else { . bug!("Attempting to decode interpret::AllocId without CrateMetadata") . } . } . } . . impl<'a, 'tcx> Decodable> for CrateNum { . fn decode(d: &mut DecodeContext<'a, 'tcx>) -> CrateNum { 64,727 ( 0.01%) let cnum = CrateNum::from_u32(d.read_u32()); . d.map_encoded_cnum_to_current(cnum) . } . } . . impl<'a, 'tcx> Decodable> for DefIndex { 1,287 ( 0.00%) fn decode(d: &mut DecodeContext<'a, 'tcx>) -> DefIndex { 111,951 ( 0.02%) DefIndex::from_u32(d.read_u32()) 2,574 ( 0.00%) } . } . . impl<'a, 'tcx> Decodable> for ExpnIndex { . fn decode(d: &mut DecodeContext<'a, 'tcx>) -> ExpnIndex { . ExpnIndex::from_u32(d.read_u32()) . } . } . . impl<'a, 'tcx> Decodable> for SyntaxContext { 116,020 ( 0.02%) fn decode(decoder: &mut DecodeContext<'a, 'tcx>) -> SyntaxContext { 29,005 ( 0.01%) let cdata = decoder.cdata(); 29,005 ( 0.01%) let sess = decoder.sess.unwrap(); 58,010 ( 0.01%) let cname = cdata.root.name; . rustc_span::hygiene::decode_syntax_context(decoder, &cdata.hygiene_context, |_, id| { . debug!("SpecializedDecoder: decoding {}", id); 732 ( 0.00%) cdata . .root . .syntax_contexts . .get(cdata, id) . .unwrap_or_else(|| panic!("Missing SyntaxContext {:?} for crate {:?}", id, cname)) . .decode((cdata, sess)) . }) 145,025 ( 0.03%) } . } . . impl<'a, 'tcx> Decodable> for ExpnId { 5,894 ( 0.00%) fn decode(decoder: &mut DecodeContext<'a, 'tcx>) -> ExpnId { 842 ( 0.00%) let local_cdata = decoder.cdata(); 842 ( 0.00%) let sess = decoder.sess.unwrap(); . . let cnum = CrateNum::decode(decoder); . let index = u32::decode(decoder); . . let expn_id = rustc_span::hygiene::decode_expn_id(cnum, index, |expn_id| { . let ExpnId { krate: cnum, local_id: index } = expn_id; . // Lookup local `ExpnData`s in our own crate data. Foreign `ExpnData`s . // are stored in the owning crate, to avoid duplication. . debug_assert_ne!(cnum, LOCAL_CRATE); 224 ( 0.00%) let crate_data = if cnum == local_cdata.cnum { . local_cdata . } else { . local_cdata.cstore.get_crate_data(cnum) . }; 1,120 ( 0.00%) let expn_data = crate_data . .root . .expn_data . .get(crate_data, index) . .unwrap() . .decode((crate_data, sess)); 1,120 ( 0.00%) let expn_hash = crate_data . .root . .expn_hashes . .get(crate_data, index) . .unwrap() . .decode((crate_data, sess)); 2,240 ( 0.00%) (expn_data, expn_hash) . }); . expn_id 6,736 ( 0.00%) } . } . . impl<'a, 'tcx> Decodable> for Span { 227,648 ( 0.05%) fn decode(decoder: &mut DecodeContext<'a, 'tcx>) -> Span { 28,456 ( 0.01%) let ctxt = SyntaxContext::decode(decoder); . let tag = u8::decode(decoder); . 56,912 ( 0.01%) if tag == TAG_PARTIAL_SPAN { . return DUMMY_SP.with_ctxt(ctxt); . } . . debug_assert!(tag == TAG_VALID_SPAN_LOCAL || tag == TAG_VALID_SPAN_FOREIGN); . . let lo = BytePos::decode(decoder); . let len = BytePos::decode(decoder); . let hi = lo + len; . 85,332 ( 0.02%) let Some(sess) = decoder.sess else { . bug!("Cannot decode Span without Session.") . }; . . // There are two possibilities here: . // 1. This is a 'local span', which is located inside a `SourceFile` . // that came from this crate. In this case, we use the source map data . // encoded in this crate. This branch should be taken nearly all of the time. . // 2. This is a 'foreign span', which is located inside a `SourceFile` -- line 476 ---------------------------------------- -- line 492 ---------------------------------------- . // a transitive dependency. . // . // When we encode a foreign span, we adjust its 'lo' and 'high' values . // to be based on the *foreign* crate (e.g. crate C), not the crate . // we are writing metadata for (e.g. crate B). This allows us to . // treat the 'local' and 'foreign' cases almost identically during deserialization: . // we can call `imported_source_files` for the proper crate, and binary search . // through the returned slice using our span. 73,925 ( 0.02%) let imported_source_files = if tag == TAG_VALID_SPAN_LOCAL { . decoder.cdata().imported_source_files(sess) . } else { . // When we encode a proc-macro crate, all `Span`s should be encoded . // with `TAG_VALID_SPAN_LOCAL` 154 ( 0.00%) if decoder.cdata().root.is_proc_macro_crate() { . // Decode `CrateNum` as u32 - using `CrateNum::decode` will ICE . // since we don't have `cnum_map` populated. . let cnum = u32::decode(decoder); . panic!( . "Decoding of crate {:?} tried to access proc-macro dep {:?}", . decoder.cdata().root.name, . cnum . ); -- line 513 ---------------------------------------- -- line 518 ---------------------------------------- . "SpecializedDecoder::specialized_decode: loading source files from cnum {:?}", . cnum . ); . . // Decoding 'foreign' spans should be rare enough that it's . // not worth it to maintain a per-CrateNum cache for `last_source_file_index`. . // We just set it to 0, to ensure that we don't try to access something out . // of bounds for our initial 'guess' 308 ( 0.00%) decoder.last_source_file_index = 0; . 154 ( 0.00%) let foreign_data = decoder.cdata().cstore.get_crate_data(cnum); . foreign_data.imported_source_files(sess) . }; . . let source_file = { . // Optimize for the case that most spans within a translated item . // originate from the same source_file. 85,332 ( 0.02%) let last_source_file = &imported_source_files[decoder.last_source_file_index]; . 108,604 ( 0.02%) if lo >= last_source_file.original_start_pos && lo <= last_source_file.original_end_pos . { . last_source_file . } else { . let index = imported_source_files . .binary_search_by_key(&lo, |source_file| source_file.original_start_pos) 34,072 ( 0.01%) .unwrap_or_else(|index| index - 1); . . // Don't try to cache the index for foreign spans, . // as this would require a map from CrateNums to indices 17,037 ( 0.00%) if tag == TAG_VALID_SPAN_LOCAL { 17,011 ( 0.00%) decoder.last_source_file_index = index; . } 34,074 ( 0.01%) &imported_source_files[index] . } . }; . . // Make sure our binary search above is correct. . debug_assert!( . lo >= source_file.original_start_pos && lo <= source_file.original_end_pos, . "Bad binary search: lo={:?} source_file.original_start_pos={:?} source_file.original_end_pos={:?}", . lo, -- line 558 ---------------------------------------- -- line 565 ---------------------------------------- . hi >= source_file.original_start_pos && hi <= source_file.original_end_pos, . "Bad binary search: hi={:?} source_file.original_start_pos={:?} source_file.original_end_pos={:?}", . hi, . source_file.original_start_pos, . source_file.original_end_pos . ); . . let lo = 113,776 ( 0.02%) (lo + source_file.translated_source_file.start_pos) - source_file.original_start_pos; . let hi = . (hi + source_file.translated_source_file.start_pos) - source_file.original_start_pos; . . // Do not try to decode parent for foreign spans. . Span::new(lo, hi, ctxt, None) 256,104 ( 0.05%) } . } . . impl<'a, 'tcx> Decodable> for &'tcx [thir::abstract_const::Node<'tcx>] { . fn decode(d: &mut DecodeContext<'a, 'tcx>) -> Self { . ty::codec::RefDecodable::decode(d) . } . } . -- line 587 ---------------------------------------- -- line 590 ---------------------------------------- . ty::codec::RefDecodable::decode(d) . } . } . . impl<'a, 'tcx, T: Decodable>> Decodable> . for Lazy . { . fn decode(decoder: &mut DecodeContext<'a, 'tcx>) -> Self { 9,820 ( 0.00%) decoder.read_lazy_with_meta(()) . } . } . . impl<'a, 'tcx, T: Decodable>> Decodable> . for Lazy<[T]> . { . fn decode(decoder: &mut DecodeContext<'a, 'tcx>) -> Self { . let len = decoder.read_usize(); 14,788 ( 0.00%) if len == 0 { Lazy::empty() } else { decoder.read_lazy_with_meta(len) } . } . } . . impl<'a, 'tcx, I: Idx, T: Decodable>> Decodable> . for Lazy> . where . Option: FixedSizeEncoding, . { . fn decode(decoder: &mut DecodeContext<'a, 'tcx>) -> Self { . let len = decoder.read_usize(); 12,337 ( 0.00%) decoder.read_lazy_with_meta(len) . } . } . . implement_ty_decoder!(DecodeContext<'a, 'tcx>); . . impl<'tcx> MetadataBlob { . crate fn new(metadata_ref: MetadataRef) -> MetadataBlob { . MetadataBlob(Lrc::new(metadata_ref)) -- line 626 ---------------------------------------- -- line 630 ---------------------------------------- . self.blob().starts_with(METADATA_HEADER) . } . . crate fn get_rustc_version(&self) -> String { . Lazy::::from_position(NonZeroUsize::new(METADATA_HEADER.len() + 4).unwrap()) . .decode(self) . } . 219 ( 0.00%) crate fn get_root(&self) -> CrateRoot<'tcx> { 73 ( 0.00%) let slice = &self.blob()[..]; . let offset = METADATA_HEADER.len(); 365 ( 0.00%) let pos = (((slice[offset + 0] as u32) << 24) 146 ( 0.00%) | ((slice[offset + 1] as u32) << 16) 292 ( 0.00%) | ((slice[offset + 2] as u32) << 8) 219 ( 0.00%) | ((slice[offset + 3] as u32) << 0)) as usize; . Lazy::>::from_position(NonZeroUsize::new(pos).unwrap()).decode(self) 365 ( 0.00%) } . . crate fn list_crate_metadata(&self, out: &mut dyn io::Write) -> io::Result<()> { . let root = self.get_root(); . writeln!(out, "Crate info:")?; . writeln!(out, "name {}{}", root.name, root.extra_filename)?; . writeln!(out, "hash {} stable_crate_id {:?}", root.hash, root.stable_crate_id)?; . writeln!(out, "proc_macro {:?}", root.proc_macro_data.is_some())?; . writeln!(out, "=External Dependencies=")?; -- line 654 ---------------------------------------- -- line 666 ---------------------------------------- . } . write!(out, "\n")?; . Ok(()) . } . } . . impl CrateRoot<'_> { . crate fn is_proc_macro_crate(&self) -> bool { 48 ( 0.00%) self.proc_macro_data.is_some() . } . . crate fn name(&self) -> Symbol { 4 ( 0.00%) self.name . } . . crate fn hash(&self) -> Svh { 24 ( 0.00%) self.hash . } . . crate fn stable_crate_id(&self) -> StableCrateId { 48 ( 0.00%) self.stable_crate_id . } . . crate fn triple(&self) -> &TargetTriple { . &self.triple . } . . crate fn decode_crate_deps<'a>( . &self, . metadata: &'a MetadataBlob, . ) -> impl ExactSizeIterator + Captures<'a> { 48 ( 0.00%) self.crate_deps.decode(metadata) . } . } . . impl<'a, 'tcx> CrateMetadataRef<'a> { . fn raw_proc_macro(self, id: DefIndex) -> &'a ProcMacro { . // DefIndex's in root.proc_macro_data have a one-to-one correspondence . // with items in 'raw_proc_macros'. . let pos = self -- line 705 ---------------------------------------- -- line 709 ---------------------------------------- . .unwrap() . .macros . .decode(self) . .position(|i| i == id) . .unwrap(); . &self.raw_proc_macros.unwrap()[pos] . } . 44,957 ( 0.01%) fn opt_item_ident(self, item_index: DefIndex, sess: &Session) -> Option { 34,366 ( 0.01%) let name = self.def_key(item_index).disambiguated_data.data.get_opt_name()?; 22,071 ( 0.00%) let span = match self.root.tables.ident_span.get(self, item_index) { . Some(lazy_span) => lazy_span.decode((self, sess)), . None => { . // FIXME: this weird case of a name with no span is specific to `extern crate` . // items, which are supposed to be treated like `use` items and only be encoded . // to metadata as `Export`s, return `None` because that's what all the callers . // expect in this case. 81 ( 0.00%) assert_eq!(self.def_kind(item_index), DefKind::ExternCrate); . return None; . } . }; . Some(Ident::new(name, span)) 57,886 ( 0.01%) } . 1,268 ( 0.00%) fn item_ident(self, item_index: DefIndex, sess: &Session) -> Ident { 5,305 ( 0.00%) self.opt_item_ident(item_index, sess).expect("no encoded ident for item") 2,536 ( 0.00%) } . . fn maybe_kind(self, item_id: DefIndex) -> Option { 10,364 ( 0.00%) self.root.tables.kind.get(self, item_id).map(|k| k.decode(self)) . } . 41,456 ( 0.01%) fn kind(self, item_id: DefIndex) -> EntryKind { . self.maybe_kind(item_id).unwrap_or_else(|| { . bug!( . "CrateMetadata::kind({:?}): id not found, in crate {:?} with number {}", . item_id, . self.root.name, . self.cnum, . ) . }) 31,092 ( 0.01%) } . 9,821 ( 0.00%) fn def_kind(self, item_id: DefIndex) -> DefKind { 8,932 ( 0.00%) self.root.tables.def_kind.get(self, item_id).map(|k| k.decode(self)).unwrap_or_else(|| { . bug!( . "CrateMetadata::def_kind({:?}): id not found, in crate {:?} with number {}", . item_id, . self.root.name, . self.cnum, . ) . }) 5,612 ( 0.00%) } . 11,412 ( 0.00%) fn get_span(self, index: DefIndex, sess: &Session) -> Span { 11,129 ( 0.00%) self.root . .tables . .span . .get(self, index) . .unwrap_or_else(|| panic!("Missing span for {:?}", index)) . .decode((self, sess)) 6,340 ( 0.00%) } . . fn load_proc_macro(self, id: DefIndex, sess: &Session) -> SyntaxExtension { . let (name, kind, helper_attrs) = match *self.raw_proc_macro(id) { . ProcMacro::CustomDerive { trait_name, attributes, client } => { . let helper_attrs = . attributes.iter().cloned().map(Symbol::intern).collect::>(); . ( . trait_name, -- line 778 ---------------------------------------- -- line 795 ---------------------------------------- . self.get_span(id, sess), . helper_attrs, . self.root.edition, . Symbol::intern(name), . &attrs, . ) . } . 351 ( 0.00%) fn get_trait_def(self, item_id: DefIndex, sess: &Session) -> ty::TraitDef { 108 ( 0.00%) match self.kind(item_id) { 27 ( 0.00%) EntryKind::Trait(data) => { 162 ( 0.00%) let data = data.decode((self, sess)); 432 ( 0.00%) ty::TraitDef::new( . self.local_def_id(item_id), 54 ( 0.00%) data.unsafety, 54 ( 0.00%) data.paren_sugar, 54 ( 0.00%) data.has_auto_impl, 54 ( 0.00%) data.is_marker, 27 ( 0.00%) data.skip_array_during_method_dispatch, 27 ( 0.00%) data.specialization_kind, 108 ( 0.00%) self.def_path_hash(item_id), 27 ( 0.00%) data.must_implement_one_of, . ) . } . EntryKind::TraitAlias => ty::TraitDef::new( . self.local_def_id(item_id), . hir::Unsafety::Normal, . false, . false, . false, . false, . ty::trait_def::TraitSpecializationKind::None, . self.def_path_hash(item_id), . None, . ), . _ => bug!("def-index does not refer to trait or trait alias"), . } 243 ( 0.00%) } . 2,682 ( 0.00%) fn get_variant( . self, . kind: &EntryKind, . index: DefIndex, . parent_did: DefId, . sess: &Session, . ) -> ty::VariantDef { 1,192 ( 0.00%) let data = match kind { . EntryKind::Variant(data) | EntryKind::Struct(data, _) | EntryKind::Union(data, _) => { 1,490 ( 0.00%) data.decode(self) . } . _ => bug!(), . }; . 702 ( 0.00%) let adt_kind = match kind { . EntryKind::Variant(_) => ty::AdtKind::Enum, . EntryKind::Struct(..) => ty::AdtKind::Struct, . EntryKind::Union(..) => ty::AdtKind::Union, . _ => bug!(), . }; . . let variant_did = . if adt_kind == ty::AdtKind::Enum { Some(self.local_def_id(index)) } else { None }; 596 ( 0.00%) let ctor_did = data.ctor.map(|index| self.local_def_id(index)); . 6,556 ( 0.00%) ty::VariantDef::new( . self.item_ident(index, sess).name, . variant_did, . ctor_did, 1,192 ( 0.00%) data.discr, 1,788 ( 0.00%) self.root . .tables . .children . .get(self, index) . .unwrap_or_else(Lazy::empty) . .decode(self) . .map(|index| ty::FieldDef { 122 ( 0.00%) did: self.local_def_id(index), 244 ( 0.00%) name: self.item_ident(index, sess).name, 244 ( 0.00%) vis: self.get_visibility(index), . }) . .collect(), . data.ctor_kind, . adt_kind, . parent_did, . false, 298 ( 0.00%) data.is_non_exhaustive, . ) 2,682 ( 0.00%) } . 1,005 ( 0.00%) fn get_adt_def(self, item_id: DefIndex, tcx: TyCtxt<'tcx>) -> &'tcx ty::AdtDef { 268 ( 0.00%) let kind = self.kind(item_id); 134 ( 0.00%) let did = self.local_def_id(item_id); . 307 ( 0.00%) let (adt_kind, repr) = match kind { . EntryKind::Enum(repr) => (ty::AdtKind::Enum, repr), . EntryKind::Struct(_, repr) => (ty::AdtKind::Struct, repr), . EntryKind::Union(_, repr) => (ty::AdtKind::Union, repr), . _ => bug!("get_adt_def called on a non-ADT {:?}", did), . }; . . let variants = if let ty::AdtKind::Enum = adt_kind { 70 ( 0.00%) self.root . .tables . .children . .get(self, item_id) . .unwrap_or_else(Lazy::empty) . .decode(self) 4,165 ( 0.00%) .map(|index| self.get_variant(&self.kind(index), index, did, tcx.sess)) . .collect() . } else { 424 ( 0.00%) std::iter::once(self.get_variant(&kind, item_id, did, tcx.sess)).collect() . }; . 1,008 ( 0.00%) tcx.alloc_adt_def(did, adt_kind, variants, repr) 469 ( 0.00%) } . . fn get_explicit_predicates( . self, . item_id: DefIndex, . tcx: TyCtxt<'tcx>, . ) -> ty::GenericPredicates<'tcx> { 2,310 ( 0.00%) self.root.tables.explicit_predicates.get(self, item_id).unwrap().decode((self, tcx)) . } . . fn get_inferred_outlives( . self, . item_id: DefIndex, . tcx: TyCtxt<'tcx>, . ) -> &'tcx [(ty::Predicate<'tcx>, Span)] { 2,310 ( 0.00%) self.root . .tables . .inferred_outlives . .get(self, item_id) 12 ( 0.00%) .map(|predicates| tcx.arena.alloc_from_iter(predicates.decode((self, tcx)))) . .unwrap_or_default() . } . . fn get_super_predicates( . self, . item_id: DefIndex, . tcx: TyCtxt<'tcx>, . ) -> ty::GenericPredicates<'tcx> { 115 ( 0.00%) self.root.tables.super_predicates.get(self, item_id).unwrap().decode((self, tcx)) . } . . fn get_explicit_item_bounds( . self, . item_id: DefIndex, . tcx: TyCtxt<'tcx>, . ) -> &'tcx [(ty::Predicate<'tcx>, Span)] { 5 ( 0.00%) self.root . .tables . .explicit_item_bounds . .get(self, item_id) 4 ( 0.00%) .map(|bounds| tcx.arena.alloc_from_iter(bounds.decode((self, tcx)))) . .unwrap_or_default() . } . . fn get_generics(self, item_id: DefIndex, sess: &Session) -> ty::Generics { 2,153 ( 0.00%) self.root.tables.generics.get(self, item_id).unwrap().decode((self, sess)) . } . . fn get_type(self, id: DefIndex, tcx: TyCtxt<'tcx>) -> Ty<'tcx> { 10,925 ( 0.00%) self.root . .tables . .ty . .get(self, id) . .unwrap_or_else(|| panic!("Not a type: {:?}", id)) . .decode((self, tcx)) . } . . fn get_stability(self, id: DefIndex) -> Option { 300 ( 0.00%) self.root.tables.stability.get(self, id).map(|stab| stab.decode(self)) . } . . fn get_const_stability(self, id: DefIndex) -> Option { 5 ( 0.00%) self.root.tables.const_stability.get(self, id).map(|stab| stab.decode(self)) . } . . fn get_deprecation(self, id: DefIndex) -> Option { 385 ( 0.00%) self.root.tables.deprecation.get(self, id).map(|depr| depr.decode(self)) . } . 114 ( 0.00%) fn get_visibility(self, id: DefIndex) -> ty::Visibility { 5,382 ( 0.00%) self.root.tables.visibility.get(self, id).unwrap().decode(self) 76 ( 0.00%) } . . fn get_impl_data(self, id: DefIndex) -> ImplData { 15,480 ( 0.00%) match self.kind(id) { 12,900 ( 0.00%) EntryKind::Impl(data) => data.decode(self), . _ => bug!(), . } . } . . fn get_parent_impl(self, id: DefIndex) -> Option { 7,491 ( 0.00%) self.get_impl_data(id).parent_impl . } . . fn get_impl_polarity(self, id: DefIndex) -> ty::ImplPolarity { 59 ( 0.00%) self.get_impl_data(id).polarity . } . . fn get_impl_defaultness(self, id: DefIndex) -> hir::Defaultness { 17 ( 0.00%) self.get_impl_data(id).defaultness . } . . fn get_impl_constness(self, id: DefIndex) -> hir::Constness { 7 ( 0.00%) self.get_impl_data(id).constness . } . . fn get_trait_item_def_id(self, id: DefIndex) -> Option { 2,125 ( 0.00%) self.root.tables.trait_item_def_id.get(self, id).map(|d| d.decode(self)) . } . . fn get_coerce_unsized_info(self, id: DefIndex) -> Option { . self.get_impl_data(id).coerce_unsized_info . } . . fn get_impl_trait(self, id: DefIndex, tcx: TyCtxt<'tcx>) -> Option> { 30,575 ( 0.01%) self.root.tables.impl_trait_ref.get(self, id).map(|tr| tr.decode((self, tcx))) . } . . fn get_expn_that_defined(self, id: DefIndex, sess: &Session) -> ExpnId { 2,175 ( 0.00%) self.root.tables.expn_that_defined.get(self, id).unwrap().decode((self, sess)) . } . . fn get_const_param_default( . self, . tcx: TyCtxt<'tcx>, . id: DefIndex, . ) -> rustc_middle::ty::Const<'tcx> { . self.root.tables.const_defaults.get(self, id).unwrap().decode((self, tcx)) -- line 1026 ---------------------------------------- -- line 1028 ---------------------------------------- . . /// Iterates over all the stability attributes in the given crate. . fn get_lib_features(self, tcx: TyCtxt<'tcx>) -> &'tcx [(Symbol, Option)] { . tcx.arena.alloc_from_iter(self.root.lib_features.decode(self)) . } . . /// Iterates over the language items in the given crate. . fn get_lang_items(self, tcx: TyCtxt<'tcx>) -> &'tcx [(DefId, usize)] { 72 ( 0.00%) tcx.arena.alloc_from_iter( 48 ( 0.00%) self.root . .lang_items . .decode(self) 564 ( 0.00%) .map(|(def_index, index)| (self.local_def_id(def_index), index)), . ) . } . . /// Iterates over the diagnostic items in the given crate. . fn get_diagnostic_items(self) -> DiagnosticItems { . let mut id_to_name = FxHashMap::default(); 48 ( 0.00%) let name_to_id = self . .root . .diagnostic_items . .decode(self) . .map(|(name, def_index)| { 130 ( 0.00%) let id = self.local_def_id(def_index); 130 ( 0.00%) id_to_name.insert(id, name); . (name, id) . }) . .collect(); 216 ( 0.00%) DiagnosticItems { id_to_name, name_to_id } . } . . /// Iterates over all named children of the given module, . /// including both proper items and reexports. . /// Module here is understood in name resolution sense - it can be a `mod` item, . /// or a crate root, or an enum, or a trait. . fn for_each_module_child( . self, . id: DefIndex, . mut callback: impl FnMut(ModChild), . sess: &Session, . ) { 228 ( 0.00%) if let Some(data) = &self.root.proc_macro_data { . // If we are loading as a proc macro, we want to return . // the view of this crate as a proc macro crate. . if id == CRATE_DEF_INDEX { . for def_index in data.macros.decode(self) { . let raw_macro = self.raw_proc_macro(def_index); . let res = Res::Def( . DefKind::Macro(macro_kind(raw_macro)), . self.local_def_id(def_index), -- line 1078 ---------------------------------------- -- line 1085 ---------------------------------------- . span: ident.span, . }); . } . } . return; . } . . // Iterate over all children. 456 ( 0.00%) if let Some(children) = self.root.tables.children.get(self, id) { 912 ( 0.00%) for child_index in children.decode((self, sess)) { 12,096 ( 0.00%) if let Some(ident) = self.opt_item_ident(child_index, sess) { 2,718 ( 0.00%) let kind = self.def_kind(child_index); 1,812 ( 0.00%) if matches!(kind, DefKind::Macro(..)) { . // FIXME: Macros are currently encoded twice, once as items and once as . // reexports. We ignore the items here and only use the reexports. . continue; . } . let def_id = self.local_def_id(child_index); . let res = Res::Def(kind, def_id); . let vis = self.get_visibility(child_index); . let span = self.get_span(child_index, sess); . . callback(ModChild { ident, res, vis, span }); . . // For non-re-export structs and variants add their constructors to children. . // Re-export lists automatically contain constructors when necessary. 3,134 ( 0.00%) match kind { . DefKind::Struct => { 53 ( 0.00%) if let Some((ctor_def_id, ctor_kind)) = . self.get_ctor_def_id_and_kind(child_index) . { . let ctor_res = . Res::Def(DefKind::Ctor(CtorOf::Struct, ctor_kind), ctor_def_id); . let vis = self.get_visibility(ctor_def_id.index); . callback(ModChild { ident, res: ctor_res, vis, span }); . } . } . DefKind::Variant => { . // Braced variants, unlike structs, generate unusable names in . // value namespace, they are reserved for possible future use. . // It's ok to use the variant's id as a ctor id since an . // error will be reported on any use of such resolution anyway. 687 ( 0.00%) let (ctor_def_id, ctor_kind) = self . .get_ctor_def_id_and_kind(child_index) . .unwrap_or((def_id, CtorKind::Fictive)); . let ctor_res = . Res::Def(DefKind::Ctor(CtorOf::Variant, ctor_kind), ctor_def_id); . let mut vis = self.get_visibility(ctor_def_id.index); 229 ( 0.00%) if ctor_def_id == def_id && vis.is_public() { . // For non-exhaustive variants lower the constructor visibility to . // within the crate. We only need this for fictive constructors, . // for other constructors correct visibilities . // were already encoded in metadata. . let mut attrs = self.get_item_attrs(def_id.index, sess); . if attrs.any(|item| item.has_name(sym::non_exhaustive)) { . let crate_def_id = self.local_def_id(CRATE_DEF_INDEX); . vis = ty::Visibility::Restricted(crate_def_id); -- line 1141 ---------------------------------------- -- line 1144 ---------------------------------------- . callback(ModChild { ident, res: ctor_res, vis, span }); . } . _ => {} . } . } . } . } . 559 ( 0.00%) match self.kind(id) { 44 ( 0.00%) EntryKind::Mod(exports) => { 11,618 ( 0.00%) for exp in exports.decode((self, sess)) { . callback(exp); . } . } . EntryKind::Enum(..) | EntryKind::Trait(..) => {} . _ => bug!("`for_each_module_child` is called on a non-module: {:?}", self.def_kind(id)), . } . } . . fn is_ctfe_mir_available(self, id: DefIndex) -> bool { 1,400 ( 0.00%) self.root.tables.mir_for_ctfe.get(self, id).is_some() . } . . fn is_item_mir_available(self, id: DefIndex) -> bool { 204 ( 0.00%) self.root.tables.mir.get(self, id).is_some() . } . . fn module_expansion(self, id: DefIndex, sess: &Session) -> ExpnId { 2,898 ( 0.00%) match self.kind(id) { . EntryKind::Mod(_) | EntryKind::Enum(_) | EntryKind::Trait(_) => { . self.get_expn_that_defined(id, sess) . } . _ => panic!("Expected module, found {:?}", self.local_def_id(id)), . } . } . . fn get_optimized_mir(self, tcx: TyCtxt<'tcx>, id: DefIndex) -> Body<'tcx> { 250 ( 0.00%) self.root . .tables . .mir . .get(self, id) . .unwrap_or_else(|| { . bug!("get_optimized_mir: missing MIR for `{:?}`", self.local_def_id(id)) . }) . .decode((self, tcx)) . } . . fn get_mir_for_ctfe(self, tcx: TyCtxt<'tcx>, id: DefIndex) -> Body<'tcx> { 1,750 ( 0.00%) self.root . .tables . .mir_for_ctfe . .get(self, id) . .unwrap_or_else(|| { . bug!("get_mir_for_ctfe: missing MIR for `{:?}`", self.local_def_id(id)) . }) . .decode((self, tcx)) . } -- line 1200 ---------------------------------------- -- line 1216 ---------------------------------------- . .tables . .unused_generic_params . .get(self, id) . .map(|params| params.decode(self)) . .unwrap_or_default() . } . . fn get_promoted_mir(self, tcx: TyCtxt<'tcx>, id: DefIndex) -> IndexVec> { 10 ( 0.00%) self.root . .tables . .promoted_mir . .get(self, id) . .unwrap_or_else(|| { . bug!("get_promoted_mir: missing MIR for `{:?}`", self.local_def_id(id)) . }) . .decode((self, tcx)) . } -- line 1232 ---------------------------------------- -- line 1242 ---------------------------------------- . qualif, . _, . ) => qualif, . _ => bug!("mir_const_qualif: unexpected kind"), . } . } . . fn get_fn_has_self_parameter(self, id: DefIndex) -> bool { 585 ( 0.00%) match self.kind(id) { 468 ( 0.00%) EntryKind::AssocFn(data) => data.decode(self).has_self, . _ => false, . } . } . . fn get_associated_item_def_ids(self, tcx: TyCtxt<'tcx>, id: DefIndex) -> &'tcx [DefId] { 455 ( 0.00%) if let Some(children) = self.root.tables.children.get(self, id) { 195 ( 0.00%) tcx.arena.alloc_from_iter( 65 ( 0.00%) children.decode((self, tcx.sess)).map(|child_index| self.local_def_id(child_index)), . ) . } else { . &[] . } . } . 4,536 ( 0.00%) fn get_associated_item(self, id: DefIndex, sess: &Session) -> ty::AssocItem { 2,646 ( 0.00%) let def_key = self.def_key(id); . let parent = self.local_def_id(def_key.parent.unwrap()); . let ident = self.item_ident(id, sess); . 2,808 ( 0.00%) let (kind, container, has_self) = match self.kind(id) { 136 ( 0.00%) EntryKind::AssocConst(container, _, _) => (ty::AssocKind::Const, container, false), 331 ( 0.00%) EntryKind::AssocFn(data) => { . let data = data.decode(self); 2,317 ( 0.00%) (ty::AssocKind::Fn, data.container, data.has_self) . } 26 ( 0.00%) EntryKind::AssocType(container) => (ty::AssocKind::Type, container, false), . _ => bug!("cannot get associated-item of `{:?}`", def_key), . }; . 6,804 ( 0.00%) ty::AssocItem { . name: ident.name, . kind, . vis: self.get_visibility(id), . defaultness: container.defaultness(), . def_id: self.local_def_id(id), . trait_item_def_id: self.get_trait_item_def_id(id), . container: container.with_def_id(parent), . fn_has_self_parameter: has_self, . } 3,402 ( 0.00%) } . . fn get_item_variances(self, id: DefIndex) -> impl Iterator + 'a { 150 ( 0.00%) self.root.tables.variances.get(self, id).unwrap_or_else(Lazy::empty).decode(self) . } . . fn get_ctor_def_id_and_kind(self, node_id: DefIndex) -> Option<(DefId, CtorKind)> { 5,084 ( 0.00%) match self.kind(node_id) { . EntryKind::Struct(data, _) | EntryKind::Variant(data) => { 2,121 ( 0.00%) let vdata = data.decode(self); 1,657 ( 0.00%) vdata.ctor.map(|index| (self.local_def_id(index), vdata.ctor_kind)) . } . _ => None, . } 3,029 ( 0.00%) } . 228 ( 0.00%) fn get_item_attrs( . self, . id: DefIndex, . sess: &'a Session, . ) -> impl Iterator + 'a { 1,167 ( 0.00%) self.root . .tables . .attributes . .get(self, id) . .unwrap_or_else(|| { . // Structure and variant constructors don't have any attributes encoded for them, . // but we assume that someone passing a constructor ID actually wants to look at . // the attributes on the corresponding struct or variant. . let def_key = self.def_key(id); -- line 1320 ---------------------------------------- -- line 1322 ---------------------------------------- . let parent_id = def_key.parent.expect("no parent for a constructor"); . self.root . .tables . .attributes . .get(self, parent_id) . .expect("no encoded attributes for a structure or variant") . }) . .decode((self, sess)) 171 ( 0.00%) } . . fn get_struct_field_names( . self, . id: DefIndex, . sess: &'a Session, . ) -> impl Iterator> + 'a { 1,665 ( 0.00%) self.root . .tables . .children . .get(self, id) . .unwrap_or_else(Lazy::empty) . .decode(self) 12,680 ( 0.00%) .map(move |index| respan(self.get_span(index, sess), self.item_ident(index, sess).name)) . } . . fn get_struct_field_visibilities(self, id: DefIndex) -> impl Iterator + 'a { 120 ( 0.00%) self.root . .tables . .children . .get(self, id) . .unwrap_or_else(Lazy::empty) . .decode(self) 38 ( 0.00%) .map(move |field_index| self.get_visibility(field_index)) . } . . fn get_inherent_implementations_for_type( . self, . tcx: TyCtxt<'tcx>, . id: DefIndex, . ) -> &'tcx [DefId] { 56 ( 0.00%) tcx.arena.alloc_from_iter( 70 ( 0.00%) self.root . .tables . .inherent_impls . .get(self, id) . .unwrap_or_else(Lazy::empty) . .decode(self) 368 ( 0.00%) .map(|index| self.local_def_id(index)), . ) . } . . fn get_traits(self) -> impl Iterator + 'a { . self.root.traits.decode(self).map(move |index| self.local_def_id(index)) . } . . fn get_trait_impls(self) -> impl Iterator)> + 'a { -- line 1376 ---------------------------------------- -- line 1380 ---------------------------------------- . index: *trait_index, . }; . impls.decode(self).map(move |(impl_index, simplified_self_ty)| { . (trait_def_id, self.local_def_id(impl_index), simplified_self_ty) . }) . }) . } . 5,832 ( 0.00%) fn get_implementations_of_trait( . self, . tcx: TyCtxt<'tcx>, . trait_def_id: DefId, . ) -> &'tcx [(DefId, Option)] { 648 ( 0.00%) if self.trait_impls.is_empty() { . return &[]; . } . . // Do a reverse lookup beforehand to avoid touching the crate_num . // hash map in the loop below. 940 ( 0.00%) let key = match self.reverse_translate_def_id(trait_def_id) { . Some(def_id) => (def_id.krate.as_u32(), def_id.index), . None => return &[], . }; . . if let Some(impls) = self.trait_impls.get(&key) { 508 ( 0.00%) tcx.arena.alloc_from_iter( 254 ( 0.00%) impls . .decode(self) 5,348 ( 0.00%) .map(|(idx, simplified_self_ty)| (self.local_def_id(idx), simplified_self_ty)), . ) . } else { . &[] . } 5,184 ( 0.00%) } . . fn get_trait_of_item(self, id: DefIndex) -> Option { 1,008 ( 0.00%) let def_key = self.def_key(id); 1,260 ( 0.00%) match def_key.disambiguated_data.data { . DefPathData::TypeNs(..) | DefPathData::ValueNs(..) => (), . // Not an associated item . _ => return None, . } 2,259 ( 0.00%) def_key.parent.and_then(|parent_index| match self.kind(parent_index) { . EntryKind::Trait(_) | EntryKind::TraitAlias => Some(self.local_def_id(parent_index)), . _ => None, . }) . } . . fn get_native_libraries(self, sess: &'a Session) -> impl Iterator + 'a { 48 ( 0.00%) self.root.native_libraries.decode((self, sess)) . } . . fn get_proc_macro_quoted_span(self, index: usize, sess: &Session) -> Span { . self.root . .tables . .proc_macro_quoted_spans . .get(self, index) . .unwrap_or_else(|| panic!("Missing proc macro quoted span: {:?}", index)) -- line 1437 ---------------------------------------- -- line 1450 ---------------------------------------- . self.root.dylib_dependency_formats.decode(self).enumerate().flat_map(|(i, link)| { . let cnum = CrateNum::new(i + 1); . link.map(|link| (self.cnum_map[cnum], link)) . }), . ) . } . . fn get_missing_lang_items(self, tcx: TyCtxt<'tcx>) -> &'tcx [lang_items::LangItem] { 120 ( 0.00%) tcx.arena.alloc_from_iter(self.root.lang_items_missing.decode(self)) . } . . fn get_fn_param_names(self, tcx: TyCtxt<'tcx>, id: DefIndex) -> &'tcx [Ident] { . let param_names = match self.kind(id) { . EntryKind::Fn(data) | EntryKind::ForeignFn(data) => data.decode(self).param_names, . EntryKind::AssocFn(data) => data.decode(self).fn_data.param_names, . _ => Lazy::empty(), . }; . tcx.arena.alloc_from_iter(param_names.decode((self, tcx))) . } . . fn exported_symbols( . self, . tcx: TyCtxt<'tcx>, . ) -> &'tcx [(ExportedSymbol<'tcx>, SymbolExportLevel)] { 24 ( 0.00%) tcx.arena.alloc_from_iter(self.root.exported_symbols.decode((self, tcx))) . } . . fn get_rendered_const(self, id: DefIndex) -> String { . match self.kind(id) { . EntryKind::AnonConst(_, data) . | EntryKind::Const(_, data) . | EntryKind::AssocConst(_, _, data) => data.decode(self).0, . _ => bug!(), . } . } . . fn get_macro(self, id: DefIndex, sess: &Session) -> MacroDef { 30 ( 0.00%) match self.kind(id) { 5 ( 0.00%) EntryKind::MacroDef(macro_def) => macro_def.decode((self, sess)), . _ => bug!(), . } . } . . // This replicates some of the logic of the crate-local `is_const_fn_raw` query, because we . // don't serialize constness for tuple variant and tuple struct constructors. . fn is_const_fn_raw(self, id: DefIndex) -> bool { 36 ( 0.00%) let constness = match self.kind(id) { 8 ( 0.00%) EntryKind::AssocFn(data) => data.decode(self).fn_data.constness, . EntryKind::Fn(data) => data.decode(self).constness, . EntryKind::ForeignFn(data) => data.decode(self).constness, . EntryKind::Variant(..) | EntryKind::Struct(..) => hir::Constness::Const, . _ => hir::Constness::NotConst, . }; . constness == hir::Constness::Const . } . -- line 1505 ---------------------------------------- -- line 1508 ---------------------------------------- . EntryKind::Fn(data) => data.decode(self).asyncness, . EntryKind::AssocFn(data) => data.decode(self).fn_data.asyncness, . EntryKind::ForeignFn(data) => data.decode(self).asyncness, . _ => bug!("asyncness: expected function kind"), . } . } . . fn is_foreign_item(self, id: DefIndex) -> bool { 814 ( 0.00%) match self.kind(id) { . EntryKind::ForeignImmStatic | EntryKind::ForeignMutStatic | EntryKind::ForeignFn(_) => { . true . } . _ => false, . } . } . . fn static_mutability(self, id: DefIndex) -> Option { 3,850 ( 0.00%) match self.kind(id) { . EntryKind::ImmStatic | EntryKind::ForeignImmStatic => Some(hir::Mutability::Not), . EntryKind::MutStatic | EntryKind::ForeignMutStatic => Some(hir::Mutability::Mut), . _ => None, . } . } . . fn generator_kind(self, id: DefIndex) -> Option { . match self.kind(id) { . EntryKind::Generator(data) => Some(data), . _ => None, . } . } . . fn fn_sig(self, id: DefIndex, tcx: TyCtxt<'tcx>) -> ty::PolyFnSig<'tcx> { 555 ( 0.00%) self.root.tables.fn_sig.get(self, id).unwrap().decode((self, tcx)) . } . . #[inline] 49,175 ( 0.01%) fn def_key(self, index: DefIndex) -> DefKey { 14,084 ( 0.00%) *self . .def_key_cache . .lock() . .entry(index) 20,285 ( 0.00%) .or_insert_with(|| self.root.tables.def_keys.get(self, index).unwrap().decode(self)) 56,200 ( 0.01%) } . . // Returns the path leading to the thing with this `id`. . fn def_path(self, id: DefIndex) -> DefPath { . debug!("def_path(cnum={:?}, id={:?})", self.cnum, id); 28 ( 0.00%) DefPath::make(self.cnum, id, |parent| self.def_key(parent)) . } . . fn def_path_hash_unlocked( . self, . index: DefIndex, . def_path_hashes: &mut FxHashMap, . ) -> DefPathHash { 3,186 ( 0.00%) *def_path_hashes.entry(index).or_insert_with(|| { 4,120 ( 0.00%) self.root.tables.def_path_hashes.get(self, index).unwrap().decode(self) . }) . } . . #[inline] 11,151 ( 0.00%) fn def_path_hash(self, index: DefIndex) -> DefPathHash { . let mut def_path_hashes = self.def_path_hash_cache.lock(); . self.def_path_hash_unlocked(index, &mut def_path_hashes) 12,744 ( 0.00%) } . . #[inline] . fn def_path_hash_to_def_index(self, hash: DefPathHash) -> DefIndex { . self.def_path_hash_map.def_path_hash_to_def_index(&hash) . } . . fn expn_hash_to_expn_id(self, sess: &Session, index_guess: u32, hash: ExpnHash) -> ExpnId { . debug_assert_eq!(ExpnId::from_hash(hash), None); -- line 1580 ---------------------------------------- -- line 1635 ---------------------------------------- . /// Proc macro crates don't currently export spans, so this function does not have . /// to work for them. . fn imported_source_files(self, sess: &Session) -> &'a [ImportedSourceFile] { . // Translate the virtual `/rustc/$hash` prefix back to a real directory . // that should hold actual sources, where possible. . // . // NOTE: if you update this, you might need to also update bootstrap's code for generating . // the `rust-src` component in `Src::run` in `src/bootstrap/dist.rs`. 56,888 ( 0.01%) let virtual_rust_source_base_dir = option_env!("CFG_VIRTUAL_RUST_SOURCE_BASE_DIR") . .map(Path::new) . .filter(|_| { . // Only spend time on further checks if we have what to translate *to*. . sess.opts.real_rust_source_base_dir.is_some() . }) . .filter(|virtual_dir| { . // Don't translate away `/rustc/$hash` if we're still remapping to it, . // since that means we're still building `std`/`rustc` that need it, . // and we don't want the real path to leak into codegen/debuginfo. . !sess.opts.remap_path_prefix.iter().any(|(_from, to)| to == virtual_dir) . }); 56,888 ( 0.01%) let try_to_translate_virtual_to_real = |name: &mut rustc_span::FileName| { . debug!( . "try_to_translate_virtual_to_real(name={:?}): \ . virtual_rust_source_base_dir={:?}, real_rust_source_base_dir={:?}", . name, virtual_rust_source_base_dir, sess.opts.real_rust_source_base_dir, . ); . 3,170 ( 0.00%) if let Some(virtual_dir) = virtual_rust_source_base_dir { . if let Some(real_dir) = &sess.opts.real_rust_source_base_dir { . if let rustc_span::FileName::Real(old_name) = name { . if let rustc_span::RealFileName::Remapped { local_path: _, virtual_name } = . old_name . { . if let Ok(rest) = virtual_name.strip_prefix(virtual_dir) { . let virtual_name = virtual_name.clone(); . -- line 1670 ---------------------------------------- -- line 1709 ---------------------------------------- . *old_name = new_name; . } . } . } . } . } . }; . 56,734 ( 0.01%) self.cdata.source_map_import_info.get_or_init(|| { 36 ( 0.00%) let external_source_map = self.root.source_map.decode(self); . . external_source_map . .map(|source_file_to_import| { . // We can't reuse an existing SourceFile, so allocate a new one . // containing the information we need. . let rustc_span::SourceFile { 13,314 ( 0.00%) mut name, 3,804 ( 0.00%) src_hash, 634 ( 0.00%) start_pos, 1,268 ( 0.00%) end_pos, 3,170 ( 0.00%) mut lines, 3,170 ( 0.00%) mut multibyte_chars, 3,804 ( 0.00%) mut non_narrow_chars, 3,170 ( 0.00%) mut normalized_pos, 3,804 ( 0.00%) name_hash, . .. . } = source_file_to_import; . . // If this file is under $sysroot/lib/rustlib/src/ but has not been remapped . // during rust bootstrapping by `remap-debuginfo = true`, and the user . // wish to simulate that behaviour by -Z simulate-remapped-rust-src-base, . // then we change `name` to a similar state as if the rust was bootstrapped . // with `remap-debuginfo = true`. . // This is useful for testing so that tests about the effects of . // `try_to_translate_virtual_to_real` don't have to worry about how the . // compiler is bootstrapped. 1,268 ( 0.00%) if let Some(virtual_dir) = . &sess.opts.debugging_opts.simulate_remapped_rust_src_base . { . if let Some(real_dir) = &sess.opts.real_rust_source_base_dir { . if let rustc_span::FileName::Real(ref mut old_name) = name { . if let rustc_span::RealFileName::LocalPath(local) = old_name { . if let Ok(rest) = local.strip_prefix(real_dir) { . *old_name = rustc_span::RealFileName::Remapped { . local_path: None, -- line 1753 ---------------------------------------- -- line 1767 ---------------------------------------- . let source_length = (end_pos - start_pos).to_usize(); . . // Translate line-start positions and multibyte character . // position into frame of reference local to file. . // `SourceMap::new_imported_source_file()` will then translate those . // coordinates to their new global frame of reference when the . // offset of the SourceFile is known. . for pos in &mut lines { 793,058 ( 0.17%) *pos = *pos - start_pos; . } . for mbc in &mut multibyte_chars { 1,608 ( 0.00%) mbc.pos = mbc.pos - start_pos; . } . for swc in &mut non_narrow_chars { 924 ( 0.00%) *swc = *swc - start_pos; . } . for np in &mut normalized_pos { . np.pos = np.pos - start_pos; . } . 11,412 ( 0.00%) let local_version = sess.source_map().new_imported_source_file( 8,242 ( 0.00%) name, 4,438 ( 0.00%) src_hash, . name_hash, . source_length, 1,268 ( 0.00%) self.cnum, 3,170 ( 0.00%) lines, 3,804 ( 0.00%) multibyte_chars, 3,170 ( 0.00%) non_narrow_chars, 2,536 ( 0.00%) normalized_pos, . start_pos, . end_pos, . ); . debug!( . "CrateMetaData::imported_source_files alloc \ . source_file {:?} original (start_pos {:?} end_pos {:?}) \ . translated (start_pos {:?} end_pos {:?})", . local_version.name, -- line 1804 ---------------------------------------- -- line 1808 ---------------------------------------- . local_version.end_pos . ); . . ImportedSourceFile { . original_start_pos: start_pos, . original_end_pos: end_pos, . translated_source_file: local_version, . } 1,902 ( 0.00%) }) . .collect() . }) . } . } . . impl CrateMetadata { 288 ( 0.00%) crate fn new( . sess: &Session, . blob: MetadataBlob, . root: CrateRoot<'static>, . raw_proc_macros: Option<&'static [ProcMacro]>, . cnum: CrateNum, . cnum_map: CrateNumMap, . dep_kind: CrateDepKind, . source: CrateSource, . private_dep: bool, . host_hash: Option, . ) -> CrateMetadata { 72 ( 0.00%) let trait_impls = root . .impls . .decode((&blob, sess)) . .map(|trait_impls| (trait_impls.trait_id, trait_impls.impls)) . .collect(); . let alloc_decoding_state = 72 ( 0.00%) AllocDecodingState::new(root.interpret_alloc_index.decode(&blob).collect()); . let dependencies = Lock::new(cnum_map.iter().cloned().collect()); . . // Pre-decode the DefPathHash->DefIndex table. This is a cheap operation . // that does not copy any data. It just does some data verification. 24 ( 0.00%) let def_path_hash_map = root.def_path_hash_map.decode(&blob); . 552 ( 0.00%) CrateMetadata { 24 ( 0.00%) blob, 72 ( 0.00%) root, 96 ( 0.00%) trait_impls, . raw_proc_macros, . source_map_import_info: OnceCell::new(), 96 ( 0.00%) def_path_hash_map, . expn_hash_map: Default::default(), 168 ( 0.00%) alloc_decoding_state, . cnum, 96 ( 0.00%) cnum_map, 120 ( 0.00%) dependencies, . dep_kind: Lock::new(dep_kind), 288 ( 0.00%) source, . private_dep, . host_hash, . extern_crate: Lock::new(None), . hygiene_context: Default::default(), . def_key_cache: Default::default(), . def_path_hash_cache: Default::default(), . } 216 ( 0.00%) } . . crate fn dependencies(&self) -> LockGuard<'_, Vec> { . self.dependencies.borrow() . } . . crate fn add_dependency(&self, cnum: CrateNum) { . self.dependencies.borrow_mut().push(cnum); . } . . crate fn update_extern_crate(&self, new_extern_crate: ExternCrate) -> bool { . let mut extern_crate = self.extern_crate.borrow_mut(); . let update = Some(new_extern_crate.rank()) > extern_crate.as_ref().map(ExternCrate::rank); . if update { 120 ( 0.00%) *extern_crate = Some(new_extern_crate); . } . update . } . . crate fn source(&self) -> &CrateSource { 184 ( 0.00%) &self.source . } . . crate fn dep_kind(&self) -> CrateDepKind { . *self.dep_kind.lock() . } . . crate fn update_dep_kind(&self, f: impl FnOnce(CrateDepKind) -> CrateDepKind) { 145 ( 0.00%) self.dep_kind.with_lock(|dep_kind| *dep_kind = f(*dep_kind)) . } . . crate fn panic_strategy(&self) -> PanicStrategy { . self.root.panic_strategy . } . . crate fn needs_panic_runtime(&self) -> bool { . self.root.needs_panic_runtime -- line 1905 ---------------------------------------- -- line 1909 ---------------------------------------- . self.root.panic_runtime . } . . crate fn is_profiler_runtime(&self) -> bool { . self.root.profiler_runtime . } . . crate fn needs_allocator(&self) -> bool { 5 ( 0.00%) self.root.needs_allocator . } . . crate fn has_global_allocator(&self) -> bool { . self.root.has_global_allocator . } . . crate fn has_default_lib_allocator(&self) -> bool { . self.root.has_default_lib_allocator . } . . crate fn is_proc_macro_crate(&self) -> bool { . self.root.is_proc_macro_crate() . } . . crate fn name(&self) -> Symbol { 407 ( 0.00%) self.root.name . } . . crate fn stable_crate_id(&self) -> StableCrateId { . self.root.stable_crate_id . } . . crate fn hash(&self) -> Svh { . self.root.hash . } . . fn num_def_ids(&self) -> usize { . self.root.tables.def_keys.size() . } . . fn local_def_id(&self, index: DefIndex) -> DefId { 13,627 ( 0.00%) DefId { krate: self.cnum, index } . } . . // Translate a DefId from the current compilation environment to a DefId . // for an external crate. . fn reverse_translate_def_id(&self, did: DefId) -> Option { . for (local, &global) in self.cnum_map.iter_enumerated() { 1,914 ( 0.00%) if global == did.krate { . return Some(DefId { krate: local, index: did.index }); . } . } . . None . } . } . -- line 1964 ---------------------------------------- 630,602 ( 0.13%) -------------------------------------------------------------------------------- -- Auto-annotated source: /usr/home/liquid/rust/worktree-benchmarking/library/core/src/slice/iter/macros.rs -------------------------------------------------------------------------------- Ir -- line 70 ---------------------------------------- . . impl<'a, T> $name<'a, T> { . // Helper function for creating a slice from the iterator. . #[inline(always)] . fn make_slice(&self) -> &'a [T] { . // SAFETY: the iterator was created from a slice with pointer . // `self.ptr` and length `len!(self)`. This guarantees that all . // the prerequisites for `from_raw_parts` are fulfilled. 3,274 ( 0.00%) unsafe { from_raw_parts(self.ptr.as_ptr(), len!(self)) } . } . . // Helper function for moving the start of the iterator forwards by `offset` elements, . // returning the old start. . // Unsafe because the offset must not exceed `self.len()`. . #[inline(always)] . unsafe fn post_inc_start(&mut self, offset: isize) -> * $raw_mut T { . if mem::size_of::() == 0 { . zst_shrink!(self, offset); . self.ptr.as_ptr() . } else { . let old = self.ptr.as_ptr(); . // SAFETY: the caller guarantees that `offset` doesn't exceed `self.len()`, . // so this new pointer is inside `self` and thus guaranteed to be non-null. 31,544 ( 0.01%) self.ptr = unsafe { NonNull::new_unchecked(self.ptr.as_ptr().offset(offset)) }; . old . } . } . . // Helper function for moving the end of the iterator backwards by `offset` elements, . // returning the new end. . // Unsafe because the offset must not exceed `self.len()`. . #[inline(always)] -- line 101 ---------------------------------------- -- line 102 ---------------------------------------- . unsafe fn pre_dec_end(&mut self, offset: isize) -> * $raw_mut T { . if mem::size_of::() == 0 { . zst_shrink!(self, offset); . self.ptr.as_ptr() . } else { . // SAFETY: the caller guarantees that `offset` doesn't exceed `self.len()`, . // which is guaranteed to not overflow an `isize`. Also, the resulting pointer . // is in bounds of `slice`, which fulfills the other requirements for `offset`. 152 ( 0.00%) self.end = unsafe { self.end.offset(-offset) }; . self.end . } . } . } . . #[stable(feature = "rust1", since = "1.0.0")] . impl ExactSizeIterator for $name<'_, T> { . #[inline(always)] . fn len(&self) -> usize { 3,785 ( 0.00%) len!(self) . } . . #[inline(always)] . fn is_empty(&self) -> bool { . is_empty!(self) . } . } . -- line 128 ---------------------------------------- -- line 134 ---------------------------------------- . fn next(&mut self) -> Option<$elem> { . // could be implemented with slices, but this avoids bounds checks . . // SAFETY: `assume` calls are safe since a slice's start pointer . // must be non-null, and slices over non-ZSTs must also have a . // non-null end pointer. The call to `next_unchecked!` is safe . // since we check if the iterator is empty first. . unsafe { 23,008 ( 0.00%) assume(!self.ptr.as_ptr().is_null()); . if mem::size_of::() != 0 { 4,409 ( 0.00%) assume(!self.end.is_null()); . } 3,956,338 ( 0.84%) if is_empty!(self) { . None . } else { . Some(next_unchecked!(self)) . } . } . } . . #[inline] . fn size_hint(&self) -> (usize, Option) { 45,891 ( 0.01%) let exact = len!(self); . (exact, Some(exact)) . } . . #[inline] . fn count(self) -> usize { . len!(self) . } . . #[inline] . fn nth(&mut self, n: usize) -> Option<$elem> { 621 ( 0.00%) if n >= len!(self) { . // This iterator is now empty. . if mem::size_of::() == 0 { . // We have to do it this way as `ptr` may never be 0, but `end` . // could be (due to wrapping). . self.end = self.ptr.as_ptr(); . } else { . // SAFETY: end can't be 0 if T isn't ZST because ptr isn't 0 and end >= ptr . unsafe { -- line 175 ---------------------------------------- -- line 203 ---------------------------------------- . // faster to compile. . #[inline] . fn for_each(mut self, mut f: F) . where . Self: Sized, . F: FnMut(Self::Item), . { . while let Some(x) = self.next() { 141 ( 0.00%) f(x); . } . } . . // We override the default implementation, which uses `try_fold`, . // because this simple implementation generates less LLVM IR and is . // faster to compile. . #[inline] . fn all(&mut self, mut f: F) -> bool . where . Self: Sized, . F: FnMut(Self::Item) -> bool, . { 201 ( 0.00%) while let Some(x) = self.next() { 4,438 ( 0.00%) if !f(x) { . return false; . } . } . true . } . . // We override the default implementation, which uses `try_fold`, . // because this simple implementation generates less LLVM IR and is . // faster to compile. . #[inline] 112 ( 0.00%) fn any(&mut self, mut f: F) -> bool . where . Self: Sized, . F: FnMut(Self::Item) -> bool, . { 17,723 ( 0.00%) while let Some(x) = self.next() { 19,382 ( 0.00%) if f(x) { . return true; . } . } . false 112 ( 0.00%) } . . // We override the default implementation, which uses `try_fold`, . // because this simple implementation generates less LLVM IR and is . // faster to compile. . #[inline] . fn find

(&mut self, mut predicate: P) -> Option . where . Self: Sized, . P: FnMut(&Self::Item) -> bool, . { 952 ( 0.00%) while let Some(x) = self.next() { 8,046 ( 0.00%) if predicate(&x) { . return Some(x); . } . } . None 1,430 ( 0.00%) } . . // We override the default implementation, which uses `try_fold`, . // because this simple implementation generates less LLVM IR and is . // faster to compile. . #[inline] . fn find_map(&mut self, mut f: F) -> Option . where . Self: Sized, . F: FnMut(Self::Item) -> Option, . { 2,799 ( 0.00%) while let Some(x) = self.next() { 6,051 ( 0.00%) if let Some(y) = f(x) { 473 ( 0.00%) return Some(y); . } . } . None 265 ( 0.00%) } . . // We override the default implementation, which uses `try_fold`, . // because this simple implementation generates less LLVM IR and is . // faster to compile. Also, the `assume` avoids a bounds check. . #[inline] . #[rustc_inherit_overflow_checks] . fn position

(&mut self, mut predicate: P) -> Option where . Self: Sized, . P: FnMut(Self::Item) -> bool, . { . let n = len!(self); . let mut i = 0; . while let Some(x) = self.next() { 10,499 ( 0.00%) if predicate(x) { . // SAFETY: we are guaranteed to be in bounds by the loop invariant: . // when `i >= n`, `self.next()` returns `None` and the loop breaks. . unsafe { assume(i < n) }; . return Some(i); . } . i += 1; . } . None -- line 303 ---------------------------------------- -- line 308 ---------------------------------------- . // faster to compile. Also, the `assume` avoids a bounds check. . #[inline] . fn rposition

(&mut self, mut predicate: P) -> Option where . P: FnMut(Self::Item) -> bool, . Self: Sized + ExactSizeIterator + DoubleEndedIterator . { . let n = len!(self); . let mut i = n; 2,413 ( 0.00%) while let Some(x) = self.next_back() { 39,656 ( 0.01%) i -= 1; 33,888 ( 0.01%) if predicate(x) { . // SAFETY: `i` must be lower than `n` since it starts at `n` . // and is only decreasing. . unsafe { assume(i < n) }; . return Some(i); . } . } . None . } -- line 326 ---------------------------------------- -- line 332 ---------------------------------------- . // the returned references is guaranteed to refer to an element . // of the slice and thus guaranteed to be valid. . // . // Also note that the caller also guarantees that we're never . // called with the same index again, and that no other methods . // that will access this subslice are called, so it is valid . // for the returned reference to be mutable in the case of . // `IterMut` 2,467 ( 0.00%) unsafe { & $( $mut_ )? * self.ptr.as_ptr().add(idx) } . } . . $($extra)* . } . . #[stable(feature = "rust1", since = "1.0.0")] . impl<'a, T> DoubleEndedIterator for $name<'a, T> { . #[inline] -- line 348 ---------------------------------------- -- line 349 ---------------------------------------- . fn next_back(&mut self) -> Option<$elem> { . // could be implemented with slices, but this avoids bounds checks . . // SAFETY: `assume` calls are safe since a slice's start pointer must be non-null, . // and slices over non-ZSTs must also have a non-null end pointer. . // The call to `next_back_unchecked!` is safe since we check if the iterator is . // empty first. . unsafe { 188 ( 0.00%) assume(!self.ptr.as_ptr().is_null()); . if mem::size_of::() != 0 { 213 ( 0.00%) assume(!self.end.is_null()); . } 48,785 ( 0.01%) if is_empty!(self) { . None . } else { . Some(next_back_unchecked!(self)) . } . } . } . . #[inline] -- line 369 ---------------------------------------- 347,324 ( 0.07%) -------------------------------------------------------------------------------- -- Auto-annotated source: /usr/home/liquid/rust/worktree-benchmarking/compiler/rustc_resolve/src/imports.rs -------------------------------------------------------------------------------- Ir -- line 109 ---------------------------------------- . /// The resolution of `module_path`. . pub imported_module: Cell>>, . pub vis: Cell, . pub used: Cell, . } . . impl<'a> Import<'a> { . pub fn is_glob(&self) -> bool { 89 ( 0.00%) matches!(self.kind, ImportKind::Glob { .. }) . } . . pub fn is_nested(&self) -> bool { . match self.kind { . ImportKind::Single { nested, .. } => nested, . _ => false, . } . } . . crate fn crate_lint(&self) -> CrateLint { 61 ( 0.00%) CrateLint::UsePath { root_id: self.root_id, root_span: self.root_span } 115 ( 0.00%) } . } . 21,264 ( 0.00%) #[derive(Clone, Default, Debug)] . /// Records information about the resolution of a name in a namespace of a module. . pub struct NameResolution<'a> { . /// Single imports that may define the name in the namespace. . /// Imports are arena-allocated, so it's ok to use pointers as keys. . single_imports: FxHashSet>>, . /// The least shadowable known binding for this name, or None if there are no known bindings. . pub binding: Option<&'a NameBinding<'a>>, . shadowed_glob: Option<&'a NameBinding<'a>>, . } . . impl<'a> NameResolution<'a> { . // Returns the binding for the name if it is known or None if it not known. . pub(crate) fn binding(&self) -> Option<&'a NameBinding<'a>> { 6,945 ( 0.00%) self.binding.and_then(|binding| { 32 ( 0.00%) if !binding.is_glob_import() || self.single_imports.is_empty() { . Some(binding) . } else { . None . } . }) . } . . crate fn add_single_import(&mut self, import: &'a Import<'a>) { -- line 155 ---------------------------------------- -- line 168 ---------------------------------------- . .. . }, . ) => import.vis.get().is_public(), . _ => false, . } . } . . impl<'a> Resolver<'a> { 91 ( 0.00%) crate fn resolve_ident_in_module_unadjusted( . &mut self, . module: ModuleOrUniformRoot<'a>, . ident: Ident, . ns: Namespace, . parent_scope: &ParentScope<'a>, . record_used: bool, . path_span: Span, . ) -> Result<&'a NameBinding<'a>, Determinacy> { 2,177 ( 0.00%) self.resolve_ident_in_module_unadjusted_ext( . module, . ident, . ns, . parent_scope, . false, . record_used, . path_span, . ) . .map_err(|(determinacy, _)| determinacy) 26 ( 0.00%) } . . /// Attempts to resolve `ident` in namespaces `ns` of `module`. . /// Invariant: if `record_used` is `Some`, expansion and import resolution must be complete. 16,107 ( 0.00%) crate fn resolve_ident_in_module_unadjusted_ext( . &mut self, . module: ModuleOrUniformRoot<'a>, . ident: Ident, . ns: Namespace, . parent_scope: &ParentScope<'a>, . restricted_shadowing: bool, . record_used: bool, . path_span: Span, -- line 207 ---------------------------------------- -- line 259 ---------------------------------------- . path_span, . ); . return binding.map_err(|determinacy| (determinacy, Weak::No)); . } . }; . . let key = self.new_key(ident, ns); . let resolution = 6,136 ( 0.00%) self.resolution(module, key).try_borrow_mut().map_err(|_| (Determined, Weak::No))?; // This happens when there is a cycle of imports. . 2,935 ( 0.00%) if let Some(binding) = resolution.binding { 306 ( 0.00%) if !restricted_shadowing && binding.expansion != LocalExpnId::ROOT { 106 ( 0.00%) if let NameBindingKind::Res(_, true) = binding.kind { . self.macro_expanded_macro_export_errors.insert((path_span, binding.span)); . } . } . } . 3,077 ( 0.00%) let check_usable = |this: &mut Self, binding: &'a NameBinding<'a>| { 634 ( 0.00%) if let Some(unusable_binding) = this.unusable_binding { . if ptr::eq(binding, unusable_binding) { . return Err((Determined, Weak::No)); . } . } 318 ( 0.00%) let usable = this.is_accessible_from(binding.vis, parent_scope.module); 364 ( 0.00%) if usable { Ok(binding) } else { Err((Determined, Weak::No)) } 11 ( 0.00%) }; . 1,534 ( 0.00%) if record_used { 712 ( 0.00%) return resolution . .binding . .and_then(|binding| { . // If the primary binding is unusable, search further and return the shadowed glob . // binding if it exists. What we really want here is having two separate scopes in . // a module - one for non-globs and one for globs, but until that's done use this . // hack to avoid inconsistent resolution ICEs during import validation. 1,104 ( 0.00%) if let Some(unusable_binding) = self.unusable_binding { . if ptr::eq(binding, unusable_binding) { . return resolution.shadowed_glob; . } . } . Some(binding) . }) . .ok_or((Determined, Weak::No)) . .and_then(|binding| { 464 ( 0.00%) if self.last_import_segment && check_usable(self, binding).is_err() { . Err((Determined, Weak::No)) . } else { 1,308 ( 0.00%) self.record_use(ident, binding, restricted_shadowing); . 654 ( 0.00%) if let Some(shadowed_glob) = resolution.shadowed_glob { . // Forbid expanded shadowing to avoid time travel. . if restricted_shadowing . && binding.expansion != LocalExpnId::ROOT . && binding.res() != shadowed_glob.res() . { . self.ambiguity_errors.push(AmbiguityError { . kind: AmbiguityKind::GlobVsExpanded, . ident, -- line 317 ---------------------------------------- -- line 318 ---------------------------------------- . b1: binding, . b2: shadowed_glob, . misc1: AmbiguityErrorMisc::None, . misc2: AmbiguityErrorMisc::None, . }); . } . } . 1,090 ( 0.00%) if !self.is_accessible_from(binding.vis, parent_scope.module) { . self.privacy_errors.push(PrivacyError { . ident, . binding, . dedup_span: path_span, . }); . } . . Ok(binding) . } . }); . } . . // Items and single imports are not shadowable, if we have one, then it's determined. 1,233 ( 0.00%) if let Some(binding) = resolution.binding { 9 ( 0.00%) if !binding.is_glob_import() { . return check_usable(self, binding); . } . } . . // --- From now on we either have a glob resolution or no resolution. --- . . // Check if one of single imports can still define the name, . // if it can then our result is not determined and can be invalidated. 3,138 ( 0.00%) for single_import in &resolution.single_imports { 24 ( 0.00%) if !self.is_accessible_from(single_import.vis.get(), parent_scope.module) { . continue; . } . let module = match single_import.imported_module.get() { . Some(x) => x, . None => return Err((Undetermined, Weak::No)), . }; . let ident = match single_import.kind { . ImportKind::Single { source, .. } => source, -- line 359 ---------------------------------------- -- line 385 ---------------------------------------- . // Additionally, any macro in any module can plant names in the root module if it creates . // `macro_export` macros, so the root module effectively has unresolved invocations if any . // module has unresolved invocations. . // However, it causes resolution/expansion to stuck too often (#53144), so, to make . // progress, we have to ignore those potential unresolved invocations from other modules . // and prohibit access to macro-expanded `macro_export` macros instead (unless restricted . // shadowing is enabled, see `macro_expanded_macro_export_errors`). . let unexpanded_macros = !module.unexpanded_invocations.borrow().is_empty(); 1,565 ( 0.00%) if let Some(binding) = resolution.binding { 2 ( 0.00%) if !unexpanded_macros || ns == MacroNS || restricted_shadowing { 10 ( 0.00%) return check_usable(self, binding); . } else { . return Err((Undetermined, Weak::No)); . } . } . . // --- From now on we have no resolution. --- . . // Now we are in situation when new item/import can appear only from a glob or a macro . // expansion. With restricted shadowing names from globs and macro expansions cannot . // shadow names from outer scopes, so we can freely fallback from module search to search . // in outer scopes. For `early_resolve_ident_in_lexical_scope` to continue search in outer . // scopes we return `Undetermined` with `Weak::Yes`. . . // Check if one of unexpanded macros can still define the name, . // if it can then our "no resolution" result is not determined and can be invalidated. 312 ( 0.00%) if unexpanded_macros { . return Err((Undetermined, Weak::Yes)); . } . . // Check if one of glob imports can still define the name, . // if it can then our "no resolution" result is not determined and can be invalidated. 1,144 ( 0.00%) for glob_import in module.globs.borrow().iter() { 68 ( 0.00%) if !self.is_accessible_from(glob_import.vis.get(), parent_scope.module) { . continue; . } 76 ( 0.00%) let module = match glob_import.imported_module.get() { . Some(ModuleOrUniformRoot::Module(module)) => module, . Some(_) => continue, . None => return Err((Undetermined, Weak::Yes)), . }; . let tmp_parent_scope; 26 ( 0.00%) let (mut adjusted_parent_scope, mut ident) = 26 ( 0.00%) (parent_scope, ident.normalize_to_macros_2_0()); 104 ( 0.00%) match ident.span.glob_adjust(module.expansion, glob_import.span) { . Some(Some(def)) => { . tmp_parent_scope = . ParentScope { module: self.expn_def_scope(def), ..*parent_scope }; . adjusted_parent_scope = &tmp_parent_scope; . } . Some(None) => {} . None => continue, . }; 91 ( 0.00%) let result = self.resolve_ident_in_module_unadjusted( . ModuleOrUniformRoot::Module(module), . ident, 65 ( 0.00%) ns, . adjusted_parent_scope, . false, . path_span, . ); . 52 ( 0.00%) match result { . Err(Determined) => continue, . Ok(binding) . if !self.is_accessible_from(binding.vis, glob_import.parent_scope.module) => . { . continue; . } . Ok(_) | Err(Undetermined) => return Err((Undetermined, Weak::Yes)), . } . } . . // No resolution and no one else can define the name - determinate error. . Err((Determined, Weak::No)) 13,806 ( 0.00%) } . . // Given a binding and an import that resolves to it, . // return the corresponding binding defined by the import. 710 ( 0.00%) crate fn import( . &self, . binding: &'a NameBinding<'a>, . import: &'a Import<'a>, . ) -> &'a NameBinding<'a> { 710 ( 0.00%) let vis = if binding.vis.is_at_least(import.vis.get(), self) . || pub_use_of_private_extern_crate_hack(import, binding) . { . import.vis.get() . } else { . binding.vis . }; . 213 ( 0.00%) if let ImportKind::Glob { ref max_vis, .. } = import.kind { 5 ( 0.00%) if vis == import.vis.get() || vis.is_at_least(max_vis.get(), self) { . max_vis.set(vis) . } . } . 71 ( 0.00%) self.arenas.alloc_name_binding(NameBinding { . kind: NameBindingKind::Import { binding, import, used: Cell::new(false) }, . ambiguity: None, 71 ( 0.00%) span: import.span, . vis, 71 ( 0.00%) expansion: import.parent_scope.expansion, . }) 568 ( 0.00%) } . . // Define the name or return the existing binding if there is a collision. 82,044 ( 0.02%) crate fn try_define( . &mut self, . module: Module<'a>, . key: BindingKey, . binding: &'a NameBinding<'a>, . ) -> Result<(), &'a NameBinding<'a>> { . let res = binding.res(); 123,066 ( 0.03%) self.check_reserved_macro_name(key.ident, res); . self.set_binding_parent_module(binding, module); 41,022 ( 0.01%) self.update_resolution(module, key, |this, resolution| { . if let Some(old_binding) = resolution.binding { . if res == Res::Err { . // Do not override real bindings with `Res::Err`s from error recovery. . return Ok(()); . } . match (old_binding.is_glob_import(), binding.is_glob_import()) { . (true, true) => { . if res != old_binding.res() { -- line 510 ---------------------------------------- -- line 535 ---------------------------------------- . } . resolution.shadowed_glob = Some(glob_binding); . } . (false, false) => { . return Err(old_binding); . } . } . } else { 20,511 ( 0.00%) resolution.binding = Some(binding); . } . . Ok(()) . }) 61,533 ( 0.01%) } . . fn ambiguity( . &self, . kind: AmbiguityKind, . primary_binding: &'a NameBinding<'a>, . secondary_binding: &'a NameBinding<'a>, . ) -> &'a NameBinding<'a> { . self.arenas.alloc_name_binding(NameBinding { -- line 556 ---------------------------------------- -- line 563 ---------------------------------------- . // If the resolution becomes a success, define it in the module's glob importers. . fn update_resolution(&mut self, module: Module<'a>, key: BindingKey, f: F) -> T . where . F: FnOnce(&mut Resolver<'a>, &mut NameResolution<'a>) -> T, . { . // Ensure that `resolution` isn't borrowed when defining in the module's glob importers, . // during which the resolution might end up getting re-defined via a glob cycle. . let (binding, t) = { 109,616 ( 0.02%) let resolution = &mut *self.resolution(module, key).borrow_mut(); . let old_binding = resolution.binding(); . . let t = f(self, resolution); . . match resolution.binding() { . _ if old_binding.is_some() => return t, . None => return t, . Some(binding) => match old_binding { -- line 579 ---------------------------------------- -- line 580 ---------------------------------------- . Some(old_binding) if ptr::eq(old_binding, binding) => return t, . _ => (binding, t), . }, . } . }; . . // Define `binding` in `module`s glob importers. . for import in module.glob_importers.borrow_mut().iter() { 8 ( 0.00%) let mut ident = key.ident; 18 ( 0.00%) let scope = match ident.span.reverse_glob_adjust(module.expansion, import.span) { . Some(Some(def)) => self.expn_def_scope(def), 4 ( 0.00%) Some(None) => import.parent_scope.module, . None => continue, . }; 4 ( 0.00%) if self.is_accessible_from(binding.vis, scope) { 4 ( 0.00%) let imported_binding = self.import(binding, import); . let key = BindingKey { ident, ..key }; 15 ( 0.00%) let _ = self.try_define(import.parent_scope.module, key, imported_binding); . } . } . . t . } . . // Define a "dummy" resolution containing a Res::Err as a placeholder for a . // failed resolution -- line 605 ---------------------------------------- -- line 643 ---------------------------------------- . // This is a fixed-point algorithm. We resolve imports until our efforts . // are stymied by an unresolved import; then we bail out of the current . // module and continue. We terminate successfully once no more imports . // remain or unsuccessfully when no forward progress in resolving imports . // is made. . . /// Resolves all imports for the crate. This method performs the fixed- . /// point iteration. 16 ( 0.00%) pub fn resolve_imports(&mut self) { 10 ( 0.00%) let mut prev_num_indeterminates = self.r.indeterminate_imports.len() + 1; 29 ( 0.00%) while self.r.indeterminate_imports.len() < prev_num_indeterminates { . prev_num_indeterminates = self.r.indeterminate_imports.len(); 50 ( 0.00%) for import in mem::take(&mut self.r.indeterminate_imports) { . match self.resolve_import(&import) { 20 ( 0.00%) true => self.r.determined_imports.push(import), 8 ( 0.00%) false => self.r.indeterminate_imports.push(import), . } . } . } 16 ( 0.00%) } . 7 ( 0.00%) pub fn finalize_imports(&mut self) { 2 ( 0.00%) for module in self.r.arenas.local_modules().iter() { 10 ( 0.00%) self.finalize_resolutions_in(module); . } . . let mut seen_spans = FxHashSet::default(); . let mut errors = vec![]; . let mut prev_root_id: NodeId = NodeId::from_u32(0); 1 ( 0.00%) let determined_imports = mem::take(&mut self.r.determined_imports); 1 ( 0.00%) let indeterminate_imports = mem::take(&mut self.r.indeterminate_imports); . 12 ( 0.00%) for (is_indeterminate, import) in determined_imports . .into_iter() . .map(|i| (false, i)) . .chain(indeterminate_imports.into_iter().map(|i| (true, i))) . { 102 ( 0.00%) if let Some(err) = self.finalize_import(import) { . if let ImportKind::Single { source, ref source_bindings, .. } = import.kind { . if source.name == kw::SelfLower { . // Silence `unresolved import` error if E0429 is already emitted . if let Err(Determined) = source_bindings.value_ns.get() { . continue; . } . } . } -- line 688 ---------------------------------------- -- line 703 ---------------------------------------- . let path = import_path_to_string( . &import.module_path.iter().map(|seg| seg.ident).collect::>(), . &import.kind, . err.span, . ); . errors.push((path, err)); . prev_root_id = import.root_id; . } 34 ( 0.00%) } else if is_indeterminate { . // Consider erroneous imports used to avoid duplicate diagnostics. . self.r.used_imports.insert(import.id); . let path = import_path_to_string( . &import.module_path.iter().map(|seg| seg.ident).collect::>(), . &import.kind, . import.span, . ); . let err = UnresolvedImportError { -- line 719 ---------------------------------------- -- line 721 ---------------------------------------- . label: None, . note: Vec::new(), . suggestion: None, . }; . errors.push((path, err)); . } . } . 1 ( 0.00%) if !errors.is_empty() { . self.throw_unresolved_import_error(errors, None); . } 8 ( 0.00%) } . . fn throw_unresolved_import_error( . &self, . errors: Vec<(String, UnresolvedImportError)>, . span: Option, . ) { . /// Upper limit on the number of `span_label` messages. . const MAX_LABEL_COUNT: usize = 10; -- line 740 ---------------------------------------- -- line 776 ---------------------------------------- . /// If successful, the resolved bindings are written into the module. . fn resolve_import(&mut self, import: &'b Import<'b>) -> bool { . debug!( . "(resolving import for module) resolving import `{}::...` in `{}`", . Segment::names_to_string(&import.module_path), . module_to_string(import.parent_scope.module).unwrap_or_else(|| "???".to_string()), . ); . 83 ( 0.00%) let module = if let Some(module) = import.imported_module.get() { 16 ( 0.00%) module . } else { . // For better failure detection, pretend that the import will . // not define any names while resolving its module path. . let orig_vis = import.vis.replace(ty::Visibility::Invisible); . let path_res = self.r.resolve_path( . &import.module_path, . None, 17 ( 0.00%) &import.parent_scope, . false, . import.span, . import.crate_lint(), . ); . import.vis.set(orig_vis); . 119 ( 0.00%) match path_res { 85 ( 0.00%) PathResult::Module(module) => module, . PathResult::Indeterminate => return false, . PathResult::NonModule(..) | PathResult::Failed { .. } => return true, . } . }; . . import.imported_module.set(Some(module)); 279 ( 0.00%) let (source, target, source_bindings, target_bindings, type_ns_only) = match import.kind { . ImportKind::Single { . source, . target, 22 ( 0.00%) ref source_bindings, 22 ( 0.00%) ref target_bindings, 22 ( 0.00%) type_ns_only, . .. . } => (source, target, source_bindings, target_bindings, type_ns_only), . ImportKind::Glob { .. } => { . self.resolve_glob_import(import); . return true; . } . _ => unreachable!(), . }; . 22 ( 0.00%) let mut indeterminate = false; 946 ( 0.00%) self.r.per_ns(|this, ns| { 198 ( 0.00%) if !type_ns_only || ns == TypeNS { 268 ( 0.00%) if let Err(Undetermined) = source_bindings[ns].get() { . // For better failure detection, pretend that the import will . // not define any names while resolving its module path. 62 ( 0.00%) let orig_vis = import.vis.replace(ty::Visibility::Invisible); . let binding = this.resolve_ident_in_module( 186 ( 0.00%) module, 186 ( 0.00%) source, . ns, 62 ( 0.00%) &import.parent_scope, . false, 62 ( 0.00%) import.span, . ); . import.vis.set(orig_vis); . source_bindings[ns].set(binding); . } else { . return; . }; . 62 ( 0.00%) let parent = import.parent_scope.module; 158 ( 0.00%) match source_bindings[ns].get() { 60 ( 0.00%) Err(Undetermined) => indeterminate = true, . // Don't update the resolution, because it was never added. 56 ( 0.00%) Err(Determined) if target.name == kw::Underscore => {} . Err(Determined) => { 56 ( 0.00%) let key = this.new_key(target, ns); . this.update_resolution(parent, key, |_, resolution| { 56 ( 0.00%) resolution.single_imports.remove(&PtrKey(import)); . }); . } . Ok(binding) if !binding.is_importable() => { . let msg = format!("`{}` is not directly importable", target); . struct_span_err!(this.session, import.span, E0253, "{}", &msg) . .span_label(import.span, "cannot be imported directly") . .emit(); . // Do not import this illegal binding. Import a dummy binding and pretend . // everything is fine . this.import_dummy_binding(import); . } . Ok(binding) => { 56 ( 0.00%) let imported_binding = this.import(binding, import); 14 ( 0.00%) target_bindings[ns].set(Some(imported_binding)); 42 ( 0.00%) this.define(parent, target, ns, imported_binding); . } . } . } 528 ( 0.00%) }); . 58 ( 0.00%) !indeterminate . } . . /// Performs final import resolution, consistency checks and error reporting. . /// . /// Optionally returns an unresolved import error. This error is buffered and used to . /// consolidate multiple unresolved import errors into a single diagnostic. 170 ( 0.00%) fn finalize_import(&mut self, import: &'b Import<'b>) -> Option { . let orig_vis = import.vis.replace(ty::Visibility::Invisible); 68 ( 0.00%) let orig_unusable_binding = match &import.kind { . ImportKind::Single { target_bindings, .. } => { 28 ( 0.00%) Some(mem::replace(&mut self.r.unusable_binding, target_bindings[TypeNS].get())) . } . _ => None, . }; 34 ( 0.00%) let prev_ambiguity_errors_len = self.r.ambiguity_errors.len(); . let path_res = self.r.resolve_path( . &import.module_path, . None, 17 ( 0.00%) &import.parent_scope, . true, 17 ( 0.00%) import.span, . import.crate_lint(), . ); 17 ( 0.00%) let no_ambiguity = self.r.ambiguity_errors.len() == prev_ambiguity_errors_len; 17 ( 0.00%) if let Some(orig_unusable_binding) = orig_unusable_binding { 14 ( 0.00%) self.r.unusable_binding = orig_unusable_binding; . } . import.vis.set(orig_vis); 85 ( 0.00%) if let PathResult::Failed { .. } | PathResult::NonModule(..) = path_res { . // Consider erroneous imports used to avoid duplicate diagnostics. . self.r.used_imports.insert(import.id); . } . let module = match path_res { 51 ( 0.00%) PathResult::Module(module) => { . // Consistency checks, analogous to `finalize_macro_resolutions`. 51 ( 0.00%) if let Some(initial_module) = import.imported_module.get() { 51 ( 0.00%) if !ModuleOrUniformRoot::same_def(module, initial_module) && no_ambiguity { . span_bug!(import.span, "inconsistent resolution for an import"); . } . } else if self.r.privacy_errors.is_empty() { . let msg = "cannot determine resolution for the import"; . let msg_note = "import resolution is stuck, try simplifying other imports"; . self.r.session.struct_span_err(import.span, msg).note(msg_note).emit(); . } . 51 ( 0.00%) module . } . PathResult::Failed { is_error_from_last_segment: false, span, label, suggestion } => { . if no_ambiguity { . assert!(import.imported_module.get().is_none()); . self.r . .report_error(span, ResolutionError::FailedToResolve { label, suggestion }); . } . return None; -- line 928 ---------------------------------------- -- line 961 ---------------------------------------- . assert!(import.imported_module.get().is_none()); . } . // The error was already reported earlier. . return None; . } . PathResult::Indeterminate => unreachable!(), . }; . 183 ( 0.00%) let (ident, target, source_bindings, target_bindings, type_ns_only) = match import.kind { . ImportKind::Single { . source, . target, 14 ( 0.00%) ref source_bindings, 14 ( 0.00%) ref target_bindings, 14 ( 0.00%) type_ns_only, . .. . } => (source, target, source_bindings, target_bindings, type_ns_only), 6 ( 0.00%) ImportKind::Glob { is_prelude, ref max_vis } => { 9 ( 0.00%) if import.module_path.len() <= 1 { . // HACK(eddyb) `lint_if_path_starts_with_module` needs at least . // 2 segments, so the `resolve_path` above won't trigger it. . let mut full_path = import.module_path.clone(); . full_path.push(Segment::from_ident(Ident::empty())); 4 ( 0.00%) self.r.lint_if_path_starts_with_module( . import.crate_lint(), . &full_path, 1 ( 0.00%) import.span, . None, . ); . } . 13 ( 0.00%) if let ModuleOrUniformRoot::Module(module) = module { 3 ( 0.00%) if ptr::eq(module, import.parent_scope.module) { . // Importing a module into itself is not allowed. . return Some(UnresolvedImportError { . span: import.span, . label: Some(String::from("cannot glob-import a module into itself")), . note: Vec::new(), . suggestion: None, . }); . } . } 7 ( 0.00%) if !is_prelude && 2 ( 0.00%) max_vis.get() != ty::Visibility::Invisible && // Allow empty globs. 2 ( 0.00%) !max_vis.get().is_at_least(import.vis.get(), &*self) . { . let msg = "glob import doesn't reexport anything because no candidate is public enough"; . self.r.lint_buffer.buffer_lint(UNUSED_IMPORTS, import.id, import.span, msg); . } . return None; . } . _ => unreachable!(), . }; . 28 ( 0.00%) let mut all_ns_err = true; 644 ( 0.00%) self.r.per_ns(|this, ns| { 126 ( 0.00%) if !type_ns_only || ns == TypeNS { 42 ( 0.00%) let orig_vis = import.vis.replace(ty::Visibility::Invisible); . let orig_unusable_binding = 84 ( 0.00%) mem::replace(&mut this.unusable_binding, target_bindings[ns].get()); . let orig_last_import_segment = mem::replace(&mut this.last_import_segment, true); . let binding = this.resolve_ident_in_module( 126 ( 0.00%) module, 168 ( 0.00%) ident, . ns, 42 ( 0.00%) &import.parent_scope, . true, . import.span, . ); 84 ( 0.00%) this.last_import_segment = orig_last_import_segment; 84 ( 0.00%) this.unusable_binding = orig_unusable_binding; . import.vis.set(orig_vis); . . match binding { . Ok(binding) => { . // Consistency checks, analogous to `finalize_macro_resolutions`. 42 ( 0.00%) let initial_res = source_bindings[ns].get().map(|initial_binding| { 14 ( 0.00%) all_ns_err = false; 42 ( 0.00%) if let Some(target_binding) = target_bindings[ns].get() { 14 ( 0.00%) if target.name == kw::Underscore . && initial_binding.is_extern_crate() . && !initial_binding.is_import() . { . this.record_use( . ident, . target_binding, . import.module_path.is_empty(), . ); . } . } . initial_binding.res() . }); . let res = binding.res(); 196 ( 0.00%) if let Ok(initial_res) = initial_res { 42 ( 0.00%) if res != initial_res && this.ambiguity_errors.is_empty() { . span_bug!(import.span, "inconsistent resolution for an import"); . } . } else if res != Res::Err . && this.ambiguity_errors.is_empty() . && this.privacy_errors.is_empty() . { . let msg = "cannot determine resolution for the import"; . let msg_note = -- line 1063 ---------------------------------------- -- line 1070 ---------------------------------------- . // single import (see test `issue-55884-2.rs`). In theory single imports should . // always block globs, even if they are not yet resolved, so that this kind of . // self-inconsistent resolution never happens. . // Re-enable the assert when the issue is fixed. . // assert!(result[ns].get().is_err()); . } . } . } 336 ( 0.00%) }); . 28 ( 0.00%) if all_ns_err { . let mut all_ns_failed = true; . self.r.per_ns(|this, ns| { . if !type_ns_only || ns == TypeNS { . let binding = this.resolve_ident_in_module( . module, . ident, . ns, . &import.parent_scope, -- line 1088 ---------------------------------------- -- line 1175 ---------------------------------------- . }) . } else { . // `resolve_ident_in_module` reported a privacy error. . self.r.import_dummy_binding(import); . None . }; . } . 14 ( 0.00%) let mut reexport_error = None; 14 ( 0.00%) let mut any_successful_reexport = false; 28 ( 0.00%) let mut crate_private_reexport = false; 168 ( 0.00%) self.r.per_ns(|this, ns| { 98 ( 0.00%) if let Ok(binding) = source_bindings[ns].get() { 42 ( 0.00%) let vis = import.vis.get(); 84 ( 0.00%) if !binding.vis.is_at_least(vis, &*this) { . reexport_error = Some((ns, binding)); . if let ty::Visibility::Restricted(binding_def_id) = binding.vis { . if binding_def_id.is_top_level_module() { . crate_private_reexport = true; . } . } . } else { 28 ( 0.00%) any_successful_reexport = true; . } . } . }); . . // All namespaces must be re-exported with extra visibility for an error to occur. 28 ( 0.00%) if !any_successful_reexport { . let (ns, binding) = reexport_error.unwrap(); . if pub_use_of_private_extern_crate_hack(import, binding) { . let msg = format!( . "extern crate `{}` is private, and cannot be \ . re-exported (error E0365), consider declaring with \ . `pub`", . ident . ); -- line 1211 ---------------------------------------- -- line 1241 ---------------------------------------- . format!("consider marking `{}` as `pub` in the imported module", ident); . struct_span_err!(self.r.session, import.span, E0364, "{}", error_msg) . .span_note(import.span, ¬e_msg) . .emit(); . } . } . } . 42 ( 0.00%) if import.module_path.len() <= 1 { . // HACK(eddyb) `lint_if_path_starts_with_module` needs at least . // 2 segments, so the `resolve_path` above won't trigger it. . let mut full_path = import.module_path.clone(); 27 ( 0.00%) full_path.push(Segment::from_ident(ident)); 63 ( 0.00%) self.r.per_ns(|this, ns| { 54 ( 0.00%) if let Ok(binding) = source_bindings[ns].get() { 27 ( 0.00%) this.lint_if_path_starts_with_module( . import.crate_lint(), . &full_path, 9 ( 0.00%) import.span, . Some(binding), . ); . } . }); . } . . // Record what this import resolves to for later uses in documentation, . // this may resolve to either a value or a type, but for documentation . // purposes it's good enough to just favor one over the other. 392 ( 0.00%) self.r.per_ns(|this, ns| { 168 ( 0.00%) if let Ok(binding) = source_bindings[ns].get() { 224 ( 0.00%) this.import_res_map.entry(import.id).or_default()[ns] = Some(binding.res()); . } 336 ( 0.00%) }); . 126 ( 0.00%) self.check_for_redundant_imports(ident, import, source_bindings, target_bindings, target); . . debug!("(resolving single import) successfully resolved import"); . None 136 ( 0.00%) } . . fn check_for_redundant_imports( . &mut self, . ident: Ident, . import: &'b Import<'b>, . source_bindings: &PerNS, Determinacy>>>, . target_bindings: &PerNS>>>, . target: Ident, . ) { . // Skip if the import was produced by a macro. 14 ( 0.00%) if import.parent_scope.expansion != LocalExpnId::ROOT { . return; . } . . // Skip if we are inside a named module (in contrast to an anonymous . // module defined by a block). 42 ( 0.00%) if let ModuleKind::Def(..) = import.parent_scope.module.kind { . return; . } . 6 ( 0.00%) let mut is_redundant = PerNS { value_ns: None, type_ns: None, macro_ns: None }; . 12 ( 0.00%) let mut redundant_span = PerNS { value_ns: None, type_ns: None, macro_ns: None }; . 132 ( 0.00%) self.r.per_ns(|this, ns| { 33 ( 0.00%) if let Ok(binding) = source_bindings[ns].get() { 15 ( 0.00%) if binding.res() == Res::Err { . return; . } . . let orig_unusable_binding = 6 ( 0.00%) mem::replace(&mut this.unusable_binding, target_bindings[ns].get()); . 36 ( 0.00%) match this.early_resolve_ident_in_lexical_scope( 9 ( 0.00%) target, . ScopeSet::All(ns, false), 6 ( 0.00%) &import.parent_scope, . false, . false, . import.span, . ) { . Ok(other_binding) => { . is_redundant[ns] = Some( . binding.res() == other_binding.res() && !other_binding.is_ambiguity(), . ); . redundant_span[ns] = Some((other_binding.span, other_binding.is_import())); . } 6 ( 0.00%) Err(_) => is_redundant[ns] = Some(false), . } . 6 ( 0.00%) this.unusable_binding = orig_unusable_binding; . } 72 ( 0.00%) }); . 24 ( 0.00%) if !is_redundant.is_empty() && is_redundant.present_items().all(|is_redundant| is_redundant) . { . let mut redundant_spans: Vec<_> = redundant_span.present_items().collect(); . redundant_spans.sort(); . redundant_spans.dedup(); . self.r.lint_buffer.buffer_lint_with_diagnostic( . UNUSED_IMPORTS, . import.id, . import.span, -- line 1342 ---------------------------------------- -- line 1350 ---------------------------------------- . let module = match import.imported_module.get().unwrap() { . ModuleOrUniformRoot::Module(module) => module, . _ => { . self.r.session.span_err(import.span, "cannot glob-import all possible crates"); . return; . } . }; . 9 ( 0.00%) if module.is_trait() { . self.r.session.span_err(import.span, "items in traits are not importable"); . return; 3 ( 0.00%) } else if ptr::eq(module, import.parent_scope.module) { . return; 12 ( 0.00%) } else if let ImportKind::Glob { is_prelude: true, .. } = import.kind { 2 ( 0.00%) self.r.prelude = Some(module); . return; . } . . // Add to module's glob_importers . module.glob_importers.borrow_mut().push(import); . . // Ensure that `resolutions` isn't borrowed during `try_define`, . // since it might get updated via a glob cycle. 6 ( 0.00%) let bindings = self . .r . .resolutions(module) . .borrow() . .iter() . .filter_map(|(key, resolution)| { 51 ( 0.00%) resolution.borrow().binding().map(|binding| (*key, binding)) . }) . .collect::>(); 20 ( 0.00%) for (mut key, binding) in bindings { 90 ( 0.00%) let scope = match key.ident.span.reverse_glob_adjust(module.expansion, import.span) { . Some(Some(def)) => self.r.expn_def_scope(def), 20 ( 0.00%) Some(None) => import.parent_scope.module, . None => continue, . }; 80 ( 0.00%) if self.r.is_accessible_from(binding.vis, scope) { 8 ( 0.00%) let imported_binding = self.r.import(binding, import); 44 ( 0.00%) let _ = self.r.try_define(import.parent_scope.module, key, imported_binding); . } . } . . // Record the destination of this import . self.r.record_partial_res(import.id, PartialRes::new(module.res().unwrap())); . } . . // Miscellaneous post-processing, including recording re-exports, . // reporting conflicts, and reporting unresolved imports. . fn finalize_resolutions_in(&mut self, module: Module<'b>) { . // Since import resolution is finished, globs will not define any more names. 40 ( 0.00%) *module.globs.borrow_mut() = Vec::new(); . . let mut reexports = Vec::new(); . 50 ( 0.00%) module.for_each_child(self.r, |_, ident, _, binding| { . // Filter away ambiguous imports and anything that has def-site hygiene. . // FIXME: Implement actual cross-crate hygiene. . let is_good_import = 67 ( 0.00%) binding.is_import() && !binding.is_ambiguity() && !ident.span.from_expansion(); 134 ( 0.00%) if is_good_import || binding.is_macro_def() { . let res = binding.res().expect_non_local(); . if res != def::Res::Err { 76 ( 0.00%) reexports.push(ModChild { ident, res, vis: binding.vis, span: binding.span }); . } . } . }); . 10 ( 0.00%) if !reexports.is_empty() { 17 ( 0.00%) if let Some(def_id) = module.opt_def_id() { . // Call to `expect_local` should be fine because current . // code is only called for local modules. . self.r.reexport_map.insert(def_id.expect_local(), reexports); . } . } . } . } . -- line 1428 ---------------------------------------- 45,251 ( 0.01%) -------------------------------------------------------------------------------- The following files chosen for auto-annotation could not be found: -------------------------------------------------------------------------------- ./elf/../sysdeps/x86_64/dl-machine.h ./elf/dl-lookup.c ./elf/do-rel.h ./malloc/malloc.c ./nptl/pthread_rwlock_common.c ./string/../sysdeps/x86_64/multiarch/memcmp-avx2-movbe.S ./string/../sysdeps/x86_64/multiarch/memmove-vec-unaligned-erms.S ./string/../sysdeps/x86_64/multiarch/memset-vec-unaligned-erms.S ./string/../sysdeps/x86_64/multiarch/strcmp-avx2.S ./string/../sysdeps/x86_64/multiarch/strlen-avx2.S ./string/../sysdeps/x86_64/strcmp.S /tmp/gcc-build/x86_64-unknown-linux-gnu/libstdc++-v3/libsupc++/../../../../gcc-5.5.0/libstdc++-v3/libsupc++/new_op.cc -------------------------------------------------------------------------------- Ir -------------------------------------------------------------------------------- 48,804,324 (10.31%) events annotated