-------------------------------------------------------------------------------- I1 cache: 65536 B, 64 B, 4-way associative D1 cache: 32768 B, 64 B, 8-way associative LL cache: 67108864 B, 64 B, 64-way associative Command: /usr/home/liquid/.rustup/toolchains/w-profiling/bin/rustc --crate-name wasm_bindgen_test --edition=2018 src/lib.rs --error-format=json --json=diagnostic-rendered-ansi,artifacts,future-incompat --crate-type lib --emit=dep-info,metadata,link -C embed-bitcode=no -C debuginfo=2 -C metadata=345b49354309a51d -C extra-filename=-345b49354309a51d --out-dir /usr/home/liquid/tmp/.tmpCkLpfM/target/debug/deps -L dependency=/usr/home/liquid/tmp/.tmpCkLpfM/target/debug/deps --extern console_error_panic_hook=/usr/home/liquid/tmp/.tmpCkLpfM/target/debug/deps/libconsole_error_panic_hook-d4455a82a29bb2b7.rmeta --extern js_sys=/usr/home/liquid/tmp/.tmpCkLpfM/target/debug/deps/libjs_sys-18ed9659890a73d6.rmeta --extern scoped_tls=/usr/home/liquid/tmp/.tmpCkLpfM/target/debug/deps/libscoped_tls-4a19484de14f3899.rmeta --extern wasm_bindgen=/usr/home/liquid/tmp/.tmpCkLpfM/target/debug/deps/libwasm_bindgen-257a3b187db57b05.rmeta --extern wasm_bindgen_futures=/usr/home/liquid/tmp/.tmpCkLpfM/target/debug/deps/libwasm_bindgen_futures-54536c0591d6c006.rmeta --extern wasm_bindgen_test_macro=/usr/home/liquid/tmp/.tmpCkLpfM/target/debug/deps/libwasm_bindgen_test_macro-16d4b10740ae8307.so -Adeprecated -Aunknown-lints -Zincremental-verify-ich Data file: results/cgout-w-profiling-wasm-bindgen-test-0.3.29-Debug-Full Events recorded: Ir Events shown: Ir Event sort order: Ir Thresholds: 0.1 Include dirs: User annotated: Auto-annotation: on -------------------------------------------------------------------------------- Ir -------------------------------------------------------------------------------- 2,528,315,059 (100.0%) PROGRAM TOTALS -------------------------------------------------------------------------------- Ir file:function -------------------------------------------------------------------------------- 126,352,473 ( 5.00%) ???:llvm::FPPassManager::runOnFunction(llvm::Function&) 70,201,193 ( 2.78%) ./malloc/malloc.c:_int_free 69,520,149 ( 2.75%) ./malloc/malloc.c:_int_malloc 56,210,941 ( 2.22%) ???:llvm::SelectionDAG::Combine(llvm::CombineLevel, llvm::AAResults*, llvm::CodeGenOpt::Level) 42,648,784 ( 1.69%) ./malloc/malloc.c:malloc 27,602,979 ( 1.09%) ???:llvm::AttributeList::addAttributes(llvm::LLVMContext&, unsigned int, llvm::AttrBuilder const&) const 26,358,395 ( 1.04%) ???:llvm::PMDataManager::verifyPreservedAnalysis(llvm::Pass*) 24,696,717 ( 0.98%) ./string/../sysdeps/x86_64/multiarch/memmove-vec-unaligned-erms.S:__memcpy_avx_unaligned_erms 24,345,666 ( 0.96%) ???:(anonymous namespace)::Verifier::visitInstruction(llvm::Instruction&) 23,116,864 ( 0.91%) ./elf/dl-lookup.c:_dl_lookup_symbol_x 21,819,666 ( 0.86%) ???:llvm::raw_svector_ostream::write_impl(char const*, unsigned long) 20,796,889 ( 0.82%) ./malloc/malloc.c:free 20,653,113 ( 0.82%) ???:llvm::MachineInstr::addOperand(llvm::MachineFunction&, llvm::MachineOperand const&) 20,451,773 ( 0.81%) ???:(anonymous namespace)::Verifier::visitMDNode(llvm::MDNode const&, (anonymous namespace)::Verifier::AreDebugLocsAllowed) 19,470,905 ( 0.77%) ???:(anonymous namespace)::RegAllocFast::allocateBasicBlock(llvm::MachineBasicBlock&) 17,245,812 ( 0.68%) ???:llvm::SelectionDAGISel::SelectCodeCommon(llvm::SDNode*, unsigned char const*, unsigned int) 17,109,393 ( 0.68%) ???:llvm::MCExpr::evaluateAsRelocatableImpl(llvm::MCValue&, llvm::MCAssembler const*, llvm::MCAsmLayout const*, llvm::MCFixup const*, llvm::DenseMap, llvm::detail::DenseMapPair > const*, bool) const 17,089,617 ( 0.68%) ./elf/dl-lookup.c:do_lookup_x 15,744,687 ( 0.62%) ./string/../sysdeps/x86_64/multiarch/memset-vec-unaligned-erms.S:__memset_avx2_erms 12,256,008 ( 0.48%) ???:llvm::ScheduleDAGSDNodes::BuildSchedUnits() 11,062,256 ( 0.44%) ???:llvm::StringMapImpl::LookupBucketFor(llvm::StringRef) 10,906,900 ( 0.43%) ???:llvm::MCAsmLayout::getSymbolOffset(llvm::MCSymbol const&) const 10,842,127 ( 0.43%) ???:(anonymous namespace)::VarLocBasedLDV::process(llvm::MachineInstr&, (anonymous namespace)::VarLocBasedLDV::OpenRangesSet&, (anonymous namespace)::VarLocBasedLDV::VarLocMap&, llvm::SmallVector<(anonymous namespace)::VarLocBasedLDV::TransferDebugPair, 4u>&) 10,570,123 ( 0.42%) ???:llvm::coro::declaresIntrinsics(llvm::Module const&, std::initializer_list) 10,494,618 ( 0.42%) ???:llvm::TargetRegisterInfo::checkAllSuperRegsMarked(llvm::BitVector const&, llvm::ArrayRef) const 10,157,868 ( 0.40%) ???:llvm::TargetLoweringBase::getTypeConversion(llvm::LLVMContext&, llvm::EVT) const 9,904,196 ( 0.39%) ./string/../sysdeps/x86_64/multiarch/memcmp-avx2-movbe.S:__memcmp_avx2_movbe 9,816,998 ( 0.39%) ./malloc/malloc.c:malloc_consolidate 9,781,742 ( 0.39%) ???:llvm::MCObjectStreamer::emitBytes(llvm::StringRef) 9,554,787 ( 0.38%) ./string/../sysdeps/x86_64/multiarch/strcmp-avx2.S:__strncmp_avx2 8,592,636 ( 0.34%) ???:llvm::TargetRegisterInfo::shouldRealignStack(llvm::MachineFunction const&) const 8,311,203 ( 0.33%) ???:llvm::MCAsmLayout::getFragmentOffset(llvm::MCFragment const*) const 8,069,697 ( 0.32%) ./string/../sysdeps/x86_64/multiarch/memmove-vec-unaligned-erms.S:__memcpy_sse2_unaligned_erms 8,022,823 ( 0.32%) ???:multikeySort(llvm::MutableArrayRef*>, int) 7,900,753 ( 0.31%) ./malloc/malloc.c:unlink_chunk.constprop.0 7,872,872 ( 0.31%) ???:llvm::SmallSet >::insert(llvm::Register const&) 7,782,944 ( 0.31%) ???:(anonymous namespace)::Verifier::verifyFunctionAttrs(llvm::FunctionType*, llvm::AttributeList, llvm::Value const*, bool) 7,621,397 ( 0.30%) ???:(anonymous namespace)::Verifier::verify(llvm::Function const&) [clone .llvm.4153962086227604281] 7,608,484 ( 0.30%) ???:(anonymous namespace)::Verifier::visitCallBase(llvm::CallBase&) 7,464,693 ( 0.30%) ???:(anonymous namespace)::TwoAddressInstructionPass::runOnMachineFunction(llvm::MachineFunction&) 7,234,888 ( 0.29%) ???:llvm::DAGTypeLegalizer::run() 7,234,086 ( 0.29%) ???:llvm::InstrEmitter::EmitMachineNode(llvm::SDNode*, bool, bool, llvm::DenseMap, llvm::detail::DenseMapPair >&) 7,188,201 ( 0.28%) ???:llvm::FoldingSetNodeID::AddInteger(unsigned int) 7,096,636 ( 0.28%) ???:llvm::FoldingSetBase::FindNodeOrInsertPos(llvm::FoldingSetNodeID const&, void*&, llvm::FoldingSetBase::FoldingSetInfo const&) 7,087,032 ( 0.28%) ???:(anonymous namespace)::SelectionDAGLegalize::LegalizeOp(llvm::SDNode*) [clone .llvm.8386621111310650999] 7,024,145 ( 0.28%) ???:llvm::SelectionDAG::MorphNodeTo(llvm::SDNode*, unsigned int, llvm::SDVTList, llvm::ArrayRef) 6,899,487 ( 0.27%) ???:llvm::DomTreeBuilder::SemiNCAInfo >::CalculateFromScratch(llvm::DominatorTreeBase&, llvm::DomTreeBuilder::SemiNCAInfo >::BatchUpdateInfo*) 6,822,985 ( 0.27%) ???:(anonymous namespace)::X86MCCodeEmitter::emitPrefixImpl(unsigned int&, llvm::MCInst const&, llvm::MCSubtargetInfo const&, llvm::raw_ostream&) const 6,557,033 ( 0.26%) ???:llvm::MCContext::getOrCreateSymbol(llvm::Twine const&) 6,422,769 ( 0.25%) ???:llvm::SelectionDAG::Legalize() 6,419,648 ( 0.25%) ???:llvm::MetadataTracking::track(void*, llvm::Metadata&, llvm::PointerUnion) 6,187,329 ( 0.24%) ???:(anonymous namespace)::VarLocBasedLDV::ExtendRanges(llvm::MachineFunction&, llvm::TargetPassConfig*) [clone .llvm.4451506318407214204] 5,745,134 ( 0.23%) ???:llvm::PMDataManager::removeNotPreservedAnalysis(llvm::Pass*) 5,675,185 ( 0.22%) ???:llvm::SelectionDAG::getConstant(llvm::ConstantInt const&, llvm::SDLoc const&, llvm::EVT, bool, bool) 5,506,998 ( 0.22%) ???:(anonymous namespace)::ELFObjectWriter::recordRelocation(llvm::MCAssembler&, llvm::MCAsmLayout const&, llvm::MCFragment const*, llvm::MCFixup const&, llvm::MCValue, unsigned long&) [clone .llvm.14145361893594770252] 5,498,730 ( 0.22%) ???:(anonymous namespace)::PEI::runOnMachineFunction(llvm::MachineFunction&) 5,430,102 ( 0.21%) ./elf/../sysdeps/x86_64/dl-machine.h:_dl_relocate_object 5,419,312 ( 0.21%) ???:(anonymous namespace)::DAGCombiner::combine(llvm::SDNode*) 5,405,071 ( 0.21%) /usr/home/liquid/.cargo/registry/src/github.com-1ecc6299db9ec823/hashbrown-0.12.0/src/raw/mod.rs:, (), core::hash::BuildHasherDefault>>::from_hash::>::{closure#0}> 5,322,344 ( 0.21%) ???:(anonymous namespace)::ScheduleDAGRRList::Schedule() [clone .llvm.6953762222372402862] 5,188,180 ( 0.21%) /usr/home/liquid/rust/worktree-benchmarking/library/proc_macro/src/bridge/server.rs:> as proc_macro::bridge::server::DispatcherTrait>::dispatch 5,008,114 ( 0.20%) ???:llvm::calculateDbgEntityHistory(llvm::MachineFunction const*, llvm::TargetRegisterInfo const*, llvm::DbgValueHistoryMap&, llvm::DbgLabelInstrMap&) 4,821,300 ( 0.19%) /usr/home/liquid/rust/worktree-benchmarking/library/core/src/hash/sip.rs:::write 4,743,785 ( 0.19%) ./stdlib/msort.c:msort_with_tmp.part.0 4,681,123 ( 0.19%) ???:llvm::FoldingSet::NodeEquals(llvm::FoldingSetBase const*, llvm::FoldingSetBase::Node*, llvm::FoldingSetNodeID const&, unsigned int, llvm::FoldingSetNodeID&) 4,566,253 ( 0.18%) ???:llvm::X86FrameLowering::hasFP(llvm::MachineFunction const&) const 4,495,186 ( 0.18%) ???:(anonymous namespace)::ELFWriter::writeObject(llvm::MCAssembler&, llvm::MCAsmLayout const&) 4,488,840 ( 0.18%) ???:(anonymous namespace)::X86MCInstLower::Lower(llvm::MachineInstr const*, llvm::MCInst&) const 4,488,144 ( 0.18%) ???:llvm::MDNode::MDNode(llvm::LLVMContext&, unsigned int, llvm::Metadata::StorageType, llvm::ArrayRef, llvm::ArrayRef) 4,348,527 ( 0.17%) ./malloc/malloc.c:realloc 4,254,372 ( 0.17%) ???:llvm::raw_ostream::write(char const*, unsigned long) 4,251,008 ( 0.17%) ???:llvm::PassRegistry::enumerateWith(llvm::PassRegistrationListener*) 4,231,880 ( 0.17%) ???:llvm::FunctionLoweringInfo::set(llvm::Function const&, llvm::MachineFunction&, llvm::SelectionDAG*) 4,179,285 ( 0.17%) ???:(anonymous namespace)::Verifier::visitFunction(llvm::Function const&)::$_3::operator()(llvm::Instruction const&, llvm::MDNode const*) const 4,132,216 ( 0.16%) /tmp/gcc-build/x86_64-unknown-linux-gnu/libstdc++-v3/libsupc++/../../../../gcc-5.5.0/libstdc++-v3/libsupc++/new_op.cc:operator new(unsigned long) 4,110,825 ( 0.16%) ???:llvm::MCAssembler::layout(llvm::MCAsmLayout&) 4,012,639 ( 0.16%) ???:(anonymous namespace)::X86MCCodeEmitter::encodeInstruction(llvm::MCInst const&, llvm::raw_ostream&, llvm::SmallVectorImpl&, llvm::MCSubtargetInfo const&) const 4,009,380 ( 0.16%) ???:llvm::MachineFunction::CreateMachineInstr(llvm::MCInstrDesc const&, llvm::DebugLoc const&, bool) 3,935,274 ( 0.16%) ???:llvm::ScheduleDAGSDNodes::AddSchedEdges() 3,916,740 ( 0.15%) ???:llvm::DwarfDebug::beginInstruction(llvm::MachineInstr const*) 3,898,248 ( 0.15%) ???:llvm::MachineFunctionPass::runOnFunction(llvm::Function&) 3,882,547 ( 0.15%) ???:llvm::Twine::printOneChild(llvm::raw_ostream&, llvm::Twine::Child, llvm::Twine::NodeKind) const 3,875,328 ( 0.15%) ???:llvm::MDString::get(llvm::LLVMContext&, llvm::StringRef) 3,859,683 ( 0.15%) ./string/../sysdeps/x86_64/multiarch/memset-vec-unaligned-erms.S:__memset_avx2_unaligned_erms 3,834,356 ( 0.15%) ???:??? 3,719,672 ( 0.15%) ???:llvm::X86TargetMachine::getTargetTransformInfo(llvm::Function const&) 3,706,596 ( 0.15%) /usr/home/liquid/rust/worktree-benchmarking/library/core/src/slice/mod.rs:::lookup_source_file_idx 3,690,302 ( 0.15%) ???:llvm::SelectionDAGISel::SelectAllBasicBlocks(llvm::Function const&) 3,654,086 ( 0.14%) ???:llvm::DataLayout::getAlignment(llvm::Type*, bool) const 3,650,726 ( 0.14%) ???:llvm::LLVMContextImpl::~LLVMContextImpl() 3,639,165 ( 0.14%) ???:(anonymous namespace)::CFIInstrInserter::runOnMachineFunction(llvm::MachineFunction&) 3,638,229 ( 0.14%) ???:llvm::MCELFStreamer::emitLabel(llvm::MCSymbol*, llvm::SMLoc) 3,585,284 ( 0.14%) ???:llvm::SelectionDAG::AssignTopologicalOrder() 3,568,496 ( 0.14%) ???:llvm::X86TargetLowering::X86TargetLowering(llvm::X86TargetMachine const&, llvm::X86Subtarget const&) 3,416,159 ( 0.14%) ???:(anonymous namespace)::RegAllocFast::markRegUsedInInstr(unsigned short) 3,408,264 ( 0.13%) /usr/home/liquid/rust/worktree-benchmarking/library/proc_macro/src/bridge/client.rs:proc_macro::bridge::scoped_cell::ScopedCell::replace 3,386,965 ( 0.13%) ???:llvm::SelectionDAG::getRegister(unsigned int, llvm::EVT) 3,308,992 ( 0.13%) ???:llvm::AttributeSetNode::get(llvm::LLVMContext&, llvm::AttrBuilder const&) 3,286,776 ( 0.13%) ???:llvm::MCStreamer::emitULEB128IntValue(unsigned long, unsigned int) 3,144,364 ( 0.12%) ???:llvm::SelectionDAG::getNode(unsigned int, llvm::SDLoc const&, llvm::SDVTList, llvm::ArrayRef) 3,121,651 ( 0.12%) /usr/home/liquid/rust/worktree-benchmarking/library/std/src/sys/unix/alloc.rs:__rdl_alloc 3,110,547 ( 0.12%) ???:llvm::AsmPrinter::emitFunctionBody() 3,084,124 ( 0.12%) ???:(anonymous namespace)::X86AsmBackend::applyFixup(llvm::MCAssembler const&, llvm::MCFixup const&, llvm::MCValue const&, llvm::MutableArrayRef, unsigned long, bool, llvm::MCSubtargetInfo const*) const 3,073,032 ( 0.12%) ???:llvm::SmallPtrSetImplBase::insert_imp_big(void const*) 3,052,960 ( 0.12%) ???:llvm::TargetLoweringBase::computeRegisterProperties(llvm::TargetRegisterInfo const*) 3,041,800 ( 0.12%) ./malloc/malloc.c:_int_realloc 3,040,209 ( 0.12%) /usr/home/liquid/rust/worktree-benchmarking/library/core/src/slice/mod.rs:::line_begin_pos 3,040,209 ( 0.12%) /usr/home/liquid/rust/worktree-benchmarking/library/core/src/slice/mod.rs:::lookup_line 3,021,533 ( 0.12%) ???:llvm::MCELFStreamer::emitInstToData(llvm::MCInst const&, llvm::MCSubtargetInfo const&) 2,986,228 ( 0.12%) /usr/home/liquid/rust/worktree-benchmarking/library/core/src/num/uint_macros.rs:::short_write_process_buffer:: 2,971,219 ( 0.12%) ./string/../sysdeps/x86_64/multiarch/memmove-vec-unaligned-erms.S:memcpy@GLIBC_2.2.5 2,936,789 ( 0.12%) ???:llvm::MCDwarfLineAddr::Encode(llvm::MCContext&, llvm::MCDwarfLineTableParams, long, unsigned long, llvm::raw_ostream&) 2,910,809 ( 0.12%) ???:SetImpliedBits(llvm::FeatureBitset&, llvm::FeatureBitset const&, llvm::ArrayRef) 2,890,527 ( 0.11%) ???:llvm::DomTreeBuilder::SemiNCAInfo >::runSemiNCA(llvm::DominatorTreeBase&, unsigned int) 2,879,505 ( 0.11%) /usr/home/liquid/rust/worktree-benchmarking/compiler/rustc_serialize/src/opaque.rs:>::decode 2,873,612 ( 0.11%) ???:(anonymous namespace)::X86MCCodeEmitter::emitMemModRMByte(llvm::MCInst const&, unsigned int, unsigned int, unsigned long, bool, unsigned long, llvm::raw_ostream&, llvm::SmallVectorImpl&, llvm::MCSubtargetInfo const&, bool) const 2,872,328 ( 0.11%) ???:llvm::BumpPtrAllocatorImpl::Allocate(unsigned long, llvm::Align) 2,848,941 ( 0.11%) ???:llvm::DILocation::getImpl(llvm::LLVMContext&, unsigned int, unsigned int, llvm::Metadata*, llvm::Metadata*, bool, llvm::Metadata::StorageType, bool) 2,838,538 ( 0.11%) ???:(anonymous namespace)::VarLocBasedLDV::collectIDsForRegs(llvm::SmallSet >&, llvm::SmallSet > const&, llvm::CoalescingBitVector const&, (anonymous namespace)::VarLocBasedLDV::VarLocMap const&) 2,829,818 ( 0.11%) ???:llvm::ScheduleDAGSDNodes::EmitSchedule(llvm::MachineInstrBundleIterator&) 2,782,116 ( 0.11%) ???:llvm::X86RegisterInfo::getReservedRegs(llvm::MachineFunction const&) const 2,777,444 ( 0.11%) ???:llvm::MCObjectStreamer::emitInstruction(llvm::MCInst const&, llvm::MCSubtargetInfo const&) 2,764,500 ( 0.11%) /usr/home/liquid/rust/worktree-benchmarking/compiler/rustc_metadata/src/rmeta/decoder.rs:>::decode 2,684,106 ( 0.11%) ???:llvm::MCAssembler::layoutSectionOnce(llvm::MCAsmLayout&, llvm::MCSection&) 2,660,926 ( 0.11%) ./malloc/malloc.c:calloc 2,660,738 ( 0.11%) ???:llvm::X86RegisterInfo::eliminateFrameIndex(llvm::MachineInstrBundleIterator, int, unsigned int, llvm::RegScavenger*) const 2,602,427 ( 0.10%) ./nptl/pthread_rwlock_common.c:pthread_rwlock_unlock 2,570,632 ( 0.10%) ???:llvm::X86TargetMachine::getSubtargetImpl(llvm::Function const&) const 2,560,879 ( 0.10%) /usr/home/liquid/rust/worktree-benchmarking/library/alloc/src/vec/mod.rs:>::decode 2,529,998 ( 0.10%) ./elf/do-rel.h:_dl_relocate_object -------------------------------------------------------------------------------- -- Auto-annotated source: /usr/home/liquid/rust/worktree-benchmarking/library/proc_macro/src/bridge/server.rs -------------------------------------------------------------------------------- Ir -- line 28 ---------------------------------------- . (type SourceFile: 'static + Clone;); . (type MultiSpan) => . (type MultiSpan: 'static;); . (type Diagnostic) => . (type Diagnostic: 'static;); . (type Span) => . (type Span: 'static + Copy + Eq + Hash;); . (fn drop(&mut self, $arg:ident: $arg_ty:ty)) => 7,680 ( 0.00%) (fn drop(&mut self, $arg: $arg_ty) { mem::drop($arg) }); . (fn clone(&mut self, $arg:ident: $arg_ty:ty) -> $ret_ty:ty) => 570 ( 0.00%) (fn clone(&mut self, $arg: $arg_ty) -> $ret_ty { $arg.clone() }); . ($($item:tt)*) => ($($item)*;) . } . . macro_rules! declare_server_traits { . ($($name:ident { . $(fn $method:ident($($arg:ident: $arg_ty:ty),* $(,)?) $(-> $ret_ty:ty)?;)* . }),* $(,)?) => { . pub trait Types { -- line 46 ---------------------------------------- -- line 63 ---------------------------------------- . ($($name:ident { . $(fn $method:ident($($arg:ident: $arg_ty:ty),* $(,)?) $(-> $ret_ty:ty)?;)* . }),* $(,)?) => { . impl Types for MarkedTypes { . $(type $name = Marked;)* . } . . $(impl $name for MarkedTypes { 1,578 ( 0.00%) $(fn $method(&mut self, $($arg: $arg_ty),*) $(-> $ret_ty)? { 356,732 ( 0.01%) <_>::mark($name::$method(&mut self.0, $($arg.unmark()),*)) 1,578 ( 0.00%) })* . })* . } . } . with_api!(Self, self_, define_mark_types_impls); . . struct Dispatcher { . handle_store: HandleStore, . server: S, -- line 81 ---------------------------------------- -- line 89 ---------------------------------------- . pub trait DispatcherTrait { . // HACK(eddyb) these are here to allow `Self::$name` to work below. . $(type $name;)* . fn dispatch(&mut self, b: Buffer) -> Buffer; . } . . impl DispatcherTrait for Dispatcher> { . $(type $name = as Types>::$name;)* 898,310 ( 0.04%) fn dispatch(&mut self, mut b: Buffer) -> Buffer { . let Dispatcher { handle_store, server } = self; . 269,493 ( 0.01%) let mut reader = &b[..]; 628,817 ( 0.02%) match api_tags::Method::decode(&mut reader, &mut ()) { 433,787 ( 0.02%) $(api_tags::Method::$name(m) => match m { . $(api_tags::$name::$method => { . let mut call_method = || { 276,507 ( 0.01%) reverse_decode!(reader, handle_store; $($arg: $arg_ty),*); 789 ( 0.00%) $name::$method(server, $($arg),*) . }; . // HACK(eddyb) don't use `panic::catch_unwind` in a panic. . // If client and server happen to use the same `libstd`, . // `catch_unwind` asserts that the panic counter was 0, . // even when the closure passed to it didn't panic. . let r = if thread::panicking() { . Ok(call_method()) . } else { 419,298 ( 0.02%) panic::catch_unwind(panic::AssertUnwindSafe(call_method)) . .map_err(PanicMessage::from) . }; . . b.clear(); 989,557 ( 0.04%) r.encode(&mut b, handle_store); . })* . }),* . } 538,986 ( 0.02%) b 808,479 ( 0.03%) } . } . } . } . with_api!(Self, self_, define_dispatcher_impl); . . pub trait ExecutionStrategy { . fn run_bridge_and_client( . &self, -- line 133 ---------------------------------------- -- line 145 ---------------------------------------- . fn run_bridge_and_client( . &self, . dispatcher: &mut impl DispatcherTrait, . input: Buffer, . run_client: extern "C" fn(Bridge<'_>, D) -> Buffer, . client_data: D, . force_show_panics: bool, . ) -> Buffer { 808,494 ( 0.03%) let mut dispatch = |b| dispatcher.dispatch(b); . 150 ( 0.00%) run_client( 165 ( 0.00%) Bridge { cached_buffer: input, dispatch: (&mut dispatch).into(), force_show_panics }, . client_data, . ) . } . } . . // NOTE(eddyb) Two implementations are provided, the second one is a bit . // faster but neither is anywhere near as fast as same-thread execution. . -- line 164 ---------------------------------------- -- line 263 ---------------------------------------- . *state.lock().unwrap() = State::Res(b); . join_handle.thread().unpark(); . } . . join_handle.join().unwrap() . } . } . 210 ( 0.00%) fn run_server< . S: Server, . I: Encode>>, . O: for<'a, 's> DecodeMut<'a, 's, HandleStore>>, . D: Copy + Send + 'static, . >( . strategy: &impl ExecutionStrategy, . handle_counters: &'static client::HandleCounters, . server: S, . input: I, . run_client: extern "C" fn(Bridge<'_>, D) -> Buffer, . client_data: D, . force_show_panics: bool, . ) -> Result { . let mut dispatcher = 255 ( 0.00%) Dispatcher { handle_store: HandleStore::new(handle_counters), server: MarkedTypes(server) }; . . let mut b = Buffer::new(); 60 ( 0.00%) input.encode(&mut b, &mut dispatcher.handle_store); . 90 ( 0.00%) b = strategy.run_bridge_and_client( . &mut dispatcher, . b, . run_client, . client_data, . force_show_panics, . ); . 75 ( 0.00%) Result::decode(&mut &b[..], &mut dispatcher.handle_store) 150 ( 0.00%) } . . impl client::Client crate::TokenStream> { . pub fn run( . &self, . strategy: &impl ExecutionStrategy, . server: S, . input: S::TokenStream, . force_show_panics: bool, -- line 308 ---------------------------------------- -- line 325 ---------------------------------------- . pub fn run( . &self, . strategy: &impl ExecutionStrategy, . server: S, . input: S::TokenStream, . input2: S::TokenStream, . force_show_panics: bool, . ) -> Result { 30 ( 0.00%) let client::Client { get_handle_counters, run, f } = *self; 195 ( 0.00%) run_server( . strategy, 15 ( 0.00%) get_handle_counters(), 150 ( 0.00%) server, . ( . as Types>::TokenStream::mark(input), . as Types>::TokenStream::mark(input2), . ), . run, . f, . force_show_panics, . ) -- line 345 ---------------------------------------- 207,561 ( 0.01%) -------------------------------------------------------------------------------- -- Auto-annotated source: /usr/home/liquid/rust/worktree-benchmarking/library/core/src/slice/mod.rs -------------------------------------------------------------------------------- Ir -- line 118 ---------------------------------------- . /// let a = [1, 2, 3]; . /// assert_eq!(a.len(), 3); . /// ``` . #[lang = "slice_len_fn"] . #[stable(feature = "rust1", since = "1.0.0")] . #[rustc_const_stable(feature = "const_slice_len", since = "1.39.0")] . #[inline] . // SAFETY: const sound because we transmute out the length field as a usize (which it must be) 10,383 ( 0.00%) pub const fn len(&self) -> usize { . // FIXME: Replace with `crate::ptr::metadata(self)` when that is const-stable. . // As of this writing this causes a "Const-stable functions can only call other . // const-stable functions" error. . . // SAFETY: Accessing the value from the `PtrRepr` union is safe since *const T . // and PtrComponents have the same memory layouts. Only std can make this . // guarantee. 10,383 ( 0.00%) unsafe { crate::ptr::PtrRepr { const_ptr: self }.components.metadata } 6,922 ( 0.00%) } . . /// Returns `true` if the slice has a length of 0. . /// . /// # Examples . /// . /// ``` . /// let a = [1, 2, 3]; . /// assert!(!a.is_empty()); . /// ``` . #[stable(feature = "rust1", since = "1.0.0")] . #[rustc_const_stable(feature = "const_slice_is_empty", since = "1.39.0")] . #[inline] . pub const fn is_empty(&self) -> bool { 236,818 ( 0.01%) self.len() == 0 . } . . /// Returns the first element of the slice, or `None` if it is empty. . /// . /// # Examples . /// . /// ``` . /// let v = [10, 40, 30]; -- line 157 ---------------------------------------- -- line 158 ---------------------------------------- . /// assert_eq!(Some(&10), v.first()); . /// . /// let w: &[i32] = &[]; . /// assert_eq!(None, w.first()); . /// ``` . #[stable(feature = "rust1", since = "1.0.0")] . #[rustc_const_stable(feature = "const_slice_first_last_not_mut", since = "1.56.0")] . #[inline] 20 ( 0.00%) pub const fn first(&self) -> Option<&T> { 2,840 ( 0.00%) if let [first, ..] = self { Some(first) } else { None } 15 ( 0.00%) } . . /// Returns a mutable pointer to the first element of the slice, or `None` if it is empty. . /// . /// # Examples . /// . /// ``` . /// let x = &mut [0, 1, 2]; . /// -- line 176 ---------------------------------------- -- line 178 ---------------------------------------- . /// *first = 5; . /// } . /// assert_eq!(x, &[5, 1, 2]); . /// ``` . #[stable(feature = "rust1", since = "1.0.0")] . #[rustc_const_unstable(feature = "const_slice_first_last", issue = "83570")] . #[inline] . pub const fn first_mut(&mut self) -> Option<&mut T> { 160 ( 0.00%) if let [first, ..] = self { Some(first) } else { None } . } . . /// Returns the first and all the rest of the elements of the slice, or `None` if it is empty. . /// . /// # Examples . /// . /// ``` . /// let x = &[0, 1, 2]; -- line 194 ---------------------------------------- -- line 197 ---------------------------------------- . /// assert_eq!(first, &0); . /// assert_eq!(elements, &[1, 2]); . /// } . /// ``` . #[stable(feature = "slice_splits", since = "1.5.0")] . #[rustc_const_stable(feature = "const_slice_first_last_not_mut", since = "1.56.0")] . #[inline] . pub const fn split_first(&self) -> Option<(&T, &[T])> { 224 ( 0.00%) if let [first, tail @ ..] = self { Some((first, tail)) } else { None } . } . . /// Returns the first and all the rest of the elements of the slice, or `None` if it is empty. . /// . /// # Examples . /// . /// ``` . /// let x = &mut [0, 1, 2]; -- line 213 ---------------------------------------- -- line 237 ---------------------------------------- . /// assert_eq!(last, &2); . /// assert_eq!(elements, &[0, 1]); . /// } . /// ``` . #[stable(feature = "slice_splits", since = "1.5.0")] . #[rustc_const_stable(feature = "const_slice_first_last_not_mut", since = "1.56.0")] . #[inline] . pub const fn split_last(&self) -> Option<(&T, &[T])> { 7,055 ( 0.00%) if let [init @ .., last] = self { Some((last, init)) } else { None } . } . . /// Returns the last and all the rest of the elements of the slice, or `None` if it is empty. . /// . /// # Examples . /// . /// ``` . /// let x = &mut [0, 1, 2]; -- line 253 ---------------------------------------- -- line 275 ---------------------------------------- . /// assert_eq!(Some(&30), v.last()); . /// . /// let w: &[i32] = &[]; . /// assert_eq!(None, w.last()); . /// ``` . #[stable(feature = "rust1", since = "1.0.0")] . #[rustc_const_stable(feature = "const_slice_first_last_not_mut", since = "1.56.0")] . #[inline] 5 ( 0.00%) pub const fn last(&self) -> Option<&T> { 94,209 ( 0.00%) if let [.., last] = self { Some(last) } else { None } 3 ( 0.00%) } . . /// Returns a mutable pointer to the last item in the slice. . /// . /// # Examples . /// . /// ``` . /// let x = &mut [0, 1, 2]; . /// -- line 293 ---------------------------------------- -- line 295 ---------------------------------------- . /// *last = 10; . /// } . /// assert_eq!(x, &[0, 1, 10]); . /// ``` . #[stable(feature = "rust1", since = "1.0.0")] . #[rustc_const_unstable(feature = "const_slice_first_last", issue = "83570")] . #[inline] . pub const fn last_mut(&mut self) -> Option<&mut T> { 244,649 ( 0.01%) if let [.., last] = self { Some(last) } else { None } . } . . /// Returns a reference to an element or subslice depending on the type of . /// index. . /// . /// - If given a position, returns a reference to the element at that . /// position or `None` if out of bounds. . /// - If given a range, returns the subslice corresponding to that range, -- line 311 ---------------------------------------- -- line 317 ---------------------------------------- . /// let v = [10, 40, 30]; . /// assert_eq!(Some(&40), v.get(1)); . /// assert_eq!(Some(&[10, 40][..]), v.get(0..2)); . /// assert_eq!(None, v.get(3)); . /// assert_eq!(None, v.get(0..4)); . /// ``` . #[stable(feature = "rust1", since = "1.0.0")] . #[inline] 6,872 ( 0.00%) pub fn get(&self, index: I) -> Option<&I::Output> . where . I: SliceIndex, . { 1,718 ( 0.00%) index.get(self) 1,718 ( 0.00%) } . . /// Returns a mutable reference to an element or subslice depending on the . /// type of index (see [`get`]) or `None` if the index is out of bounds. . /// . /// [`get`]: slice::get . /// . /// # Examples . /// -- line 338 ---------------------------------------- -- line 372 ---------------------------------------- . /// let x = &[1, 2, 4]; . /// . /// unsafe { . /// assert_eq!(x.get_unchecked(1), &2); . /// } . /// ``` . #[stable(feature = "rust1", since = "1.0.0")] . #[inline] 1,000 ( 0.00%) pub unsafe fn get_unchecked(&self, index: I) -> &I::Output . where . I: SliceIndex, . { . // SAFETY: the caller must uphold most of the safety requirements for `get_unchecked`; . // the slice is dereferenceable because `self` is a safe reference. . // The returned pointer is safe because impls of `SliceIndex` have to guarantee that it is. 250 ( 0.00%) unsafe { &*index.get_unchecked(self) } 250 ( 0.00%) } . . /// Returns a mutable reference to an element or subslice, without doing . /// bounds checking. . /// . /// For a safe alternative see [`get_mut`]. . /// . /// # Safety . /// -- line 396 ---------------------------------------- -- line 447 ---------------------------------------- . /// } . /// } . /// ``` . /// . /// [`as_mut_ptr`]: slice::as_mut_ptr . #[stable(feature = "rust1", since = "1.0.0")] . #[rustc_const_stable(feature = "const_slice_as_ptr", since = "1.32.0")] . #[inline] 155,388 ( 0.01%) pub const fn as_ptr(&self) -> *const T { 346,330 ( 0.01%) self as *const [T] as *const T 77,694 ( 0.00%) } . . /// Returns an unsafe mutable pointer to the slice's buffer. . /// . /// The caller must ensure that the slice outlives the pointer this . /// function returns, or else it will end up pointing to garbage. . /// . /// Modifying the container referenced by this slice may cause its buffer . /// to be reallocated, which would also make any pointers to it invalid. -- line 465 ---------------------------------------- -- line 475 ---------------------------------------- . /// *x_ptr.add(i) += 2; . /// } . /// } . /// assert_eq!(x, &[3, 4, 6]); . /// ``` . #[stable(feature = "rust1", since = "1.0.0")] . #[rustc_const_unstable(feature = "const_ptr_offset", issue = "71499")] . #[inline] 228 ( 0.00%) pub const fn as_mut_ptr(&mut self) -> *mut T { 8 ( 0.00%) self as *mut [T] as *mut T 114 ( 0.00%) } . . /// Returns the two raw pointers spanning the slice. . /// . /// The returned range is half-open, which means that the end pointer . /// points *one past* the last element of the slice. This way, an empty . /// slice is represented by two equal pointers, and the difference between . /// the two pointers represents the size of the slice. . /// -- line 493 ---------------------------------------- -- line 556 ---------------------------------------- . /// [`as_mut_ptr`]: slice::as_mut_ptr . #[stable(feature = "slice_ptr_range", since = "1.48.0")] . #[rustc_const_unstable(feature = "const_ptr_offset", issue = "71499")] . #[inline] . pub const fn as_mut_ptr_range(&mut self) -> Range<*mut T> { . let start = self.as_mut_ptr(); . // SAFETY: See as_ptr_range() above for why `add` here is safe. . let end = unsafe { start.add(self.len()) }; 1 ( 0.00%) start..end . } . . /// Swaps two elements in the slice. . /// . /// # Arguments . /// . /// * a - The index of the first element . /// * b - The index of the second element -- line 572 ---------------------------------------- -- line 582 ---------------------------------------- . /// v.swap(2, 4); . /// assert!(v == ["a", "b", "e", "d", "c"]); . /// ``` . #[stable(feature = "rust1", since = "1.0.0")] . #[rustc_const_unstable(feature = "const_swap", issue = "83163")] . #[inline] . #[track_caller] . pub const fn swap(&mut self, a: usize, b: usize) { 7,352 ( 0.00%) let _ = &self[a]; 18,210 ( 0.00%) let _ = &self[b]; . . // SAFETY: we just checked that both `a` and `b` are in bounds . unsafe { self.swap_unchecked(a, b) } . } . . /// Swaps two elements in the slice, without doing bounds checking. . /// . /// For a safe alternative see [`swap`]. -- line 599 ---------------------------------------- -- line 644 ---------------------------------------- . /// ``` . /// let mut v = [1, 2, 3]; . /// v.reverse(); . /// assert!(v == [3, 2, 1]); . /// ``` . #[stable(feature = "rust1", since = "1.0.0")] . #[inline] . pub fn reverse(&mut self) { 2 ( 0.00%) let half_len = self.len() / 2; . let Range { start, end } = self.as_mut_ptr_range(); . . // These slices will skip the middle item for an odd length, . // since that one doesn't need to move. . let (front_half, back_half) = . // SAFETY: Both are subparts of the original slice, so the memory . // range is valid, and they don't overlap because they're each only . // half (or less) of the original slice. -- line 660 ---------------------------------------- -- line 677 ---------------------------------------- . . // Because this function is first compiled in isolation, . // this check tells LLVM that the indexing below is . // in-bounds. Then after inlining -- once the actual . // lengths of the slices are known -- it's removed. . let (a, b) = (&mut a[..n], &mut b[..n]); . . for i in 0..n { 6,640 ( 0.00%) mem::swap(&mut a[i], &mut b[n - 1 - i]); . } . } . } . . /// Returns an iterator over the slice. . /// . /// # Examples . /// -- line 693 ---------------------------------------- -- line 697 ---------------------------------------- . /// . /// assert_eq!(iterator.next(), Some(&1)); . /// assert_eq!(iterator.next(), Some(&2)); . /// assert_eq!(iterator.next(), Some(&4)); . /// assert_eq!(iterator.next(), None); . /// ``` . #[stable(feature = "rust1", since = "1.0.0")] . #[inline] 55,113 ( 0.00%) pub fn iter(&self) -> Iter<'_, T> { 55,113 ( 0.00%) Iter::new(self) 36,742 ( 0.00%) } . . /// Returns an iterator that allows modifying each value. . /// . /// # Examples . /// . /// ``` . /// let x = &mut [1, 2, 4]; . /// for elem in x.iter_mut() { . /// *elem += 2; . /// } . /// assert_eq!(x, &[3, 4, 6]); . /// ``` . #[stable(feature = "rust1", since = "1.0.0")] . #[inline] 81 ( 0.00%) pub fn iter_mut(&mut self) -> IterMut<'_, T> { 81 ( 0.00%) IterMut::new(self) 54 ( 0.00%) } . . /// Returns an iterator over all contiguous windows of length . /// `size`. The windows overlap. If the slice is shorter than . /// `size`, the iterator returns no values. . /// . /// # Panics . /// . /// Panics if `size` is 0. -- line 732 ---------------------------------------- -- line 1499 ---------------------------------------- . /// assert_eq!(left, [1, 2, 3, 4, 5, 6]); . /// assert_eq!(right, []); . /// } . /// ``` . #[stable(feature = "rust1", since = "1.0.0")] . #[inline] . #[track_caller] . pub fn split_at(&self, mid: usize) -> (&[T], &[T]) { 312 ( 0.00%) assert!(mid <= self.len()); . // SAFETY: `[ptr; mid]` and `[mid; len]` are inside `self`, which . // fulfills the requirements of `from_raw_parts_mut`. . unsafe { self.split_at_unchecked(mid) } . } . . /// Divides one mutable slice into two at an index. . /// . /// The first will contain all indices from `[0, mid)` (excluding -- line 1515 ---------------------------------------- -- line 1530 ---------------------------------------- . /// left[1] = 2; . /// right[1] = 4; . /// assert_eq!(v, [1, 2, 3, 4, 5, 6]); . /// ``` . #[stable(feature = "rust1", since = "1.0.0")] . #[inline] . #[track_caller] . pub fn split_at_mut(&mut self, mid: usize) -> (&mut [T], &mut [T]) { 13,062 ( 0.00%) assert!(mid <= self.len()); . // SAFETY: `[ptr; mid]` and `[mid; len]` are inside `self`, which . // fulfills the requirements of `from_raw_parts_mut`. . unsafe { self.split_at_mut_unchecked(mid) } . } . . /// Divides one slice into two at an index, without doing bounds checking. . /// . /// The first will contain all indices from `[0, mid)` (excluding -- line 1546 ---------------------------------------- -- line 1628 ---------------------------------------- . pub unsafe fn split_at_mut_unchecked(&mut self, mid: usize) -> (&mut [T], &mut [T]) { . let len = self.len(); . let ptr = self.as_mut_ptr(); . . // SAFETY: Caller has to check that `0 <= mid <= self.len()`. . // . // `[ptr; mid]` and `[mid; len]` are not overlapping, so returning a mutable reference . // is fine. 2,127 ( 0.00%) unsafe { (from_raw_parts_mut(ptr, mid), from_raw_parts_mut(ptr.add(mid), len - mid)) } . } . . /// Divides one slice into an array and a remainder slice at an index. . /// . /// The array will contain all indices from `[0, N)` (excluding . /// the index `N` itself) and the slice will contain all . /// indices from `[N, len)` (excluding the index `len` itself). . /// -- line 1644 ---------------------------------------- -- line 2113 ---------------------------------------- . /// assert!(!v.iter().any(|e| e == "hi")); . /// ``` . #[stable(feature = "rust1", since = "1.0.0")] . #[inline] . pub fn contains(&self, x: &T) -> bool . where . T: PartialEq, . { 94 ( 0.00%) cmp::SliceContains::slice_contains(x, self) . } . . /// Returns `true` if `needle` is a prefix of the slice. . /// . /// # Examples . /// . /// ``` . /// let v = [10, 40, 30]; -- line 2129 ---------------------------------------- -- line 2137 ---------------------------------------- . /// . /// ``` . /// let v = &[10, 40, 30]; . /// assert!(v.starts_with(&[])); . /// let v: &[u8] = &[]; . /// assert!(v.starts_with(&[])); . /// ``` . #[stable(feature = "rust1", since = "1.0.0")] 23,583 ( 0.00%) pub fn starts_with(&self, needle: &[T]) -> bool . where . T: PartialEq, . { 10,107 ( 0.00%) let n = needle.len(); 97,258 ( 0.00%) self.len() >= n && needle == &self[..n] 16,845 ( 0.00%) } . . /// Returns `true` if `needle` is a suffix of the slice. . /// . /// # Examples . /// . /// ``` . /// let v = [10, 40, 30]; . /// assert!(v.ends_with(&[30])); -- line 2159 ---------------------------------------- -- line 2166 ---------------------------------------- . /// . /// ``` . /// let v = &[10, 40, 30]; . /// assert!(v.ends_with(&[])); . /// let v: &[u8] = &[]; . /// assert!(v.ends_with(&[])); . /// ``` . #[stable(feature = "rust1", since = "1.0.0")] 343 ( 0.00%) pub fn ends_with(&self, needle: &[T]) -> bool . where . T: PartialEq, . { 441 ( 0.00%) let (m, n) = (self.len(), needle.len()); 31,027 ( 0.00%) m >= n && needle == &self[m - n..] 245 ( 0.00%) } . . /// Returns a subslice with the prefix removed. . /// . /// If the slice starts with `prefix`, returns the subslice after the prefix, wrapped in `Some`. . /// If `prefix` is empty, simply returns the original slice. . /// . /// If the slice does not start with `prefix`, returns `None`. . /// -- line 2188 ---------------------------------------- -- line 2293 ---------------------------------------- . /// s.insert(idx, num); . /// assert_eq!(s, [0, 1, 1, 1, 1, 2, 3, 5, 8, 13, 21, 34, 42, 55]); . /// ``` . #[stable(feature = "rust1", since = "1.0.0")] . pub fn binary_search(&self, x: &T) -> Result . where . T: Ord, . { 3 ( 0.00%) self.binary_search_by(|p| p.cmp(x)) . } . . /// Binary searches this sorted slice with a comparator function. . /// . /// The comparator function should implement an order consistent . /// with the sort order of the underlying slice, returning an . /// order code that indicates whether its argument is `Less`, . /// `Equal` or `Greater` the desired target. -- line 2309 ---------------------------------------- -- line 2345 ---------------------------------------- . #[inline] . pub fn binary_search_by<'a, F>(&'a self, mut f: F) -> Result . where . F: FnMut(&'a T) -> Ordering, . { . let mut size = self.len(); . let mut left = 0; . let mut right = size; 1,827,592 ( 0.07%) while left < right { 3,045,620 ( 0.12%) let mid = left + size / 2; . . // SAFETY: the call is made safe by the following invariants: . // - `mid >= 0` . // - `mid < size`: `mid` is limited by `[left; right)` bound. 350,842 ( 0.01%) let cmp = f(unsafe { self.get_unchecked(mid) }); . . // The reason why we use if/else control flow rather than match . // is because match reorders comparison operations, which is perf sensitive. . // This is x86 asm for u8: https://rust.godbolt.org/z/8Y8Pra. 788,667 ( 0.03%) if cmp == Less { 1,649,152 ( 0.07%) left = mid + 1; 420,134 ( 0.02%) } else if cmp == Greater { . right = mid; . } else { . // SAFETY: same as the `get_unchecked` above . unsafe { crate::intrinsics::assume(mid < self.len()) }; . return Ok(mid); . } . 2,307,707 ( 0.09%) size = right - left; . } . Err(left) . } . . /// Binary searches this sorted slice with a key extraction function. . /// . /// Assumes that the slice is sorted by the key, for instance with . /// [`sort_by_key`] using the same key extraction function. -- line 2382 ---------------------------------------- -- line 3187 ---------------------------------------- . /// assert_eq!(slice, [4, 5, 3, 4, 5]); . /// ``` . /// . /// [`clone_from_slice`]: slice::clone_from_slice . /// [`split_at_mut`]: slice::split_at_mut . #[doc(alias = "memcpy")] . #[stable(feature = "copy_from_slice", since = "1.9.0")] . #[track_caller] 300 ( 0.00%) pub fn copy_from_slice(&mut self, src: &[T]) . where . T: Copy, . { . // The panic code path was put into a cold function to not bloat the . // call site. . #[inline(never)] . #[cold] . #[track_caller] . fn len_mismatch_fail(dst_len: usize, src_len: usize) -> ! { . panic!( . "source slice length ({}) does not match destination slice length ({})", . src_len, dst_len, . ); . } . 42,742 ( 0.00%) if self.len() != src.len() { . len_mismatch_fail(self.len(), src.len()); . } . . // SAFETY: `self` is valid for `self.len()` elements by definition, and `src` was . // checked to have the same length. The slices cannot overlap because . // mutable references are exclusive. . unsafe { 180 ( 0.00%) ptr::copy_nonoverlapping(src.as_ptr(), self.as_mut_ptr(), self.len()); . } 60 ( 0.00%) } . . /// Copies elements from one part of the slice to another part of itself, . /// using a memmove. . /// . /// `src` is the range within `self` to copy from. `dest` is the starting . /// index of the range within `self` to copy to, which will have the same . /// length as `src`. The two ranges may overlap. The ends of the two ranges . /// must be less than or equal to `self.len()`. -- line 3229 ---------------------------------------- -- line 3380 ---------------------------------------- . } . a << k . } . let gcd: usize = gcd(mem::size_of::(), mem::size_of::()); . let ts: usize = mem::size_of::() / gcd; . let us: usize = mem::size_of::() / gcd; . . // Armed with this knowledge, we can find how many `U`s we can fit! 4 ( 0.00%) let us_len = self.len() / ts * us; . // And how many `T`s will be in the trailing slice! 10,738 ( 0.00%) let ts_len = self.len() % ts; . (us_len, ts_len) . } . . /// Transmute the slice to a slice of another type, ensuring alignment of the types is . /// maintained. . /// . /// This method splits the slice into three distinct slices: prefix, correctly aligned middle . /// slice of a new type, and the suffix slice. The method may make the middle slice the greatest -- line 3398 ---------------------------------------- -- line 3429 ---------------------------------------- . return (self, &[], &[]); . } . . // First, find at what point do we split between the first and 2nd slice. Easy with . // ptr.align_offset. . let ptr = self.as_ptr(); . // SAFETY: See the `align_to_mut` method for the detailed safety comment. . let offset = unsafe { crate::ptr::align_offset(ptr, mem::align_of::()) }; 10,738 ( 0.00%) if offset > self.len() { . (self, &[], &[]) . } else { . let (left, rest) = self.split_at(offset); . let (us_len, ts_len) = rest.align_to_offsets::(); . // SAFETY: now `rest` is definitely aligned, so `from_raw_parts` below is okay, . // since the caller guarantees that we can transmute `T` to `U` safely. . unsafe { . ( . left, . from_raw_parts(rest.as_ptr() as *const U, us_len), 10,742 ( 0.00%) from_raw_parts(rest.as_ptr().add(rest.len() - ts_len), ts_len), . ) . } . } . } . . /// Transmute the slice to a slice of another type, ensuring alignment of the types is . /// maintained. . /// -- line 3456 ---------------------------------------- -- line 3731 ---------------------------------------- . /// let v = [1, 2, 3, 3, 5, 6, 7]; . /// let i = v.partition_point(|&x| x < 5); . /// . /// assert_eq!(i, 4); . /// assert!(v[..i].iter().all(|&x| x < 5)); . /// assert!(v[i..].iter().all(|&x| !(x < 5))); . /// ``` . #[stable(feature = "partition_point", since = "1.52.0")] 1 ( 0.00%) pub fn partition_point

(&self, mut pred: P) -> usize . where . P: FnMut(&T) -> bool, . { . self.binary_search_by(|x| if pred(x) { Less } else { Greater }).unwrap_or_else(|i| i) 2 ( 0.00%) } . . /// Removes the subslice corresponding to the given range . /// and returns a reference to it. . /// . /// Returns `None` and does not modify the slice if the given . /// range is out of bounds. . /// . /// Note that this method only accepts one-sided ranges such as -- line 3752 ---------------------------------------- 1,417,707 ( 0.06%) -------------------------------------------------------------------------------- -- Auto-annotated source: /usr/home/liquid/rust/worktree-benchmarking/library/alloc/src/vec/mod.rs -------------------------------------------------------------------------------- Ir -- line 416 ---------------------------------------- . /// ``` . /// # #![allow(unused_mut)] . /// let mut vec: Vec = Vec::new(); . /// ``` . #[inline] . #[rustc_const_stable(feature = "const_vec_new", since = "1.39.0")] . #[stable(feature = "rust1", since = "1.0.0")] . #[must_use] 7,018 ( 0.00%) pub const fn new() -> Self { 1,492,063 ( 0.06%) Vec { buf: RawVec::NEW, len: 0 } 7,172 ( 0.00%) } . . /// Constructs a new, empty `Vec` with the specified capacity. . /// . /// The vector will be able to hold exactly `capacity` elements without . /// reallocating. If `capacity` is 0, the vector will not allocate. . /// . /// It is important to note that although the returned vector has the . /// *capacity* specified, the vector will have a zero *length*. For an -- line 434 ---------------------------------------- -- line 461 ---------------------------------------- . /// vec.push(11); . /// assert_eq!(vec.len(), 11); . /// assert!(vec.capacity() >= 11); . /// ``` . #[cfg(not(no_global_oom_handling))] . #[inline] . #[stable(feature = "rust1", since = "1.0.0")] . #[must_use] 796 ( 0.00%) pub fn with_capacity(capacity: usize) -> Self { 199 ( 0.00%) Self::with_capacity_in(capacity, Global) 398 ( 0.00%) } . . /// Creates a `Vec` directly from the raw components of another vector. . /// . /// # Safety . /// . /// This is highly unsafe, due to the number of invariants that aren't . /// checked: . /// -- line 479 ---------------------------------------- -- line 529 ---------------------------------------- . /// . /// // Put everything back together into a Vec . /// let rebuilt = Vec::from_raw_parts(p, len, cap); . /// assert_eq!(rebuilt, [4, 5, 6]); . /// } . /// ``` . #[inline] . #[stable(feature = "rust1", since = "1.0.0")] 270 ( 0.00%) pub unsafe fn from_raw_parts(ptr: *mut T, length: usize, capacity: usize) -> Self { 45 ( 0.00%) unsafe { Self::from_raw_parts_in(ptr, length, capacity, Global) } 90 ( 0.00%) } . } . . impl Vec { . /// Constructs a new, empty `Vec`. . /// . /// The vector will not allocate until elements are pushed onto it. . /// . /// # Examples -- line 547 ---------------------------------------- -- line 600 ---------------------------------------- . /// // ...but this may make the vector reallocate . /// vec.push(11); . /// assert_eq!(vec.len(), 11); . /// assert!(vec.capacity() >= 11); . /// ``` . #[cfg(not(no_global_oom_handling))] . #[inline] . #[unstable(feature = "allocator_api", issue = "32838")] 7,637 ( 0.00%) pub fn with_capacity_in(capacity: usize, alloc: A) -> Self { 531,289 ( 0.02%) Vec { buf: RawVec::with_capacity_in(capacity, alloc), len: 0 } 2,182 ( 0.00%) } . . /// Creates a `Vec` directly from the raw components of another vector. . /// . /// # Safety . /// . /// This is highly unsafe, due to the number of invariants that aren't . /// checked: . /// -- line 618 ---------------------------------------- -- line 676 ---------------------------------------- . /// . /// // Put everything back together into a Vec . /// let rebuilt = Vec::from_raw_parts_in(p, len, cap, alloc.clone()); . /// assert_eq!(rebuilt, [4, 5, 6]); . /// } . /// ``` . #[inline] . #[unstable(feature = "allocator_api", issue = "32838")] 7,150 ( 0.00%) pub unsafe fn from_raw_parts_in(ptr: *mut T, length: usize, capacity: usize, alloc: A) -> Self { 54,647 ( 0.00%) unsafe { Vec { buf: RawVec::from_raw_parts_in(ptr, capacity, alloc), len: length } } 1,100 ( 0.00%) } . . /// Decomposes a `Vec` into its raw components. . /// . /// Returns the raw pointer to the underlying data, the length of . /// the vector (in elements), and the allocated capacity of the . /// data (in elements). These are the same arguments in the same . /// order as the arguments to [`from_raw_parts`]. . /// -- line 694 ---------------------------------------- -- line 777 ---------------------------------------- . /// # Examples . /// . /// ``` . /// let vec: Vec = Vec::with_capacity(10); . /// assert_eq!(vec.capacity(), 10); . /// ``` . #[inline] . #[stable(feature = "rust1", since = "1.0.0")] 2,300 ( 0.00%) pub fn capacity(&self) -> usize { 83,755 ( 0.00%) self.buf.capacity() 1,150 ( 0.00%) } . . /// Reserves capacity for at least `additional` more elements to be inserted . /// in the given `Vec`. The collection may reserve more space to avoid . /// frequent reallocations. After calling `reserve`, capacity will be . /// greater than or equal to `self.len() + additional`. Does nothing if . /// capacity is already sufficient. . /// . /// # Panics -- line 795 ---------------------------------------- -- line 800 ---------------------------------------- . /// . /// ``` . /// let mut vec = vec![1]; . /// vec.reserve(10); . /// assert!(vec.capacity() >= 11); . /// ``` . #[cfg(not(no_global_oom_handling))] . #[stable(feature = "rust1", since = "1.0.0")] 10,425 ( 0.00%) pub fn reserve(&mut self, additional: usize) { 238,107 ( 0.01%) self.buf.reserve(self.len, additional); 4,170 ( 0.00%) } . . /// Reserves the minimum capacity for exactly `additional` more elements to . /// be inserted in the given `Vec`. After calling `reserve_exact`, . /// capacity will be greater than or equal to `self.len() + additional`. . /// Does nothing if the capacity is already sufficient. . /// . /// Note that the allocator may give the collection more space than it . /// requests. Therefore, capacity can not be relied upon to be precisely -- line 818 ---------------------------------------- -- line 829 ---------------------------------------- . /// ``` . /// let mut vec = vec![1]; . /// vec.reserve_exact(10); . /// assert!(vec.capacity() >= 11); . /// ``` . #[cfg(not(no_global_oom_handling))] . #[stable(feature = "rust1", since = "1.0.0")] . pub fn reserve_exact(&mut self, additional: usize) { 4,452 ( 0.00%) self.buf.reserve_exact(self.len, additional); . } . . /// Tries to reserve capacity for at least `additional` more elements to be inserted . /// in the given `Vec`. The collection may reserve more space to avoid . /// frequent reallocations. After calling `try_reserve`, capacity will be . /// greater than or equal to `self.len() + additional`. Does nothing if . /// capacity is already sufficient. . /// -- line 845 ---------------------------------------- -- line 926 ---------------------------------------- . /// let mut vec = Vec::with_capacity(10); . /// vec.extend([1, 2, 3]); . /// assert_eq!(vec.capacity(), 10); . /// vec.shrink_to_fit(); . /// assert!(vec.capacity() >= 3); . /// ``` . #[cfg(not(no_global_oom_handling))] . #[stable(feature = "rust1", since = "1.0.0")] 1,278 ( 0.00%) pub fn shrink_to_fit(&mut self) { . // The capacity is never less than the length, and there's nothing to do when . // they are equal, so we can avoid the panic case in `RawVec::shrink_to_fit` . // by only calling it with a greater capacity. 11,280 ( 0.00%) if self.capacity() > self.len { 7,661 ( 0.00%) self.buf.shrink_to_fit(self.len); . } 852 ( 0.00%) } . . /// Shrinks the capacity of the vector with a lower bound. . /// . /// The capacity will remain at least as large as both the length . /// and the supplied value. . /// . /// If the current capacity is less than the lower limit, this is a no-op. . /// -- line 949 ---------------------------------------- -- line 987 ---------------------------------------- . /// vec.extend([1, 2, 3]); . /// . /// assert_eq!(vec.capacity(), 10); . /// let slice = vec.into_boxed_slice(); . /// assert_eq!(slice.into_vec().capacity(), 3); . /// ``` . #[cfg(not(no_global_oom_handling))] . #[stable(feature = "rust1", since = "1.0.0")] 852 ( 0.00%) pub fn into_boxed_slice(mut self) -> Box<[T], A> { . unsafe { 2,556 ( 0.00%) self.shrink_to_fit(); 5,414 ( 0.00%) let me = ManuallyDrop::new(self); 4,260 ( 0.00%) let buf = ptr::read(&me.buf); 2,130 ( 0.00%) let len = me.len(); 5,538 ( 0.00%) buf.into_box(len).assume_init() 426 ( 0.00%) } 852 ( 0.00%) } . . /// Shortens the vector, keeping the first `len` elements and dropping . /// the rest. . /// . /// If `len` is greater than the vector's current length, this has no . /// effect. . /// . /// The [`drain`] method can emulate `truncate`, but causes the excess -- line 1011 ---------------------------------------- -- line 1040 ---------------------------------------- . /// let mut vec = vec![1, 2, 3]; . /// vec.truncate(0); . /// assert_eq!(vec, []); . /// ``` . /// . /// [`clear`]: Vec::clear . /// [`drain`]: Vec::drain . #[stable(feature = "rust1", since = "1.0.0")] 108,008 ( 0.00%) pub fn truncate(&mut self, len: usize) { . // This is safe because: . // . // * the slice passed to `drop_in_place` is valid; the `len > self.len` . // case avoids creating an invalid slice, and . // * the `len` of the vector is shrunk before calling `drop_in_place`, . // such that no value will be dropped twice in case `drop_in_place` . // were to panic once (if it panics twice, the program aborts). . unsafe { . // Note: It's intentional that this is `>` and not `>=`. . // Changing it to `>=` has negative performance . // implications in some cases. See #78884 for more. 83,099 ( 0.00%) if len > self.len { . return; . } 32 ( 0.00%) let remaining_len = self.len - len; 24 ( 0.00%) let s = ptr::slice_from_raw_parts_mut(self.as_mut_ptr().add(len), remaining_len); 80,982 ( 0.00%) self.len = len; 2,042 ( 0.00%) ptr::drop_in_place(s); . } 129,578 ( 0.01%) } . . /// Extracts a slice containing the entire vector. . /// . /// Equivalent to `&s[..]`. . /// . /// # Examples . /// . /// ``` . /// use std::io::{self, Write}; . /// let buffer = vec![1, 2, 3, 5, 8]; . /// io::sink().write(buffer.as_slice()).unwrap(); . /// ``` . #[inline] . #[stable(feature = "vec_as_slice", since = "1.7.0")] 36 ( 0.00%) pub fn as_slice(&self) -> &[T] { 54 ( 0.00%) self 36 ( 0.00%) } . . /// Extracts a mutable slice of the entire vector. . /// . /// Equivalent to `&mut s[..]`. . /// . /// # Examples . /// . /// ``` -- line 1093 ---------------------------------------- -- line 1123 ---------------------------------------- . /// assert_eq!(*x_ptr.add(i), 1 << i); . /// } . /// } . /// ``` . /// . /// [`as_mut_ptr`]: Vec::as_mut_ptr . #[stable(feature = "vec_as_ptr", since = "1.37.0")] . #[inline] 18,452 ( 0.00%) pub fn as_ptr(&self) -> *const T { . // We shadow the slice method of the same name to avoid going through . // `deref`, which creates an intermediate reference. 1,598,962 ( 0.06%) let ptr = self.buf.ptr(); . unsafe { 9,226 ( 0.00%) assume(!ptr.is_null()); . } . ptr 18,452 ( 0.00%) } . . /// Returns an unsafe mutable pointer to the vector's buffer. . /// . /// The caller must ensure that the vector outlives the pointer this . /// function returns, or else it will end up pointing to garbage. . /// Modifying the vector may cause its buffer to be reallocated, . /// which would also make any pointers to it invalid. . /// -- line 1147 ---------------------------------------- -- line 1159 ---------------------------------------- . /// *x_ptr.add(i) = i as i32; . /// } . /// x.set_len(size); . /// } . /// assert_eq!(&*x, &[0, 1, 2, 3]); . /// ``` . #[stable(feature = "vec_as_ptr", since = "1.37.0")] . #[inline] 69,344 ( 0.00%) pub fn as_mut_ptr(&mut self) -> *mut T { . // We shadow the slice method of the same name to avoid going through . // `deref_mut`, which creates an intermediate reference. 1,662,472 ( 0.07%) let ptr = self.buf.ptr(); . unsafe { 34,672 ( 0.00%) assume(!ptr.is_null()); . } . ptr 69,344 ( 0.00%) } . . /// Returns a reference to the underlying allocator. . #[unstable(feature = "allocator_api", issue = "32838")] . #[inline] 2,836 ( 0.00%) pub fn allocator(&self) -> &A { 2,836 ( 0.00%) self.buf.allocator() 2,836 ( 0.00%) } . . /// Forces the length of the vector to `new_len`. . /// . /// This is a low-level operation that maintains none of the normal . /// invariants of the type. Normally changing the length of a vector . /// is done using one of the safe operations instead, such as . /// [`truncate`], [`resize`], [`extend`], or [`clear`]. . /// -- line 1190 ---------------------------------------- -- line 1256 ---------------------------------------- . /// vec.set_len(0); . /// } . /// ``` . /// . /// Normally, here, one would use [`clear`] instead to correctly drop . /// the contents and thus not leak memory. . #[inline] . #[stable(feature = "rust1", since = "1.0.0")] 9,090 ( 0.00%) pub unsafe fn set_len(&mut self, new_len: usize) { . debug_assert!(new_len <= self.capacity()); . 233,672 ( 0.01%) self.len = new_len; 11,644 ( 0.00%) } . . /// Removes an element from the vector and returns it. . /// . /// The removed element is replaced by the last element of the vector. . /// . /// This does not preserve ordering, but is *O*(1). . /// If you need to preserve the element order, use [`remove`] instead. . /// -- line 1276 ---------------------------------------- -- line 1305 ---------------------------------------- . assert_failed(index, len); . } . unsafe { . // We replace self[index] with the last element. Note that if the . // bounds check above succeeds there must be a last element (which . // can be self[index] itself). . let value = ptr::read(self.as_ptr().add(index)); . let base_ptr = self.as_mut_ptr(); 4 ( 0.00%) ptr::copy(base_ptr.add(len - 1), base_ptr.add(index), 1); . self.set_len(len - 1); . value . } . } . . /// Inserts an element at position `index` within the vector, shifting all . /// elements after it to the right. . /// -- line 1321 ---------------------------------------- -- line 1329 ---------------------------------------- . /// let mut vec = vec![1, 2, 3]; . /// vec.insert(1, 4); . /// assert_eq!(vec, [1, 4, 2, 3]); . /// vec.insert(4, 5); . /// assert_eq!(vec, [1, 4, 2, 3, 5]); . /// ``` . #[cfg(not(no_global_oom_handling))] . #[stable(feature = "rust1", since = "1.0.0")] 2,503 ( 0.00%) pub fn insert(&mut self, index: usize, element: T) { . #[cold] . #[inline(never)] . fn assert_failed(index: usize, len: usize) -> ! { . panic!("insertion index (is {}) should be <= len (is {})", index, len); . } . 383 ( 0.00%) let len = self.len(); 909 ( 0.00%) if index > len { . assert_failed(index, len); . } . . // space for the new element 1,638 ( 0.00%) if len == self.buf.capacity() { . self.reserve(1); . } . . unsafe { . // infallible . // The spot to put the new value . { 21 ( 0.00%) let p = self.as_mut_ptr().add(index); . // Shift everything over to make space. (Duplicating the . // `index`th element into two consecutive places.) 2,424 ( 0.00%) ptr::copy(p, p.offset(1), len - index); . // Write it in, overwriting the first copy of the `index`th . // element. 57 ( 0.00%) ptr::write(p, element); . } 1,656 ( 0.00%) self.set_len(len + 1); . } 2,136 ( 0.00%) } . . /// Removes and returns the element at position `index` within the vector, . /// shifting all elements after it to the left. . /// . /// Note: Because this shifts over the remaining elements, it has a . /// worst-case performance of *O*(*n*). If you don't need the order of elements . /// to be preserved, use [`swap_remove`] instead. If you'd like to remove . /// elements from the beginning of the `Vec`, consider using -- line 1376 ---------------------------------------- -- line 1387 ---------------------------------------- . /// . /// ``` . /// let mut v = vec![1, 2, 3]; . /// assert_eq!(v.remove(1), 2); . /// assert_eq!(v, [1, 3]); . /// ``` . #[stable(feature = "rust1", since = "1.0.0")] . #[track_caller] 2,920 ( 0.00%) pub fn remove(&mut self, index: usize) -> T { . #[cold] . #[inline(never)] . #[track_caller] . fn assert_failed(index: usize, len: usize) -> ! { . panic!("removal index (is {}) should be < len (is {})", index, len); . } . 1,273 ( 0.00%) let len = self.len(); 3,914 ( 0.00%) if index >= len { . assert_failed(index, len); . } . unsafe { . // infallible . let ret; . { . // the place we are taking from. 50 ( 0.00%) let ptr = self.as_mut_ptr().add(index); . // copy it out, unsafely having a copy of the value on . // the stack and in the vector at the same time. 7,318 ( 0.00%) ret = ptr::read(ptr); . . // Shift everything down to fill in that spot. 2,692 ( 0.00%) ptr::copy(ptr.offset(1), ptr, len - index - 1); . } 1,385 ( 0.00%) self.set_len(len - 1); 10 ( 0.00%) ret 10 ( 0.00%) } 4,720 ( 0.00%) } . . /// Retains only the elements specified by the predicate. . /// . /// In other words, remove all elements `e` such that `f(&e)` returns `false`. . /// This method operates in place, visiting each element exactly once in the . /// original order, and preserves the order of the retained elements. . /// . /// # Examples -- line 1431 ---------------------------------------- -- line 1442 ---------------------------------------- . /// ``` . /// let mut vec = vec![1, 2, 3, 4, 5]; . /// let keep = [false, true, true, false, true]; . /// let mut iter = keep.iter(); . /// vec.retain(|_| *iter.next().unwrap()); . /// assert_eq!(vec, [2, 3, 5]); . /// ``` . #[stable(feature = "rust1", since = "1.0.0")] 132,639 ( 0.01%) pub fn retain(&mut self, mut f: F) . where . F: FnMut(&T) -> bool, . { 4,205 ( 0.00%) self.retain_mut(|elem| f(elem)); 129,168 ( 0.01%) } . . /// Retains only the elements specified by the predicate, passing a mutable reference to it. . /// . /// In other words, remove all elements `e` such that `f(&mut e)` returns `false`. . /// This method operates in place, visiting each element exactly once in the . /// original order, and preserves the order of the retained elements. . /// . /// # Examples -- line 1463 ---------------------------------------- -- line 1474 ---------------------------------------- . /// }); . /// assert_eq!(vec, [2, 3, 4]); . /// ``` . #[unstable(feature = "vec_retain_mut", issue = "90829")] . pub fn retain_mut(&mut self, mut f: F) . where . F: FnMut(&mut T) -> bool, . { 16,587 ( 0.00%) let original_len = self.len(); . // Avoid double drop if the drop guard is not executed, . // since we may make some holes during the process. . unsafe { self.set_len(0) }; . . // Vec: [Kept, Kept, Hole, Hole, Hole, Hole, Unchecked, Unchecked] . // |<- processed len ->| ^- next to check . // |<- deleted cnt ->| . // |<- original_len ->| -- line 1490 ---------------------------------------- -- line 1499 ---------------------------------------- . v: &'a mut Vec, . processed_len: usize, . deleted_cnt: usize, . original_len: usize, . } . . impl Drop for BackshiftOnDrop<'_, T, A> { . fn drop(&mut self) { 17,132 ( 0.00%) if self.deleted_cnt > 0 { . // SAFETY: Trailing unchecked items must be valid since we never touch them. . unsafe { . ptr::copy( . self.v.as_ptr().add(self.processed_len), 12,853 ( 0.00%) self.v.as_mut_ptr().add(self.processed_len - self.deleted_cnt), . self.original_len - self.processed_len, . ); . } . } . // SAFETY: After filling holes, all items are in contiguous memory. . unsafe { 2 ( 0.00%) self.v.set_len(self.original_len - self.deleted_cnt); . } . } . } . . let mut g = BackshiftOnDrop { v: self, processed_len: 0, deleted_cnt: 0, original_len }; . . fn process_loop( . original_len: usize, . f: &mut F, . g: &mut BackshiftOnDrop<'_, T, A>, . ) where . F: FnMut(&mut T) -> bool, . { 161,172 ( 0.01%) while g.processed_len != original_len { . // SAFETY: Unchecked element must be valid. . let cur = unsafe { &mut *g.v.as_mut_ptr().add(g.processed_len) }; 40,502 ( 0.00%) if !f(cur) { . // Advance early to avoid double drop if `drop_in_place` panicked. 7,786 ( 0.00%) g.processed_len += 1; 3,786 ( 0.00%) g.deleted_cnt += 1; . // SAFETY: We never touch this element again after dropped. 48 ( 0.00%) unsafe { ptr::drop_in_place(cur) }; . // We already advanced the counter. . if DELETED { . continue; . } else { . break; . } . } . if DELETED { . // SAFETY: `deleted_cnt` > 0, so the hole slot must not overlap with current element. . // We use copy for move, and never touch this element again. . unsafe { 5,481 ( 0.00%) let hole_slot = g.v.as_mut_ptr().add(g.processed_len - g.deleted_cnt); . ptr::copy_nonoverlapping(cur, hole_slot, 1); . } . } 5,186 ( 0.00%) g.processed_len += 1; . } . } . . // Stage 1: Nothing was deleted. . process_loop::(original_len, &mut f, &mut g); . . // Stage 2: Some elements were deleted. . process_loop::(original_len, &mut f, &mut g); -- line 1565 ---------------------------------------- -- line 1606 ---------------------------------------- . /// ``` . /// let mut vec = vec!["foo", "bar", "Bar", "baz", "bar"]; . /// . /// vec.dedup_by(|a, b| a.eq_ignore_ascii_case(b)); . /// . /// assert_eq!(vec, ["foo", "bar", "baz", "bar"]); . /// ``` . #[stable(feature = "dedup_by", since = "1.16.0")] 3,423 ( 0.00%) pub fn dedup_by(&mut self, mut same_bucket: F) . where . F: FnMut(&mut T, &mut T) -> bool, . { 974 ( 0.00%) let len = self.len(); 2,826 ( 0.00%) if len <= 1 { . return; . } . . /* INVARIANT: vec.len() > read >= write > write-1 >= 0 */ . struct FillGapOnDrop<'a, T, A: core::alloc::Allocator> { . /* Offset of the element we want to check if it is duplicate */ . read: usize, . -- line 1627 ---------------------------------------- -- line 1670 ---------------------------------------- . let ptr = gap.vec.as_mut_ptr(); . . /* Drop items while going through Vec, it should be more efficient than . * doing slice partition_dedup + truncate */ . . /* SAFETY: Because of the invariant, read_ptr, prev_ptr and write_ptr . * are always in-bounds and read_ptr never aliases prev_ptr */ . unsafe { 3,947 ( 0.00%) while gap.read < len { . let read_ptr = ptr.add(gap.read); . let prev_ptr = ptr.add(gap.write.wrapping_sub(1)); . 102 ( 0.00%) if same_bucket(&mut *read_ptr, &mut *prev_ptr) { . // Increase `gap.read` now since the drop may panic. . gap.read += 1; . /* We have found duplicate, drop it in-place */ . ptr::drop_in_place(read_ptr); . } else { . let write_ptr = ptr.add(gap.write); . . /* Because `read_ptr` can be equal to `write_ptr`, we either . * have to use `copy` or conditional `copy_nonoverlapping`. . * Looks like the first option is faster. */ . ptr::copy(read_ptr, write_ptr, 1); . . /* We have filled that place, so go further */ 2,507 ( 0.00%) gap.write += 1; . gap.read += 1; . } . } . . /* Technically we could let `gap` clean up with its Drop, but . * when `same_bucket` is guaranteed to not panic, this bloats a little . * the codegen, so we just do it manually */ . gap.vec.set_len(gap.write); . mem::forget(gap); . } 3,912 ( 0.00%) } . . /// Appends an element to the back of a collection. . /// . /// # Panics . /// . /// Panics if the new capacity exceeds `isize::MAX` bytes. . /// . /// # Examples -- line 1715 ---------------------------------------- -- line 1717 ---------------------------------------- . /// ``` . /// let mut vec = vec![1, 2]; . /// vec.push(3); . /// assert_eq!(vec, [1, 2, 3]); . /// ``` . #[cfg(not(no_global_oom_handling))] . #[inline] . #[stable(feature = "rust1", since = "1.0.0")] 96,239 ( 0.00%) pub fn push(&mut self, value: T) { . // This will panic or abort if we would allocate > isize::MAX bytes . // or if the length increment would overflow for zero-sized types. 2,829,628 ( 0.11%) if self.len == self.buf.capacity() { 209,200 ( 0.01%) self.buf.reserve_for_push(self.len); . } . unsafe { 250,694 ( 0.01%) let end = self.as_mut_ptr().add(self.len); 123,412 ( 0.00%) ptr::write(end, value); 3,520,697 ( 0.14%) self.len += 1; . } 33,994 ( 0.00%) } . . /// Removes the last element from a vector and returns it, or [`None`] if it . /// is empty. . /// . /// If you'd like to pop the first element, consider using . /// [`VecDeque::pop_front`] instead. . /// . /// [`VecDeque::pop_front`]: crate::collections::VecDeque::pop_front -- line 1744 ---------------------------------------- -- line 1748 ---------------------------------------- . /// ``` . /// let mut vec = vec![1, 2, 3]; . /// assert_eq!(vec.pop(), Some(3)); . /// assert_eq!(vec, [1, 2]); . /// ``` . #[inline] . #[stable(feature = "rust1", since = "1.0.0")] . pub fn pop(&mut self) -> Option { 501,807 ( 0.02%) if self.len == 0 { 25 ( 0.00%) None . } else { . unsafe { 376,025 ( 0.01%) self.len -= 1; 75,832 ( 0.00%) Some(ptr::read(self.as_ptr().add(self.len()))) . } . } . } . . /// Moves all the elements of `other` into `Self`, leaving `other` empty. . /// . /// # Panics . /// -- line 1769 ---------------------------------------- -- line 1776 ---------------------------------------- . /// let mut vec2 = vec![4, 5, 6]; . /// vec.append(&mut vec2); . /// assert_eq!(vec, [1, 2, 3, 4, 5, 6]); . /// assert_eq!(vec2, []); . /// ``` . #[cfg(not(no_global_oom_handling))] . #[inline] . #[stable(feature = "append", since = "1.4.0")] 837 ( 0.00%) pub fn append(&mut self, other: &mut Self) { . unsafe { 72 ( 0.00%) self.append_elements(other.as_slice() as _); 54 ( 0.00%) other.set_len(0); . } 630 ( 0.00%) } . . /// Appends elements to `Self` from other buffer. . #[cfg(not(no_global_oom_handling))] . #[inline] 18,920 ( 0.00%) unsafe fn append_elements(&mut self, other: *const [T]) { 1,892 ( 0.00%) let count = unsafe { (*other).len() }; 1,892 ( 0.00%) self.reserve(count); 81,849 ( 0.00%) let len = self.len(); 5,756 ( 0.00%) unsafe { ptr::copy_nonoverlapping(other as *const T, self.as_mut_ptr().add(len), count) }; 413,343 ( 0.02%) self.len += count; 3,784 ( 0.00%) } . . /// Creates a draining iterator that removes the specified range in the vector . /// and yields the removed items. . /// . /// When the iterator **is** dropped, all elements in the range are removed . /// from the vector, even if the iterator was not fully consumed. If the . /// iterator **is not** dropped (with [`mem::forget`] for example), it is . /// unspecified how many elements are removed. -- line 1808 ---------------------------------------- -- line 1820 ---------------------------------------- . /// assert_eq!(v, &[1]); . /// assert_eq!(u, &[2, 3]); . /// . /// // A full range clears the vector . /// v.drain(..); . /// assert_eq!(v, &[]); . /// ``` . #[stable(feature = "drain", since = "1.6.0")] 13,818 ( 0.00%) pub fn drain(&mut self, range: R) -> Drain<'_, T, A> . where . R: RangeBounds, . { . // Memory safety . // . // When the Drain is first created, it shortens the length of . // the source vector to make sure no uninitialized or moved-from elements . // are accessible at all if the Drain's destructor never gets to run. . // . // Drain will ptr::read out the values to remove. . // When finished, remaining tail of the vec is copied back to cover . // the hole, and the vector length is restored to the new length. . // 20,797 ( 0.00%) let len = self.len(); 29,610 ( 0.00%) let Range { start, end } = slice::range(range, ..len); . . unsafe { . // set self.vec length's to start, to be safe in case Drain is leaked 5,922 ( 0.00%) self.set_len(start); . // Use the borrow in the IterMut to indicate borrowing behavior of the . // whole Drain iterator (like &mut T). 31,769 ( 0.00%) let range_slice = slice::from_raw_parts_mut(self.as_mut_ptr().add(start), end - start); 10,694 ( 0.00%) Drain { . tail_start: end, 10,028 ( 0.00%) tail_len: len - end, 9,870 ( 0.00%) iter: range_slice.iter(), 7,896 ( 0.00%) vec: NonNull::from(self), . } . } 3,948 ( 0.00%) } . . /// Clears the vector, removing all values. . /// . /// Note that this method has no effect on the allocated capacity . /// of the vector. . /// . /// # Examples . /// -- line 1866 ---------------------------------------- -- line 1869 ---------------------------------------- . /// . /// v.clear(); . /// . /// assert!(v.is_empty()); . /// ``` . #[inline] . #[stable(feature = "rust1", since = "1.0.0")] . pub fn clear(&mut self) { 190,733 ( 0.01%) self.truncate(0) . } . . /// Returns the number of elements in the vector, also referred to . /// as its 'length'. . /// . /// # Examples . /// . /// ``` . /// let a = vec![1, 2, 3]; . /// assert_eq!(a.len(), 3); . /// ``` . #[inline] . #[stable(feature = "rust1", since = "1.0.0")] 19,604 ( 0.00%) pub fn len(&self) -> usize { 9,802 ( 0.00%) self.len 19,604 ( 0.00%) } . . /// Returns `true` if the vector contains no elements. . /// . /// # Examples . /// . /// ``` . /// let mut v = Vec::new(); . /// assert!(v.is_empty()); . /// . /// v.push(1); . /// assert!(!v.is_empty()); . /// ``` . #[stable(feature = "rust1", since = "1.0.0")] 7,378 ( 0.00%) pub fn is_empty(&self) -> bool { 194,225 ( 0.01%) self.len() == 0 14,756 ( 0.00%) } . . /// Splits the collection into two at the given index. . /// . /// Returns a newly allocated vector containing the elements in the range . /// `[at, len)`. After the call, the original vector will be left containing . /// the elements `[0, at)` with its previous capacity unchanged. . /// . /// # Panics -- line 1917 ---------------------------------------- -- line 1935 ---------------------------------------- . A: Clone, . { . #[cold] . #[inline(never)] . fn assert_failed(at: usize, len: usize) -> ! { . panic!("`at` split index (is {}) should be <= len (is {})", at, len); . } . 879 ( 0.00%) if at > self.len() { . assert_failed(at, self.len()); . } . 679 ( 0.00%) if at == 0 { . // the new vector can take over the original buffer and avoid the copy . return mem::replace( . self, . Vec::with_capacity_in(self.capacity(), self.allocator().clone()), . ); . } . 4 ( 0.00%) let other_len = self.len - at; . let mut other = Vec::with_capacity_in(other_len, self.allocator().clone()); . . // Unsafely `set_len` and copy items to `other`. . unsafe { . self.set_len(at); . other.set_len(other_len); . . ptr::copy_nonoverlapping(self.as_ptr().add(at), other.as_mut_ptr(), other.len()); . } 8 ( 0.00%) other . } . . /// Resizes the `Vec` in-place so that `len` is equal to `new_len`. . /// . /// If `new_len` is greater than `len`, the `Vec` is extended by the . /// difference, with each additional slot filled with the result of . /// calling the closure `f`. The return values from `f` will end up . /// in the `Vec` in the order they have been generated. -- line 1973 ---------------------------------------- -- line 1988 ---------------------------------------- . /// . /// let mut vec = vec![]; . /// let mut p = 1; . /// vec.resize_with(4, || { p *= 2; p }); . /// assert_eq!(vec, [2, 4, 8, 16]); . /// ``` . #[cfg(not(no_global_oom_handling))] . #[stable(feature = "vec_resize_with", since = "1.33.0")] 9,777 ( 0.00%) pub fn resize_with(&mut self, new_len: usize, f: F) . where . F: FnMut() -> T, . { 1,311 ( 0.00%) let len = self.len(); 2,622 ( 0.00%) if new_len > len { 9,648 ( 0.00%) self.extend_with(new_len - len, ExtendFunc(f)); . } else { . self.truncate(new_len); . } 8,466 ( 0.00%) } . . /// Consumes and leaks the `Vec`, returning a mutable reference to the contents, . /// `&'a mut [T]`. Note that the type `T` must outlive the chosen lifetime . /// `'a`. If the type has only static references, or none at all, then this . /// may be chosen to be `'static`. . /// . /// As of Rust 1.57, this method does not reallocate or shrink the `Vec`, . /// so the leaked allocation may include unused capacity that is not part -- line 2014 ---------------------------------------- -- line 2064 ---------------------------------------- . /// unsafe { . /// v.set_len(3); . /// } . /// . /// assert_eq!(&v, &[0, 1, 2]); . /// ``` . #[stable(feature = "vec_spare_capacity", since = "1.60.0")] . #[inline] 2,337 ( 0.00%) pub fn spare_capacity_mut(&mut self) -> &mut [MaybeUninit] { . // Note: . // This method is not implemented in terms of `split_at_spare_mut`, . // to prevent invalidation of pointers to the buffer. . unsafe { 3,116 ( 0.00%) slice::from_raw_parts_mut( 5,455 ( 0.00%) self.as_mut_ptr().add(self.len) as *mut MaybeUninit, 3,968 ( 0.00%) self.buf.capacity() - self.len, . ) . } 1,558 ( 0.00%) } . . /// Returns vector content as a slice of `T`, along with the remaining spare . /// capacity of the vector as a slice of `MaybeUninit`. . /// . /// The returned spare capacity slice can be used to fill the vector with data . /// (e.g. by reading from a file) before marking the data as initialized using . /// the [`set_len`] method. . /// -- line 2090 ---------------------------------------- -- line 2189 ---------------------------------------- . /// assert_eq!(vec, ["hello", "world", "world"]); . /// . /// let mut vec = vec![1, 2, 3, 4]; . /// vec.resize(2, 0); . /// assert_eq!(vec, [1, 2]); . /// ``` . #[cfg(not(no_global_oom_handling))] . #[stable(feature = "vec_resize", since = "1.5.0")] 10,706 ( 0.00%) pub fn resize(&mut self, new_len: usize, value: T) { 1,121 ( 0.00%) let len = self.len(); . 3,757 ( 0.00%) if new_len > len { 2,227 ( 0.00%) self.extend_with(new_len - len, ExtendElement(value)) . } else { 12 ( 0.00%) self.truncate(new_len); . } 1,361 ( 0.00%) } . . /// Clones and appends all elements in a slice to the `Vec`. . /// . /// Iterates over the slice `other`, clones each element, and then appends . /// it to this `Vec`. The `other` slice is traversed in-order. . /// . /// Note that this function is same as [`extend`] except that it is . /// specialized to work with slices instead. If and when Rust gets -- line 2213 ---------------------------------------- -- line 2220 ---------------------------------------- . /// let mut vec = vec![1]; . /// vec.extend_from_slice(&[2, 3, 4]); . /// assert_eq!(vec, [1, 2, 3, 4]); . /// ``` . /// . /// [`extend`]: Vec::extend . #[cfg(not(no_global_oom_handling))] . #[stable(feature = "vec_extend_from_slice", since = "1.6.0")] 18,150 ( 0.00%) pub fn extend_from_slice(&mut self, other: &[T]) { 6,600 ( 0.00%) self.spec_extend(other.iter()) 3,300 ( 0.00%) } . . /// Copies elements from `src` range to the end of the vector. . /// . /// # Panics . /// . /// Panics if the starting point is greater than the end point or if . /// the end point is greater than the length of the vector. . /// -- line 2238 ---------------------------------------- -- line 2270 ---------------------------------------- . // This code generalizes `extend_with_{element,default}`. . trait ExtendWith { . fn next(&mut self) -> T; . fn last(self) -> T; . } . . struct ExtendElement(T); . impl ExtendWith for ExtendElement { 8 ( 0.00%) fn next(&mut self) -> T { 187 ( 0.00%) self.0.clone() 8 ( 0.00%) } 12 ( 0.00%) fn last(self) -> T { . self.0 8 ( 0.00%) } . } . . struct ExtendFunc(F); . impl T> ExtendWith for ExtendFunc { . fn next(&mut self) -> T { . (self.0)() . } . fn last(mut self) -> T { . (self.0)() . } . } . . impl Vec { . #[cfg(not(no_global_oom_handling))] . /// Extend the vector by `n` values, using the given generator. 100,012 ( 0.00%) fn extend_with>(&mut self, n: usize, mut value: E) { 20 ( 0.00%) self.reserve(n); . . unsafe { 2,537 ( 0.00%) let mut ptr = self.as_mut_ptr().add(self.len()); . // Use SetLenOnDrop to work around bug where compiler . // might not realize the store through `ptr` through self.set_len() . // don't alias. 28 ( 0.00%) let mut local_len = SetLenOnDrop::new(&mut self.len); . . // Write all elements except the last one 148 ( 0.00%) for _ in 1..n { 44 ( 0.00%) ptr::write(ptr, value.next()); 24 ( 0.00%) ptr = ptr.offset(1); . // Increment the length in every step in case next() panics 8 ( 0.00%) local_len.increment_len(1); . } . 120,189 ( 0.00%) if n > 0 { . // We can write the last element directly without cloning needlessly 48 ( 0.00%) ptr::write(ptr, value.last()); 8 ( 0.00%) local_len.increment_len(1); . } . . // len set by scope guard 16 ( 0.00%) } 79,539 ( 0.00%) } . } . . impl Vec { . /// Removes consecutive repeated elements in the vector according to the . /// [`PartialEq`] trait implementation. . /// . /// If the vector is sorted, this removes all duplicates. . /// -- line 2333 ---------------------------------------- -- line 2338 ---------------------------------------- . /// . /// vec.dedup(); . /// . /// assert_eq!(vec, [1, 2, 3, 2]); . /// ``` . #[stable(feature = "rust1", since = "1.0.0")] . #[inline] . pub fn dedup(&mut self) { 490 ( 0.00%) self.dedup_by(|a, b| a == b) . } . } . . //////////////////////////////////////////////////////////////////////////////// . // Internal methods and functions . //////////////////////////////////////////////////////////////////////////////// . . #[doc(hidden)] . #[cfg(not(no_global_oom_handling))] . #[stable(feature = "rust1", since = "1.0.0")] 35,125 ( 0.00%) pub fn from_elem(elem: T, n: usize) -> Vec { 150,518 ( 0.01%) ::from_elem(elem, n, Global) 39,843 ( 0.00%) } . . #[doc(hidden)] . #[cfg(not(no_global_oom_handling))] . #[unstable(feature = "allocator_api", issue = "32838")] . pub fn from_elem_in(elem: T, n: usize, alloc: A) -> Vec { . ::from_elem(elem, n, alloc) . } . -- line 2367 ---------------------------------------- -- line 2423 ---------------------------------------- . //////////////////////////////////////////////////////////////////////////////// . // Common trait implementations for Vec . //////////////////////////////////////////////////////////////////////////////// . . #[stable(feature = "rust1", since = "1.0.0")] . impl ops::Deref for Vec { . type Target = [T]; . 27,678 ( 0.00%) fn deref(&self) -> &[T] { 2,511,290 ( 0.10%) unsafe { slice::from_raw_parts(self.as_ptr(), self.len) } 18,608 ( 0.00%) } . } . . #[stable(feature = "rust1", since = "1.0.0")] . impl ops::DerefMut for Vec { 273 ( 0.00%) fn deref_mut(&mut self) -> &mut [T] { 551,574 ( 0.02%) unsafe { slice::from_raw_parts_mut(self.as_mut_ptr(), self.len) } 182 ( 0.00%) } . } . . #[cfg(not(no_global_oom_handling))] . trait SpecCloneFrom { . fn clone_from(this: &mut Self, other: &Self); . } . . #[cfg(not(no_global_oom_handling))] -- line 2448 ---------------------------------------- -- line 2468 ---------------------------------------- . this.extend_from_slice(other); . } . } . . #[cfg(not(no_global_oom_handling))] . #[stable(feature = "rust1", since = "1.0.0")] . impl Clone for Vec { . #[cfg(not(test))] 89,223 ( 0.00%) fn clone(&self) -> Self { 3,890 ( 0.00%) let alloc = self.allocator().clone(); 7,004 ( 0.00%) <[T]>::to_vec_in(&**self, alloc) 108,341 ( 0.00%) } . . // HACK(japaric): with cfg(test) the inherent `[T]::to_vec` method, which is . // required for this method definition, is not available. Instead use the . // `slice::to_vec` function which is only available with cfg(test) . // NB see the slice::hack module in slice.rs for more information . #[cfg(test)] . fn clone(&self) -> Self { . let alloc = self.allocator().clone(); -- line 2487 ---------------------------------------- -- line 2517 ---------------------------------------- . #[rustc_on_unimplemented( . message = "vector indices are of type `usize` or ranges of `usize`", . label = "vector indices are of type `usize` or ranges of `usize`" . )] . impl, A: Allocator> Index for Vec { . type Output = I::Output; . . #[inline] 26 ( 0.00%) fn index(&self, index: I) -> &Self::Output { 55,703 ( 0.00%) Index::index(&**self, index) 12 ( 0.00%) } . } . . #[stable(feature = "rust1", since = "1.0.0")] . #[rustc_on_unimplemented( . message = "vector indices are of type `usize` or ranges of `usize`", . label = "vector indices are of type `usize` or ranges of `usize`" . )] . impl, A: Allocator> IndexMut for Vec { . #[inline] 320 ( 0.00%) fn index_mut(&mut self, index: I) -> &mut Self::Output { 904 ( 0.00%) IndexMut::index_mut(&mut **self, index) 128 ( 0.00%) } . } . . #[cfg(not(no_global_oom_handling))] . #[stable(feature = "rust1", since = "1.0.0")] . impl FromIterator for Vec { . #[inline] 735 ( 0.00%) fn from_iter>(iter: I) -> Vec { 727,229 ( 0.03%) >::from_iter(iter.into_iter()) 362 ( 0.00%) } . } . . #[stable(feature = "rust1", since = "1.0.0")] . impl IntoIterator for Vec { . type Item = T; . type IntoIter = IntoIter; . . /// Creates a consuming iterator, that is, one that moves each value out of -- line 2556 ---------------------------------------- -- line 2562 ---------------------------------------- . /// ``` . /// let v = vec!["a".to_string(), "b".to_string()]; . /// for s in v.into_iter() { . /// // s has type String, not &String . /// println!("{}", s); . /// } . /// ``` . #[inline] 1,920 ( 0.00%) fn into_iter(self) -> IntoIter { . unsafe { 237,805 ( 0.01%) let mut me = ManuallyDrop::new(self); 1,280 ( 0.00%) let alloc = ptr::read(me.allocator()); 3,200 ( 0.00%) let begin = me.as_mut_ptr(); 1,920 ( 0.00%) let end = if mem::size_of::() == 0 { . arith_offset(begin as *const i8, me.len() as isize) as *const T . } else { 3,200 ( 0.00%) begin.add(me.len()) as *const T . }; 640 ( 0.00%) let cap = me.buf.capacity(); 98,284 ( 0.00%) IntoIter { 2,560 ( 0.00%) buf: NonNull::new_unchecked(begin), . phantom: PhantomData, . cap, . alloc, . ptr: begin, 640 ( 0.00%) end, . } . } 1,280 ( 0.00%) } . } . . #[stable(feature = "rust1", since = "1.0.0")] . impl<'a, T, A: Allocator> IntoIterator for &'a Vec { . type Item = &'a T; . type IntoIter = slice::Iter<'a, T>; . 430 ( 0.00%) fn into_iter(self) -> slice::Iter<'a, T> { 1,811 ( 0.00%) self.iter() 430 ( 0.00%) } . } . . #[stable(feature = "rust1", since = "1.0.0")] . impl<'a, T, A: Allocator> IntoIterator for &'a mut Vec { . type Item = &'a mut T; . type IntoIter = slice::IterMut<'a, T>; . 8 ( 0.00%) fn into_iter(self) -> slice::IterMut<'a, T> { 1,132 ( 0.00%) self.iter_mut() 8 ( 0.00%) } . } . . #[cfg(not(no_global_oom_handling))] . #[stable(feature = "rust1", since = "1.0.0")] . impl Extend for Vec { . #[inline] 3,767 ( 0.00%) fn extend>(&mut self, iter: I) { 260,260 ( 0.01%) >::spec_extend(self, iter.into_iter()) 3,442 ( 0.00%) } . . #[inline] . fn extend_one(&mut self, item: T) { 64 ( 0.00%) self.push(item); . } . . #[inline] . fn extend_reserve(&mut self, additional: usize) { . self.reserve(additional); . } . } . . impl Vec { . // leaf method to which various SpecFrom/SpecExtend implementations delegate when . // they have no further optimizations to apply . #[cfg(not(no_global_oom_handling))] 296 ( 0.00%) fn extend_desugared>(&mut self, mut iterator: I) { . // This is the case for a general iterator. . // . // This function should be the moral equivalent of: . // . // for item in iterator { . // self.push(item); . // } 61,812 ( 0.00%) while let Some(element) = iterator.next() { 7,383 ( 0.00%) let len = self.len(); 131,526 ( 0.01%) if len == self.capacity() { 1,054 ( 0.00%) let (lower, _) = iterator.size_hint(); 45 ( 0.00%) self.reserve(lower.saturating_add(1)); . } . unsafe { 843 ( 0.00%) ptr::write(self.as_mut_ptr().add(len), element); . // Since next() executes user code which can panic we have to bump the length . // after each step. . // NB can't overflow since we would have had to alloc the address space 97,661 ( 0.00%) self.set_len(len + 1); . } 59 ( 0.00%) } 1,111 ( 0.00%) } . . /// Creates a splicing iterator that replaces the specified range in the vector . /// with the given `replace_with` iterator and yields the removed items. . /// `replace_with` does not need to be the same length as `range`. . /// . /// `range` is removed even if the iterator is not consumed until the end. . /// . /// It is unspecified how many elements are removed from the vector -- line 2666 ---------------------------------------- -- line 2693 ---------------------------------------- . #[cfg(not(no_global_oom_handling))] . #[inline] . #[stable(feature = "vec_splice", since = "1.21.0")] . pub fn splice(&mut self, range: R, replace_with: I) -> Splice<'_, I::IntoIter, A> . where . R: RangeBounds, . I: IntoIterator, . { 407 ( 0.00%) Splice { drain: self.drain(range), replace_with: replace_with.into_iter() } . } . . /// Creates an iterator which uses a closure to determine if an element should be removed. . /// . /// If the closure returns true, then the element is removed and yielded. . /// If the closure returns false, the element will remain in the vector and will not be yielded . /// by the iterator. . /// -- line 2709 ---------------------------------------- -- line 2745 ---------------------------------------- . /// assert_eq!(evens, vec![2, 4, 6, 8, 14]); . /// assert_eq!(odds, vec![1, 3, 5, 9, 11, 13, 15]); . /// ``` . #[unstable(feature = "drain_filter", reason = "recently added", issue = "43244")] . pub fn drain_filter(&mut self, filter: F) -> DrainFilter<'_, T, F, A> . where . F: FnMut(&mut T) -> bool, . { 2,275 ( 0.00%) let old_len = self.len(); . . // Guard against us getting leaked (leak amplification) . unsafe { . self.set_len(0); . } . 7,912 ( 0.00%) DrainFilter { vec: self, idx: 0, del: 0, old_len, pred: filter, panic_flag: false } . } . } . . /// Extend implementation that copies elements out of references before pushing them onto the Vec. . /// . /// This implementation is specialized for slice iterators, where it uses [`copy_from_slice`] to . /// append the entire slice at once. . /// . /// [`copy_from_slice`]: slice::copy_from_slice . #[cfg(not(no_global_oom_handling))] . #[stable(feature = "extend_ref", since = "1.2.0")] . impl<'a, T: Copy + 'a, A: Allocator + 'a> Extend<&'a T> for Vec { 7,112 ( 0.00%) fn extend>(&mut self, iter: I) { . self.spec_extend(iter.into_iter()) 5,334 ( 0.00%) } . . #[inline] . fn extend_one(&mut self, &item: &'a T) { . self.push(item); . } . . #[inline] . fn extend_reserve(&mut self, additional: usize) { -- line 2783 ---------------------------------------- -- line 2803 ---------------------------------------- . #[inline] . fn cmp(&self, other: &Self) -> Ordering { . Ord::cmp(&**self, &**other) . } . } . . #[stable(feature = "rust1", since = "1.0.0")] . unsafe impl<#[may_dangle] T, A: Allocator> Drop for Vec { 389,000 ( 0.02%) fn drop(&mut self) { . unsafe { . // use drop for [T] . // use a raw slice to refer to the elements of the vector as weakest necessary type; . // could avoid questions of validity in certain cases 249,974 ( 0.01%) ptr::drop_in_place(ptr::slice_from_raw_parts_mut(self.as_mut_ptr(), self.len)) . } . // RawVec handles deallocation 438,394 ( 0.02%) } . } . . #[stable(feature = "rust1", since = "1.0.0")] . #[rustc_const_unstable(feature = "const_default_impls", issue = "87864")] . impl const Default for Vec { . /// Creates an empty `Vec`. 270 ( 0.00%) fn default() -> Vec { 90 ( 0.00%) Vec::new() 180 ( 0.00%) } . } . . #[stable(feature = "rust1", since = "1.0.0")] . impl fmt::Debug for Vec { . fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { . fmt::Debug::fmt(&**self, f) . } . } -- line 2836 ---------------------------------------- -- line 2976 ---------------------------------------- . /// newly-allocated buffer with exactly the right capacity. . /// . /// # Examples . /// . /// ``` . /// assert_eq!(Box::from(vec![1, 2, 3]), vec![1, 2, 3].into_boxed_slice()); . /// ``` . fn from(v: Vec) -> Self { 4 ( 0.00%) v.into_boxed_slice() . } . } . . #[cfg(not(no_global_oom_handling))] . #[stable(feature = "rust1", since = "1.0.0")] . impl From<&str> for Vec { . /// Allocate a `Vec` and fill it with a UTF-8 string. . /// -- line 2992 ---------------------------------------- 1,053,333 ( 0.04%) -------------------------------------------------------------------------------- -- Auto-annotated source: /usr/home/liquid/.cargo/registry/src/github.com-1ecc6299db9ec823/hashbrown-0.12.0/src/raw/mod.rs -------------------------------------------------------------------------------- Ir -- line 110 ---------------------------------------- . /// Control byte value for an empty bucket. . const EMPTY: u8 = 0b1111_1111; . . /// Control byte value for a deleted bucket. . const DELETED: u8 = 0b1000_0000; . . /// Checks whether a control byte represents a full bucket (top bit is clear). . #[inline] 258 ( 0.00%) fn is_full(ctrl: u8) -> bool { 1,035,582 ( 0.04%) ctrl & 0x80 == 0 344 ( 0.00%) } . . /// Checks whether a control byte represents a special value (top bit is set). . #[inline] . fn is_special(ctrl: u8) -> bool { . ctrl & 0x80 != 0 . } . . /// Checks whether a special control value is EMPTY (just check 1 bit). . #[inline] 90 ( 0.00%) fn special_is_empty(ctrl: u8) -> bool { . debug_assert!(is_special(ctrl)); 91,922 ( 0.00%) ctrl & 0x01 != 0 120 ( 0.00%) } . . /// Primary hash function, used to select the initial bucket to probe from. . #[inline] . #[allow(clippy::cast_possible_truncation)] 231 ( 0.00%) fn h1(hash: u64) -> usize { . // On 32-bit platforms we simply ignore the higher hash bits. . hash as usize 154 ( 0.00%) } . . /// Secondary hash function, saved in the low 7 bits of the control byte. . #[inline] . #[allow(clippy::cast_possible_truncation)] 210 ( 0.00%) fn h2(hash: u64) -> u8 { . // Grab the top 7 bits of the hash. While the hash is normally a full 64-bit . // value, some hash functions (such as FxHash) produce a usize result . // instead, which means that the top 32 bits are 0 on 32-bit platforms. 420 ( 0.00%) let hash_len = usize::min(mem::size_of::(), mem::size_of::()); 14,659,207 ( 0.58%) let top7 = hash >> (hash_len * 8 - 7); 70 ( 0.00%) (top7 & 0x7f) as u8 // truncation 140 ( 0.00%) } . . /// Probe sequence based on triangular numbers, which is guaranteed (since our . /// table size is a power of two) to visit every group of elements exactly once. . /// . /// A triangular probe has us jump by 1 more group every time. So first we . /// jump by 1 group (meaning we just continue our linear scan), then 2 groups . /// (skipping over 1 group), then 3 groups (skipping over 2 groups), and so on. . /// -- line 161 ---------------------------------------- -- line 170 ---------------------------------------- . #[inline] . fn move_next(&mut self, bucket_mask: usize) { . // We should have found an empty bucket by now and ended the probe. . debug_assert!( . self.stride <= bucket_mask, . "Went past end of probe sequence" . ); . 132,520 ( 0.01%) self.stride += Group::WIDTH; 132,520 ( 0.01%) self.pos += self.stride; 109,564 ( 0.00%) self.pos &= bucket_mask; . } . } . . /// Returns the number of buckets needed to hold the given number of items, . /// taking the maximum load factor into account. . /// . /// Returns `None` if an overflow occurs. . // Workaround for emscripten bug emscripten-core/emscripten-fastcomp#258 . #[cfg_attr(target_os = "emscripten", inline(never))] . #[cfg_attr(not(target_os = "emscripten"), inline)] 21 ( 0.00%) fn capacity_to_buckets(cap: usize) -> Option { . debug_assert_ne!(cap, 0); . . // For small tables we require at least 1 empty bucket so that lookups are . // guaranteed to terminate if an element doesn't exist in the table. 57,160 ( 0.00%) if cap < 8 { . // We don't bother with a table size of 2 buckets since that can only . // hold a single element. Instead we skip directly to a 4 bucket table . // which can hold 3 elements. 117,191 ( 0.00%) return Some(if cap < 4 { 4 } else { 8 }); . } . . // Otherwise require 1/8 buckets to be empty (87.5% load) . // . // Be careful when modifying this, calculate_layout relies on the . // overflow check here. 30,884 ( 0.00%) let adjusted_cap = cap.checked_mul(8)? / 7; . . // Any overflows will have been caught by the checked_mul. Also, any . // rounding errors from the division above will be cleaned up by . // next_power_of_two (which can't overflow because of the previous division). 10 ( 0.00%) Some(adjusted_cap.next_power_of_two()) 30 ( 0.00%) } . . /// Returns the maximum effective capacity for the given bucket mask, taking . /// the maximum load factor into account. . #[inline] 42 ( 0.00%) fn bucket_mask_to_capacity(bucket_mask: usize) -> usize { 153,808 ( 0.01%) if bucket_mask < 8 { . // For tables with 1/2/4/8 buckets, we always reserve one empty slot. . // Keep in mind that the bucket mask is one less than the bucket count. 11 ( 0.00%) bucket_mask . } else { . // For larger tables we reserve 12.5% of the slots as empty. 38,429 ( 0.00%) ((bucket_mask + 1) / 8) * 7 . } 42 ( 0.00%) } . . /// Helper which allows the max calculation for ctrl_align to be statically computed for each T . /// while keeping the rest of `calculate_layout_for` independent of `T` . #[derive(Copy, Clone)] . struct TableLayout { . size: usize, . ctrl_align: usize, . } . . impl TableLayout { . #[inline] 10 ( 0.00%) fn new() -> Self { 40 ( 0.00%) let layout = Layout::new::(); 20 ( 0.00%) Self { 30 ( 0.00%) size: layout.size(), 60 ( 0.00%) ctrl_align: usize::max(layout.align(), Group::WIDTH), . } 40 ( 0.00%) } . . #[inline] 169 ( 0.00%) fn calculate_layout_for(self, buckets: usize) -> Option<(Layout, usize)> { . debug_assert!(buckets.is_power_of_two()); . 26 ( 0.00%) let TableLayout { size, ctrl_align } = self; . // Manual layout calculation since Layout methods are not yet stable. . let ctrl_offset = 111,530 ( 0.00%) size.checked_mul(buckets)?.checked_add(ctrl_align - 1)? & !(ctrl_align - 1); 153,280 ( 0.01%) let len = ctrl_offset.checked_add(buckets + Group::WIDTH)?; . 117 ( 0.00%) Some(( 52 ( 0.00%) unsafe { Layout::from_size_align_unchecked(len, ctrl_align) }, . ctrl_offset, . )) 39 ( 0.00%) } . } . . /// Returns a Layout which describes the allocation required for a hash table, . /// and the offset of the control bytes in the allocation. . /// (the offset is also one past last element of buckets) . /// . /// Returns `None` if an overflow occurs. . #[cfg_attr(feature = "inline-more", inline)] -- line 269 ---------------------------------------- -- line 292 ---------------------------------------- . #[inline] . fn clone(&self) -> Self { . Self { ptr: self.ptr } . } . } . . impl Bucket { . #[inline] 405 ( 0.00%) unsafe fn from_base_index(base: NonNull, index: usize) -> Self { 324 ( 0.00%) let ptr = if mem::size_of::() == 0 { . // won't overflow because index must be less than length . (index + 1) as *mut T . } else { 324 ( 0.00%) base.as_ptr().sub(index) . }; 81 ( 0.00%) Self { 324 ( 0.00%) ptr: NonNull::new_unchecked(ptr), . } 243 ( 0.00%) } . #[inline] . unsafe fn to_base_index(&self, base: NonNull) -> usize { . if mem::size_of::() == 0 { . self.ptr.as_ptr() as usize - 1 . } else { . offset_from(base.as_ptr(), self.ptr.as_ptr()) . } . } . #[inline] 159 ( 0.00%) pub fn as_ptr(&self) -> *mut T { 212 ( 0.00%) if mem::size_of::() == 0 { . // Just return an arbitrary ZST pointer which is properly aligned . mem::align_of::() as *mut T . } else { 318 ( 0.00%) unsafe { self.ptr.as_ptr().sub(1) } . } 159 ( 0.00%) } . #[inline] 25 ( 0.00%) unsafe fn next_n(&self, offset: usize) -> Self { 20 ( 0.00%) let ptr = if mem::size_of::() == 0 { . (self.ptr.as_ptr() as usize + offset) as *mut T . } else { 25 ( 0.00%) self.ptr.as_ptr().sub(offset) . }; 5 ( 0.00%) Self { 20 ( 0.00%) ptr: NonNull::new_unchecked(ptr), . } 15 ( 0.00%) } . #[cfg_attr(feature = "inline-more", inline)] 10 ( 0.00%) pub unsafe fn drop(&self) { 10 ( 0.00%) self.as_ptr().drop_in_place(); 10 ( 0.00%) } . #[inline] . pub unsafe fn read(&self) -> T { 278 ( 0.00%) self.as_ptr().read() . } . #[inline] 69 ( 0.00%) pub unsafe fn write(&self, val: T) { 273 ( 0.00%) self.as_ptr().write(val); 46 ( 0.00%) } . #[inline] 50 ( 0.00%) pub unsafe fn as_ref<'a>(&self) -> &'a T { 50 ( 0.00%) &*self.as_ptr() 50 ( 0.00%) } . #[inline] . pub unsafe fn as_mut<'a>(&self) -> &'a mut T { . &mut *self.as_ptr() . } . #[cfg(feature = "raw")] . #[inline] . pub unsafe fn copy_from_nonoverlapping(&self, other: &Self) { . self.as_ptr().copy_from_nonoverlapping(other.as_ptr(), 1); -- line 362 ---------------------------------------- -- line 392 ---------------------------------------- . . impl RawTable { . /// Creates a new empty hash table without allocating any memory. . /// . /// In effect this returns a table with exactly 1 bucket. However we can . /// leave the data pointer dangling since that bucket is never written to . /// due to our load factor forcing us to always have at least 1 free bucket. . #[inline] 90 ( 0.00%) pub const fn new() -> Self { 240 ( 0.00%) Self { 60 ( 0.00%) table: RawTableInner::new_in(Global), . marker: PhantomData, . } 60 ( 0.00%) } . . /// Attempts to allocate a new hash table with at least enough capacity . /// for inserting the given number of elements without reallocating. . #[cfg(feature = "raw")] . pub fn try_with_capacity(capacity: usize) -> Result { . Self::try_with_capacity_in(capacity, Global) . } . -- line 413 ---------------------------------------- -- line 421 ---------------------------------------- . impl RawTable { . /// Creates a new empty hash table without allocating any memory, using the . /// given allocator. . /// . /// In effect this returns a table with exactly 1 bucket. However we can . /// leave the data pointer dangling since that bucket is never written to . /// due to our load factor forcing us to always have at least 1 free bucket. . #[inline] 3 ( 0.00%) pub fn new_in(alloc: A) -> Self { 371 ( 0.00%) Self { 2 ( 0.00%) table: RawTableInner::new_in(alloc), . marker: PhantomData, . } 2 ( 0.00%) } . . /// Allocates a new hash table with the given number of buckets. . /// . /// The control bytes are left uninitialized. . #[cfg_attr(feature = "inline-more", inline)] . unsafe fn new_uninitialized( . alloc: A, . buckets: usize, . fallibility: Fallibility, . ) -> Result { . debug_assert!(buckets.is_power_of_two()); . . Ok(Self { 110 ( 0.00%) table: RawTableInner::new_uninitialized( . alloc, . TableLayout::new::(), . buckets, . fallibility, . )?, . marker: PhantomData, . }) . } -- line 456 ---------------------------------------- -- line 458 ---------------------------------------- . /// Attempts to allocate a new hash table with at least enough capacity . /// for inserting the given number of elements without reallocating. . fn fallible_with_capacity( . alloc: A, . capacity: usize, . fallibility: Fallibility, . ) -> Result { . Ok(Self { 7,556 ( 0.00%) table: RawTableInner::fallible_with_capacity( . alloc, . TableLayout::new::(), . capacity, . fallibility, . )?, . marker: PhantomData, . }) . } -- line 474 ---------------------------------------- -- line 493 ---------------------------------------- . /// Returns a reference to the underlying allocator. . #[inline] . pub fn allocator(&self) -> &A { . &self.table.alloc . } . . /// Deallocates the table without dropping any entries. . #[cfg_attr(feature = "inline-more", inline)] 9 ( 0.00%) unsafe fn free_buckets(&mut self) { 12 ( 0.00%) self.table.free_buckets(TableLayout::new::()); 6 ( 0.00%) } . . /// Returns pointer to one past last element of data table. . #[inline] 114 ( 0.00%) pub unsafe fn data_end(&self) -> NonNull { 285 ( 0.00%) NonNull::new_unchecked(self.table.ctrl.as_ptr().cast()) 114 ( 0.00%) } . . /// Returns pointer to start of data table. . #[inline] . #[cfg(feature = "nightly")] . pub unsafe fn data_start(&self) -> *mut T { . self.data_end().as_ptr().wrapping_sub(self.buckets()) . } . -- line 517 ---------------------------------------- -- line 518 ---------------------------------------- . /// Returns the index of a bucket from a `Bucket`. . #[inline] . pub unsafe fn bucket_index(&self, bucket: &Bucket) -> usize { . bucket.to_base_index(self.data_end()) . } . . /// Returns a pointer to an element in the table. . #[inline] 96 ( 0.00%) pub unsafe fn bucket(&self, index: usize) -> Bucket { . debug_assert_ne!(self.table.bucket_mask, 0); . debug_assert!(index < self.buckets()); 96 ( 0.00%) Bucket::from_base_index(self.data_end(), index) 48 ( 0.00%) } . . /// Erases an element from the table without dropping it. . #[cfg_attr(feature = "inline-more", inline)] . #[deprecated(since = "0.8.1", note = "use erase or remove instead")] 6,366 ( 0.00%) pub unsafe fn erase_no_drop(&mut self, item: &Bucket) { 6,366 ( 0.00%) let index = self.bucket_index(item); . self.table.erase(index); 12,732 ( 0.00%) } . . /// Erases an element from the table, dropping it in place. . #[cfg_attr(feature = "inline-more", inline)] . #[allow(clippy::needless_pass_by_value)] . #[allow(deprecated)] . pub unsafe fn erase(&mut self, item: Bucket) { . // Erase the element from the table first since drop might panic. 2,156 ( 0.00%) self.erase_no_drop(&item); . item.drop(); . } . . /// Finds and erases an element from the table, dropping it in place. . /// Returns true if an element was found. . #[cfg(feature = "raw")] . #[cfg_attr(feature = "inline-more", inline)] . pub fn erase_entry(&mut self, hash: u64, eq: impl FnMut(&T) -> bool) -> bool { -- line 554 ---------------------------------------- -- line 563 ---------------------------------------- . } . } . . /// Removes an element from the table, returning it. . #[cfg_attr(feature = "inline-more", inline)] . #[allow(clippy::needless_pass_by_value)] . #[allow(deprecated)] . pub unsafe fn remove(&mut self, item: Bucket) -> T { 10,576 ( 0.00%) self.erase_no_drop(&item); 56 ( 0.00%) item.read() . } . . /// Finds and removes an element from the table, returning it. . #[cfg_attr(feature = "inline-more", inline)] 335,045 ( 0.01%) pub fn remove_entry(&mut self, hash: u64, eq: impl FnMut(&T) -> bool) -> Option { . // Avoid `Option::map` because it bloats LLVM IR. 1,550 ( 0.00%) match self.find(hash, eq) { 2,343 ( 0.00%) Some(bucket) => Some(unsafe { self.remove(bucket) }), 35,849 ( 0.00%) None => None, . } 486,604 ( 0.02%) } . . /// Marks all table buckets as empty without dropping their contents. . #[cfg_attr(feature = "inline-more", inline)] . pub fn clear_no_drop(&mut self) { . self.table.clear_no_drop(); . } . . /// Removes all elements from the table without freeing the backing memory. . #[cfg_attr(feature = "inline-more", inline)] . pub fn clear(&mut self) { . // Ensure that the table is reset even if one of the drops panic . let mut self_ = guard(self, |self_| self_.clear_no_drop()); . unsafe { 1 ( 0.00%) self_.drop_elements(); . } . } . 16 ( 0.00%) unsafe fn drop_elements(&mut self) { 6,590 ( 0.00%) if mem::needs_drop::() && !self.is_empty() { 116 ( 0.00%) for item in self.iter() { 10 ( 0.00%) item.drop(); . } . } 14 ( 0.00%) } . . /// Shrinks the table to fit `max(self.len(), min_size)` elements. . #[cfg_attr(feature = "inline-more", inline)] . pub fn shrink_to(&mut self, min_size: usize, hasher: impl Fn(&T) -> u64) { . // Calculate the minimal number of elements that we need to reserve . // space for. . let min_size = usize::max(self.table.items, min_size); . if min_size == 0 { -- line 615 ---------------------------------------- -- line 641 ---------------------------------------- . } . } . } . } . . /// Ensures that at least `additional` items can be inserted into the table . /// without reallocation. . #[cfg_attr(feature = "inline-more", inline)] 49 ( 0.00%) pub fn reserve(&mut self, additional: usize, hasher: impl Fn(&T) -> u64) { 214,301 ( 0.01%) if additional > self.table.growth_left { . // Avoid `Result::unwrap_or_else` because it bloats LLVM IR. 95,993 ( 0.00%) if self 14 ( 0.00%) .reserve_rehash(additional, hasher, Fallibility::Infallible) . .is_err() . { . unsafe { hint::unreachable_unchecked() } . } . } 35 ( 0.00%) } . . /// Tries to ensure that at least `additional` items can be inserted into . /// the table without reallocation. . #[cfg_attr(feature = "inline-more", inline)] . pub fn try_reserve( . &mut self, . additional: usize, . hasher: impl Fn(&T) -> u64, -- line 667 ---------------------------------------- -- line 671 ---------------------------------------- . } else { . Ok(()) . } . } . . /// Out-of-line slow path for `reserve` and `try_reserve`. . #[cold] . #[inline(never)] 226,286 ( 0.01%) fn reserve_rehash( . &mut self, . additional: usize, . hasher: impl Fn(&T) -> u64, . fallibility: Fallibility, . ) -> Result<(), TryReserveError> { . unsafe { 84 ( 0.00%) self.table.reserve_rehash_inner( . additional, 573 ( 0.00%) &|table, index| hasher(table.bucket::(index).as_ref()), . fallibility, 35 ( 0.00%) TableLayout::new::(), 42 ( 0.00%) if mem::needs_drop::() { 35 ( 0.00%) Some(mem::transmute(ptr::drop_in_place:: as unsafe fn(*mut T))) . } else { . None . }, . ) . } 156,269 ( 0.01%) } . . /// Allocates a new table of a different size and moves the contents of the . /// current table into it. . fn resize( . &mut self, . capacity: usize, . hasher: impl Fn(&T) -> u64, . fallibility: Fallibility, -- line 706 ---------------------------------------- -- line 714 ---------------------------------------- . ) . } . } . . /// Inserts a new element into the table, and returns its raw bucket. . /// . /// This does not check if the given element already exists in the table. . #[cfg_attr(feature = "inline-more", inline)] 1,402,800 ( 0.06%) pub fn insert(&mut self, hash: u64, value: T, hasher: impl Fn(&T) -> u64) -> Bucket { . unsafe { 184 ( 0.00%) let mut index = self.table.find_insert_slot(hash); . . // We can avoid growing the table once we have reached our load . // factor if we are replacing a tombstone. This works since the . // number of EMPTY slots does not change in this case. 2,606 ( 0.00%) let old_ctrl = *self.table.ctrl(index); 961,297 ( 0.04%) if unlikely(self.table.growth_left == 0 && special_is_empty(old_ctrl)) { 35 ( 0.00%) self.reserve(1, hasher); 35 ( 0.00%) index = self.table.find_insert_slot(hash); . } . 115 ( 0.00%) self.table.record_item_insert_at(index, old_ctrl, hash); . 138 ( 0.00%) let bucket = self.bucket(index); 252 ( 0.00%) bucket.write(value); 46 ( 0.00%) bucket . } 1,034,013 ( 0.04%) } . . /// Attempts to insert a new element without growing the table and return its raw bucket. . /// . /// Returns an `Err` containing the given element if inserting it would require growing the . /// table. . /// . /// This does not check if the given element already exists in the table. . #[cfg(feature = "raw")] -- line 749 ---------------------------------------- -- line 760 ---------------------------------------- . } . } . } . . /// Inserts a new element into the table, and returns a mutable reference to it. . /// . /// This does not check if the given element already exists in the table. . #[cfg_attr(feature = "inline-more", inline)] 349,592 ( 0.01%) pub fn insert_entry(&mut self, hash: u64, value: T, hasher: impl Fn(&T) -> u64) -> &mut T { 155 ( 0.00%) unsafe { self.insert(hash, value, hasher).as_mut() } 262,194 ( 0.01%) } . . /// Inserts a new element into the table, without growing the table. . /// . /// There must be enough space in the table to insert the new element. . /// . /// This does not check if the given element already exists in the table. . #[cfg_attr(feature = "inline-more", inline)] . #[cfg(any(feature = "raw", feature = "rustc-internal-api"))] 829 ( 0.00%) pub unsafe fn insert_no_grow(&mut self, hash: u64, value: T) -> Bucket { 272,825 ( 0.01%) let (index, old_ctrl) = self.table.prepare_insert_slot(hash); 3,234 ( 0.00%) let bucket = self.table.bucket(index); . . // If we are replacing a DELETED entry then we don't need to update . // the load counter. 550,734 ( 0.02%) self.table.growth_left -= special_is_empty(old_ctrl) as usize; . . bucket.write(value); 435,064 ( 0.02%) self.table.items += 1; . bucket 1,581 ( 0.00%) } . . /// Temporary removes a bucket, applying the given function to the removed . /// element and optionally put back the returned value in the same bucket. . /// . /// Returns `true` if the bucket still contains an element . /// . /// This does not check if the given bucket is actually occupied. . #[cfg_attr(feature = "inline-more", inline)] -- line 798 ---------------------------------------- -- line 813 ---------------------------------------- . true . } else { . false . } . } . . /// Searches for an element in the table. . #[inline] 31,642 ( 0.00%) pub fn find(&self, hash: u64, mut eq: impl FnMut(&T) -> bool) -> Option> { 3,581 ( 0.00%) let result = self.table.find_inner(hash, &mut |index| unsafe { 15,740 ( 0.00%) eq(self.bucket(index).as_ref()) 918 ( 0.00%) }); . . // Avoid `Option::map` because it bloats LLVM IR. 69 ( 0.00%) match result { 383 ( 0.00%) Some(index) => Some(unsafe { self.bucket(index) }), 46 ( 0.00%) None => None, . } 35,978 ( 0.00%) } . . /// Gets a reference to an element in the table. . #[inline] . pub fn get(&self, hash: u64, eq: impl FnMut(&T) -> bool) -> Option<&T> { . // Avoid `Option::map` because it bloats LLVM IR. 11,444 ( 0.00%) match self.find(hash, eq) { . Some(bucket) => Some(unsafe { bucket.as_ref() }), . None => None, . } . } . . /// Gets a mutable reference to an element in the table. . #[inline] 3,261 ( 0.00%) pub fn get_mut(&mut self, hash: u64, eq: impl FnMut(&T) -> bool) -> Option<&mut T> { . // Avoid `Option::map` because it bloats LLVM IR. 16,221 ( 0.00%) match self.find(hash, eq) { . Some(bucket) => Some(unsafe { bucket.as_mut() }), 46 ( 0.00%) None => None, . } 3,725 ( 0.00%) } . . /// Attempts to get mutable references to `N` entries in the table at once. . /// . /// Returns an array of length `N` with the results of each query. . /// . /// At most one mutable reference will be returned to any entry. `None` will be returned if any . /// of the hashes are duplicates. `None` will be returned if the hash is not found. . /// -- line 859 ---------------------------------------- -- line 913 ---------------------------------------- . /// more, but is guaranteed to be able to hold at least this many. . #[inline] . pub fn capacity(&self) -> usize { . self.table.items + self.table.growth_left . } . . /// Returns the number of elements in the table. . #[inline] 8 ( 0.00%) pub fn len(&self) -> usize { 4 ( 0.00%) self.table.items 8 ( 0.00%) } . . /// Returns `true` if the table contains no elements. . #[inline] 8 ( 0.00%) pub fn is_empty(&self) -> bool { 354,914 ( 0.01%) self.len() == 0 16 ( 0.00%) } . . /// Returns the number of buckets in the table. . #[inline] . pub fn buckets(&self) -> usize { . self.table.bucket_mask + 1 . } . . /// Returns an iterator over every element in the table. It is up to . /// the caller to ensure that the `RawTable` outlives the `RawIter`. . /// Because we cannot make the `next` method unsafe on the `RawIter` . /// struct, we have to make the `iter` method unsafe. . #[inline] 231 ( 0.00%) pub unsafe fn iter(&self) -> RawIter { 231 ( 0.00%) let data = Bucket::from_base_index(self.data_end(), 0); 297 ( 0.00%) RawIter { 231 ( 0.00%) iter: RawIterRange::new(self.table.ctrl.as_ptr(), data, self.table.buckets()), 40,252 ( 0.00%) items: self.table.items, . } 66 ( 0.00%) } . . /// Returns an iterator over occupied buckets that could match a given hash. . /// . /// `RawTable` only stores 7 bits of the hash value, so this iterator may . /// return items that have a hash value different than the one provided. You . /// should always validate the returned values before using them. . /// . /// It is up to the caller to ensure that the `RawTable` outlives the -- line 956 ---------------------------------------- -- line 995 ---------------------------------------- . /// Iteration starts at the provided iterator's current location. . /// . /// It is up to the caller to ensure that the iterator is valid for this . /// `RawTable` and covers all items that remain in the table. . pub unsafe fn into_iter_from(self, iter: RawIter) -> RawIntoIter { . debug_assert_eq!(iter.len(), self.len()); . . let alloc = self.table.alloc.clone(); 4,040 ( 0.00%) let allocation = self.into_allocation(); 3,030 ( 0.00%) RawIntoIter { 5,050 ( 0.00%) iter, . allocation, . marker: PhantomData, . alloc, . } . } . . /// Converts the table into a raw allocation. The contents of the table . /// should be dropped using a `RawIter` before freeing the allocation. . #[cfg_attr(feature = "inline-more", inline)] . pub(crate) fn into_allocation(self) -> Option<(NonNull, Layout)> { 1,800 ( 0.00%) let alloc = if self.table.is_empty_singleton() { . None . } else { . // Avoid `Option::unwrap_or_else` because it bloats LLVM IR. . let (layout, ctrl_offset) = match calculate_layout::(self.table.buckets()) { . Some(lco) => lco, . None => unsafe { hint::unreachable_unchecked() }, . }; . Some(( 320 ( 0.00%) unsafe { NonNull::new_unchecked(self.table.ctrl.as_ptr().sub(ctrl_offset)) }, . layout, . )) . }; . mem::forget(self); . alloc . } . } . -- line 1033 ---------------------------------------- -- line 1041 ---------------------------------------- . where . T: Sync, . A: Sync, . { . } . . impl RawTableInner { . #[inline] 93 ( 0.00%) const fn new_in(alloc: A) -> Self { 317,018 ( 0.01%) Self { . // Be careful to cast the entire slice to a raw pointer. 217 ( 0.00%) ctrl: unsafe { NonNull::new_unchecked(Group::static_empty() as *const _ as *mut u8) }, . bucket_mask: 0, . items: 0, . growth_left: 0, . alloc, . } 62 ( 0.00%) } . } . . impl RawTableInner { . #[cfg_attr(feature = "inline-more", inline)] 186,176 ( 0.01%) unsafe fn new_uninitialized( . alloc: A, . table_layout: TableLayout, . buckets: usize, . fallibility: Fallibility, . ) -> Result { . debug_assert!(buckets.is_power_of_two()); . . // Avoid `Option::ok_or_else` because it bloats LLVM IR. 112 ( 0.00%) let (layout, ctrl_offset) = match table_layout.calculate_layout_for(buckets) { 84 ( 0.00%) Some(lco) => lco, . None => return Err(fallibility.capacity_overflow()), . }; . . // We need an additional check to ensure that the allocation doesn't . // exceed `isize::MAX`. We can skip this check on 64-bit systems since . // such allocations will never succeed anyways. . // . // This mirrors what Vec does in the standard library. 28 ( 0.00%) if mem::size_of::() < 8 && layout.size() > isize::MAX as usize { . return Err(fallibility.capacity_overflow()); . } . 48,443 ( 0.00%) let ptr: NonNull = match do_alloc(&alloc, layout) { 49 ( 0.00%) Ok(block) => block.cast(), . Err(_) => return Err(fallibility.alloc_err(layout)), . }; . 63 ( 0.00%) let ctrl = NonNull::new_unchecked(ptr.as_ptr().add(ctrl_offset)); 104,641 ( 0.00%) Ok(Self { . ctrl, 46,055 ( 0.00%) bucket_mask: buckets - 1, . items: 0, 63 ( 0.00%) growth_left: bucket_mask_to_capacity(buckets - 1), . alloc, . }) 135,692 ( 0.01%) } . . #[inline] 20,444 ( 0.00%) fn fallible_with_capacity( . alloc: A, . table_layout: TableLayout, . capacity: usize, . fallibility: Fallibility, . ) -> Result { 5,250 ( 0.00%) if capacity == 0 { 5,119 ( 0.00%) Ok(Self::new_in(alloc)) . } else { . unsafe { . let buckets = 147 ( 0.00%) capacity_to_buckets(capacity).ok_or_else(|| fallibility.capacity_overflow())?; . 84,911 ( 0.00%) let result = Self::new_uninitialized(alloc, table_layout, buckets, fallibility)?; 84 ( 0.00%) result.ctrl(0).write_bytes(EMPTY, result.num_ctrl_bytes()); . 9,432 ( 0.00%) Ok(result) . } . } 20,381 ( 0.00%) } . . /// Searches for an empty or deleted bucket which is suitable for inserting . /// a new element and sets the hash for that slot. . /// . /// There must be at least 1 empty bucket in the table. . #[inline] 91,182 ( 0.00%) unsafe fn prepare_insert_slot(&self, hash: u64) -> (usize, u8) { 72 ( 0.00%) let index = self.find_insert_slot(hash); 91,182 ( 0.00%) let old_ctrl = *self.ctrl(index); 24 ( 0.00%) self.set_ctrl_h2(index, hash); 48 ( 0.00%) (index, old_ctrl) 182,220 ( 0.01%) } . . /// Searches for an empty or deleted bucket which is suitable for inserting . /// a new element. . /// . /// There must be at least 1 empty bucket in the table. . #[inline] 216 ( 0.00%) fn find_insert_slot(&self, hash: u64) -> usize { 216 ( 0.00%) let mut probe_seq = self.probe_seq(hash); 54 ( 0.00%) loop { . unsafe { 432 ( 0.00%) let group = Group::load(self.ctrl(probe_seq.pos)); 681,151 ( 0.03%) if let Some(bit) = group.match_empty_or_deleted().lowest_set_bit() { 1,984,002 ( 0.08%) let result = (probe_seq.pos + bit) & self.bucket_mask; . . // In tables smaller than the group width, trailing control . // bytes outside the range of the table are filled with . // EMPTY entries. These will unfortunately trigger a . // match, but once masked may point to a full bucket that . // is already occupied. We detect this situation here and . // perform a second scan starting at the beginning of the . // table. This second scan is guaranteed to find an empty . // slot (due to the load factor) before hitting the trailing . // control bytes (containing EMPTY). 901,916 ( 0.04%) if unlikely(is_full(*self.ctrl(result))) { . debug_assert!(self.bucket_mask < Group::WIDTH); . debug_assert_ne!(probe_seq.pos, 0); . return Group::load_aligned(self.ctrl(0)) . .match_empty_or_deleted() . .lowest_set_bit_nonzero(); . } . 108 ( 0.00%) return result; . } . } . probe_seq.move_next(self.bucket_mask); . } 162 ( 0.00%) } . . /// Searches for an element in the table. This uses dynamic dispatch to reduce the amount of . /// code generated, but it is eliminated by LLVM optimizations. . #[inline] 276 ( 0.00%) fn find_inner(&self, hash: u64, eq: &mut dyn FnMut(usize) -> bool) -> Option { 69 ( 0.00%) let h2_hash = h2(hash); 92 ( 0.00%) let mut probe_seq = self.probe_seq(hash); . 47,963 ( 0.00%) loop { 184 ( 0.00%) let group = unsafe { Group::load(self.ctrl(probe_seq.pos)) }; . 1,749,643 ( 0.07%) for bit in group.match_byte(h2_hash) { 3,334,812 ( 0.13%) let index = (probe_seq.pos + bit) & self.bucket_mask; . 2,453,020 ( 0.10%) if likely(eq(index)) { . return Some(index); . } . } . 579,197 ( 0.02%) if likely(group.match_empty().any_bit_set()) { 23 ( 0.00%) return None; . } . . probe_seq.move_next(self.bucket_mask); . } 92 ( 0.00%) } . . #[allow(clippy::mut_mut)] . #[inline] . unsafe fn prepare_rehash_in_place(&mut self) { . // Bulk convert all full control bytes to DELETED, and all DELETED . // control bytes to EMPTY. This effectively frees up all buckets . // containing a DELETED entry. . for i in (0..self.buckets()).step_by(Group::WIDTH) { -- line 1204 ---------------------------------------- -- line 1214 ---------------------------------------- . .copy_to(self.ctrl(Group::WIDTH), self.buckets()); . } else { . self.ctrl(0) . .copy_to(self.ctrl(self.buckets()), Group::WIDTH); . } . } . . #[inline] 96 ( 0.00%) unsafe fn bucket(&self, index: usize) -> Bucket { . debug_assert_ne!(self.bucket_mask, 0); . debug_assert!(index < self.buckets()); 96 ( 0.00%) Bucket::from_base_index(self.data_end(), index) 48 ( 0.00%) } . . #[inline] 288 ( 0.00%) unsafe fn bucket_ptr(&self, index: usize, size_of: usize) -> *mut u8 { . debug_assert_ne!(self.bucket_mask, 0); . debug_assert!(index < self.buckets()); 240 ( 0.00%) let base: *mut u8 = self.data_end().as_ptr(); 2,470,800 ( 0.10%) base.sub((index + 1) * size_of) 96 ( 0.00%) } . . #[inline] 144 ( 0.00%) unsafe fn data_end(&self) -> NonNull { 360 ( 0.00%) NonNull::new_unchecked(self.ctrl.as_ptr().cast()) 144 ( 0.00%) } . . /// Returns an iterator-like object for a probe sequence on the table. . /// . /// This iterator never terminates, but is guaranteed to visit each bucket . /// group exactly once. The loop using `probe_seq` must terminate upon . /// reaching a group containing an empty bucket. . #[inline] 539 ( 0.00%) fn probe_seq(&self, hash: u64) -> ProbeSeq { 154 ( 0.00%) ProbeSeq { 10,898,945 ( 0.43%) pos: h1(hash) & self.bucket_mask, . stride: 0, . } 308 ( 0.00%) } . . /// Returns the index of a bucket for which a value must be inserted if there is enough rooom . /// in the table, otherwise returns error . #[cfg(feature = "raw")] . #[inline] . unsafe fn prepare_insert_no_grow(&mut self, hash: u64) -> Result { . let index = self.find_insert_slot(hash); . let old_ctrl = *self.ctrl(index); -- line 1260 ---------------------------------------- -- line 1262 ---------------------------------------- . Err(()) . } else { . self.record_item_insert_at(index, old_ctrl, hash); . Ok(index) . } . } . . #[inline] 207 ( 0.00%) unsafe fn record_item_insert_at(&mut self, index: usize, old_ctrl: u8, hash: u64) { 1,203,905 ( 0.05%) self.growth_left -= special_is_empty(old_ctrl) as usize; 46 ( 0.00%) self.set_ctrl_h2(index, hash); 961,017 ( 0.04%) self.items += 1; 46 ( 0.00%) } . . #[inline] . fn is_in_same_group(&self, i: usize, new_i: usize, hash: u64) -> bool { . let probe_seq_pos = self.probe_seq(hash).pos; . let probe_index = . |pos: usize| (pos.wrapping_sub(probe_seq_pos) & self.bucket_mask) / Group::WIDTH; . probe_index(i) == probe_index(new_i) . } . . /// Sets a control byte to the hash, and possibly also the replicated control byte at . /// the end of the array. . #[inline] 423 ( 0.00%) unsafe fn set_ctrl_h2(&self, index: usize, hash: u64) { 188 ( 0.00%) self.set_ctrl(index, h2(hash)); 94 ( 0.00%) } . . #[inline] . unsafe fn replace_ctrl_h2(&self, index: usize, hash: u64) -> u8 { . let prev_ctrl = *self.ctrl(index); . self.set_ctrl_h2(index, hash); . prev_ctrl . } . . /// Sets a control byte, and possibly also the replicated control byte at . /// the end of the array. . #[inline] 470 ( 0.00%) unsafe fn set_ctrl(&self, index: usize, ctrl: u8) { . // Replicate the first Group::WIDTH control bytes at the end of . // the array without using a branch: . // - If index >= Group::WIDTH then index == index2. . // - Otherwise index2 == self.bucket_mask + 1 + index. . // . // The very last replicated control byte is never actually read because . // we mask the initial index for unaligned loads, but we write it . // anyways because it makes the set_ctrl implementation simpler. -- line 1309 ---------------------------------------- -- line 1312 ---------------------------------------- . // replicate the buckets at the end of the trailing group. For example . // with 2 buckets and a group size of 4, the control bytes will look . // like this: . // . // Real | Replicated . // --------------------------------------------- . // | [A] | [B] | [EMPTY] | [EMPTY] | [A] | [B] | . // --------------------------------------------- 1,924,733 ( 0.08%) let index2 = ((index.wrapping_sub(Group::WIDTH)) & self.bucket_mask) + Group::WIDTH; . 641,000 ( 0.03%) *self.ctrl(index) = ctrl; 641,011 ( 0.03%) *self.ctrl(index2) = ctrl; 94 ( 0.00%) } . . /// Returns a pointer to a control byte. . #[inline] 1,244 ( 0.00%) unsafe fn ctrl(&self, index: usize) -> *mut u8 { . debug_assert!(index < self.num_ctrl_bytes()); 933 ( 0.00%) self.ctrl.as_ptr().add(index) 622 ( 0.00%) } . . #[inline] 92 ( 0.00%) fn buckets(&self) -> usize { 133,133 ( 0.01%) self.bucket_mask + 1 92 ( 0.00%) } . . #[inline] 14 ( 0.00%) fn num_ctrl_bytes(&self) -> usize { 102,851 ( 0.00%) self.bucket_mask + 1 + Group::WIDTH 14 ( 0.00%) } . . #[inline] 74 ( 0.00%) fn is_empty_singleton(&self) -> bool { 595,664 ( 0.02%) self.bucket_mask == 0 148 ( 0.00%) } . . #[allow(clippy::mut_mut)] . #[inline] 119 ( 0.00%) unsafe fn prepare_resize( . &self, . table_layout: TableLayout, . capacity: usize, . fallibility: Fallibility, . ) -> Result, TryReserveError> { 7 ( 0.00%) debug_assert!(self.items <= capacity); . . // Allocate and initialize the new table. 2,230 ( 0.00%) let mut new_table = RawTableInner::fallible_with_capacity( 7 ( 0.00%) self.alloc.clone(), . table_layout, . capacity, . fallibility, . )?; 59,922 ( 0.00%) new_table.growth_left -= self.items; 14 ( 0.00%) new_table.items = self.items; . . // The hash function may panic, in which case we simply free the new . // table without dropping any elements that may have been copied into . // it. . // . // This guard is also used to free the old table on success, see . // the comment at the bottom of this function. 196 ( 0.00%) Ok(guard(new_table, move |self_| { 27,736 ( 0.00%) if !self_.is_empty_singleton() { 9 ( 0.00%) self_.free_buckets(table_layout); . } 14 ( 0.00%) })) 28 ( 0.00%) } . . /// Reserves or rehashes to make room for `additional` more elements. . /// . /// This uses dynamic dispatch to reduce the amount of . /// code generated, but it is eliminated by LLVM optimizations when inlined. . #[allow(clippy::inline_always)] . #[inline(always)] . unsafe fn reserve_rehash_inner( -- line 1387 ---------------------------------------- -- line 1388 ---------------------------------------- . &mut self, . additional: usize, . hasher: &dyn Fn(&mut Self, usize) -> u64, . fallibility: Fallibility, . layout: TableLayout, . drop: Option, . ) -> Result<(), TryReserveError> { . // Avoid `Option::ok_or_else` because it bloats LLVM IR. 55,484 ( 0.00%) let new_items = match self.items.checked_add(additional) { 28 ( 0.00%) Some(new_items) => new_items, . None => return Err(fallibility.capacity_overflow()), . }; 55,431 ( 0.00%) let full_capacity = bucket_mask_to_capacity(self.bucket_mask); 115,788 ( 0.00%) if new_items <= full_capacity / 2 { . // Rehash in-place without re-allocating if we have plenty of spare . // capacity that is locked up due to DELETED entries. . self.rehash_in_place(hasher, layout.size, drop); . Ok(()) . } else { . // Otherwise, conservatively resize to at least the next size up . // to avoid churning deletes into frequent rehashes. . self.resize_inner( 27,768 ( 0.00%) usize::max(new_items, full_capacity + 1), . hasher, . fallibility, . layout, . ) . } 14 ( 0.00%) } . . /// Allocates a new table of a different size and moves the contents of the . /// current table into it. . /// . /// This uses dynamic dispatch to reduce the amount of . /// code generated, but it is eliminated by LLVM optimizations when inlined. . #[allow(clippy::inline_always)] . #[inline(always)] . unsafe fn resize_inner( . &mut self, . capacity: usize, . hasher: &dyn Fn(&mut Self, usize) -> u64, . fallibility: Fallibility, . layout: TableLayout, . ) -> Result<(), TryReserveError> { 5,276 ( 0.00%) let mut new_table = self.prepare_resize(layout, capacity, fallibility)?; . . // Copy all elements to the new table. 718 ( 0.00%) for i in 0..self.buckets() { 374,676 ( 0.01%) if !is_full(*self.ctrl(i)) { . continue; . } . . // This may panic. 216 ( 0.00%) let hash = hasher(self, i); . . // We can use a simpler version of insert() here since: . // - there are no DELETED entries. . // - we know there is enough space in the table. . // - all elements are unique. 288 ( 0.00%) let (index, _) = new_table.prepare_insert_slot(hash); . 72 ( 0.00%) ptr::copy_nonoverlapping( 96 ( 0.00%) self.bucket_ptr(i, layout.size), 216 ( 0.00%) new_table.bucket_ptr(index, layout.size), . layout.size, . ); . } . . // We successfully copied all elements without panicking. Now replace . // self with the new table. The old table will have its memory freed but . // the items will not be dropped (since they have been moved into the . // new table). 56 ( 0.00%) mem::swap(self, &mut new_table); . 27,705 ( 0.00%) Ok(()) 42 ( 0.00%) } . . /// Rehashes the contents of the table in place (i.e. without changing the . /// allocation). . /// . /// If `hasher` panics then some the table's contents may be lost. . /// . /// This uses dynamic dispatch to reduce the amount of . /// code generated, but it is eliminated by LLVM optimizations when inlined. -- line 1471 ---------------------------------------- -- line 1547 ---------------------------------------- . } . . guard.growth_left = bucket_mask_to_capacity(guard.bucket_mask) - guard.items; . . mem::forget(guard); . } . . #[inline] 42 ( 0.00%) unsafe fn free_buckets(&mut self, table_layout: TableLayout) { . // Avoid `Option::unwrap_or_else` because it bloats LLVM IR. 120 ( 0.00%) let (layout, ctrl_offset) = match table_layout.calculate_layout_for(self.buckets()) { 72 ( 0.00%) Some(lco) => lco, . None => hint::unreachable_unchecked(), . }; 18 ( 0.00%) self.alloc.deallocate( 18,593 ( 0.00%) NonNull::new_unchecked(self.ctrl.as_ptr().sub(ctrl_offset)), . layout, . ); 12 ( 0.00%) } . . /// Marks all table buckets as empty without dropping their contents. . #[inline] . fn clear_no_drop(&mut self) { 3,735 ( 0.00%) if !self.is_empty_singleton() { . unsafe { . self.ctrl(0).write_bytes(EMPTY, self.num_ctrl_bytes()); . } . } 4,970 ( 0.00%) self.items = 0; 3,808 ( 0.00%) self.growth_left = bucket_mask_to_capacity(self.bucket_mask); . } . . #[inline] . unsafe fn erase(&mut self, index: usize) { . debug_assert!(is_full(*self.ctrl(index))); 75,770 ( 0.00%) let index_before = index.wrapping_sub(Group::WIDTH) & self.bucket_mask; . let empty_before = Group::load(self.ctrl(index_before)).match_empty(); . let empty_after = Group::load(self.ctrl(index)).match_empty(); . . // If we are inside a continuous block of Group::WIDTH full or deleted . // cells then a probe window may have seen a full block when trying to . // insert. We therefore need to keep that block non-empty so that . // lookups will continue searching to the next probe window. . // . // Note that in this context `leading_zeros` refers to the bytes at the . // end of a group, while `trailing_zeros` refers to the bytes at the . // beginning of a group. 303,080 ( 0.01%) let ctrl = if empty_before.leading_zeros() + empty_after.trailing_zeros() >= Group::WIDTH { . DELETED . } else { 376,730 ( 0.01%) self.growth_left += 1; . EMPTY . }; . self.set_ctrl(index, ctrl); 303,080 ( 0.01%) self.items -= 1; . } . } . . impl Clone for RawTable { 984 ( 0.00%) fn clone(&self) -> Self { 149 ( 0.00%) if self.table.is_empty_singleton() { . Self::new_in(self.table.alloc.clone()) . } else { . unsafe { . let mut new_table = ManuallyDrop::new( . // Avoid `Result::ok_or_else` because it bloats LLVM IR. . match Self::new_uninitialized( . self.table.alloc.clone(), . self.table.buckets(), -- line 1615 ---------------------------------------- -- line 1624 ---------------------------------------- . // We need to free the memory allocated for the new table. . new_table.free_buckets(); . }); . . // Return the newly created table. . ManuallyDrop::into_inner(new_table) . } . } 1,107 ( 0.00%) } . . fn clone_from(&mut self, source: &Self) { . if source.table.is_empty_singleton() { . *self = Self::new_in(self.table.alloc.clone()); . } else { . unsafe { . // First, drop all our elements without clearing the control bytes. . self.drop_elements(); -- line 1640 ---------------------------------------- -- line 1687 ---------------------------------------- . .table . .ctrl(0) . .copy_to_nonoverlapping(self.table.ctrl(0), self.table.num_ctrl_bytes()); . source . .data_start() . .copy_to_nonoverlapping(self.data_start(), self.table.buckets()); . . self.table.items = source.table.items; 44 ( 0.00%) self.table.growth_left = source.table.growth_left; . } . } . . impl RawTable { . /// Common code for clone and clone_from. Assumes `self.buckets() == source.buckets()`. . #[cfg_attr(feature = "inline-more", inline)] . unsafe fn clone_from_impl(&mut self, source: &Self, mut on_panic: impl FnMut(&mut Self)) { . // Copy the control bytes unchanged. We do this in a single pass -- line 1703 ---------------------------------------- -- line 1790 ---------------------------------------- . fn default() -> Self { . Self::new_in(Default::default()) . } . } . . #[cfg(feature = "nightly")] . unsafe impl<#[may_dangle] T, A: Allocator + Clone> Drop for RawTable { . #[cfg_attr(feature = "inline-more", inline)] 356,092 ( 0.01%) fn drop(&mut self) { 261,665 ( 0.01%) if !self.table.is_empty_singleton() { . unsafe { 3 ( 0.00%) self.drop_elements(); 3 ( 0.00%) self.free_buckets(); . } . } 384,290 ( 0.02%) } . } . #[cfg(not(feature = "nightly"))] . impl Drop for RawTable { . #[cfg_attr(feature = "inline-more", inline)] . fn drop(&mut self) { . if !self.table.is_empty_singleton() { . unsafe { . self.drop_elements(); -- line 1813 ---------------------------------------- -- line 1817 ---------------------------------------- . } . } . . impl IntoIterator for RawTable { . type Item = T; . type IntoIter = RawIntoIter; . . #[cfg_attr(feature = "inline-more", inline)] 4,040 ( 0.00%) fn into_iter(self) -> RawIntoIter { . unsafe { . let iter = self.iter(); . self.into_iter_from(iter) . } 5,050 ( 0.00%) } . } . . /// Iterator over a sub-range of a table. Unlike `RawIter` this iterator does . /// not track an item count. . pub(crate) struct RawIterRange { . // Mask of full buckets in the current group. Bits are cleared from this . // mask as each element is processed. . current_group: BitMask, -- line 1838 ---------------------------------------- -- line 1848 ---------------------------------------- . end: *const u8, . } . . impl RawIterRange { . /// Returns a `RawIterRange` covering a subset of a table. . /// . /// The control byte address must be aligned to the group size. . #[cfg_attr(feature = "inline-more", inline)] 396 ( 0.00%) unsafe fn new(ctrl: *const u8, data: Bucket, len: usize) -> Self { . debug_assert_ne!(len, 0); . debug_assert_eq!(ctrl as usize % Group::WIDTH, 0); 33 ( 0.00%) let end = ctrl.add(len); . . // Load the first group and advance ctrl to point to the next group 198 ( 0.00%) let current_group = Group::load_aligned(ctrl).match_full(); 33 ( 0.00%) let next_ctrl = ctrl.add(Group::WIDTH); . 132 ( 0.00%) Self { . current_group, . data, . next_ctrl, . end, . } 66 ( 0.00%) } . . /// Splits a `RawIterRange` into two halves. . /// . /// Returns `None` if the remaining range is smaller than or equal to the . /// group width. . #[cfg_attr(feature = "inline-more", inline)] . #[cfg(feature = "rayon")] . pub(crate) fn split(mut self) -> (Self, Option>) { -- line 1879 ---------------------------------------- -- line 1931 ---------------------------------------- . } . } . } . . impl Iterator for RawIterRange { . type Item = Bucket; . . #[cfg_attr(feature = "inline-more", inline)] 152 ( 0.00%) fn next(&mut self) -> Option> { . unsafe { . loop { 117,675 ( 0.00%) if let Some(index) = self.current_group.lowest_set_bit() { 15,484 ( 0.00%) self.current_group = self.current_group.remove_lowest_bit(); 25,523 ( 0.00%) return Some(self.data.next_n(index)); . } . 89,806 ( 0.00%) if self.next_ctrl >= self.end { 33 ( 0.00%) return None; . } . . // We might read past self.end up to the next group boundary, . // but this is fine because it only occurs on tables smaller . // than the group size where the trailing control bytes are all . // EMPTY. On larger tables self.end is guaranteed to be aligned . // to the group size (since tables are power-of-two sized). 2,450 ( 0.00%) self.current_group = Group::load_aligned(self.next_ctrl).match_full(); 1,877 ( 0.00%) self.data = self.data.next_n(Group::WIDTH); 5,062 ( 0.00%) self.next_ctrl = self.next_ctrl.add(Group::WIDTH); . } . } 114 ( 0.00%) } . . #[inline] . fn size_hint(&self) -> (usize, Option) { . // We don't have an item count, so just guess based on the range size. . ( . 0, . Some(unsafe { offset_from(self.end, self.next_ctrl) + Group::WIDTH }), . ) -- line 1969 ---------------------------------------- -- line 2102 ---------------------------------------- . } . } else { . // We must have already iterated past the removed item. . } . } . } . . unsafe fn drop_elements(&mut self) { 941 ( 0.00%) if mem::needs_drop::() && self.len() != 0 { . for item in self { . item.drop(); . } . } . } . } . . impl Clone for RawIter { -- line 2118 ---------------------------------------- -- line 2124 ---------------------------------------- . } . } . } . . impl Iterator for RawIter { . type Item = Bucket; . . #[cfg_attr(feature = "inline-more", inline)] 18,882 ( 0.00%) fn next(&mut self) -> Option> { 70,264 ( 0.00%) if let Some(b) = self.iter.next() { 154,602 ( 0.01%) self.items -= 1; 5 ( 0.00%) Some(b) . } else { . // We don't check against items == 0 here to allow the . // compiler to optimize away the item count entirely if the . // iterator length is never queried. . debug_assert_eq!(self.items, 0); 33 ( 0.00%) None . } 37,650 ( 0.00%) } . . #[inline] . fn size_hint(&self) -> (usize, Option) { . (self.items, Some(self.items)) . } . } . . impl ExactSizeIterator for RawIter {} -- line 2151 ---------------------------------------- -- line 2177 ---------------------------------------- . T: Sync, . A: Sync, . { . } . . #[cfg(feature = "nightly")] . unsafe impl<#[may_dangle] T, A: Allocator + Clone> Drop for RawIntoIter { . #[cfg_attr(feature = "inline-more", inline)] 1,958 ( 0.00%) fn drop(&mut self) { . unsafe { . // Drop all remaining elements . self.iter.drop_elements(); . . // Free the table 8,389 ( 0.00%) if let Some((ptr, layout)) = self.allocation { . self.alloc.deallocate(ptr, layout); . } . } 406 ( 0.00%) } . } . #[cfg(not(feature = "nightly"))] . impl Drop for RawIntoIter { . #[cfg_attr(feature = "inline-more", inline)] . fn drop(&mut self) { . unsafe { . // Drop all remaining elements . self.iter.drop_elements(); -- line 2203 ---------------------------------------- -- line 2209 ---------------------------------------- . } . } . } . . impl Iterator for RawIntoIter { . type Item = T; . . #[cfg_attr(feature = "inline-more", inline)] 1,131 ( 0.00%) fn next(&mut self) -> Option { 108 ( 0.00%) unsafe { Some(self.iter.next()?.read()) } 3,149 ( 0.00%) } . . #[inline] . fn size_hint(&self) -> (usize, Option) { 4 ( 0.00%) self.iter.size_hint() . } . } . . impl ExactSizeIterator for RawIntoIter {} . impl FusedIterator for RawIntoIter {} . . /// Iterator which consumes elements without freeing the table storage. . pub struct RawDrain<'a, T, A: Allocator + Clone = Global> { -- line 2231 ---------------------------------------- -- line 2259 ---------------------------------------- . where . T: Sync, . A: Sync, . { . } . . impl Drop for RawDrain<'_, T, A> { . #[cfg_attr(feature = "inline-more", inline)] 304 ( 0.00%) fn drop(&mut self) { . unsafe { . // Drop all remaining elements. Note that this may panic. . self.iter.drop_elements(); . . // Reset the contents of the table now that all elements have been . // dropped. . self.table.clear_no_drop(); . . // Move the now empty table back to its original location. 38 ( 0.00%) self.orig_table . .as_ptr() . .copy_from_nonoverlapping(&*self.table, 1); . } 304 ( 0.00%) } . } . . impl Iterator for RawDrain<'_, T, A> { . type Item = T; . . #[cfg_attr(feature = "inline-more", inline)] . fn next(&mut self) -> Option { . unsafe { -- line 2289 ---------------------------------------- 6,067,621 ( 0.24%) -------------------------------------------------------------------------------- -- Auto-annotated source: /usr/home/liquid/rust/worktree-benchmarking/library/std/src/sys/unix/alloc.rs -------------------------------------------------------------------------------- Ir -- line 5 ---------------------------------------- . #[stable(feature = "alloc_system_type", since = "1.28.0")] . unsafe impl GlobalAlloc for System { . #[inline] . unsafe fn alloc(&self, layout: Layout) -> *mut u8 { . // jemalloc provides alignment less than MIN_ALIGN for small allocations. . // So only rely on MIN_ALIGN if size >= align. . // Also see and . // . 1,783,798 ( 0.07%) if layout.align() <= MIN_ALIGN && layout.align() <= layout.size() { 1,337,847 ( 0.05%) libc::malloc(layout.size()) as *mut u8 . } else { . #[cfg(target_os = "macos")] . { . if layout.align() > (1 << 31) { . return ptr::null_mut(); . } . } . aligned_malloc(&layout) . } . } . . #[inline] . unsafe fn alloc_zeroed(&self, layout: Layout) -> *mut u8 { . // See the comment above in `alloc` for why this check looks the way it does. 91,688 ( 0.00%) if layout.align() <= MIN_ALIGN && layout.align() <= layout.size() { 137,532 ( 0.01%) libc::calloc(layout.size(), 1) as *mut u8 . } else { . let ptr = self.alloc(layout); . if !ptr.is_null() { . ptr::write_bytes(ptr, 0, layout.size()); . } . ptr . } . } . . #[inline] . unsafe fn dealloc(&self, ptr: *mut u8, _layout: Layout) { 468,840 ( 0.02%) libc::free(ptr as *mut libc::c_void) . } . . #[inline] . unsafe fn realloc(&self, ptr: *mut u8, layout: Layout, new_size: usize) -> *mut u8 { 165,664 ( 0.01%) if layout.align() <= MIN_ALIGN && layout.align() <= new_size { 331,328 ( 0.01%) libc::realloc(ptr as *mut libc::c_void, new_size) as *mut u8 . } else { . realloc_fallback(self, ptr, layout, new_size) . } . } . } . . cfg_if::cfg_if! { . if #[cfg(any( -- line 56 ---------------------------------------- -- line 84 ---------------------------------------- . } else if #[cfg(target_os = "wasi")] { . #[inline] . unsafe fn aligned_malloc(layout: &Layout) -> *mut u8 { . libc::aligned_alloc(layout.align(), layout.size()) as *mut u8 . } . } else { . #[inline] . unsafe fn aligned_malloc(layout: &Layout) -> *mut u8 { 1 ( 0.00%) let mut out = ptr::null_mut(); . // posix_memalign requires that the alignment be a multiple of `sizeof(void*)`. . // Since these are all powers of 2, we can just use max. . let align = layout.align().max(crate::mem::size_of::()); 2 ( 0.00%) let ret = libc::posix_memalign(&mut out, align, layout.size()); 2 ( 0.00%) if ret != 0 { ptr::null_mut() } else { out as *mut u8 } . } . } . } 1 ( 0.00%) -------------------------------------------------------------------------------- -- Auto-annotated source: /usr/home/liquid/rust/worktree-benchmarking/compiler/rustc_metadata/src/rmeta/decoder.rs -------------------------------------------------------------------------------- Ir -- line 176 ---------------------------------------- . } . fn sess(self) -> Option<&'tcx Session> { . None . } . fn tcx(self) -> Option> { . None . } . 14,982 ( 0.00%) fn decoder(self, pos: usize) -> DecodeContext<'a, 'tcx> { 4,994 ( 0.00%) let tcx = self.tcx(); 551,420 ( 0.02%) DecodeContext { 59,865 ( 0.00%) opaque: opaque::Decoder::new(self.blob(), pos), 4,994 ( 0.00%) cdata: self.cdata(), . blob: self.blob(), 23,264 ( 0.00%) sess: self.sess().or(tcx.map(|tcx| tcx.sess)), . tcx, . last_source_file_index: 0, . lazy_state: LazyState::NoNode, . alloc_decoding_session: self . .cdata() 52,639 ( 0.00%) .map(|cdata| cdata.cdata.alloc_decoding_state.new_decoding_session()), . } 14,982 ( 0.00%) } . } . . impl<'a, 'tcx> Metadata<'a, 'tcx> for &'a MetadataBlob { . #[inline] . fn blob(self) -> &'a MetadataBlob { . self . } . } -- line 206 ---------------------------------------- -- line 255 ---------------------------------------- . } . #[inline] . fn tcx(self) -> Option> { . Some(self.1) . } . } . . impl<'a, 'tcx, T: Decodable>> Lazy { 79,358 ( 0.00%) fn decode>(self, metadata: M) -> T { 54,722 ( 0.00%) let mut dcx = metadata.decoder(self.position.get()); 161,929 ( 0.01%) dcx.lazy_state = LazyState::NodeStart(self.position); 50,947 ( 0.00%) T::decode(&mut dcx) 101,477 ( 0.00%) } . } . . impl<'a: 'x, 'tcx: 'x, 'x, T: Decodable>> Lazy<[T]> { . fn decode>( . self, . metadata: M, . ) -> impl ExactSizeIterator + Captures<'a> + Captures<'tcx> + 'x { 336 ( 0.00%) let mut dcx = metadata.decoder(self.position.get()); 96 ( 0.00%) dcx.lazy_state = LazyState::NodeStart(self.position); 60,589 ( 0.00%) (0..self.meta).map(move |_| T::decode(&mut dcx)) . } . } . . impl<'a, 'tcx> DecodeContext<'a, 'tcx> { . #[inline] . fn tcx(&self) -> TyCtxt<'tcx> { . debug_assert!(self.tcx.is_some(), "missing TyCtxt in DecodeContext"); . self.tcx.unwrap() -- line 285 ---------------------------------------- -- line 292 ---------------------------------------- . . #[inline] . pub fn cdata(&self) -> CrateMetadataRef<'a> { . debug_assert!(self.cdata.is_some(), "missing CrateMetadata in DecodeContext"); . self.cdata.unwrap() . } . . fn map_encoded_cnum_to_current(&self, cnum: CrateNum) -> CrateNum { 151,547 ( 0.01%) if cnum == LOCAL_CRATE { self.cdata().cnum } else { self.cdata().cnum_map[cnum] } . } . 59,166 ( 0.00%) fn read_lazy_with_meta(&mut self, meta: T::Meta) -> Lazy { . let distance = self.read_usize(); 55,710 ( 0.00%) let position = match self.lazy_state { . LazyState::NoNode => bug!("read_lazy_with_meta: outside of a metadata node"), 11,940 ( 0.00%) LazyState::NodeStart(start) => { . let start = start.get(); 23,880 ( 0.00%) assert!(distance <= start); . start - distance . } 11,934 ( 0.00%) LazyState::Previous(last_pos) => last_pos.get() + distance, . }; 31,836 ( 0.00%) self.lazy_state = LazyState::Previous(NonZeroUsize::new(position).unwrap()); . Lazy::from_position_and_meta(NonZeroUsize::new(position).unwrap(), meta) 59,166 ( 0.00%) } . . #[inline] . pub fn read_raw_bytes(&mut self, len: usize) -> &'a [u8] { . self.opaque.read_raw_bytes(len) . } . } . . impl<'a, 'tcx> TyDecoder<'tcx> for DecodeContext<'a, 'tcx> { -- line 324 ---------------------------------------- -- line 326 ---------------------------------------- . . #[inline] . fn tcx(&self) -> TyCtxt<'tcx> { . self.tcx.expect("missing TyCtxt in DecodeContext") . } . . #[inline] . fn peek_byte(&self) -> u8 { 168,180 ( 0.01%) self.opaque.data[self.opaque.position()] . } . . #[inline] . fn position(&self) -> usize { 1 ( 0.00%) self.opaque.position() . } . 150,066 ( 0.01%) fn cached_ty_for_shorthand(&mut self, shorthand: usize, or_insert_with: F) -> Ty<'tcx> . where . F: FnOnce(&mut Self) -> Ty<'tcx>, . { 21,438 ( 0.00%) let tcx = self.tcx(); . 21,438 ( 0.00%) let key = ty::CReaderCacheKey { cnum: Some(self.cdata().cnum), pos: shorthand }; . 68,084 ( 0.00%) if let Some(&ty) = tcx.ty_rcache.borrow().get(&key) { . return ty; . } . 4,417 ( 0.00%) let ty = or_insert_with(self); . tcx.ty_rcache.borrow_mut().insert(key, ty); . ty 192,942 ( 0.01%) } . 4 ( 0.00%) fn with_position(&mut self, pos: usize, f: F) -> R . where . F: FnOnce(&mut Self) -> R, . { . let new_opaque = opaque::Decoder::new(self.opaque.data, pos); . let old_opaque = mem::replace(&mut self.opaque, new_opaque); . let old_state = mem::replace(&mut self.lazy_state, LazyState::NoNode); 2 ( 0.00%) let r = f(self); 24,830 ( 0.00%) self.opaque = old_opaque; 4,966 ( 0.00%) self.lazy_state = old_state; . r 5 ( 0.00%) } . . fn decode_alloc_id(&mut self) -> rustc_middle::mir::interpret::AllocId { 7 ( 0.00%) if let Some(alloc_decoding_session) = self.alloc_decoding_session { 2 ( 0.00%) alloc_decoding_session.decode_alloc_id(self) . } else { . bug!("Attempting to decode interpret::AllocId without CrateMetadata") . } . } . } . . impl<'a, 'tcx> Decodable> for CrateNum { . fn decode(d: &mut DecodeContext<'a, 'tcx>) -> CrateNum { 80,199 ( 0.00%) let cnum = CrateNum::from_u32(d.read_u32()); . d.map_encoded_cnum_to_current(cnum) . } . } . . impl<'a, 'tcx> Decodable> for DefIndex { 363 ( 0.00%) fn decode(d: &mut DecodeContext<'a, 'tcx>) -> DefIndex { 155,415 ( 0.01%) DefIndex::from_u32(d.read_u32()) 726 ( 0.00%) } . } . . impl<'a, 'tcx> Decodable> for ExpnIndex { . fn decode(d: &mut DecodeContext<'a, 'tcx>) -> ExpnIndex { . ExpnIndex::from_u32(d.read_u32()) . } . } . . impl<'a, 'tcx> Decodable> for SyntaxContext { 206,404 ( 0.01%) fn decode(decoder: &mut DecodeContext<'a, 'tcx>) -> SyntaxContext { 51,601 ( 0.00%) let cdata = decoder.cdata(); 51,601 ( 0.00%) let sess = decoder.sess.unwrap(); 103,202 ( 0.00%) let cname = cdata.root.name; . rustc_span::hygiene::decode_syntax_context(decoder, &cdata.hygiene_context, |_, id| { . debug!("SpecializedDecoder: decoding {}", id); 1,168 ( 0.00%) cdata . .root . .syntax_contexts . .get(cdata, id) . .unwrap_or_else(|| panic!("Missing SyntaxContext {:?} for crate {:?}", id, cname)) . .decode((cdata, sess)) . }) 258,005 ( 0.01%) } . } . . impl<'a, 'tcx> Decodable> for ExpnId { 7,616 ( 0.00%) fn decode(decoder: &mut DecodeContext<'a, 'tcx>) -> ExpnId { 1,088 ( 0.00%) let local_cdata = decoder.cdata(); 1,088 ( 0.00%) let sess = decoder.sess.unwrap(); . . let cnum = CrateNum::decode(decoder); . let index = u32::decode(decoder); . . let expn_id = rustc_span::hygiene::decode_expn_id(cnum, index, |expn_id| { . let ExpnId { krate: cnum, local_id: index } = expn_id; . // Lookup local `ExpnData`s in our own crate data. Foreign `ExpnData`s . // are stored in the owning crate, to avoid duplication. . debug_assert_ne!(cnum, LOCAL_CRATE); 314 ( 0.00%) let crate_data = if cnum == local_cdata.cnum { . local_cdata . } else { . local_cdata.cstore.get_crate_data(cnum) . }; 1,570 ( 0.00%) let expn_data = crate_data . .root . .expn_data . .get(crate_data, index) . .unwrap() . .decode((crate_data, sess)); 1,570 ( 0.00%) let expn_hash = crate_data . .root . .expn_hashes . .get(crate_data, index) . .unwrap() . .decode((crate_data, sess)); 3,140 ( 0.00%) (expn_data, expn_hash) . }); . expn_id 8,704 ( 0.00%) } . } . . impl<'a, 'tcx> Decodable> for Span { 405,800 ( 0.02%) fn decode(decoder: &mut DecodeContext<'a, 'tcx>) -> Span { 50,725 ( 0.00%) let ctxt = SyntaxContext::decode(decoder); . let tag = u8::decode(decoder); . 101,450 ( 0.00%) if tag == TAG_PARTIAL_SPAN { . return DUMMY_SP.with_ctxt(ctxt); . } . . debug_assert!(tag == TAG_VALID_SPAN_LOCAL || tag == TAG_VALID_SPAN_FOREIGN); . . let lo = BytePos::decode(decoder); . let len = BytePos::decode(decoder); . let hi = lo + len; . 151,773 ( 0.01%) let Some(sess) = decoder.sess else { . bug!("Cannot decode Span without Session.") . }; . . // There are two possibilities here: . // 1. This is a 'local span', which is located inside a `SourceFile` . // that came from this crate. In this case, we use the source map data . // encoded in this crate. This branch should be taken nearly all of the time. . // 2. This is a 'foreign span', which is located inside a `SourceFile` -- line 476 ---------------------------------------- -- line 492 ---------------------------------------- . // a transitive dependency. . // . // When we encode a foreign span, we adjust its 'lo' and 'high' values . // to be based on the *foreign* crate (e.g. crate C), not the crate . // we are writing metadata for (e.g. crate B). This allows us to . // treat the 'local' and 'foreign' cases almost identically during deserialization: . // we can call `imported_source_files` for the proper crate, and binary search . // through the returned slice using our span. 107,687 ( 0.00%) let imported_source_files = if tag == TAG_VALID_SPAN_LOCAL { . decoder.cdata().imported_source_files(sess) . } else { . // When we encode a proc-macro crate, all `Span`s should be encoded . // with `TAG_VALID_SPAN_LOCAL` 241 ( 0.00%) if decoder.cdata().root.is_proc_macro_crate() { . // Decode `CrateNum` as u32 - using `CrateNum::decode` will ICE . // since we don't have `cnum_map` populated. . let cnum = u32::decode(decoder); . panic!( . "Decoding of crate {:?} tried to access proc-macro dep {:?}", . decoder.cdata().root.name, . cnum . ); -- line 513 ---------------------------------------- -- line 518 ---------------------------------------- . "SpecializedDecoder::specialized_decode: loading source files from cnum {:?}", . cnum . ); . . // Decoding 'foreign' spans should be rare enough that it's . // not worth it to maintain a per-CrateNum cache for `last_source_file_index`. . // We just set it to 0, to ensure that we don't try to access something out . // of bounds for our initial 'guess' 482 ( 0.00%) decoder.last_source_file_index = 0; . 241 ( 0.00%) let foreign_data = decoder.cdata().cstore.get_crate_data(cnum); . foreign_data.imported_source_files(sess) . }; . . let source_file = { . // Optimize for the case that most spans within a translated item . // originate from the same source_file. 151,773 ( 0.01%) let last_source_file = &imported_source_files[decoder.last_source_file_index]; . 202,148 ( 0.01%) if lo >= last_source_file.original_start_pos && lo <= last_source_file.original_end_pos . { . last_source_file . } else { . let index = imported_source_files . .binary_search_by_key(&lo, |source_file| source_file.original_start_pos) 12,998 ( 0.00%) .unwrap_or_else(|index| index - 1); . . // Don't try to cache the index for foreign spans, . // as this would require a map from CrateNums to indices 6,505 ( 0.00%) if tag == TAG_VALID_SPAN_LOCAL { 6,337 ( 0.00%) decoder.last_source_file_index = index; . } 13,010 ( 0.00%) &imported_source_files[index] . } . }; . . // Make sure our binary search above is correct. . debug_assert!( . lo >= source_file.original_start_pos && lo <= source_file.original_end_pos, . "Bad binary search: lo={:?} source_file.original_start_pos={:?} source_file.original_end_pos={:?}", . lo, -- line 558 ---------------------------------------- -- line 565 ---------------------------------------- . hi >= source_file.original_start_pos && hi <= source_file.original_end_pos, . "Bad binary search: hi={:?} source_file.original_start_pos={:?} source_file.original_end_pos={:?}", . hi, . source_file.original_start_pos, . source_file.original_end_pos . ); . . let lo = 202,364 ( 0.01%) (lo + source_file.translated_source_file.start_pos) - source_file.original_start_pos; . let hi = . (hi + source_file.translated_source_file.start_pos) - source_file.original_start_pos; . . // Do not try to decode parent for foreign spans. . Span::new(lo, hi, ctxt, None) 456,525 ( 0.02%) } . } . . impl<'a, 'tcx> Decodable> for &'tcx [thir::abstract_const::Node<'tcx>] { . fn decode(d: &mut DecodeContext<'a, 'tcx>) -> Self { . ty::codec::RefDecodable::decode(d) . } . } . -- line 587 ---------------------------------------- -- line 590 ---------------------------------------- . ty::codec::RefDecodable::decode(d) . } . } . . impl<'a, 'tcx, T: Decodable>> Decodable> . for Lazy . { . fn decode(decoder: &mut DecodeContext<'a, 'tcx>) -> Self { 20,911 ( 0.00%) decoder.read_lazy_with_meta(()) . } . } . . impl<'a, 'tcx, T: Decodable>> Decodable> . for Lazy<[T]> . { . fn decode(decoder: &mut DecodeContext<'a, 'tcx>) -> Self { . let len = decoder.read_usize(); 21,226 ( 0.00%) if len == 0 { Lazy::empty() } else { decoder.read_lazy_with_meta(len) } . } . } . . impl<'a, 'tcx, I: Idx, T: Decodable>> Decodable> . for Lazy> . where . Option: FixedSizeEncoding, . { . fn decode(decoder: &mut DecodeContext<'a, 'tcx>) -> Self { . let len = decoder.read_usize(); 14,534 ( 0.00%) decoder.read_lazy_with_meta(len) . } . } . . implement_ty_decoder!(DecodeContext<'a, 'tcx>); . . impl<'tcx> MetadataBlob { . crate fn new(metadata_ref: MetadataRef) -> MetadataBlob { . MetadataBlob(Lrc::new(metadata_ref)) -- line 626 ---------------------------------------- -- line 630 ---------------------------------------- . self.blob().starts_with(METADATA_HEADER) . } . . crate fn get_rustc_version(&self) -> String { . Lazy::::from_position(NonZeroUsize::new(METADATA_HEADER.len() + 4).unwrap()) . .decode(self) . } . 258 ( 0.00%) crate fn get_root(&self) -> CrateRoot<'tcx> { 86 ( 0.00%) let slice = &self.blob()[..]; . let offset = METADATA_HEADER.len(); 430 ( 0.00%) let pos = (((slice[offset + 0] as u32) << 24) 172 ( 0.00%) | ((slice[offset + 1] as u32) << 16) 344 ( 0.00%) | ((slice[offset + 2] as u32) << 8) 258 ( 0.00%) | ((slice[offset + 3] as u32) << 0)) as usize; . Lazy::>::from_position(NonZeroUsize::new(pos).unwrap()).decode(self) 430 ( 0.00%) } . . crate fn list_crate_metadata(&self, out: &mut dyn io::Write) -> io::Result<()> { . let root = self.get_root(); . writeln!(out, "Crate info:")?; . writeln!(out, "name {}{}", root.name, root.extra_filename)?; . writeln!(out, "hash {} stable_crate_id {:?}", root.hash, root.stable_crate_id)?; . writeln!(out, "proc_macro {:?}", root.proc_macro_data.is_some())?; . writeln!(out, "=External Dependencies=")?; -- line 654 ---------------------------------------- -- line 666 ---------------------------------------- . } . write!(out, "\n")?; . Ok(()) . } . } . . impl CrateRoot<'_> { . crate fn is_proc_macro_crate(&self) -> bool { 52 ( 0.00%) self.proc_macro_data.is_some() . } . . crate fn name(&self) -> Symbol { 12 ( 0.00%) self.name . } . . crate fn hash(&self) -> Svh { 27 ( 0.00%) self.hash . } . . crate fn stable_crate_id(&self) -> StableCrateId { 60 ( 0.00%) self.stable_crate_id . } . . crate fn triple(&self) -> &TargetTriple { . &self.triple . } . . crate fn decode_crate_deps<'a>( . &self, . metadata: &'a MetadataBlob, . ) -> impl ExactSizeIterator + Captures<'a> { 50 ( 0.00%) self.crate_deps.decode(metadata) . } . } . . impl<'a, 'tcx> CrateMetadataRef<'a> { 21 ( 0.00%) fn raw_proc_macro(self, id: DefIndex) -> &'a ProcMacro { . // DefIndex's in root.proc_macro_data have a one-to-one correspondence . // with items in 'raw_proc_macros'. 6 ( 0.00%) let pos = self . .root . .proc_macro_data . .as_ref() . .unwrap() . .macros . .decode(self) . .position(|i| i == id) . .unwrap(); 12 ( 0.00%) &self.raw_proc_macros.unwrap()[pos] 27 ( 0.00%) } . 57,024 ( 0.00%) fn opt_item_ident(self, item_index: DefIndex, sess: &Session) -> Option { 42,572 ( 0.00%) let name = self.def_key(item_index).disambiguated_data.data.get_opt_name()?; 21,119 ( 0.00%) let span = match self.root.tables.ident_span.get(self, item_index) { . Some(lazy_span) => lazy_span.decode((self, sess)), . None => { . // FIXME: this weird case of a name with no span is specific to `extern crate` . // items, which are supposed to be treated like `use` items and only be encoded . // to metadata as `Export`s, return `None` because that's what all the callers . // expect in this case. 162 ( 0.00%) assert_eq!(self.def_kind(item_index), DefKind::ExternCrate); . return None; . } . }; . Some(Ident::new(name, span)) 73,016 ( 0.00%) } . 340 ( 0.00%) fn item_ident(self, item_index: DefIndex, sess: &Session) -> Ident { 9,914 ( 0.00%) self.opt_item_ident(item_index, sess).expect("no encoded ident for item") 680 ( 0.00%) } . . fn maybe_kind(self, item_id: DefIndex) -> Option { 21,456 ( 0.00%) self.root.tables.kind.get(self, item_id).map(|k| k.decode(self)) . } . 85,824 ( 0.00%) fn kind(self, item_id: DefIndex) -> EntryKind { . self.maybe_kind(item_id).unwrap_or_else(|| { . bug!( . "CrateMetadata::kind({:?}): id not found, in crate {:?} with number {}", . item_id, . self.root.name, . self.cnum, . ) . }) 64,368 ( 0.00%) } . 11,473 ( 0.00%) fn def_kind(self, item_id: DefIndex) -> DefKind { 13,861 ( 0.00%) self.root.tables.def_kind.get(self, item_id).map(|k| k.decode(self)).unwrap_or_else(|| { . bug!( . "CrateMetadata::def_kind({:?}): id not found, in crate {:?} with number {}", . item_id, . self.root.name, . self.cnum, . ) . }) 6,556 ( 0.00%) } . 3,060 ( 0.00%) fn get_span(self, index: DefIndex, sess: &Session) -> Span { 8,777 ( 0.00%) self.root . .tables . .span . .get(self, index) . .unwrap_or_else(|| panic!("Missing span for {:?}", index)) . .decode((self, sess)) 1,700 ( 0.00%) } . . fn load_proc_macro(self, id: DefIndex, sess: &Session) -> SyntaxExtension { 36 ( 0.00%) let (name, kind, helper_attrs) = match *self.raw_proc_macro(id) { . ProcMacro::CustomDerive { trait_name, attributes, client } => { . let helper_attrs = . attributes.iter().cloned().map(Symbol::intern).collect::>(); . ( . trait_name, . SyntaxExtensionKind::Derive(Box::new(ProcMacroDerive { client })), . helper_attrs, . ) . } 8 ( 0.00%) ProcMacro::Attr { name, client } => { 8 ( 0.00%) (name, SyntaxExtensionKind::Attr(Box::new(AttrProcMacro { client })), Vec::new()) . } . ProcMacro::Bang { name, client } => { . (name, SyntaxExtensionKind::Bang(Box::new(BangProcMacro { client })), Vec::new()) . } . }; . 10 ( 0.00%) let attrs: Vec<_> = self.get_item_attrs(id, sess).collect(); 24 ( 0.00%) SyntaxExtension::new( . sess, 10 ( 0.00%) kind, . self.get_span(id, sess), 8 ( 0.00%) helper_attrs, 6 ( 0.00%) self.root.edition, 2 ( 0.00%) Symbol::intern(name), . &attrs, . ) . } . 806 ( 0.00%) fn get_trait_def(self, item_id: DefIndex, sess: &Session) -> ty::TraitDef { 248 ( 0.00%) match self.kind(item_id) { 62 ( 0.00%) EntryKind::Trait(data) => { 372 ( 0.00%) let data = data.decode((self, sess)); 992 ( 0.00%) ty::TraitDef::new( . self.local_def_id(item_id), 124 ( 0.00%) data.unsafety, 124 ( 0.00%) data.paren_sugar, 124 ( 0.00%) data.has_auto_impl, 124 ( 0.00%) data.is_marker, 62 ( 0.00%) data.skip_array_during_method_dispatch, 62 ( 0.00%) data.specialization_kind, 248 ( 0.00%) self.def_path_hash(item_id), 62 ( 0.00%) data.must_implement_one_of, . ) . } . EntryKind::TraitAlias => ty::TraitDef::new( . self.local_def_id(item_id), . hir::Unsafety::Normal, . false, . false, . false, . false, . ty::trait_def::TraitSpecializationKind::None, . self.def_path_hash(item_id), . None, . ), . _ => bug!("def-index does not refer to trait or trait alias"), . } 558 ( 0.00%) } . 1,035 ( 0.00%) fn get_variant( . self, . kind: &EntryKind, . index: DefIndex, . parent_did: DefId, . sess: &Session, . ) -> ty::VariantDef { 460 ( 0.00%) let data = match kind { . EntryKind::Variant(data) | EntryKind::Struct(data, _) | EntryKind::Union(data, _) => { 575 ( 0.00%) data.decode(self) . } . _ => bug!(), . }; . 416 ( 0.00%) let adt_kind = match kind { . EntryKind::Variant(_) => ty::AdtKind::Enum, . EntryKind::Struct(..) => ty::AdtKind::Struct, . EntryKind::Union(..) => ty::AdtKind::Union, . _ => bug!(), . }; . . let variant_did = . if adt_kind == ty::AdtKind::Enum { Some(self.local_def_id(index)) } else { None }; 230 ( 0.00%) let ctor_did = data.ctor.map(|index| self.local_def_id(index)); . 2,530 ( 0.00%) ty::VariantDef::new( . self.item_ident(index, sess).name, . variant_did, . ctor_did, 460 ( 0.00%) data.discr, 690 ( 0.00%) self.root . .tables . .children . .get(self, index) . .unwrap_or_else(Lazy::empty) . .decode(self) . .map(|index| ty::FieldDef { 186 ( 0.00%) did: self.local_def_id(index), 372 ( 0.00%) name: self.item_ident(index, sess).name, 372 ( 0.00%) vis: self.get_visibility(index), . }) . .collect(), . data.ctor_kind, . adt_kind, . parent_did, . false, 115 ( 0.00%) data.is_non_exhaustive, . ) 1,035 ( 0.00%) } . 1,560 ( 0.00%) fn get_adt_def(self, item_id: DefIndex, tcx: TyCtxt<'tcx>) -> &'tcx ty::AdtDef { 416 ( 0.00%) let kind = self.kind(item_id); 208 ( 0.00%) let did = self.local_def_id(item_id); . 498 ( 0.00%) let (adt_kind, repr) = match kind { . EntryKind::Enum(repr) => (ty::AdtKind::Enum, repr), . EntryKind::Struct(_, repr) => (ty::AdtKind::Struct, repr), . EntryKind::Union(_, repr) => (ty::AdtKind::Union, repr), . _ => bug!("get_adt_def called on a non-ADT {:?}", did), . }; . . let variants = if let ty::AdtKind::Enum = adt_kind { 60 ( 0.00%) self.root . .tables . .children . .get(self, item_id) . .unwrap_or_else(Lazy::empty) . .decode(self) 391 ( 0.00%) .map(|index| self.get_variant(&self.kind(index), index, did, tcx.sess)) . .collect() . } else { 736 ( 0.00%) std::iter::once(self.get_variant(&kind, item_id, did, tcx.sess)).collect() . }; . 1,516 ( 0.00%) tcx.alloc_adt_def(did, adt_kind, variants, repr) 728 ( 0.00%) } . . fn get_explicit_predicates( . self, . item_id: DefIndex, . tcx: TyCtxt<'tcx>, . ) -> ty::GenericPredicates<'tcx> { 6,415 ( 0.00%) self.root.tables.explicit_predicates.get(self, item_id).unwrap().decode((self, tcx)) . } . . fn get_inferred_outlives( . self, . item_id: DefIndex, . tcx: TyCtxt<'tcx>, . ) -> &'tcx [(ty::Predicate<'tcx>, Span)] { 6,415 ( 0.00%) self.root . .tables . .inferred_outlives . .get(self, item_id) 40 ( 0.00%) .map(|predicates| tcx.arena.alloc_from_iter(predicates.decode((self, tcx)))) . .unwrap_or_default() . } . . fn get_super_predicates( . self, . item_id: DefIndex, . tcx: TyCtxt<'tcx>, . ) -> ty::GenericPredicates<'tcx> { 280 ( 0.00%) self.root.tables.super_predicates.get(self, item_id).unwrap().decode((self, tcx)) . } . . fn get_explicit_item_bounds( . self, . item_id: DefIndex, . tcx: TyCtxt<'tcx>, . ) -> &'tcx [(ty::Predicate<'tcx>, Span)] { 60 ( 0.00%) self.root . .tables . .explicit_item_bounds . .get(self, item_id) 44 ( 0.00%) .map(|bounds| tcx.arena.alloc_from_iter(bounds.decode((self, tcx)))) . .unwrap_or_default() . } . . fn get_generics(self, item_id: DefIndex, sess: &Session) -> ty::Generics { 8,334 ( 0.00%) self.root.tables.generics.get(self, item_id).unwrap().decode((self, sess)) . } . . fn get_type(self, id: DefIndex, tcx: TyCtxt<'tcx>) -> Ty<'tcx> { 12,975 ( 0.00%) self.root . .tables . .ty . .get(self, id) . .unwrap_or_else(|| panic!("Not a type: {:?}", id)) . .decode((self, tcx)) . } . . fn get_stability(self, id: DefIndex) -> Option { 535 ( 0.00%) self.root.tables.stability.get(self, id).map(|stab| stab.decode(self)) . } . . fn get_const_stability(self, id: DefIndex) -> Option { 10 ( 0.00%) self.root.tables.const_stability.get(self, id).map(|stab| stab.decode(self)) . } . . fn get_deprecation(self, id: DefIndex) -> Option { 810 ( 0.00%) self.root.tables.deprecation.get(self, id).map(|depr| depr.decode(self)) . } . 138 ( 0.00%) fn get_visibility(self, id: DefIndex) -> ty::Visibility { 10,782 ( 0.00%) self.root.tables.visibility.get(self, id).unwrap().decode(self) 92 ( 0.00%) } . . fn get_impl_data(self, id: DefIndex) -> ImplData { 42,343 ( 0.00%) match self.kind(id) { 35,285 ( 0.00%) EntryKind::Impl(data) => data.decode(self), . _ => bug!(), . } . } . . fn get_parent_impl(self, id: DefIndex) -> Option { 19,716 ( 0.00%) self.get_impl_data(id).parent_impl . } . . fn get_impl_polarity(self, id: DefIndex) -> ty::ImplPolarity { 340 ( 0.00%) self.get_impl_data(id).polarity . } . . fn get_impl_defaultness(self, id: DefIndex) -> hir::Defaultness { 133 ( 0.00%) self.get_impl_data(id).defaultness . } . . fn get_impl_constness(self, id: DefIndex) -> hir::Constness { 11 ( 0.00%) self.get_impl_data(id).constness . } . . fn get_trait_item_def_id(self, id: DefIndex) -> Option { 8,985 ( 0.00%) self.root.tables.trait_item_def_id.get(self, id).map(|d| d.decode(self)) . } . . fn get_coerce_unsized_info(self, id: DefIndex) -> Option { 1 ( 0.00%) self.get_impl_data(id).coerce_unsized_info . } . . fn get_impl_trait(self, id: DefIndex, tcx: TyCtxt<'tcx>) -> Option> { 80,011 ( 0.00%) self.root.tables.impl_trait_ref.get(self, id).map(|tr| tr.decode((self, tcx))) . } . . fn get_expn_that_defined(self, id: DefIndex, sess: &Session) -> ExpnId { 2,410 ( 0.00%) self.root.tables.expn_that_defined.get(self, id).unwrap().decode((self, sess)) . } . . fn get_const_param_default( . self, . tcx: TyCtxt<'tcx>, . id: DefIndex, . ) -> rustc_middle::ty::Const<'tcx> { . self.root.tables.const_defaults.get(self, id).unwrap().decode((self, tcx)) -- line 1026 ---------------------------------------- -- line 1028 ---------------------------------------- . . /// Iterates over all the stability attributes in the given crate. . fn get_lib_features(self, tcx: TyCtxt<'tcx>) -> &'tcx [(Symbol, Option)] { . tcx.arena.alloc_from_iter(self.root.lib_features.decode(self)) . } . . /// Iterates over the language items in the given crate. . fn get_lang_items(self, tcx: TyCtxt<'tcx>) -> &'tcx [(DefId, usize)] { 81 ( 0.00%) tcx.arena.alloc_from_iter( 54 ( 0.00%) self.root . .lang_items . .decode(self) 564 ( 0.00%) .map(|(def_index, index)| (self.local_def_id(def_index), index)), . ) . } . . /// Iterates over the diagnostic items in the given crate. . fn get_diagnostic_items(self) -> DiagnosticItems { . let mut id_to_name = FxHashMap::default(); 54 ( 0.00%) let name_to_id = self . .root . .diagnostic_items . .decode(self) . .map(|(name, def_index)| { 130 ( 0.00%) let id = self.local_def_id(def_index); 130 ( 0.00%) id_to_name.insert(id, name); . (name, id) . }) . .collect(); 243 ( 0.00%) DiagnosticItems { id_to_name, name_to_id } . } . . /// Iterates over all named children of the given module, . /// including both proper items and reexports. . /// Module here is understood in name resolution sense - it can be a `mod` item, . /// or a crate root, or an enum, or a trait. . fn for_each_module_child( . self, . id: DefIndex, . mut callback: impl FnMut(ModChild), . sess: &Session, . ) { 308 ( 0.00%) if let Some(data) = &self.root.proc_macro_data { . // If we are loading as a proc macro, we want to return . // the view of this crate as a proc macro crate. 1 ( 0.00%) if id == CRATE_DEF_INDEX { 15 ( 0.00%) for def_index in data.macros.decode(self) { 3 ( 0.00%) let raw_macro = self.raw_proc_macro(def_index); . let res = Res::Def( . DefKind::Macro(macro_kind(raw_macro)), . self.local_def_id(def_index), . ); . let ident = self.item_ident(def_index, sess); . callback(ModChild { . ident, . res, -- line 1083 ---------------------------------------- -- line 1085 ---------------------------------------- . span: ident.span, . }); . } . } . return; . } . . // Iterate over all children. 608 ( 0.00%) if let Some(children) = self.root.tables.children.get(self, id) { 1,216 ( 0.00%) for child_index in children.decode((self, sess)) { 18,642 ( 0.00%) if let Some(ident) = self.opt_item_ident(child_index, sess) { 2,436 ( 0.00%) let kind = self.def_kind(child_index); 1,624 ( 0.00%) if matches!(kind, DefKind::Macro(..)) { . // FIXME: Macros are currently encoded twice, once as items and once as . // reexports. We ignore the items here and only use the reexports. . continue; . } . let def_id = self.local_def_id(child_index); . let res = Res::Def(kind, def_id); . let vis = self.get_visibility(child_index); . let span = self.get_span(child_index, sess); . . callback(ModChild { ident, res, vis, span }); . . // For non-re-export structs and variants add their constructors to children. . // Re-export lists automatically contain constructors when necessary. 3,102 ( 0.00%) match kind { . DefKind::Struct => { 91 ( 0.00%) if let Some((ctor_def_id, ctor_kind)) = . self.get_ctor_def_id_and_kind(child_index) . { . let ctor_res = . Res::Def(DefKind::Ctor(CtorOf::Struct, ctor_kind), ctor_def_id); . let vis = self.get_visibility(ctor_def_id.index); . callback(ModChild { ident, res: ctor_res, vis, span }); . } . } . DefKind::Variant => { . // Braced variants, unlike structs, generate unusable names in . // value namespace, they are reserved for possible future use. . // It's ok to use the variant's id as a ctor id since an . // error will be reported on any use of such resolution anyway. 21 ( 0.00%) let (ctor_def_id, ctor_kind) = self . .get_ctor_def_id_and_kind(child_index) . .unwrap_or((def_id, CtorKind::Fictive)); . let ctor_res = . Res::Def(DefKind::Ctor(CtorOf::Variant, ctor_kind), ctor_def_id); . let mut vis = self.get_visibility(ctor_def_id.index); 7 ( 0.00%) if ctor_def_id == def_id && vis.is_public() { . // For non-exhaustive variants lower the constructor visibility to . // within the crate. We only need this for fictive constructors, . // for other constructors correct visibilities . // were already encoded in metadata. . let mut attrs = self.get_item_attrs(def_id.index, sess); . if attrs.any(|item| item.has_name(sym::non_exhaustive)) { . let crate_def_id = self.local_def_id(CRATE_DEF_INDEX); . vis = ty::Visibility::Restricted(crate_def_id); -- line 1141 ---------------------------------------- -- line 1144 ---------------------------------------- . callback(ModChild { ident, res: ctor_res, vis, span }); . } . _ => {} . } . } . } . } . 758 ( 0.00%) match self.kind(id) { 66 ( 0.00%) EntryKind::Mod(exports) => { 2,120 ( 0.00%) for exp in exports.decode((self, sess)) { . callback(exp); . } . } . EntryKind::Enum(..) | EntryKind::Trait(..) => {} . _ => bug!("`for_each_module_child` is called on a non-module: {:?}", self.def_kind(id)), . } . } . . fn is_ctfe_mir_available(self, id: DefIndex) -> bool { 216 ( 0.00%) self.root.tables.mir_for_ctfe.get(self, id).is_some() . } . . fn is_item_mir_available(self, id: DefIndex) -> bool { 1,568 ( 0.00%) self.root.tables.mir.get(self, id).is_some() . } . . fn module_expansion(self, id: DefIndex, sess: &Session) -> ExpnId { 2,912 ( 0.00%) match self.kind(id) { . EntryKind::Mod(_) | EntryKind::Enum(_) | EntryKind::Trait(_) => { . self.get_expn_that_defined(id, sess) . } . _ => panic!("Expected module, found {:?}", self.local_def_id(id)), . } . } . . fn get_optimized_mir(self, tcx: TyCtxt<'tcx>, id: DefIndex) -> Body<'tcx> { 1,955 ( 0.00%) self.root . .tables . .mir . .get(self, id) . .unwrap_or_else(|| { . bug!("get_optimized_mir: missing MIR for `{:?}`", self.local_def_id(id)) . }) . .decode((self, tcx)) . } . . fn get_mir_for_ctfe(self, tcx: TyCtxt<'tcx>, id: DefIndex) -> Body<'tcx> { 270 ( 0.00%) self.root . .tables . .mir_for_ctfe . .get(self, id) . .unwrap_or_else(|| { . bug!("get_mir_for_ctfe: missing MIR for `{:?}`", self.local_def_id(id)) . }) . .decode((self, tcx)) . } -- line 1200 ---------------------------------------- -- line 1216 ---------------------------------------- . .tables . .unused_generic_params . .get(self, id) . .map(|params| params.decode(self)) . .unwrap_or_default() . } . . fn get_promoted_mir(self, tcx: TyCtxt<'tcx>, id: DefIndex) -> IndexVec> { 25 ( 0.00%) self.root . .tables . .promoted_mir . .get(self, id) . .unwrap_or_else(|| { . bug!("get_promoted_mir: missing MIR for `{:?}`", self.local_def_id(id)) . }) . .decode((self, tcx)) . } -- line 1232 ---------------------------------------- -- line 1242 ---------------------------------------- . qualif, . _, . ) => qualif, . _ => bug!("mir_const_qualif: unexpected kind"), . } . } . . fn get_fn_has_self_parameter(self, id: DefIndex) -> bool { 665 ( 0.00%) match self.kind(id) { 532 ( 0.00%) EntryKind::AssocFn(data) => data.decode(self).has_self, . _ => false, . } . } . . fn get_associated_item_def_ids(self, tcx: TyCtxt<'tcx>, id: DefIndex) -> &'tcx [DefId] { 2,590 ( 0.00%) if let Some(children) = self.root.tables.children.get(self, id) { 1,110 ( 0.00%) tcx.arena.alloc_from_iter( 370 ( 0.00%) children.decode((self, tcx.sess)).map(|child_index| self.local_def_id(child_index)), . ) . } else { . &[] . } . } . 17,064 ( 0.00%) fn get_associated_item(self, id: DefIndex, sess: &Session) -> ty::AssocItem { 9,954 ( 0.00%) let def_key = self.def_key(id); . let parent = self.local_def_id(def_key.parent.unwrap()); . let ident = self.item_ident(id, sess); . 10,190 ( 0.00%) let (kind, container, has_self) = match self.kind(id) { 28 ( 0.00%) EntryKind::AssocConst(container, _, _) => (ty::AssocKind::Const, container, false), 1,311 ( 0.00%) EntryKind::AssocFn(data) => { . let data = data.decode(self); 9,177 ( 0.00%) (ty::AssocKind::Fn, data.container, data.has_self) . } 208 ( 0.00%) EntryKind::AssocType(container) => (ty::AssocKind::Type, container, false), . _ => bug!("cannot get associated-item of `{:?}`", def_key), . }; . 25,596 ( 0.00%) ty::AssocItem { . name: ident.name, . kind, . vis: self.get_visibility(id), . defaultness: container.defaultness(), . def_id: self.local_def_id(id), . trait_item_def_id: self.get_trait_item_def_id(id), . container: container.with_def_id(parent), . fn_has_self_parameter: has_self, . } 12,798 ( 0.00%) } . . fn get_item_variances(self, id: DefIndex) -> impl Iterator + 'a { 255 ( 0.00%) self.root.tables.variances.get(self, id).unwrap_or_else(Lazy::empty).decode(self) . } . . fn get_ctor_def_id_and_kind(self, node_id: DefIndex) -> Option<(DefId, CtorKind)> { 2,415 ( 0.00%) match self.kind(node_id) { . EntryKind::Struct(data, _) | EntryKind::Variant(data) => { 1,046 ( 0.00%) let vdata = data.decode(self); 500 ( 0.00%) vdata.ctor.map(|index| (self.local_def_id(index), vdata.ctor_kind)) . } . _ => None, . } 1,132 ( 0.00%) } . 1,884 ( 0.00%) fn get_item_attrs( . self, . id: DefIndex, . sess: &'a Session, . ) -> impl Iterator + 'a { 5,577 ( 0.00%) self.root . .tables . .attributes . .get(self, id) . .unwrap_or_else(|| { . // Structure and variant constructors don't have any attributes encoded for them, . // but we assume that someone passing a constructor ID actually wants to look at . // the attributes on the corresponding struct or variant. . let def_key = self.def_key(id); -- line 1320 ---------------------------------------- -- line 1322 ---------------------------------------- . let parent_id = def_key.parent.expect("no parent for a constructor"); . self.root . .tables . .attributes . .get(self, parent_id) . .expect("no encoded attributes for a structure or variant") . }) . .decode((self, sess)) 1,413 ( 0.00%) } . . fn get_struct_field_names( . self, . id: DefIndex, . sess: &'a Session, . ) -> impl Iterator> + 'a { 1,100 ( 0.00%) self.root . .tables . .children . .get(self, id) . .unwrap_or_else(Lazy::empty) . .decode(self) 3,400 ( 0.00%) .map(move |index| respan(self.get_span(index, sess), self.item_ident(index, sess).name)) . } . . fn get_struct_field_visibilities(self, id: DefIndex) -> impl Iterator + 'a { 150 ( 0.00%) self.root . .tables . .children . .get(self, id) . .unwrap_or_else(Lazy::empty) . .decode(self) 46 ( 0.00%) .map(move |field_index| self.get_visibility(field_index)) . } . . fn get_inherent_implementations_for_type( . self, . tcx: TyCtxt<'tcx>, . id: DefIndex, . ) -> &'tcx [DefId] { 108 ( 0.00%) tcx.arena.alloc_from_iter( 135 ( 0.00%) self.root . .tables . .inherent_impls . .get(self, id) . .unwrap_or_else(Lazy::empty) . .decode(self) 1,448 ( 0.00%) .map(|index| self.local_def_id(index)), . ) . } . . fn get_traits(self) -> impl Iterator + 'a { . self.root.traits.decode(self).map(move |index| self.local_def_id(index)) . } . . fn get_trait_impls(self) -> impl Iterator)> + 'a { -- line 1376 ---------------------------------------- -- line 1380 ---------------------------------------- . index: *trait_index, . }; . impls.decode(self).map(move |(impl_index, simplified_self_ty)| { . (trait_def_id, self.local_def_id(impl_index), simplified_self_ty) . }) . }) . } . 15,552 ( 0.00%) fn get_implementations_of_trait( . self, . tcx: TyCtxt<'tcx>, . trait_def_id: DefId, . ) -> &'tcx [(DefId, Option)] { 1,728 ( 0.00%) if self.trait_impls.is_empty() { . return &[]; . } . . // Do a reverse lookup beforehand to avoid touching the crate_num . // hash map in the loop below. 2,104 ( 0.00%) let key = match self.reverse_translate_def_id(trait_def_id) { . Some(def_id) => (def_id.krate.as_u32(), def_id.index), . None => return &[], . }; . . if let Some(impls) = self.trait_impls.get(&key) { 1,052 ( 0.00%) tcx.arena.alloc_from_iter( 526 ( 0.00%) impls . .decode(self) 8,604 ( 0.00%) .map(|(idx, simplified_self_ty)| (self.local_def_id(idx), simplified_self_ty)), . ) . } else { . &[] . } 13,824 ( 0.00%) } . . fn get_trait_of_item(self, id: DefIndex) -> Option { 2,252 ( 0.00%) let def_key = self.def_key(id); 2,815 ( 0.00%) match def_key.disambiguated_data.data { . DefPathData::TypeNs(..) | DefPathData::ValueNs(..) => (), . // Not an associated item . _ => return None, . } 4,950 ( 0.00%) def_key.parent.and_then(|parent_index| match self.kind(parent_index) { . EntryKind::Trait(_) | EntryKind::TraitAlias => Some(self.local_def_id(parent_index)), . _ => None, . }) . } . . fn get_native_libraries(self, sess: &'a Session) -> impl Iterator + 'a { 54 ( 0.00%) self.root.native_libraries.decode((self, sess)) . } . . fn get_proc_macro_quoted_span(self, index: usize, sess: &Session) -> Span { . self.root . .tables . .proc_macro_quoted_spans . .get(self, index) . .unwrap_or_else(|| panic!("Missing proc macro quoted span: {:?}", index)) -- line 1437 ---------------------------------------- -- line 1450 ---------------------------------------- . self.root.dylib_dependency_formats.decode(self).enumerate().flat_map(|(i, link)| { . let cnum = CrateNum::new(i + 1); . link.map(|link| (self.cnum_map[cnum], link)) . }), . ) . } . . fn get_missing_lang_items(self, tcx: TyCtxt<'tcx>) -> &'tcx [lang_items::LangItem] { 135 ( 0.00%) tcx.arena.alloc_from_iter(self.root.lang_items_missing.decode(self)) . } . . fn get_fn_param_names(self, tcx: TyCtxt<'tcx>, id: DefIndex) -> &'tcx [Ident] { . let param_names = match self.kind(id) { . EntryKind::Fn(data) | EntryKind::ForeignFn(data) => data.decode(self).param_names, . EntryKind::AssocFn(data) => data.decode(self).fn_data.param_names, . _ => Lazy::empty(), . }; . tcx.arena.alloc_from_iter(param_names.decode((self, tcx))) . } . . fn exported_symbols( . self, . tcx: TyCtxt<'tcx>, . ) -> &'tcx [(ExportedSymbol<'tcx>, SymbolExportLevel)] { 162 ( 0.00%) tcx.arena.alloc_from_iter(self.root.exported_symbols.decode((self, tcx))) . } . . fn get_rendered_const(self, id: DefIndex) -> String { . match self.kind(id) { . EntryKind::AnonConst(_, data) . | EntryKind::Const(_, data) . | EntryKind::AssocConst(_, _, data) => data.decode(self).0, . _ => bug!(), . } . } . . fn get_macro(self, id: DefIndex, sess: &Session) -> MacroDef { 78 ( 0.00%) match self.kind(id) { 13 ( 0.00%) EntryKind::MacroDef(macro_def) => macro_def.decode((self, sess)), . _ => bug!(), . } . } . . // This replicates some of the logic of the crate-local `is_const_fn_raw` query, because we . // don't serialize constness for tuple variant and tuple struct constructors. . fn is_const_fn_raw(self, id: DefIndex) -> bool { 225 ( 0.00%) let constness = match self.kind(id) { 44 ( 0.00%) EntryKind::AssocFn(data) => data.decode(self).fn_data.constness, 18 ( 0.00%) EntryKind::Fn(data) => data.decode(self).constness, . EntryKind::ForeignFn(data) => data.decode(self).constness, . EntryKind::Variant(..) | EntryKind::Struct(..) => hir::Constness::Const, . _ => hir::Constness::NotConst, . }; . constness == hir::Constness::Const . } . . fn asyncness(self, id: DefIndex) -> hir::IsAsync { -- line 1506 ---------------------------------------- -- line 1508 ---------------------------------------- . EntryKind::Fn(data) => data.decode(self).asyncness, . EntryKind::AssocFn(data) => data.decode(self).fn_data.asyncness, . EntryKind::ForeignFn(data) => data.decode(self).asyncness, . _ => bug!("asyncness: expected function kind"), . } . } . . fn is_foreign_item(self, id: DefIndex) -> bool { 5,170 ( 0.00%) match self.kind(id) { . EntryKind::ForeignImmStatic | EntryKind::ForeignMutStatic | EntryKind::ForeignFn(_) => { . true . } . _ => false, . } . } . . fn static_mutability(self, id: DefIndex) -> Option { 539 ( 0.00%) match self.kind(id) { . EntryKind::ImmStatic | EntryKind::ForeignImmStatic => Some(hir::Mutability::Not), . EntryKind::MutStatic | EntryKind::ForeignMutStatic => Some(hir::Mutability::Mut), . _ => None, . } . } . . fn generator_kind(self, id: DefIndex) -> Option { 120 ( 0.00%) match self.kind(id) { . EntryKind::Generator(data) => Some(data), . _ => None, . } . } . . fn fn_sig(self, id: DefIndex, tcx: TyCtxt<'tcx>) -> ty::PolyFnSig<'tcx> { 2,810 ( 0.00%) self.root.tables.fn_sig.get(self, id).unwrap().decode((self, tcx)) . } . . #[inline] 268,002 ( 0.01%) fn def_key(self, index: DefIndex) -> DefKey { 76,900 ( 0.00%) *self . .def_key_cache . .lock() . .entry(index) 26,100 ( 0.00%) .or_insert_with(|| self.root.tables.def_keys.get(self, index).unwrap().decode(self)) 306,288 ( 0.01%) } . . // Returns the path leading to the thing with this `id`. . fn def_path(self, id: DefIndex) -> DefPath { . debug!("def_path(cnum={:?}, id={:?})", self.cnum, id); 256 ( 0.00%) DefPath::make(self.cnum, id, |parent| self.def_key(parent)) . } . . fn def_path_hash_unlocked( . self, . index: DefIndex, . def_path_hashes: &mut FxHashMap, . ) -> DefPathHash { 13,264 ( 0.00%) *def_path_hashes.entry(index).or_insert_with(|| { 5,310 ( 0.00%) self.root.tables.def_path_hashes.get(self, index).unwrap().decode(self) . }) . } . . #[inline] 46,424 ( 0.00%) fn def_path_hash(self, index: DefIndex) -> DefPathHash { . let mut def_path_hashes = self.def_path_hash_cache.lock(); . self.def_path_hash_unlocked(index, &mut def_path_hashes) 53,056 ( 0.00%) } . . #[inline] . fn def_path_hash_to_def_index(self, hash: DefPathHash) -> DefIndex { . self.def_path_hash_map.def_path_hash_to_def_index(&hash) . } . . fn expn_hash_to_expn_id(self, sess: &Session, index_guess: u32, hash: ExpnHash) -> ExpnId { . debug_assert_eq!(ExpnId::from_hash(hash), None); -- line 1580 ---------------------------------------- -- line 1635 ---------------------------------------- . /// Proc macro crates don't currently export spans, so this function does not have . /// to work for them. . fn imported_source_files(self, sess: &Session) -> &'a [ImportedSourceFile] { . // Translate the virtual `/rustc/$hash` prefix back to a real directory . // that should hold actual sources, where possible. . // . // NOTE: if you update this, you might need to also update bootstrap's code for generating . // the `rust-src` component in `Src::run` in `src/bootstrap/dist.rs`. 101,182 ( 0.00%) let virtual_rust_source_base_dir = option_env!("CFG_VIRTUAL_RUST_SOURCE_BASE_DIR") . .map(Path::new) . .filter(|_| { . // Only spend time on further checks if we have what to translate *to*. . sess.opts.real_rust_source_base_dir.is_some() . }) . .filter(|virtual_dir| { . // Don't translate away `/rustc/$hash` if we're still remapping to it, . // since that means we're still building `std`/`rustc` that need it, . // and we don't want the real path to leak into codegen/debuginfo. . !sess.opts.remap_path_prefix.iter().any(|(_from, to)| to == virtual_dir) . }); 101,182 ( 0.00%) let try_to_translate_virtual_to_real = |name: &mut rustc_span::FileName| { . debug!( . "try_to_translate_virtual_to_real(name={:?}): \ . virtual_rust_source_base_dir={:?}, real_rust_source_base_dir={:?}", . name, virtual_rust_source_base_dir, sess.opts.real_rust_source_base_dir, . ); . 2,770 ( 0.00%) if let Some(virtual_dir) = virtual_rust_source_base_dir { . if let Some(real_dir) = &sess.opts.real_rust_source_base_dir { . if let rustc_span::FileName::Real(old_name) = name { . if let rustc_span::RealFileName::Remapped { local_path: _, virtual_name } = . old_name . { . if let Ok(rest) = virtual_name.strip_prefix(virtual_dir) { . let virtual_name = virtual_name.clone(); . -- line 1670 ---------------------------------------- -- line 1709 ---------------------------------------- . *old_name = new_name; . } . } . } . } . } . }; . 100,941 ( 0.00%) self.cdata.source_map_import_info.get_or_init(|| { 48 ( 0.00%) let external_source_map = self.root.source_map.decode(self); . . external_source_map . .map(|source_file_to_import| { . // We can't reuse an existing SourceFile, so allocate a new one . // containing the information we need. . let rustc_span::SourceFile { 11,634 ( 0.00%) mut name, 3,324 ( 0.00%) src_hash, 554 ( 0.00%) start_pos, 1,108 ( 0.00%) end_pos, 2,770 ( 0.00%) mut lines, 2,770 ( 0.00%) mut multibyte_chars, 3,324 ( 0.00%) mut non_narrow_chars, 2,770 ( 0.00%) mut normalized_pos, 3,324 ( 0.00%) name_hash, . .. . } = source_file_to_import; . . // If this file is under $sysroot/lib/rustlib/src/ but has not been remapped . // during rust bootstrapping by `remap-debuginfo = true`, and the user . // wish to simulate that behaviour by -Z simulate-remapped-rust-src-base, . // then we change `name` to a similar state as if the rust was bootstrapped . // with `remap-debuginfo = true`. . // This is useful for testing so that tests about the effects of . // `try_to_translate_virtual_to_real` don't have to worry about how the . // compiler is bootstrapped. 1,108 ( 0.00%) if let Some(virtual_dir) = . &sess.opts.debugging_opts.simulate_remapped_rust_src_base . { . if let Some(real_dir) = &sess.opts.real_rust_source_base_dir { . if let rustc_span::FileName::Real(ref mut old_name) = name { . if let rustc_span::RealFileName::LocalPath(local) = old_name { . if let Ok(rest) = local.strip_prefix(real_dir) { . *old_name = rustc_span::RealFileName::Remapped { . local_path: None, -- line 1753 ---------------------------------------- -- line 1767 ---------------------------------------- . let source_length = (end_pos - start_pos).to_usize(); . . // Translate line-start positions and multibyte character . // position into frame of reference local to file. . // `SourceMap::new_imported_source_file()` will then translate those . // coordinates to their new global frame of reference when the . // offset of the SourceFile is known. . for pos in &mut lines { 730,730 ( 0.03%) *pos = *pos - start_pos; . } . for mbc in &mut multibyte_chars { 1,656 ( 0.00%) mbc.pos = mbc.pos - start_pos; . } . for swc in &mut non_narrow_chars { 924 ( 0.00%) *swc = *swc - start_pos; . } . for np in &mut normalized_pos { . np.pos = np.pos - start_pos; . } . 9,972 ( 0.00%) let local_version = sess.source_map().new_imported_source_file( 7,202 ( 0.00%) name, 3,878 ( 0.00%) src_hash, . name_hash, . source_length, 1,108 ( 0.00%) self.cnum, 2,770 ( 0.00%) lines, 3,324 ( 0.00%) multibyte_chars, 2,770 ( 0.00%) non_narrow_chars, 2,216 ( 0.00%) normalized_pos, . start_pos, . end_pos, . ); . debug!( . "CrateMetaData::imported_source_files alloc \ . source_file {:?} original (start_pos {:?} end_pos {:?}) \ . translated (start_pos {:?} end_pos {:?})", . local_version.name, -- line 1804 ---------------------------------------- -- line 1808 ---------------------------------------- . local_version.end_pos . ); . . ImportedSourceFile { . original_start_pos: start_pos, . original_end_pos: end_pos, . translated_source_file: local_version, . } 1,662 ( 0.00%) }) . .collect() . }) . } . } . . impl CrateMetadata { 324 ( 0.00%) crate fn new( . sess: &Session, . blob: MetadataBlob, . root: CrateRoot<'static>, . raw_proc_macros: Option<&'static [ProcMacro]>, . cnum: CrateNum, . cnum_map: CrateNumMap, . dep_kind: CrateDepKind, . source: CrateSource, . private_dep: bool, . host_hash: Option, . ) -> CrateMetadata { 81 ( 0.00%) let trait_impls = root . .impls . .decode((&blob, sess)) . .map(|trait_impls| (trait_impls.trait_id, trait_impls.impls)) . .collect(); . let alloc_decoding_state = 81 ( 0.00%) AllocDecodingState::new(root.interpret_alloc_index.decode(&blob).collect()); . let dependencies = Lock::new(cnum_map.iter().cloned().collect()); . . // Pre-decode the DefPathHash->DefIndex table. This is a cheap operation . // that does not copy any data. It just does some data verification. 27 ( 0.00%) let def_path_hash_map = root.def_path_hash_map.decode(&blob); . 621 ( 0.00%) CrateMetadata { 27 ( 0.00%) blob, 81 ( 0.00%) root, 108 ( 0.00%) trait_impls, . raw_proc_macros, . source_map_import_info: OnceCell::new(), 108 ( 0.00%) def_path_hash_map, . expn_hash_map: Default::default(), 189 ( 0.00%) alloc_decoding_state, . cnum, 108 ( 0.00%) cnum_map, 135 ( 0.00%) dependencies, . dep_kind: Lock::new(dep_kind), 324 ( 0.00%) source, . private_dep, . host_hash, . extern_crate: Lock::new(None), . hygiene_context: Default::default(), . def_key_cache: Default::default(), . def_path_hash_cache: Default::default(), . } 243 ( 0.00%) } . . crate fn dependencies(&self) -> LockGuard<'_, Vec> { . self.dependencies.borrow() . } . . crate fn add_dependency(&self, cnum: CrateNum) { . self.dependencies.borrow_mut().push(cnum); . } . . crate fn update_extern_crate(&self, new_extern_crate: ExternCrate) -> bool { . let mut extern_crate = self.extern_crate.borrow_mut(); . let update = Some(new_extern_crate.rank()) > extern_crate.as_ref().map(ExternCrate::rank); . if update { 145 ( 0.00%) *extern_crate = Some(new_extern_crate); . } . update . } . . crate fn source(&self) -> &CrateSource { 1,568 ( 0.00%) &self.source . } . . crate fn dep_kind(&self) -> CrateDepKind { . *self.dep_kind.lock() . } . . crate fn update_dep_kind(&self, f: impl FnOnce(CrateDepKind) -> CrateDepKind) { 565 ( 0.00%) self.dep_kind.with_lock(|dep_kind| *dep_kind = f(*dep_kind)) . } . . crate fn panic_strategy(&self) -> PanicStrategy { . self.root.panic_strategy . } . . crate fn needs_panic_runtime(&self) -> bool { . self.root.needs_panic_runtime -- line 1905 ---------------------------------------- -- line 1909 ---------------------------------------- . self.root.panic_runtime . } . . crate fn is_profiler_runtime(&self) -> bool { . self.root.profiler_runtime . } . . crate fn needs_allocator(&self) -> bool { 5 ( 0.00%) self.root.needs_allocator . } . . crate fn has_global_allocator(&self) -> bool { . self.root.has_global_allocator . } . . crate fn has_default_lib_allocator(&self) -> bool { . self.root.has_default_lib_allocator . } . . crate fn is_proc_macro_crate(&self) -> bool { . self.root.is_proc_macro_crate() . } . . crate fn name(&self) -> Symbol { 8,495 ( 0.00%) self.root.name . } . . crate fn stable_crate_id(&self) -> StableCrateId { . self.root.stable_crate_id . } . . crate fn hash(&self) -> Svh { . self.root.hash . } . . fn num_def_ids(&self) -> usize { . self.root.tables.def_keys.size() . } . . fn local_def_id(&self, index: DefIndex) -> DefId { 19,318 ( 0.00%) DefId { krate: self.cnum, index } . } . . // Translate a DefId from the current compilation environment to a DefId . // for an external crate. . fn reverse_translate_def_id(&self, did: DefId) -> Option { . for (local, &global) in self.cnum_map.iter_enumerated() { 7,013 ( 0.00%) if global == did.krate { . return Some(DefId { krate: local, index: did.index }); . } . } . . None . } . } . . // Cannot be implemented on 'ProcMacro', as libproc_macro . // does not depend on librustc_ast . fn macro_kind(raw: &ProcMacro) -> MacroKind { 3 ( 0.00%) match raw { . ProcMacro::CustomDerive { .. } => MacroKind::Derive, . ProcMacro::Attr { .. } => MacroKind::Attr, . ProcMacro::Bang { .. } => MacroKind::Bang, . } . } 928,124 ( 0.04%) -------------------------------------------------------------------------------- -- Auto-annotated source: /usr/home/liquid/rust/worktree-benchmarking/compiler/rustc_serialize/src/opaque.rs -------------------------------------------------------------------------------- Ir -- line 14 ---------------------------------------- . . pub type EncodeResult = Result<(), !>; . . pub struct Encoder { . pub data: Vec, . } . . impl Encoder { 1 ( 0.00%) pub fn new(data: Vec) -> Encoder { 4 ( 0.00%) Encoder { data } 1 ( 0.00%) } . 1 ( 0.00%) pub fn into_inner(self) -> Vec { 4 ( 0.00%) self.data 1 ( 0.00%) } . . #[inline] . pub fn position(&self) -> usize { 9,684 ( 0.00%) self.data.len() . } . } . . macro_rules! write_leb128 { . ($enc:expr, $value:expr, $int_ty:ty, $fun:ident) => {{ . const MAX_ENCODED_LEN: usize = max_leb128_len!($int_ty); . let old_len = $enc.data.len(); . -- line 40 ---------------------------------------- -- line 67 ---------------------------------------- . . #[inline] . fn emit_unit(&mut self) -> EncodeResult { . Ok(()) . } . . #[inline] . fn emit_usize(&mut self, v: usize) -> EncodeResult { 237,449 ( 0.01%) write_leb128!(self, v, usize, write_usize_leb128) . } . . #[inline] . fn emit_u128(&mut self, v: u128) -> EncodeResult { 1,736 ( 0.00%) write_leb128!(self, v, u128, write_u128_leb128) . } . . #[inline] . fn emit_u64(&mut self, v: u64) -> EncodeResult { 1,379 ( 0.00%) write_leb128!(self, v, u64, write_u64_leb128) . } . . #[inline] . fn emit_u32(&mut self, v: u32) -> EncodeResult { 192,394 ( 0.01%) write_leb128!(self, v, u32, write_u32_leb128) . } . . #[inline] . fn emit_u16(&mut self, v: u16) -> EncodeResult { . self.data.extend_from_slice(&v.to_le_bytes()); . Ok(()) . } . -- line 98 ---------------------------------------- -- line 131 ---------------------------------------- . #[inline] . fn emit_i8(&mut self, v: i8) -> EncodeResult { . let as_u8: u8 = unsafe { std::mem::transmute(v) }; . self.emit_u8(as_u8) . } . . #[inline] . fn emit_bool(&mut self, v: bool) -> EncodeResult { 2,732 ( 0.00%) self.emit_u8(if v { 1 } else { 0 }) . } . . #[inline] . fn emit_f64(&mut self, v: f64) -> EncodeResult { . let as_u64: u64 = v.to_bits(); . self.emit_u64(as_u64) . } . -- line 147 ---------------------------------------- -- line 549 ---------------------------------------- . #[inline] . pub fn advance(&mut self, bytes: usize) { . self.position += bytes; . } . . #[inline] . pub fn read_raw_bytes(&mut self, bytes: usize) -> &'a [u8] { . let start = self.position; 27 ( 0.00%) self.position += bytes; 27 ( 0.00%) &self.data[start..self.position] . } . } . . macro_rules! read_leb128 { . ($dec:expr, $fun:ident) => {{ leb128::$fun($dec.data, &mut $dec.position) }}; . } . . impl<'a> serialize::Decoder for Decoder<'a> { . #[inline] . fn read_unit(&mut self) -> () { . () . } . . #[inline] . fn read_u128(&mut self) -> u128 { 2,004 ( 0.00%) read_leb128!(self, read_u128_leb128) . } . . #[inline] . fn read_u64(&mut self) -> u64 { 4,382 ( 0.00%) read_leb128!(self, read_u64_leb128) . } . . #[inline] 79,242 ( 0.00%) fn read_u32(&mut self) -> u32 { 346,483 ( 0.01%) read_leb128!(self, read_u32_leb128) 158,484 ( 0.01%) } . . #[inline] . fn read_u16(&mut self) -> u16 { 562,492 ( 0.02%) let bytes = [self.data[self.position], self.data[self.position + 1]]; 80,356 ( 0.00%) let value = u16::from_le_bytes(bytes); 160,726 ( 0.01%) self.position += 2; . value . } . . #[inline] . fn read_u8(&mut self) -> u8 { 2,225,160 ( 0.09%) let value = self.data[self.position]; 474,404 ( 0.02%) self.position += 1; . value . } . . #[inline] 441,817 ( 0.02%) fn read_usize(&mut self) -> usize { 442,225 ( 0.02%) read_leb128!(self, read_usize_leb128) 883,634 ( 0.03%) } . . #[inline] . fn read_i128(&mut self) -> i128 { . read_leb128!(self, read_i128_leb128) . } . . #[inline] . fn read_i64(&mut self) -> i64 { -- line 613 ---------------------------------------- -- line 637 ---------------------------------------- . #[inline] . fn read_isize(&mut self) -> isize { . read_leb128!(self, read_isize_leb128) . } . . #[inline] . fn read_bool(&mut self) -> bool { . let value = self.read_u8(); 42,884 ( 0.00%) value != 0 . } . . #[inline] . fn read_f64(&mut self) -> f64 { . let bits = self.read_u64(); . f64::from_bits(bits) . } . -- line 653 ---------------------------------------- -- line 660 ---------------------------------------- . #[inline] . fn read_char(&mut self) -> char { . let bits = self.read_u32(); . std::char::from_u32(bits).unwrap() . } . . #[inline] . fn read_str(&mut self) -> Cow<'_, str> { 56,337 ( 0.00%) let len = self.read_usize(); 323,808 ( 0.01%) let sentinel = self.data[self.position + len]; 81,258 ( 0.00%) assert!(sentinel == STR_SENTINEL); . let s = unsafe { . std::str::from_utf8_unchecked(&self.data[self.position..self.position + len]) . }; 203,145 ( 0.01%) self.position += len + 1; 45,993 ( 0.00%) Cow::Borrowed(s) . } . . #[inline] . fn read_raw_bytes_into(&mut self, s: &mut [u8]) { 1,062 ( 0.00%) let start = self.position; 5,190 ( 0.00%) self.position += s.len(); 1,062 ( 0.00%) s.copy_from_slice(&self.data[start..self.position]); . } . } . . // Specializations for contiguous byte sequences follow. The default implementations for slices . // encode and decode each element individually. This isn't necessary for `u8` slices when using . // opaque encoders and decoders, because each `u8` is unchanged by encoding and decoding. . // Therefore, we can use more efficient implementations that process the entire sequence at once. . -- line 690 ---------------------------------------- 1,013,838 ( 0.04%) -------------------------------------------------------------------------------- -- Auto-annotated source: /usr/home/liquid/rust/worktree-benchmarking/library/core/src/num/uint_macros.rs -------------------------------------------------------------------------------- Ir -- line 57 ---------------------------------------- . /// # Examples . /// . /// Basic usage: . /// . /// ``` . #[doc = concat!("assert_eq!(", stringify!($SelfT), "::from_str_radix(\"A\", 16), Ok(10));")] . /// ``` . #[stable(feature = "rust1", since = "1.0.0")] 278 ( 0.00%) pub fn from_str_radix(src: &str, radix: u32) -> Result { 139 ( 0.00%) from_str_radix(src, radix) 417 ( 0.00%) } . . /// Returns the number of ones in the binary representation of `self`. . /// . /// # Examples . /// . /// Basic usage: . /// . /// ``` -- line 75 ---------------------------------------- -- line 80 ---------------------------------------- . #[stable(feature = "rust1", since = "1.0.0")] . #[rustc_const_stable(feature = "const_math", since = "1.32.0")] . #[doc(alias = "popcount")] . #[doc(alias = "popcnt")] . #[must_use = "this returns the result of the operation, \ . without modifying the original"] . #[inline(always)] . pub const fn count_ones(self) -> u32 { 117,900 ( 0.00%) intrinsics::ctpop(self as $ActualT) as u32 . } . . /// Returns the number of zeros in the binary representation of `self`. . /// . /// # Examples . /// . /// Basic usage: . /// -- line 96 ---------------------------------------- -- line 118 ---------------------------------------- . /// assert_eq!(n.leading_zeros(), 2); . /// ``` . #[stable(feature = "rust1", since = "1.0.0")] . #[rustc_const_stable(feature = "const_math", since = "1.32.0")] . #[must_use = "this returns the result of the operation, \ . without modifying the original"] . #[inline(always)] . pub const fn leading_zeros(self) -> u32 { 309,256 ( 0.01%) intrinsics::ctlz(self as $ActualT) as u32 . } . . /// Returns the number of trailing zeros in the binary representation . /// of `self`. . /// . /// # Examples . /// . /// Basic usage: -- line 134 ---------------------------------------- -- line 139 ---------------------------------------- . /// assert_eq!(n.trailing_zeros(), 3); . /// ``` . #[stable(feature = "rust1", since = "1.0.0")] . #[rustc_const_stable(feature = "const_math", since = "1.32.0")] . #[must_use = "this returns the result of the operation, \ . without modifying the original"] . #[inline(always)] . pub const fn trailing_zeros(self) -> u32 { 283,885 ( 0.01%) intrinsics::cttz(self) as u32 . } . . /// Returns the number of leading ones in the binary representation of `self`. . /// . /// # Examples . /// . /// Basic usage: . /// -- line 155 ---------------------------------------- -- line 204 ---------------------------------------- . #[doc = concat!("assert_eq!(n.rotate_left(", $rot, "), m);")] . /// ``` . #[stable(feature = "rust1", since = "1.0.0")] . #[rustc_const_stable(feature = "const_math", since = "1.32.0")] . #[must_use = "this returns the result of the operation, \ . without modifying the original"] . #[inline(always)] . pub const fn rotate_left(self, n: u32) -> Self { 5,821,525 ( 0.23%) intrinsics::rotate_left(self, n as $SelfT) . } . . /// Shifts the bits to the right by a specified amount, `n`, . /// wrapping the truncated bits to the beginning of the resulting . /// integer. . /// . /// Please note this isn't the same operation as the `>>` shifting operator! . /// -- line 220 ---------------------------------------- -- line 429 ---------------------------------------- . )] . #[doc = concat!("assert_eq!((", stringify!($SelfT), "::MAX - 2).checked_add(3), None);")] . /// ``` . #[stable(feature = "rust1", since = "1.0.0")] . #[rustc_const_stable(feature = "const_checked_int_methods", since = "1.47.0")] . #[must_use = "this returns the result of the operation, \ . without modifying the original"] . #[inline] 28,960 ( 0.00%) pub const fn checked_add(self, rhs: Self) -> Option { 23,180 ( 0.00%) let (a, b) = self.overflowing_add(rhs); 57,920 ( 0.00%) if unlikely!(b) {None} else {Some(a)} 23,168 ( 0.00%) } . . /// Unchecked integer addition. Computes `self + rhs`, assuming overflow . /// cannot occur. . /// . /// # Safety . /// . /// This results in undefined behavior when . #[doc = concat!("`self + rhs > ", stringify!($SelfT), "::MAX` or `self + rhs < ", stringify!($SelfT), "::MIN`,")] -- line 448 ---------------------------------------- -- line 456 ---------------------------------------- . )] . #[must_use = "this returns the result of the operation, \ . without modifying the original"] . #[rustc_const_unstable(feature = "const_inherent_unchecked_arith", issue = "85122")] . #[inline(always)] . pub const unsafe fn unchecked_add(self, rhs: Self) -> Self { . // SAFETY: the caller must uphold the safety contract for . // `unchecked_add`. 557,197 ( 0.02%) unsafe { intrinsics::unchecked_add(self, rhs) } . } . . /// Checked addition with a signed integer. Computes `self + rhs`, . /// returning `None` if overflow occurred. . /// . /// # Examples . /// . /// Basic usage: -- line 472 ---------------------------------------- -- line 498 ---------------------------------------- . #[doc = concat!("assert_eq!(1", stringify!($SelfT), ".checked_sub(1), Some(0));")] . #[doc = concat!("assert_eq!(0", stringify!($SelfT), ".checked_sub(1), None);")] . /// ``` . #[stable(feature = "rust1", since = "1.0.0")] . #[rustc_const_stable(feature = "const_checked_int_methods", since = "1.47.0")] . #[must_use = "this returns the result of the operation, \ . without modifying the original"] . #[inline] 135 ( 0.00%) pub const fn checked_sub(self, rhs: Self) -> Option { 108 ( 0.00%) let (a, b) = self.overflowing_sub(rhs); 270 ( 0.00%) if unlikely!(b) {None} else {Some(a)} 108 ( 0.00%) } . . /// Unchecked integer subtraction. Computes `self - rhs`, assuming overflow . /// cannot occur. . /// . /// # Safety . /// . /// This results in undefined behavior when . #[doc = concat!("`self - rhs > ", stringify!($SelfT), "::MAX` or `self - rhs < ", stringify!($SelfT), "::MIN`,")] -- line 517 ---------------------------------------- -- line 525 ---------------------------------------- . )] . #[must_use = "this returns the result of the operation, \ . without modifying the original"] . #[rustc_const_unstable(feature = "const_inherent_unchecked_arith", issue = "85122")] . #[inline(always)] . pub const unsafe fn unchecked_sub(self, rhs: Self) -> Self { . // SAFETY: the caller must uphold the safety contract for . // `unchecked_sub`. 22,759 ( 0.00%) unsafe { intrinsics::unchecked_sub(self, rhs) } . } . . /// Checked integer multiplication. Computes `self * rhs`, returning . /// `None` if overflow occurred. . /// . /// # Examples . /// . /// Basic usage: -- line 541 ---------------------------------------- -- line 544 ---------------------------------------- . #[doc = concat!("assert_eq!(5", stringify!($SelfT), ".checked_mul(1), Some(5));")] . #[doc = concat!("assert_eq!(", stringify!($SelfT), "::MAX.checked_mul(2), None);")] . /// ``` . #[stable(feature = "rust1", since = "1.0.0")] . #[rustc_const_stable(feature = "const_checked_int_methods", since = "1.47.0")] . #[must_use = "this returns the result of the operation, \ . without modifying the original"] . #[inline] 41,010 ( 0.00%) pub const fn checked_mul(self, rhs: Self) -> Option { 27,340 ( 0.00%) let (a, b) = self.overflowing_mul(rhs); 68,350 ( 0.00%) if unlikely!(b) {None} else {Some(a)} 27,340 ( 0.00%) } . . /// Unchecked integer multiplication. Computes `self * rhs`, assuming overflow . /// cannot occur. . /// . /// # Safety . /// . /// This results in undefined behavior when . #[doc = concat!("`self * rhs > ", stringify!($SelfT), "::MAX` or `self * rhs < ", stringify!($SelfT), "::MIN`,")] -- line 563 ---------------------------------------- -- line 596 ---------------------------------------- . without modifying the original"] . #[inline] . pub const fn checked_div(self, rhs: Self) -> Option { . if unlikely!(rhs == 0) { . None . } else { . // SAFETY: div by zero has been checked above and unsigned types have no other . // failure modes for division 2,760 ( 0.00%) Some(unsafe { intrinsics::unchecked_div(self, rhs) }) . } . } . . /// Checked Euclidean division. Computes `self.div_euclid(rhs)`, returning `None` . /// if `rhs == 0`. . /// . /// # Examples . /// -- line 612 ---------------------------------------- -- line 1035 ---------------------------------------- . #[doc = concat!("assert_eq!(", stringify!($SelfT), "::MAX.saturating_add(127), ", stringify!($SelfT), "::MAX);")] . /// ``` . #[stable(feature = "rust1", since = "1.0.0")] . #[must_use = "this returns the result of the operation, \ . without modifying the original"] . #[rustc_const_stable(feature = "const_saturating_int_methods", since = "1.47.0")] . #[inline(always)] . pub const fn saturating_add(self, rhs: Self) -> Self { 96,440 ( 0.00%) intrinsics::saturating_add(self, rhs) 41 ( 0.00%) } . . /// Saturating addition with a signed integer. Computes `self + rhs`, . /// saturating at the numeric bounds instead of overflowing. . /// . /// # Examples . /// . /// Basic usage: . /// -- line 1052 ---------------------------------------- -- line 1084 ---------------------------------------- . #[doc = concat!("assert_eq!(13", stringify!($SelfT), ".saturating_sub(127), 0);")] . /// ``` . #[stable(feature = "rust1", since = "1.0.0")] . #[must_use = "this returns the result of the operation, \ . without modifying the original"] . #[rustc_const_stable(feature = "const_saturating_int_methods", since = "1.47.0")] . #[inline(always)] . pub const fn saturating_sub(self, rhs: Self) -> Self { 6,842 ( 0.00%) intrinsics::saturating_sub(self, rhs) . } . . /// Saturating integer multiplication. Computes `self * rhs`, . /// saturating at the numeric bounds instead of overflowing. . /// . /// # Examples . /// . /// Basic usage: -- line 1100 ---------------------------------------- -- line 1175 ---------------------------------------- . #[doc = concat!("assert_eq!(200", stringify!($SelfT), ".wrapping_add(", stringify!($SelfT), "::MAX), 199);")] . /// ``` . #[stable(feature = "rust1", since = "1.0.0")] . #[rustc_const_stable(feature = "const_wrapping_math", since = "1.32.0")] . #[must_use = "this returns the result of the operation, \ . without modifying the original"] . #[inline(always)] . pub const fn wrapping_add(self, rhs: Self) -> Self { 2,806,624 ( 0.11%) intrinsics::wrapping_add(self, rhs) . } . . /// Wrapping (modular) addition with a signed integer. Computes . /// `self + rhs`, wrapping around at the boundary of the type. . /// . /// # Examples . /// . /// Basic usage: -- line 1191 ---------------------------------------- -- line 1217 ---------------------------------------- . #[doc = concat!("assert_eq!(100", stringify!($SelfT), ".wrapping_sub(", stringify!($SelfT), "::MAX), 101);")] . /// ``` . #[stable(feature = "rust1", since = "1.0.0")] . #[rustc_const_stable(feature = "const_wrapping_math", since = "1.32.0")] . #[must_use = "this returns the result of the operation, \ . without modifying the original"] . #[inline(always)] . pub const fn wrapping_sub(self, rhs: Self) -> Self { 1,238,039 ( 0.05%) intrinsics::wrapping_sub(self, rhs) . } . . /// Wrapping (modular) multiplication. Computes `self * . /// rhs`, wrapping around at the boundary of the type. . /// . /// # Examples . /// . /// Basic usage: -- line 1233 ---------------------------------------- -- line 1240 ---------------------------------------- . /// assert_eq!(25u8.wrapping_mul(12), 44); . /// ``` . #[stable(feature = "rust1", since = "1.0.0")] . #[rustc_const_stable(feature = "const_wrapping_math", since = "1.32.0")] . #[must_use = "this returns the result of the operation, \ . without modifying the original"] . #[inline(always)] . pub const fn wrapping_mul(self, rhs: Self) -> Self { 3,501,330 ( 0.14%) intrinsics::wrapping_mul(self, rhs) . } . . /// Wrapping (modular) division. Computes `self / rhs`. . /// Wrapped division on unsigned types is just normal division. . /// There's no way wrapping could ever happen. . /// This function exists, so that all operations . /// are accounted for in the wrapping operations. . /// -- line 1256 ---------------------------------------- -- line 1397 ---------------------------------------- . #[rustc_const_stable(feature = "const_wrapping_math", since = "1.32.0")] . #[must_use = "this returns the result of the operation, \ . without modifying the original"] . #[inline(always)] . pub const fn wrapping_shl(self, rhs: u32) -> Self { . // SAFETY: the masking by the bitsize of the type ensures that we do not shift . // out of bounds . unsafe { 8 ( 0.00%) intrinsics::unchecked_shl(self, (rhs & ($BITS - 1)) as $SelfT) . } . } . . /// Panic-free bitwise shift-right; yields `self >> mask(rhs)`, . /// where `mask` removes any high-order bits of `rhs` that . /// would cause the shift to exceed the bitwidth of the type. . /// . /// Note that this is *not* the same as a rotate-right; the -- line 1413 ---------------------------------------- -- line 1491 ---------------------------------------- . #[doc = concat!("assert_eq!(5", stringify!($SelfT), ".overflowing_add(2), (7, false));")] . #[doc = concat!("assert_eq!(", stringify!($SelfT), "::MAX.overflowing_add(1), (0, true));")] . /// ``` . #[stable(feature = "wrapping", since = "1.7.0")] . #[rustc_const_stable(feature = "const_wrapping_math", since = "1.32.0")] . #[must_use = "this returns the result of the operation, \ . without modifying the original"] . #[inline(always)] 28 ( 0.00%) pub const fn overflowing_add(self, rhs: Self) -> (Self, bool) { 498,441 ( 0.02%) let (a, b) = intrinsics::add_with_overflow(self as $ActualT, rhs as $ActualT); 11,584 ( 0.00%) (a as Self, b) 23,224 ( 0.00%) } . . /// Calculates `self + rhs + carry` without the ability to overflow. . /// . /// Performs "ternary addition" which takes in an extra bit to add, and may return an . /// additional bit of overflow. This allows for chaining together multiple additions . /// to create "big integers" which represent larger values. . /// . #[doc = concat!("This can be thought of as a ", stringify!($BITS), "-bit \"full adder\", in the electronics sense.")] -- line 1510 ---------------------------------------- -- line 1588 ---------------------------------------- . #[doc = concat!("assert_eq!(0", stringify!($SelfT), ".overflowing_sub(1), (", stringify!($SelfT), "::MAX, true));")] . /// ``` . #[stable(feature = "wrapping", since = "1.7.0")] . #[rustc_const_stable(feature = "const_wrapping_math", since = "1.32.0")] . #[must_use = "this returns the result of the operation, \ . without modifying the original"] . #[inline(always)] . pub const fn overflowing_sub(self, rhs: Self) -> (Self, bool) { 121,178 ( 0.00%) let (a, b) = intrinsics::sub_with_overflow(self as $ActualT, rhs as $ActualT); 54 ( 0.00%) (a as Self, b) 108 ( 0.00%) } . . /// Calculates `self - rhs - borrow` without the ability to overflow. . /// . /// Performs "ternary subtraction" which takes in an extra bit to subtract, and may return . /// an additional bit of overflow. This allows for chaining together multiple subtractions . /// to create "big integers" which represent larger values. . /// . /// # Examples -- line 1606 ---------------------------------------- -- line 1674 ---------------------------------------- . /// assert_eq!(1_000_000_000u32.overflowing_mul(10), (1410065408, true)); . /// ``` . #[stable(feature = "wrapping", since = "1.7.0")] . #[rustc_const_stable(feature = "const_wrapping_math", since = "1.32.0")] . #[must_use = "this returns the result of the operation, \ . without modifying the original"] . #[inline(always)] . pub const fn overflowing_mul(self, rhs: Self) -> (Self, bool) { 817,637 ( 0.03%) let (a, b) = intrinsics::mul_with_overflow(self as $ActualT, rhs as $ActualT); 13,670 ( 0.00%) (a as Self, b) 27,340 ( 0.00%) } . . /// Calculates the divisor when `self` is divided by `rhs`. . /// . /// Returns a tuple of the divisor along with a boolean indicating . /// whether an arithmetic overflow would occur. Note that for unsigned . /// integers overflow never occurs, so the second value is always . /// `false`. . /// -- line 1692 ---------------------------------------- -- line 2132 ---------------------------------------- . #[doc = concat!("assert!(16", stringify!($SelfT), ".is_power_of_two());")] . #[doc = concat!("assert!(!10", stringify!($SelfT), ".is_power_of_two());")] . /// ``` . #[must_use] . #[stable(feature = "rust1", since = "1.0.0")] . #[rustc_const_stable(feature = "const_is_power_of_two", since = "1.32.0")] . #[inline(always)] . pub const fn is_power_of_two(self) -> bool { 339 ( 0.00%) self.count_ones() == 1 . } . . // Returns one less than next power of two. . // (For 8u8 next power of two is 8u8 and for 6u8 it is 8u8) . // . // 8u8.one_less_than_next_power_of_two() == 7 . // 6u8.one_less_than_next_power_of_two() == 7 . // . // This method cannot overflow, as in the `next_power_of_two` . // overflow cases it instead ends up returning the maximum value . // of the type, and can return 0 for 0. . #[inline] . #[rustc_const_stable(feature = "const_int_pow", since = "1.50.0")] 51 ( 0.00%) const fn one_less_than_next_power_of_two(self) -> Self { 7,990 ( 0.00%) if self <= 1 { return 0; } . 17,158 ( 0.00%) let p = self - 1; . // SAFETY: Because `p > 0`, it cannot consist entirely of leading zeros. . // That means the shift is always in-bounds, and some processors . // (such as intel pre-haswell) have more efficient ctlz . // intrinsics when the argument is non-zero. 51,175 ( 0.00%) let z = unsafe { intrinsics::ctlz_nonzero(p) }; 17,207 ( 0.00%) <$SelfT>::MAX >> z 68 ( 0.00%) } . . /// Returns the smallest power of two greater than or equal to `self`. . /// . /// When return value overflows (i.e., `self > (1 << (N-1))` for type . /// `uN`), it panics in debug mode and the return value is wrapped to 0 in . /// release mode (the only situation in which method can return 0). . /// . /// # Examples -- line 2172 ---------------------------------------- -- line 2178 ---------------------------------------- . #[doc = concat!("assert_eq!(3", stringify!($SelfT), ".next_power_of_two(), 4);")] . /// ``` . #[stable(feature = "rust1", since = "1.0.0")] . #[rustc_const_stable(feature = "const_int_pow", since = "1.50.0")] . #[must_use = "this returns the result of the operation, \ . without modifying the original"] . #[inline] . #[rustc_inherit_overflow_checks] 34 ( 0.00%) pub const fn next_power_of_two(self) -> Self { 26,201 ( 0.00%) self.one_less_than_next_power_of_two() + 1 34 ( 0.00%) } . . /// Returns the smallest power of two greater than or equal to `n`. If . /// the next power of two is greater than the type's maximum value, . /// `None` is returned, otherwise the power of two is wrapped in `Some`. . /// . /// # Examples . /// . /// Basic usage: -- line 2196 ---------------------------------------- -- line 2264 ---------------------------------------- . #[doc = concat!("let bytes = ", $swap_op, stringify!($SelfT), ".to_le_bytes();")] . #[doc = concat!("assert_eq!(bytes, ", $le_bytes, ");")] . /// ``` . #[stable(feature = "int_to_from_bytes", since = "1.32.0")] . #[rustc_const_stable(feature = "const_int_conversion", since = "1.44.0")] . #[must_use = "this returns the result of the operation, \ . without modifying the original"] . #[inline] 75 ( 0.00%) pub const fn to_le_bytes(self) -> [u8; mem::size_of::()] { 60 ( 0.00%) self.to_le().to_ne_bytes() 45 ( 0.00%) } . . /// Return the memory representation of this integer as a byte array in . /// native byte order. . /// . /// As the target platform's native endianness is used, portable code . /// should use [`to_be_bytes`] or [`to_le_bytes`], as appropriate, . /// instead. . /// -- line 2282 ---------------------------------------- -- line 2300 ---------------------------------------- . /// ``` . #[stable(feature = "int_to_from_bytes", since = "1.32.0")] . #[rustc_const_stable(feature = "const_int_conversion", since = "1.44.0")] . #[must_use = "this returns the result of the operation, \ . without modifying the original"] . // SAFETY: const sound because integers are plain old datatypes so we can always . // transmute them to arrays of bytes . #[inline] 198 ( 0.00%) pub const fn to_ne_bytes(self) -> [u8; mem::size_of::()] { . // SAFETY: integers are plain old datatypes so we can always transmute them to . // arrays of bytes 99 ( 0.00%) unsafe { mem::transmute(self) } 297 ( 0.00%) } . . /// Create a native endian integer value from its representation . /// as a byte array in big endian. . /// . #[doc = $from_xe_bytes_doc] . /// . /// # Examples . /// -- line 2320 ---------------------------------------- -- line 2360 ---------------------------------------- . /// *input = rest; . #[doc = concat!(" ", stringify!($SelfT), "::from_le_bytes(int_bytes.try_into().unwrap())")] . /// } . /// ``` . #[stable(feature = "int_to_from_bytes", since = "1.32.0")] . #[rustc_const_stable(feature = "const_int_conversion", since = "1.44.0")] . #[must_use] . #[inline] 120 ( 0.00%) pub const fn from_le_bytes(bytes: [u8; mem::size_of::()]) -> Self { 150 ( 0.00%) Self::from_le(Self::from_ne_bytes(bytes)) 60 ( 0.00%) } . . /// Create a native endian integer value from its memory representation . /// as a byte array in native endianness. . /// . /// As the target platform's native endianness is used, portable code . /// likely wants to use [`from_be_bytes`] or [`from_le_bytes`], as . /// appropriate instead. . /// -- line 2378 ---------------------------------------- -- line 2402 ---------------------------------------- . /// } . /// ``` . #[stable(feature = "int_to_from_bytes", since = "1.32.0")] . #[rustc_const_stable(feature = "const_int_conversion", since = "1.44.0")] . #[must_use] . // SAFETY: const sound because integers are plain old datatypes so we can always . // transmute to them . #[inline] 120 ( 0.00%) pub const fn from_ne_bytes(bytes: [u8; mem::size_of::()]) -> Self { . // SAFETY: integers are plain old datatypes so we can always transmute to them 180 ( 0.00%) unsafe { mem::transmute(bytes) } 60 ( 0.00%) } . . /// New code should prefer to use . #[doc = concat!("[`", stringify!($SelfT), "::MIN", "`] instead.")] . /// . /// Returns the smallest value that can be represented by this integer type. . #[stable(feature = "rust1", since = "1.0.0")] . #[rustc_promotable] . #[inline(always)] -- line 2421 ---------------------------------------- 123,413 ( 0.00%) -------------------------------------------------------------------------------- -- Auto-annotated source: /usr/home/liquid/rust/worktree-benchmarking/library/proc_macro/src/bridge/client.rs -------------------------------------------------------------------------------- Ir -- line 18 ---------------------------------------- . // FIXME(eddyb) use a reference to the `static COUNTERS`, instead of . // a wrapper `fn` pointer, once `const fn` can reference `static`s. . extern "C" fn get() -> &'static Self { . static COUNTERS: HandleCounters = HandleCounters { . $($oty: AtomicUsize::new(1),)* . $($ity: AtomicUsize::new(1),)* . }; . &COUNTERS 30 ( 0.00%) } . } . . // FIXME(eddyb) generate the definition of `HandleStore` in `server.rs`. . #[repr(C)] . #[allow(non_snake_case)] . pub(super) struct HandleStore { . $($oty: handle::OwnedStore,)* . $($ity: handle::InternedStore,)* . } . . impl HandleStore { 90 ( 0.00%) pub(super) fn new(handle_counters: &'static HandleCounters) -> Self { 885 ( 0.00%) HandleStore { . $($oty: handle::OwnedStore::new(&handle_counters.$oty),)* . $($ity: handle::InternedStore::new(&handle_counters.$ity),)* . } 120 ( 0.00%) } . } . . $( . #[repr(C)] . pub(crate) struct $oty(handle::Handle); . impl !Send for $oty {} . impl !Sync for $oty {} . . // Forward `Drop::drop` to the inherent `drop` method. . impl Drop for $oty { 1,212 ( 0.00%) fn drop(&mut self) { 10,204 ( 0.00%) $oty(self.0).drop(); 1,818 ( 0.00%) } . } . . impl Encode for $oty { 120 ( 0.00%) fn encode(self, w: &mut Writer, s: &mut S) { 45 ( 0.00%) let handle = self.0; 60 ( 0.00%) mem::forget(self); 15 ( 0.00%) handle.encode(w, s); 30 ( 0.00%) } . } . . impl DecodeMut<'_, '_, HandleStore>> . for Marked . { 76,065 ( 0.00%) fn decode(r: &mut Reader<'_>, s: &mut HandleStore>) -> Self { . s.$oty.take(handle::Handle::decode(r, &mut ())) 101,420 ( 0.00%) } . } . . impl Encode for &$oty { . fn encode(self, w: &mut Writer, s: &mut S) { 28,050 ( 0.00%) self.0.encode(w, s); . } . } . . impl<'s, S: server::Types> Decode<'_, 's, HandleStore>> . for &'s Marked . { 9,350 ( 0.00%) fn decode(r: &mut Reader<'_>, s: &'s HandleStore>) -> Self { . &s.$oty[handle::Handle::decode(r, &mut ())] 18,700 ( 0.00%) } . } . . impl Encode for &mut $oty { . fn encode(self, w: &mut Writer, s: &mut S) { 51,453 ( 0.00%) self.0.encode(w, s); . } . } . . impl<'s, S: server::Types> DecodeMut<'_, 's, HandleStore>> . for &'s mut Marked . { 17,151 ( 0.00%) fn decode( . r: &mut Reader<'_>, . s: &'s mut HandleStore> . ) -> Self { . &mut s.$oty[handle::Handle::decode(r, &mut ())] 34,302 ( 0.00%) } . } . . impl Encode>> . for Marked . { 152,213 ( 0.01%) fn encode(self, w: &mut Writer, s: &mut HandleStore>) { 26,234 ( 0.00%) s.$oty.alloc(self).encode(w, s); . } . } . . impl DecodeMut<'_, '_, S> for $oty { 90 ( 0.00%) fn decode(r: &mut Reader<'_>, s: &mut S) -> Self { 90 ( 0.00%) $oty(handle::Handle::decode(r, s)) 90 ( 0.00%) } . } . )* . . $( . #[repr(C)] 58,849 ( 0.00%) #[derive(Copy, Clone, PartialEq, Eq, Hash)] . pub(crate) struct $ity(handle::Handle); . impl !Send for $ity {} . impl !Sync for $ity {} . . impl Encode for $ity { . fn encode(self, w: &mut Writer, s: &mut S) { . self.0.encode(w, s); . } . } . . impl DecodeMut<'_, '_, HandleStore>> . for Marked . { 16,019 ( 0.00%) fn decode(r: &mut Reader<'_>, s: &mut HandleStore>) -> Self { . s.$ity.copy(handle::Handle::decode(r, &mut ())) 32,038 ( 0.00%) } . } . . impl Encode>> . for Marked . { . fn encode(self, w: &mut Writer, s: &mut HandleStore>) { 135,443 ( 0.01%) s.$ity.alloc(self).encode(w, s); . } . } . . impl DecodeMut<'_, '_, S> for $ity { . fn decode(r: &mut Reader<'_>, s: &mut S) -> Self { . $ity(handle::Handle::decode(r, s)) . } . } -- line 154 ---------------------------------------- -- line 175 ---------------------------------------- . . // FIXME(eddyb) generate these impls by pattern-matching on the . // names of methods - also could use the presence of `fn drop` . // to distinguish between 'owned and 'interned, above. . // Alternatively, special 'modes" could be listed of types in with_api . // instead of pattern matching on methods, here and in server decl. . . impl Clone for TokenStream { 1,710 ( 0.00%) fn clone(&self) -> Self { . self.clone() 1,710 ( 0.00%) } . } . . impl Clone for TokenStreamIter { . fn clone(&self) -> Self { . self.clone() . } . } . . impl Clone for Group { 1,209 ( 0.00%) fn clone(&self) -> Self { . self.clone() 1,209 ( 0.00%) } . } . . impl Clone for Literal { 423 ( 0.00%) fn clone(&self) -> Self { . self.clone() 423 ( 0.00%) } . } . . impl fmt::Debug for Literal { . fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { . f.debug_struct("Literal") . // format the kind without quotes, as in `kind: Float` . .field("kind", &format_args!("{}", &self.debug_kind())) . .field("symbol", &self.symbol()) -- line 211 ---------------------------------------- -- line 228 ---------------------------------------- . } . } . . macro_rules! define_client_side { . ($($name:ident { . $(fn $method:ident($($arg:ident: $arg_ty:ty),* $(,)?) $(-> $ret_ty:ty)*;)* . }),* $(,)?) => { . $(impl $name { 88,062 ( 0.00%) $(pub(crate) fn $method($($arg: $arg_ty),*) $(-> $ret_ty)* { 45,904 ( 0.00%) Bridge::with(|bridge| { . let mut b = bridge.cached_buffer.take(); . . b.clear(); 269,493 ( 0.01%) api_tags::Method::$name(api_tags::$name::$method).encode(&mut b, &mut ()); 64,962 ( 0.00%) reverse_encode!(b; $($arg),*); . 538,986 ( 0.02%) b = bridge.dispatch.call(b); . 375,289 ( 0.01%) let r = Result::<_, PanicMessage>::decode(&mut &b[..], &mut ()); . 1,077,972 ( 0.04%) bridge.cached_buffer = b; . 138,173 ( 0.01%) r.unwrap_or_else(|e| panic::resume_unwind(e.into())) . }) 73,385 ( 0.00%) })* . })* . } . } . with_api!(self, self, define_client_side); . . enum BridgeState<'a> { . /// No server is currently connected to this client. . NotConnected, -- line 260 ---------------------------------------- -- line 270 ---------------------------------------- . enum BridgeStateL {} . . impl<'a> scoped_cell::ApplyL<'a> for BridgeStateL { . type Out = BridgeState<'a>; . } . . thread_local! { . static BRIDGE_STATE: scoped_cell::ScopedCell = 3 ( 0.00%) scoped_cell::ScopedCell::new(BridgeState::NotConnected); . } . . impl BridgeState<'_> { . /// Take exclusive control of the thread-local . /// `BridgeState`, and pass it to `f`, mutably. . /// The state will be restored after `f` exits, even . /// by panic, including modifications made to it by `f`. . /// . /// N.B., while `f` is running, the thread-local state . /// is `BridgeState::InUse`. 58,870 ( 0.00%) fn with(f: impl FnOnce(&mut BridgeState<'_>) -> R) -> R { 60,409 ( 0.00%) BRIDGE_STATE.with(|state| { 485,960 ( 0.02%) state.replace(BridgeState::InUse, |mut state| { . // FIXME(#52812) pass `f` directly to `replace` when `RefMutL` is gone 360 ( 0.00%) f(&mut *state) 60 ( 0.00%) }) 60 ( 0.00%) }) 49,863 ( 0.00%) } . } . . impl Bridge<'_> { . pub(crate) fn is_available() -> bool { 3 ( 0.00%) BridgeState::with(|state| match state { . BridgeState::Connected(_) | BridgeState::InUse => true, . BridgeState::NotConnected => false, . }) . } . 90 ( 0.00%) fn enter(self, f: impl FnOnce() -> R) -> R { 105 ( 0.00%) let force_show_panics = self.force_show_panics; . // Hide the default panic output within `proc_macro` expansions. . // NB. the server can't do this because it may use a different libstd. . static HIDE_PANICS_DURING_EXPANSION: Once = Once::new(); 77 ( 0.00%) HIDE_PANICS_DURING_EXPANSION.call_once(|| { 7 ( 0.00%) panic::update_hook(move |prev, info| { . let show = BridgeState::with(|state| match state { . BridgeState::NotConnected => true, . BridgeState::Connected(_) | BridgeState::InUse => force_show_panics, . }); . if show { . prev(info) . } . }); 2 ( 0.00%) }); . 555 ( 0.00%) BRIDGE_STATE.with(|state| state.set(BridgeState::Connected(self), f)) 30 ( 0.00%) } . 75 ( 0.00%) fn with(f: impl FnOnce(&mut Bridge<'_>) -> R) -> R { 638,484 ( 0.03%) BridgeState::with(|state| match state { . BridgeState::NotConnected => { . panic!("procedural macro API is used outside of a procedural macro"); . } . BridgeState::InUse => { . panic!("procedural macro API is used while it's already in use"); . } 195 ( 0.00%) BridgeState::Connected(bridge) => f(bridge), 60 ( 0.00%) }) 60 ( 0.00%) } . } . . /// A client-side "global object" (usually a function pointer), . /// which may be using a different `proc_macro` from the one . /// used by the server, but can be interacted with compatibly. . /// . /// N.B., `F` must have FFI-friendly memory layout (e.g., a pointer). . /// The call ABI of function pointers used for `F` doesn't -- line 345 ---------------------------------------- -- line 353 ---------------------------------------- . pub(super) get_handle_counters: extern "C" fn() -> &'static HandleCounters, . pub(super) run: extern "C" fn(Bridge<'_>, F) -> Buffer, . pub(super) f: F, . } . . /// Client-side helper for handling client panics, entering the bridge, . /// deserializing input and serializing output. . // FIXME(eddyb) maybe replace `Bridge::enter` with this? 105 ( 0.00%) fn run_client DecodeMut<'a, 's, ()>, R: Encode<()>>( . mut bridge: Bridge<'_>, . f: impl FnOnce(A) -> R, . ) -> Buffer { . // The initial `cached_buffer` contains the input. 90 ( 0.00%) let mut b = bridge.cached_buffer.take(); . 540 ( 0.00%) panic::catch_unwind(panic::AssertUnwindSafe(|| { 225 ( 0.00%) bridge.enter(|| { 285 ( 0.00%) let reader = &mut &b[..]; 105 ( 0.00%) let input = A::decode(reader, &mut ()); . . // Put the `cached_buffer` back in the `Bridge`, for requests. 300 ( 0.00%) Bridge::with(|bridge| bridge.cached_buffer = b.take()); . 240 ( 0.00%) let output = f(input); . . // Take the `cached_buffer` back out, for the output value. 285 ( 0.00%) b = Bridge::with(|bridge| bridge.cached_buffer.take()); . . // HACK(eddyb) Separate encoding a success value (`Ok(output)`) . // from encoding a panic (`Err(e: PanicMessage)`) to avoid . // having handles outside the `bridge.enter(|| ...)` scope, and . // to catch panics that could happen while encoding the success. . // . // Note that panics should be impossible beyond this point, but . // this is defensively trying to avoid any accidental panicking . // reaching the `extern "C"` (which should `abort` but might not . // at the moment, so this is also potentially preventing UB). 45 ( 0.00%) b.clear(); 150 ( 0.00%) Ok::<_, ()>(output).encode(&mut b, &mut ()); 60 ( 0.00%) }) 30 ( 0.00%) })) . .map_err(PanicMessage::from) 15 ( 0.00%) .unwrap_or_else(|e| { . b.clear(); . Err::<(), _>(e).encode(&mut b, &mut ()); . }); . b 30 ( 0.00%) } . . impl Client crate::TokenStream> { . #[rustc_allow_const_fn_unstable(const_fn)] 1 ( 0.00%) pub const fn expand1(f: fn(crate::TokenStream) -> crate::TokenStream) -> Self { . extern "C" fn run( . bridge: Bridge<'_>, . f: impl FnOnce(crate::TokenStream) -> crate::TokenStream, . ) -> Buffer { . run_client(bridge, |input| f(crate::TokenStream(input)).0) . } 5 ( 0.00%) Client { get_handle_counters: HandleCounters::get, run, f } 1 ( 0.00%) } . } . . impl Client crate::TokenStream> { . #[rustc_allow_const_fn_unstable(const_fn)] . pub const fn expand2( . f: fn(crate::TokenStream, crate::TokenStream) -> crate::TokenStream, . ) -> Self { 105 ( 0.00%) extern "C" fn run( . bridge: Bridge<'_>, . f: impl FnOnce(crate::TokenStream, crate::TokenStream) -> crate::TokenStream, . ) -> Buffer { 270 ( 0.00%) run_client(bridge, |(input, input2)| { 300 ( 0.00%) f(crate::TokenStream(input), crate::TokenStream(input2)).0 45 ( 0.00%) }) 30 ( 0.00%) } . Client { get_handle_counters: HandleCounters::get, run, f } . } . } . . #[repr(C)] . #[derive(Copy, Clone)] . pub enum ProcMacro { . CustomDerive { -- line 435 ---------------------------------------- 241,162 ( 0.01%) -------------------------------------------------------------------------------- -- Auto-annotated source: /usr/home/liquid/rust/worktree-benchmarking/library/core/src/hash/sip.rs -------------------------------------------------------------------------------- Ir -- line 120 ---------------------------------------- . } . . /// Loads a u64 using up to 7 bytes of a byte slice. It looks clumsy but the . /// `copy_nonoverlapping` calls that occur (via `load_int_le!`) all have fixed . /// sizes and avoid calling `memcpy`, which is good for speed. . /// . /// Unsafe because: unchecked indexing at start..start+len . #[inline] 2,439 ( 0.00%) unsafe fn u8to64_le(buf: &[u8], start: usize, len: usize) -> u64 { . debug_assert!(len < 8); 271 ( 0.00%) let mut i = 0; // current byte index (from LSB) in the output u64 271 ( 0.00%) let mut out = 0; 292,241 ( 0.01%) if i + 3 < len { . // SAFETY: `i` cannot be greater than `len`, and the caller must guarantee . // that the index start..start+len is in bounds. 1,404 ( 0.00%) out = unsafe { load_int_le!(buf, start + i, u32) } as u64; 546 ( 0.00%) i += 4; . } 608,257 ( 0.02%) if i + 1 < len { . // SAFETY: same as above. 304,152 ( 0.01%) out |= (unsafe { load_int_le!(buf, start + i, u16) } as u64) << (i * 8); 151,464 ( 0.01%) i += 2 . } 339,264 ( 0.01%) if i < len { . // SAFETY: same as above. 472,962 ( 0.02%) out |= (unsafe { *buf.get_unchecked(start + i) } as u64) << (i * 8); 1,000 ( 0.00%) i += 1; . } . debug_assert_eq!(i, len); . out 813 ( 0.00%) } . . impl SipHasher { . /// Creates a new `SipHasher` with the two initial keys set to 0. . #[inline] . #[stable(feature = "rust1", since = "1.0.0")] . #[rustc_deprecated( . since = "1.13.0", . reason = "use `std::collections::hash_map::DefaultHasher` instead" -- line 158 ---------------------------------------- -- line 189 ---------------------------------------- . . /// Creates a `SipHasher13` that is keyed off the provided keys. . #[inline] . #[unstable(feature = "hashmap_internals", issue = "none")] . #[rustc_deprecated( . since = "1.13.0", . reason = "use `std::collections::hash_map::DefaultHasher` instead" . )] 235 ( 0.00%) pub fn new_with_keys(key0: u64, key1: u64) -> SipHasher13 { 59,001 ( 0.00%) SipHasher13 { hasher: Hasher::new_with_keys(key0, key1) } 94 ( 0.00%) } . } . . impl Hasher { . #[inline] 282 ( 0.00%) fn new_with_keys(key0: u64, key1: u64) -> Hasher { 758 ( 0.00%) let mut state = Hasher { . k0: key0, . k1: key1, . length: 0, 188 ( 0.00%) state: State { v0: 0, v1: 0, v2: 0, v3: 0 }, . tail: 0, . ntail: 0, . _marker: PhantomData, . }; 47 ( 0.00%) state.reset(); . state 94 ( 0.00%) } . . #[inline] 94 ( 0.00%) fn reset(&mut self) { 47 ( 0.00%) self.length = 0; 88,146 ( 0.00%) self.state.v0 = self.k0 ^ 0x736f6d6570736575; 94,629 ( 0.00%) self.state.v1 = self.k1 ^ 0x646f72616e646f6d; 141 ( 0.00%) self.state.v2 = self.k0 ^ 0x6c7967656e657261; 141 ( 0.00%) self.state.v3 = self.k1 ^ 0x7465646279746573; 47 ( 0.00%) self.ntail = 0; 94 ( 0.00%) } . } . . #[stable(feature = "rust1", since = "1.0.0")] . impl super::Hasher for SipHasher { . #[inline] . fn write(&mut self, msg: &[u8]) { . self.0.hasher.write(msg) . } -- line 234 ---------------------------------------- -- line 237 ---------------------------------------- . fn finish(&self) -> u64 { . self.0.hasher.finish() . } . } . . #[unstable(feature = "hashmap_internals", issue = "none")] . impl super::Hasher for SipHasher13 { . #[inline] 968 ( 0.00%) fn write(&mut self, msg: &[u8]) { 242 ( 0.00%) self.hasher.write(msg) 484 ( 0.00%) } . . #[inline] 186 ( 0.00%) fn finish(&self) -> u64 { 186 ( 0.00%) self.hasher.finish() 186 ( 0.00%) } . } . . impl super::Hasher for Hasher { . // Note: no integer hashing methods (`write_u*`, `write_i*`) are defined . // for this type. We could add them, copy the `short_write` implementation . // in librustc_data_structures/sip128.rs, and add `write_u*`/`write_i*` . // methods to `SipHasher`, `SipHasher13`, and `DefaultHasher`. This would . // greatly speed up integer hashing by those hashers, at the cost of . // slightly slowing down compile speeds on some benchmarks. See #69152 for . // details. . #[inline] 1,694 ( 0.00%) fn write(&mut self, msg: &[u8]) { 242 ( 0.00%) let length = msg.len(); 373,118 ( 0.01%) self.length += length; . 242 ( 0.00%) let mut needed = 0; . 365,898 ( 0.01%) if self.ntail != 0 { 185,292 ( 0.01%) needed = 8 - self.ntail; . // SAFETY: `cmp::min(length, needed)` is guaranteed to not be over `length` 925,668 ( 0.04%) self.tail |= unsafe { u8to64_le(msg, 0, cmp::min(length, needed)) } << (8 * self.ntail); 184,843 ( 0.01%) if length < needed { 125,728 ( 0.00%) self.ntail += length; . return; . } else { 29,731 ( 0.00%) self.state.v3 ^= self.tail; 87 ( 0.00%) S::c_rounds(&mut self.state); 59,375 ( 0.00%) self.state.v0 ^= self.tail; 29,673 ( 0.00%) self.ntail = 0; . } . } . . // Buffered tail is now flushed, process new input. 106,195 ( 0.00%) let len = length - needed; 105,675 ( 0.00%) let left = len & 0x7; // len % 8 . 354 ( 0.00%) let mut i = needed; 162,135 ( 0.01%) while i < len - left { . // SAFETY: because `len - left` is the biggest multiple of 8 under . // `len`, and because `i` starts at `needed` where `len` is `length - needed`, . // `i + 8` is guaranteed to be less than or equal to `length`. 3,424 ( 0.00%) let mi = unsafe { load_int_le!(msg, i, u64) }; . 594 ( 0.00%) self.state.v3 ^= mi; 428 ( 0.00%) S::c_rounds(&mut self.state); 594 ( 0.00%) self.state.v0 ^= mi; . 1,830 ( 0.00%) i += 8; . } . . // SAFETY: `i` is now `needed + len.div_euclid(8) * 8`, . // so `i + left` = `needed + len` = `length`, which is by . // definition equal to `msg.len()`. 112,501 ( 0.00%) self.tail = unsafe { u8to64_le(msg, i, left) }; 177 ( 0.00%) self.ntail = left; 726 ( 0.00%) } . . #[inline] 279 ( 0.00%) fn finish(&self) -> u64 { 30,082 ( 0.00%) let mut state = self.state; . 147,341 ( 0.01%) let b: u64 = ((self.length as u64 & 0xff) << 56) | self.tail; . 29,524 ( 0.00%) state.v3 ^= b; 186 ( 0.00%) S::c_rounds(&mut state); 29,524 ( 0.00%) state.v0 ^= b; . 29,617 ( 0.00%) state.v2 ^= 0xff; 186 ( 0.00%) S::d_rounds(&mut state); . 59,048 ( 0.00%) state.v0 ^ state.v1 ^ state.v2 ^ state.v3 186 ( 0.00%) } . } . . impl Clone for Hasher { . #[inline] . fn clone(&self) -> Hasher { . Hasher { . k0: self.k0, . k1: self.k1, -- line 332 ---------------------------------------- -- line 353 ---------------------------------------- . fn d_rounds(_: &mut State); . } . . #[derive(Debug, Clone, Default)] . struct Sip13Rounds; . . impl Sip for Sip13Rounds { . #[inline] 1,008 ( 0.00%) fn c_rounds(state: &mut State) { 403,868 ( 0.02%) compress!(state); 672 ( 0.00%) } . . #[inline] 279 ( 0.00%) fn d_rounds(state: &mut State) { 122,560 ( 0.00%) compress!(state); 122,560 ( 0.00%) compress!(state); 93,222 ( 0.00%) compress!(state); 186 ( 0.00%) } . } . . #[derive(Debug, Clone, Default)] . struct Sip24Rounds; . . impl Sip for Sip24Rounds { . #[inline] . fn c_rounds(state: &mut State) { -- line 378 ---------------------------------------- 872,966 ( 0.03%) -------------------------------------------------------------------------------- The following files chosen for auto-annotation could not be found: -------------------------------------------------------------------------------- ./elf/../sysdeps/x86_64/dl-machine.h ./elf/dl-lookup.c ./elf/do-rel.h ./malloc/malloc.c ./nptl/pthread_rwlock_common.c ./stdlib/msort.c ./string/../sysdeps/x86_64/multiarch/memcmp-avx2-movbe.S ./string/../sysdeps/x86_64/multiarch/memmove-vec-unaligned-erms.S ./string/../sysdeps/x86_64/multiarch/memset-vec-unaligned-erms.S ./string/../sysdeps/x86_64/multiarch/strcmp-avx2.S /tmp/gcc-build/x86_64-unknown-linux-gnu/libstdc++-v3/libsupc++/../../../../gcc-5.5.0/libstdc++-v3/libsupc++/new_op.cc -------------------------------------------------------------------------------- Ir -------------------------------------------------------------------------------- 149,051,158 ( 5.90%) events annotated