rustc -C opt-level=3 generates bad assembly code for `Vec` by default

Created on 9 Sep 2017  Â·  28Comments  Â·  Source: rust-lang/rust

Compiling the following C++ snippet with clang++ -O3 and g++ -O3 (see here):

#include <vector>

unsigned foo() {
  std::vector<unsigned> a;
  a.push_back(314);
  return a[0];
}

generates this assembly on x86_64:

foo(): # @foo()
  push rax
  mov edi, 4
  call operator new(unsigned long)
  mov rdi, rax
  call operator delete(void*)
  mov eax, 314
  pop rcx
  ret

(note: clang generates perfect assembly even with multiple push backs, the only thing that seems to trip it is a reallocation)

This snippet compiled with rustc --C opt-level=3 (see here):

pub fn foo() -> u32 {
  let mut v: Vec<u32> = Vec::new();
  v.push(0);
  v[0]
}

generates the following assembly:

<alloc::raw_vec::RawVec<T, A>>::double:
        push    rbp
        mov     rbp, rsp
        push    r14
        push    rbx
        sub     rsp, 64
        mov     r14, rdi
        mov     rbx, qword ptr [r14 + 8]
        test    rbx, rbx
        je      .LBB0_6
        lea     rsi, [4*rbx]
        lea     rcx, [8*rbx]
        mov     rdi, qword ptr [r14]
        lea     r9, [rbp - 40]
        mov     edx, 4
        mov     r8d, 4
        call    __rust_realloc@PLT
        test    rax, rax
        je      .LBB0_4
        add     rbx, rbx
        jmp     .LBB0_3
.LBB0_6:
        lea     rdx, [rbp - 40]
        mov     edi, 16
        mov     esi, 4
        call    __rust_alloc@PLT
        test    rax, rax
        je      .LBB0_8
        mov     ebx, 4
.LBB0_3:
        mov     qword ptr [r14], rax
        mov     qword ptr [r14 + 8], rbx
        add     rsp, 64
        pop     rbx
        pop     r14
        pop     rbp
        ret
.LBB0_4:
        mov     rax, qword ptr [rbp - 40]
        movups  xmm0, xmmword ptr [rbp - 32]
        movaps  xmmword ptr [rbp - 64], xmm0
        mov     qword ptr [rbp - 40], rax
        movaps  xmm0, xmmword ptr [rbp - 64]
        jmp     .LBB0_5
.LBB0_8:
        movups  xmm0, xmmword ptr [rbp - 32]
        movaps  xmmword ptr [rbp - 64], xmm0
        movaps  xmm0, xmmword ptr [rbp - 64]
        movaps  xmmword ptr [rbp - 80], xmm0
        movaps  xmm0, xmmword ptr [rbp - 80]
.LBB0_5:
        movups  xmmword ptr [rbp - 32], xmm0
        lea     rdi, [rbp - 40]
        call    <alloc::heap::Heap as alloc::allocator::Alloc>::oom
core::ptr::drop_in_place:
        push    rbp
        mov     rbp, rsp
        mov     rsi, qword ptr [rdi + 8]
        test    rsi, rsi
        je      .LBB1_1
        mov     rdi, qword ptr [rdi]
        shl     rsi, 2
        mov     edx, 4
        pop     rbp
        jmp     __rust_dealloc@PLT
.LBB1_1:
        pop     rbp
        ret
<alloc::heap::Heap as alloc::allocator::Alloc>::oom:
        push    rbp
        mov     rbp, rsp
        sub     rsp, 32
        mov     rax, qword ptr [rdi + 16]
        mov     qword ptr [rbp - 16], rax
        movups  xmm0, xmmword ptr [rdi]
        movaps  xmmword ptr [rbp - 32], xmm0
        lea     rdi, [rbp - 32]
        call    __rust_oom@PLT

example::foo:
        push    rbp
        mov     rbp, rsp
        push    rbx
        sub     rsp, 24
        mov     qword ptr [rbp - 32], 4
        xorps   xmm0, xmm0
        movups  xmmword ptr [rbp - 24], xmm0
        lea     rdi, [rbp - 32]
        call    <alloc::raw_vec::RawVec<T, A>>::double
        mov     rdi, qword ptr [rbp - 32]
        mov     rax, qword ptr [rbp - 16]
        mov     dword ptr [rdi + 4*rax], 0
        inc     rax
        mov     qword ptr [rbp - 16], rax
        je      .LBB3_2
        mov     ebx, dword ptr [rdi]
        mov     rsi, qword ptr [rbp - 24]
        test    rsi, rsi
        je      .LBB3_6
        shl     rsi, 2
        mov     edx, 4
        call    __rust_dealloc@PLT
.LBB3_6:
        mov     eax, ebx
        add     rsp, 24
        pop     rbx
        pop     rbp
        ret
.LBB3_2:
        lea     rdi, [rip + panic_bounds_check_loc.2]
        xor     esi, esi
        xor     edx, edx
        call    core::panicking::panic_bounds_check@PLT
        mov     rbx, rax
        lea     rdi, [rbp - 32]
        call    core::ptr::drop_in_place
        mov     rdi, rbx
        call    _Unwind_Resume@PLT
GCC_except_table3:
        .byte   255
        .byte   155
        .asciz  "\234"
        .byte   3
        .byte   26
        .long   .Ltmp29-.Lfunc_begin3
        .long   .Ltmp32-.Ltmp29
        .long   .Ltmp33-.Lfunc_begin3
        .byte   0
        .long   .Ltmp32-.Lfunc_begin3
        .long   .Lfunc_end3-.Ltmp32
        .long   0
        .byte   0
str.1:
        .ascii  "/checkout/src/liballoc/vec.rs"
panic_bounds_check_loc.2:
        .quad   str.1
        .quad   29
        .long   1555
        .long   10
DW.ref.rust_eh_personality:
        .quad   rust_eh_personality

I've tried adding -lto and -C panic=abort to rustc without much luck. I've also tried replacing [0] with unsafe { *v.get_unchecked(0) } without any luck. The only thing that makes it generate good assembly is using Vec::with_capacity(N) (see here):

pub fn foo() -> u32 {
  let mut v: Vec<u32> = Vec::with_capacity(3);
  v.push(7);
  v.push(4);
  v[1]
}

generates

example::foo:
        push    rbp
        mov     rbp, rsp
        mov     eax, 4
        pop     rbp
        ret
A-codegen C-enhancement I-slow T-compiler

All 28 comments

This is the asm I am seeing:

foo:
    pushq   %rsi
    subq    $64, %rsp
    movq    $4, 40(%rsp)
    vxorps  %xmm0, %xmm0, %xmm0
    vmovups %xmm0, 48(%rsp)
    leaq    40(%rsp), %rcx
    callq   _ZN49_$LT$alloc..raw_vec..RawVec$LT$T$C$$u20$A$GT$$GT$6double17h82e5d6919d342c2fE
    movq    40(%rsp), %rcx
    movq    56(%rsp), %rax
    movl    $0, (%rcx,%rax,4)
    addq    $1, %rax
    movq    %rax, 56(%rsp)
    je  .LBB3_2
    movl    (%rcx), %esi
    movq    48(%rsp), %rdx
    testq   %rdx, %rdx
    je  .LBB3_8
    shlq    $2, %rdx
    movl    $4, %r8d
    callq   __rust_dealloc
.LBB3_8:
    movl    %esi, %eax
    addq    $64, %rsp
    popq    %rsi
    retq
.LBB3_2:
    leaq    panic_bounds_check_loc.2(%rip), %rcx
    xorl    %edx, %edx
    xorl    %r8d, %r8d
    callq   _ZN4core9panicking18panic_bounds_check17hbf9952bd7dce1212E
    ud2
.LBB3_4:
    movq    %rax, %rcx
    callq   rust_eh_unwind_resume
    ud2
.LBB3_9:
    movq    %rax, %rsi
    leaq    40(%rsp), %rcx
    callq   _ZN4core3ptr13drop_in_place17h731ac0064b2a28c1E
    movq    %rsi, %rcx
    callq   rust_eh_unwind_resume
    ud2

This is the asm I am seeing on the playground (link):

.text
    .intel_syntax noprefix
    .file   "playground.cgu-0.rs"
    .section    ".text.cold._ZN49_$LT$alloc..raw_vec..RawVec$LT$T$C$$u20$A$GT$$GT$6double17h1882b1047d8074ffE","ax",@progbits
    .p2align    4, 0x90
    .type   _ZN49_$LT$alloc..raw_vec..RawVec$LT$T$C$$u20$A$GT$$GT$6double17h1882b1047d8074ffE,@function
_ZN49_$LT$alloc..raw_vec..RawVec$LT$T$C$$u20$A$GT$$GT$6double17h1882b1047d8074ffE:
    .cfi_startproc
    push    r14
.Lcfi0:
    .cfi_def_cfa_offset 16
    push    rbx
.Lcfi1:
    .cfi_def_cfa_offset 24
    sub rsp, 72
.Lcfi2:
    .cfi_def_cfa_offset 96
.Lcfi3:
    .cfi_offset rbx, -24
.Lcfi4:
    .cfi_offset r14, -16
    mov r14, rdi
    mov rbx, qword ptr [r14 + 8]
    test    rbx, rbx
    je  .LBB0_6
    lea rsi, [4*rbx]
    lea rcx, [8*rbx]
    mov rdi, qword ptr [r14]
    lea r9, [rsp + 8]
    mov edx, 4
    mov r8d, 4
    call    __rust_realloc@PLT
    test    rax, rax
    je  .LBB0_4
    add rbx, rbx
    jmp .LBB0_3
.LBB0_6:
    lea rdx, [rsp + 8]
    mov edi, 16
    mov esi, 4
    call    __rust_alloc@PLT
    test    rax, rax
    je  .LBB0_8
    mov ebx, 4
.LBB0_3:
    mov qword ptr [r14], rax
    mov qword ptr [r14 + 8], rbx
    add rsp, 72
    pop rbx
    pop r14
    ret
.LBB0_4:
    mov rax, qword ptr [rsp + 8]
    movups  xmm0, xmmword ptr [rsp + 16]
    movaps  xmmword ptr [rsp + 32], xmm0
    mov qword ptr [rsp + 8], rax
    movaps  xmm0, xmmword ptr [rsp + 32]
    jmp .LBB0_5
.LBB0_8:
    movups  xmm0, xmmword ptr [rsp + 16]
    movaps  xmmword ptr [rsp + 32], xmm0
    movaps  xmm0, xmmword ptr [rsp + 32]
    movaps  xmmword ptr [rsp + 48], xmm0
    movaps  xmm0, xmmword ptr [rsp + 48]
.LBB0_5:
    movups  xmmword ptr [rsp + 16], xmm0
    lea rdi, [rsp + 8]
    call    _ZN61_$LT$alloc..heap..Heap$u20$as$u20$alloc..allocator..Alloc$GT$3oom17h8b5f820f2dd0e667E
.Lfunc_end0:
    .size   _ZN49_$LT$alloc..raw_vec..RawVec$LT$T$C$$u20$A$GT$$GT$6double17h1882b1047d8074ffE, .Lfunc_end0-_ZN49_$LT$alloc..raw_vec..RawVec$LT$T$C$$u20$A$GT$$GT$6double17h1882b1047d8074ffE
    .cfi_endproc

    .section    .text._ZN4core3ptr13drop_in_place17h62d05f1818f0f9d8E,"ax",@progbits
    .p2align    4, 0x90
    .type   _ZN4core3ptr13drop_in_place17h62d05f1818f0f9d8E,@function
_ZN4core3ptr13drop_in_place17h62d05f1818f0f9d8E:
    .cfi_startproc
    mov rsi, qword ptr [rdi + 8]
    test    rsi, rsi
    je  .LBB1_1
    mov rdi, qword ptr [rdi]
    shl rsi, 2
    mov edx, 4
    jmp __rust_dealloc@PLT
.LBB1_1:
    ret
.Lfunc_end1:
    .size   _ZN4core3ptr13drop_in_place17h62d05f1818f0f9d8E, .Lfunc_end1-_ZN4core3ptr13drop_in_place17h62d05f1818f0f9d8E
    .cfi_endproc

    .section    ".text.cold._ZN61_$LT$alloc..heap..Heap$u20$as$u20$alloc..allocator..Alloc$GT$3oom17h8b5f820f2dd0e667E","ax",@progbits
    .p2align    4, 0x90
    .type   _ZN61_$LT$alloc..heap..Heap$u20$as$u20$alloc..allocator..Alloc$GT$3oom17h8b5f820f2dd0e667E,@function
_ZN61_$LT$alloc..heap..Heap$u20$as$u20$alloc..allocator..Alloc$GT$3oom17h8b5f820f2dd0e667E:
    .cfi_startproc
    sub rsp, 24
.Lcfi5:
    .cfi_def_cfa_offset 32
    mov rax, qword ptr [rdi + 16]
    mov qword ptr [rsp + 16], rax
    movups  xmm0, xmmword ptr [rdi]
    movaps  xmmword ptr [rsp], xmm0
    mov rdi, rsp
    call    __rust_oom@PLT
.Lfunc_end2:
    .size   _ZN61_$LT$alloc..heap..Heap$u20$as$u20$alloc..allocator..Alloc$GT$3oom17h8b5f820f2dd0e667E, .Lfunc_end2-_ZN61_$LT$alloc..heap..Heap$u20$as$u20$alloc..allocator..Alloc$GT$3oom17h8b5f820f2dd0e667E
    .cfi_endproc

    .section    .text._ZN10playground4main17hf21e17cd5a16d5d9E,"ax",@progbits
    .p2align    4, 0x90
    .type   _ZN10playground4main17hf21e17cd5a16d5d9E,@function
_ZN10playground4main17hf21e17cd5a16d5d9E:
.Lfunc_begin0:
    .cfi_startproc
    .cfi_personality 155, DW.ref.rust_eh_personality
    .cfi_lsda 27, .Lexception0
    push    rbx
.Lcfi6:
    .cfi_def_cfa_offset 16
    sub rsp, 32
.Lcfi7:
    .cfi_def_cfa_offset 48
.Lcfi8:
    .cfi_offset rbx, -16
    mov qword ptr [rsp + 8], 4
    xorps   xmm0, xmm0
    movups  xmmword ptr [rsp + 16], xmm0
.Ltmp0:
    lea rdi, [rsp + 8]
    call    _ZN49_$LT$alloc..raw_vec..RawVec$LT$T$C$$u20$A$GT$$GT$6double17h1882b1047d8074ffE
.Ltmp1:
    mov rdi, qword ptr [rsp + 8]
    mov rax, qword ptr [rsp + 24]
    mov dword ptr [rdi + 4*rax], 0
    inc rax
    mov qword ptr [rsp + 24], rax
    je  .LBB3_2
    mov rsi, qword ptr [rsp + 16]
    test    rsi, rsi
    je  .LBB3_6
    shl rsi, 2
    mov edx, 4
    call    __rust_dealloc@PLT
.LBB3_6:
    add rsp, 32
    pop rbx
    ret
.LBB3_2:
.Ltmp2:
    lea rdi, [rip + panic_bounds_check_loc.2]
    xor esi, esi
    xor edx, edx
    call    _ZN4core9panicking18panic_bounds_check17h78beadfd8229dc37E@PLT
.Ltmp3:
.LBB3_7:
.Ltmp4:
    mov rbx, rax
    lea rdi, [rsp + 8]
    call    _ZN4core3ptr13drop_in_place17h62d05f1818f0f9d8E
    mov rdi, rbx
    call    _Unwind_Resume@PLT
.Lfunc_end3:
    .size   _ZN10playground4main17hf21e17cd5a16d5d9E, .Lfunc_end3-_ZN10playground4main17hf21e17cd5a16d5d9E
    .cfi_endproc
    .section    .gcc_except_table,"a",@progbits
    .p2align    2
GCC_except_table3:
.Lexception0:
    .byte   255
    .byte   155
    .asciz  "\234"
    .byte   3
    .byte   26
    .long   .Ltmp0-.Lfunc_begin0
    .long   .Ltmp3-.Ltmp0
    .long   .Ltmp4-.Lfunc_begin0
    .byte   0
    .long   .Ltmp3-.Lfunc_begin0
    .long   .Lfunc_end3-.Ltmp3
    .long   0
    .byte   0
    .p2align    2

    .section    .text.main,"ax",@progbits
    .globl  main
    .p2align    4, 0x90
    .type   main,@function
main:
    .cfi_startproc
    mov rax, rsi
    mov rcx, rdi
    lea rdi, [rip + _ZN10playground4main17hf21e17cd5a16d5d9E]
    mov rsi, rcx
    mov rdx, rax
    jmp _ZN3std2rt10lang_start17h573cecb903a42a26E@PLT
.Lfunc_end4:
    .size   main, .Lfunc_end4-main
    .cfi_endproc

    .type   str.1,@object
    .section    .rodata.str.1,"a",@progbits
    .p2align    4
str.1:
    .ascii  "/checkout/src/liballoc/vec.rs"
    .size   str.1, 29

    .type   panic_bounds_check_loc.2,@object
    .section    .data.rel.ro.panic_bounds_check_loc.2,"aw",@progbits
    .p2align    3
panic_bounds_check_loc.2:
    .quad   str.1
    .quad   29
    .long   1555
    .long   10
    .size   panic_bounds_check_loc.2, 24

    .hidden DW.ref.rust_eh_personality
    .weak   DW.ref.rust_eh_personality
    .section    .data.DW.ref.rust_eh_personality,"aGw",@progbits,DW.ref.rust_eh_personality,comdat
    .p2align    3
    .type   DW.ref.rust_eh_personality,@object
    .size   DW.ref.rust_eh_personality, 8
DW.ref.rust_eh_personality:
    .quad   rust_eh_personality

    .section    ".note.GNU-stack","",@progbits

Not all of it is due to the vector code (it seems the play ground doesn't support creating an object file for code without main).

One difference between using with_capacity and Vec::new is that RawVec::double which is called by Vec::push is annotated with #[cold] and #[inline(never)] which will prevent the compiler from optimizing it out (at least without lto). Normally preventing it from being inlined is probably ideal, but this is a bit of an edge case. With with_capacity, the code around the allocation is inlined, and the compiler is able to figure out that it's redundant and also avoids the call to double entirely since the cap was just set larger than the length.

It looks like in the C++ version, manages optimize out the zeroing of length/capacity. The double call probably prevents this in the rust version. There is also an bounds check, not sure why this isn't optimized out. Maybe the compiler can't prove that length field (which is part of Vec, not RawVec) doesn't change in RawVec::double

My (possibly incorrect) attempt at labelling what's happening:

...

        ; Function prologue (save stack pointer and reserve space)
        push    rbp
        mov     rbp, rsp
        push    rbx
        sub     rsp, 24
        ; This sets the internal ptr to the (later) heap allocated data to 4 (The pointer is set to the alignment of the type initially if the vector is empty.)
        mov     qword ptr [rbp - 32], 4
        ; And this sets the both the capacity and length to 0 using vector instructions.
        xorps   xmm0, xmm0
        movups  xmmword ptr [rbp - 24], xmm0

        ; Load pointer to the RawVec to pass to the double function (I think)
        lea     rdi, [rbp - 32]
        ; Call to capacity doubling function.
        call    <alloc::raw_vec::RawVec<T, A>>::double
        ; Load ptr to heap memory into %rdi
        mov     rdi, qword ptr [rbp - 32]
        ; Load length into %rax
        mov     rax, qword ptr [rbp - 16]
        ; Set first element to 0 (I suppose it could be possible to avoid doing the 4*rax bit here since the length should always be 0 at this point.)
        mov     dword ptr [rdi + 4*rax], 0
        ; Increment length
        inc     rax
        ; Store the length back to the length field in the vector instance
        mov     qword ptr [rbp - 16], rax
        ; Jump to bounds check panic if incrementing length wraps to zero(I think)
        je      .LBB3_2
        ; Load value of the first element into %ebx
        mov     ebx, dword ptr [rdi]
        ; Load the capacity into %rsi
        mov     rsi, qword ptr [rbp - 24]
        ; Check if capacity == 0
        test    rsi, rsi
        ; Jump to end and skip deallocation if the capacity of the vector was 0.
        je      .LBB3_6
        ; Set up arguments to rust_dealloc (I think)
        shl     rsi, 2
        mov     edx, 4
        ; Deallocate vector memory
        call    __rust_dealloc@PLT
.LBB3_6:
        ; Load value from %eax into return register
        mov     eax, ebx
        ; Function epilogue
        add     rsp, 24
        pop     rbx
        pop     rbp
        ret
...

When using ::with_capacity the Rust version:

example::foo:
        push    rbp
        mov     rbp, rsp
        mov     eax, 4
        pop     rbp
        ret

is way better than the C++ version

foo(): # @foo()
  push rax
  mov edi, 4
  ;; -------------------------------------
  ;; what's the point of this memory allocation?
  call operator new(unsigned long)
  mov rdi, rax
  call operator delete(void*)
  ;; -------------------------------------
  mov eax, 314
  pop rcx
  ret

Ideally the Rust versions with and without ::with_capacity would beat the C++ version in both situations.

By altering Vec::push to:

        let len = self.len;
        if len == self.buf.cap() {
            self.buf.double();
        }
        unsafe {
            let end = self.as_mut_ptr().offset(len as isize);
            ptr::write(end, value);
            self.len = len + 1;
        }

I got rid of the bounds check.

I then tried to remove #[inline(never)] from RawVec::double(), and lo and behold, the example with 0 resolved to this:

    xorl    %eax, %eax
    retq

Using other (non-0) values in push seem to work fine too.

I don't know whether that could have some negative in other cases though.

Is RawVec::double() still [cold] in your case?

Yeah, I didn't remove that.

So... two questions:

  • _why_ is double marked #[inline(never)]? I understand the desire of making it cold is to avoid the compiler from inlining it and trashing the instruction cache for the "rare" situation in which the Vec grows. However, #[inline(never)] is a hammer that prevents some optimizations for these other examples. Isn't cold enough?

  • does double actually double the size of a Vec or does it use a memory-reuse-friendly growth factor like the one stated here and here or probably more relevant, at facebook (they are using jemalloc internally as well)? (in most vector implementations I know double is called grow instead to denote that the growth factor is an implementation detail)

Double is simply doubling.
There is an issue about the growth strategy of vectors: #29931

why is double marked #[inline(never)]?

See #23670, reducing the size of Vec::push was the motivation given.

Seems there is a bit of a tradeoff here.

Here's one silly benchmark:

I don't think the benchmark provided is silly. It shows that something wrong is going on with #[cold]. Why is a cold function being inlined when it is not profitable to do so?

It would be nice to know what happens in that benchmark w/o #[inline(never)] in that case today, and check the assembly.


EDIT: maybe cold is not enough and we need to annotate the call to double as unlikely ? (using LLVM's branch prediction hints?)

Extend doesn't use push anymore, so that benchmark isn't all that useful for current rust.

EDIT: If I use push manually, the non-inlined version is significantly faster:

#[bench]
fn x(b: &mut Bencher) {
    let mut v = Vec::with_capacity(100);
    b.iter(|| {
        for n in 0..100 {
            v.push(n);
        }
        v.truncate(0);
    }
    );
}
with inline(never)
test x ... bench:         202 ns/iter (+/- 5)

without:
test x ... bench:         334 ns/iter (+/- 29)

Actually, completely optimizing away the allocation would technically not be correct, since that would strictly speaking be changing the behaviour of the function, as the allocation could theoretically fail and cause a panic/exception. So in other words, the C++ version having a memory allocation call is correct.

It might still be possible to improve the Rust implementation optimize down to something similar to the C++ one though.

The first change in my previous comment makes this benchmark significantly faster (I got about 120-130 ns, compared to around 200 for the current version.) Though, it increases the code size a small bit, so for functions where nothing is known about the length of the vector it might cause a slight performance increase.

I'm tinkering around to see if I can reproduce it with just using the assume intrinsic, I managed to get rid of the bounds check in the original example with that at least without any code size increase..

I'm tinkering around to see if I can reproduce it with just using the assume intrinsic,

There is an llvm.expect intrinsic to tell LLVM that we expect the branch to not be taken in most cases.

I'm tinkering around to see if I can reproduce it with just using the assume intrinsic,

I managed to avoid the code size increase,
EDIT: Apparently there was still one extra instruction.
so at least I now have an improvement that avoids the bounds check and improves the benchmark (it seems llvm manages to merge the loop counter and length value after my changes.) I don't know if it's possible to reduce it further without letting double be inlined, but I'm not sure how to manage that without increasing the code size and slowing down things. (Using unlikely(), just prevented it from being inlined for the code in the OP, but not for a function that only did a push, so that wasn't very useful.)

I'm preparing a PR with my changes, so people can test if they are of any use.
EDIT:
There seems to be a few tradeoffs between different ways of altering the push version. Not sure what cases it's best to optimize for. Putting push inside a tight loop on it's own may not be the most real-world relevant example since one would normally use extend which avoids push entirely and is way faster as it can avoid checking the capacity inside the loop.

@oyvindln how is getting rid of a potential panic "changing the behavior"? I mean, if the optimized vector doesn't fail with OOM when the unoptimized one would, is that really bad? Surely it doesn't affect algorithmic correctness… Memory allocation, OOM conditions, etc. are mostly implementation details (and consequently very chaotic on modern OSes) anyway.

If someone is trying to test OOM by allocating and destroying single-element vectors, s/he is better off calling malloc() directly, no? And I fail to see any other case where removing an error condition via optimization could be called "changing observable behavior". Saying this would be akin to not allowing register allocation "because local variables are supposed to be on the stack". It does not quite seem reasonable to me, frankly.

@H2CO3 Ah sorry, I worded myself badly. I don't see a problem with optimizing out the allocations in practice (and normally that's what you would want), I was just pondering why the allocations weren't removed in the C++ version.

I tried various combinations of using the assume intrinsic calls on the length and on the capacity, ordering of the increment and the pointer write, and using unlikely on the in Vec<T>::push() but it's a bit hard to say which combination would be the ideal one without a benchmark that is more relevant to real-world code. (Also using assume a lot is a bit iffy as it can cause problems if used incorrectly.)

This was one compromise I came up with:

    pub fn push(&mut self, value: T) {
        unsafe {
            // This will panic or abort if we would allocate > isize::MAX bytes
            // or if the length increment would overflow for zero-sized types.
            if unlikely(self.len == self.buf.cap()) {
                let len = self.len;
                self.buf.double();
                // Let llvm know that the length didn't change
                // so if we push to a new vector of length 0,
                // we don't have to calculate the offset from
                // the start of the array.
                // This might be worth revisiting once #31681 is solved.
                // See also #44452.
                assume(self.len() == len);
            }

            let end = self.as_mut_ptr().offset(self.len as isize);
            // Copy len to a local before writing `value` to the vector. This seems to enable
            // llvm to merge the length and loop counter when writing to a fresh vector.
            let len = self.len;
            ptr::write(end, value);
            // Let llvm know that the length doesn't change in the write call.
            assume(self.len() == len);
            self.len += 1;
            // Let llvm know that len didn't overflow.
            // We can safely assume that there was no overflow,
            // since if len == usize::MAX, growing the vector would fail.
            // We already made the assumption that the invariant len <= capacity
            // holds when checked if we should grow the vector.
            //
            // Letting llvm know that there was no overflow can help llvm
            // avoid bounds checks after a push in some situations.
            assume(self.len() != 0);
        }
    }

It does result in one extra instruction for a lone raw push call though.
It's a bit harder to experiment with doing changes to RawVec::double as it uses some rust internals so it's difficult to test as a local function.

@oyvindln Oh, that does make sense. Sorry, I thought you were writing in favor of not optimizing away the allocation.

@oyvindln

unlikely(self.len == self.buf.cap())

this is what I had in mind, but I expected the code to look like this:

if unlikely(full()) {
  grow();
}

and grow should be marked #[cold], so that the cold applies to everything in the branch.

I know that cold calls are never inlined by llvm, but they _just have a lower threshold in the MIR inliner_. Is the MIR inliner the reason why it's conventional in the rust stdlib to mark functions as both cold and inline(never) when cold already prevents inlining in LLVM?

Would dropping cold functions for consideration in MIR inlining still prevent the double function from being inlined while still allowing LLVM to fully optimize the call? Even if it doesn't fix the issue we should probably prevent cold functions from being inlined for consistency with LLVM's behavior.

@Techcable MIR inlining is not currently enabled by default, and LLVM does inline coldcc functions (source). Also note that #[cold] translates to the cold attribute, not the coldcc calling convention, though that doesn't prevent inlining either.

Thanks for the clarification @rkruppe, you'd think the LLVM docs would be right! Do you think using the coldcc calling convention instead help prevent the issue? At the very least it'd help performance by helping reduce the call overhead.

On nightly the original code:

pub fn foo() -> u32 {
  let mut v: Vec<u32> = Vec::new();
  v.push(0);
  v[0]
}

now generates this:

example::foo:
  xor eax, eax
  ret

https://godbolt.org/g/bo4WaM

Nice! A second push still makes it trip: https://godbolt.org/g/TZ1Mbo =/

clang appears to handle a couple of push_backs before tripping.

No idea if it'll help, but this sounds like the optimizer limitation that this C++ lightning talk is about: https://www.youtube.com/watch?v=s4wnuiCwTGU Dunno why Rust would fare worse at this.

The only thing that makes it generate good assembly is using Vec::with_capacity(N) (see here):

pub fn foo() -> u32 {
  let mut v: Vec<u32> = Vec::with_capacity(3);
  v.push(7);
  v.push(4);
  v[1]
}

generates

example::foo:
        push    rbp
        mov     rbp, rsp
        mov     eax, 4
        pop     rbp
        ret

It doesn't anymore, it has regressed from 1.42 to 1.43

Was this page helpful?
0 / 5 - 0 ratings