about summary refs log tree commit diff
path: root/compiler/rustc_codegen_llvm/src
diff options
context:
space:
mode:
authorScott McMurray <scottmcm@users.noreply.github.com>2022-02-20 23:25:18 -0800
committerScott McMurray <scottmcm@users.noreply.github.com>2022-02-21 00:54:02 -0800
commit8ca47d7ae4e068c94b4ab7b25cc0ccc38d01d52c (patch)
tree60de7dd6dc79daf75dd18b25ec5d92ea44a85222 /compiler/rustc_codegen_llvm/src
parent73a7423e77b49a99e270531fbadda5b8899df3f6 (diff)
downloadrust-8ca47d7ae4e068c94b4ab7b25cc0ccc38d01d52c.tar.gz
rust-8ca47d7ae4e068c94b4ab7b25cc0ccc38d01d52c.zip
Stop manually SIMDing in swap_nonoverlapping
Like I previously did for `reverse`, this leaves it to LLVM to pick how to vectorize it, since it can know better the chunk size to use, compared to the "32 bytes always" approach we currently have.

It does still need logic to type-erase where appropriate, though, as while LLVM is now smart enough to vectorize over slices of things like `[u8; 4]`, it fails to do so over slices of `[u8; 3]`.

As a bonus, this also means one no longer gets the spurious `memcpy`(s?) at the end up swapping a slice of `__m256`s: <https://rust.godbolt.org/z/joofr4v8Y>
Diffstat (limited to 'compiler/rustc_codegen_llvm/src')
0 files changed, 0 insertions, 0 deletions