diff options
| author | Corey Farwell <coreyf@rwell.org> | 2017-03-25 09:30:32 -0700 |
|---|---|---|
| committer | GitHub <noreply@github.com> | 2017-03-25 09:30:32 -0700 |
| commit | 2bdbcb061806e8e03507ecdfa22493d9195bf25c (patch) | |
| tree | 20bbed9b435f5c215a4d38a7ee202b18838d9c39 /src/libcore | |
| parent | 22f914a09a95fdd20126660af1889c543755b800 (diff) | |
| parent | 2c816f7fb6772fe0b77af04da9355aab6c2d3fb2 (diff) | |
| download | rust-2bdbcb061806e8e03507ecdfa22493d9195bf25c.tar.gz rust-2bdbcb061806e8e03507ecdfa22493d9195bf25c.zip | |
Rollup merge of #40807 - stjepang:optimize-insertion-sort, r=alexcrichton
Optimize insertion sort This change slightly changes the main iteration loop so that LLVM can optimize it more efficiently. Benchmark: ``` name before ns/iter after ns/iter diff ns/iter diff % slice::sort_unstable_small_ascending 39 (2051 MB/s) 38 (2105 MB/s) -1 -2.56% slice::sort_unstable_small_big_random 579 (2210 MB/s) 575 (2226 MB/s) -4 -0.69% slice::sort_unstable_small_descending 80 (1000 MB/s) 70 (1142 MB/s) -10 -12.50% slice::sort_unstable_small_random 396 (202 MB/s) 386 -10 -2.53% ``` The benchmark is not a fluke. I can see that performance on `small_descending` is consistently better after this change. I'm not 100% sure why this makes things faster, but my guess would be that `v.len()+1` to the compiler looks like it could in theory overflow.
Diffstat (limited to 'src/libcore')
| -rw-r--r-- | src/libcore/slice/sort.rs | 4 |
1 files changed, 2 insertions, 2 deletions
diff --git a/src/libcore/slice/sort.rs b/src/libcore/slice/sort.rs index d13d537d993..307e4974d97 100644 --- a/src/libcore/slice/sort.rs +++ b/src/libcore/slice/sort.rs @@ -152,8 +152,8 @@ fn partial_insertion_sort<T, F>(v: &mut [T], is_less: &mut F) -> bool fn insertion_sort<T, F>(v: &mut [T], is_less: &mut F) where F: FnMut(&T, &T) -> bool { - for i in 2..v.len()+1 { - shift_tail(&mut v[..i], is_less); + for i in 1..v.len() { + shift_tail(&mut v[..i+1], is_less); } } |
