-rw-r--r-- 100467 djbsort-20260210/int32/neon/sort.c raw
/* WARNING: auto-generated (by autogen/sort); do not edit */
#include <arm_neon.h>
#include "int32_sort.h"
#define int32 int32_t
#define int32_largest 0x7fffffff
#include "crypto_int32.h"
#define int32_min crypto_int32_min
#define int32_MINMAX(a,b) crypto_int32_minmax(&(a),&(b))
#define NOINLINE __attribute__((noinline))
#include "crypto_int8.h"
#define int8 crypto_int8
#define int8_min crypto_int8_min
#define int8x16 int8x16_t
#include "crypto_uint8.h"
#define uint8 crypto_uint8
#define uint8x16 uint8x16_t
#define int32x4 int32x4_t
#define uint32x4 uint32x4_t
#define int32x4_load vld1q_s32
#define int32x4_store vst1q_s32
#define int32x4_ifthenelse vbslq_s32
#define int32x4_smaller_umask vcltq_s32
#define int32x4_min vminq_s32
#define int32x4_max vmaxq_s32
#define int32x4_MINMAX(a,b) \
do { \
int32x4 c = int32x4_min(a,b); \
b = int32x4_max(a,b); \
a = c; \
} while(0)
#define int8x16_load vld1q_s8
#define int8x16_varextract vqtbl1q_s8
#define int8x16_add vaddq_s8
#define int8x16_sub vsubq_s8
#define int8x16_broadcast vdupq_n_s8
#define uint8x16_load vld1q_u8
#define uint8x16_add vaddq_u8
#define uint8x16_sub vsubq_u8
#define uint8x16_broadcast vdupq_n_u8
#define int8x16_from_int32x4 vreinterpretq_s8_s32
#define uint32x4_from_int32x4 vreinterpretq_u32_s32
#define int32x4_from_uint32x4 vreinterpretq_s32_u32
#define int32x4_from_int8x16 vreinterpretq_s32_s8
#define int32x4_load vld1q_s32
#define int32x4_add vaddq_s32
#define int32x4_sub vsubq_s32
#define int32x4_broadcast vdupq_n_s32
static inline uint8x16 uint8x16_set(uint8 x0,uint8 x1,uint8 x2,uint8 x3,uint8 x4,uint8 x5,uint8 x6,uint8 x7,uint8 x8,uint8 x9,uint8 x10,uint8 x11,uint8 x12,uint8 x13,uint8 x14,uint8 x15)
{
uint8 x[16] = {x0,x1,x2,x3,x4,x5,x6,x7,x8,x9,x10,x11,x12,x13,x14,x15};
return uint8x16_load(x);
}
static inline int32x4 int32x4_set(int32 x0,int32 x1,int32 x2,int32 x3)
{
int32 x[4] = {x0,x1,x2,x3};
return int32x4_load(x);
}
#define int32x4_1032 vrev64q_s32
#define int32x4_2301(v) vextq_s32(v,v,2)
#define int32x4_3210(v) int32x4_1032(int32x4_2301(v))
#define int32x4_a0b0a2b2 vtrn1q_s32
#define int32x4_a1b1a3b3 vtrn2q_s32
#define int32x4_a0b0a1b1 vzip1q_s32
#define int32x4_a2b2a3b3 vzip2q_s32
#define int32x4_leftleft(a,b) vreinterpretq_s32_s64(vzip1q_s64(vreinterpretq_s64_s32(a),vreinterpretq_s64_s32(b)))
#define int32x4_rightright(a,b) vreinterpretq_s32_s64(vzip2q_s64(vreinterpretq_s64_s32(a),vreinterpretq_s64_s32(b)))
NOINLINE
static void int32_sort_3through7(int32 *x,long long n)
{
if (n >= 4) {
int32 x0 = x[0];
int32 x1 = x[1];
int32 x2 = x[2];
int32 x3 = x[3];
int32_MINMAX(x0,x1);
int32_MINMAX(x2,x3);
int32_MINMAX(x0,x2);
int32_MINMAX(x1,x3);
int32_MINMAX(x1,x2);
if (n >= 5) {
if (n == 5) {
int32 x4 = x[4];
int32_MINMAX(x0,x4);
int32_MINMAX(x2,x4);
int32_MINMAX(x1,x2);
int32_MINMAX(x3,x4);
x[4] = x4;
} else {
int32 x4 = x[4];
int32 x5 = x[5];
int32_MINMAX(x4,x5);
if (n == 6) {
int32_MINMAX(x0,x4);
int32_MINMAX(x2,x4);
int32_MINMAX(x1,x5);
int32_MINMAX(x3,x5);
} else {
int32 x6 = x[6];
int32_MINMAX(x4,x6);
int32_MINMAX(x5,x6);
int32_MINMAX(x0,x4);
int32_MINMAX(x2,x6);
int32_MINMAX(x2,x4);
int32_MINMAX(x1,x5);
int32_MINMAX(x3,x5);
int32_MINMAX(x5,x6);
x[6] = x6;
}
int32_MINMAX(x1,x2);
int32_MINMAX(x3,x4);
x[4] = x4;
x[5] = x5;
}
}
x[0] = x0;
x[1] = x1;
x[2] = x2;
x[3] = x3;
} else {
int32 x0 = x[0];
int32 x1 = x[1];
int32 x2 = x[2];
int32_MINMAX(x0,x1);
int32_MINMAX(x0,x2);
int32_MINMAX(x1,x2);
x[0] = x0;
x[1] = x1;
x[2] = x2;
}
}
NOINLINE
static void int32_sort_8through16(int32 *x,long long n)
{
int32_t pos2,pos3;
int32x4 infty,x0,x1,x2,x3,y0,y1,y2,y3;
x0 = int32x4_load(x+0); // 0 4 8 12
x1 = int32x4_load(x+4); // 1 5 9 13
infty = int32x4_broadcast(int32_largest);
pos2 = int32_min(12,n);
x2 = int32x4_ifthenelse(int32x4_smaller_umask(int32x4_set(8,9,10,11),int32x4_broadcast(pos2)),int32x4_from_int8x16(int8x16_varextract(int8x16_from_int32x4(int32x4_load(x+pos2-4)),uint8x16_add(uint8x16_set(0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15),uint8x16_broadcast(4*((-pos2)&3))))),infty); // 2 6 10 14
pos3 = int32_min(16,n);
x3 = int32x4_ifthenelse(int32x4_smaller_umask(int32x4_set(12,13,14,15),int32x4_broadcast(pos3)),int32x4_from_int8x16(int8x16_varextract(int8x16_from_int32x4(int32x4_load(x+pos3-4)),uint8x16_add(uint8x16_set(0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15),uint8x16_broadcast(4*((-pos3)&3))))),infty); // 3 7 11 15
// stage (1,0) 0:1 2:3 4:5 6:7 8:9 10:11 12:13 14:15
y0 = int32x4_min(x0,x1); // 0 4 8 12
x1 = int32x4_max(x0,x1); // 1 5 9 13
y2 = int32x4_min(x2,x3); // 2 6 10 14
x3 = int32x4_max(x2,x3); // 3 7 11 15
// stage (2,1) 0:2 1:3 4:6 5:7 8:10 9:11 12:14 13:15
x0 = int32x4_min(y0,y2); // 0 4 8 12
y2 = int32x4_max(y0,y2); // 2 6 10 14
y1 = int32x4_min(x1,x3); // 1 5 9 13
x3 = int32x4_max(x1,x3); // 3 7 11 15
// stage (2,0) 1:2 5:6 9:10 13:14
x2 = int32x4_min(y2,y1); // 1 5 9 13
y1 = int32x4_max(y2,y1); // 2 6 10 14
// stage (3,2) 0:7 1:6 2:5 3:4 8:15 9:14 10:13 11:12
y1 = int32x4_1032(y1); // 6 2 14 10
x3 = int32x4_1032(x3); // 7 3 15 11
y2 = int32x4_min(x2,y1); // 1 2 9 10
y1 = int32x4_max(x2,y1); // 6 5 14 13
y0 = int32x4_min(x0,x3); // 0 3 8 11
x3 = int32x4_max(x0,x3); // 7 4 15 12
// stage (3,1) 0:2 1:3 4:6 5:7 8:10 9:11 12:14 13:15
y0 = int32x4_1032(y0); // 3 0 11 8
x3 = int32x4_1032(x3); // 4 7 12 15
x2 = int32x4_min(y2,y0); // 1 0 9 8
y0 = int32x4_max(y2,y0); // 3 2 11 10
x1 = int32x4_min(y1,x3); // 4 5 12 13
x3 = int32x4_max(y1,x3); // 6 7 14 15
// stage (3,0) 0:1 2:3 4:5 6:7 8:9 10:11 12:13 14:15
y2 = int32x4_a0b0a2b2(x2,y0); // 1 3 9 11
y0 = int32x4_a1b1a3b3(x2,y0); // 0 2 8 10
y1 = int32x4_a0b0a2b2(x1,x3); // 4 6 12 14
x3 = int32x4_a1b1a3b3(x1,x3); // 5 7 13 15
x2 = int32x4_min(y2,y0); // 0 2 8 10
y0 = int32x4_max(y2,y0); // 1 3 9 11
x1 = int32x4_min(y1,x3); // 4 6 12 14
x3 = int32x4_max(y1,x3); // 5 7 13 15
// stage (4,3) 0:15 1:14 2:13 3:12 4:11 5:10 6:9 7:8
x3 = int32x4_3210(x3); // 15 13 7 5
y0 = int32x4_3210(y0); // 11 9 3 1
y2 = int32x4_min(x2,x3); // 0 2 7 5
x3 = int32x4_max(x2,x3); // 15 13 8 10
y1 = int32x4_min(x1,y0); // 4 6 3 1
y0 = int32x4_max(x1,y0); // 11 9 12 14
// stage (4,2) 0:4 1:5 2:6 3:7 8:12 9:13 10:14 11:15
x2 = int32x4_min(y2,y1); // 0 2 3 1
y1 = int32x4_max(y2,y1); // 4 6 7 5
y3 = int32x4_min(x3,y0); // 11 9 8 10
y0 = int32x4_max(x3,y0); // 15 13 12 14
// stage (4,1) 0:2 1:3 4:6 5:7 8:10 9:11 12:14 13:15
y2 = int32x4_a0b0a2b2(x2,y1); // 0 4 3 7
y1 = int32x4_a1b1a3b3(x2,y1); // 2 6 1 5
x3 = int32x4_a0b0a2b2(y3,y0); // 11 15 8 12
y0 = int32x4_a1b1a3b3(y3,y0); // 9 13 10 14
x2 = int32x4_min(y2,y1); // 0 4 1 5
y1 = int32x4_max(y2,y1); // 2 6 3 7
y3 = int32x4_min(x3,y0); // 9 13 8 12
y0 = int32x4_max(x3,y0); // 11 15 10 14
// stage (4,0) 0:1 2:3 4:5 6:7 8:9 10:11 12:13 14:15
y2 = int32x4_leftleft(x2,y1); // 0 4 2 6
y1 = int32x4_rightright(x2,y1); // 1 5 3 7
x3 = int32x4_leftleft(y3,y0); // 9 13 11 15
y0 = int32x4_rightright(y3,y0); // 8 12 10 14
x2 = int32x4_min(y2,y1); // 0 4 2 6
y1 = int32x4_max(y2,y1); // 1 5 3 7
y3 = int32x4_min(x3,y0); // 8 12 10 14
y0 = int32x4_max(x3,y0); // 9 13 11 15
y2 = int32x4_a0b0a2b2(x2,y1); // 0 1 2 3
y1 = int32x4_a1b1a3b3(x2,y1); // 4 5 6 7
x3 = int32x4_a0b0a2b2(y3,y0); // 8 9 10 11
y0 = int32x4_a1b1a3b3(y3,y0); // 12 13 14 15
int32x4_store(x+pos3-4,int32x4_from_int8x16(int8x16_varextract(int8x16_from_int32x4(y0),uint8x16_sub(uint8x16_set(0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15),uint8x16_broadcast(4*((-pos3)&3))))));
int32x4_store(x+pos2-4,int32x4_from_int8x16(int8x16_varextract(int8x16_from_int32x4(x3),uint8x16_sub(uint8x16_set(0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15),uint8x16_broadcast(4*((-pos2)&3))))));
int32x4_store(x+4,y1);
int32x4_store(x+0,y2);
}
NOINLINE
static void int32_sort_16through32(int32 *x,long long n)
{
int32_t pos4,pos5,pos6,pos7;
int32x4 infty,x0,x1,x2,x3,x4,x5,x6,x7,y0,y1,y2,y3,y4,y5,y6,y7;
x0 = int32x4_load(x+0); // 0 8 16 24
x1 = int32x4_load(x+4); // 1 9 17 25
x2 = int32x4_load(x+8); // 2 10 18 26
x3 = int32x4_load(x+12); // 3 11 19 27
infty = int32x4_broadcast(int32_largest);
pos4 = int32_min(20,n);
x4 = int32x4_ifthenelse(int32x4_smaller_umask(int32x4_set(16,17,18,19),int32x4_broadcast(pos4)),int32x4_from_int8x16(int8x16_varextract(int8x16_from_int32x4(int32x4_load(x+pos4-4)),uint8x16_add(uint8x16_set(0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15),uint8x16_broadcast(4*((-pos4)&3))))),infty); // 4 12 20 28
pos5 = int32_min(24,n);
x5 = int32x4_ifthenelse(int32x4_smaller_umask(int32x4_set(20,21,22,23),int32x4_broadcast(pos5)),int32x4_from_int8x16(int8x16_varextract(int8x16_from_int32x4(int32x4_load(x+pos5-4)),uint8x16_add(uint8x16_set(0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15),uint8x16_broadcast(4*((-pos5)&3))))),infty); // 5 13 21 29
pos6 = int32_min(28,n);
x6 = int32x4_ifthenelse(int32x4_smaller_umask(int32x4_set(24,25,26,27),int32x4_broadcast(pos6)),int32x4_from_int8x16(int8x16_varextract(int8x16_from_int32x4(int32x4_load(x+pos6-4)),uint8x16_add(uint8x16_set(0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15),uint8x16_broadcast(4*((-pos6)&3))))),infty); // 6 14 22 30
pos7 = int32_min(32,n);
x7 = int32x4_ifthenelse(int32x4_smaller_umask(int32x4_set(28,29,30,31),int32x4_broadcast(pos7)),int32x4_from_int8x16(int8x16_varextract(int8x16_from_int32x4(int32x4_load(x+pos7-4)),uint8x16_add(uint8x16_set(0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15),uint8x16_broadcast(4*((-pos7)&3))))),infty); // 7 15 23 31
// stage (1,0) 0:1 2:3 4:5 6:7 8:9 10:11 12:13 14:15 16:17 18:19 20:21 22:23 24:25 26:27 28:29 30:31
y0 = int32x4_min(x0,x1); // 0 8 16 24
x1 = int32x4_max(x0,x1); // 1 9 17 25
y2 = int32x4_min(x2,x3); // 2 10 18 26
x3 = int32x4_max(x2,x3); // 3 11 19 27
y4 = int32x4_min(x4,x5); // 4 12 20 28
x5 = int32x4_max(x4,x5); // 5 13 21 29
y6 = int32x4_min(x6,x7); // 6 14 22 30
x7 = int32x4_max(x6,x7); // 7 15 23 31
// stage (2,1) 0:2 1:3 4:6 5:7 8:10 9:11 12:14 13:15 16:18 17:19 20:22 21:23 24:26 25:27 28:30 29:31
x0 = int32x4_min(y0,y2); // 0 8 16 24
y2 = int32x4_max(y0,y2); // 2 10 18 26
y1 = int32x4_min(x1,x3); // 1 9 17 25
x3 = int32x4_max(x1,x3); // 3 11 19 27
x4 = int32x4_min(y4,y6); // 4 12 20 28
y6 = int32x4_max(y4,y6); // 6 14 22 30
y5 = int32x4_min(x5,x7); // 5 13 21 29
x7 = int32x4_max(x5,x7); // 7 15 23 31
// stage (2,0) 1:2 5:6 9:10 13:14 17:18 21:22 25:26 29:30
x2 = int32x4_min(y2,y1); // 1 9 17 25
y1 = int32x4_max(y2,y1); // 2 10 18 26
x6 = int32x4_min(y6,y5); // 5 13 21 29
y5 = int32x4_max(y6,y5); // 6 14 22 30
// stage (3,2) 0:4 1:5 2:6 3:7 8:12 9:13 10:14 11:15 16:20 17:21 18:22 19:23 24:28 25:29 26:30 27:31
y2 = int32x4_min(x2,x6); // 1 9 17 25
x6 = int32x4_max(x2,x6); // 5 13 21 29
y0 = int32x4_min(x0,x4); // 0 8 16 24
x4 = int32x4_max(x0,x4); // 4 12 20 28
x1 = int32x4_min(y1,y5); // 2 10 18 26
y5 = int32x4_max(y1,y5); // 6 14 22 30
y3 = int32x4_min(x3,x7); // 3 11 19 27
x7 = int32x4_max(x3,x7); // 7 15 23 31
// stage (3,1) 2:4 3:5 10:12 11:13 18:20 19:21 26:28 27:29
y6 = int32x4_min(x6,y3); // 3 11 19 27
y3 = int32x4_max(x6,y3); // 5 13 21 29
y4 = int32x4_min(x4,x1); // 2 10 18 26
x1 = int32x4_max(x4,x1); // 4 12 20 28
// stage (3,0) 1:2 3:4 5:6 9:10 11:12 13:14 17:18 19:20 21:22 25:26 27:28 29:30
x6 = int32x4_min(y6,x1); // 3 11 19 27
x1 = int32x4_max(y6,x1); // 4 12 20 28
x4 = int32x4_min(y4,y2); // 1 9 17 25
y2 = int32x4_max(y4,y2); // 2 10 18 26
x5 = int32x4_min(y5,y3); // 5 13 21 29
y3 = int32x4_max(y5,y3); // 6 14 22 30
// stage (4,3) 0:15 1:14 2:13 3:12 4:11 5:10 6:9 7:8 16:31 17:30 18:29 19:28 20:27 21:26 22:25 23:24
x1 = int32x4_1032(x1); // 12 4 28 20
y3 = int32x4_1032(y3); // 14 6 30 22
y2 = int32x4_1032(y2); // 10 2 26 18
x7 = int32x4_1032(x7); // 15 7 31 23
y6 = int32x4_min(x6,x1); // 3 4 19 20
x1 = int32x4_max(x6,x1); // 12 11 28 27
y4 = int32x4_min(x4,y3); // 1 6 17 22
y3 = int32x4_max(x4,y3); // 14 9 30 25
y5 = int32x4_min(x5,y2); // 5 2 21 18
y2 = int32x4_max(x5,y2); // 10 13 26 29
x0 = int32x4_min(y0,x7); // 0 7 16 23
x7 = int32x4_max(y0,x7); // 15 8 31 24
// stage (4,2) 0:4 1:5 2:6 3:7 8:12 9:13 10:14 11:15 16:20 17:21 18:22 19:23 24:28 25:29 26:30 27:31
x0 = int32x4_1032(x0); // 7 0 23 16
x7 = int32x4_1032(x7); // 8 15 24 31
x6 = int32x4_min(y6,x0); // 3 0 19 16
x0 = int32x4_max(y6,x0); // 7 4 23 20
x4 = int32x4_min(y4,y5); // 1 2 17 18
y5 = int32x4_max(y4,y5); // 5 6 21 22
x3 = int32x4_min(y3,y2); // 10 9 26 25
y2 = int32x4_max(y3,y2); // 14 13 30 29
y1 = int32x4_min(x1,x7); // 8 11 24 27
x7 = int32x4_max(x1,x7); // 12 15 28 31
// stage (4,1) 0:2 1:3 4:6 5:7 8:10 9:11 12:14 13:15 16:18 17:19 20:22 21:23 24:26 25:27 28:30 29:31
y6 = int32x4_min(x6,x4); // 1 0 17 16
x4 = int32x4_max(x6,x4); // 3 2 19 18
y0 = int32x4_min(x0,y5); // 5 4 21 20
y5 = int32x4_max(x0,y5); // 7 6 23 22
y3 = int32x4_min(x3,y1); // 8 9 24 25
y1 = int32x4_max(x3,y1); // 10 11 26 27
x2 = int32x4_min(y2,x7); // 12 13 28 29
x7 = int32x4_max(y2,x7); // 14 15 30 31
// stage (4,0) 0:1 2:3 4:5 6:7 8:9 10:11 12:13 14:15 16:17 18:19 20:21 22:23 24:25 26:27 28:29 30:31
x6 = int32x4_a0b0a2b2(y6,x4); // 1 3 17 19
x4 = int32x4_a1b1a3b3(y6,x4); // 0 2 16 18
x0 = int32x4_a0b0a2b2(y0,y5); // 5 7 21 23
y5 = int32x4_a1b1a3b3(y0,y5); // 4 6 20 22
x3 = int32x4_a0b0a2b2(y3,y1); // 8 10 24 26
y1 = int32x4_a1b1a3b3(y3,y1); // 9 11 25 27
y2 = int32x4_a0b0a2b2(x2,x7); // 12 14 28 30
x7 = int32x4_a1b1a3b3(x2,x7); // 13 15 29 31
y6 = int32x4_min(x6,x4); // 0 2 16 18
x4 = int32x4_max(x6,x4); // 1 3 17 19
y0 = int32x4_min(x0,y5); // 4 6 20 22
y5 = int32x4_max(x0,y5); // 5 7 21 23
y3 = int32x4_min(x3,y1); // 8 10 24 26
y1 = int32x4_max(x3,y1); // 9 11 25 27
x2 = int32x4_min(y2,x7); // 12 14 28 30
x7 = int32x4_max(y2,x7); // 13 15 29 31
// stage (5,4) 0:31 1:30 2:29 3:28 4:27 5:26 6:25 7:24 8:23 9:22 10:21 11:20 12:19 13:18 14:17 15:16
x7 = int32x4_3210(x7); // 31 29 15 13
y1 = int32x4_3210(y1); // 27 25 11 9
y5 = int32x4_3210(y5); // 23 21 7 5
x4 = int32x4_3210(x4); // 19 17 3 1
x6 = int32x4_min(y6,x7); // 0 2 15 13
x7 = int32x4_max(y6,x7); // 31 29 16 18
x0 = int32x4_min(y0,y1); // 4 6 11 9
y1 = int32x4_max(y0,y1); // 27 25 20 22
x3 = int32x4_min(y3,y5); // 8 10 7 5
y5 = int32x4_max(y3,y5); // 23 21 24 26
y2 = int32x4_min(x2,x4); // 12 14 3 1
x4 = int32x4_max(x2,x4); // 19 17 28 30
// stage (5,3) 0:8 1:9 2:10 3:11 4:12 5:13 6:14 7:15 16:24 17:25 18:26 19:27 20:28 21:29 22:30 23:31
y6 = int32x4_min(x6,x3); // 0 2 7 5
x3 = int32x4_max(x6,x3); // 8 10 15 13
y0 = int32x4_min(x0,y2); // 4 6 3 1
y2 = int32x4_max(x0,y2); // 12 14 11 9
y7 = int32x4_min(x7,y5); // 23 21 16 18
y5 = int32x4_max(x7,y5); // 31 29 24 26
x1 = int32x4_min(y1,x4); // 19 17 20 22
x4 = int32x4_max(y1,x4); // 27 25 28 30
// stage (5,2) 0:4 1:5 2:6 3:7 8:12 9:13 10:14 11:15 16:20 17:21 18:22 19:23 24:28 25:29 26:30 27:31
x6 = int32x4_min(y6,y0); // 0 2 3 1
y0 = int32x4_max(y6,y0); // 4 6 7 5
y3 = int32x4_min(x3,y2); // 8 10 11 9
y2 = int32x4_max(x3,y2); // 12 14 15 13
x7 = int32x4_min(y7,x1); // 19 17 16 18
x1 = int32x4_max(y7,x1); // 23 21 20 22
x5 = int32x4_min(y5,x4); // 27 25 24 26
x4 = int32x4_max(y5,x4); // 31 29 28 30
// stage (5,1) 0:2 1:3 4:6 5:7 8:10 9:11 12:14 13:15 16:18 17:19 20:22 21:23 24:26 25:27 28:30 29:31
y6 = int32x4_a0b0a2b2(x6,y0); // 0 4 3 7
y0 = int32x4_a1b1a3b3(x6,y0); // 2 6 1 5
x3 = int32x4_a0b0a2b2(y3,y2); // 8 12 11 15
y2 = int32x4_a1b1a3b3(y3,y2); // 10 14 9 13
y7 = int32x4_a0b0a2b2(x7,x1); // 19 23 16 20
x1 = int32x4_a1b1a3b3(x7,x1); // 17 21 18 22
y5 = int32x4_a0b0a2b2(x5,x4); // 27 31 24 28
x4 = int32x4_a1b1a3b3(x5,x4); // 25 29 26 30
x6 = int32x4_min(y6,y0); // 0 4 1 5
y0 = int32x4_max(y6,y0); // 2 6 3 7
y3 = int32x4_min(x3,y2); // 8 12 9 13
y2 = int32x4_max(x3,y2); // 10 14 11 15
x7 = int32x4_min(y7,x1); // 17 21 16 20
x1 = int32x4_max(y7,x1); // 19 23 18 22
x5 = int32x4_min(y5,x4); // 25 29 24 28
x4 = int32x4_max(y5,x4); // 27 31 26 30
// stage (5,0) 0:1 2:3 4:5 6:7 8:9 10:11 12:13 14:15 16:17 18:19 20:21 22:23 24:25 26:27 28:29 30:31
y6 = int32x4_leftleft(x6,y0); // 0 4 2 6
y0 = int32x4_rightright(x6,y0); // 1 5 3 7
x3 = int32x4_leftleft(y3,y2); // 8 12 10 14
y2 = int32x4_rightright(y3,y2); // 9 13 11 15
y7 = int32x4_leftleft(x7,x1); // 17 21 19 23
x1 = int32x4_rightright(x7,x1); // 16 20 18 22
y5 = int32x4_leftleft(x5,x4); // 25 29 27 31
x4 = int32x4_rightright(x5,x4); // 24 28 26 30
x6 = int32x4_min(y6,y0); // 0 4 2 6
y0 = int32x4_max(y6,y0); // 1 5 3 7
y3 = int32x4_min(x3,y2); // 8 12 10 14
y2 = int32x4_max(x3,y2); // 9 13 11 15
x7 = int32x4_min(y7,x1); // 16 20 18 22
x1 = int32x4_max(y7,x1); // 17 21 19 23
x5 = int32x4_min(y5,x4); // 24 28 26 30
x4 = int32x4_max(y5,x4); // 25 29 27 31
y6 = int32x4_a0b0a2b2(x6,y0); // 0 1 2 3
y0 = int32x4_a1b1a3b3(x6,y0); // 4 5 6 7
x3 = int32x4_a0b0a2b2(y3,y2); // 8 9 10 11
y2 = int32x4_a1b1a3b3(y3,y2); // 12 13 14 15
y7 = int32x4_a0b0a2b2(x7,x1); // 16 17 18 19
x1 = int32x4_a1b1a3b3(x7,x1); // 20 21 22 23
y5 = int32x4_a0b0a2b2(x5,x4); // 24 25 26 27
x4 = int32x4_a1b1a3b3(x5,x4); // 28 29 30 31
int32x4_store(x+pos7-4,int32x4_from_int8x16(int8x16_varextract(int8x16_from_int32x4(x4),uint8x16_sub(uint8x16_set(0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15),uint8x16_broadcast(4*((-pos7)&3))))));
int32x4_store(x+pos6-4,int32x4_from_int8x16(int8x16_varextract(int8x16_from_int32x4(y5),uint8x16_sub(uint8x16_set(0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15),uint8x16_broadcast(4*((-pos6)&3))))));
int32x4_store(x+pos5-4,int32x4_from_int8x16(int8x16_varextract(int8x16_from_int32x4(x1),uint8x16_sub(uint8x16_set(0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15),uint8x16_broadcast(4*((-pos5)&3))))));
int32x4_store(x+pos4-4,int32x4_from_int8x16(int8x16_varextract(int8x16_from_int32x4(y7),uint8x16_sub(uint8x16_set(0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15),uint8x16_broadcast(4*((-pos4)&3))))));
int32x4_store(x+12,y2);
int32x4_store(x+8,x3);
int32x4_store(x+4,y0);
int32x4_store(x+0,y6);
}
NOINLINE
static void int32_sort_32_xor(int32 *x,int32 xor)
{
int32x4 vecxor,x0,x1,x2,x3,x4,x5,x6,x7,y0,y1,y2,y3,y4,y5,y6,y7;
vecxor = int32x4_broadcast(xor);
x0 = vecxor^int32x4_load(x+0); // 0 8 16 24
x1 = vecxor^int32x4_load(x+4); // 1 9 17 25
x2 = vecxor^int32x4_load(x+8); // 2 10 18 26
x3 = vecxor^int32x4_load(x+12); // 3 11 19 27
x4 = vecxor^int32x4_load(x+16); // 4 12 20 28
x5 = vecxor^int32x4_load(x+20); // 5 13 21 29
x6 = vecxor^int32x4_load(x+24); // 6 14 22 30
x7 = vecxor^int32x4_load(x+28); // 7 15 23 31
// stage (1,0) 0:1 2:3 4:5 6:7 8:9 10:11 12:13 14:15 16:17 18:19 20:21 22:23 24:25 26:27 28:29 30:31
y0 = int32x4_min(x0,x1); // 0 8 16 24
x1 = int32x4_max(x0,x1); // 1 9 17 25
y2 = int32x4_min(x2,x3); // 2 10 18 26
x3 = int32x4_max(x2,x3); // 3 11 19 27
y4 = int32x4_min(x4,x5); // 4 12 20 28
x5 = int32x4_max(x4,x5); // 5 13 21 29
y6 = int32x4_min(x6,x7); // 6 14 22 30
x7 = int32x4_max(x6,x7); // 7 15 23 31
// stage (2,1) 0:2 1:3 4:6 5:7 8:10 9:11 12:14 13:15 16:18 17:19 20:22 21:23 24:26 25:27 28:30 29:31
x0 = int32x4_min(y0,y2); // 0 8 16 24
y2 = int32x4_max(y0,y2); // 2 10 18 26
y1 = int32x4_min(x1,x3); // 1 9 17 25
x3 = int32x4_max(x1,x3); // 3 11 19 27
x4 = int32x4_min(y4,y6); // 4 12 20 28
y6 = int32x4_max(y4,y6); // 6 14 22 30
y5 = int32x4_min(x5,x7); // 5 13 21 29
x7 = int32x4_max(x5,x7); // 7 15 23 31
// stage (2,0) 1:2 5:6 9:10 13:14 17:18 21:22 25:26 29:30
x2 = int32x4_min(y2,y1); // 1 9 17 25
y1 = int32x4_max(y2,y1); // 2 10 18 26
x6 = int32x4_min(y6,y5); // 5 13 21 29
y5 = int32x4_max(y6,y5); // 6 14 22 30
// stage (3,2) 0:4 1:5 2:6 3:7 8:12 9:13 10:14 11:15 16:20 17:21 18:22 19:23 24:28 25:29 26:30 27:31
y2 = int32x4_min(x2,x6); // 1 9 17 25
x6 = int32x4_max(x2,x6); // 5 13 21 29
y0 = int32x4_min(x0,x4); // 0 8 16 24
x4 = int32x4_max(x0,x4); // 4 12 20 28
x1 = int32x4_min(y1,y5); // 2 10 18 26
y5 = int32x4_max(y1,y5); // 6 14 22 30
y3 = int32x4_min(x3,x7); // 3 11 19 27
x7 = int32x4_max(x3,x7); // 7 15 23 31
// stage (3,1) 2:4 3:5 10:12 11:13 18:20 19:21 26:28 27:29
y6 = int32x4_min(x6,y3); // 3 11 19 27
y3 = int32x4_max(x6,y3); // 5 13 21 29
y4 = int32x4_min(x4,x1); // 2 10 18 26
x1 = int32x4_max(x4,x1); // 4 12 20 28
// stage (3,0) 1:2 3:4 5:6 9:10 11:12 13:14 17:18 19:20 21:22 25:26 27:28 29:30
x6 = int32x4_min(y6,x1); // 3 11 19 27
x1 = int32x4_max(y6,x1); // 4 12 20 28
x4 = int32x4_min(y4,y2); // 1 9 17 25
y2 = int32x4_max(y4,y2); // 2 10 18 26
x5 = int32x4_min(y5,y3); // 5 13 21 29
y3 = int32x4_max(y5,y3); // 6 14 22 30
// stage (4,3) 0:15 1:14 2:13 3:12 4:11 5:10 6:9 7:8 16:31 17:30 18:29 19:28 20:27 21:26 22:25 23:24
x1 = int32x4_1032(x1); // 12 4 28 20
y3 = int32x4_1032(y3); // 14 6 30 22
y2 = int32x4_1032(y2); // 10 2 26 18
x7 = int32x4_1032(x7); // 15 7 31 23
y6 = int32x4_min(x6,x1); // 3 4 19 20
x1 = int32x4_max(x6,x1); // 12 11 28 27
y4 = int32x4_min(x4,y3); // 1 6 17 22
y3 = int32x4_max(x4,y3); // 14 9 30 25
y5 = int32x4_min(x5,y2); // 5 2 21 18
y2 = int32x4_max(x5,y2); // 10 13 26 29
x0 = int32x4_min(y0,x7); // 0 7 16 23
x7 = int32x4_max(y0,x7); // 15 8 31 24
// stage (4,2) 0:4 1:5 2:6 3:7 8:12 9:13 10:14 11:15 16:20 17:21 18:22 19:23 24:28 25:29 26:30 27:31
x0 = int32x4_1032(x0); // 7 0 23 16
x7 = int32x4_1032(x7); // 8 15 24 31
x6 = int32x4_min(y6,x0); // 3 0 19 16
x0 = int32x4_max(y6,x0); // 7 4 23 20
x4 = int32x4_min(y4,y5); // 1 2 17 18
y5 = int32x4_max(y4,y5); // 5 6 21 22
x3 = int32x4_min(y3,y2); // 10 9 26 25
y2 = int32x4_max(y3,y2); // 14 13 30 29
y1 = int32x4_min(x1,x7); // 8 11 24 27
x7 = int32x4_max(x1,x7); // 12 15 28 31
// stage (4,1) 0:2 1:3 4:6 5:7 8:10 9:11 12:14 13:15 16:18 17:19 20:22 21:23 24:26 25:27 28:30 29:31
y6 = int32x4_min(x6,x4); // 1 0 17 16
x4 = int32x4_max(x6,x4); // 3 2 19 18
y0 = int32x4_min(x0,y5); // 5 4 21 20
y5 = int32x4_max(x0,y5); // 7 6 23 22
y3 = int32x4_min(x3,y1); // 8 9 24 25
y1 = int32x4_max(x3,y1); // 10 11 26 27
x2 = int32x4_min(y2,x7); // 12 13 28 29
x7 = int32x4_max(y2,x7); // 14 15 30 31
// stage (4,0) 0:1 2:3 4:5 6:7 8:9 10:11 12:13 14:15 16:17 18:19 20:21 22:23 24:25 26:27 28:29 30:31
x6 = int32x4_a0b0a2b2(y6,x4); // 1 3 17 19
x4 = int32x4_a1b1a3b3(y6,x4); // 0 2 16 18
x0 = int32x4_a0b0a2b2(y0,y5); // 5 7 21 23
y5 = int32x4_a1b1a3b3(y0,y5); // 4 6 20 22
x3 = int32x4_a0b0a2b2(y3,y1); // 8 10 24 26
y1 = int32x4_a1b1a3b3(y3,y1); // 9 11 25 27
y2 = int32x4_a0b0a2b2(x2,x7); // 12 14 28 30
x7 = int32x4_a1b1a3b3(x2,x7); // 13 15 29 31
y6 = int32x4_min(x6,x4); // 0 2 16 18
x4 = int32x4_max(x6,x4); // 1 3 17 19
y0 = int32x4_min(x0,y5); // 4 6 20 22
y5 = int32x4_max(x0,y5); // 5 7 21 23
y3 = int32x4_min(x3,y1); // 8 10 24 26
y1 = int32x4_max(x3,y1); // 9 11 25 27
x2 = int32x4_min(y2,x7); // 12 14 28 30
x7 = int32x4_max(y2,x7); // 13 15 29 31
// stage (5,4) 0:31 1:30 2:29 3:28 4:27 5:26 6:25 7:24 8:23 9:22 10:21 11:20 12:19 13:18 14:17 15:16
x7 = int32x4_3210(x7); // 31 29 15 13
y1 = int32x4_3210(y1); // 27 25 11 9
y5 = int32x4_3210(y5); // 23 21 7 5
x4 = int32x4_3210(x4); // 19 17 3 1
x6 = int32x4_min(y6,x7); // 0 2 15 13
x7 = int32x4_max(y6,x7); // 31 29 16 18
x0 = int32x4_min(y0,y1); // 4 6 11 9
y1 = int32x4_max(y0,y1); // 27 25 20 22
x3 = int32x4_min(y3,y5); // 8 10 7 5
y5 = int32x4_max(y3,y5); // 23 21 24 26
y2 = int32x4_min(x2,x4); // 12 14 3 1
x4 = int32x4_max(x2,x4); // 19 17 28 30
// stage (5,3) 0:8 1:9 2:10 3:11 4:12 5:13 6:14 7:15 16:24 17:25 18:26 19:27 20:28 21:29 22:30 23:31
y6 = int32x4_min(x6,x3); // 0 2 7 5
x3 = int32x4_max(x6,x3); // 8 10 15 13
y0 = int32x4_min(x0,y2); // 4 6 3 1
y2 = int32x4_max(x0,y2); // 12 14 11 9
y7 = int32x4_min(x7,y5); // 23 21 16 18
y5 = int32x4_max(x7,y5); // 31 29 24 26
x1 = int32x4_min(y1,x4); // 19 17 20 22
x4 = int32x4_max(y1,x4); // 27 25 28 30
// stage (5,2) 0:4 1:5 2:6 3:7 8:12 9:13 10:14 11:15 16:20 17:21 18:22 19:23 24:28 25:29 26:30 27:31
x6 = int32x4_min(y6,y0); // 0 2 3 1
y0 = int32x4_max(y6,y0); // 4 6 7 5
y3 = int32x4_min(x3,y2); // 8 10 11 9
y2 = int32x4_max(x3,y2); // 12 14 15 13
x7 = int32x4_min(y7,x1); // 19 17 16 18
x1 = int32x4_max(y7,x1); // 23 21 20 22
x5 = int32x4_min(y5,x4); // 27 25 24 26
x4 = int32x4_max(y5,x4); // 31 29 28 30
// stage (5,1) 0:2 1:3 4:6 5:7 8:10 9:11 12:14 13:15 16:18 17:19 20:22 21:23 24:26 25:27 28:30 29:31
y6 = int32x4_a0b0a2b2(x6,y0); // 0 4 3 7
y0 = int32x4_a1b1a3b3(x6,y0); // 2 6 1 5
x3 = int32x4_a0b0a2b2(y3,y2); // 8 12 11 15
y2 = int32x4_a1b1a3b3(y3,y2); // 10 14 9 13
y7 = int32x4_a0b0a2b2(x7,x1); // 19 23 16 20
x1 = int32x4_a1b1a3b3(x7,x1); // 17 21 18 22
y5 = int32x4_a0b0a2b2(x5,x4); // 27 31 24 28
x4 = int32x4_a1b1a3b3(x5,x4); // 25 29 26 30
x6 = int32x4_min(y6,y0); // 0 4 1 5
y0 = int32x4_max(y6,y0); // 2 6 3 7
y3 = int32x4_min(x3,y2); // 8 12 9 13
y2 = int32x4_max(x3,y2); // 10 14 11 15
x7 = int32x4_min(y7,x1); // 17 21 16 20
x1 = int32x4_max(y7,x1); // 19 23 18 22
x5 = int32x4_min(y5,x4); // 25 29 24 28
x4 = int32x4_max(y5,x4); // 27 31 26 30
// stage (5,0) 0:1 2:3 4:5 6:7 8:9 10:11 12:13 14:15 16:17 18:19 20:21 22:23 24:25 26:27 28:29 30:31
y6 = int32x4_leftleft(x6,y0); // 0 4 2 6
y0 = int32x4_rightright(x6,y0); // 1 5 3 7
x3 = int32x4_leftleft(y3,y2); // 8 12 10 14
y2 = int32x4_rightright(y3,y2); // 9 13 11 15
y7 = int32x4_leftleft(x7,x1); // 17 21 19 23
x1 = int32x4_rightright(x7,x1); // 16 20 18 22
y5 = int32x4_leftleft(x5,x4); // 25 29 27 31
x4 = int32x4_rightright(x5,x4); // 24 28 26 30
x6 = int32x4_min(y6,y0); // 0 4 2 6
y0 = int32x4_max(y6,y0); // 1 5 3 7
y3 = int32x4_min(x3,y2); // 8 12 10 14
y2 = int32x4_max(x3,y2); // 9 13 11 15
x7 = int32x4_min(y7,x1); // 16 20 18 22
x1 = int32x4_max(y7,x1); // 17 21 19 23
x5 = int32x4_min(y5,x4); // 24 28 26 30
x4 = int32x4_max(y5,x4); // 25 29 27 31
y6 = int32x4_a0b0a2b2(x6,y0); // 0 1 2 3
y0 = int32x4_a1b1a3b3(x6,y0); // 4 5 6 7
x3 = int32x4_a0b0a2b2(y3,y2); // 8 9 10 11
y2 = int32x4_a1b1a3b3(y3,y2); // 12 13 14 15
y7 = int32x4_a0b0a2b2(x7,x1); // 16 17 18 19
x1 = int32x4_a1b1a3b3(x7,x1); // 20 21 22 23
y5 = int32x4_a0b0a2b2(x5,x4); // 24 25 26 27
x4 = int32x4_a1b1a3b3(x5,x4); // 28 29 30 31
int32x4_store(x+0,vecxor^y6);
int32x4_store(x+4,vecxor^y0);
int32x4_store(x+8,vecxor^x3);
int32x4_store(x+12,vecxor^y2);
int32x4_store(x+16,vecxor^y7);
int32x4_store(x+20,vecxor^x1);
int32x4_store(x+24,vecxor^y5);
int32x4_store(x+28,vecxor^x4);
}
NOINLINE
static void int32_sort_64_xor(int32 *x,int32 xor)
{
int32x4 vecxor,x0,x1,x10,x11,x12,x13,x14,x15,x2,x3,x4,x5,x6,x7,x8,x9,y0,y1,y10,y11,y12,y13,y14,y15,y2,y3,y4,y5,y6,y7,y8,y9;
vecxor = int32x4_broadcast(xor);
x0 = vecxor^int32x4_load(x+0); // 0 16 32 48
x1 = vecxor^int32x4_load(x+4); // 1 17 33 49
x2 = vecxor^int32x4_load(x+8); // 2 18 34 50
x3 = vecxor^int32x4_load(x+12); // 3 19 35 51
x4 = vecxor^int32x4_load(x+16); // 4 20 36 52
x5 = vecxor^int32x4_load(x+20); // 5 21 37 53
x6 = vecxor^int32x4_load(x+24); // 6 22 38 54
x7 = vecxor^int32x4_load(x+28); // 7 23 39 55
x8 = vecxor^int32x4_load(x+32); // 8 24 40 56
x9 = vecxor^int32x4_load(x+36); // 9 25 41 57
x10 = vecxor^int32x4_load(x+40); // 10 26 42 58
x11 = vecxor^int32x4_load(x+44); // 11 27 43 59
x12 = vecxor^int32x4_load(x+48); // 12 28 44 60
x13 = vecxor^int32x4_load(x+52); // 13 29 45 61
x14 = vecxor^int32x4_load(x+56); // 14 30 46 62
x15 = vecxor^int32x4_load(x+60); // 15 31 47 63
// stage (1,0) 0:1 2:3 4:5 6:7 8:9 10:11 12:13 14:15 16:17 18:19 20:21 22:23 24:25 26:27 28:29 30:31 32:33 34:35 36:37 38:39 40:41 42:43 44:45 46:47 48:49 50:51 52:53 54:55 56:57 58:59 60:61 62:63
y0 = int32x4_min(x0,x1); // 0 16 32 48
x1 = int32x4_max(x0,x1); // 1 17 33 49
y2 = int32x4_min(x2,x3); // 2 18 34 50
x3 = int32x4_max(x2,x3); // 3 19 35 51
y4 = int32x4_min(x4,x5); // 4 20 36 52
x5 = int32x4_max(x4,x5); // 5 21 37 53
y6 = int32x4_min(x6,x7); // 6 22 38 54
x7 = int32x4_max(x6,x7); // 7 23 39 55
y8 = int32x4_min(x8,x9); // 8 24 40 56
x9 = int32x4_max(x8,x9); // 9 25 41 57
y10 = int32x4_min(x10,x11); // 10 26 42 58
x11 = int32x4_max(x10,x11); // 11 27 43 59
y12 = int32x4_min(x12,x13); // 12 28 44 60
x13 = int32x4_max(x12,x13); // 13 29 45 61
y14 = int32x4_min(x14,x15); // 14 30 46 62
x15 = int32x4_max(x14,x15); // 15 31 47 63
// stage (2,1) 0:2 1:3 4:6 5:7 8:10 9:11 12:14 13:15 16:18 17:19 20:22 21:23 24:26 25:27 28:30 29:31 32:34 33:35 36:38 37:39 40:42 41:43 44:46 45:47 48:50 49:51 52:54 53:55 56:58 57:59 60:62 61:63
x0 = int32x4_min(y0,y2); // 0 16 32 48
y2 = int32x4_max(y0,y2); // 2 18 34 50
y1 = int32x4_min(x1,x3); // 1 17 33 49
x3 = int32x4_max(x1,x3); // 3 19 35 51
x4 = int32x4_min(y4,y6); // 4 20 36 52
y6 = int32x4_max(y4,y6); // 6 22 38 54
y5 = int32x4_min(x5,x7); // 5 21 37 53
x7 = int32x4_max(x5,x7); // 7 23 39 55
x8 = int32x4_min(y8,y10); // 8 24 40 56
y10 = int32x4_max(y8,y10); // 10 26 42 58
y9 = int32x4_min(x9,x11); // 9 25 41 57
x11 = int32x4_max(x9,x11); // 11 27 43 59
x12 = int32x4_min(y12,y14); // 12 28 44 60
y14 = int32x4_max(y12,y14); // 14 30 46 62
y13 = int32x4_min(x13,x15); // 13 29 45 61
x15 = int32x4_max(x13,x15); // 15 31 47 63
// stage (2,0) 1:2 5:6 9:10 13:14 17:18 21:22 25:26 29:30 33:34 37:38 41:42 45:46 49:50 53:54 57:58 61:62
x2 = int32x4_min(y2,y1); // 1 17 33 49
y1 = int32x4_max(y2,y1); // 2 18 34 50
x6 = int32x4_min(y6,y5); // 5 21 37 53
y5 = int32x4_max(y6,y5); // 6 22 38 54
x10 = int32x4_min(y10,y9); // 9 25 41 57
y9 = int32x4_max(y10,y9); // 10 26 42 58
x14 = int32x4_min(y14,y13); // 13 29 45 61
y13 = int32x4_max(y14,y13); // 14 30 46 62
// stage (3,2) 0:4 1:5 2:6 3:7 8:12 9:13 10:14 11:15 16:20 17:21 18:22 19:23 24:28 25:29 26:30 27:31 32:36 33:37 34:38 35:39 40:44 41:45 42:46 43:47 48:52 49:53 50:54 51:55 56:60 57:61 58:62 59:63
y2 = int32x4_min(x2,x6); // 1 17 33 49
x6 = int32x4_max(x2,x6); // 5 21 37 53
y0 = int32x4_min(x0,x4); // 0 16 32 48
x4 = int32x4_max(x0,x4); // 4 20 36 52
x1 = int32x4_min(y1,y5); // 2 18 34 50
y5 = int32x4_max(y1,y5); // 6 22 38 54
y3 = int32x4_min(x3,x7); // 3 19 35 51
x7 = int32x4_max(x3,x7); // 7 23 39 55
y10 = int32x4_min(x10,x14); // 9 25 41 57
x14 = int32x4_max(x10,x14); // 13 29 45 61
y8 = int32x4_min(x8,x12); // 8 24 40 56
x12 = int32x4_max(x8,x12); // 12 28 44 60
x9 = int32x4_min(y9,y13); // 10 26 42 58
y13 = int32x4_max(y9,y13); // 14 30 46 62
y11 = int32x4_min(x11,x15); // 11 27 43 59
x15 = int32x4_max(x11,x15); // 15 31 47 63
// stage (3,1) 2:4 3:5 10:12 11:13 18:20 19:21 26:28 27:29 34:36 35:37 42:44 43:45 50:52 51:53 58:60 59:61
y6 = int32x4_min(x6,y3); // 3 19 35 51
y3 = int32x4_max(x6,y3); // 5 21 37 53
y4 = int32x4_min(x4,x1); // 2 18 34 50
x1 = int32x4_max(x4,x1); // 4 20 36 52
y14 = int32x4_min(x14,y11); // 11 27 43 59
y11 = int32x4_max(x14,y11); // 13 29 45 61
y12 = int32x4_min(x12,x9); // 10 26 42 58
x9 = int32x4_max(x12,x9); // 12 28 44 60
// stage (3,0) 1:2 3:4 5:6 9:10 11:12 13:14 17:18 19:20 21:22 25:26 27:28 29:30 33:34 35:36 37:38 41:42 43:44 45:46 49:50 51:52 53:54 57:58 59:60 61:62
x6 = int32x4_min(y6,x1); // 3 19 35 51
x1 = int32x4_max(y6,x1); // 4 20 36 52
x4 = int32x4_min(y4,y2); // 1 17 33 49
y2 = int32x4_max(y4,y2); // 2 18 34 50
x5 = int32x4_min(y5,y3); // 5 21 37 53
y3 = int32x4_max(y5,y3); // 6 22 38 54
x14 = int32x4_min(y14,x9); // 11 27 43 59
x9 = int32x4_max(y14,x9); // 12 28 44 60
x12 = int32x4_min(y12,y10); // 9 25 41 57
y10 = int32x4_max(y12,y10); // 10 26 42 58
x13 = int32x4_min(y13,y11); // 13 29 45 61
y11 = int32x4_max(y13,y11); // 14 30 46 62
// stage (4,3) 0:8 1:9 2:10 3:11 4:12 5:13 6:14 7:15 16:24 17:25 18:26 19:27 20:28 21:29 22:30 23:31 32:40 33:41 34:42 35:43 36:44 37:45 38:46 39:47 48:56 49:57 50:58 51:59 52:60 53:61 54:62 55:63
y6 = int32x4_min(x6,x14); // 3 19 35 51
x14 = int32x4_max(x6,x14); // 11 27 43 59
y4 = int32x4_min(x4,x12); // 1 17 33 49
x12 = int32x4_max(x4,x12); // 9 25 41 57
y5 = int32x4_min(x5,x13); // 5 21 37 53
x13 = int32x4_max(x5,x13); // 13 29 45 61
x0 = int32x4_min(y0,y8); // 0 16 32 48
y8 = int32x4_max(y0,y8); // 8 24 40 56
y1 = int32x4_min(x1,x9); // 4 20 36 52
x9 = int32x4_max(x1,x9); // 12 28 44 60
x2 = int32x4_min(y2,y10); // 2 18 34 50
y10 = int32x4_max(y2,y10); // 10 26 42 58
x3 = int32x4_min(y3,y11); // 6 22 38 54
y11 = int32x4_max(y3,y11); // 14 30 46 62
y7 = int32x4_min(x7,x15); // 7 23 39 55
x15 = int32x4_max(x7,x15); // 15 31 47 63
// stage (4,2) 4:8 5:9 6:10 7:11 20:24 21:25 22:26 23:27 36:40 37:41 38:42 39:43 52:56 53:57 54:58 55:59
y14 = int32x4_min(x14,y7); // 7 23 39 55
y7 = int32x4_max(x14,y7); // 11 27 43 59
y12 = int32x4_min(x12,y5); // 5 21 37 53
y5 = int32x4_max(x12,y5); // 9 25 41 57
x8 = int32x4_min(y8,y1); // 4 20 36 52
y1 = int32x4_max(y8,y1); // 8 24 40 56
x10 = int32x4_min(y10,x3); // 6 22 38 54
x3 = int32x4_max(y10,x3); // 10 26 42 58
// stage (4,1) 2:4 3:5 6:8 7:9 10:12 11:13 18:20 19:21 22:24 23:25 26:28 27:29 34:36 35:37 38:40 39:41 42:44 43:45 50:52 51:53 54:56 55:57 58:60 59:61
x14 = int32x4_min(y14,y5); // 7 23 39 55
y5 = int32x4_max(y14,y5); // 9 25 41 57
x12 = int32x4_min(y12,y6); // 3 19 35 51
y6 = int32x4_max(y12,y6); // 5 21 37 53
y13 = int32x4_min(x13,y7); // 11 27 43 59
y7 = int32x4_max(x13,y7); // 13 29 45 61
y8 = int32x4_min(x8,x2); // 2 18 34 50
x2 = int32x4_max(x8,x2); // 4 20 36 52
y9 = int32x4_min(x9,x3); // 10 26 42 58
x3 = int32x4_max(x9,x3); // 12 28 44 60
y10 = int32x4_min(x10,y1); // 6 22 38 54
y1 = int32x4_max(x10,y1); // 8 24 40 56
// stage (4,0) 1:2 3:4 5:6 7:8 9:10 11:12 13:14 17:18 19:20 21:22 23:24 25:26 27:28 29:30 33:34 35:36 37:38 39:40 41:42 43:44 45:46 49:50 51:52 53:54 55:56 57:58 59:60 61:62
y14 = int32x4_min(x14,y1); // 7 23 39 55
y1 = int32x4_max(x14,y1); // 8 24 40 56
y12 = int32x4_min(x12,x2); // 3 19 35 51
x2 = int32x4_max(x12,x2); // 4 20 36 52
x13 = int32x4_min(y13,x3); // 11 27 43 59
x3 = int32x4_max(y13,x3); // 12 28 44 60
x8 = int32x4_min(y8,y4); // 1 17 33 49
y4 = int32x4_max(y8,y4); // 2 18 34 50
x9 = int32x4_min(y9,y5); // 9 25 41 57
y5 = int32x4_max(y9,y5); // 10 26 42 58
x10 = int32x4_min(y10,y6); // 5 21 37 53
y6 = int32x4_max(y10,y6); // 6 22 38 54
x11 = int32x4_min(y11,y7); // 13 29 45 61
y7 = int32x4_max(y11,y7); // 14 30 46 62
// stage (5,4) 0:31 1:30 2:29 3:28 4:27 5:26 6:25 7:24 8:23 9:22 10:21 11:20 12:19 13:18 14:17 15:16 32:63 33:62 34:61 35:60 36:59 37:58 38:57 39:56 40:55 41:54 42:53 43:52 44:51 45:50 46:49 47:48
y1 = int32x4_1032(y1); // 24 8 56 40
x3 = int32x4_1032(x3); // 28 12 60 44
x2 = int32x4_1032(x2); // 20 4 52 36
y7 = int32x4_1032(y7); // 30 14 62 46
y6 = int32x4_1032(y6); // 22 6 54 38
y5 = int32x4_1032(y5); // 26 10 58 42
y4 = int32x4_1032(y4); // 18 2 50 34
x15 = int32x4_1032(x15); // 31 15 63 47
x14 = int32x4_min(y14,y1); // 7 8 39 40
y1 = int32x4_max(y14,y1); // 24 23 56 55
x12 = int32x4_min(y12,x3); // 3 12 35 44
x3 = int32x4_max(y12,x3); // 28 19 60 51
y13 = int32x4_min(x13,x2); // 11 4 43 36
x2 = int32x4_max(x13,x2); // 20 27 52 59
y8 = int32x4_min(x8,y7); // 1 14 33 46
y7 = int32x4_max(x8,y7); // 30 17 62 49
y9 = int32x4_min(x9,y6); // 9 6 41 38
y6 = int32x4_max(x9,y6); // 22 25 54 57
y10 = int32x4_min(x10,y5); // 5 10 37 42
y5 = int32x4_max(x10,y5); // 26 21 58 53
y11 = int32x4_min(x11,y4); // 13 2 45 34
y4 = int32x4_max(x11,y4); // 18 29 50 61
y0 = int32x4_min(x0,x15); // 0 15 32 47
x15 = int32x4_max(x0,x15); // 31 16 63 48
// stage (5,3) 0:8 1:9 2:10 3:11 4:12 5:13 6:14 7:15 16:24 17:25 18:26 19:27 20:28 21:29 22:30 23:31 32:40 33:41 34:42 35:43 36:44 37:45 38:46 39:47 48:56 49:57 50:58 51:59 52:60 53:61 54:62 55:63
y0 = int32x4_1032(y0); // 15 0 47 32
x15 = int32x4_1032(x15); // 16 31 48 63
y14 = int32x4_min(x14,y0); // 7 0 39 32
y0 = int32x4_max(x14,y0); // 15 8 47 40
y12 = int32x4_min(x12,y13); // 3 4 35 36
y13 = int32x4_max(x12,y13); // 11 12 43 44
y3 = int32x4_min(x3,x2); // 20 19 52 51
x2 = int32x4_max(x3,x2); // 28 27 60 59
x8 = int32x4_min(y8,y9); // 1 6 33 38
y9 = int32x4_max(y8,y9); // 9 14 41 46
x7 = int32x4_min(y7,y6); // 22 17 54 49
y6 = int32x4_max(y7,y6); // 30 25 62 57
x10 = int32x4_min(y10,y11); // 5 2 37 34
y11 = int32x4_max(y10,y11); // 13 10 45 42
x5 = int32x4_min(y5,y4); // 18 21 50 53
y4 = int32x4_max(y5,y4); // 26 29 58 61
x1 = int32x4_min(y1,x15); // 16 23 48 55
x15 = int32x4_max(y1,x15); // 24 31 56 63
// stage (5,2) 0:4 1:5 2:6 3:7 8:12 9:13 10:14 11:15 16:20 17:21 18:22 19:23 24:28 25:29 26:30 27:31 32:36 33:37 34:38 35:39 40:44 41:45 42:46 43:47 48:52 49:53 50:54 51:55 56:60 57:61 58:62 59:63
x14 = int32x4_min(y14,y12); // 3 0 35 32
y12 = int32x4_max(y14,y12); // 7 4 39 36
x0 = int32x4_min(y0,y13); // 11 8 43 40
y13 = int32x4_max(y0,y13); // 15 12 47 44
x3 = int32x4_min(y3,x1); // 16 19 48 51
x1 = int32x4_max(y3,x1); // 20 23 52 55
y8 = int32x4_min(x8,x10); // 1 2 33 34
x10 = int32x4_max(x8,x10); // 5 6 37 38
y7 = int32x4_min(x7,x5); // 18 17 50 49
x5 = int32x4_max(x7,x5); // 22 21 54 53
x9 = int32x4_min(y9,y11); // 9 10 41 42
y11 = int32x4_max(y9,y11); // 13 14 45 46
x6 = int32x4_min(y6,y4); // 26 25 58 57
y4 = int32x4_max(y6,y4); // 30 29 62 61
y2 = int32x4_min(x2,x15); // 24 27 56 59
x15 = int32x4_max(x2,x15); // 28 31 60 63
// stage (5,1) 0:2 1:3 4:6 5:7 8:10 9:11 12:14 13:15 16:18 17:19 20:22 21:23 24:26 25:27 28:30 29:31 32:34 33:35 36:38 37:39 40:42 41:43 44:46 45:47 48:50 49:51 52:54 53:55 56:58 57:59 60:62 61:63
y14 = int32x4_min(x14,y8); // 1 0 33 32
y8 = int32x4_max(x14,y8); // 3 2 35 34
y0 = int32x4_min(x0,x9); // 9 8 41 40
x9 = int32x4_max(x0,x9); // 11 10 43 42
y3 = int32x4_min(x3,y7); // 16 17 48 49
y7 = int32x4_max(x3,y7); // 18 19 50 51
x12 = int32x4_min(y12,x10); // 5 4 37 36
x10 = int32x4_max(y12,x10); // 7 6 39 38
y1 = int32x4_min(x1,x5); // 20 21 52 53
x5 = int32x4_max(x1,x5); // 22 23 54 55
x13 = int32x4_min(y13,y11); // 13 12 45 44
y11 = int32x4_max(y13,y11); // 15 14 47 46
y6 = int32x4_min(x6,y2); // 24 25 56 57
y2 = int32x4_max(x6,y2); // 26 27 58 59
x4 = int32x4_min(y4,x15); // 28 29 60 61
x15 = int32x4_max(y4,x15); // 30 31 62 63
// stage (5,0) 0:1 2:3 4:5 6:7 8:9 10:11 12:13 14:15 16:17 18:19 20:21 22:23 24:25 26:27 28:29 30:31 32:33 34:35 36:37 38:39 40:41 42:43 44:45 46:47 48:49 50:51 52:53 54:55 56:57 58:59 60:61 62:63
x14 = int32x4_a0b0a2b2(y14,y8); // 1 3 33 35
y8 = int32x4_a1b1a3b3(y14,y8); // 0 2 32 34
x0 = int32x4_a0b0a2b2(y0,x9); // 9 11 41 43
x9 = int32x4_a1b1a3b3(y0,x9); // 8 10 40 42
x3 = int32x4_a0b0a2b2(y3,y7); // 16 18 48 50
y7 = int32x4_a1b1a3b3(y3,y7); // 17 19 49 51
y12 = int32x4_a0b0a2b2(x12,x10); // 5 7 37 39
x10 = int32x4_a1b1a3b3(x12,x10); // 4 6 36 38
x1 = int32x4_a0b0a2b2(y1,x5); // 20 22 52 54
x5 = int32x4_a1b1a3b3(y1,x5); // 21 23 53 55
y13 = int32x4_a0b0a2b2(x13,y11); // 13 15 45 47
y11 = int32x4_a1b1a3b3(x13,y11); // 12 14 44 46
x6 = int32x4_a0b0a2b2(y6,y2); // 24 26 56 58
y2 = int32x4_a1b1a3b3(y6,y2); // 25 27 57 59
y4 = int32x4_a0b0a2b2(x4,x15); // 28 30 60 62
x15 = int32x4_a1b1a3b3(x4,x15); // 29 31 61 63
y14 = int32x4_min(x14,y8); // 0 2 32 34
y8 = int32x4_max(x14,y8); // 1 3 33 35
y0 = int32x4_min(x0,x9); // 8 10 40 42
x9 = int32x4_max(x0,x9); // 9 11 41 43
y3 = int32x4_min(x3,y7); // 16 18 48 50
y7 = int32x4_max(x3,y7); // 17 19 49 51
x12 = int32x4_min(y12,x10); // 4 6 36 38
x10 = int32x4_max(y12,x10); // 5 7 37 39
y1 = int32x4_min(x1,x5); // 20 22 52 54
x5 = int32x4_max(x1,x5); // 21 23 53 55
x13 = int32x4_min(y13,y11); // 12 14 44 46
y11 = int32x4_max(y13,y11); // 13 15 45 47
y6 = int32x4_min(x6,y2); // 24 26 56 58
y2 = int32x4_max(x6,y2); // 25 27 57 59
x4 = int32x4_min(y4,x15); // 28 30 60 62
x15 = int32x4_max(y4,x15); // 29 31 61 63
// stage (6,5) 0:63 1:62 2:61 3:60 4:59 5:58 6:57 7:56 8:55 9:54 10:53 11:52 12:51 13:50 14:49 15:48 16:47 17:46 18:45 19:44 20:43 21:42 22:41 23:40 24:39 25:38 26:37 27:36 28:35 29:34 30:33 31:32
x15 = int32x4_3210(x15); // 63 61 31 29
x5 = int32x4_3210(x5); // 55 53 23 21
y11 = int32x4_3210(y11); // 47 45 15 13
y2 = int32x4_3210(y2); // 59 57 27 25
x9 = int32x4_3210(x9); // 43 41 11 9
y7 = int32x4_3210(y7); // 51 49 19 17
x10 = int32x4_3210(x10); // 39 37 7 5
y8 = int32x4_3210(y8); // 35 33 3 1
x14 = int32x4_min(y14,x15); // 0 2 31 29
x15 = int32x4_max(y14,x15); // 63 61 32 34
x0 = int32x4_min(y0,x5); // 8 10 23 21
x5 = int32x4_max(y0,x5); // 55 53 40 42
x3 = int32x4_min(y3,y11); // 16 18 15 13
y11 = int32x4_max(y3,y11); // 47 45 48 50
y12 = int32x4_min(x12,y2); // 4 6 27 25
y2 = int32x4_max(x12,y2); // 59 57 36 38
x1 = int32x4_min(y1,x9); // 20 22 11 9
x9 = int32x4_max(y1,x9); // 43 41 52 54
y13 = int32x4_min(x13,y7); // 12 14 19 17
y7 = int32x4_max(x13,y7); // 51 49 44 46
x6 = int32x4_min(y6,x10); // 24 26 7 5
x10 = int32x4_max(y6,x10); // 39 37 56 58
y4 = int32x4_min(x4,y8); // 28 30 3 1
y8 = int32x4_max(x4,y8); // 35 33 60 62
// stage (6,4) 0:16 1:17 2:18 3:19 4:20 5:21 6:22 7:23 8:24 9:25 10:26 11:27 12:28 13:29 14:30 15:31 32:48 33:49 34:50 35:51 36:52 37:53 38:54 39:55 40:56 41:57 42:58 43:59 44:60 45:61 46:62 47:63
y14 = int32x4_min(x14,x3); // 0 2 15 13
x3 = int32x4_max(x14,x3); // 16 18 31 29
y0 = int32x4_min(x0,x6); // 8 10 7 5
x6 = int32x4_max(x0,x6); // 24 26 23 21
y15 = int32x4_min(x15,y11); // 47 45 32 34
y11 = int32x4_max(x15,y11); // 63 61 48 50
x12 = int32x4_min(y12,x1); // 4 6 11 9
x1 = int32x4_max(y12,x1); // 20 22 27 25
x2 = int32x4_min(y2,x9); // 43 41 36 38
x9 = int32x4_max(y2,x9); // 59 57 52 54
x13 = int32x4_min(y13,y4); // 12 14 3 1
y4 = int32x4_max(y13,y4); // 28 30 19 17
y5 = int32x4_min(x5,x10); // 39 37 40 42
x10 = int32x4_max(x5,x10); // 55 53 56 58
x7 = int32x4_min(y7,y8); // 35 33 44 46
y8 = int32x4_max(y7,y8); // 51 49 60 62
// stage (6,3) 0:8 1:9 2:10 3:11 4:12 5:13 6:14 7:15 16:24 17:25 18:26 19:27 20:28 21:29 22:30 23:31 32:40 33:41 34:42 35:43 36:44 37:45 38:46 39:47 48:56 49:57 50:58 51:59 52:60 53:61 54:62 55:63
x14 = int32x4_min(y14,y0); // 0 2 7 5
y0 = int32x4_max(y14,y0); // 8 10 15 13
y3 = int32x4_min(x3,x6); // 16 18 23 21
x6 = int32x4_max(x3,x6); // 24 26 31 29
x15 = int32x4_min(y15,y5); // 39 37 32 34
y5 = int32x4_max(y15,y5); // 47 45 40 42
y12 = int32x4_min(x12,x13); // 4 6 3 1
x13 = int32x4_max(x12,x13); // 12 14 11 9
y2 = int32x4_min(x2,x7); // 35 33 36 38
x7 = int32x4_max(x2,x7); // 43 41 44 46
y1 = int32x4_min(x1,y4); // 20 22 19 17
y4 = int32x4_max(x1,y4); // 28 30 27 25
x11 = int32x4_min(y11,x10); // 55 53 48 50
x10 = int32x4_max(y11,x10); // 63 61 56 58
y9 = int32x4_min(x9,y8); // 51 49 52 54
y8 = int32x4_max(x9,y8); // 59 57 60 62
// stage (6,2) 0:4 1:5 2:6 3:7 8:12 9:13 10:14 11:15 16:20 17:21 18:22 19:23 24:28 25:29 26:30 27:31 32:36 33:37 34:38 35:39 40:44 41:45 42:46 43:47 48:52 49:53 50:54 51:55 56:60 57:61 58:62 59:63
y14 = int32x4_min(x14,y12); // 0 2 3 1
y12 = int32x4_max(x14,y12); // 4 6 7 5
x3 = int32x4_min(y3,y1); // 16 18 19 17
y1 = int32x4_max(y3,y1); // 20 22 23 21
y15 = int32x4_min(x15,y2); // 35 33 32 34
y2 = int32x4_max(x15,y2); // 39 37 36 38
x0 = int32x4_min(y0,x13); // 8 10 11 9
x13 = int32x4_max(y0,x13); // 12 14 15 13
x5 = int32x4_min(y5,x7); // 43 41 40 42
x7 = int32x4_max(y5,x7); // 47 45 44 46
y6 = int32x4_min(x6,y4); // 24 26 27 25
y4 = int32x4_max(x6,y4); // 28 30 31 29
y11 = int32x4_min(x11,y9); // 51 49 48 50
y9 = int32x4_max(x11,y9); // 55 53 52 54
y10 = int32x4_min(x10,y8); // 59 57 56 58
y8 = int32x4_max(x10,y8); // 63 61 60 62
// stage (6,1) 0:2 1:3 4:6 5:7 8:10 9:11 12:14 13:15 16:18 17:19 20:22 21:23 24:26 25:27 28:30 29:31 32:34 33:35 36:38 37:39 40:42 41:43 44:46 45:47 48:50 49:51 52:54 53:55 56:58 57:59 60:62 61:63
x14 = int32x4_a0b0a2b2(y14,y12); // 0 4 3 7
y12 = int32x4_a1b1a3b3(y14,y12); // 2 6 1 5
y3 = int32x4_a0b0a2b2(x3,y1); // 16 20 19 23
y1 = int32x4_a1b1a3b3(x3,y1); // 18 22 17 21
x15 = int32x4_a0b0a2b2(y15,y2); // 35 39 32 36
y2 = int32x4_a1b1a3b3(y15,y2); // 33 37 34 38
y0 = int32x4_a0b0a2b2(x0,x13); // 8 12 11 15
x13 = int32x4_a1b1a3b3(x0,x13); // 10 14 9 13
y5 = int32x4_a0b0a2b2(x5,x7); // 43 47 40 44
x7 = int32x4_a1b1a3b3(x5,x7); // 41 45 42 46
x6 = int32x4_a0b0a2b2(y6,y4); // 24 28 27 31
y4 = int32x4_a1b1a3b3(y6,y4); // 26 30 25 29
x11 = int32x4_a0b0a2b2(y11,y9); // 51 55 48 52
y9 = int32x4_a1b1a3b3(y11,y9); // 49 53 50 54
x10 = int32x4_a0b0a2b2(y10,y8); // 59 63 56 60
y8 = int32x4_a1b1a3b3(y10,y8); // 57 61 58 62
y14 = int32x4_min(x14,y12); // 0 4 1 5
y12 = int32x4_max(x14,y12); // 2 6 3 7
x3 = int32x4_min(y3,y1); // 16 20 17 21
y1 = int32x4_max(y3,y1); // 18 22 19 23
y15 = int32x4_min(x15,y2); // 33 37 32 36
y2 = int32x4_max(x15,y2); // 35 39 34 38
x0 = int32x4_min(y0,x13); // 8 12 9 13
x13 = int32x4_max(y0,x13); // 10 14 11 15
x5 = int32x4_min(y5,x7); // 41 45 40 44
x7 = int32x4_max(y5,x7); // 43 47 42 46
y6 = int32x4_min(x6,y4); // 24 28 25 29
y4 = int32x4_max(x6,y4); // 26 30 27 31
y11 = int32x4_min(x11,y9); // 49 53 48 52
y9 = int32x4_max(x11,y9); // 51 55 50 54
y10 = int32x4_min(x10,y8); // 57 61 56 60
y8 = int32x4_max(x10,y8); // 59 63 58 62
// stage (6,0) 0:1 2:3 4:5 6:7 8:9 10:11 12:13 14:15 16:17 18:19 20:21 22:23 24:25 26:27 28:29 30:31 32:33 34:35 36:37 38:39 40:41 42:43 44:45 46:47 48:49 50:51 52:53 54:55 56:57 58:59 60:61 62:63
x14 = int32x4_leftleft(y14,y12); // 0 4 2 6
y12 = int32x4_rightright(y14,y12); // 1 5 3 7
y3 = int32x4_leftleft(x3,y1); // 16 20 18 22
y1 = int32x4_rightright(x3,y1); // 17 21 19 23
x15 = int32x4_leftleft(y15,y2); // 33 37 35 39
y2 = int32x4_rightright(y15,y2); // 32 36 34 38
y0 = int32x4_leftleft(x0,x13); // 8 12 10 14
x13 = int32x4_rightright(x0,x13); // 9 13 11 15
y5 = int32x4_leftleft(x5,x7); // 41 45 43 47
x7 = int32x4_rightright(x5,x7); // 40 44 42 46
x6 = int32x4_leftleft(y6,y4); // 24 28 26 30
y4 = int32x4_rightright(y6,y4); // 25 29 27 31
x11 = int32x4_leftleft(y11,y9); // 49 53 51 55
y9 = int32x4_rightright(y11,y9); // 48 52 50 54
x10 = int32x4_leftleft(y10,y8); // 57 61 59 63
y8 = int32x4_rightright(y10,y8); // 56 60 58 62
y14 = int32x4_min(x14,y12); // 0 4 2 6
y12 = int32x4_max(x14,y12); // 1 5 3 7
x3 = int32x4_min(y3,y1); // 16 20 18 22
y1 = int32x4_max(y3,y1); // 17 21 19 23
y15 = int32x4_min(x15,y2); // 32 36 34 38
y2 = int32x4_max(x15,y2); // 33 37 35 39
x0 = int32x4_min(y0,x13); // 8 12 10 14
x13 = int32x4_max(y0,x13); // 9 13 11 15
x5 = int32x4_min(y5,x7); // 40 44 42 46
x7 = int32x4_max(y5,x7); // 41 45 43 47
y6 = int32x4_min(x6,y4); // 24 28 26 30
y4 = int32x4_max(x6,y4); // 25 29 27 31
y11 = int32x4_min(x11,y9); // 48 52 50 54
y9 = int32x4_max(x11,y9); // 49 53 51 55
y10 = int32x4_min(x10,y8); // 56 60 58 62
y8 = int32x4_max(x10,y8); // 57 61 59 63
x14 = int32x4_a0b0a2b2(y14,y12); // 0 1 2 3
y12 = int32x4_a1b1a3b3(y14,y12); // 4 5 6 7
y0 = int32x4_a0b0a2b2(x0,x13); // 8 9 10 11
x13 = int32x4_a1b1a3b3(x0,x13); // 12 13 14 15
y3 = int32x4_a0b0a2b2(x3,y1); // 16 17 18 19
y1 = int32x4_a1b1a3b3(x3,y1); // 20 21 22 23
x6 = int32x4_a0b0a2b2(y6,y4); // 24 25 26 27
y4 = int32x4_a1b1a3b3(y6,y4); // 28 29 30 31
x15 = int32x4_a0b0a2b2(y15,y2); // 32 33 34 35
y2 = int32x4_a1b1a3b3(y15,y2); // 36 37 38 39
y5 = int32x4_a0b0a2b2(x5,x7); // 40 41 42 43
x7 = int32x4_a1b1a3b3(x5,x7); // 44 45 46 47
x11 = int32x4_a0b0a2b2(y11,y9); // 48 49 50 51
y9 = int32x4_a1b1a3b3(y11,y9); // 52 53 54 55
x10 = int32x4_a0b0a2b2(y10,y8); // 56 57 58 59
y8 = int32x4_a1b1a3b3(y10,y8); // 60 61 62 63
int32x4_store(x+0,vecxor^x14);
int32x4_store(x+4,vecxor^y12);
int32x4_store(x+8,vecxor^y0);
int32x4_store(x+12,vecxor^x13);
int32x4_store(x+16,vecxor^y3);
int32x4_store(x+20,vecxor^y1);
int32x4_store(x+24,vecxor^x6);
int32x4_store(x+28,vecxor^y4);
int32x4_store(x+32,vecxor^x15);
int32x4_store(x+36,vecxor^y2);
int32x4_store(x+40,vecxor^y5);
int32x4_store(x+44,vecxor^x7);
int32x4_store(x+48,vecxor^x11);
int32x4_store(x+52,vecxor^y9);
int32x4_store(x+56,vecxor^x10);
int32x4_store(x+60,vecxor^y8);
}
NOINLINE
static void int32_V_sort_8through16(int32 *x,long long n)
{
int32_t pos2,pos3;
int32x4 infty,x0,x1,x2,x3,y0,y1,y2;
x0 = int32x4_load(x+0); // 7 6 5 4
x1 = int32x4_load(x+4); // 3 2 1 0
infty = int32x4_broadcast(int32_largest);
pos2 = int32_min(12,n);
x2 = int32x4_ifthenelse(int32x4_smaller_umask(int32x4_set(8,9,10,11),int32x4_broadcast(pos2)),int32x4_from_int8x16(int8x16_varextract(int8x16_from_int32x4(int32x4_load(x+pos2-4)),uint8x16_add(uint8x16_set(0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15),uint8x16_broadcast(4*((-pos2)&3))))),infty); // 8 9 10 11
pos3 = int32_min(16,n);
x3 = int32x4_ifthenelse(int32x4_smaller_umask(int32x4_set(12,13,14,15),int32x4_broadcast(pos3)),int32x4_from_int8x16(int8x16_varextract(int8x16_from_int32x4(int32x4_load(x+pos3-4)),uint8x16_add(uint8x16_set(0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15),uint8x16_broadcast(4*((-pos3)&3))))),infty); // 12 13 14 15
// stage (4,3) 0:15 1:14 2:13 3:12 4:11 5:10 6:9 7:8
y0 = int32x4_min(x0,x2); // 7 6 5 4
x2 = int32x4_max(x0,x2); // 8 9 10 11
y1 = int32x4_min(x1,x3); // 3 2 1 0
x3 = int32x4_max(x1,x3); // 12 13 14 15
// stage (4,2) 0:4 1:5 2:6 3:7 8:12 9:13 10:14 11:15
x0 = int32x4_min(y0,y1); // 3 2 1 0
y1 = int32x4_max(y0,y1); // 7 6 5 4
y2 = int32x4_min(x2,x3); // 8 9 10 11
x3 = int32x4_max(x2,x3); // 12 13 14 15
// stage (4,1) 0:2 1:3 4:6 5:7 8:10 9:11 12:14 13:15
y0 = int32x4_leftleft(x0,y1); // 3 2 7 6
y1 = int32x4_rightright(x0,y1); // 1 0 5 4
x2 = int32x4_leftleft(y2,x3); // 8 9 12 13
x3 = int32x4_rightright(y2,x3); // 10 11 14 15
x0 = int32x4_min(y0,y1); // 1 0 5 4
y1 = int32x4_max(y0,y1); // 3 2 7 6
y2 = int32x4_min(x2,x3); // 8 9 12 13
x3 = int32x4_max(x2,x3); // 10 11 14 15
// stage (4,0) 0:1 2:3 4:5 6:7 8:9 10:11 12:13 14:15
y0 = int32x4_a0b0a2b2(x0,y1); // 1 3 5 7
y1 = int32x4_a1b1a3b3(x0,y1); // 0 2 4 6
x2 = int32x4_a0b0a2b2(y2,x3); // 8 10 12 14
x3 = int32x4_a1b1a3b3(y2,x3); // 9 11 13 15
x0 = int32x4_min(y0,y1); // 0 2 4 6
y1 = int32x4_max(y0,y1); // 1 3 5 7
y2 = int32x4_min(x2,x3); // 8 10 12 14
x3 = int32x4_max(x2,x3); // 9 11 13 15
y0 = int32x4_a0b0a1b1(x0,y1); // 0 1 2 3
y1 = int32x4_a2b2a3b3(x0,y1); // 4 5 6 7
x2 = int32x4_a0b0a1b1(y2,x3); // 8 9 10 11
x3 = int32x4_a2b2a3b3(y2,x3); // 12 13 14 15
int32x4_store(x+pos3-4,int32x4_from_int8x16(int8x16_varextract(int8x16_from_int32x4(x3),uint8x16_sub(uint8x16_set(0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15),uint8x16_broadcast(4*((-pos3)&3))))));
int32x4_store(x+pos2-4,int32x4_from_int8x16(int8x16_varextract(int8x16_from_int32x4(x2),uint8x16_sub(uint8x16_set(0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15),uint8x16_broadcast(4*((-pos2)&3))))));
int32x4_store(x+4,y1);
int32x4_store(x+0,y0);
}
NOINLINE
static void int32_V_sort_16through32(int32 *x,long long n)
{
int32_t pos4,pos5,pos6,pos7;
int32x4 infty,x0,x1,x2,x3,x4,x5,x6,x7,y0,y1,y2,y3,y4,y5,y6;
x0 = int32x4_load(x+0); // 15 14 13 12
x1 = int32x4_load(x+4); // 11 10 9 8
x2 = int32x4_load(x+8); // 7 6 5 4
x3 = int32x4_load(x+12); // 3 2 1 0
infty = int32x4_broadcast(int32_largest);
pos4 = int32_min(20,n);
x4 = int32x4_ifthenelse(int32x4_smaller_umask(int32x4_set(16,17,18,19),int32x4_broadcast(pos4)),int32x4_from_int8x16(int8x16_varextract(int8x16_from_int32x4(int32x4_load(x+pos4-4)),uint8x16_add(uint8x16_set(0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15),uint8x16_broadcast(4*((-pos4)&3))))),infty); // 16 17 18 19
pos5 = int32_min(24,n);
x5 = int32x4_ifthenelse(int32x4_smaller_umask(int32x4_set(20,21,22,23),int32x4_broadcast(pos5)),int32x4_from_int8x16(int8x16_varextract(int8x16_from_int32x4(int32x4_load(x+pos5-4)),uint8x16_add(uint8x16_set(0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15),uint8x16_broadcast(4*((-pos5)&3))))),infty); // 20 21 22 23
pos6 = int32_min(28,n);
x6 = int32x4_ifthenelse(int32x4_smaller_umask(int32x4_set(24,25,26,27),int32x4_broadcast(pos6)),int32x4_from_int8x16(int8x16_varextract(int8x16_from_int32x4(int32x4_load(x+pos6-4)),uint8x16_add(uint8x16_set(0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15),uint8x16_broadcast(4*((-pos6)&3))))),infty); // 24 25 26 27
pos7 = int32_min(32,n);
x7 = int32x4_ifthenelse(int32x4_smaller_umask(int32x4_set(28,29,30,31),int32x4_broadcast(pos7)),int32x4_from_int8x16(int8x16_varextract(int8x16_from_int32x4(int32x4_load(x+pos7-4)),uint8x16_add(uint8x16_set(0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15),uint8x16_broadcast(4*((-pos7)&3))))),infty); // 28 29 30 31
// stage (5,4) 0:31 1:30 2:29 3:28 4:27 5:26 6:25 7:24 8:23 9:22 10:21 11:20 12:19 13:18 14:17 15:16
y0 = int32x4_min(x0,x4); // 15 14 13 12
x4 = int32x4_max(x0,x4); // 16 17 18 19
y2 = int32x4_min(x2,x6); // 7 6 5 4
x6 = int32x4_max(x2,x6); // 24 25 26 27
y1 = int32x4_min(x1,x5); // 11 10 9 8
x5 = int32x4_max(x1,x5); // 20 21 22 23
y3 = int32x4_min(x3,x7); // 3 2 1 0
x7 = int32x4_max(x3,x7); // 28 29 30 31
// stage (5,3) 0:8 1:9 2:10 3:11 4:12 5:13 6:14 7:15 16:24 17:25 18:26 19:27 20:28 21:29 22:30 23:31
x0 = int32x4_min(y0,y2); // 7 6 5 4
y2 = int32x4_max(y0,y2); // 15 14 13 12
y4 = int32x4_min(x4,x6); // 16 17 18 19
x6 = int32x4_max(x4,x6); // 24 25 26 27
x1 = int32x4_min(y1,y3); // 3 2 1 0
y3 = int32x4_max(y1,y3); // 11 10 9 8
y5 = int32x4_min(x5,x7); // 20 21 22 23
x7 = int32x4_max(x5,x7); // 28 29 30 31
// stage (5,2) 0:4 1:5 2:6 3:7 8:12 9:13 10:14 11:15 16:20 17:21 18:22 19:23 24:28 25:29 26:30 27:31
y0 = int32x4_min(x0,x1); // 3 2 1 0
x1 = int32x4_max(x0,x1); // 7 6 5 4
x4 = int32x4_min(y4,y5); // 16 17 18 19
y5 = int32x4_max(y4,y5); // 20 21 22 23
x2 = int32x4_min(y2,y3); // 11 10 9 8
y3 = int32x4_max(y2,y3); // 15 14 13 12
y6 = int32x4_min(x6,x7); // 24 25 26 27
x7 = int32x4_max(x6,x7); // 28 29 30 31
// stage (5,1) 0:2 1:3 4:6 5:7 8:10 9:11 12:14 13:15 16:18 17:19 20:22 21:23 24:26 25:27 28:30 29:31
x0 = int32x4_leftleft(y0,x1); // 3 2 7 6
x1 = int32x4_rightright(y0,x1); // 1 0 5 4
y4 = int32x4_leftleft(x4,y5); // 16 17 20 21
y5 = int32x4_rightright(x4,y5); // 18 19 22 23
y2 = int32x4_leftleft(x2,y3); // 11 10 15 14
y3 = int32x4_rightright(x2,y3); // 9 8 13 12
x6 = int32x4_leftleft(y6,x7); // 24 25 28 29
x7 = int32x4_rightright(y6,x7); // 26 27 30 31
y0 = int32x4_min(x0,x1); // 1 0 5 4
x1 = int32x4_max(x0,x1); // 3 2 7 6
x4 = int32x4_min(y4,y5); // 16 17 20 21
y5 = int32x4_max(y4,y5); // 18 19 22 23
x2 = int32x4_min(y2,y3); // 9 8 13 12
y3 = int32x4_max(y2,y3); // 11 10 15 14
y6 = int32x4_min(x6,x7); // 24 25 28 29
x7 = int32x4_max(x6,x7); // 26 27 30 31
// stage (5,0) 0:1 2:3 4:5 6:7 8:9 10:11 12:13 14:15 16:17 18:19 20:21 22:23 24:25 26:27 28:29 30:31
x0 = int32x4_a0b0a2b2(y0,x1); // 1 3 5 7
x1 = int32x4_a1b1a3b3(y0,x1); // 0 2 4 6
y4 = int32x4_a0b0a2b2(x4,y5); // 16 18 20 22
y5 = int32x4_a1b1a3b3(x4,y5); // 17 19 21 23
y2 = int32x4_a0b0a2b2(x2,y3); // 9 11 13 15
y3 = int32x4_a1b1a3b3(x2,y3); // 8 10 12 14
x6 = int32x4_a0b0a2b2(y6,x7); // 24 26 28 30
x7 = int32x4_a1b1a3b3(y6,x7); // 25 27 29 31
y0 = int32x4_min(x0,x1); // 0 2 4 6
x1 = int32x4_max(x0,x1); // 1 3 5 7
x4 = int32x4_min(y4,y5); // 16 18 20 22
y5 = int32x4_max(y4,y5); // 17 19 21 23
x2 = int32x4_min(y2,y3); // 8 10 12 14
y3 = int32x4_max(y2,y3); // 9 11 13 15
y6 = int32x4_min(x6,x7); // 24 26 28 30
x7 = int32x4_max(x6,x7); // 25 27 29 31
x0 = int32x4_a0b0a1b1(y0,x1); // 0 1 2 3
x1 = int32x4_a2b2a3b3(y0,x1); // 4 5 6 7
y2 = int32x4_a0b0a1b1(x2,y3); // 8 9 10 11
y3 = int32x4_a2b2a3b3(x2,y3); // 12 13 14 15
y4 = int32x4_a0b0a1b1(x4,y5); // 16 17 18 19
y5 = int32x4_a2b2a3b3(x4,y5); // 20 21 22 23
x6 = int32x4_a0b0a1b1(y6,x7); // 24 25 26 27
x7 = int32x4_a2b2a3b3(y6,x7); // 28 29 30 31
int32x4_store(x+pos7-4,int32x4_from_int8x16(int8x16_varextract(int8x16_from_int32x4(x7),uint8x16_sub(uint8x16_set(0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15),uint8x16_broadcast(4*((-pos7)&3))))));
int32x4_store(x+pos6-4,int32x4_from_int8x16(int8x16_varextract(int8x16_from_int32x4(x6),uint8x16_sub(uint8x16_set(0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15),uint8x16_broadcast(4*((-pos6)&3))))));
int32x4_store(x+pos5-4,int32x4_from_int8x16(int8x16_varextract(int8x16_from_int32x4(y5),uint8x16_sub(uint8x16_set(0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15),uint8x16_broadcast(4*((-pos5)&3))))));
int32x4_store(x+pos4-4,int32x4_from_int8x16(int8x16_varextract(int8x16_from_int32x4(y4),uint8x16_sub(uint8x16_set(0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15),uint8x16_broadcast(4*((-pos4)&3))))));
int32x4_store(x+12,y3);
int32x4_store(x+8,y2);
int32x4_store(x+4,x1);
int32x4_store(x+0,x0);
}
NOINLINE
static void int32_V_sort_32through64(int32 *x,long long n)
{
int32_t pos10,pos11,pos12,pos13,pos14,pos15,pos8,pos9;
int32x4 infty,x0,x1,x10,x11,x12,x13,x14,x15,x2,x3,x4,x5,x6,x7,x8,x9,y0,y1,y10,y11,y12,y13,y14,y2,y3,y4,y5,y6,y7,y8,y9;
x0 = int32x4_load(x+0); // 31 30 29 28
x1 = int32x4_load(x+4); // 27 26 25 24
x2 = int32x4_load(x+8); // 23 22 21 20
x3 = int32x4_load(x+12); // 19 18 17 16
x4 = int32x4_load(x+16); // 15 14 13 12
x5 = int32x4_load(x+20); // 11 10 9 8
x6 = int32x4_load(x+24); // 7 6 5 4
x7 = int32x4_load(x+28); // 3 2 1 0
infty = int32x4_broadcast(int32_largest);
pos8 = int32_min(36,n);
x8 = int32x4_ifthenelse(int32x4_smaller_umask(int32x4_set(32,33,34,35),int32x4_broadcast(pos8)),int32x4_from_int8x16(int8x16_varextract(int8x16_from_int32x4(int32x4_load(x+pos8-4)),uint8x16_add(uint8x16_set(0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15),uint8x16_broadcast(4*((-pos8)&3))))),infty); // 32 33 34 35
pos9 = int32_min(40,n);
x9 = int32x4_ifthenelse(int32x4_smaller_umask(int32x4_set(36,37,38,39),int32x4_broadcast(pos9)),int32x4_from_int8x16(int8x16_varextract(int8x16_from_int32x4(int32x4_load(x+pos9-4)),uint8x16_add(uint8x16_set(0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15),uint8x16_broadcast(4*((-pos9)&3))))),infty); // 36 37 38 39
pos10 = int32_min(44,n);
x10 = int32x4_ifthenelse(int32x4_smaller_umask(int32x4_set(40,41,42,43),int32x4_broadcast(pos10)),int32x4_from_int8x16(int8x16_varextract(int8x16_from_int32x4(int32x4_load(x+pos10-4)),uint8x16_add(uint8x16_set(0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15),uint8x16_broadcast(4*((-pos10)&3))))),infty); // 40 41 42 43
pos11 = int32_min(48,n);
x11 = int32x4_ifthenelse(int32x4_smaller_umask(int32x4_set(44,45,46,47),int32x4_broadcast(pos11)),int32x4_from_int8x16(int8x16_varextract(int8x16_from_int32x4(int32x4_load(x+pos11-4)),uint8x16_add(uint8x16_set(0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15),uint8x16_broadcast(4*((-pos11)&3))))),infty); // 44 45 46 47
pos12 = int32_min(52,n);
x12 = int32x4_ifthenelse(int32x4_smaller_umask(int32x4_set(48,49,50,51),int32x4_broadcast(pos12)),int32x4_from_int8x16(int8x16_varextract(int8x16_from_int32x4(int32x4_load(x+pos12-4)),uint8x16_add(uint8x16_set(0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15),uint8x16_broadcast(4*((-pos12)&3))))),infty); // 48 49 50 51
pos13 = int32_min(56,n);
x13 = int32x4_ifthenelse(int32x4_smaller_umask(int32x4_set(52,53,54,55),int32x4_broadcast(pos13)),int32x4_from_int8x16(int8x16_varextract(int8x16_from_int32x4(int32x4_load(x+pos13-4)),uint8x16_add(uint8x16_set(0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15),uint8x16_broadcast(4*((-pos13)&3))))),infty); // 52 53 54 55
pos14 = int32_min(60,n);
x14 = int32x4_ifthenelse(int32x4_smaller_umask(int32x4_set(56,57,58,59),int32x4_broadcast(pos14)),int32x4_from_int8x16(int8x16_varextract(int8x16_from_int32x4(int32x4_load(x+pos14-4)),uint8x16_add(uint8x16_set(0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15),uint8x16_broadcast(4*((-pos14)&3))))),infty); // 56 57 58 59
pos15 = int32_min(64,n);
x15 = int32x4_ifthenelse(int32x4_smaller_umask(int32x4_set(60,61,62,63),int32x4_broadcast(pos15)),int32x4_from_int8x16(int8x16_varextract(int8x16_from_int32x4(int32x4_load(x+pos15-4)),uint8x16_add(uint8x16_set(0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15),uint8x16_broadcast(4*((-pos15)&3))))),infty); // 60 61 62 63
// stage (6,5) 0:63 1:62 2:61 3:60 4:59 5:58 6:57 7:56 8:55 9:54 10:53 11:52 12:51 13:50 14:49 15:48 16:47 17:46 18:45 19:44 20:43 21:42 22:41 23:40 24:39 25:38 26:37 27:36 28:35 29:34 30:33 31:32
y0 = int32x4_min(x0,x8); // 31 30 29 28
x8 = int32x4_max(x0,x8); // 32 33 34 35
y2 = int32x4_min(x2,x10); // 23 22 21 20
x10 = int32x4_max(x2,x10); // 40 41 42 43
y4 = int32x4_min(x4,x12); // 15 14 13 12
x12 = int32x4_max(x4,x12); // 48 49 50 51
y6 = int32x4_min(x6,x14); // 7 6 5 4
x14 = int32x4_max(x6,x14); // 56 57 58 59
y1 = int32x4_min(x1,x9); // 27 26 25 24
x9 = int32x4_max(x1,x9); // 36 37 38 39
y3 = int32x4_min(x3,x11); // 19 18 17 16
x11 = int32x4_max(x3,x11); // 44 45 46 47
y5 = int32x4_min(x5,x13); // 11 10 9 8
x13 = int32x4_max(x5,x13); // 52 53 54 55
y7 = int32x4_min(x7,x15); // 3 2 1 0
x15 = int32x4_max(x7,x15); // 60 61 62 63
// stage (6,4) 0:16 1:17 2:18 3:19 4:20 5:21 6:22 7:23 8:24 9:25 10:26 11:27 12:28 13:29 14:30 15:31 32:48 33:49 34:50 35:51 36:52 37:53 38:54 39:55 40:56 41:57 42:58 43:59 44:60 45:61 46:62 47:63
x0 = int32x4_min(y0,y4); // 15 14 13 12
y4 = int32x4_max(y0,y4); // 31 30 29 28
x2 = int32x4_min(y2,y6); // 7 6 5 4
y6 = int32x4_max(y2,y6); // 23 22 21 20
y8 = int32x4_min(x8,x12); // 32 33 34 35
x12 = int32x4_max(x8,x12); // 48 49 50 51
y10 = int32x4_min(x10,x14); // 40 41 42 43
x14 = int32x4_max(x10,x14); // 56 57 58 59
x1 = int32x4_min(y1,y5); // 11 10 9 8
y5 = int32x4_max(y1,y5); // 27 26 25 24
x3 = int32x4_min(y3,y7); // 3 2 1 0
y7 = int32x4_max(y3,y7); // 19 18 17 16
y9 = int32x4_min(x9,x13); // 36 37 38 39
x13 = int32x4_max(x9,x13); // 52 53 54 55
y11 = int32x4_min(x11,x15); // 44 45 46 47
x15 = int32x4_max(x11,x15); // 60 61 62 63
// stage (6,3) 0:8 1:9 2:10 3:11 4:12 5:13 6:14 7:15 16:24 17:25 18:26 19:27 20:28 21:29 22:30 23:31 32:40 33:41 34:42 35:43 36:44 37:45 38:46 39:47 48:56 49:57 50:58 51:59 52:60 53:61 54:62 55:63
y0 = int32x4_min(x0,x2); // 7 6 5 4
x2 = int32x4_max(x0,x2); // 15 14 13 12
x4 = int32x4_min(y4,y6); // 23 22 21 20
y6 = int32x4_max(y4,y6); // 31 30 29 28
x8 = int32x4_min(y8,y10); // 32 33 34 35
y10 = int32x4_max(y8,y10); // 40 41 42 43
y12 = int32x4_min(x12,x14); // 48 49 50 51
x14 = int32x4_max(x12,x14); // 56 57 58 59
y1 = int32x4_min(x1,x3); // 3 2 1 0
x3 = int32x4_max(x1,x3); // 11 10 9 8
x5 = int32x4_min(y5,y7); // 19 18 17 16
y7 = int32x4_max(y5,y7); // 27 26 25 24
x9 = int32x4_min(y9,y11); // 36 37 38 39
y11 = int32x4_max(y9,y11); // 44 45 46 47
y13 = int32x4_min(x13,x15); // 52 53 54 55
x15 = int32x4_max(x13,x15); // 60 61 62 63
// stage (6,2) 0:4 1:5 2:6 3:7 8:12 9:13 10:14 11:15 16:20 17:21 18:22 19:23 24:28 25:29 26:30 27:31 32:36 33:37 34:38 35:39 40:44 41:45 42:46 43:47 48:52 49:53 50:54 51:55 56:60 57:61 58:62 59:63
x0 = int32x4_min(y0,y1); // 3 2 1 0
y1 = int32x4_max(y0,y1); // 7 6 5 4
y4 = int32x4_min(x4,x5); // 19 18 17 16
x5 = int32x4_max(x4,x5); // 23 22 21 20
y8 = int32x4_min(x8,x9); // 32 33 34 35
x9 = int32x4_max(x8,x9); // 36 37 38 39
x12 = int32x4_min(y12,y13); // 48 49 50 51
y13 = int32x4_max(y12,y13); // 52 53 54 55
y2 = int32x4_min(x2,x3); // 11 10 9 8
x3 = int32x4_max(x2,x3); // 15 14 13 12
x6 = int32x4_min(y6,y7); // 27 26 25 24
y7 = int32x4_max(y6,y7); // 31 30 29 28
x10 = int32x4_min(y10,y11); // 40 41 42 43
y11 = int32x4_max(y10,y11); // 44 45 46 47
y14 = int32x4_min(x14,x15); // 56 57 58 59
x15 = int32x4_max(x14,x15); // 60 61 62 63
// stage (6,1) 0:2 1:3 4:6 5:7 8:10 9:11 12:14 13:15 16:18 17:19 20:22 21:23 24:26 25:27 28:30 29:31 32:34 33:35 36:38 37:39 40:42 41:43 44:46 45:47 48:50 49:51 52:54 53:55 56:58 57:59 60:62 61:63
y0 = int32x4_leftleft(x0,y1); // 3 2 7 6
y1 = int32x4_rightright(x0,y1); // 1 0 5 4
x4 = int32x4_leftleft(y4,x5); // 19 18 23 22
x5 = int32x4_rightright(y4,x5); // 17 16 21 20
x8 = int32x4_leftleft(y8,x9); // 32 33 36 37
x9 = int32x4_rightright(y8,x9); // 34 35 38 39
y12 = int32x4_leftleft(x12,y13); // 48 49 52 53
y13 = int32x4_rightright(x12,y13); // 50 51 54 55
x2 = int32x4_leftleft(y2,x3); // 11 10 15 14
x3 = int32x4_rightright(y2,x3); // 9 8 13 12
y6 = int32x4_leftleft(x6,y7); // 27 26 31 30
y7 = int32x4_rightright(x6,y7); // 25 24 29 28
y10 = int32x4_leftleft(x10,y11); // 40 41 44 45
y11 = int32x4_rightright(x10,y11); // 42 43 46 47
x14 = int32x4_leftleft(y14,x15); // 56 57 60 61
x15 = int32x4_rightright(y14,x15); // 58 59 62 63
x0 = int32x4_min(y0,y1); // 1 0 5 4
y1 = int32x4_max(y0,y1); // 3 2 7 6
y4 = int32x4_min(x4,x5); // 17 16 21 20
x5 = int32x4_max(x4,x5); // 19 18 23 22
y8 = int32x4_min(x8,x9); // 32 33 36 37
x9 = int32x4_max(x8,x9); // 34 35 38 39
x12 = int32x4_min(y12,y13); // 48 49 52 53
y13 = int32x4_max(y12,y13); // 50 51 54 55
y2 = int32x4_min(x2,x3); // 9 8 13 12
x3 = int32x4_max(x2,x3); // 11 10 15 14
x6 = int32x4_min(y6,y7); // 25 24 29 28
y7 = int32x4_max(y6,y7); // 27 26 31 30
x10 = int32x4_min(y10,y11); // 40 41 44 45
y11 = int32x4_max(y10,y11); // 42 43 46 47
y14 = int32x4_min(x14,x15); // 56 57 60 61
x15 = int32x4_max(x14,x15); // 58 59 62 63
// stage (6,0) 0:1 2:3 4:5 6:7 8:9 10:11 12:13 14:15 16:17 18:19 20:21 22:23 24:25 26:27 28:29 30:31 32:33 34:35 36:37 38:39 40:41 42:43 44:45 46:47 48:49 50:51 52:53 54:55 56:57 58:59 60:61 62:63
y0 = int32x4_a0b0a2b2(x0,y1); // 1 3 5 7
y1 = int32x4_a1b1a3b3(x0,y1); // 0 2 4 6
x4 = int32x4_a0b0a2b2(y4,x5); // 17 19 21 23
x5 = int32x4_a1b1a3b3(y4,x5); // 16 18 20 22
x8 = int32x4_a0b0a2b2(y8,x9); // 32 34 36 38
x9 = int32x4_a1b1a3b3(y8,x9); // 33 35 37 39
y12 = int32x4_a0b0a2b2(x12,y13); // 48 50 52 54
y13 = int32x4_a1b1a3b3(x12,y13); // 49 51 53 55
x2 = int32x4_a0b0a2b2(y2,x3); // 9 11 13 15
x3 = int32x4_a1b1a3b3(y2,x3); // 8 10 12 14
y6 = int32x4_a0b0a2b2(x6,y7); // 25 27 29 31
y7 = int32x4_a1b1a3b3(x6,y7); // 24 26 28 30
y10 = int32x4_a0b0a2b2(x10,y11); // 40 42 44 46
y11 = int32x4_a1b1a3b3(x10,y11); // 41 43 45 47
x14 = int32x4_a0b0a2b2(y14,x15); // 56 58 60 62
x15 = int32x4_a1b1a3b3(y14,x15); // 57 59 61 63
x0 = int32x4_min(y0,y1); // 0 2 4 6
y1 = int32x4_max(y0,y1); // 1 3 5 7
y4 = int32x4_min(x4,x5); // 16 18 20 22
x5 = int32x4_max(x4,x5); // 17 19 21 23
y8 = int32x4_min(x8,x9); // 32 34 36 38
x9 = int32x4_max(x8,x9); // 33 35 37 39
x12 = int32x4_min(y12,y13); // 48 50 52 54
y13 = int32x4_max(y12,y13); // 49 51 53 55
y2 = int32x4_min(x2,x3); // 8 10 12 14
x3 = int32x4_max(x2,x3); // 9 11 13 15
x6 = int32x4_min(y6,y7); // 24 26 28 30
y7 = int32x4_max(y6,y7); // 25 27 29 31
x10 = int32x4_min(y10,y11); // 40 42 44 46
y11 = int32x4_max(y10,y11); // 41 43 45 47
y14 = int32x4_min(x14,x15); // 56 58 60 62
x15 = int32x4_max(x14,x15); // 57 59 61 63
y0 = int32x4_a0b0a1b1(x0,y1); // 0 1 2 3
y1 = int32x4_a2b2a3b3(x0,y1); // 4 5 6 7
x2 = int32x4_a0b0a1b1(y2,x3); // 8 9 10 11
x3 = int32x4_a2b2a3b3(y2,x3); // 12 13 14 15
x4 = int32x4_a0b0a1b1(y4,x5); // 16 17 18 19
x5 = int32x4_a2b2a3b3(y4,x5); // 20 21 22 23
y6 = int32x4_a0b0a1b1(x6,y7); // 24 25 26 27
y7 = int32x4_a2b2a3b3(x6,y7); // 28 29 30 31
x8 = int32x4_a0b0a1b1(y8,x9); // 32 33 34 35
x9 = int32x4_a2b2a3b3(y8,x9); // 36 37 38 39
y10 = int32x4_a0b0a1b1(x10,y11); // 40 41 42 43
y11 = int32x4_a2b2a3b3(x10,y11); // 44 45 46 47
y12 = int32x4_a0b0a1b1(x12,y13); // 48 49 50 51
y13 = int32x4_a2b2a3b3(x12,y13); // 52 53 54 55
x14 = int32x4_a0b0a1b1(y14,x15); // 56 57 58 59
x15 = int32x4_a2b2a3b3(y14,x15); // 60 61 62 63
int32x4_store(x+pos15-4,int32x4_from_int8x16(int8x16_varextract(int8x16_from_int32x4(x15),uint8x16_sub(uint8x16_set(0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15),uint8x16_broadcast(4*((-pos15)&3))))));
int32x4_store(x+pos14-4,int32x4_from_int8x16(int8x16_varextract(int8x16_from_int32x4(x14),uint8x16_sub(uint8x16_set(0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15),uint8x16_broadcast(4*((-pos14)&3))))));
int32x4_store(x+pos13-4,int32x4_from_int8x16(int8x16_varextract(int8x16_from_int32x4(y13),uint8x16_sub(uint8x16_set(0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15),uint8x16_broadcast(4*((-pos13)&3))))));
int32x4_store(x+pos12-4,int32x4_from_int8x16(int8x16_varextract(int8x16_from_int32x4(y12),uint8x16_sub(uint8x16_set(0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15),uint8x16_broadcast(4*((-pos12)&3))))));
int32x4_store(x+pos11-4,int32x4_from_int8x16(int8x16_varextract(int8x16_from_int32x4(y11),uint8x16_sub(uint8x16_set(0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15),uint8x16_broadcast(4*((-pos11)&3))))));
int32x4_store(x+pos10-4,int32x4_from_int8x16(int8x16_varextract(int8x16_from_int32x4(y10),uint8x16_sub(uint8x16_set(0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15),uint8x16_broadcast(4*((-pos10)&3))))));
int32x4_store(x+pos9-4,int32x4_from_int8x16(int8x16_varextract(int8x16_from_int32x4(x9),uint8x16_sub(uint8x16_set(0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15),uint8x16_broadcast(4*((-pos9)&3))))));
int32x4_store(x+pos8-4,int32x4_from_int8x16(int8x16_varextract(int8x16_from_int32x4(x8),uint8x16_sub(uint8x16_set(0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15),uint8x16_broadcast(4*((-pos8)&3))))));
int32x4_store(x+28,y7);
int32x4_store(x+24,y6);
int32x4_store(x+20,x5);
int32x4_store(x+16,x4);
int32x4_store(x+12,x3);
int32x4_store(x+8,x2);
int32x4_store(x+4,y1);
int32x4_store(x+0,y0);
}
NOINLINE
static void int32_V_sort_16_xor(int32 *x,int32 xor)
{
int32x4 vecxor,x0,x1,x2,x3,y0,y1,y2;
vecxor = int32x4_broadcast(xor);
x0 = vecxor^int32x4_load(x+0); // 7 6 5 4
x1 = vecxor^int32x4_load(x+4); // 3 2 1 0
x2 = vecxor^int32x4_load(x+8); // 8 9 10 11
x3 = vecxor^int32x4_load(x+12); // 12 13 14 15
// stage (4,3) 0:15 1:14 2:13 3:12 4:11 5:10 6:9 7:8
y0 = int32x4_min(x0,x2); // 7 6 5 4
x2 = int32x4_max(x0,x2); // 8 9 10 11
y1 = int32x4_min(x1,x3); // 3 2 1 0
x3 = int32x4_max(x1,x3); // 12 13 14 15
// stage (4,2) 0:4 1:5 2:6 3:7 8:12 9:13 10:14 11:15
x0 = int32x4_min(y0,y1); // 3 2 1 0
y1 = int32x4_max(y0,y1); // 7 6 5 4
y2 = int32x4_min(x2,x3); // 8 9 10 11
x3 = int32x4_max(x2,x3); // 12 13 14 15
// stage (4,1) 0:2 1:3 4:6 5:7 8:10 9:11 12:14 13:15
y0 = int32x4_leftleft(x0,y1); // 3 2 7 6
y1 = int32x4_rightright(x0,y1); // 1 0 5 4
x2 = int32x4_leftleft(y2,x3); // 8 9 12 13
x3 = int32x4_rightright(y2,x3); // 10 11 14 15
x0 = int32x4_min(y0,y1); // 1 0 5 4
y1 = int32x4_max(y0,y1); // 3 2 7 6
y2 = int32x4_min(x2,x3); // 8 9 12 13
x3 = int32x4_max(x2,x3); // 10 11 14 15
// stage (4,0) 0:1 2:3 4:5 6:7 8:9 10:11 12:13 14:15
y0 = int32x4_a0b0a2b2(x0,y1); // 1 3 5 7
y1 = int32x4_a1b1a3b3(x0,y1); // 0 2 4 6
x2 = int32x4_a0b0a2b2(y2,x3); // 8 10 12 14
x3 = int32x4_a1b1a3b3(y2,x3); // 9 11 13 15
x0 = int32x4_min(y0,y1); // 0 2 4 6
y1 = int32x4_max(y0,y1); // 1 3 5 7
y2 = int32x4_min(x2,x3); // 8 10 12 14
x3 = int32x4_max(x2,x3); // 9 11 13 15
y0 = int32x4_a0b0a1b1(x0,y1); // 0 1 2 3
y1 = int32x4_a2b2a3b3(x0,y1); // 4 5 6 7
x2 = int32x4_a0b0a1b1(y2,x3); // 8 9 10 11
x3 = int32x4_a2b2a3b3(y2,x3); // 12 13 14 15
int32x4_store(x+0,vecxor^y0);
int32x4_store(x+4,vecxor^y1);
int32x4_store(x+8,vecxor^x2);
int32x4_store(x+12,vecxor^x3);
}
NOINLINE
static void int32_V_sort_32_xor(int32 *x,int32 xor)
{
int32x4 vecxor,x0,x1,x2,x3,x4,x5,x6,x7,y0,y1,y2,y3,y4,y5,y6;
vecxor = int32x4_broadcast(xor);
x0 = vecxor^int32x4_load(x+0); // 15 14 13 12
x1 = vecxor^int32x4_load(x+4); // 11 10 9 8
x2 = vecxor^int32x4_load(x+8); // 7 6 5 4
x3 = vecxor^int32x4_load(x+12); // 3 2 1 0
x4 = vecxor^int32x4_load(x+16); // 16 17 18 19
x5 = vecxor^int32x4_load(x+20); // 20 21 22 23
x6 = vecxor^int32x4_load(x+24); // 24 25 26 27
x7 = vecxor^int32x4_load(x+28); // 28 29 30 31
// stage (5,4) 0:31 1:30 2:29 3:28 4:27 5:26 6:25 7:24 8:23 9:22 10:21 11:20 12:19 13:18 14:17 15:16
y0 = int32x4_min(x0,x4); // 15 14 13 12
x4 = int32x4_max(x0,x4); // 16 17 18 19
y2 = int32x4_min(x2,x6); // 7 6 5 4
x6 = int32x4_max(x2,x6); // 24 25 26 27
y1 = int32x4_min(x1,x5); // 11 10 9 8
x5 = int32x4_max(x1,x5); // 20 21 22 23
y3 = int32x4_min(x3,x7); // 3 2 1 0
x7 = int32x4_max(x3,x7); // 28 29 30 31
// stage (5,3) 0:8 1:9 2:10 3:11 4:12 5:13 6:14 7:15 16:24 17:25 18:26 19:27 20:28 21:29 22:30 23:31
x0 = int32x4_min(y0,y2); // 7 6 5 4
y2 = int32x4_max(y0,y2); // 15 14 13 12
y4 = int32x4_min(x4,x6); // 16 17 18 19
x6 = int32x4_max(x4,x6); // 24 25 26 27
x1 = int32x4_min(y1,y3); // 3 2 1 0
y3 = int32x4_max(y1,y3); // 11 10 9 8
y5 = int32x4_min(x5,x7); // 20 21 22 23
x7 = int32x4_max(x5,x7); // 28 29 30 31
// stage (5,2) 0:4 1:5 2:6 3:7 8:12 9:13 10:14 11:15 16:20 17:21 18:22 19:23 24:28 25:29 26:30 27:31
y0 = int32x4_min(x0,x1); // 3 2 1 0
x1 = int32x4_max(x0,x1); // 7 6 5 4
x4 = int32x4_min(y4,y5); // 16 17 18 19
y5 = int32x4_max(y4,y5); // 20 21 22 23
x2 = int32x4_min(y2,y3); // 11 10 9 8
y3 = int32x4_max(y2,y3); // 15 14 13 12
y6 = int32x4_min(x6,x7); // 24 25 26 27
x7 = int32x4_max(x6,x7); // 28 29 30 31
// stage (5,1) 0:2 1:3 4:6 5:7 8:10 9:11 12:14 13:15 16:18 17:19 20:22 21:23 24:26 25:27 28:30 29:31
x0 = int32x4_leftleft(y0,x1); // 3 2 7 6
x1 = int32x4_rightright(y0,x1); // 1 0 5 4
y4 = int32x4_leftleft(x4,y5); // 16 17 20 21
y5 = int32x4_rightright(x4,y5); // 18 19 22 23
y2 = int32x4_leftleft(x2,y3); // 11 10 15 14
y3 = int32x4_rightright(x2,y3); // 9 8 13 12
x6 = int32x4_leftleft(y6,x7); // 24 25 28 29
x7 = int32x4_rightright(y6,x7); // 26 27 30 31
y0 = int32x4_min(x0,x1); // 1 0 5 4
x1 = int32x4_max(x0,x1); // 3 2 7 6
x4 = int32x4_min(y4,y5); // 16 17 20 21
y5 = int32x4_max(y4,y5); // 18 19 22 23
x2 = int32x4_min(y2,y3); // 9 8 13 12
y3 = int32x4_max(y2,y3); // 11 10 15 14
y6 = int32x4_min(x6,x7); // 24 25 28 29
x7 = int32x4_max(x6,x7); // 26 27 30 31
// stage (5,0) 0:1 2:3 4:5 6:7 8:9 10:11 12:13 14:15 16:17 18:19 20:21 22:23 24:25 26:27 28:29 30:31
x0 = int32x4_a0b0a2b2(y0,x1); // 1 3 5 7
x1 = int32x4_a1b1a3b3(y0,x1); // 0 2 4 6
y4 = int32x4_a0b0a2b2(x4,y5); // 16 18 20 22
y5 = int32x4_a1b1a3b3(x4,y5); // 17 19 21 23
y2 = int32x4_a0b0a2b2(x2,y3); // 9 11 13 15
y3 = int32x4_a1b1a3b3(x2,y3); // 8 10 12 14
x6 = int32x4_a0b0a2b2(y6,x7); // 24 26 28 30
x7 = int32x4_a1b1a3b3(y6,x7); // 25 27 29 31
y0 = int32x4_min(x0,x1); // 0 2 4 6
x1 = int32x4_max(x0,x1); // 1 3 5 7
x4 = int32x4_min(y4,y5); // 16 18 20 22
y5 = int32x4_max(y4,y5); // 17 19 21 23
x2 = int32x4_min(y2,y3); // 8 10 12 14
y3 = int32x4_max(y2,y3); // 9 11 13 15
y6 = int32x4_min(x6,x7); // 24 26 28 30
x7 = int32x4_max(x6,x7); // 25 27 29 31
x0 = int32x4_a0b0a1b1(y0,x1); // 0 1 2 3
x1 = int32x4_a2b2a3b3(y0,x1); // 4 5 6 7
y2 = int32x4_a0b0a1b1(x2,y3); // 8 9 10 11
y3 = int32x4_a2b2a3b3(x2,y3); // 12 13 14 15
y4 = int32x4_a0b0a1b1(x4,y5); // 16 17 18 19
y5 = int32x4_a2b2a3b3(x4,y5); // 20 21 22 23
x6 = int32x4_a0b0a1b1(y6,x7); // 24 25 26 27
x7 = int32x4_a2b2a3b3(y6,x7); // 28 29 30 31
int32x4_store(x+0,vecxor^x0);
int32x4_store(x+4,vecxor^x1);
int32x4_store(x+8,vecxor^y2);
int32x4_store(x+12,vecxor^y3);
int32x4_store(x+16,vecxor^y4);
int32x4_store(x+20,vecxor^y5);
int32x4_store(x+24,vecxor^x6);
int32x4_store(x+28,vecxor^x7);
}
NOINLINE
static void int32_V_sort_64_xor(int32 *x,int32 xor)
{
int32x4 vecxor,x0,x1,x10,x11,x12,x13,x14,x15,x2,x3,x4,x5,x6,x7,x8,x9,y0,y1,y10,y11,y12,y13,y14,y2,y3,y4,y5,y6,y7,y8,y9;
vecxor = int32x4_broadcast(xor);
x0 = vecxor^int32x4_load(x+0); // 31 30 29 28
x1 = vecxor^int32x4_load(x+4); // 27 26 25 24
x2 = vecxor^int32x4_load(x+8); // 23 22 21 20
x3 = vecxor^int32x4_load(x+12); // 19 18 17 16
x4 = vecxor^int32x4_load(x+16); // 15 14 13 12
x5 = vecxor^int32x4_load(x+20); // 11 10 9 8
x6 = vecxor^int32x4_load(x+24); // 7 6 5 4
x7 = vecxor^int32x4_load(x+28); // 3 2 1 0
x8 = vecxor^int32x4_load(x+32); // 32 33 34 35
x9 = vecxor^int32x4_load(x+36); // 36 37 38 39
x10 = vecxor^int32x4_load(x+40); // 40 41 42 43
x11 = vecxor^int32x4_load(x+44); // 44 45 46 47
x12 = vecxor^int32x4_load(x+48); // 48 49 50 51
x13 = vecxor^int32x4_load(x+52); // 52 53 54 55
x14 = vecxor^int32x4_load(x+56); // 56 57 58 59
x15 = vecxor^int32x4_load(x+60); // 60 61 62 63
// stage (6,5) 0:63 1:62 2:61 3:60 4:59 5:58 6:57 7:56 8:55 9:54 10:53 11:52 12:51 13:50 14:49 15:48 16:47 17:46 18:45 19:44 20:43 21:42 22:41 23:40 24:39 25:38 26:37 27:36 28:35 29:34 30:33 31:32
y0 = int32x4_min(x0,x8); // 31 30 29 28
x8 = int32x4_max(x0,x8); // 32 33 34 35
y2 = int32x4_min(x2,x10); // 23 22 21 20
x10 = int32x4_max(x2,x10); // 40 41 42 43
y4 = int32x4_min(x4,x12); // 15 14 13 12
x12 = int32x4_max(x4,x12); // 48 49 50 51
y6 = int32x4_min(x6,x14); // 7 6 5 4
x14 = int32x4_max(x6,x14); // 56 57 58 59
y1 = int32x4_min(x1,x9); // 27 26 25 24
x9 = int32x4_max(x1,x9); // 36 37 38 39
y3 = int32x4_min(x3,x11); // 19 18 17 16
x11 = int32x4_max(x3,x11); // 44 45 46 47
y5 = int32x4_min(x5,x13); // 11 10 9 8
x13 = int32x4_max(x5,x13); // 52 53 54 55
y7 = int32x4_min(x7,x15); // 3 2 1 0
x15 = int32x4_max(x7,x15); // 60 61 62 63
// stage (6,4) 0:16 1:17 2:18 3:19 4:20 5:21 6:22 7:23 8:24 9:25 10:26 11:27 12:28 13:29 14:30 15:31 32:48 33:49 34:50 35:51 36:52 37:53 38:54 39:55 40:56 41:57 42:58 43:59 44:60 45:61 46:62 47:63
x0 = int32x4_min(y0,y4); // 15 14 13 12
y4 = int32x4_max(y0,y4); // 31 30 29 28
x2 = int32x4_min(y2,y6); // 7 6 5 4
y6 = int32x4_max(y2,y6); // 23 22 21 20
y8 = int32x4_min(x8,x12); // 32 33 34 35
x12 = int32x4_max(x8,x12); // 48 49 50 51
y10 = int32x4_min(x10,x14); // 40 41 42 43
x14 = int32x4_max(x10,x14); // 56 57 58 59
x1 = int32x4_min(y1,y5); // 11 10 9 8
y5 = int32x4_max(y1,y5); // 27 26 25 24
x3 = int32x4_min(y3,y7); // 3 2 1 0
y7 = int32x4_max(y3,y7); // 19 18 17 16
y9 = int32x4_min(x9,x13); // 36 37 38 39
x13 = int32x4_max(x9,x13); // 52 53 54 55
y11 = int32x4_min(x11,x15); // 44 45 46 47
x15 = int32x4_max(x11,x15); // 60 61 62 63
// stage (6,3) 0:8 1:9 2:10 3:11 4:12 5:13 6:14 7:15 16:24 17:25 18:26 19:27 20:28 21:29 22:30 23:31 32:40 33:41 34:42 35:43 36:44 37:45 38:46 39:47 48:56 49:57 50:58 51:59 52:60 53:61 54:62 55:63
y0 = int32x4_min(x0,x2); // 7 6 5 4
x2 = int32x4_max(x0,x2); // 15 14 13 12
x4 = int32x4_min(y4,y6); // 23 22 21 20
y6 = int32x4_max(y4,y6); // 31 30 29 28
x8 = int32x4_min(y8,y10); // 32 33 34 35
y10 = int32x4_max(y8,y10); // 40 41 42 43
y12 = int32x4_min(x12,x14); // 48 49 50 51
x14 = int32x4_max(x12,x14); // 56 57 58 59
y1 = int32x4_min(x1,x3); // 3 2 1 0
x3 = int32x4_max(x1,x3); // 11 10 9 8
x5 = int32x4_min(y5,y7); // 19 18 17 16
y7 = int32x4_max(y5,y7); // 27 26 25 24
x9 = int32x4_min(y9,y11); // 36 37 38 39
y11 = int32x4_max(y9,y11); // 44 45 46 47
y13 = int32x4_min(x13,x15); // 52 53 54 55
x15 = int32x4_max(x13,x15); // 60 61 62 63
// stage (6,2) 0:4 1:5 2:6 3:7 8:12 9:13 10:14 11:15 16:20 17:21 18:22 19:23 24:28 25:29 26:30 27:31 32:36 33:37 34:38 35:39 40:44 41:45 42:46 43:47 48:52 49:53 50:54 51:55 56:60 57:61 58:62 59:63
x0 = int32x4_min(y0,y1); // 3 2 1 0
y1 = int32x4_max(y0,y1); // 7 6 5 4
y4 = int32x4_min(x4,x5); // 19 18 17 16
x5 = int32x4_max(x4,x5); // 23 22 21 20
y8 = int32x4_min(x8,x9); // 32 33 34 35
x9 = int32x4_max(x8,x9); // 36 37 38 39
x12 = int32x4_min(y12,y13); // 48 49 50 51
y13 = int32x4_max(y12,y13); // 52 53 54 55
y2 = int32x4_min(x2,x3); // 11 10 9 8
x3 = int32x4_max(x2,x3); // 15 14 13 12
x6 = int32x4_min(y6,y7); // 27 26 25 24
y7 = int32x4_max(y6,y7); // 31 30 29 28
x10 = int32x4_min(y10,y11); // 40 41 42 43
y11 = int32x4_max(y10,y11); // 44 45 46 47
y14 = int32x4_min(x14,x15); // 56 57 58 59
x15 = int32x4_max(x14,x15); // 60 61 62 63
// stage (6,1) 0:2 1:3 4:6 5:7 8:10 9:11 12:14 13:15 16:18 17:19 20:22 21:23 24:26 25:27 28:30 29:31 32:34 33:35 36:38 37:39 40:42 41:43 44:46 45:47 48:50 49:51 52:54 53:55 56:58 57:59 60:62 61:63
y0 = int32x4_leftleft(x0,y1); // 3 2 7 6
y1 = int32x4_rightright(x0,y1); // 1 0 5 4
x4 = int32x4_leftleft(y4,x5); // 19 18 23 22
x5 = int32x4_rightright(y4,x5); // 17 16 21 20
x8 = int32x4_leftleft(y8,x9); // 32 33 36 37
x9 = int32x4_rightright(y8,x9); // 34 35 38 39
y12 = int32x4_leftleft(x12,y13); // 48 49 52 53
y13 = int32x4_rightright(x12,y13); // 50 51 54 55
x2 = int32x4_leftleft(y2,x3); // 11 10 15 14
x3 = int32x4_rightright(y2,x3); // 9 8 13 12
y6 = int32x4_leftleft(x6,y7); // 27 26 31 30
y7 = int32x4_rightright(x6,y7); // 25 24 29 28
y10 = int32x4_leftleft(x10,y11); // 40 41 44 45
y11 = int32x4_rightright(x10,y11); // 42 43 46 47
x14 = int32x4_leftleft(y14,x15); // 56 57 60 61
x15 = int32x4_rightright(y14,x15); // 58 59 62 63
x0 = int32x4_min(y0,y1); // 1 0 5 4
y1 = int32x4_max(y0,y1); // 3 2 7 6
y4 = int32x4_min(x4,x5); // 17 16 21 20
x5 = int32x4_max(x4,x5); // 19 18 23 22
y8 = int32x4_min(x8,x9); // 32 33 36 37
x9 = int32x4_max(x8,x9); // 34 35 38 39
x12 = int32x4_min(y12,y13); // 48 49 52 53
y13 = int32x4_max(y12,y13); // 50 51 54 55
y2 = int32x4_min(x2,x3); // 9 8 13 12
x3 = int32x4_max(x2,x3); // 11 10 15 14
x6 = int32x4_min(y6,y7); // 25 24 29 28
y7 = int32x4_max(y6,y7); // 27 26 31 30
x10 = int32x4_min(y10,y11); // 40 41 44 45
y11 = int32x4_max(y10,y11); // 42 43 46 47
y14 = int32x4_min(x14,x15); // 56 57 60 61
x15 = int32x4_max(x14,x15); // 58 59 62 63
// stage (6,0) 0:1 2:3 4:5 6:7 8:9 10:11 12:13 14:15 16:17 18:19 20:21 22:23 24:25 26:27 28:29 30:31 32:33 34:35 36:37 38:39 40:41 42:43 44:45 46:47 48:49 50:51 52:53 54:55 56:57 58:59 60:61 62:63
y0 = int32x4_a0b0a2b2(x0,y1); // 1 3 5 7
y1 = int32x4_a1b1a3b3(x0,y1); // 0 2 4 6
x4 = int32x4_a0b0a2b2(y4,x5); // 17 19 21 23
x5 = int32x4_a1b1a3b3(y4,x5); // 16 18 20 22
x8 = int32x4_a0b0a2b2(y8,x9); // 32 34 36 38
x9 = int32x4_a1b1a3b3(y8,x9); // 33 35 37 39
y12 = int32x4_a0b0a2b2(x12,y13); // 48 50 52 54
y13 = int32x4_a1b1a3b3(x12,y13); // 49 51 53 55
x2 = int32x4_a0b0a2b2(y2,x3); // 9 11 13 15
x3 = int32x4_a1b1a3b3(y2,x3); // 8 10 12 14
y6 = int32x4_a0b0a2b2(x6,y7); // 25 27 29 31
y7 = int32x4_a1b1a3b3(x6,y7); // 24 26 28 30
y10 = int32x4_a0b0a2b2(x10,y11); // 40 42 44 46
y11 = int32x4_a1b1a3b3(x10,y11); // 41 43 45 47
x14 = int32x4_a0b0a2b2(y14,x15); // 56 58 60 62
x15 = int32x4_a1b1a3b3(y14,x15); // 57 59 61 63
x0 = int32x4_min(y0,y1); // 0 2 4 6
y1 = int32x4_max(y0,y1); // 1 3 5 7
y4 = int32x4_min(x4,x5); // 16 18 20 22
x5 = int32x4_max(x4,x5); // 17 19 21 23
y8 = int32x4_min(x8,x9); // 32 34 36 38
x9 = int32x4_max(x8,x9); // 33 35 37 39
x12 = int32x4_min(y12,y13); // 48 50 52 54
y13 = int32x4_max(y12,y13); // 49 51 53 55
y2 = int32x4_min(x2,x3); // 8 10 12 14
x3 = int32x4_max(x2,x3); // 9 11 13 15
x6 = int32x4_min(y6,y7); // 24 26 28 30
y7 = int32x4_max(y6,y7); // 25 27 29 31
x10 = int32x4_min(y10,y11); // 40 42 44 46
y11 = int32x4_max(y10,y11); // 41 43 45 47
y14 = int32x4_min(x14,x15); // 56 58 60 62
x15 = int32x4_max(x14,x15); // 57 59 61 63
y0 = int32x4_a0b0a1b1(x0,y1); // 0 1 2 3
y1 = int32x4_a2b2a3b3(x0,y1); // 4 5 6 7
x2 = int32x4_a0b0a1b1(y2,x3); // 8 9 10 11
x3 = int32x4_a2b2a3b3(y2,x3); // 12 13 14 15
x4 = int32x4_a0b0a1b1(y4,x5); // 16 17 18 19
x5 = int32x4_a2b2a3b3(y4,x5); // 20 21 22 23
y6 = int32x4_a0b0a1b1(x6,y7); // 24 25 26 27
y7 = int32x4_a2b2a3b3(x6,y7); // 28 29 30 31
x8 = int32x4_a0b0a1b1(y8,x9); // 32 33 34 35
x9 = int32x4_a2b2a3b3(y8,x9); // 36 37 38 39
y10 = int32x4_a0b0a1b1(x10,y11); // 40 41 42 43
y11 = int32x4_a2b2a3b3(x10,y11); // 44 45 46 47
y12 = int32x4_a0b0a1b1(x12,y13); // 48 49 50 51
y13 = int32x4_a2b2a3b3(x12,y13); // 52 53 54 55
x14 = int32x4_a0b0a1b1(y14,x15); // 56 57 58 59
x15 = int32x4_a2b2a3b3(y14,x15); // 60 61 62 63
int32x4_store(x+0,vecxor^y0);
int32x4_store(x+4,vecxor^y1);
int32x4_store(x+8,vecxor^x2);
int32x4_store(x+12,vecxor^x3);
int32x4_store(x+16,vecxor^x4);
int32x4_store(x+20,vecxor^x5);
int32x4_store(x+24,vecxor^y6);
int32x4_store(x+28,vecxor^y7);
int32x4_store(x+32,vecxor^x8);
int32x4_store(x+36,vecxor^x9);
int32x4_store(x+40,vecxor^y10);
int32x4_store(x+44,vecxor^y11);
int32x4_store(x+48,vecxor^y12);
int32x4_store(x+52,vecxor^y13);
int32x4_store(x+56,vecxor^x14);
int32x4_store(x+60,vecxor^x15);
}
NOINLINE
static void int32_threestages_8_up(int32 *x,long long p,long long n)
{
long long i;
for (i = 0;i+4 <= n;i += 4) {
int32x4 x0 = int32x4_load(&x[i]);
int32x4 x1 = int32x4_load(&x[p+i]);
int32x4 x2 = int32x4_load(&x[2*p+i]);
int32x4 x3 = int32x4_load(&x[3*p+i]);
int32x4 x4 = int32x4_load(&x[4*p+i]);
int32x4 x5 = int32x4_load(&x[5*p+i]);
int32x4 x6 = int32x4_load(&x[6*p+i]);
int32x4 x7 = int32x4_load(&x[7*p+i]);
int32x4_MINMAX(x0,x4);
int32x4_MINMAX(x1,x5);
int32x4_MINMAX(x2,x6);
int32x4_MINMAX(x3,x7);
int32x4_MINMAX(x0,x2);
int32x4_MINMAX(x1,x3);
int32x4_MINMAX(x4,x6);
int32x4_MINMAX(x5,x7);
int32x4_MINMAX(x0,x1);
int32x4_MINMAX(x2,x3);
int32x4_MINMAX(x4,x5);
int32x4_MINMAX(x6,x7);
int32x4_store(&x[i],x0);
int32x4_store(&x[p+i],x1);
int32x4_store(&x[2*p+i],x2);
int32x4_store(&x[3*p+i],x3);
int32x4_store(&x[4*p+i],x4);
int32x4_store(&x[5*p+i],x5);
int32x4_store(&x[6*p+i],x6);
int32x4_store(&x[7*p+i],x7);
}
for (;i < n;++i) {
int32 x0 = x[i];
int32 x1 = x[p+i];
int32 x2 = x[2*p+i];
int32 x3 = x[3*p+i];
int32 x4 = x[4*p+i];
int32 x5 = x[5*p+i];
int32 x6 = x[6*p+i];
int32 x7 = x[7*p+i];
int32_MINMAX(x0,x4);
int32_MINMAX(x1,x5);
int32_MINMAX(x2,x6);
int32_MINMAX(x3,x7);
int32_MINMAX(x0,x2);
int32_MINMAX(x1,x3);
int32_MINMAX(x4,x6);
int32_MINMAX(x5,x7);
int32_MINMAX(x0,x1);
int32_MINMAX(x2,x3);
int32_MINMAX(x4,x5);
int32_MINMAX(x6,x7);
x[i] = x0;
x[p+i] = x1;
x[2*p+i] = x2;
x[3*p+i] = x3;
x[4*p+i] = x4;
x[5*p+i] = x5;
x[6*p+i] = x6;
x[7*p+i] = x7;
}
}
NOINLINE
static void int32_threestages_7_up(int32 *x,long long p,long long n)
{
long long i;
for (i = 0;i+4 <= n;i += 4) {
int32x4 x0 = int32x4_load(&x[i]);
int32x4 x1 = int32x4_load(&x[p+i]);
int32x4 x2 = int32x4_load(&x[2*p+i]);
int32x4 x3 = int32x4_load(&x[3*p+i]);
int32x4 x4 = int32x4_load(&x[4*p+i]);
int32x4 x5 = int32x4_load(&x[5*p+i]);
int32x4 x6 = int32x4_load(&x[6*p+i]);
int32x4_MINMAX(x0,x4);
int32x4_MINMAX(x1,x5);
int32x4_MINMAX(x2,x6);
int32x4_MINMAX(x0,x2);
int32x4_MINMAX(x1,x3);
int32x4_MINMAX(x4,x6);
int32x4_MINMAX(x0,x1);
int32x4_MINMAX(x2,x3);
int32x4_MINMAX(x4,x5);
int32x4_store(&x[i],x0);
int32x4_store(&x[p+i],x1);
int32x4_store(&x[2*p+i],x2);
int32x4_store(&x[3*p+i],x3);
int32x4_store(&x[4*p+i],x4);
int32x4_store(&x[5*p+i],x5);
int32x4_store(&x[6*p+i],x6);
}
for (;i < n;++i) {
int32 x0 = x[i];
int32 x1 = x[p+i];
int32 x2 = x[2*p+i];
int32 x3 = x[3*p+i];
int32 x4 = x[4*p+i];
int32 x5 = x[5*p+i];
int32 x6 = x[6*p+i];
int32_MINMAX(x0,x4);
int32_MINMAX(x1,x5);
int32_MINMAX(x2,x6);
int32_MINMAX(x0,x2);
int32_MINMAX(x1,x3);
int32_MINMAX(x4,x6);
int32_MINMAX(x0,x1);
int32_MINMAX(x2,x3);
int32_MINMAX(x4,x5);
x[i] = x0;
x[p+i] = x1;
x[2*p+i] = x2;
x[3*p+i] = x3;
x[4*p+i] = x4;
x[5*p+i] = x5;
x[6*p+i] = x6;
}
}
NOINLINE
static void int32_threestages_6_up(int32 *x,long long p,long long n)
{
long long i;
for (i = 0;i+4 <= n;i += 4) {
int32x4 x0 = int32x4_load(&x[i]);
int32x4 x1 = int32x4_load(&x[p+i]);
int32x4 x2 = int32x4_load(&x[2*p+i]);
int32x4 x3 = int32x4_load(&x[3*p+i]);
int32x4 x4 = int32x4_load(&x[4*p+i]);
int32x4 x5 = int32x4_load(&x[5*p+i]);
int32x4_MINMAX(x0,x4);
int32x4_MINMAX(x1,x5);
int32x4_MINMAX(x0,x2);
int32x4_MINMAX(x1,x3);
int32x4_MINMAX(x0,x1);
int32x4_MINMAX(x2,x3);
int32x4_MINMAX(x4,x5);
int32x4_store(&x[i],x0);
int32x4_store(&x[p+i],x1);
int32x4_store(&x[2*p+i],x2);
int32x4_store(&x[3*p+i],x3);
int32x4_store(&x[4*p+i],x4);
int32x4_store(&x[5*p+i],x5);
}
for (;i < n;++i) {
int32 x0 = x[i];
int32 x1 = x[p+i];
int32 x2 = x[2*p+i];
int32 x3 = x[3*p+i];
int32 x4 = x[4*p+i];
int32 x5 = x[5*p+i];
int32_MINMAX(x0,x4);
int32_MINMAX(x1,x5);
int32_MINMAX(x0,x2);
int32_MINMAX(x1,x3);
int32_MINMAX(x0,x1);
int32_MINMAX(x2,x3);
int32_MINMAX(x4,x5);
x[i] = x0;
x[p+i] = x1;
x[2*p+i] = x2;
x[3*p+i] = x3;
x[4*p+i] = x4;
x[5*p+i] = x5;
}
}
NOINLINE
static void int32_threestages_5_up(int32 *x,long long p,long long n)
{
long long i;
for (i = 0;i+4 <= n;i += 4) {
int32x4 x0 = int32x4_load(&x[i]);
int32x4 x1 = int32x4_load(&x[p+i]);
int32x4 x2 = int32x4_load(&x[2*p+i]);
int32x4 x3 = int32x4_load(&x[3*p+i]);
int32x4 x4 = int32x4_load(&x[4*p+i]);
int32x4_MINMAX(x0,x4);
int32x4_MINMAX(x0,x2);
int32x4_MINMAX(x1,x3);
int32x4_MINMAX(x0,x1);
int32x4_MINMAX(x2,x3);
int32x4_store(&x[i],x0);
int32x4_store(&x[p+i],x1);
int32x4_store(&x[2*p+i],x2);
int32x4_store(&x[3*p+i],x3);
int32x4_store(&x[4*p+i],x4);
}
for (;i < n;++i) {
int32 x0 = x[i];
int32 x1 = x[p+i];
int32 x2 = x[2*p+i];
int32 x3 = x[3*p+i];
int32 x4 = x[4*p+i];
int32_MINMAX(x0,x4);
int32_MINMAX(x0,x2);
int32_MINMAX(x1,x3);
int32_MINMAX(x0,x1);
int32_MINMAX(x2,x3);
x[i] = x0;
x[p+i] = x1;
x[2*p+i] = x2;
x[3*p+i] = x3;
x[4*p+i] = x4;
}
}
NOINLINE
static void int32_threestages_4_up(int32 *x,long long p,long long n)
{
long long i;
for (i = 0;i+4 <= n;i += 4) {
int32x4 x0 = int32x4_load(&x[i]);
int32x4 x1 = int32x4_load(&x[p+i]);
int32x4 x2 = int32x4_load(&x[2*p+i]);
int32x4 x3 = int32x4_load(&x[3*p+i]);
int32x4_MINMAX(x0,x2);
int32x4_MINMAX(x1,x3);
int32x4_MINMAX(x0,x1);
int32x4_MINMAX(x2,x3);
int32x4_store(&x[i],x0);
int32x4_store(&x[p+i],x1);
int32x4_store(&x[2*p+i],x2);
int32x4_store(&x[3*p+i],x3);
}
for (;i < n;++i) {
int32 x0 = x[i];
int32 x1 = x[p+i];
int32 x2 = x[2*p+i];
int32 x3 = x[3*p+i];
int32_MINMAX(x0,x2);
int32_MINMAX(x1,x3);
int32_MINMAX(x0,x1);
int32_MINMAX(x2,x3);
x[i] = x0;
x[p+i] = x1;
x[2*p+i] = x2;
x[3*p+i] = x3;
}
}
NOINLINE
static void int32_threestages_8_up_16(int32 *x)
{
long long i;
long long p = 16;
long long n = p;
for (i = 0;i+4 <= n;i += 4) {
int32x4 x0 = int32x4_load(&x[i]);
int32x4 x1 = int32x4_load(&x[p+i]);
int32x4 x2 = int32x4_load(&x[2*p+i]);
int32x4 x3 = int32x4_load(&x[3*p+i]);
int32x4 x4 = int32x4_load(&x[4*p+i]);
int32x4 x5 = int32x4_load(&x[5*p+i]);
int32x4 x6 = int32x4_load(&x[6*p+i]);
int32x4 x7 = int32x4_load(&x[7*p+i]);
int32x4_MINMAX(x0,x4);
int32x4_MINMAX(x1,x5);
int32x4_MINMAX(x2,x6);
int32x4_MINMAX(x3,x7);
int32x4_MINMAX(x0,x2);
int32x4_MINMAX(x1,x3);
int32x4_MINMAX(x4,x6);
int32x4_MINMAX(x5,x7);
int32x4_MINMAX(x0,x1);
int32x4_MINMAX(x2,x3);
int32x4_MINMAX(x4,x5);
int32x4_MINMAX(x6,x7);
int32x4_store(&x[i],x0);
int32x4_store(&x[p+i],x1);
int32x4_store(&x[2*p+i],x2);
int32x4_store(&x[3*p+i],x3);
int32x4_store(&x[4*p+i],x4);
int32x4_store(&x[5*p+i],x5);
int32x4_store(&x[6*p+i],x6);
int32x4_store(&x[7*p+i],x7);
}
}
NOINLINE
static void int32_threestages_8_down_16(int32 *x)
{
long long i;
long long p = 16;
long long n = p;
for (i = 0;i+4 <= n;i += 4) {
int32x4 x0 = int32x4_load(&x[i]);
int32x4 x1 = int32x4_load(&x[p+i]);
int32x4 x2 = int32x4_load(&x[2*p+i]);
int32x4 x3 = int32x4_load(&x[3*p+i]);
int32x4 x4 = int32x4_load(&x[4*p+i]);
int32x4 x5 = int32x4_load(&x[5*p+i]);
int32x4 x6 = int32x4_load(&x[6*p+i]);
int32x4 x7 = int32x4_load(&x[7*p+i]);
int32x4_MINMAX(x4,x0);
int32x4_MINMAX(x5,x1);
int32x4_MINMAX(x6,x2);
int32x4_MINMAX(x7,x3);
int32x4_MINMAX(x2,x0);
int32x4_MINMAX(x3,x1);
int32x4_MINMAX(x6,x4);
int32x4_MINMAX(x7,x5);
int32x4_MINMAX(x1,x0);
int32x4_MINMAX(x3,x2);
int32x4_MINMAX(x5,x4);
int32x4_MINMAX(x7,x6);
int32x4_store(&x[i],x0);
int32x4_store(&x[p+i],x1);
int32x4_store(&x[2*p+i],x2);
int32x4_store(&x[3*p+i],x3);
int32x4_store(&x[4*p+i],x4);
int32x4_store(&x[5*p+i],x5);
int32x4_store(&x[6*p+i],x6);
int32x4_store(&x[7*p+i],x7);
}
}
NOINLINE
static void int32_threestages_8_down_atleast32(int32 *x,long long p)
{
long long i;
long long n = p;
for (i = 0;i+4 <= n;i += 4) {
int32x4 x0 = int32x4_load(&x[i]);
int32x4 x1 = int32x4_load(&x[p+i]);
int32x4 x2 = int32x4_load(&x[2*p+i]);
int32x4 x3 = int32x4_load(&x[3*p+i]);
int32x4 x4 = int32x4_load(&x[4*p+i]);
int32x4 x5 = int32x4_load(&x[5*p+i]);
int32x4 x6 = int32x4_load(&x[6*p+i]);
int32x4 x7 = int32x4_load(&x[7*p+i]);
int32x4_MINMAX(x4,x0);
int32x4_MINMAX(x5,x1);
int32x4_MINMAX(x6,x2);
int32x4_MINMAX(x7,x3);
int32x4_MINMAX(x2,x0);
int32x4_MINMAX(x3,x1);
int32x4_MINMAX(x6,x4);
int32x4_MINMAX(x7,x5);
int32x4_MINMAX(x1,x0);
int32x4_MINMAX(x3,x2);
int32x4_MINMAX(x5,x4);
int32x4_MINMAX(x7,x6);
int32x4_store(&x[i],x0);
int32x4_store(&x[p+i],x1);
int32x4_store(&x[2*p+i],x2);
int32x4_store(&x[3*p+i],x3);
int32x4_store(&x[4*p+i],x4);
int32x4_store(&x[5*p+i],x5);
int32x4_store(&x[6*p+i],x6);
int32x4_store(&x[7*p+i],x7);
}
}
NOINLINE
static void int32_threestages_6_down(int32 *x,long long p,long long n)
{
long long i;
for (i = 0;i+4 <= n;i += 4) {
int32x4 x0 = int32x4_load(&x[i]);
int32x4 x1 = int32x4_load(&x[p+i]);
int32x4 x2 = int32x4_load(&x[2*p+i]);
int32x4 x3 = int32x4_load(&x[3*p+i]);
int32x4 x4 = int32x4_load(&x[4*p+i]);
int32x4 x5 = int32x4_load(&x[5*p+i]);
int32x4_MINMAX(x4,x0);
int32x4_MINMAX(x5,x1);
int32x4_MINMAX(x2,x0);
int32x4_MINMAX(x3,x1);
int32x4_MINMAX(x1,x0);
int32x4_MINMAX(x3,x2);
int32x4_MINMAX(x5,x4);
int32x4_store(&x[i],x0);
int32x4_store(&x[p+i],x1);
int32x4_store(&x[2*p+i],x2);
int32x4_store(&x[3*p+i],x3);
int32x4_store(&x[4*p+i],x4);
int32x4_store(&x[5*p+i],x5);
}
for (;i < n;++i) {
int32 x0 = x[i];
int32 x1 = x[p+i];
int32 x2 = x[2*p+i];
int32 x3 = x[3*p+i];
int32 x4 = x[4*p+i];
int32 x5 = x[5*p+i];
int32_MINMAX(x4,x0);
int32_MINMAX(x5,x1);
int32_MINMAX(x2,x0);
int32_MINMAX(x3,x1);
int32_MINMAX(x1,x0);
int32_MINMAX(x3,x2);
int32_MINMAX(x5,x4);
x[i] = x0;
x[p+i] = x1;
x[2*p+i] = x2;
x[3*p+i] = x3;
x[4*p+i] = x4;
x[5*p+i] = x5;
}
}
// XXX: currently xor must be 0 or -1
NOINLINE
static void int32_V_sort_2poweratleast16_xor(int32 *x,long long n,int32 xor)
{
if (n == 16) { int32_V_sort_16_xor(x,xor); return; }
if (n == 32) { int32_V_sort_32_xor(x,xor); return; }
if (n == 64) { int32_V_sort_64_xor(x,xor); return; }
if (n == 128) {
if (xor)
int32_threestages_8_down_16(x);
else
int32_threestages_8_up_16(x);
for (long long i = 0;i < 8;++i)
int32_V_sort_16_xor(x+16*i,xor);
return;
}
if (xor)
int32_threestages_8_down_atleast32(x,n>>3);
else
int32_threestages_8_up(x,n>>3,n>>3);
for (long long i = 0;i < 8;++i)
int32_V_sort_2poweratleast16_xor(x+(n>>3)*i,n>>3,xor);
}
/* q is power of 2; want only merge stages q,q/2,q/4,...,1 */
// XXX: assuming 8 <= q < n <= 2q; q is a power of 2
NOINLINE
static void int32_V_sort(int32 *x,long long q,long long n)
{
if (!(n & (n - 1))) { int32_V_sort_2poweratleast16_xor(x,n,0); return; }
if (n <= 16) { int32_V_sort_8through16(x,n); return; }
if (n <= 32) { int32_V_sort_16through32(x,n); return; }
if (n <= 64) { int32_V_sort_32through64(x,n); return; }
// 64 <= q < n < 2q
q >>= 2;
// 64 <= 4q < n < 8q
if (7*q < n) {
int32_threestages_8_up(x,q,n-7*q);
int32_threestages_7_up(x+n-7*q,q,8*q-n);
} else if (6*q < n) {
int32_threestages_7_up(x,q,n-6*q);
int32_threestages_6_up(x+n-6*q,q,7*q-n);
} else if (5*q < n) {
int32_threestages_6_up(x,q,n-5*q);
int32_threestages_5_up(x+n-5*q,q,6*q-n);
} else {
int32_threestages_5_up(x,q,n-4*q);
int32_threestages_4_up(x+n-4*q,q,5*q-n);
}
// now want to handle each batch of q entries separately
int32_V_sort(x,q>>1,q);
int32_V_sort(x+q,q>>1,q);
int32_V_sort(x+2*q,q>>1,q);
int32_V_sort(x+3*q,q>>1,q);
x += 4*q;
n -= 4*q;
while (n >= q) {
int32_V_sort(x,q>>1,q);
x += q;
n -= q;
}
// have n entries left in last batch, with 0 <= n < q
if (n <= 1) return;
while (q >= n) q >>= 1; // empty merge stage
// now 1 <= q < n <= 2q
if (q >= 8) { int32_V_sort(x,q,n); return; }
if (n == 8) {
int32_MINMAX(x[0],x[4]);
int32_MINMAX(x[1],x[5]);
int32_MINMAX(x[2],x[6]);
int32_MINMAX(x[3],x[7]);
int32_MINMAX(x[0],x[2]);
int32_MINMAX(x[1],x[3]);
int32_MINMAX(x[0],x[1]);
int32_MINMAX(x[2],x[3]);
int32_MINMAX(x[4],x[6]);
int32_MINMAX(x[5],x[7]);
int32_MINMAX(x[4],x[5]);
int32_MINMAX(x[6],x[7]);
return;
}
if (4 <= n) {
for (long long i = 0;i < n-4;++i)
int32_MINMAX(x[i],x[4+i]);
int32_MINMAX(x[0],x[2]);
int32_MINMAX(x[1],x[3]);
int32_MINMAX(x[0],x[1]);
int32_MINMAX(x[2],x[3]);
n -= 4;
x += 4;
}
if (3 <= n)
int32_MINMAX(x[0],x[2]);
if (2 <= n)
int32_MINMAX(x[0],x[1]);
}
// XXX: currently xor must be 0 or -1
NOINLINE
static void int32_sort_2poweratleast32_xor(int32 *x,long long n,int32 xor)
{
if (n == 32) { int32_sort_32_xor(x,xor); return; }
if (n == 64) { int32_sort_64_xor(x,xor); return; }
int32_sort_2poweratleast32_xor(x,n>>1,~xor);
int32_sort_2poweratleast32_xor(x+(n>>1),n>>1,xor);
int32_V_sort_2poweratleast16_xor(x,n,xor);
}
void int32_sort(int32 *x,long long n)
{ long long q;
if (n <= 1) return;
if (n == 2) { int32_MINMAX(x[0],x[1]); return; }
if (n <= 7) { int32_sort_3through7(x,n); return; }
if (n <= 16) { int32_sort_8through16(x,n); return; }
if (!(n & (n - 1))) { int32_sort_2poweratleast32_xor(x,n,0); return; }
if (n <= 32) { int32_sort_16through32(x,n); return; }
q = 32;
while (q < n - q) q += q;
// 32 <= q < n < 2q
if (512 <= n && n <= (7*q)>>2) {
long long m = (3*q)>>2; // strategy: sort m, sort n-m, merge
long long r = q>>3; // at least 32 since q is at least 256
int32_sort_2poweratleast32_xor(x,4*r,0);
int32_sort_2poweratleast32_xor(x+4*r,r,0);
int32_sort_2poweratleast32_xor(x+5*r,r,-1);
int32_V_sort_2poweratleast16_xor(x+4*r,2*r,-1);
int32_threestages_6_down(x,r,r);
for (long long i = 0;i < 6;++i)
int32_V_sort_2poweratleast16_xor(x+i*r,r,-1);
int32_sort(x+m,n-m);
} else if (64 <= q && n == (3*q)>>1) {
// strategy: sort q, sort q/2, merge
long long r = q>>2; // at least 16 since q is at least 64
int32_sort_2poweratleast32_xor(x,4*r,-1);
int32_sort_2poweratleast32_xor(x+4*r,2*r,0);
int32_threestages_6_up(x,r,r);
for (long long i = 0;i < 6;++i)
int32_V_sort_2poweratleast16_xor(x+i*r,r,0);
return;
} else {
int32_sort_2poweratleast32_xor(x,q,-1);
int32_sort(x+q,n-q);
}
int32_V_sort(x,q,n);
}