v0 = v0 + nonce + vec2(0u, u32(v0.x + nonce.x < v0.x));
// d = rotr64(d ^ a, 32)
- xor = v12 ^ v0;
- v12 = xor.yx;
+ v12 = v12.yx ^ v0.yx;
// c = c + d
v8 = v8 + v12 + vec2(0u, u32(v8.x + v12.x < v8.x));
v1 = v1 + m2 + vec2(0u, u32(v1.x + m2.x < v1.x));
// d = rotr64(d ^ a, 32)
- xor = v13 ^ v1;
- v13 = xor.yx;
+ v13 = v13.yx ^ v1.yx;
// c = c + d
v9 = v9 + v13 + vec2(0u, u32(v9.x + v13.x < v9.x));
v2 = v2 + m4 + vec2(0u, u32(v2.x + m4.x < v2.x));
// d = rotr64(d ^ a, 32)
- xor = v14 ^ v2;
- v14 = xor.yx;
+ v14 = v14.yx ^ v2.yx;
// c = c + d
v10 = v10 + v14 + vec2(0u, u32(v10.x + v14.x < v10.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v15 ^ v3;
- v15 = xor.yx;
+ v15 = v15.yx ^ v3.yx;
// c = c + d
v11 = v11 + v15 + vec2(0u, u32(v11.x + v15.x < v11.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v15 ^ v0;
- v15 = xor.yx;
+ v15 = v15.yx ^ v0.yx;
// c = c + d
v10 = v10 + v15 + vec2(0u, u32(v10.x + v15.x < v10.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v12 ^ v1;
- v12 = xor.yx;
+ v12 = v12.yx ^ v1.yx;
// c = c + d
v11 = v11 + v12 + vec2(0u, u32(v11.x + v12.x < v11.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v13 ^ v2;
- v13 = xor.yx;
+ v13 = v13.yx ^ v2.yx;
// c = c + d
v8 = v8 + v13 + vec2(0u, u32(v8.x + v13.x < v8.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v14 ^ v3;
- v14 = xor.yx;
+ v14 = v14.yx ^ v3.yx;
// c = c + d
v9 = v9 + v14 + vec2(0u, u32(v9.x + v14.x < v9.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v12 ^ v0;
- v12 = xor.yx;
+ v12 = v12.yx ^ v0.yx;
// c = c + d
v8 = v8 + v12 + vec2(0u, u32(v8.x + v12.x < v8.x));
v1 = v1 + m4 + vec2(0u, u32(v1.x + m4.x < v1.x));
// d = rotr64(d ^ a, 32)
- xor = v13 ^ v1;
- v13 = xor.yx;
+ v13 = v13.yx ^ v1.yx;
// c = c + d
v9 = v9 + v13 + vec2(0u, u32(v9.x + v13.x < v9.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v14 ^ v2;
- v14 = xor.yx;
+ v14 = v14.yx ^ v2.yx;
// c = c + d
v10 = v10 + v14 + vec2(0u, u32(v10.x + v14.x < v10.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v15 ^ v3;
- v15 = xor.yx;
+ v15 = v15.yx ^ v3.yx;
// c = c + d
v11 = v11 + v15 + vec2(0u, u32(v11.x + v15.x < v11.x));
v0 = v0 + m1 + vec2(0u, u32(v0.x + m1.x < v0.x));
// d = rotr64(d ^ a, 32)
- xor = v15 ^ v0;
- v15 = xor.yx;
+ v15 = v15.yx ^ v0.yx;
// c = c + d
v10 = v10 + v15 + vec2(0u, u32(v10.x + v15.x < v10.x));
v1 = v1 + nonce + vec2(0u, u32(v1.x + nonce.x < v1.x));
// d = rotr64(d ^ a, 32)
- xor = v12 ^ v1;
- v12 = xor.yx;
+ v12 = v12.yx ^ v1.yx;
// c = c + d
v11 = v11 + v12 + vec2(0u, u32(v11.x + v12.x < v11.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v13 ^ v2;
- v13 = xor.yx;
+ v13 = v13.yx ^ v2.yx;
// c = c + d
v8 = v8 + v13 + vec2(0u, u32(v8.x + v13.x < v8.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v14 ^ v3;
- v14 = xor.yx;
+ v14 = v14.yx ^ v3.yx;
// c = c + d
v9 = v9 + v14 + vec2(0u, u32(v9.x + v14.x < v9.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v12 ^ v0;
- v12 = xor.yx;
+ v12 = v12.yx ^ v0.yx;
// c = c + d
v8 = v8 + v12 + vec2(0u, u32(v8.x + v12.x < v8.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v13 ^ v1;
- v13 = xor.yx;
+ v13 = v13.yx ^ v1.yx;
// c = c + d
v9 = v9 + v13 + vec2(0u, u32(v9.x + v13.x < v9.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v14 ^ v2;
- v14 = xor.yx;
+ v14 = v14.yx ^ v2.yx;
// c = c + d
v10 = v10 + v14 + vec2(0u, u32(v10.x + v14.x < v10.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v15 ^ v3;
- v15 = xor.yx;
+ v15 = v15.yx ^ v3.yx;
// c = c + d
v11 = v11 + v15 + vec2(0u, u32(v11.x + v15.x < v11.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v15 ^ v0;
- v15 = xor.yx;
+ v15 = v15.yx ^ v0.yx;
// c = c + d
v10 = v10 + v15 + vec2(0u, u32(v10.x + v15.x < v10.x));
v1 = v1 + m3 + vec2(0u, u32(v1.x + m3.x < v1.x));
// d = rotr64(d ^ a, 32)
- xor = v12 ^ v1;
- v12 = xor.yx;
+ v12 = v12.yx ^ v1.yx;
// c = c + d
v11 = v11 + v12 + vec2(0u, u32(v11.x + v12.x < v11.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v13 ^ v2;
- v13 = xor.yx;
+ v13 = v13.yx ^ v2.yx;
// c = c + d
v8 = v8 + v13 + vec2(0u, u32(v8.x + v13.x < v8.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v14 ^ v3;
- v14 = xor.yx;
+ v14 = v14.yx ^ v3.yx;
// c = c + d
v9 = v9 + v14 + vec2(0u, u32(v9.x + v14.x < v9.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v12 ^ v0;
- v12 = xor.yx;
+ v12 = v12.yx ^ v0.yx;
// c = c + d
v8 = v8 + v12 + vec2(0u, u32(v8.x + v12.x < v8.x));
v1 = v1 + m3 + vec2(0u, u32(v1.x + m3.x < v1.x));
// d = rotr64(d ^ a, 32)
- xor = v13 ^ v1;
- v13 = xor.yx;
+ v13 = v13.yx ^ v1.yx;
// c = c + d
v9 = v9 + v13 + vec2(0u, u32(v9.x + v13.x < v9.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v14 ^ v2;
- v14 = xor.yx;
+ v14 = v14.yx ^ v2.yx;
// c = c + d
v10 = v10 + v14 + vec2(0u, u32(v10.x + v14.x < v10.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v15 ^ v3;
- v15 = xor.yx;
+ v15 = v15.yx ^ v3.yx;
// c = c + d
v11 = v11 + v15 + vec2(0u, u32(v11.x + v15.x < v11.x));
v0 = v0 + m2 + vec2(0u, u32(v0.x + m2.x < v0.x));
// d = rotr64(d ^ a, 32)
- xor = v15 ^ v0;
- v15 = xor.yx;
+ v15 = v15.yx ^ v0.yx;
// c = c + d
v10 = v10 + v15 + vec2(0u, u32(v10.x + v15.x < v10.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v12 ^ v1;
- v12 = xor.yx;
+ v12 = v12.yx ^ v1.yx;
// c = c + d
v11 = v11 + v12 + vec2(0u, u32(v11.x + v12.x < v11.x));
v2 = v2 + m4 + vec2(0u, u32(v2.x + m4.x < v2.x));
// d = rotr64(d ^ a, 32)
- xor = v13 ^ v2;
- v13 = xor.yx;
+ v13 = v13.yx ^ v2.yx;
// c = c + d
v8 = v8 + v13 + vec2(0u, u32(v8.x + v13.x < v8.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v14 ^ v3;
- v14 = xor.yx;
+ v14 = v14.yx ^ v3.yx;
// c = c + d
v9 = v9 + v14 + vec2(0u, u32(v9.x + v14.x < v9.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v12 ^ v0;
- v12 = xor.yx;
+ v12 = v12.yx ^ v0.yx;
// c = c + d
v8 = v8 + v12 + vec2(0u, u32(v8.x + v12.x < v8.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v13 ^ v1;
- v13 = xor.yx;
+ v13 = v13.yx ^ v1.yx;
// c = c + d
v9 = v9 + v13 + vec2(0u, u32(v9.x + v13.x < v9.x));
v2 = v2 + m2 + vec2(0u, u32(v2.x + m2.x < v2.x));
// d = rotr64(d ^ a, 32)
- xor = v14 ^ v2;
- v14 = xor.yx;
+ v14 = v14.yx ^ v2.yx;
// c = c + d
v10 = v10 + v14 + vec2(0u, u32(v10.x + v14.x < v10.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v15 ^ v3;
- v15 = xor.yx;
+ v15 = v15.yx ^ v3.yx;
// c = c + d
v11 = v11 + v15 + vec2(0u, u32(v11.x + v15.x < v11.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v15 ^ v0;
- v15 = xor.yx;
+ v15 = v15.yx ^ v0.yx;
// c = c + d
v10 = v10 + v15 + vec2(0u, u32(v10.x + v15.x < v10.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v12 ^ v1;
- v12 = xor.yx;
+ v12 = v12.yx ^ v1.yx;
// c = c + d
v11 = v11 + v12 + vec2(0u, u32(v11.x + v12.x < v11.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v13 ^ v2;
- v13 = xor.yx;
+ v13 = v13.yx ^ v2.yx;
// c = c + d
v8 = v8 + v13 + vec2(0u, u32(v8.x + v13.x < v8.x));
v3 = v3 + m3 + vec2(0u, u32(v3.x + m3.x < v3.x));
// d = rotr64(d ^ a, 32)
- xor = v14 ^ v3;
- v14 = xor.yx;
+ v14 = v14.yx ^ v3.yx;
// c = c + d
v9 = v9 + v14 + vec2(0u, u32(v9.x + v14.x < v9.x));
v0 = v0 + m2 + vec2(0u, u32(v0.x + m2.x < v0.x));
// d = rotr64(d ^ a, 32)
- xor = v12 ^ v0;
- v12 = xor.yx;
+ v12 = v12.yx ^ v0.yx;
// c = c + d
v8 = v8 + v12 + vec2(0u, u32(v8.x + v12.x < v8.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v13 ^ v1;
- v13 = xor.yx;
+ v13 = v13.yx ^ v1.yx;
// c = c + d
v9 = v9 + v13 + vec2(0u, u32(v9.x + v13.x < v9.x));
v2 = v2 + nonce + vec2(0u, u32(v2.x + nonce.x < v2.x));
// d = rotr64(d ^ a, 32)
- xor = v14 ^ v2;
- v14 = xor.yx;
+ v14 = v14.yx ^ v2.yx;
// c = c + d
v10 = v10 + v14 + vec2(0u, u32(v10.x + v14.x < v10.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v15 ^ v3;
- v15 = xor.yx;
+ v15 = v15.yx ^ v3.yx;
// c = c + d
v11 = v11 + v15 + vec2(0u, u32(v11.x + v15.x < v11.x));
v0 = v0 + m4 + vec2(0u, u32(v0.x + m4.x < v0.x));
// d = rotr64(d ^ a, 32)
- xor = v15 ^ v0;
- v15 = xor.yx;
+ v15 = v15.yx ^ v0.yx;
// c = c + d
v10 = v10 + v15 + vec2(0u, u32(v10.x + v15.x < v10.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v12 ^ v1;
- v12 = xor.yx;
+ v12 = v12.yx ^ v1.yx;
// c = c + d
v11 = v11 + v12 + vec2(0u, u32(v11.x + v12.x < v11.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v13 ^ v2;
- v13 = xor.yx;
+ v13 = v13.yx ^ v2.yx;
// c = c + d
v8 = v8 + v13 + vec2(0u, u32(v8.x + v13.x < v8.x));
v3 = v3 + m1 + vec2(0u, u32(v3.x + m1.x < v3.x));
// d = rotr64(d ^ a, 32)
- xor = v14 ^ v3;
- v14 = xor.yx;
+ v14 = v14.yx ^ v3.yx;
// c = c + d
v9 = v9 + v14 + vec2(0u, u32(v9.x + v14.x < v9.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v12 ^ v0;
- v12 = xor.yx;
+ v12 = v12.yx ^ v0.yx;
// c = c + d
v8 = v8 + v12 + vec2(0u, u32(v8.x + v12.x < v8.x));
v1 = v1 + m1 + vec2(0u, u32(v1.x + m1.x < v1.x));
// d = rotr64(d ^ a, 32)
- xor = v13 ^ v1;
- v13 = xor.yx;
+ v13 = v13.yx ^ v1.yx;
// c = c + d
v9 = v9 + v13 + vec2(0u, u32(v9.x + v13.x < v9.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v14 ^ v2;
- v14 = xor.yx;
+ v14 = v14.yx ^ v2.yx;
// c = c + d
v10 = v10 + v14 + vec2(0u, u32(v10.x + v14.x < v10.x));
v3 = v3 + m4 + vec2(0u, u32(v3.x + m4.x < v3.x));
// d = rotr64(d ^ a, 32)
- xor = v15 ^ v3;
- v15 = xor.yx;
+ v15 = v15.yx ^ v3.yx;
// c = c + d
v11 = v11 + v15 + vec2(0u, u32(v11.x + v15.x < v11.x));
v0 = v0 + nonce + vec2(0u, u32(v0.x + nonce.x < v0.x));
// d = rotr64(d ^ a, 32)
- xor = v15 ^ v0;
- v15 = xor.yx;
+ v15 = v15.yx ^ v0.yx;
// c = c + d
v10 = v10 + v15 + vec2(0u, u32(v10.x + v15.x < v10.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v12 ^ v1;
- v12 = xor.yx;
+ v12 = v12.yx ^ v1.yx;
// c = c + d
v11 = v11 + v12 + vec2(0u, u32(v11.x + v12.x < v11.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v13 ^ v2;
- v13 = xor.yx;
+ v13 = v13.yx ^ v2.yx;
// c = c + d
v8 = v8 + v13 + vec2(0u, u32(v8.x + v13.x < v8.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v14 ^ v3;
- v14 = xor.yx;
+ v14 = v14.yx ^ v3.yx;
// c = c + d
v9 = v9 + v14 + vec2(0u, u32(v9.x + v14.x < v9.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v12 ^ v0;
- v12 = xor.yx;
+ v12 = v12.yx ^ v0.yx;
// c = c + d
v8 = v8 + v12 + vec2(0u, u32(v8.x + v12.x < v8.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v13 ^ v1;
- v13 = xor.yx;
+ v13 = v13.yx ^ v1.yx;
// c = c + d
v9 = v9 + v13 + vec2(0u, u32(v9.x + v13.x < v9.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v14 ^ v2;
- v14 = xor.yx;
+ v14 = v14.yx ^ v2.yx;
// c = c + d
v10 = v10 + v14 + vec2(0u, u32(v10.x + v14.x < v10.x));
v3 = v3 + m3 + vec2(0u, u32(v3.x + m3.x < v3.x));
// d = rotr64(d ^ a, 32)
- xor = v15 ^ v3;
- v15 = xor.yx;
+ v15 = v15.yx ^ v3.yx;
// c = c + d
v11 = v11 + v15 + vec2(0u, u32(v11.x + v15.x < v11.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v15 ^ v0;
- v15 = xor.yx;
+ v15 = v15.yx ^ v0.yx;
// c = c + d
v10 = v10 + v15 + vec2(0u, u32(v10.x + v15.x < v10.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v12 ^ v1;
- v12 = xor.yx;
+ v12 = v12.yx ^ v1.yx;
// c = c + d
v11 = v11 + v12 + vec2(0u, u32(v11.x + v12.x < v11.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v13 ^ v2;
- v13 = xor.yx;
+ v13 = v13.yx ^ v2.yx;
// c = c + d
v8 = v8 + v13 + vec2(0u, u32(v8.x + v13.x < v8.x));
v3 = v3 + m2 + vec2(0u, u32(v3.x + m2.x < v3.x));
// d = rotr64(d ^ a, 32)
- xor = v14 ^ v3;
- v14 = xor.yx;
+ v14 = v14.yx ^ v3.yx;
// c = c + d
v9 = v9 + v14 + vec2(0u, u32(v9.x + v14.x < v9.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v12 ^ v0;
- v12 = xor.yx;
+ v12 = v12.yx ^ v0.yx;
// c = c + d
v8 = v8 + v12 + vec2(0u, u32(v8.x + v12.x < v8.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v13 ^ v1;
- v13 = xor.yx;
+ v13 = v13.yx ^ v1.yx;
// c = c + d
v9 = v9 + v13 + vec2(0u, u32(v9.x + v13.x < v9.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v14 ^ v2;
- v14 = xor.yx;
+ v14 = v14.yx ^ v2.yx;
// c = c + d
v10 = v10 + v14 + vec2(0u, u32(v10.x + v14.x < v10.x));
v3 = v3 + nonce + vec2(0u, u32(v3.x + nonce.x < v3.x));
// d = rotr64(d ^ a, 32)
- xor = v15 ^ v3;
- v15 = xor.yx;
+ v15 = v15.yx ^ v3.yx;
// c = c + d
v11 = v11 + v15 + vec2(0u, u32(v11.x + v15.x < v11.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v15 ^ v0;
- v15 = xor.yx;
+ v15 = v15.yx ^ v0.yx;
// c = c + d
v10 = v10 + v15 + vec2(0u, u32(v10.x + v15.x < v10.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v12 ^ v1;
- v12 = xor.yx;
+ v12 = v12.yx ^ v1.yx;
// c = c + d
v11 = v11 + v12 + vec2(0u, u32(v11.x + v12.x < v11.x));
v2 = v2 + m1 + vec2(0u, u32(v2.x + m1.x < v2.x));
// d = rotr64(d ^ a, 32)
- xor = v13 ^ v2;
- v13 = xor.yx;
+ v13 = v13.yx ^ v2.yx;
// c = c + d
v8 = v8 + v13 + vec2(0u, u32(v8.x + v13.x < v8.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v14 ^ v3;
- v14 = xor.yx;
+ v14 = v14.yx ^ v3.yx;
// c = c + d
v9 = v9 + v14 + vec2(0u, u32(v9.x + v14.x < v9.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v12 ^ v0;
- v12 = xor.yx;
+ v12 = v12.yx ^ v0.yx;
// c = c + d
v8 = v8 + v12 + vec2(0u, u32(v8.x + v12.x < v8.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v13 ^ v1;
- v13 = xor.yx;
+ v13 = v13.yx ^ v1.yx;
// c = c + d
v9 = v9 + v13 + vec2(0u, u32(v9.x + v13.x < v9.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v14 ^ v2;
- v14 = xor.yx;
+ v14 = v14.yx ^ v2.yx;
// c = c + d
v10 = v10 + v14 + vec2(0u, u32(v10.x + v14.x < v10.x));
v3 = v3 + m1 + vec2(0u, u32(v3.x + m1.x < v3.x));
// d = rotr64(d ^ a, 32)
- xor = v15 ^ v3;
- v15 = xor.yx;
+ v15 = v15.yx ^ v3.yx;
// c = c + d
v11 = v11 + v15 + vec2(0u, u32(v11.x + v15.x < v11.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v15 ^ v0;
- v15 = xor.yx;
+ v15 = v15.yx ^ v0.yx;
// c = c + d
v10 = v10 + v15 + vec2(0u, u32(v10.x + v15.x < v10.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v12 ^ v1;
- v12 = xor.yx;
+ v12 = v12.yx ^ v1.yx;
// c = c + d
v11 = v11 + v12 + vec2(0u, u32(v11.x + v12.x < v11.x));
v2 = v2 + m3 + vec2(0u, u32(v2.x + m3.x < v2.x));
// d = rotr64(d ^ a, 32)
- xor = v13 ^ v2;
- v13 = xor.yx;
+ v13 = v13.yx ^ v2.yx;
// c = c + d
v8 = v8 + v13 + vec2(0u, u32(v8.x + v13.x < v8.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v14 ^ v3;
- v14 = xor.yx;
+ v14 = v14.yx ^ v3.yx;
// c = c + d
v9 = v9 + v14 + vec2(0u, u32(v9.x + v14.x < v9.x));
v0 = v0 + nonce + vec2(0u, u32(v0.x + nonce.x < v0.x));
// d = rotr64(d ^ a, 32)
- xor = v12 ^ v0;
- v12 = xor.yx;
+ v12 = v12.yx ^ v0.yx;
// c = c + d
v8 = v8 + v12 + vec2(0u, u32(v8.x + v12.x < v8.x));
v1 = v1 + m2 + vec2(0u, u32(v1.x + m2.x < v1.x));
// d = rotr64(d ^ a, 32)
- xor = v13 ^ v1;
- v13 = xor.yx;
+ v13 = v13.yx ^ v1.yx;
// c = c + d
v9 = v9 + v13 + vec2(0u, u32(v9.x + v13.x < v9.x));
v2 = v2 + m4 + vec2(0u, u32(v2.x + m4.x < v2.x));
// d = rotr64(d ^ a, 32)
- xor = v14 ^ v2;
- v14 = xor.yx;
+ v14 = v14.yx ^ v2.yx;
// c = c + d
v10 = v10 + v14 + vec2(0u, u32(v10.x + v14.x < v10.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v15 ^ v3;
- v15 = xor.yx;
+ v15 = v15.yx ^ v3.yx;
// c = c + d
v11 = v11 + v15 + vec2(0u, u32(v11.x + v15.x < v11.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v15 ^ v0;
- v15 = xor.yx;
+ v15 = v15.yx ^ v0.yx;
// c = c + d
v10 = v10 + v15 + vec2(0u, u32(v10.x + v15.x < v10.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v12 ^ v1;
- v12 = xor.yx;
+ v12 = v12.yx ^ v1.yx;
// c = c + d
v11 = v11 + v12 + vec2(0u, u32(v11.x + v12.x < v11.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v13 ^ v2;
- v13 = xor.yx;
+ v13 = v13.yx ^ v2.yx;
// c = c + d
v8 = v8 + v13 + vec2(0u, u32(v8.x + v13.x < v8.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v14 ^ v3;
- v14 = xor.yx;
+ v14 = v14.yx ^ v3.yx;
// c = c + d
v9 = v9 + v14 + vec2(0u, u32(v9.x + v14.x < v9.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v12 ^ v0;
- v12 = xor.yx;
+ v12 = v12.yx ^ v0.yx;
// c = c + d
v8 = v8 + v12 + vec2(0u, u32(v8.x + v12.x < v8.x));
v1 = v1 + m4 + vec2(0u, u32(v1.x + m4.x < v1.x));
// d = rotr64(d ^ a, 32)
- xor = v13 ^ v1;
- v13 = xor.yx;
+ v13 = v13.yx ^ v1.yx;
// c = c + d
v9 = v9 + v13 + vec2(0u, u32(v9.x + v13.x < v9.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v14 ^ v2;
- v14 = xor.yx;
+ v14 = v14.yx ^ v2.yx;
// c = c + d
v10 = v10 + v14 + vec2(0u, u32(v10.x + v14.x < v10.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v15 ^ v3;
- v15 = xor.yx;
+ v15 = v15.yx ^ v3.yx;
// c = c + d
v11 = v11 + v15 + vec2(0u, u32(v11.x + v15.x < v11.x));
v0 = v0 + m1 + vec2(0u, u32(v0.x + m1.x < v0.x));
// d = rotr64(d ^ a, 32)
- xor = v15 ^ v0;
- v15 = xor.yx;
+ v15 = v15.yx ^ v0.yx;
// c = c + d
v10 = v10 + v15 + vec2(0u, u32(v10.x + v15.x < v10.x));
v1 = v1 + nonce + vec2(0u, u32(v1.x + nonce.x < v1.x));
// d = rotr64(d ^ a, 32)
- xor = v12 ^ v1;
- v12 = xor.yx;
+ v12 = v12.yx ^ v1.yx;
// c = c + d
v11 = v11 + v12 + vec2(0u, u32(v11.x + v12.x < v11.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v13 ^ v2;
- v13 = xor.yx;
+ v13 = v13.yx ^ v2.yx;
// c = c + d
v8 = v8 + v13 + vec2(0u, u32(v8.x + v13.x < v8.x));
// skip since adding 0u does nothing
// d = rotr64(d ^ a, 32)
- xor = v14 ^ v3;
- v14 = xor.yx;
+ v14 = v14.yx ^ v3.yx;
// c = c + d
v9 = v9 + v14 + vec2(0u, u32(v9.x + v14.x < v9.x));