diff options
-rw-r--r-- | aezref/Cargo.lock | 25 | ||||
-rw-r--r-- | aezref/Cargo.toml | 10 | ||||
-rw-r--r-- | aezref/aezv5/aesni/api.h | 4 | ||||
-rw-r--r-- | aezref/aezv5/aesni/encrypt.c | 944 | ||||
-rw-r--r-- | aezref/aezv5/ref/api.h | 4 | ||||
-rw-r--r-- | aezref/aezv5/ref/blake2b.c | 180 | ||||
-rw-r--r-- | aezref/aezv5/ref/blake2b.h | 38 | ||||
-rw-r--r-- | aezref/aezv5/ref/crypto_aead.h | 17 | ||||
-rw-r--r-- | aezref/aezv5/ref/encrypt.c | 440 | ||||
-rw-r--r-- | aezref/aezv5/ref/rijndael-alg-fst.c | 1400 | ||||
-rw-r--r-- | aezref/aezv5/ref/rijndael-alg-fst.h | 52 | ||||
-rw-r--r-- | aezref/build.rs | 7 | ||||
-rw-r--r-- | aezref/src/lib.rs | 155 | ||||
-rw-r--r-- | fuzz/Cargo.lock | 8 | ||||
-rw-r--r-- | fuzz/Cargo.toml | 4 | ||||
-rw-r--r-- | fuzz/fuzz_targets/zears_vs_aez.rs | 35 |
16 files changed, 3301 insertions, 22 deletions
diff --git a/aezref/Cargo.lock b/aezref/Cargo.lock new file mode 100644 index 0000000..216470b --- /dev/null +++ b/aezref/Cargo.lock @@ -0,0 +1,25 @@ +# This file is automatically @generated by Cargo. +# It is not intended for manual editing. +version = 4 + +[[package]] +name = "aezref" +version = "0.1.0" +dependencies = [ + "cc", +] + +[[package]] +name = "cc" +version = "1.2.19" +source = "registry+https://github.com/rust-lang/crates.io-index" +checksum = "8e3a13707ac958681c13b39b458c073d0d9bc8a22cb1b2f4c8e55eb72c13f362" +dependencies = [ + "shlex", +] + +[[package]] +name = "shlex" +version = "1.3.0" +source = "registry+https://github.com/rust-lang/crates.io-index" +checksum = "0fda2ff0d084019ba4d7c6f371c95d8fd75ce3524c3cb8fb653a3023f6323e64" diff --git a/aezref/Cargo.toml b/aezref/Cargo.toml new file mode 100644 index 0000000..51a864f --- /dev/null +++ b/aezref/Cargo.toml @@ -0,0 +1,10 @@ +[package] +name = "aezref" +version = "0.1.0" +edition = "2024" +build = "build.rs" + +[dependencies] + +[build-dependencies] +cc = "1.2" diff --git a/aezref/aezv5/aesni/api.h b/aezref/aezv5/aesni/api.h new file mode 100644 index 0000000..abcd511 --- /dev/null +++ b/aezref/aezv5/aesni/api.h @@ -0,0 +1,4 @@ +#define CRYPTO_KEYBYTES 48 +#define CRYPTO_NSECBYTES 0 +#define CRYPTO_NPUBBYTES 12 +#define CRYPTO_ABYTES 16 diff --git a/aezref/aezv5/aesni/encrypt.c b/aezref/aezv5/aesni/encrypt.c new file mode 100644 index 0000000..9c86bcf --- /dev/null +++ b/aezref/aezv5/aesni/encrypt.c @@ -0,0 +1,944 @@ +/* + // AEZ v5 AES-NI version. AEZ info: http://www.cs.ucdavis.edu/~rogaway/aez + // + // REQUIREMENTS: - Intel or ARM CPU supporting AES instructions + // - Faster if all pointers are 16-byte aligned. + // - Max 16 byte nonce, 16 byte authenticator + // - Single AD (AEZ spec allows vector AD but this code doesn't) + // - Max 2^32-1 byte buffers allowed (due to using unsigned int) + // + // Written by Ted Krovetz (ted@krovetz.net). Last modified 21 March 2017. + // + // This is free and unencumbered software released into the public domain. + // + // Anyone is free to copy, modify, publish, use, compile, sell, or + // distribute this software, either in source code form or as a compiled + // binary, for any purpose, commercial or non-commercial, and by any + // means. + // + // In jurisdictions that recognize copyright laws, the author or authors + // of this software dedicate any and all copyright interest in the + // software to the public domain. We make this dedication for the benefit + // of the public at large and to the detriment of our heirs and + // successors. We intend this dedication to be an overt act of + // relinquishment in perpetuity of all present and future rights to this + // software under copyright law. + // + // THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, + // EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF + // MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. + // IN NO EVENT SHALL THE AUTHORS BE LIABLE FOR ANY CLAIM, DAMAGES OR + // OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, + // ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR + // OTHER DEALINGS IN THE SOFTWARE. + // + // For more information, please refer to <http://unlicense.org/> + */ + +#include "crypto_aead.h" +#include <stdint.h> +#include <stddef.h> + +/* ------------------------------------------------------------------------- */ +#if __AES__ /* Defined by gcc/clang when compiling for AES-NI */ +/* ------------------------------------------------------------------------- */ + +#include <smmintrin.h> +#include <wmmintrin.h> +#define block __m128i + +/* ------------------------------------------------------------------------- */ + +#define zero _mm_setzero_si128() +#define vadd(x,y) _mm_add_epi8(x,y) +#define vand(x,y) _mm_and_si128(x,y) +#define vandnot(x,y) _mm_andnot_si128(x,y) /* (~x)&y */ +#define vor(x,y) _mm_or_si128(x,y) +#define vxor(x,y) _mm_xor_si128(x,y) + +static int is_zero(block x) { return _mm_testz_si128(x,x); } /* 0 or 1 */ + +static block sll4(block x) { + return vor(_mm_srli_epi64(x, 4), _mm_slli_epi64(_mm_srli_si128(x, 8), 60)); +} + +static block srl4(block x) { + return vor(_mm_slli_epi64(x, 4), _mm_srli_epi64(_mm_slli_si128(x, 8), 60)); +} + +static __m128i bswap16(__m128i b) { + const __m128i t = _mm_set_epi8(0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15); + return _mm_shuffle_epi8(b,t); +} + +static __m128i double_block(__m128i bl) { + const __m128i mask = _mm_set_epi32(135,1,1,1); + __m128i tmp = _mm_srai_epi32(bl, 31); + tmp = _mm_and_si128(tmp, mask); + tmp = _mm_shuffle_epi32(tmp, _MM_SHUFFLE(2,1,0,3)); + bl = _mm_slli_epi32(bl, 1); + return _mm_xor_si128(bl,tmp); +} + +static __m128i aes(__m128i *key, __m128i in, __m128i first_key) { + in = vxor(in, first_key); + in = _mm_aesenc_si128 (in,key[0]); + in = _mm_aesenc_si128 (in,key[2]); + in = _mm_aesenc_si128 (in,key[5]); + in = _mm_aesenc_si128 (in,key[0]); + in = _mm_aesenc_si128 (in,key[2]); + in = _mm_aesenc_si128 (in,key[5]); + in = _mm_aesenc_si128 (in,key[0]); + in = _mm_aesenc_si128 (in,key[2]); + in = _mm_aesenc_si128 (in,key[5]); + return _mm_aesenc_si128 (in,key[0]); +} + +static __m128i aes4(__m128i in, __m128i a, __m128i b, + __m128i c, __m128i d, __m128i e) { + in = _mm_aesenc_si128(vxor(in,a),b); + in = _mm_aesenc_si128(in,c); + in = _mm_aesenc_si128(in,d); + return _mm_aesenc_si128 (in,e); +} + +#define aes4pre(in,a,b,c,d) aes4(in,a,b,c,d,zero) + +static __m128i loadu(const void *p) { return _mm_loadu_si128((__m128i*)p); } +static void storeu(const void *p, __m128i x) {_mm_storeu_si128((__m128i*)p,x);} + +#define load loadu /* Intel with AES-NI has fast unaligned loads/stores */ +#define store storeu + +/* ------------------------------------------------------------------------- */ +#elif __ARM_FEATURE_CRYPTO +/* ------------------------------------------------------------------------- */ + +#include <arm_neon.h> +#define block uint8x16_t + +#define zero vmovq_n_u8(0) +#define vadd(x,y) vaddq_u8(x,y) +#define vand(x,y) vandq_u8(x,y) +#define vandnot(x,y) vbicq_u8(y,x) /* (~x)&y */ +#define vor(x,y) vorrq_u8(x,y) +#define vxor(x,y) veorq_u8(x,y) + +static int is_zero(block x) { /* 0 or 1 */ + uint8x8_t t = vorr_u8(vget_high_u8(x), vget_low_u8(x)); + return vget_lane_u64(vreinterpret_u64_u8(t),0) == 0; +} + +static block srl4(block x) { + const block mask = {15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,0}; + uint8x16_t tmp = vandq_u8(vshrq_n_u8(vextq_u8(x, x, 1),4),mask); + return veorq_u8(tmp,vshlq_n_u8(x,4)); +} + +static block sll4(block x) { + const block mask = {0,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15}; + uint8x16_t tmp = vshlq_n_u8(vandq_u8(vextq_u8(x, x, 15),mask),4); + return veorq_u8(tmp,vshrq_n_u8(x,4)); +} + +static uint8x16_t bswap16(uint8x16_t b) { return b; } /* Not with uint8x16_t */ + +static block double_block(block b) { + const block mask = {135,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1}; + block tmp = (block)vshrq_n_s8((int8x16_t)b,7); + tmp = vandq_u8(tmp, mask); + tmp = vextq_u8(tmp, tmp, 1); /* Rotate high byte to low end */ + b = vshlq_n_u8(b,1); + return veorq_u8(tmp,b); +} + +static uint8x16_t aes(uint8x16_t *key, uint8x16_t in, uint8x16_t first_key) { + in = vaesmcq_u8(vaeseq_u8(in, first_key)); + in = vaesmcq_u8(vaeseq_u8(in, key[0])); + in = vaesmcq_u8(vaeseq_u8(in, key[2])); + in = vaesmcq_u8(vaeseq_u8(in, key[5])); + in = vaesmcq_u8(vaeseq_u8(in, key[0])); + in = vaesmcq_u8(vaeseq_u8(in, key[2])); + in = vaesmcq_u8(vaeseq_u8(in, key[5])); + in = vaesmcq_u8(vaeseq_u8(in, key[0])); + in = vaesmcq_u8(vaeseq_u8(in, key[2])); + in = vaesmcq_u8(vaeseq_u8(in, key[5])); + return vxor(in, key[0]); +} + +static uint8x16_t aes4pre(uint8x16_t in, uint8x16_t a, uint8x16_t b, + uint8x16_t c, uint8x16_t d) { + in = vaesmcq_u8(vaeseq_u8(in, a)); + in = vaesmcq_u8(vaeseq_u8(in, b)); + in = vaesmcq_u8(vaeseq_u8(in, c)); + return vaesmcq_u8(vaeseq_u8(in, d)); +} + +#define aes4(in,a,b,c,d,e) vxor(aes4pre(in,a,b,c,d),e) + +static uint8x16_t load(const void *p) { return *(uint8x16_t *)p; } +static void store(void *p, uint8x16_t x) { *(uint8x16_t *)p = x; } + +#define loadu load /* ARMv8 allows unaligned loads/stores */ +#define storeu store /* ARMv8 allows unaligned stores */ + +/* ------------------------------------------------------------------------- */ +#else +#error - This implementation requires __AES__ or __ARM_FEATURE_CRYPTO +#endif +/* ------------------------------------------------------------------------- */ + +#define vxor3(x,y,z) vxor(vxor(x,y),z) +#define vxor4(w,x,y,z) vxor(vxor(w,x),vxor(y,z)) +#define load_partial(p,n) loadu(p) + +/* +Might need a version like this if, for example, we want to load a 12-byte nonce +into a 16-byte block. + +static block load_partial(const void *p, unsigned n) { + if ((intptr_t)p % 16 == 0) return load(p); + else { + block tmp; unsigned i; + for (i=0; i<n; i++) ((char*)&tmp)[i] = ((char*)p)[i]; + return tmp; + } +} +*/ + +static const unsigned char pad[] = {0xff,0xff,0xff,0xff,0xff,0xff,0xff,0xff, + 0xff,0xff,0xff,0xff,0xff,0xff,0xff,0xff, + 0x00,0x00,0x00,0x00,0x00,0x00,0x00,0x00, + 0x00,0x00,0x00,0x00,0x00,0x00,0x00,0x00, + 0x80,0x00,0x00,0x00,0x00,0x00,0x00,0x00, + 0x00,0x00,0x00,0x00,0x00,0x00,0x00,0x00}; + +static block zero_pad(block x, unsigned zero_bytes) { + return vand(x, loadu(pad + zero_bytes)); +} + +static block one_zero_pad(block x, unsigned one_zero_bytes) { + block *p = (block*)(pad + one_zero_bytes); + return vor(vand(x, loadu(p)), loadu(p+1)); +} + +static block zero_set_byte(char val, unsigned idx) { + block tmp = zero; ((char *)&tmp)[idx] = val; return tmp; +} + +/* ------------------------------------------------------------------------- */ + +typedef struct { /* All data memory-correct except 2I register-correct */ + block I[2]; /* 1I, 2I */ + block J[3]; /* 1J,2J,4J */ + block L[3]; /* 1L,2L,4L */ + block delta3_cache; +} aez_ctx_t; + +/* ------------------------------------------------------------------------- */ + +static int blake2b(void *out, size_t outlen, + const void *key, size_t keylen, + const void *in, size_t inlen); + +/* ------------------------------------------------------------------------- */ + +void aez_setup(unsigned char *key, unsigned keylen, aez_ctx_t *ctx) { + block tmp; + if (keylen==48) { + ctx->I[0] = loadu(key); + ctx->J[0] = loadu(key+16); + ctx->L[0] = loadu(key+32); + } else { + blake2b(ctx, 48, 0, 0, key, keylen); /* Puts IJL into ctx */ + ctx->L[0] = ctx->J[0]; /* Rearrange. */ + ctx->J[0] = ctx->I[1]; /* Rearrange. */ + } + /* Fill remaining ctx locations with doublings */ + ctx->I[1] = double_block(bswap16(ctx->I[0])); /* No post-bswap */ + ctx->J[1] = bswap16(tmp = double_block(bswap16(ctx->J[0]))); + ctx->J[2] = bswap16(double_block(tmp)); + ctx->L[1] = bswap16(tmp = double_block(bswap16(ctx->L[0]))); + ctx->L[2] = bswap16(double_block(tmp)); + ctx->delta3_cache = zero; +} + +/* ------------------------------------------------------------------------- */ + +/* !! Warning !! Only handles nbytes <= 16 and abytes <= 16 */ +static block aez_hash(aez_ctx_t *ctx, char *n, unsigned nbytes, char *ad, + unsigned adbytes, unsigned abytes) { + block o1, o2, o3, o4, o5, o6, o7, o8, sum, offset, tmp; + block I=ctx->I[0], Ifordoubling = ctx->I[1], I2 = bswap16(Ifordoubling); + block L=ctx->L[0], L2=ctx->L[1],L4=ctx->L[2]; + block J=ctx->J[0], J2 = ctx->J[1], J4 = ctx->J[2], J5 = vxor(J,J4); + + /* Process abytes and nonce */ + offset = vxor4(J, J2, I2, L); + tmp = zero_set_byte((char)(8*abytes),15); + sum = aes4pre(offset,tmp,J,I,L); + + if (nbytes==16) sum = aes4(vxor(loadu(n), J4), vxor(I2, L),J,I,L,sum); + else sum = aes4(vxor(J4, I), + one_zero_pad(load_partial(n,nbytes),16-nbytes),J,I,L,sum); + + if (ad) { /* Possible easy misuse: ad==null && adbytes==0 */ + if (adbytes==0) { + ctx->delta3_cache = aes4pre(vxor(J5, I), loadu(pad+32),J,I,L); + } else { + block delta3 = zero; + offset = vxor(J5, I2); + while (adbytes >= 8*16) { + o1 = vxor(offset,L); + o2 = vxor(offset,L2); + o3 = vxor(o1,L2); + o4 = vxor(offset,L4); + o5 = vxor(o1,L4); + o6 = vxor(o2,L4); + o7 = vxor(o3,L4); + o8 = offset; + Ifordoubling = double_block(Ifordoubling); + offset = vxor(J5, bswap16(Ifordoubling)); + delta3 = vxor(delta3, aes4pre(load(ad+ 0), o1, J, I, L)); + delta3 = vxor(delta3, aes4pre(load(ad+ 16), o2, J, I, L)); + delta3 = vxor(delta3, aes4pre(load(ad+ 32), o3, J, I, L)); + delta3 = vxor(delta3, aes4pre(load(ad+ 48), o4, J, I, L)); + delta3 = vxor(delta3, aes4pre(load(ad+ 64), o5, J, I, L)); + delta3 = vxor(delta3, aes4pre(load(ad+ 80), o6, J, I, L)); + delta3 = vxor(delta3, aes4pre(load(ad+ 96), o7, J, I, L)); + delta3 = vxor(delta3, aes4pre(load(ad+112), o8, J, I, L)); + adbytes-=8*16; ad+=8*16; + } + if (adbytes >= 4*16) { + o1 = vxor(offset,L); + o2 = vxor(offset,L2); + o3 = vxor(o1,L2); + o4 = offset = vxor(offset,L4); + delta3 = vxor(delta3, aes4pre(load(ad+ 0), o1, J, I, L)); + delta3 = vxor(delta3, aes4pre(load(ad+ 16), o2, J, I, L)); + delta3 = vxor(delta3, aes4pre(load(ad+ 32), o3, J, I, L)); + delta3 = vxor(delta3, aes4pre(load(ad+ 48), o4, J, I, L)); + adbytes-=4*16; ad+=4*16; + } + if (adbytes >= 2*16) { + o1 = vxor(offset,L); + o2 = offset = vxor(offset,L2); + delta3 = vxor(delta3, aes4pre(load(ad+ 0), o1, J, I, L)); + delta3 = vxor(delta3, aes4pre(load(ad+ 16), o2, J, I, L)); + adbytes-=2*16; ad+=2*16; + } + if (adbytes >= 1*16) { + o1 = vxor(offset,L); + delta3 = vxor(delta3, aes4pre(load(ad+ 0), o1, J, I, L)); + adbytes-=1*16; ad+=1*16; + } + if (adbytes) { + tmp = vxor3(J5, I, one_zero_pad(load(ad),16-adbytes)); + delta3 = aes4(vxor(J5, I), one_zero_pad(load(ad),16-adbytes), + J, I, L, delta3); + } + ctx->delta3_cache = delta3; + } + } + return vxor(sum,ctx->delta3_cache); +} + +/* ------------------------------------------------------------------------- */ + +static block pass_one(aez_ctx_t *ctx, block *src, unsigned bytes, block *dst) { + block o1, o2, o3, o4, o5, o6, o7, o8, offset, tmp, sum=zero; + block I=ctx->I[0], Ifordoubling = ctx->I[1]; + block L=ctx->L[0], L2=ctx->L[1],L4=ctx->L[2]; + block J=ctx->J[0]; + offset = vxor(J, bswap16(Ifordoubling)); + while (bytes >= 16*16) { + o1 = vxor(offset,L); + o2 = vxor(offset,L2); + o3 = vxor(o1,L2); + o4 = vxor(offset,L4); + o5 = vxor(o1,L4); + o6 = vxor(o2,L4); + o7 = vxor(o3,L4); + o8 = offset; + Ifordoubling = double_block(Ifordoubling); + offset = vxor(J,bswap16(Ifordoubling)); + store(dst+ 0, aes4(load(src + 1),o1, J, I, L, load(src+ 0))); + store(dst+ 2, aes4(load(src + 3),o2, J, I, L, load(src+ 2))); + store(dst+ 4, aes4(load(src + 5),o3, J, I, L, load(src+ 4))); + store(dst+ 6, aes4(load(src + 7),o4, J, I, L, load(src+ 6))); + store(dst+ 8, aes4(load(src + 9),o5, J, I, L, load(src+ 8))); + store(dst+10, aes4(load(src +11),o6, J, I, L, load(src+10))); + store(dst+12, aes4(load(src +13),o7, J, I, L, load(src+12))); + store(dst+14, aes4(load(src +15),o8, J, I, L, load(src+14))); + tmp=aes4(I,load(dst+ 0),J,I,L,load(src+ 1));store(dst+ 1,tmp); + sum=vxor(sum,tmp); + tmp=aes4(I,load(dst+ 2),J,I,L,load(src+ 3)); + store(dst+ 3,tmp);sum=vxor(sum,tmp); + tmp=aes4(I,load(dst+ 4),J,I,L,load(src+ 5)); + store(dst+ 5,tmp);sum=vxor(sum,tmp); + tmp=aes4(I,load(dst+ 6),J,I,L,load(src+ 7)); + store(dst+ 7,tmp);sum=vxor(sum,tmp); + tmp=aes4(I,load(dst+ 8),J,I,L,load(src+ 9)); + store(dst+ 9,tmp);sum=vxor(sum,tmp); + tmp=aes4(I,load(dst+10),J,I,L,load(src+11)); + store(dst+11,tmp);sum=vxor(sum,tmp); + tmp=aes4(I,load(dst+12),J,I,L,load(src+13)); + store(dst+13,tmp);sum=vxor(sum,tmp); + tmp=aes4(I,load(dst+14),J,I,L,load(src+15)); + store(dst+15,tmp);sum=vxor(sum,tmp); + bytes -= 16*16; dst += 16; src += 16; + } + if (bytes >= 8*16) { + o1 = vxor(offset,L); + o2 = vxor(offset,L2); + o3 = vxor(o1,L2); + o4 = offset = vxor(offset,L4); + store(dst+ 0, aes4(load(src + 1),o1, J, I, L, load(src+ 0))); + store(dst+ 2, aes4(load(src + 3),o2, J, I, L, load(src+ 2))); + store(dst+ 4, aes4(load(src + 5),o3, J, I, L, load(src+ 4))); + store(dst+ 6, aes4(load(src + 7),o4, J, I, L, load(src+ 6))); + tmp=aes4(I,load(dst+ 0),J,I,L,load(src+ 1)); + store(dst+ 1,tmp);sum=vxor(sum,tmp); + tmp=aes4(I,load(dst+ 2),J,I,L,load(src+ 3)); + store(dst+ 3,tmp);sum=vxor(sum,tmp); + tmp=aes4(I,load(dst+ 4),J,I,L,load(src+ 5)); + store(dst+ 5,tmp);sum=vxor(sum,tmp); + tmp=aes4(I,load(dst+ 6),J,I,L,load(src+ 7)); + store(dst+ 7,tmp);sum=vxor(sum,tmp); + bytes -= 8*16; dst += 8; src += 8; + } + if (bytes >= 4*16) { + o1 = vxor(offset,L); + o2 = offset = vxor(offset,L2); + store(dst+ 0, aes4(load(src + 1),o1, J, I, L, load(src+ 0))); + store(dst+ 2, aes4(load(src + 3),o2, J, I, L, load(src+ 2))); + tmp=aes4(I,load(dst+ 0),J,I,L,load(src+ 1)); + store(dst+ 1,tmp);sum=vxor(sum,tmp); + tmp=aes4(I,load(dst+ 2),J,I,L,load(src+ 3)); + store(dst+ 3,tmp);sum=vxor(sum,tmp); + bytes -= 4*16; dst += 4; src += 4; + } + if (bytes) { + o1 = vxor(offset,L); + store(dst+ 0, aes4(load(src + 1),o1, J, I, L, load(src+ 0))); + tmp=aes4(I,load(dst+ 0),J,I,L,load(src+ 1)); + store(dst+ 1,tmp);sum=vxor(sum,tmp); + } + return sum; +} + +/* ------------------------------------------------------------------------- */ + +static block pass_two(aez_ctx_t *ctx, block s, unsigned bytes, block *dst) { + block o1, o2, o3, o4, o5, o6, o7, o8, sum=zero, offset, fs[8], tmp[8]; + block I=ctx->I[0], Ifordoubling = ctx->I[1]; + block L=ctx->L[0], L2=ctx->L[1],L4=ctx->L[2]; + block J=ctx->J[0], J2=ctx->J[1], J3=vxor(J,J2); + offset = vxor(J2, bswap16(Ifordoubling)); + while (bytes >= 16*16) { + o1 = vxor(offset,L); + o2 = vxor(offset,L2); + o3 = vxor(o1,L2); + o4 = vxor(offset,L4); + o5 = vxor(o1,L4); + o6 = vxor(o2,L4); + o7 = vxor(o3,L4); + o8 = offset; + Ifordoubling = double_block(Ifordoubling); + offset = vxor(J2, bswap16(Ifordoubling)); + fs[0] = aes4pre(s,o1,J,I,L); fs[1] = aes4pre(s,o2,J,I,L); + fs[2] = aes4pre(s,o3,J,I,L); fs[3] = aes4pre(s,o4,J,I,L); + fs[4] = aes4pre(s,o5,J,I,L); fs[5] = aes4pre(s,o6,J,I,L); + fs[6] = aes4pre(s,o7,J,I,L); fs[7] = aes4pre(s,o8,J,I,L); + o1 = vxor(J3,o1); o2 = vxor(J3,o2); + o3 = vxor(J3,o3); o4 = vxor(J3,o4); + o5 = vxor(J3,o5); o6 = vxor(J3,o6); + o7 = vxor(J3,o7); o8 = vxor(J3,o8); + tmp[0] = vxor(load(dst+ 0),fs[0]); sum = vxor(sum,tmp[0]); + store(dst+ 0,vxor(load(dst+ 1),fs[0])); + tmp[1] = vxor(load(dst+ 2),fs[1]); sum = vxor(sum,tmp[1]); + store(dst+ 2,vxor(load(dst+ 3),fs[1])); + tmp[2] = vxor(load(dst+ 4),fs[2]); sum = vxor(sum,tmp[2]); + store(dst+ 4,vxor(load(dst+ 5),fs[2])); + tmp[3] = vxor(load(dst+ 6),fs[3]); sum = vxor(sum,tmp[3]); + store(dst+ 6,vxor(load(dst+ 7),fs[3])); + tmp[4] = vxor(load(dst+ 8),fs[4]); sum = vxor(sum,tmp[4]); + store(dst+ 8,vxor(load(dst+ 9),fs[4])); + tmp[5] = vxor(load(dst+10),fs[5]); sum = vxor(sum,tmp[5]); + store(dst+10,vxor(load(dst+11),fs[5])); + tmp[6] = vxor(load(dst+12),fs[6]); sum = vxor(sum,tmp[6]); + store(dst+12,vxor(load(dst+13),fs[6])); + tmp[7] = vxor(load(dst+14),fs[7]); sum = vxor(sum,tmp[7]); + store(dst+14,vxor(load(dst+15),fs[7])); + store(dst+ 1, aes4(I,load(dst+ 0), J, I, L, tmp[0])); + store(dst+ 3, aes4(I,load(dst+ 2), J, I, L, tmp[1])); + store(dst+ 5, aes4(I,load(dst+ 4), J, I, L, tmp[2])); + store(dst+ 7, aes4(I,load(dst+ 6), J, I, L, tmp[3])); + store(dst+ 9, aes4(I,load(dst+ 8), J, I, L, tmp[4])); + store(dst+11, aes4(I,load(dst+10), J, I, L, tmp[5])); + store(dst+13, aes4(I,load(dst+12), J, I, L, tmp[6])); + store(dst+15, aes4(I,load(dst+14), J, I, L, tmp[7])); + store(dst+ 0, aes4(load(dst+ 1),o1, J, I, L, load(dst+ 0))); + store(dst+ 2, aes4(load(dst+ 3),o2, J, I, L, load(dst+ 2))); + store(dst+ 4, aes4(load(dst+ 5),o3, J, I, L, load(dst+ 4))); + store(dst+ 6, aes4(load(dst+ 7),o4, J, I, L, load(dst+ 6))); + store(dst+ 8, aes4(load(dst+ 9),o5, J, I, L, load(dst+ 8))); + store(dst+10, aes4(load(dst+11),o6, J, I, L, load(dst+10))); + store(dst+12, aes4(load(dst+13),o7, J, I, L, load(dst+12))); + store(dst+14, aes4(load(dst+15),o8, J, I, L, load(dst+14))); + bytes -= 16*16; dst += 16; + } + if (bytes >= 8*16) { + o1 = vxor(offset,L); + o2 = vxor(offset,L2); + o3 = vxor(o1,L2); + o4 = offset = vxor(offset,L4); + fs[0] = aes4pre(s,o1,J,I,L); fs[1] = aes4pre(s,o2,J,I,L); + fs[2] = aes4pre(s,o3,J,I,L); fs[3] = aes4pre(s,o4,J,I,L); + o1 = vxor(J3,o1); o2 = vxor(J3,o2); + o3 = vxor(J3,o3); o4 = vxor(J3,o4); + tmp[0] = vxor(load(dst+ 0),fs[0]); sum = vxor(sum,tmp[0]); + store(dst+ 0,vxor(load(dst+ 1),fs[0])); + tmp[1] = vxor(load(dst+ 2),fs[1]); sum = vxor(sum,tmp[1]); + store(dst+ 2,vxor(load(dst+ 3),fs[1])); + tmp[2] = vxor(load(dst+ 4),fs[2]); sum = vxor(sum,tmp[2]); + store(dst+ 4,vxor(load(dst+ 5),fs[2])); + tmp[3] = vxor(load(dst+ 6),fs[3]); sum = vxor(sum,tmp[3]); + store(dst+ 6,vxor(load(dst+ 7),fs[3])); + store(dst+ 1, aes4(I,load(dst+ 0), J, I, L, tmp[0])); + store(dst+ 3, aes4(I,load(dst+ 2), J, I, L, tmp[1])); + store(dst+ 5, aes4(I,load(dst+ 4), J, I, L, tmp[2])); + store(dst+ 7, aes4(I,load(dst+ 6), J, I, L, tmp[3])); + store(dst+ 0, aes4(load(dst+ 1),o1, J, I, L, load(dst+ 0))); + store(dst+ 2, aes4(load(dst+ 3),o2, J, I, L, load(dst+ 2))); + store(dst+ 4, aes4(load(dst+ 5),o3, J, I, L, load(dst+ 4))); + store(dst+ 6, aes4(load(dst+ 7),o4, J, I, L, load(dst+ 6))); + bytes -= 8*16; dst += 8; + } + if (bytes >= 4*16) { + o1 = vxor(offset,L); + o2 = offset = vxor(offset,L2); + fs[0] = aes4pre(s,o1,J,I,L); fs[1] = aes4pre(s,o2,J,I,L); + o1 = vxor(J3,o1); o2 = vxor(J3,o2); + tmp[0] = vxor(load(dst+ 0),fs[0]); sum = vxor(sum,tmp[0]); + store(dst+ 0,vxor(load(dst+ 1),fs[0])); + tmp[1] = vxor(load(dst+ 2),fs[1]); sum = vxor(sum,tmp[1]); + store(dst+ 2,vxor(load(dst+ 3),fs[1])); + store(dst+ 1, aes4(I,load(dst+ 0), J, I, L, tmp[0])); + store(dst+ 3, aes4(I,load(dst+ 2), J, I, L, tmp[1])); + store(dst+ 0, aes4(load(dst+ 1),o1, J, I, L, load(dst+ 0))); + store(dst+ 2, aes4(load(dst+ 3),o2, J, I, L, load(dst+ 2))); + bytes -= 4*16; dst += 4; + } + if (bytes) { + o1 = vxor(offset,L); + fs[0] = aes4pre(s,o1,J,I,L); + o1 = vxor(J3,o1); + tmp[0] = vxor(load(dst+ 0),fs[0]); sum = vxor(sum,tmp[0]); + store(dst+ 0,vxor(load(dst+ 1),fs[0])); + store(dst+ 1, aes4(I,load(dst+ 0), J, I, L, tmp[0])); + store(dst+ 0, aes4(load(dst+ 1),o1, J, I, L, load(dst+ 0))); + } + return sum; +} + +/* ------------------------------------------------------------------------- */ + +static int cipher_aez_core(aez_ctx_t *ctx, block t, int d, char *src, + unsigned bytes, unsigned abytes, char *dst) { + block s, x, y, frag0, frag1, final0, final1; + block I=ctx->I[0], L=ctx->L[0], J=ctx->J[0]; + block L4=ctx->L[2], I2 = bswap16(ctx->I[1]); + unsigned i, frag_bytes, initial_bytes; + + if (!d) bytes += abytes; + frag_bytes = bytes % 32; + initial_bytes = bytes - frag_bytes - 32; + + /* Compute x and store intermediate results */ + x = pass_one(ctx, (block*)src, initial_bytes, (block*)dst); + if (frag_bytes >= 16) { + frag0 = load(src + initial_bytes); + frag1 = one_zero_pad(load(src + initial_bytes + 16), 32-frag_bytes); + x = aes4(frag0, vxor(L4, I2), J, I, L, x); + x = vxor(x, aes4pre(frag1, vxor3(I2, L4, L), J, I, L)); + } else if (frag_bytes) { + frag0 = one_zero_pad(load(src + initial_bytes), 16-frag_bytes); + x = aes4(frag0, vxor(L4, I2), J, I, L, x); + } + + /* Calculate s and final block values (y xor'd to final1 later) */ + final0 = vxor3(loadu(src + (bytes - 32)), x, t); + if (d || !abytes) final1 = loadu(src+(bytes-32)+16); + else final1 = zero_pad(loadu(src+(bytes-32)+16), abytes); + final0 = aes4(final1, vxor(I2, ctx->L[d]), J, I, L, final0); + final1 = vxor(final1, aes((block*)ctx, final0, ctx->L[d])); + s = vxor(final0, final1); + final0 = vxor(final0, aes((block*)ctx, final1, ctx->L[d^1])); + /* Decryption: final0 should hold abytes zero bytes. If not, failure */ + if (d && !is_zero(vandnot(loadu(pad+abytes),final0))) return -1; + final1 = aes4(final0, vxor(I2, ctx->L[d^1]), J, I, L, final1); + + /* Compute y and store final results */ + y = pass_two(ctx, s, initial_bytes, (block*)dst); + if (frag_bytes >= 16) { + frag0 = vxor(frag0, aes((block*)ctx, s, L4)); + frag1 = vxor(frag1, aes((block*)ctx, s, vxor(L4, L))); + frag1 = one_zero_pad(frag1, 32-frag_bytes); + y = aes4(frag0, vxor(I2, L4), J, I, L, y); + y = vxor(y, aes4pre(frag1, vxor3(I2, L4, L), J, I, L)); + store(dst + initial_bytes, frag0); + store(dst + initial_bytes + 16, frag1); + } else if (frag_bytes) { + frag0 = vxor(frag0, aes((block*)ctx, s, L4)); + frag0 = one_zero_pad(frag0, 16-frag_bytes); + y = aes4(frag0, vxor(I2, L4), J, I, L, y); + store(dst + initial_bytes, frag0); + } + + storeu(dst + (bytes - 32), vxor3(final1, y, t)); + if (!d || !abytes) + storeu(dst + (bytes - 32) + 16, final0); + else { + for (i=0; i<16-abytes; i++) + ((char*)dst + (bytes - 16))[i] = ((char*)&final0)[i]; + } + return 0; +} + +/* ------------------------------------------------------------------------- */ + +static int cipher_aez_tiny(aez_ctx_t *ctx, block t, int d, char *src, + unsigned bytes, unsigned abytes, char *dst) { + block l, r, tmp, one, rcon, buf[2], mask_10, mask_ff; + block I=ctx->I[0], L=ctx->L[0], J=ctx->J[0], t_orig = t; + block L2=ctx->L[1], L4=ctx->L[2], I2 = bswap16(ctx->I[1]); + unsigned rnds, i; + + /* load src into buf, zero pad, update bytes for abytes */ + if (bytes >= 16) { + buf[0] = load(src); + buf[1] = zero_pad(load_partial(src+16,bytes-16),32-bytes); + } else { + buf[0] = zero_pad(load_partial(src,bytes),16-bytes); + buf[1] = zero; + } + if (!d) bytes += abytes; + + /* load l/r, create 10* padding masks, shift r 4 bits if odd length */ + l = buf[0]; + r = loadu((char*)buf+bytes/2); + mask_ff = loadu(pad+16-bytes/2); + mask_10 = loadu(pad+32-bytes/2); + if (bytes&1) { /* Odd length. Deal with nibbles. */ + mask_10 = sll4(mask_10); + ((char*)&mask_ff)[bytes/2] = (char)0xf0; + r = bswap16(r); + r = srl4(r); + r = bswap16(r); + } + r = vor(vand(r, mask_ff), mask_10); + + /* Add tweak offset into t, and determine the number of rounds */ + if (bytes >= 16) { + t = vxor4(t, I2, L2, L4); /* (0,6) offset */ + rnds = 8; + } else { + t = vxor(vxor4(t, I2, L2, L4), L); /* (0,7) offset */ + if (bytes>=3) rnds = 10; else if (bytes==2) rnds = 16; else rnds = 24; + } + + if (!d) { + one = zero_set_byte(1,15); + rcon = zero; + } else { + one = zero_set_byte(-1,15); + rcon = zero_set_byte((char)(rnds-1),15); + } + + if ((d) && (bytes < 16)) { + block offset = vxor3(I2, L, L2); + tmp = vor(l, loadu(pad+32)); + tmp = aes4pre(t_orig, vxor(tmp,offset), J, I, L); + tmp = vand(tmp, loadu(pad+32)); + l = vxor(l, tmp); + } + + /* Feistel */ + for (i=0; i<rnds; i+=2) { + l = vor(vand(aes4(t,vxor(r,rcon), J, I, L, l), mask_ff), mask_10); + rcon = vadd(rcon,one); + r = vor(vand(aes4(t,vxor(l,rcon), J, I, L, r), mask_ff), mask_10); + rcon = vadd(rcon,one); + } + buf[0] = r; + if (bytes&1) { + l = bswap16(l); + l = sll4(l); + l = bswap16(l); + r = vand(loadu((char*)buf+bytes/2), zero_set_byte((char)0xf0,0)); + l = vor(l, r); + } + storeu((char*)buf+bytes/2, l); + if (d) { + bytes -= abytes; + if (abytes==16) tmp = loadu((char*)buf+bytes); + else { + tmp = zero; + for (i=0; i<abytes; i++) ((char*)&tmp)[i] = ((char*)buf+bytes)[i]; + } + if (!is_zero(tmp)) return -1; + } else if (bytes < 16) { + block offset = vxor3(I2, L, L2); + tmp = vor(zero_pad(buf[0], 16-bytes), loadu(pad+32)); + tmp = aes4pre(t_orig,vxor(tmp,offset), J, I, L); + buf[0] = vxor(buf[0], vand(tmp, loadu(pad+32))); + } + for (i=0; i<bytes; i++) dst[i] = ((char*)buf)[i]; + return 0; +} + +/* ------------------------------------------------------------------------- */ + +void aez_encrypt(aez_ctx_t *ctx, char *n, unsigned nbytes, + char *ad, unsigned adbytes, unsigned abytes, + char *src, unsigned bytes, char *dst) { + + block t = aez_hash(ctx, n, nbytes, ad, adbytes, abytes); + if (bytes==0) { + unsigned i; + t = aes((block*)ctx, t, vxor(ctx->L[0], ctx->L[1])); + for (i=0; i<abytes; i++) dst[i] = ((char*)&t)[i]; + } else if (bytes+abytes < 32) + cipher_aez_tiny(ctx, t, 0, src, bytes, abytes, dst); + else + cipher_aez_core(ctx, t, 0, src, bytes, abytes, dst); +} + +/* ------------------------------------------------------------------------- */ + +int aez_decrypt(aez_ctx_t *ctx, char *n, unsigned nbytes, + char *ad, unsigned adbytes, unsigned abytes, + char *src, unsigned bytes, char *dst) { + + block t; + if (bytes < abytes) return -1; + t = aez_hash(ctx, n, nbytes, ad, adbytes, abytes); + if (bytes==abytes) { + block claimed = zero_pad(load_partial(src,abytes), 16-abytes); + t = zero_pad(aes((block*)ctx, t, vxor(ctx->L[0], ctx->L[1])), 16-abytes); + return is_zero(vandnot(t, claimed)) - 1; /* is_zero return 0 or 1 */ + } else if (bytes < 32) { + return cipher_aez_tiny(ctx, t, 1, src, bytes, abytes, dst); + } else { + return cipher_aez_core(ctx, t, 1, src, bytes, abytes, dst); + } +} + +/* ------------------------------------------------------------------------- */ +/* Reference Blake2b code, here for convenience, and not for speed. */ +/* Dowloaded Sep 2015 from https://github.com/mjosaarinen/blake2_mjosref */ + +#include <stdint.h> + +typedef struct { + uint8_t b[128]; + uint64_t h[8]; + uint64_t t[2]; + size_t c; + size_t outlen; +} blake2b_ctx; + +#ifndef ROTR64 +#define ROTR64(x, y) (((x) >> (y)) ^ ((x) << (64 - (y)))) +#endif + +#define B2B_GET64(p) \ +(((uint64_t) ((uint8_t *) (p))[0]) ^ \ +(((uint64_t) ((uint8_t *) (p))[1]) << 8) ^ \ +(((uint64_t) ((uint8_t *) (p))[2]) << 16) ^ \ +(((uint64_t) ((uint8_t *) (p))[3]) << 24) ^ \ +(((uint64_t) ((uint8_t *) (p))[4]) << 32) ^ \ +(((uint64_t) ((uint8_t *) (p))[5]) << 40) ^ \ +(((uint64_t) ((uint8_t *) (p))[6]) << 48) ^ \ +(((uint64_t) ((uint8_t *) (p))[7]) << 56)) + +#define B2B_G(a, b, c, d, x, y) { \ +v[a] = v[a] + v[b] + x; \ +v[d] = ROTR64(v[d] ^ v[a], 32); \ +v[c] = v[c] + v[d]; \ +v[b] = ROTR64(v[b] ^ v[c], 24); \ +v[a] = v[a] + v[b] + y; \ +v[d] = ROTR64(v[d] ^ v[a], 16); \ +v[c] = v[c] + v[d]; \ +v[b] = ROTR64(v[b] ^ v[c], 63); } + +static const uint64_t blake2b_iv[8] = { + 0x6A09E667F3BCC908, 0xBB67AE8584CAA73B, + 0x3C6EF372FE94F82B, 0xA54FF53A5F1D36F1, + 0x510E527FADE682D1, 0x9B05688C2B3E6C1F, + 0x1F83D9ABFB41BD6B, 0x5BE0CD19137E2179 +}; + +static void blake2b_compress(blake2b_ctx *ctx, int last) +{ + const uint8_t sigma[12][16] = { + { 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15 }, + { 14, 10, 4, 8, 9, 15, 13, 6, 1, 12, 0, 2, 11, 7, 5, 3 }, + { 11, 8, 12, 0, 5, 2, 15, 13, 10, 14, 3, 6, 7, 1, 9, 4 }, + { 7, 9, 3, 1, 13, 12, 11, 14, 2, 6, 5, 10, 4, 0, 15, 8 }, + { 9, 0, 5, 7, 2, 4, 10, 15, 14, 1, 11, 12, 6, 8, 3, 13 }, + { 2, 12, 6, 10, 0, 11, 8, 3, 4, 13, 7, 5, 15, 14, 1, 9 }, + { 12, 5, 1, 15, 14, 13, 4, 10, 0, 7, 6, 3, 9, 2, 8, 11 }, + { 13, 11, 7, 14, 12, 1, 3, 9, 5, 0, 15, 4, 8, 6, 2, 10 }, + { 6, 15, 14, 9, 11, 3, 0, 8, 12, 2, 13, 7, 1, 4, 10, 5 }, + { 10, 2, 8, 4, 7, 6, 1, 5, 15, 11, 9, 14, 3, 12, 13, 0 }, + { 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15 }, + { 14, 10, 4, 8, 9, 15, 13, 6, 1, 12, 0, 2, 11, 7, 5, 3 } + }; + int i; + uint64_t v[16], m[16]; + + for (i = 0; i < 8; i++) { + v[i] = ctx->h[i]; + v[i + 8] = blake2b_iv[i]; + } + + v[12] ^= ctx->t[0]; + v[13] ^= ctx->t[1]; + if (last) + v[14] = ~v[14]; + + for (i = 0; i < 16; i++) + m[i] = B2B_GET64(&ctx->b[8 * i]); + + for (i = 0; i < 12; i++) { + B2B_G( 0, 4, 8, 12, m[sigma[i][ 0]], m[sigma[i][ 1]]); + B2B_G( 1, 5, 9, 13, m[sigma[i][ 2]], m[sigma[i][ 3]]); + B2B_G( 2, 6, 10, 14, m[sigma[i][ 4]], m[sigma[i][ 5]]); + B2B_G( 3, 7, 11, 15, m[sigma[i][ 6]], m[sigma[i][ 7]]); + B2B_G( 0, 5, 10, 15, m[sigma[i][ 8]], m[sigma[i][ 9]]); + B2B_G( 1, 6, 11, 12, m[sigma[i][10]], m[sigma[i][11]]); + B2B_G( 2, 7, 8, 13, m[sigma[i][12]], m[sigma[i][13]]); + B2B_G( 3, 4, 9, 14, m[sigma[i][14]], m[sigma[i][15]]); + } + + for( i = 0; i < 8; ++i ) + ctx->h[i] ^= v[i] ^ v[i + 8]; +} + +static void blake2b_update(blake2b_ctx *ctx, + const void *in, size_t inlen) +{ + size_t i; + + for (i = 0; i < inlen; i++) { + if (ctx->c == 128) { + ctx->t[0] += ctx->c; + if (ctx->t[0] < ctx->c) + ctx->t[1]++; + blake2b_compress(ctx, 0); + ctx->c = 0; + } + ctx->b[ctx->c++] = ((const uint8_t *) in)[i]; + } +} + +static void blake2b_final(blake2b_ctx *ctx, void *out) +{ + size_t i; + + ctx->t[0] += ctx->c; + if (ctx->t[0] < ctx->c) + ctx->t[1]++; + + while (ctx->c < 128) + ctx->b[ctx->c++] = 0; + blake2b_compress(ctx, 1); + + for (i = 0; i < ctx->outlen; i++) { + ((uint8_t *) out)[i] = + (ctx->h[i >> 3] >> (8 * (i & 7))) & 0xFF; + } +} + +static int blake2b_init(blake2b_ctx *ctx, size_t outlen, + const void *key, size_t keylen) +{ + size_t i; + + if (outlen == 0 || outlen > 64 || keylen > 64) + return -1; + + for (i = 0; i < 8; i++) + ctx->h[i] = blake2b_iv[i]; + ctx->h[0] ^= 0x01010000 ^ (keylen << 8) ^ outlen; + + ctx->t[0] = 0; + ctx->t[1] = 0; + ctx->c = 0; + ctx->outlen = outlen; + + for (i = keylen; i < 128; i++) + ctx->b[i] = 0; + if (keylen > 0) { + blake2b_update(ctx, key, keylen); + ctx->c = 128; + } + + return 0; +} + +static int blake2b(void *out, size_t outlen, + const void *key, size_t keylen, + const void *in, size_t inlen) +{ + blake2b_ctx ctx; + + if (blake2b_init(&ctx, outlen, key, keylen)) + return -1; + blake2b_update(&ctx, in, inlen); + blake2b_final(&ctx, out); + + return 0; +} + +/* ------------------------------------------------------------------------- */ +/* aez mapping for CAESAR competition */ + +int crypto_aead_encrypt( + unsigned char *c,unsigned long long *clen, + const unsigned char *m,unsigned long long mlen, + const unsigned char *ad,unsigned long long adlen, + const unsigned char *nsec, + const unsigned char *npub, + const unsigned char *k +) +{ + aez_ctx_t ctx; + (void)nsec; + if (clen) *clen = mlen+16; + aez_setup((unsigned char *)k, 48, &ctx); + aez_encrypt(&ctx, (char *)npub, 12, + (char *)ad, (unsigned)adlen, 16, + (char *)m, (unsigned)mlen, (char *)c); + return 0; +} + +int crypto_aead_decrypt( + unsigned char *m,unsigned long long *mlen, + unsigned char *nsec, + const unsigned char *c,unsigned long long clen, + const unsigned char *ad,unsigned long long adlen, + const unsigned char *npub, + const unsigned char *k +) +{ + aez_ctx_t ctx; + (void)nsec; + if (mlen) *mlen = clen-16; + aez_setup((unsigned char *)k, 48, &ctx); + return aez_decrypt(&ctx, (char *)npub, 12, + (char *)ad, (unsigned)adlen, 16, + (char *)c, (unsigned)clen, (char *)m); +} diff --git a/aezref/aezv5/ref/api.h b/aezref/aezv5/ref/api.h new file mode 100644 index 0000000..abcd511 --- /dev/null +++ b/aezref/aezv5/ref/api.h @@ -0,0 +1,4 @@ +#define CRYPTO_KEYBYTES 48 +#define CRYPTO_NSECBYTES 0 +#define CRYPTO_NPUBBYTES 12 +#define CRYPTO_ABYTES 16 diff --git a/aezref/aezv5/ref/blake2b.c b/aezref/aezv5/ref/blake2b.c new file mode 100644 index 0000000..a54f880 --- /dev/null +++ b/aezref/aezv5/ref/blake2b.c @@ -0,0 +1,180 @@ +// blake2b.c +// A simple BLAKE2b Reference Implementation. + +#include "blake2b.h" + +// Cyclic right rotation. + +#ifndef ROTR64 +#define ROTR64(x, y) (((x) >> (y)) ^ ((x) << (64 - (y)))) +#endif + +// Little-endian byte access. + +#define B2B_GET64(p) \ + (((uint64_t) ((uint8_t *) (p))[0]) ^ \ + (((uint64_t) ((uint8_t *) (p))[1]) << 8) ^ \ + (((uint64_t) ((uint8_t *) (p))[2]) << 16) ^ \ + (((uint64_t) ((uint8_t *) (p))[3]) << 24) ^ \ + (((uint64_t) ((uint8_t *) (p))[4]) << 32) ^ \ + (((uint64_t) ((uint8_t *) (p))[5]) << 40) ^ \ + (((uint64_t) ((uint8_t *) (p))[6]) << 48) ^ \ + (((uint64_t) ((uint8_t *) (p))[7]) << 56)) + +// G Mixing function. + +#define B2B_G(a, b, c, d, x, y) { \ + v[a] = v[a] + v[b] + x; \ + v[d] = ROTR64(v[d] ^ v[a], 32); \ + v[c] = v[c] + v[d]; \ + v[b] = ROTR64(v[b] ^ v[c], 24); \ + v[a] = v[a] + v[b] + y; \ + v[d] = ROTR64(v[d] ^ v[a], 16); \ + v[c] = v[c] + v[d]; \ + v[b] = ROTR64(v[b] ^ v[c], 63); } + +// Initialization Vector. + +static const uint64_t blake2b_iv[8] = { + 0x6A09E667F3BCC908, 0xBB67AE8584CAA73B, + 0x3C6EF372FE94F82B, 0xA54FF53A5F1D36F1, + 0x510E527FADE682D1, 0x9B05688C2B3E6C1F, + 0x1F83D9ABFB41BD6B, 0x5BE0CD19137E2179 +}; + +// Compression function. "last" flag indicates last block. + +static void blake2b_compress(blake2b_ctx *ctx, int last) +{ + const uint8_t sigma[12][16] = { + { 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15 }, + { 14, 10, 4, 8, 9, 15, 13, 6, 1, 12, 0, 2, 11, 7, 5, 3 }, + { 11, 8, 12, 0, 5, 2, 15, 13, 10, 14, 3, 6, 7, 1, 9, 4 }, + { 7, 9, 3, 1, 13, 12, 11, 14, 2, 6, 5, 10, 4, 0, 15, 8 }, + { 9, 0, 5, 7, 2, 4, 10, 15, 14, 1, 11, 12, 6, 8, 3, 13 }, + { 2, 12, 6, 10, 0, 11, 8, 3, 4, 13, 7, 5, 15, 14, 1, 9 }, + { 12, 5, 1, 15, 14, 13, 4, 10, 0, 7, 6, 3, 9, 2, 8, 11 }, + { 13, 11, 7, 14, 12, 1, 3, 9, 5, 0, 15, 4, 8, 6, 2, 10 }, + { 6, 15, 14, 9, 11, 3, 0, 8, 12, 2, 13, 7, 1, 4, 10, 5 }, + { 10, 2, 8, 4, 7, 6, 1, 5, 15, 11, 9, 14, 3, 12, 13, 0 }, + { 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15 }, + { 14, 10, 4, 8, 9, 15, 13, 6, 1, 12, 0, 2, 11, 7, 5, 3 } + }; + int i; + uint64_t v[16], m[16]; + + for (i = 0; i < 8; i++) { // init work variables + v[i] = ctx->h[i]; + v[i + 8] = blake2b_iv[i]; + } + + v[12] ^= ctx->t[0]; // low 64 bits of offset + v[13] ^= ctx->t[1]; // high 64 bits + if (last) // last block flag set ? + v[14] = ~v[14]; + + for (i = 0; i < 16; i++) // get little-endian words + m[i] = B2B_GET64(&ctx->b[8 * i]); + + for (i = 0; i < 12; i++) { // twelve rounds + B2B_G( 0, 4, 8, 12, m[sigma[i][ 0]], m[sigma[i][ 1]]); + B2B_G( 1, 5, 9, 13, m[sigma[i][ 2]], m[sigma[i][ 3]]); + B2B_G( 2, 6, 10, 14, m[sigma[i][ 4]], m[sigma[i][ 5]]); + B2B_G( 3, 7, 11, 15, m[sigma[i][ 6]], m[sigma[i][ 7]]); + B2B_G( 0, 5, 10, 15, m[sigma[i][ 8]], m[sigma[i][ 9]]); + B2B_G( 1, 6, 11, 12, m[sigma[i][10]], m[sigma[i][11]]); + B2B_G( 2, 7, 8, 13, m[sigma[i][12]], m[sigma[i][13]]); + B2B_G( 3, 4, 9, 14, m[sigma[i][14]], m[sigma[i][15]]); + } + + for( i = 0; i < 8; ++i ) + ctx->h[i] ^= v[i] ^ v[i + 8]; +} + +// Initialize the hashing context "ctx" with optional key "key". +// 1 <= outlen <= 64 gives the digest size in bytes. +// Secret key (also <= 64 bytes) is optional (keylen = 0). + +int blake2b_init(blake2b_ctx *ctx, size_t outlen, + const void *key, size_t keylen) // (keylen=0: no key) +{ + size_t i; + + if (outlen == 0 || outlen > 64 || keylen > 64) + return -1; // illegal parameters + + for (i = 0; i < 8; i++) // state, "param block" + ctx->h[i] = blake2b_iv[i]; + ctx->h[0] ^= 0x01010000 ^ (keylen << 8) ^ outlen; + + ctx->t[0] = 0; // input count low word + ctx->t[1] = 0; // input count high word + ctx->c = 0; // pointer within buffer + ctx->outlen = outlen; + + for (i = keylen; i < 128; i++) // zero input block + ctx->b[i] = 0; + if (keylen > 0) { + blake2b_update(ctx, key, keylen); + ctx->c = 128; // at the end + } + + return 0; +} + +// Add "inlen" bytes from "in" into the hash. + +void blake2b_update(blake2b_ctx *ctx, + const void *in, size_t inlen) // data bytes +{ + size_t i; + + for (i = 0; i < inlen; i++) { + if (ctx->c == 128) { // buffer full ? + ctx->t[0] += ctx->c; // add counters + if (ctx->t[0] < ctx->c) // carry overflow ? + ctx->t[1]++; // high word + blake2b_compress(ctx, 0); // compress (not last) + ctx->c = 0; // counter to zero + } + ctx->b[ctx->c++] = ((const uint8_t *) in)[i]; + } +} + +// Generate the message digest (size given in init). +// Result placed in "out". + +void blake2b_final(blake2b_ctx *ctx, void *out) +{ + size_t i; + + ctx->t[0] += ctx->c; // mark last block offset + if (ctx->t[0] < ctx->c) // carry overflow + ctx->t[1]++; // high word + + while (ctx->c < 128) // fill up with zeros + ctx->b[ctx->c++] = 0; + blake2b_compress(ctx, 1); // final block flag = 1 + + // little endian convert and store + for (i = 0; i < ctx->outlen; i++) { + ((uint8_t *) out)[i] = + (ctx->h[i >> 3] >> (8 * (i & 7))) & 0xFF; + } +} + +// Convenience function for all-in-one computation. + +int blake2b(void *out, size_t outlen, + const void *key, size_t keylen, + const void *in, size_t inlen) +{ + blake2b_ctx ctx; + + if (blake2b_init(&ctx, outlen, key, keylen)) + return -1; + blake2b_update(&ctx, in, inlen); + blake2b_final(&ctx, out); + + return 0; +} diff --git a/aezref/aezv5/ref/blake2b.h b/aezref/aezv5/ref/blake2b.h new file mode 100644 index 0000000..c4ebc4a --- /dev/null +++ b/aezref/aezv5/ref/blake2b.h @@ -0,0 +1,38 @@ +// blake2b.h +// BLAKE2b Hashing Context and API Prototypes + +#ifndef BLAKE2B_H +#define BLAKE2B_H + +#include <stdint.h> +#include <stddef.h> + +// state context +typedef struct { + uint8_t b[128]; // input buffer + uint64_t h[8]; // chained state + uint64_t t[2]; // total number of bytes + size_t c; // pointer for b[] + size_t outlen; // digest size +} blake2b_ctx; + +// Initialize the hashing context "ctx" with optional key "key". +// 1 <= outlen <= 64 gives the digest size in bytes. +// Secret key (also <= 64 bytes) is optional (keylen = 0). +int blake2b_init(blake2b_ctx *ctx, size_t outlen, + const void *key, size_t keylen); // secret key + +// Add "inlen" bytes from "in" into the hash. +void blake2b_update(blake2b_ctx *ctx, // context + const void *in, size_t inlen); // data to be hashed + +// Generate the message digest (size given in init). +// Result placed in "out". +void blake2b_final(blake2b_ctx *ctx, void *out); + +// All-in-one convenience function. +int blake2b(void *out, size_t outlen, // return buffer for digest + const void *key, size_t keylen, // optional secret key + const void *in, size_t inlen); // data to be hashed + +#endif diff --git a/aezref/aezv5/ref/crypto_aead.h b/aezref/aezv5/ref/crypto_aead.h new file mode 100644 index 0000000..4d9b0e2 --- /dev/null +++ b/aezref/aezv5/ref/crypto_aead.h @@ -0,0 +1,17 @@ +int crypto_aead_encrypt( + unsigned char *c,unsigned long long *clen, + const unsigned char *m,unsigned long long mlen, + const unsigned char *ad,unsigned long long adlen, + const unsigned char *nsec, + const unsigned char *npub, + const unsigned char *k +); + +int crypto_aead_decrypt( + unsigned char *m,unsigned long long *mlen, + unsigned char *nsec, + const unsigned char *c,unsigned long long clen, + const unsigned char *ad,unsigned long long adlen, + const unsigned char *npub, + const unsigned char *k +); diff --git a/aezref/aezv5/ref/encrypt.c b/aezref/aezv5/ref/encrypt.c new file mode 100644 index 0000000..9e4eb1a --- /dev/null +++ b/aezref/aezv5/ref/encrypt.c @@ -0,0 +1,440 @@ +/* +// AEZ v5 reference code. AEZ info: http://www.cs.ucdavis.edu/~rogaway/aez +// +// ** This version is slow and susceptible to side-channel attacks. ** +// ** Do not use for any purpose other than to understand AEZ. ** +// +// Written by Ted Krovetz (ted@krovetz.net). Last modified 21 March 2017. +// +// This is free and unencumbered software released into the public domain. +// +// Anyone is free to copy, modify, publish, use, compile, sell, or +// distribute this software, either in source code form or as a compiled +// binary, for any purpose, commercial or non-commercial, and by any +// means. +// +// In jurisdictions that recognize copyright laws, the author or authors +// of this software dedicate any and all copyright interest in the +// software to the public domain. We make this dedication for the benefit +// of the public at large and to the detriment of our heirs and +// successors. We intend this dedication to be an overt act of +// relinquishment in perpetuity of all present and future rights to this +// software under copyright law. +// +// THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, +// EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF +// MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. +// IN NO EVENT SHALL THE AUTHORS BE LIABLE FOR ANY CLAIM, DAMAGES OR +// OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, +// ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR +// OTHER DEALINGS IN THE SOFTWARE. +// +// For more information, please refer to <http://unlicense.org/> +*/ +#include <stdlib.h> +#include <string.h> +/* BLAKE2b384 is used in Extract(). We use Saarinen's reference version */ +/* http://github.com/mjosaarinen/blake2_mjosref (accessed 02SEP2015) */ +#include "blake2b.h" +/* We us Rijmen, Bosselaers and Barreto's AES reference code, with */ +/* rijndael-alg-fst.h modified to define INTERMEDIATE_VALUE_KAT. This gives */ +/* access to AES4 and AES10, and allows the forcing of MixColumns in the */ +/* final round. It also defines "u32", used for internal AES keys. */ +#include "rijndael-alg-fst.h" +#include "crypto_aead.h" + +typedef unsigned char byte; + +/* ------------------------------------------------------------------------- */ + +static void write32_big_endian(unsigned x, void *ptr) { + byte *p = (byte *)ptr; + p[0] = (byte)(x>>24); p[1] = (byte)(x>>16); + p[2] = (byte)(x>> 8); p[3] = (byte)(x>> 0); +} + +/* ------------------------------------------------------------------------- */ + +/* Adjust our constructed round keys to be compatible with rijndael-alg-fst */ +static void correct_key(u32 *p, unsigned nbytes) { + unsigned i; + for (i=0; i<nbytes/4; i++) write32_big_endian(p[i], p+i); +} + +/* ------------------------------------------------------------------------- */ + +static void xor_bytes(byte *src1, byte *src2, unsigned n, byte *dst) { + while (n) { n--; dst[n] = src1[n] ^ src2[n]; } +} + +/* ------------------------------------------------------------------------- */ + +static void double_block(byte *p) { + byte i, tmp = p[0]; + for (i=0; i<15; i++) + p[i] = (p[i] << 1) | (p[i+1] >> 7); + p[15] = (p[15] << 1) ^ ((tmp >> 7)?135:0); +} + +/* ------------------------------------------------------------------------- */ + +static void mult_block(unsigned x, byte *src, byte *dst) { + byte t[16], r[16]; + memcpy(t,src,16); memset(r,0,16); + while (x != 0) { + if (x&1) xor_bytes(r,t,16,r); + double_block(t); + x>>=1; + } + memcpy(dst,r,16); +} + +/* ------------------------------------------------------------------------- */ + +static void Extract(byte *K, unsigned kbytes, byte extracted_key[3*16]) { + if (kbytes==48) memcpy(extracted_key, K, 48); + else blake2b(extracted_key, 48, NULL, 0, K, kbytes); +} + +/* ------------------------------------------------------------------------- */ + +static void E(byte *K, unsigned kbytes, int j, unsigned i, + byte src[16], byte dst[16]) { + byte extracted_key[3*16], buf[16], delta[16], I[16], J[16], L[16]; + + Extract(K, kbytes, extracted_key); + memcpy(I,extracted_key,16); + memcpy(J,extracted_key+16,16); + memcpy(L,extracted_key+32,16); + + /* Encipher */ + if (j == -1) { + u32 aes_key[4*11]; + memset(aes_key,0,16); /* 0 */ + memcpy((byte*)aes_key+ 16, extracted_key, 48); /* I J L */ + correct_key(aes_key+4,3*16); + memcpy((byte*)aes_key+ 64, (byte*)aes_key+16, 48); /* I J L */ + memcpy((byte*)aes_key+112, (byte*)aes_key+16, 48); /* I J L */ + memcpy((byte*)aes_key+160, (byte*)aes_key+16, 16); /* I */ + mult_block(i,L,delta); xor_bytes(delta,src,16,buf); + rijndaelEncryptRound(aes_key, 99, buf, 10); /*incl final MixColumns*/ + } else { + u32 aes4_key[4*5]; + memset(aes4_key,0,16); + memcpy((byte*)aes4_key+16, J, 16); + memcpy((byte*)aes4_key+32, I, 16); + memcpy((byte*)aes4_key+48, L, 16); + memset((byte*)aes4_key+64,0,16); + correct_key(aes4_key+4,3*16); + mult_block(j,J,delta); + mult_block(i%8,L,buf); xor_bytes(delta, buf, 16, delta); + for (i=(i+7)/8; i>0; i--) mult_block(2,I,I); + xor_bytes(delta, I, 16, delta); + xor_bytes(delta, src, 16, buf); + rijndaelEncryptRound(aes4_key, 99, buf, 4); + } + memcpy(dst, buf, 16); +} + +/* ------------------------------------------------------------------------- */ + +static void AEZhash(byte *K, unsigned kbytes, byte *N, unsigned nbytes, + byte *A[], unsigned abytes[], unsigned veclen, unsigned tau, byte *result) { + + byte buf[16], sum[16], *p; + unsigned i, k, bytes, empty; + + /* Initialize sum with hash of tau */ + memset(buf,0,12); write32_big_endian(tau, buf+12); + E(K,kbytes,3,1,buf,sum); + + /* Hash nonce, accumulate into sum */ + empty = (nbytes==0); + for (i=1; nbytes>=16; i++, nbytes-=16, N+=16) { + E(K,kbytes,4,i,N,buf); xor_bytes(sum, buf, 16, sum); + } + if (nbytes || empty) { + memset(buf,0,16); memcpy(buf,N,nbytes); buf[nbytes]=0x80; + E(K,kbytes,4,0,buf,buf); + xor_bytes(sum, buf, 16, sum); + } + + /* Hash each vector element, accumulate into sum */ + for (k=0; k<veclen; k++) { + p = A[k]; bytes = abytes[k]; empty = (bytes==0); + for (i=1; bytes>=16; i++, bytes-=16, p+=16) { + E(K,kbytes,5+k,i,p,buf); xor_bytes(sum, buf, 16, sum); + } + if (bytes || empty) { + memset(buf,0,16); memcpy(buf,p,bytes); buf[bytes]=0x80; + E(K,kbytes,5+k,0,buf,buf); + xor_bytes(sum, buf, 16, sum); + } + } + memcpy(result,sum,16); +} + +/* ------------------------------------------------------------------------- */ + +static void AEZprf(byte *K, unsigned kbytes, byte delta[16], + unsigned bytes, byte *result) { + + byte buf[16], ctr[16]; + memset(ctr,0,16); + for ( ; bytes >= 16; bytes-=16, result+=16) { + unsigned i=15; + xor_bytes(delta, ctr, 16, buf); + E(K,kbytes,-1,3,buf,result); + do { ctr[i]++; i--; } while (ctr[i+1]==0); /* ctr+=1 */ + } + if (bytes) { + xor_bytes(delta, ctr, 16, buf); + E(K,kbytes,-1,3,buf,buf); + memcpy(result, buf, bytes); + } +} + +/* ------------------------------------------------------------------------- */ + +/* Set d=0 for EncipherAEZcore and d=1 for DecipherAEZcore */ +static void AEZcore(byte *K, unsigned kbytes, byte delta[16], + byte *in, unsigned inbytes, unsigned d, byte *out) { + byte tmp[16], X[16], Y[16], S[16]; + byte *in_orig = in, *out_orig = out; + unsigned i, inbytes_orig = inbytes; + + memset(X,0,16); memset(Y,0,16); + + /* Pass 1 over in[0:-32], store intermediate values in out[0:-32] */ + for (i=1; inbytes >= 64; i++, inbytes-=32, in+=32, out+=32) { + E(K, kbytes, 1, i, in+16, tmp); xor_bytes(in, tmp, 16, out); + E(K, kbytes, 0, 0, out, tmp); xor_bytes(in+16, tmp, 16, out+16); + xor_bytes(out+16, X, 16, X); + } + + /* Finish X calculation */ + inbytes -= 32; /* inbytes now has fragment length 0..31 */ + if (inbytes >= 16) { + E(K, kbytes, 0, 4, in, tmp); xor_bytes(X, tmp, 16, X); + inbytes -= 16; in += 16; out += 16; + memset(tmp,0,16); memcpy(tmp,in,inbytes); tmp[inbytes] = 0x80; + E(K, kbytes, 0, 5, tmp, tmp); xor_bytes(X, tmp, 16, X); + } else if (inbytes > 0) { + memset(tmp,0,16); memcpy(tmp,in,inbytes); tmp[inbytes] = 0x80; + E(K, kbytes, 0, 4, tmp, tmp); xor_bytes(X, tmp, 16, X); + } + in += inbytes; out += inbytes; + + /* Calculate S */ + E(K, kbytes, 0, 1+d, in+16, tmp); + xor_bytes(X, in, 16, out); + xor_bytes(delta, out, 16, out); + xor_bytes(tmp, out, 16, out); + E(K, kbytes, -1, 1+d, out, tmp); + xor_bytes(in+16, tmp, 16, out+16); + xor_bytes(out, out+16, 16, S); + + /* Pass 2 over intermediate values in out[32..]. Final values written */ + inbytes = inbytes_orig; out = out_orig; in = in_orig; + for (i=1; inbytes >= 64; i++, inbytes-=32, in+=32, out+=32) { + E(K, kbytes, 2, i, S, tmp); + xor_bytes(out, tmp, 16, out); xor_bytes(out+16, tmp, 16, out+16); + xor_bytes(out, Y, 16, Y); + E(K, kbytes, 0, 0, out+16, tmp); xor_bytes(out, tmp, 16, out); + E(K, kbytes, 1, i, out, tmp); xor_bytes(out+16, tmp, 16, out+16); + memcpy(tmp, out, 16); memcpy(out, out+16, 16); memcpy(out+16, tmp, 16); + } + + /* Finish Y calculation and finish encryption of fragment bytes */ + inbytes -= 32; /* inbytes now has fragment length 0..31 */ + if (inbytes >= 16) { + E(K, kbytes, -1, 4, S, tmp); xor_bytes(in, tmp, 16, out); + E(K, kbytes, 0, 4, out, tmp); xor_bytes(Y, tmp, 16, Y); + inbytes -= 16; in += 16; out += 16; + E(K, kbytes, -1, 5, S, tmp); xor_bytes(in, tmp, inbytes, tmp); + memcpy(out,tmp,inbytes); + memset(tmp+inbytes,0,16-inbytes); tmp[inbytes] = 0x80; + E(K, kbytes, 0, 5, tmp, tmp); xor_bytes(Y, tmp, 16, Y); + } else if (inbytes > 0) { + E(K, kbytes, -1, 4, S, tmp); xor_bytes(in, tmp, inbytes, tmp); + memcpy(out,tmp,inbytes); + memset(tmp+inbytes,0,16-inbytes); tmp[inbytes] = 0x80; + E(K, kbytes, 0, 4, tmp, tmp); xor_bytes(Y, tmp, 16, Y); + } + in += inbytes; out += inbytes; + + /* Finish encryption of last two blocks */ + E(K, kbytes, -1, 2-d, out+16, tmp); + xor_bytes(out, tmp, 16, out); + E(K, kbytes, 0, 2-d, out, tmp); + xor_bytes(tmp, out+16, 16, out+16); + xor_bytes(delta, out+16, 16, out+16); + xor_bytes(Y, out+16, 16, out+16); + memcpy(tmp, out, 16); memcpy(out, out+16, 16); memcpy(out+16, tmp, 16); +} + +/* ------------------------------------------------------------------------- */ + +/* Set d=0 for EncipherAEZtiny and d=1 for DecipherAEZtiny */ +static void AEZtiny(byte *K, unsigned kbytes, byte delta[16], + byte *in, unsigned inbytes, unsigned d, byte *out) { + unsigned rounds,i=7,j,k; + int step; + byte mask=0x00, pad=0x80, L[16], R[16], buf[32]; + if (inbytes==1) rounds=24; + else if (inbytes==2) rounds=16; + else if (inbytes<16) rounds=10; + else { i=6; rounds=8; } + /* Split (inbytes*8)/2 bits into L and R. Beware: May end in nibble. */ + memcpy(L, in, (inbytes+1)/2); + memcpy(R, in+inbytes/2, (inbytes+1)/2); + if (inbytes&1) { /* Must shift R left by half a byte */ + for (k=0; k<inbytes/2; k++) + R[k] = (byte)((R[k] << 4) | (R[k+1] >> 4)); + R[inbytes/2] = (byte)(R[inbytes/2] << 4); + pad = 0x08; mask = 0xf0; + } + if (d) { + if (inbytes < 16) { + memset(buf,0,16); memcpy(buf,in,inbytes); buf[0] |= 0x80; + xor_bytes(delta, buf, 16, buf); + E(K, kbytes,0,3,buf,buf); + L[0] ^= (buf[0] & 0x80); + } + j = rounds-1; step = -1; + } else { + j = 0; step = 1; + } + for (k=0; k<rounds/2; k++,j=(unsigned)((int)j+2*step)) { + memset(buf, 0, 16); + memcpy(buf,R,(inbytes+1)/2); + buf[inbytes/2] = (buf[inbytes/2] & mask) | pad; + xor_bytes(buf, delta, 16, buf); + buf[15] ^= (byte)j; + E(K, kbytes,0,i,buf,buf); + xor_bytes(L, buf, 16, L); + + memset(buf, 0, 16); + memcpy(buf,L,(inbytes+1)/2); + buf[inbytes/2] = (buf[inbytes/2] & mask) | pad; + xor_bytes(buf, delta, 16, buf); + buf[15] ^= (byte)((int)j+step); + E(K, kbytes,0,i,buf,buf); + xor_bytes(R, buf, 16, R); + } + memcpy(buf, R, inbytes/2); + memcpy(buf+inbytes/2, L, (inbytes+1)/2); + if (inbytes&1) { + for (k=inbytes-1; k>inbytes/2; k--) + buf[k] = (byte)((buf[k] >> 4) | (buf[k-1] << 4)); + buf[inbytes/2] = (byte)((L[0] >> 4) | (R[inbytes/2] & 0xf0)); + } + memcpy(out,buf,inbytes); + if ((inbytes < 16) && !d) { + memset(buf+inbytes,0,16-inbytes); buf[0] |= 0x80; + xor_bytes(delta, buf, 16, buf); + E(K, kbytes,0,3,buf,buf); + out[0] ^= (buf[0] & 0x80); + } +} + +/* ------------------------------------------------------------------------- */ + +static void Encipher(byte *K, unsigned kbytes, byte delta[16], + byte *in, unsigned inbytes, byte *out) { + if (inbytes == 0) return; + if (inbytes < 32) AEZtiny(K, kbytes, delta, in, inbytes, 0, out); + else AEZcore(K, kbytes, delta, in, inbytes, 0, out); +} + +/* ------------------------------------------------------------------------- */ + +static void Decipher(byte *K, unsigned kbytes, byte delta[16], + byte *in, unsigned inbytes, byte *out) { + if (inbytes == 0) return; + if (inbytes < 32) AEZtiny(K, kbytes, delta, in, inbytes, 1, out); + else AEZcore(K, kbytes, delta, in, inbytes, 1, out); +} + +/* ------------------------------------------------------------------------- */ + +int Decrypt(byte *K, unsigned kbytes, + byte *N, unsigned nbytes, + byte *AD[], unsigned adbytes[], + unsigned veclen, unsigned abytes, + byte *C, unsigned cbytes, byte *M) { + byte delta[16], *X, sum=0; + unsigned i; + if (cbytes < abytes) return -1; + AEZhash(K, kbytes, N, nbytes, AD, adbytes, veclen, abytes*8, delta); + X = (byte *)malloc(cbytes); + if (cbytes==abytes) { + AEZprf(K, kbytes, delta, abytes, X); + for (i=0; i<abytes; i++) sum |= (X[i] ^ C[i]); + } else { + Decipher(K, kbytes, delta, C, cbytes, X); + for (i=0; i<abytes; i++) sum |= X[cbytes-abytes+i]; + if (sum==0) memcpy(M,X,cbytes-abytes); + } + free(X); + return (sum == 0 ? 0 : -1); /* return 0 if valid, -1 if invalid */ +} + +/* ------------------------------------------------------------------------- */ + +void Encrypt(byte *K, unsigned kbytes, + byte *N, unsigned nbytes, + byte *AD[], unsigned adbytes[], + unsigned veclen, unsigned abytes, + byte *M, unsigned mbytes, byte *C) { + byte delta[16], *X; + AEZhash(K, kbytes, N, nbytes, AD, adbytes, veclen, abytes*8, delta); + if (mbytes==0) { + AEZprf(K, kbytes, delta, abytes, C); + } else { + X = (byte *)malloc(mbytes+abytes); + memcpy(X, M, mbytes); memset(X+mbytes,0,abytes); + Encipher(K, kbytes, delta, X, mbytes+abytes, X); + memcpy(C, X, mbytes+abytes); + free(X); + } +} + +/* ------------------------------------------------------------------------- */ +/* aez mapping for CAESAR competition */ + +int crypto_aead_encrypt( + unsigned char *c,unsigned long long *clen, + const unsigned char *m,unsigned long long mlen, + const unsigned char *ad,unsigned long long adlen, + const unsigned char *nsec, + const unsigned char *npub, + const unsigned char *k +) +{ + byte *AD[] = {(byte*)ad}; + unsigned adbytes[] = {(unsigned)adlen}; + (void)nsec; + if (clen) *clen = mlen+16; + Encrypt((byte*)k, 48, (byte*)npub, 12, AD, + adbytes, 1, 16, (byte*)m, mlen, (byte*)c); + return 0; +} + +int crypto_aead_decrypt( + unsigned char *m,unsigned long long *mlen, + unsigned char *nsec, + const unsigned char *c,unsigned long long clen, + const unsigned char *ad,unsigned long long adlen, + const unsigned char *npub, + const unsigned char *k +) +{ + byte *AD[] = {(byte*)ad}; + unsigned adbytes[] = {(unsigned)adlen}; + (void)nsec; + if (mlen) *mlen = clen-16; + return Decrypt((byte*)k, 48, (byte*)npub, 12, AD, + adbytes, 1, 16, (byte*)c, clen, (byte*)m); +} + diff --git a/aezref/aezv5/ref/rijndael-alg-fst.c b/aezref/aezv5/ref/rijndael-alg-fst.c new file mode 100644 index 0000000..8a10040 --- /dev/null +++ b/aezref/aezv5/ref/rijndael-alg-fst.c @@ -0,0 +1,1400 @@ +/** + * rijndael-alg-fst.c + * + * @version 3.0 (December 2000) + * + * Optimised ANSI C code for the Rijndael cipher (now AES) + * + * @author Vincent Rijmen <vincent.rijmen@esat.kuleuven.ac.be> + * @author Antoon Bosselaers <antoon.bosselaers@esat.kuleuven.ac.be> + * @author Paulo Barreto <paulo.barreto@terra.com.br> + * + * This code is hereby placed in the public domain. + * + * THIS SOFTWARE IS PROVIDED BY THE AUTHORS ''AS IS'' AND ANY EXPRESS + * OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED + * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE + * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHORS OR CONTRIBUTORS BE + * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR + * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF + * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR + * BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, + * WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE + * OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, + * EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. + */ +#include <assert.h> +#include <stdlib.h> + +#include "rijndael-alg-fst.h" + +/* +Te0[x] = S [x].[02, 01, 01, 03]; +Te1[x] = S [x].[03, 02, 01, 01]; +Te2[x] = S [x].[01, 03, 02, 01]; +Te3[x] = S [x].[01, 01, 03, 02]; +Te4[x] = S [x].[01, 01, 01, 01]; + +Td0[x] = Si[x].[0e, 09, 0d, 0b]; +Td1[x] = Si[x].[0b, 0e, 09, 0d]; +Td2[x] = Si[x].[0d, 0b, 0e, 09]; +Td3[x] = Si[x].[09, 0d, 0b, 0e]; +Td4[x] = Si[x].[01, 01, 01, 01]; +*/ + +static const u32 Te0[256] = { + 0xc66363a5U, 0xf87c7c84U, 0xee777799U, 0xf67b7b8dU, + 0xfff2f20dU, 0xd66b6bbdU, 0xde6f6fb1U, 0x91c5c554U, + 0x60303050U, 0x02010103U, 0xce6767a9U, 0x562b2b7dU, + 0xe7fefe19U, 0xb5d7d762U, 0x4dababe6U, 0xec76769aU, + 0x8fcaca45U, 0x1f82829dU, 0x89c9c940U, 0xfa7d7d87U, + 0xeffafa15U, 0xb25959ebU, 0x8e4747c9U, 0xfbf0f00bU, + 0x41adadecU, 0xb3d4d467U, 0x5fa2a2fdU, 0x45afafeaU, + 0x239c9cbfU, 0x53a4a4f7U, 0xe4727296U, 0x9bc0c05bU, + 0x75b7b7c2U, 0xe1fdfd1cU, 0x3d9393aeU, 0x4c26266aU, + 0x6c36365aU, 0x7e3f3f41U, 0xf5f7f702U, 0x83cccc4fU, + 0x6834345cU, 0x51a5a5f4U, 0xd1e5e534U, 0xf9f1f108U, + 0xe2717193U, 0xabd8d873U, 0x62313153U, 0x2a15153fU, + 0x0804040cU, 0x95c7c752U, 0x46232365U, 0x9dc3c35eU, + 0x30181828U, 0x379696a1U, 0x0a05050fU, 0x2f9a9ab5U, + 0x0e070709U, 0x24121236U, 0x1b80809bU, 0xdfe2e23dU, + 0xcdebeb26U, 0x4e272769U, 0x7fb2b2cdU, 0xea75759fU, + 0x1209091bU, 0x1d83839eU, 0x582c2c74U, 0x341a1a2eU, + 0x361b1b2dU, 0xdc6e6eb2U, 0xb45a5aeeU, 0x5ba0a0fbU, + 0xa45252f6U, 0x763b3b4dU, 0xb7d6d661U, 0x7db3b3ceU, + 0x5229297bU, 0xdde3e33eU, 0x5e2f2f71U, 0x13848497U, + 0xa65353f5U, 0xb9d1d168U, 0x00000000U, 0xc1eded2cU, + 0x40202060U, 0xe3fcfc1fU, 0x79b1b1c8U, 0xb65b5bedU, + 0xd46a6abeU, 0x8dcbcb46U, 0x67bebed9U, 0x7239394bU, + 0x944a4adeU, 0x984c4cd4U, 0xb05858e8U, 0x85cfcf4aU, + 0xbbd0d06bU, 0xc5efef2aU, 0x4faaaae5U, 0xedfbfb16U, + 0x864343c5U, 0x9a4d4dd7U, 0x66333355U, 0x11858594U, + 0x8a4545cfU, 0xe9f9f910U, 0x04020206U, 0xfe7f7f81U, + 0xa05050f0U, 0x783c3c44U, 0x259f9fbaU, 0x4ba8a8e3U, + 0xa25151f3U, 0x5da3a3feU, 0x804040c0U, 0x058f8f8aU, + 0x3f9292adU, 0x219d9dbcU, 0x70383848U, 0xf1f5f504U, + 0x63bcbcdfU, 0x77b6b6c1U, 0xafdada75U, 0x42212163U, + 0x20101030U, 0xe5ffff1aU, 0xfdf3f30eU, 0xbfd2d26dU, + 0x81cdcd4cU, 0x180c0c14U, 0x26131335U, 0xc3ecec2fU, + 0xbe5f5fe1U, 0x359797a2U, 0x884444ccU, 0x2e171739U, + 0x93c4c457U, 0x55a7a7f2U, 0xfc7e7e82U, 0x7a3d3d47U, + 0xc86464acU, 0xba5d5de7U, 0x3219192bU, 0xe6737395U, + 0xc06060a0U, 0x19818198U, 0x9e4f4fd1U, 0xa3dcdc7fU, + 0x44222266U, 0x542a2a7eU, 0x3b9090abU, 0x0b888883U, + 0x8c4646caU, 0xc7eeee29U, 0x6bb8b8d3U, 0x2814143cU, + 0xa7dede79U, 0xbc5e5ee2U, 0x160b0b1dU, 0xaddbdb76U, + 0xdbe0e03bU, 0x64323256U, 0x743a3a4eU, 0x140a0a1eU, + 0x924949dbU, 0x0c06060aU, 0x4824246cU, 0xb85c5ce4U, + 0x9fc2c25dU, 0xbdd3d36eU, 0x43acacefU, 0xc46262a6U, + 0x399191a8U, 0x319595a4U, 0xd3e4e437U, 0xf279798bU, + 0xd5e7e732U, 0x8bc8c843U, 0x6e373759U, 0xda6d6db7U, + 0x018d8d8cU, 0xb1d5d564U, 0x9c4e4ed2U, 0x49a9a9e0U, + 0xd86c6cb4U, 0xac5656faU, 0xf3f4f407U, 0xcfeaea25U, + 0xca6565afU, 0xf47a7a8eU, 0x47aeaee9U, 0x10080818U, + 0x6fbabad5U, 0xf0787888U, 0x4a25256fU, 0x5c2e2e72U, + 0x381c1c24U, 0x57a6a6f1U, 0x73b4b4c7U, 0x97c6c651U, + 0xcbe8e823U, 0xa1dddd7cU, 0xe874749cU, 0x3e1f1f21U, + 0x964b4bddU, 0x61bdbddcU, 0x0d8b8b86U, 0x0f8a8a85U, + 0xe0707090U, 0x7c3e3e42U, 0x71b5b5c4U, 0xcc6666aaU, + 0x904848d8U, 0x06030305U, 0xf7f6f601U, 0x1c0e0e12U, + 0xc26161a3U, 0x6a35355fU, 0xae5757f9U, 0x69b9b9d0U, + 0x17868691U, 0x99c1c158U, 0x3a1d1d27U, 0x279e9eb9U, + 0xd9e1e138U, 0xebf8f813U, 0x2b9898b3U, 0x22111133U, + 0xd26969bbU, 0xa9d9d970U, 0x078e8e89U, 0x339494a7U, + 0x2d9b9bb6U, 0x3c1e1e22U, 0x15878792U, 0xc9e9e920U, + 0x87cece49U, 0xaa5555ffU, 0x50282878U, 0xa5dfdf7aU, + 0x038c8c8fU, 0x59a1a1f8U, 0x09898980U, 0x1a0d0d17U, + 0x65bfbfdaU, 0xd7e6e631U, 0x844242c6U, 0xd06868b8U, + 0x824141c3U, 0x299999b0U, 0x5a2d2d77U, 0x1e0f0f11U, + 0x7bb0b0cbU, 0xa85454fcU, 0x6dbbbbd6U, 0x2c16163aU, +}; +static const u32 Te1[256] = { + 0xa5c66363U, 0x84f87c7cU, 0x99ee7777U, 0x8df67b7bU, + 0x0dfff2f2U, 0xbdd66b6bU, 0xb1de6f6fU, 0x5491c5c5U, + 0x50603030U, 0x03020101U, 0xa9ce6767U, 0x7d562b2bU, + 0x19e7fefeU, 0x62b5d7d7U, 0xe64dababU, 0x9aec7676U, + 0x458fcacaU, 0x9d1f8282U, 0x4089c9c9U, 0x87fa7d7dU, + 0x15effafaU, 0xebb25959U, 0xc98e4747U, 0x0bfbf0f0U, + 0xec41adadU, 0x67b3d4d4U, 0xfd5fa2a2U, 0xea45afafU, + 0xbf239c9cU, 0xf753a4a4U, 0x96e47272U, 0x5b9bc0c0U, + 0xc275b7b7U, 0x1ce1fdfdU, 0xae3d9393U, 0x6a4c2626U, + 0x5a6c3636U, 0x417e3f3fU, 0x02f5f7f7U, 0x4f83ccccU, + 0x5c683434U, 0xf451a5a5U, 0x34d1e5e5U, 0x08f9f1f1U, + 0x93e27171U, 0x73abd8d8U, 0x53623131U, 0x3f2a1515U, + 0x0c080404U, 0x5295c7c7U, 0x65462323U, 0x5e9dc3c3U, + 0x28301818U, 0xa1379696U, 0x0f0a0505U, 0xb52f9a9aU, + 0x090e0707U, 0x36241212U, 0x9b1b8080U, 0x3ddfe2e2U, + 0x26cdebebU, 0x694e2727U, 0xcd7fb2b2U, 0x9fea7575U, + 0x1b120909U, 0x9e1d8383U, 0x74582c2cU, 0x2e341a1aU, + 0x2d361b1bU, 0xb2dc6e6eU, 0xeeb45a5aU, 0xfb5ba0a0U, + 0xf6a45252U, 0x4d763b3bU, 0x61b7d6d6U, 0xce7db3b3U, + 0x7b522929U, 0x3edde3e3U, 0x715e2f2fU, 0x97138484U, + 0xf5a65353U, 0x68b9d1d1U, 0x00000000U, 0x2cc1ededU, + 0x60402020U, 0x1fe3fcfcU, 0xc879b1b1U, 0xedb65b5bU, + 0xbed46a6aU, 0x468dcbcbU, 0xd967bebeU, 0x4b723939U, + 0xde944a4aU, 0xd4984c4cU, 0xe8b05858U, 0x4a85cfcfU, + 0x6bbbd0d0U, 0x2ac5efefU, 0xe54faaaaU, 0x16edfbfbU, + 0xc5864343U, 0xd79a4d4dU, 0x55663333U, 0x94118585U, + 0xcf8a4545U, 0x10e9f9f9U, 0x06040202U, 0x81fe7f7fU, + 0xf0a05050U, 0x44783c3cU, 0xba259f9fU, 0xe34ba8a8U, + 0xf3a25151U, 0xfe5da3a3U, 0xc0804040U, 0x8a058f8fU, + 0xad3f9292U, 0xbc219d9dU, 0x48703838U, 0x04f1f5f5U, + 0xdf63bcbcU, 0xc177b6b6U, 0x75afdadaU, 0x63422121U, + 0x30201010U, 0x1ae5ffffU, 0x0efdf3f3U, 0x6dbfd2d2U, + 0x4c81cdcdU, 0x14180c0cU, 0x35261313U, 0x2fc3ececU, + 0xe1be5f5fU, 0xa2359797U, 0xcc884444U, 0x392e1717U, + 0x5793c4c4U, 0xf255a7a7U, 0x82fc7e7eU, 0x477a3d3dU, + 0xacc86464U, 0xe7ba5d5dU, 0x2b321919U, 0x95e67373U, + 0xa0c06060U, 0x98198181U, 0xd19e4f4fU, 0x7fa3dcdcU, + 0x66442222U, 0x7e542a2aU, 0xab3b9090U, 0x830b8888U, + 0xca8c4646U, 0x29c7eeeeU, 0xd36bb8b8U, 0x3c281414U, + 0x79a7dedeU, 0xe2bc5e5eU, 0x1d160b0bU, 0x76addbdbU, + 0x3bdbe0e0U, 0x56643232U, 0x4e743a3aU, 0x1e140a0aU, + 0xdb924949U, 0x0a0c0606U, 0x6c482424U, 0xe4b85c5cU, + 0x5d9fc2c2U, 0x6ebdd3d3U, 0xef43acacU, 0xa6c46262U, + 0xa8399191U, 0xa4319595U, 0x37d3e4e4U, 0x8bf27979U, + 0x32d5e7e7U, 0x438bc8c8U, 0x596e3737U, 0xb7da6d6dU, + 0x8c018d8dU, 0x64b1d5d5U, 0xd29c4e4eU, 0xe049a9a9U, + 0xb4d86c6cU, 0xfaac5656U, 0x07f3f4f4U, 0x25cfeaeaU, + 0xafca6565U, 0x8ef47a7aU, 0xe947aeaeU, 0x18100808U, + 0xd56fbabaU, 0x88f07878U, 0x6f4a2525U, 0x725c2e2eU, + 0x24381c1cU, 0xf157a6a6U, 0xc773b4b4U, 0x5197c6c6U, + 0x23cbe8e8U, 0x7ca1ddddU, 0x9ce87474U, 0x213e1f1fU, + 0xdd964b4bU, 0xdc61bdbdU, 0x860d8b8bU, 0x850f8a8aU, + 0x90e07070U, 0x427c3e3eU, 0xc471b5b5U, 0xaacc6666U, + 0xd8904848U, 0x05060303U, 0x01f7f6f6U, 0x121c0e0eU, + 0xa3c26161U, 0x5f6a3535U, 0xf9ae5757U, 0xd069b9b9U, + 0x91178686U, 0x5899c1c1U, 0x273a1d1dU, 0xb9279e9eU, + 0x38d9e1e1U, 0x13ebf8f8U, 0xb32b9898U, 0x33221111U, + 0xbbd26969U, 0x70a9d9d9U, 0x89078e8eU, 0xa7339494U, + 0xb62d9b9bU, 0x223c1e1eU, 0x92158787U, 0x20c9e9e9U, + 0x4987ceceU, 0xffaa5555U, 0x78502828U, 0x7aa5dfdfU, + 0x8f038c8cU, 0xf859a1a1U, 0x80098989U, 0x171a0d0dU, + 0xda65bfbfU, 0x31d7e6e6U, 0xc6844242U, 0xb8d06868U, + 0xc3824141U, 0xb0299999U, 0x775a2d2dU, 0x111e0f0fU, + 0xcb7bb0b0U, 0xfca85454U, 0xd66dbbbbU, 0x3a2c1616U, +}; +static const u32 Te2[256] = { + 0x63a5c663U, 0x7c84f87cU, 0x7799ee77U, 0x7b8df67bU, + 0xf20dfff2U, 0x6bbdd66bU, 0x6fb1de6fU, 0xc55491c5U, + 0x30506030U, 0x01030201U, 0x67a9ce67U, 0x2b7d562bU, + 0xfe19e7feU, 0xd762b5d7U, 0xabe64dabU, 0x769aec76U, + 0xca458fcaU, 0x829d1f82U, 0xc94089c9U, 0x7d87fa7dU, + 0xfa15effaU, 0x59ebb259U, 0x47c98e47U, 0xf00bfbf0U, + 0xadec41adU, 0xd467b3d4U, 0xa2fd5fa2U, 0xafea45afU, + 0x9cbf239cU, 0xa4f753a4U, 0x7296e472U, 0xc05b9bc0U, + 0xb7c275b7U, 0xfd1ce1fdU, 0x93ae3d93U, 0x266a4c26U, + 0x365a6c36U, 0x3f417e3fU, 0xf702f5f7U, 0xcc4f83ccU, + 0x345c6834U, 0xa5f451a5U, 0xe534d1e5U, 0xf108f9f1U, + 0x7193e271U, 0xd873abd8U, 0x31536231U, 0x153f2a15U, + 0x040c0804U, 0xc75295c7U, 0x23654623U, 0xc35e9dc3U, + 0x18283018U, 0x96a13796U, 0x050f0a05U, 0x9ab52f9aU, + 0x07090e07U, 0x12362412U, 0x809b1b80U, 0xe23ddfe2U, + 0xeb26cdebU, 0x27694e27U, 0xb2cd7fb2U, 0x759fea75U, + 0x091b1209U, 0x839e1d83U, 0x2c74582cU, 0x1a2e341aU, + 0x1b2d361bU, 0x6eb2dc6eU, 0x5aeeb45aU, 0xa0fb5ba0U, + 0x52f6a452U, 0x3b4d763bU, 0xd661b7d6U, 0xb3ce7db3U, + 0x297b5229U, 0xe33edde3U, 0x2f715e2fU, 0x84971384U, + 0x53f5a653U, 0xd168b9d1U, 0x00000000U, 0xed2cc1edU, + 0x20604020U, 0xfc1fe3fcU, 0xb1c879b1U, 0x5bedb65bU, + 0x6abed46aU, 0xcb468dcbU, 0xbed967beU, 0x394b7239U, + 0x4ade944aU, 0x4cd4984cU, 0x58e8b058U, 0xcf4a85cfU, + 0xd06bbbd0U, 0xef2ac5efU, 0xaae54faaU, 0xfb16edfbU, + 0x43c58643U, 0x4dd79a4dU, 0x33556633U, 0x85941185U, + 0x45cf8a45U, 0xf910e9f9U, 0x02060402U, 0x7f81fe7fU, + 0x50f0a050U, 0x3c44783cU, 0x9fba259fU, 0xa8e34ba8U, + 0x51f3a251U, 0xa3fe5da3U, 0x40c08040U, 0x8f8a058fU, + 0x92ad3f92U, 0x9dbc219dU, 0x38487038U, 0xf504f1f5U, + 0xbcdf63bcU, 0xb6c177b6U, 0xda75afdaU, 0x21634221U, + 0x10302010U, 0xff1ae5ffU, 0xf30efdf3U, 0xd26dbfd2U, + 0xcd4c81cdU, 0x0c14180cU, 0x13352613U, 0xec2fc3ecU, + 0x5fe1be5fU, 0x97a23597U, 0x44cc8844U, 0x17392e17U, + 0xc45793c4U, 0xa7f255a7U, 0x7e82fc7eU, 0x3d477a3dU, + 0x64acc864U, 0x5de7ba5dU, 0x192b3219U, 0x7395e673U, + 0x60a0c060U, 0x81981981U, 0x4fd19e4fU, 0xdc7fa3dcU, + 0x22664422U, 0x2a7e542aU, 0x90ab3b90U, 0x88830b88U, + 0x46ca8c46U, 0xee29c7eeU, 0xb8d36bb8U, 0x143c2814U, + 0xde79a7deU, 0x5ee2bc5eU, 0x0b1d160bU, 0xdb76addbU, + 0xe03bdbe0U, 0x32566432U, 0x3a4e743aU, 0x0a1e140aU, + 0x49db9249U, 0x060a0c06U, 0x246c4824U, 0x5ce4b85cU, + 0xc25d9fc2U, 0xd36ebdd3U, 0xacef43acU, 0x62a6c462U, + 0x91a83991U, 0x95a43195U, 0xe437d3e4U, 0x798bf279U, + 0xe732d5e7U, 0xc8438bc8U, 0x37596e37U, 0x6db7da6dU, + 0x8d8c018dU, 0xd564b1d5U, 0x4ed29c4eU, 0xa9e049a9U, + 0x6cb4d86cU, 0x56faac56U, 0xf407f3f4U, 0xea25cfeaU, + 0x65afca65U, 0x7a8ef47aU, 0xaee947aeU, 0x08181008U, + 0xbad56fbaU, 0x7888f078U, 0x256f4a25U, 0x2e725c2eU, + 0x1c24381cU, 0xa6f157a6U, 0xb4c773b4U, 0xc65197c6U, + 0xe823cbe8U, 0xdd7ca1ddU, 0x749ce874U, 0x1f213e1fU, + 0x4bdd964bU, 0xbddc61bdU, 0x8b860d8bU, 0x8a850f8aU, + 0x7090e070U, 0x3e427c3eU, 0xb5c471b5U, 0x66aacc66U, + 0x48d89048U, 0x03050603U, 0xf601f7f6U, 0x0e121c0eU, + 0x61a3c261U, 0x355f6a35U, 0x57f9ae57U, 0xb9d069b9U, + 0x86911786U, 0xc15899c1U, 0x1d273a1dU, 0x9eb9279eU, + 0xe138d9e1U, 0xf813ebf8U, 0x98b32b98U, 0x11332211U, + 0x69bbd269U, 0xd970a9d9U, 0x8e89078eU, 0x94a73394U, + 0x9bb62d9bU, 0x1e223c1eU, 0x87921587U, 0xe920c9e9U, + 0xce4987ceU, 0x55ffaa55U, 0x28785028U, 0xdf7aa5dfU, + 0x8c8f038cU, 0xa1f859a1U, 0x89800989U, 0x0d171a0dU, + 0xbfda65bfU, 0xe631d7e6U, 0x42c68442U, 0x68b8d068U, + 0x41c38241U, 0x99b02999U, 0x2d775a2dU, 0x0f111e0fU, + 0xb0cb7bb0U, 0x54fca854U, 0xbbd66dbbU, 0x163a2c16U, +}; +static const u32 Te3[256] = { + + 0x6363a5c6U, 0x7c7c84f8U, 0x777799eeU, 0x7b7b8df6U, + 0xf2f20dffU, 0x6b6bbdd6U, 0x6f6fb1deU, 0xc5c55491U, + 0x30305060U, 0x01010302U, 0x6767a9ceU, 0x2b2b7d56U, + 0xfefe19e7U, 0xd7d762b5U, 0xababe64dU, 0x76769aecU, + 0xcaca458fU, 0x82829d1fU, 0xc9c94089U, 0x7d7d87faU, + 0xfafa15efU, 0x5959ebb2U, 0x4747c98eU, 0xf0f00bfbU, + 0xadadec41U, 0xd4d467b3U, 0xa2a2fd5fU, 0xafafea45U, + 0x9c9cbf23U, 0xa4a4f753U, 0x727296e4U, 0xc0c05b9bU, + 0xb7b7c275U, 0xfdfd1ce1U, 0x9393ae3dU, 0x26266a4cU, + 0x36365a6cU, 0x3f3f417eU, 0xf7f702f5U, 0xcccc4f83U, + 0x34345c68U, 0xa5a5f451U, 0xe5e534d1U, 0xf1f108f9U, + 0x717193e2U, 0xd8d873abU, 0x31315362U, 0x15153f2aU, + 0x04040c08U, 0xc7c75295U, 0x23236546U, 0xc3c35e9dU, + 0x18182830U, 0x9696a137U, 0x05050f0aU, 0x9a9ab52fU, + 0x0707090eU, 0x12123624U, 0x80809b1bU, 0xe2e23ddfU, + 0xebeb26cdU, 0x2727694eU, 0xb2b2cd7fU, 0x75759feaU, + 0x09091b12U, 0x83839e1dU, 0x2c2c7458U, 0x1a1a2e34U, + 0x1b1b2d36U, 0x6e6eb2dcU, 0x5a5aeeb4U, 0xa0a0fb5bU, + 0x5252f6a4U, 0x3b3b4d76U, 0xd6d661b7U, 0xb3b3ce7dU, + 0x29297b52U, 0xe3e33eddU, 0x2f2f715eU, 0x84849713U, + 0x5353f5a6U, 0xd1d168b9U, 0x00000000U, 0xeded2cc1U, + 0x20206040U, 0xfcfc1fe3U, 0xb1b1c879U, 0x5b5bedb6U, + 0x6a6abed4U, 0xcbcb468dU, 0xbebed967U, 0x39394b72U, + 0x4a4ade94U, 0x4c4cd498U, 0x5858e8b0U, 0xcfcf4a85U, + 0xd0d06bbbU, 0xefef2ac5U, 0xaaaae54fU, 0xfbfb16edU, + 0x4343c586U, 0x4d4dd79aU, 0x33335566U, 0x85859411U, + 0x4545cf8aU, 0xf9f910e9U, 0x02020604U, 0x7f7f81feU, + 0x5050f0a0U, 0x3c3c4478U, 0x9f9fba25U, 0xa8a8e34bU, + 0x5151f3a2U, 0xa3a3fe5dU, 0x4040c080U, 0x8f8f8a05U, + 0x9292ad3fU, 0x9d9dbc21U, 0x38384870U, 0xf5f504f1U, + 0xbcbcdf63U, 0xb6b6c177U, 0xdada75afU, 0x21216342U, + 0x10103020U, 0xffff1ae5U, 0xf3f30efdU, 0xd2d26dbfU, + 0xcdcd4c81U, 0x0c0c1418U, 0x13133526U, 0xecec2fc3U, + 0x5f5fe1beU, 0x9797a235U, 0x4444cc88U, 0x1717392eU, + 0xc4c45793U, 0xa7a7f255U, 0x7e7e82fcU, 0x3d3d477aU, + 0x6464acc8U, 0x5d5de7baU, 0x19192b32U, 0x737395e6U, + 0x6060a0c0U, 0x81819819U, 0x4f4fd19eU, 0xdcdc7fa3U, + 0x22226644U, 0x2a2a7e54U, 0x9090ab3bU, 0x8888830bU, + 0x4646ca8cU, 0xeeee29c7U, 0xb8b8d36bU, 0x14143c28U, + 0xdede79a7U, 0x5e5ee2bcU, 0x0b0b1d16U, 0xdbdb76adU, + 0xe0e03bdbU, 0x32325664U, 0x3a3a4e74U, 0x0a0a1e14U, + 0x4949db92U, 0x06060a0cU, 0x24246c48U, 0x5c5ce4b8U, + 0xc2c25d9fU, 0xd3d36ebdU, 0xacacef43U, 0x6262a6c4U, + 0x9191a839U, 0x9595a431U, 0xe4e437d3U, 0x79798bf2U, + 0xe7e732d5U, 0xc8c8438bU, 0x3737596eU, 0x6d6db7daU, + 0x8d8d8c01U, 0xd5d564b1U, 0x4e4ed29cU, 0xa9a9e049U, + 0x6c6cb4d8U, 0x5656faacU, 0xf4f407f3U, 0xeaea25cfU, + 0x6565afcaU, 0x7a7a8ef4U, 0xaeaee947U, 0x08081810U, + 0xbabad56fU, 0x787888f0U, 0x25256f4aU, 0x2e2e725cU, + 0x1c1c2438U, 0xa6a6f157U, 0xb4b4c773U, 0xc6c65197U, + 0xe8e823cbU, 0xdddd7ca1U, 0x74749ce8U, 0x1f1f213eU, + 0x4b4bdd96U, 0xbdbddc61U, 0x8b8b860dU, 0x8a8a850fU, + 0x707090e0U, 0x3e3e427cU, 0xb5b5c471U, 0x6666aaccU, + 0x4848d890U, 0x03030506U, 0xf6f601f7U, 0x0e0e121cU, + 0x6161a3c2U, 0x35355f6aU, 0x5757f9aeU, 0xb9b9d069U, + 0x86869117U, 0xc1c15899U, 0x1d1d273aU, 0x9e9eb927U, + 0xe1e138d9U, 0xf8f813ebU, 0x9898b32bU, 0x11113322U, + 0x6969bbd2U, 0xd9d970a9U, 0x8e8e8907U, 0x9494a733U, + 0x9b9bb62dU, 0x1e1e223cU, 0x87879215U, 0xe9e920c9U, + 0xcece4987U, 0x5555ffaaU, 0x28287850U, 0xdfdf7aa5U, + 0x8c8c8f03U, 0xa1a1f859U, 0x89898009U, 0x0d0d171aU, + 0xbfbfda65U, 0xe6e631d7U, 0x4242c684U, 0x6868b8d0U, + 0x4141c382U, 0x9999b029U, 0x2d2d775aU, 0x0f0f111eU, + 0xb0b0cb7bU, 0x5454fca8U, 0xbbbbd66dU, 0x16163a2cU, +}; +static const u32 Te4[256] = { + 0x63636363U, 0x7c7c7c7cU, 0x77777777U, 0x7b7b7b7bU, + 0xf2f2f2f2U, 0x6b6b6b6bU, 0x6f6f6f6fU, 0xc5c5c5c5U, + 0x30303030U, 0x01010101U, 0x67676767U, 0x2b2b2b2bU, + 0xfefefefeU, 0xd7d7d7d7U, 0xababababU, 0x76767676U, + 0xcacacacaU, 0x82828282U, 0xc9c9c9c9U, 0x7d7d7d7dU, + 0xfafafafaU, 0x59595959U, 0x47474747U, 0xf0f0f0f0U, + 0xadadadadU, 0xd4d4d4d4U, 0xa2a2a2a2U, 0xafafafafU, + 0x9c9c9c9cU, 0xa4a4a4a4U, 0x72727272U, 0xc0c0c0c0U, + 0xb7b7b7b7U, 0xfdfdfdfdU, 0x93939393U, 0x26262626U, + 0x36363636U, 0x3f3f3f3fU, 0xf7f7f7f7U, 0xccccccccU, + 0x34343434U, 0xa5a5a5a5U, 0xe5e5e5e5U, 0xf1f1f1f1U, + 0x71717171U, 0xd8d8d8d8U, 0x31313131U, 0x15151515U, + 0x04040404U, 0xc7c7c7c7U, 0x23232323U, 0xc3c3c3c3U, + 0x18181818U, 0x96969696U, 0x05050505U, 0x9a9a9a9aU, + 0x07070707U, 0x12121212U, 0x80808080U, 0xe2e2e2e2U, + 0xebebebebU, 0x27272727U, 0xb2b2b2b2U, 0x75757575U, + 0x09090909U, 0x83838383U, 0x2c2c2c2cU, 0x1a1a1a1aU, + 0x1b1b1b1bU, 0x6e6e6e6eU, 0x5a5a5a5aU, 0xa0a0a0a0U, + 0x52525252U, 0x3b3b3b3bU, 0xd6d6d6d6U, 0xb3b3b3b3U, + 0x29292929U, 0xe3e3e3e3U, 0x2f2f2f2fU, 0x84848484U, + 0x53535353U, 0xd1d1d1d1U, 0x00000000U, 0xededededU, + 0x20202020U, 0xfcfcfcfcU, 0xb1b1b1b1U, 0x5b5b5b5bU, + 0x6a6a6a6aU, 0xcbcbcbcbU, 0xbebebebeU, 0x39393939U, + 0x4a4a4a4aU, 0x4c4c4c4cU, 0x58585858U, 0xcfcfcfcfU, + 0xd0d0d0d0U, 0xefefefefU, 0xaaaaaaaaU, 0xfbfbfbfbU, + 0x43434343U, 0x4d4d4d4dU, 0x33333333U, 0x85858585U, + 0x45454545U, 0xf9f9f9f9U, 0x02020202U, 0x7f7f7f7fU, + 0x50505050U, 0x3c3c3c3cU, 0x9f9f9f9fU, 0xa8a8a8a8U, + 0x51515151U, 0xa3a3a3a3U, 0x40404040U, 0x8f8f8f8fU, + 0x92929292U, 0x9d9d9d9dU, 0x38383838U, 0xf5f5f5f5U, + 0xbcbcbcbcU, 0xb6b6b6b6U, 0xdadadadaU, 0x21212121U, + 0x10101010U, 0xffffffffU, 0xf3f3f3f3U, 0xd2d2d2d2U, + 0xcdcdcdcdU, 0x0c0c0c0cU, 0x13131313U, 0xececececU, + 0x5f5f5f5fU, 0x97979797U, 0x44444444U, 0x17171717U, + 0xc4c4c4c4U, 0xa7a7a7a7U, 0x7e7e7e7eU, 0x3d3d3d3dU, + 0x64646464U, 0x5d5d5d5dU, 0x19191919U, 0x73737373U, + 0x60606060U, 0x81818181U, 0x4f4f4f4fU, 0xdcdcdcdcU, + 0x22222222U, 0x2a2a2a2aU, 0x90909090U, 0x88888888U, + 0x46464646U, 0xeeeeeeeeU, 0xb8b8b8b8U, 0x14141414U, + 0xdedededeU, 0x5e5e5e5eU, 0x0b0b0b0bU, 0xdbdbdbdbU, + 0xe0e0e0e0U, 0x32323232U, 0x3a3a3a3aU, 0x0a0a0a0aU, + 0x49494949U, 0x06060606U, 0x24242424U, 0x5c5c5c5cU, + 0xc2c2c2c2U, 0xd3d3d3d3U, 0xacacacacU, 0x62626262U, + 0x91919191U, 0x95959595U, 0xe4e4e4e4U, 0x79797979U, + 0xe7e7e7e7U, 0xc8c8c8c8U, 0x37373737U, 0x6d6d6d6dU, + 0x8d8d8d8dU, 0xd5d5d5d5U, 0x4e4e4e4eU, 0xa9a9a9a9U, + 0x6c6c6c6cU, 0x56565656U, 0xf4f4f4f4U, 0xeaeaeaeaU, + 0x65656565U, 0x7a7a7a7aU, 0xaeaeaeaeU, 0x08080808U, + 0xbabababaU, 0x78787878U, 0x25252525U, 0x2e2e2e2eU, + 0x1c1c1c1cU, 0xa6a6a6a6U, 0xb4b4b4b4U, 0xc6c6c6c6U, + 0xe8e8e8e8U, 0xddddddddU, 0x74747474U, 0x1f1f1f1fU, + 0x4b4b4b4bU, 0xbdbdbdbdU, 0x8b8b8b8bU, 0x8a8a8a8aU, + 0x70707070U, 0x3e3e3e3eU, 0xb5b5b5b5U, 0x66666666U, + 0x48484848U, 0x03030303U, 0xf6f6f6f6U, 0x0e0e0e0eU, + 0x61616161U, 0x35353535U, 0x57575757U, 0xb9b9b9b9U, + 0x86868686U, 0xc1c1c1c1U, 0x1d1d1d1dU, 0x9e9e9e9eU, + 0xe1e1e1e1U, 0xf8f8f8f8U, 0x98989898U, 0x11111111U, + 0x69696969U, 0xd9d9d9d9U, 0x8e8e8e8eU, 0x94949494U, + 0x9b9b9b9bU, 0x1e1e1e1eU, 0x87878787U, 0xe9e9e9e9U, + 0xcecececeU, 0x55555555U, 0x28282828U, 0xdfdfdfdfU, + 0x8c8c8c8cU, 0xa1a1a1a1U, 0x89898989U, 0x0d0d0d0dU, + 0xbfbfbfbfU, 0xe6e6e6e6U, 0x42424242U, 0x68686868U, + 0x41414141U, 0x99999999U, 0x2d2d2d2dU, 0x0f0f0f0fU, + 0xb0b0b0b0U, 0x54545454U, 0xbbbbbbbbU, 0x16161616U, +}; +static const u32 Td0[256] = { + 0x51f4a750U, 0x7e416553U, 0x1a17a4c3U, 0x3a275e96U, + 0x3bab6bcbU, 0x1f9d45f1U, 0xacfa58abU, 0x4be30393U, + 0x2030fa55U, 0xad766df6U, 0x88cc7691U, 0xf5024c25U, + 0x4fe5d7fcU, 0xc52acbd7U, 0x26354480U, 0xb562a38fU, + 0xdeb15a49U, 0x25ba1b67U, 0x45ea0e98U, 0x5dfec0e1U, + 0xc32f7502U, 0x814cf012U, 0x8d4697a3U, 0x6bd3f9c6U, + 0x038f5fe7U, 0x15929c95U, 0xbf6d7aebU, 0x955259daU, + 0xd4be832dU, 0x587421d3U, 0x49e06929U, 0x8ec9c844U, + 0x75c2896aU, 0xf48e7978U, 0x99583e6bU, 0x27b971ddU, + 0xbee14fb6U, 0xf088ad17U, 0xc920ac66U, 0x7dce3ab4U, + 0x63df4a18U, 0xe51a3182U, 0x97513360U, 0x62537f45U, + 0xb16477e0U, 0xbb6bae84U, 0xfe81a01cU, 0xf9082b94U, + 0x70486858U, 0x8f45fd19U, 0x94de6c87U, 0x527bf8b7U, + 0xab73d323U, 0x724b02e2U, 0xe31f8f57U, 0x6655ab2aU, + 0xb2eb2807U, 0x2fb5c203U, 0x86c57b9aU, 0xd33708a5U, + 0x302887f2U, 0x23bfa5b2U, 0x02036abaU, 0xed16825cU, + 0x8acf1c2bU, 0xa779b492U, 0xf307f2f0U, 0x4e69e2a1U, + 0x65daf4cdU, 0x0605bed5U, 0xd134621fU, 0xc4a6fe8aU, + 0x342e539dU, 0xa2f355a0U, 0x058ae132U, 0xa4f6eb75U, + 0x0b83ec39U, 0x4060efaaU, 0x5e719f06U, 0xbd6e1051U, + 0x3e218af9U, 0x96dd063dU, 0xdd3e05aeU, 0x4de6bd46U, + 0x91548db5U, 0x71c45d05U, 0x0406d46fU, 0x605015ffU, + 0x1998fb24U, 0xd6bde997U, 0x894043ccU, 0x67d99e77U, + 0xb0e842bdU, 0x07898b88U, 0xe7195b38U, 0x79c8eedbU, + 0xa17c0a47U, 0x7c420fe9U, 0xf8841ec9U, 0x00000000U, + 0x09808683U, 0x322bed48U, 0x1e1170acU, 0x6c5a724eU, + 0xfd0efffbU, 0x0f853856U, 0x3daed51eU, 0x362d3927U, + 0x0a0fd964U, 0x685ca621U, 0x9b5b54d1U, 0x24362e3aU, + 0x0c0a67b1U, 0x9357e70fU, 0xb4ee96d2U, 0x1b9b919eU, + 0x80c0c54fU, 0x61dc20a2U, 0x5a774b69U, 0x1c121a16U, + 0xe293ba0aU, 0xc0a02ae5U, 0x3c22e043U, 0x121b171dU, + 0x0e090d0bU, 0xf28bc7adU, 0x2db6a8b9U, 0x141ea9c8U, + 0x57f11985U, 0xaf75074cU, 0xee99ddbbU, 0xa37f60fdU, + 0xf701269fU, 0x5c72f5bcU, 0x44663bc5U, 0x5bfb7e34U, + 0x8b432976U, 0xcb23c6dcU, 0xb6edfc68U, 0xb8e4f163U, + 0xd731dccaU, 0x42638510U, 0x13972240U, 0x84c61120U, + 0x854a247dU, 0xd2bb3df8U, 0xaef93211U, 0xc729a16dU, + 0x1d9e2f4bU, 0xdcb230f3U, 0x0d8652ecU, 0x77c1e3d0U, + 0x2bb3166cU, 0xa970b999U, 0x119448faU, 0x47e96422U, + 0xa8fc8cc4U, 0xa0f03f1aU, 0x567d2cd8U, 0x223390efU, + 0x87494ec7U, 0xd938d1c1U, 0x8ccaa2feU, 0x98d40b36U, + 0xa6f581cfU, 0xa57ade28U, 0xdab78e26U, 0x3fadbfa4U, + 0x2c3a9de4U, 0x5078920dU, 0x6a5fcc9bU, 0x547e4662U, + 0xf68d13c2U, 0x90d8b8e8U, 0x2e39f75eU, 0x82c3aff5U, + 0x9f5d80beU, 0x69d0937cU, 0x6fd52da9U, 0xcf2512b3U, + 0xc8ac993bU, 0x10187da7U, 0xe89c636eU, 0xdb3bbb7bU, + 0xcd267809U, 0x6e5918f4U, 0xec9ab701U, 0x834f9aa8U, + 0xe6956e65U, 0xaaffe67eU, 0x21bccf08U, 0xef15e8e6U, + 0xbae79bd9U, 0x4a6f36ceU, 0xea9f09d4U, 0x29b07cd6U, + 0x31a4b2afU, 0x2a3f2331U, 0xc6a59430U, 0x35a266c0U, + 0x744ebc37U, 0xfc82caa6U, 0xe090d0b0U, 0x33a7d815U, + 0xf104984aU, 0x41ecdaf7U, 0x7fcd500eU, 0x1791f62fU, + 0x764dd68dU, 0x43efb04dU, 0xccaa4d54U, 0xe49604dfU, + 0x9ed1b5e3U, 0x4c6a881bU, 0xc12c1fb8U, 0x4665517fU, + 0x9d5eea04U, 0x018c355dU, 0xfa877473U, 0xfb0b412eU, + 0xb3671d5aU, 0x92dbd252U, 0xe9105633U, 0x6dd64713U, + 0x9ad7618cU, 0x37a10c7aU, 0x59f8148eU, 0xeb133c89U, + 0xcea927eeU, 0xb761c935U, 0xe11ce5edU, 0x7a47b13cU, + 0x9cd2df59U, 0x55f2733fU, 0x1814ce79U, 0x73c737bfU, + 0x53f7cdeaU, 0x5ffdaa5bU, 0xdf3d6f14U, 0x7844db86U, + 0xcaaff381U, 0xb968c43eU, 0x3824342cU, 0xc2a3405fU, + 0x161dc372U, 0xbce2250cU, 0x283c498bU, 0xff0d9541U, + 0x39a80171U, 0x080cb3deU, 0xd8b4e49cU, 0x6456c190U, + 0x7bcb8461U, 0xd532b670U, 0x486c5c74U, 0xd0b85742U, +}; +static const u32 Td1[256] = { + 0x5051f4a7U, 0x537e4165U, 0xc31a17a4U, 0x963a275eU, + 0xcb3bab6bU, 0xf11f9d45U, 0xabacfa58U, 0x934be303U, + 0x552030faU, 0xf6ad766dU, 0x9188cc76U, 0x25f5024cU, + 0xfc4fe5d7U, 0xd7c52acbU, 0x80263544U, 0x8fb562a3U, + 0x49deb15aU, 0x6725ba1bU, 0x9845ea0eU, 0xe15dfec0U, + 0x02c32f75U, 0x12814cf0U, 0xa38d4697U, 0xc66bd3f9U, + 0xe7038f5fU, 0x9515929cU, 0xebbf6d7aU, 0xda955259U, + 0x2dd4be83U, 0xd3587421U, 0x2949e069U, 0x448ec9c8U, + 0x6a75c289U, 0x78f48e79U, 0x6b99583eU, 0xdd27b971U, + 0xb6bee14fU, 0x17f088adU, 0x66c920acU, 0xb47dce3aU, + 0x1863df4aU, 0x82e51a31U, 0x60975133U, 0x4562537fU, + 0xe0b16477U, 0x84bb6baeU, 0x1cfe81a0U, 0x94f9082bU, + 0x58704868U, 0x198f45fdU, 0x8794de6cU, 0xb7527bf8U, + 0x23ab73d3U, 0xe2724b02U, 0x57e31f8fU, 0x2a6655abU, + 0x07b2eb28U, 0x032fb5c2U, 0x9a86c57bU, 0xa5d33708U, + 0xf2302887U, 0xb223bfa5U, 0xba02036aU, 0x5ced1682U, + 0x2b8acf1cU, 0x92a779b4U, 0xf0f307f2U, 0xa14e69e2U, + 0xcd65daf4U, 0xd50605beU, 0x1fd13462U, 0x8ac4a6feU, + 0x9d342e53U, 0xa0a2f355U, 0x32058ae1U, 0x75a4f6ebU, + 0x390b83ecU, 0xaa4060efU, 0x065e719fU, 0x51bd6e10U, + 0xf93e218aU, 0x3d96dd06U, 0xaedd3e05U, 0x464de6bdU, + 0xb591548dU, 0x0571c45dU, 0x6f0406d4U, 0xff605015U, + 0x241998fbU, 0x97d6bde9U, 0xcc894043U, 0x7767d99eU, + 0xbdb0e842U, 0x8807898bU, 0x38e7195bU, 0xdb79c8eeU, + 0x47a17c0aU, 0xe97c420fU, 0xc9f8841eU, 0x00000000U, + 0x83098086U, 0x48322bedU, 0xac1e1170U, 0x4e6c5a72U, + 0xfbfd0effU, 0x560f8538U, 0x1e3daed5U, 0x27362d39U, + 0x640a0fd9U, 0x21685ca6U, 0xd19b5b54U, 0x3a24362eU, + 0xb10c0a67U, 0x0f9357e7U, 0xd2b4ee96U, 0x9e1b9b91U, + 0x4f80c0c5U, 0xa261dc20U, 0x695a774bU, 0x161c121aU, + 0x0ae293baU, 0xe5c0a02aU, 0x433c22e0U, 0x1d121b17U, + 0x0b0e090dU, 0xadf28bc7U, 0xb92db6a8U, 0xc8141ea9U, + 0x8557f119U, 0x4caf7507U, 0xbbee99ddU, 0xfda37f60U, + 0x9ff70126U, 0xbc5c72f5U, 0xc544663bU, 0x345bfb7eU, + 0x768b4329U, 0xdccb23c6U, 0x68b6edfcU, 0x63b8e4f1U, + 0xcad731dcU, 0x10426385U, 0x40139722U, 0x2084c611U, + 0x7d854a24U, 0xf8d2bb3dU, 0x11aef932U, 0x6dc729a1U, + 0x4b1d9e2fU, 0xf3dcb230U, 0xec0d8652U, 0xd077c1e3U, + 0x6c2bb316U, 0x99a970b9U, 0xfa119448U, 0x2247e964U, + 0xc4a8fc8cU, 0x1aa0f03fU, 0xd8567d2cU, 0xef223390U, + 0xc787494eU, 0xc1d938d1U, 0xfe8ccaa2U, 0x3698d40bU, + 0xcfa6f581U, 0x28a57adeU, 0x26dab78eU, 0xa43fadbfU, + 0xe42c3a9dU, 0x0d507892U, 0x9b6a5fccU, 0x62547e46U, + 0xc2f68d13U, 0xe890d8b8U, 0x5e2e39f7U, 0xf582c3afU, + 0xbe9f5d80U, 0x7c69d093U, 0xa96fd52dU, 0xb3cf2512U, + 0x3bc8ac99U, 0xa710187dU, 0x6ee89c63U, 0x7bdb3bbbU, + 0x09cd2678U, 0xf46e5918U, 0x01ec9ab7U, 0xa8834f9aU, + 0x65e6956eU, 0x7eaaffe6U, 0x0821bccfU, 0xe6ef15e8U, + 0xd9bae79bU, 0xce4a6f36U, 0xd4ea9f09U, 0xd629b07cU, + 0xaf31a4b2U, 0x312a3f23U, 0x30c6a594U, 0xc035a266U, + 0x37744ebcU, 0xa6fc82caU, 0xb0e090d0U, 0x1533a7d8U, + 0x4af10498U, 0xf741ecdaU, 0x0e7fcd50U, 0x2f1791f6U, + 0x8d764dd6U, 0x4d43efb0U, 0x54ccaa4dU, 0xdfe49604U, + 0xe39ed1b5U, 0x1b4c6a88U, 0xb8c12c1fU, 0x7f466551U, + 0x049d5eeaU, 0x5d018c35U, 0x73fa8774U, 0x2efb0b41U, + 0x5ab3671dU, 0x5292dbd2U, 0x33e91056U, 0x136dd647U, + 0x8c9ad761U, 0x7a37a10cU, 0x8e59f814U, 0x89eb133cU, + 0xeecea927U, 0x35b761c9U, 0xede11ce5U, 0x3c7a47b1U, + 0x599cd2dfU, 0x3f55f273U, 0x791814ceU, 0xbf73c737U, + 0xea53f7cdU, 0x5b5ffdaaU, 0x14df3d6fU, 0x867844dbU, + 0x81caaff3U, 0x3eb968c4U, 0x2c382434U, 0x5fc2a340U, + 0x72161dc3U, 0x0cbce225U, 0x8b283c49U, 0x41ff0d95U, + 0x7139a801U, 0xde080cb3U, 0x9cd8b4e4U, 0x906456c1U, + 0x617bcb84U, 0x70d532b6U, 0x74486c5cU, 0x42d0b857U, +}; +static const u32 Td2[256] = { + 0xa75051f4U, 0x65537e41U, 0xa4c31a17U, 0x5e963a27U, + 0x6bcb3babU, 0x45f11f9dU, 0x58abacfaU, 0x03934be3U, + 0xfa552030U, 0x6df6ad76U, 0x769188ccU, 0x4c25f502U, + 0xd7fc4fe5U, 0xcbd7c52aU, 0x44802635U, 0xa38fb562U, + 0x5a49deb1U, 0x1b6725baU, 0x0e9845eaU, 0xc0e15dfeU, + 0x7502c32fU, 0xf012814cU, 0x97a38d46U, 0xf9c66bd3U, + 0x5fe7038fU, 0x9c951592U, 0x7aebbf6dU, 0x59da9552U, + 0x832dd4beU, 0x21d35874U, 0x692949e0U, 0xc8448ec9U, + 0x896a75c2U, 0x7978f48eU, 0x3e6b9958U, 0x71dd27b9U, + 0x4fb6bee1U, 0xad17f088U, 0xac66c920U, 0x3ab47dceU, + 0x4a1863dfU, 0x3182e51aU, 0x33609751U, 0x7f456253U, + 0x77e0b164U, 0xae84bb6bU, 0xa01cfe81U, 0x2b94f908U, + 0x68587048U, 0xfd198f45U, 0x6c8794deU, 0xf8b7527bU, + 0xd323ab73U, 0x02e2724bU, 0x8f57e31fU, 0xab2a6655U, + 0x2807b2ebU, 0xc2032fb5U, 0x7b9a86c5U, 0x08a5d337U, + 0x87f23028U, 0xa5b223bfU, 0x6aba0203U, 0x825ced16U, + 0x1c2b8acfU, 0xb492a779U, 0xf2f0f307U, 0xe2a14e69U, + 0xf4cd65daU, 0xbed50605U, 0x621fd134U, 0xfe8ac4a6U, + 0x539d342eU, 0x55a0a2f3U, 0xe132058aU, 0xeb75a4f6U, + 0xec390b83U, 0xefaa4060U, 0x9f065e71U, 0x1051bd6eU, + + 0x8af93e21U, 0x063d96ddU, 0x05aedd3eU, 0xbd464de6U, + 0x8db59154U, 0x5d0571c4U, 0xd46f0406U, 0x15ff6050U, + 0xfb241998U, 0xe997d6bdU, 0x43cc8940U, 0x9e7767d9U, + 0x42bdb0e8U, 0x8b880789U, 0x5b38e719U, 0xeedb79c8U, + 0x0a47a17cU, 0x0fe97c42U, 0x1ec9f884U, 0x00000000U, + 0x86830980U, 0xed48322bU, 0x70ac1e11U, 0x724e6c5aU, + 0xfffbfd0eU, 0x38560f85U, 0xd51e3daeU, 0x3927362dU, + 0xd9640a0fU, 0xa621685cU, 0x54d19b5bU, 0x2e3a2436U, + 0x67b10c0aU, 0xe70f9357U, 0x96d2b4eeU, 0x919e1b9bU, + 0xc54f80c0U, 0x20a261dcU, 0x4b695a77U, 0x1a161c12U, + 0xba0ae293U, 0x2ae5c0a0U, 0xe0433c22U, 0x171d121bU, + 0x0d0b0e09U, 0xc7adf28bU, 0xa8b92db6U, 0xa9c8141eU, + 0x198557f1U, 0x074caf75U, 0xddbbee99U, 0x60fda37fU, + 0x269ff701U, 0xf5bc5c72U, 0x3bc54466U, 0x7e345bfbU, + 0x29768b43U, 0xc6dccb23U, 0xfc68b6edU, 0xf163b8e4U, + 0xdccad731U, 0x85104263U, 0x22401397U, 0x112084c6U, + 0x247d854aU, 0x3df8d2bbU, 0x3211aef9U, 0xa16dc729U, + 0x2f4b1d9eU, 0x30f3dcb2U, 0x52ec0d86U, 0xe3d077c1U, + 0x166c2bb3U, 0xb999a970U, 0x48fa1194U, 0x642247e9U, + 0x8cc4a8fcU, 0x3f1aa0f0U, 0x2cd8567dU, 0x90ef2233U, + 0x4ec78749U, 0xd1c1d938U, 0xa2fe8ccaU, 0x0b3698d4U, + 0x81cfa6f5U, 0xde28a57aU, 0x8e26dab7U, 0xbfa43fadU, + 0x9de42c3aU, 0x920d5078U, 0xcc9b6a5fU, 0x4662547eU, + 0x13c2f68dU, 0xb8e890d8U, 0xf75e2e39U, 0xaff582c3U, + 0x80be9f5dU, 0x937c69d0U, 0x2da96fd5U, 0x12b3cf25U, + 0x993bc8acU, 0x7da71018U, 0x636ee89cU, 0xbb7bdb3bU, + 0x7809cd26U, 0x18f46e59U, 0xb701ec9aU, 0x9aa8834fU, + 0x6e65e695U, 0xe67eaaffU, 0xcf0821bcU, 0xe8e6ef15U, + 0x9bd9bae7U, 0x36ce4a6fU, 0x09d4ea9fU, 0x7cd629b0U, + 0xb2af31a4U, 0x23312a3fU, 0x9430c6a5U, 0x66c035a2U, + 0xbc37744eU, 0xcaa6fc82U, 0xd0b0e090U, 0xd81533a7U, + 0x984af104U, 0xdaf741ecU, 0x500e7fcdU, 0xf62f1791U, + 0xd68d764dU, 0xb04d43efU, 0x4d54ccaaU, 0x04dfe496U, + 0xb5e39ed1U, 0x881b4c6aU, 0x1fb8c12cU, 0x517f4665U, + 0xea049d5eU, 0x355d018cU, 0x7473fa87U, 0x412efb0bU, + 0x1d5ab367U, 0xd25292dbU, 0x5633e910U, 0x47136dd6U, + 0x618c9ad7U, 0x0c7a37a1U, 0x148e59f8U, 0x3c89eb13U, + 0x27eecea9U, 0xc935b761U, 0xe5ede11cU, 0xb13c7a47U, + 0xdf599cd2U, 0x733f55f2U, 0xce791814U, 0x37bf73c7U, + 0xcdea53f7U, 0xaa5b5ffdU, 0x6f14df3dU, 0xdb867844U, + 0xf381caafU, 0xc43eb968U, 0x342c3824U, 0x405fc2a3U, + 0xc372161dU, 0x250cbce2U, 0x498b283cU, 0x9541ff0dU, + 0x017139a8U, 0xb3de080cU, 0xe49cd8b4U, 0xc1906456U, + 0x84617bcbU, 0xb670d532U, 0x5c74486cU, 0x5742d0b8U, +}; +static const u32 Td3[256] = { + 0xf4a75051U, 0x4165537eU, 0x17a4c31aU, 0x275e963aU, + 0xab6bcb3bU, 0x9d45f11fU, 0xfa58abacU, 0xe303934bU, + 0x30fa5520U, 0x766df6adU, 0xcc769188U, 0x024c25f5U, + 0xe5d7fc4fU, 0x2acbd7c5U, 0x35448026U, 0x62a38fb5U, + 0xb15a49deU, 0xba1b6725U, 0xea0e9845U, 0xfec0e15dU, + 0x2f7502c3U, 0x4cf01281U, 0x4697a38dU, 0xd3f9c66bU, + 0x8f5fe703U, 0x929c9515U, 0x6d7aebbfU, 0x5259da95U, + 0xbe832dd4U, 0x7421d358U, 0xe0692949U, 0xc9c8448eU, + 0xc2896a75U, 0x8e7978f4U, 0x583e6b99U, 0xb971dd27U, + 0xe14fb6beU, 0x88ad17f0U, 0x20ac66c9U, 0xce3ab47dU, + 0xdf4a1863U, 0x1a3182e5U, 0x51336097U, 0x537f4562U, + 0x6477e0b1U, 0x6bae84bbU, 0x81a01cfeU, 0x082b94f9U, + 0x48685870U, 0x45fd198fU, 0xde6c8794U, 0x7bf8b752U, + 0x73d323abU, 0x4b02e272U, 0x1f8f57e3U, 0x55ab2a66U, + 0xeb2807b2U, 0xb5c2032fU, 0xc57b9a86U, 0x3708a5d3U, + 0x2887f230U, 0xbfa5b223U, 0x036aba02U, 0x16825cedU, + 0xcf1c2b8aU, 0x79b492a7U, 0x07f2f0f3U, 0x69e2a14eU, + 0xdaf4cd65U, 0x05bed506U, 0x34621fd1U, 0xa6fe8ac4U, + 0x2e539d34U, 0xf355a0a2U, 0x8ae13205U, 0xf6eb75a4U, + 0x83ec390bU, 0x60efaa40U, 0x719f065eU, 0x6e1051bdU, + 0x218af93eU, 0xdd063d96U, 0x3e05aeddU, 0xe6bd464dU, + 0x548db591U, 0xc45d0571U, 0x06d46f04U, 0x5015ff60U, + 0x98fb2419U, 0xbde997d6U, 0x4043cc89U, 0xd99e7767U, + 0xe842bdb0U, 0x898b8807U, 0x195b38e7U, 0xc8eedb79U, + 0x7c0a47a1U, 0x420fe97cU, 0x841ec9f8U, 0x00000000U, + 0x80868309U, 0x2bed4832U, 0x1170ac1eU, 0x5a724e6cU, + 0x0efffbfdU, 0x8538560fU, 0xaed51e3dU, 0x2d392736U, + 0x0fd9640aU, 0x5ca62168U, 0x5b54d19bU, 0x362e3a24U, + 0x0a67b10cU, 0x57e70f93U, 0xee96d2b4U, 0x9b919e1bU, + 0xc0c54f80U, 0xdc20a261U, 0x774b695aU, 0x121a161cU, + 0x93ba0ae2U, 0xa02ae5c0U, 0x22e0433cU, 0x1b171d12U, + 0x090d0b0eU, 0x8bc7adf2U, 0xb6a8b92dU, 0x1ea9c814U, + 0xf1198557U, 0x75074cafU, 0x99ddbbeeU, 0x7f60fda3U, + 0x01269ff7U, 0x72f5bc5cU, 0x663bc544U, 0xfb7e345bU, + 0x4329768bU, 0x23c6dccbU, 0xedfc68b6U, 0xe4f163b8U, + 0x31dccad7U, 0x63851042U, 0x97224013U, 0xc6112084U, + 0x4a247d85U, 0xbb3df8d2U, 0xf93211aeU, 0x29a16dc7U, + 0x9e2f4b1dU, 0xb230f3dcU, 0x8652ec0dU, 0xc1e3d077U, + 0xb3166c2bU, 0x70b999a9U, 0x9448fa11U, 0xe9642247U, + 0xfc8cc4a8U, 0xf03f1aa0U, 0x7d2cd856U, 0x3390ef22U, + 0x494ec787U, 0x38d1c1d9U, 0xcaa2fe8cU, 0xd40b3698U, + 0xf581cfa6U, 0x7ade28a5U, 0xb78e26daU, 0xadbfa43fU, + 0x3a9de42cU, 0x78920d50U, 0x5fcc9b6aU, 0x7e466254U, + 0x8d13c2f6U, 0xd8b8e890U, 0x39f75e2eU, 0xc3aff582U, + 0x5d80be9fU, 0xd0937c69U, 0xd52da96fU, 0x2512b3cfU, + 0xac993bc8U, 0x187da710U, 0x9c636ee8U, 0x3bbb7bdbU, + 0x267809cdU, 0x5918f46eU, 0x9ab701ecU, 0x4f9aa883U, + 0x956e65e6U, 0xffe67eaaU, 0xbccf0821U, 0x15e8e6efU, + 0xe79bd9baU, 0x6f36ce4aU, 0x9f09d4eaU, 0xb07cd629U, + 0xa4b2af31U, 0x3f23312aU, 0xa59430c6U, 0xa266c035U, + 0x4ebc3774U, 0x82caa6fcU, 0x90d0b0e0U, 0xa7d81533U, + 0x04984af1U, 0xecdaf741U, 0xcd500e7fU, 0x91f62f17U, + 0x4dd68d76U, 0xefb04d43U, 0xaa4d54ccU, 0x9604dfe4U, + 0xd1b5e39eU, 0x6a881b4cU, 0x2c1fb8c1U, 0x65517f46U, + 0x5eea049dU, 0x8c355d01U, 0x877473faU, 0x0b412efbU, + 0x671d5ab3U, 0xdbd25292U, 0x105633e9U, 0xd647136dU, + 0xd7618c9aU, 0xa10c7a37U, 0xf8148e59U, 0x133c89ebU, + 0xa927eeceU, 0x61c935b7U, 0x1ce5ede1U, 0x47b13c7aU, + 0xd2df599cU, 0xf2733f55U, 0x14ce7918U, 0xc737bf73U, + 0xf7cdea53U, 0xfdaa5b5fU, 0x3d6f14dfU, 0x44db8678U, + 0xaff381caU, 0x68c43eb9U, 0x24342c38U, 0xa3405fc2U, + 0x1dc37216U, 0xe2250cbcU, 0x3c498b28U, 0x0d9541ffU, + 0xa8017139U, 0x0cb3de08U, 0xb4e49cd8U, 0x56c19064U, + 0xcb84617bU, 0x32b670d5U, 0x6c5c7448U, 0xb85742d0U, +}; +static const u32 Td4[256] = { + 0x52525252U, 0x09090909U, 0x6a6a6a6aU, 0xd5d5d5d5U, + 0x30303030U, 0x36363636U, 0xa5a5a5a5U, 0x38383838U, + 0xbfbfbfbfU, 0x40404040U, 0xa3a3a3a3U, 0x9e9e9e9eU, + 0x81818181U, 0xf3f3f3f3U, 0xd7d7d7d7U, 0xfbfbfbfbU, + 0x7c7c7c7cU, 0xe3e3e3e3U, 0x39393939U, 0x82828282U, + 0x9b9b9b9bU, 0x2f2f2f2fU, 0xffffffffU, 0x87878787U, + 0x34343434U, 0x8e8e8e8eU, 0x43434343U, 0x44444444U, + 0xc4c4c4c4U, 0xdedededeU, 0xe9e9e9e9U, 0xcbcbcbcbU, + 0x54545454U, 0x7b7b7b7bU, 0x94949494U, 0x32323232U, + 0xa6a6a6a6U, 0xc2c2c2c2U, 0x23232323U, 0x3d3d3d3dU, + 0xeeeeeeeeU, 0x4c4c4c4cU, 0x95959595U, 0x0b0b0b0bU, + 0x42424242U, 0xfafafafaU, 0xc3c3c3c3U, 0x4e4e4e4eU, + 0x08080808U, 0x2e2e2e2eU, 0xa1a1a1a1U, 0x66666666U, + 0x28282828U, 0xd9d9d9d9U, 0x24242424U, 0xb2b2b2b2U, + 0x76767676U, 0x5b5b5b5bU, 0xa2a2a2a2U, 0x49494949U, + 0x6d6d6d6dU, 0x8b8b8b8bU, 0xd1d1d1d1U, 0x25252525U, + 0x72727272U, 0xf8f8f8f8U, 0xf6f6f6f6U, 0x64646464U, + 0x86868686U, 0x68686868U, 0x98989898U, 0x16161616U, + 0xd4d4d4d4U, 0xa4a4a4a4U, 0x5c5c5c5cU, 0xccccccccU, + 0x5d5d5d5dU, 0x65656565U, 0xb6b6b6b6U, 0x92929292U, + 0x6c6c6c6cU, 0x70707070U, 0x48484848U, 0x50505050U, + 0xfdfdfdfdU, 0xededededU, 0xb9b9b9b9U, 0xdadadadaU, + 0x5e5e5e5eU, 0x15151515U, 0x46464646U, 0x57575757U, + 0xa7a7a7a7U, 0x8d8d8d8dU, 0x9d9d9d9dU, 0x84848484U, + 0x90909090U, 0xd8d8d8d8U, 0xababababU, 0x00000000U, + 0x8c8c8c8cU, 0xbcbcbcbcU, 0xd3d3d3d3U, 0x0a0a0a0aU, + 0xf7f7f7f7U, 0xe4e4e4e4U, 0x58585858U, 0x05050505U, + 0xb8b8b8b8U, 0xb3b3b3b3U, 0x45454545U, 0x06060606U, + 0xd0d0d0d0U, 0x2c2c2c2cU, 0x1e1e1e1eU, 0x8f8f8f8fU, + 0xcacacacaU, 0x3f3f3f3fU, 0x0f0f0f0fU, 0x02020202U, + 0xc1c1c1c1U, 0xafafafafU, 0xbdbdbdbdU, 0x03030303U, + 0x01010101U, 0x13131313U, 0x8a8a8a8aU, 0x6b6b6b6bU, + 0x3a3a3a3aU, 0x91919191U, 0x11111111U, 0x41414141U, + 0x4f4f4f4fU, 0x67676767U, 0xdcdcdcdcU, 0xeaeaeaeaU, + 0x97979797U, 0xf2f2f2f2U, 0xcfcfcfcfU, 0xcecececeU, + 0xf0f0f0f0U, 0xb4b4b4b4U, 0xe6e6e6e6U, 0x73737373U, + 0x96969696U, 0xacacacacU, 0x74747474U, 0x22222222U, + 0xe7e7e7e7U, 0xadadadadU, 0x35353535U, 0x85858585U, + 0xe2e2e2e2U, 0xf9f9f9f9U, 0x37373737U, 0xe8e8e8e8U, + 0x1c1c1c1cU, 0x75757575U, 0xdfdfdfdfU, 0x6e6e6e6eU, + 0x47474747U, 0xf1f1f1f1U, 0x1a1a1a1aU, 0x71717171U, + 0x1d1d1d1dU, 0x29292929U, 0xc5c5c5c5U, 0x89898989U, + 0x6f6f6f6fU, 0xb7b7b7b7U, 0x62626262U, 0x0e0e0e0eU, + 0xaaaaaaaaU, 0x18181818U, 0xbebebebeU, 0x1b1b1b1bU, + 0xfcfcfcfcU, 0x56565656U, 0x3e3e3e3eU, 0x4b4b4b4bU, + 0xc6c6c6c6U, 0xd2d2d2d2U, 0x79797979U, 0x20202020U, + 0x9a9a9a9aU, 0xdbdbdbdbU, 0xc0c0c0c0U, 0xfefefefeU, + 0x78787878U, 0xcdcdcdcdU, 0x5a5a5a5aU, 0xf4f4f4f4U, + 0x1f1f1f1fU, 0xddddddddU, 0xa8a8a8a8U, 0x33333333U, + 0x88888888U, 0x07070707U, 0xc7c7c7c7U, 0x31313131U, + 0xb1b1b1b1U, 0x12121212U, 0x10101010U, 0x59595959U, + 0x27272727U, 0x80808080U, 0xececececU, 0x5f5f5f5fU, + 0x60606060U, 0x51515151U, 0x7f7f7f7fU, 0xa9a9a9a9U, + 0x19191919U, 0xb5b5b5b5U, 0x4a4a4a4aU, 0x0d0d0d0dU, + 0x2d2d2d2dU, 0xe5e5e5e5U, 0x7a7a7a7aU, 0x9f9f9f9fU, + 0x93939393U, 0xc9c9c9c9U, 0x9c9c9c9cU, 0xefefefefU, + 0xa0a0a0a0U, 0xe0e0e0e0U, 0x3b3b3b3bU, 0x4d4d4d4dU, + 0xaeaeaeaeU, 0x2a2a2a2aU, 0xf5f5f5f5U, 0xb0b0b0b0U, + 0xc8c8c8c8U, 0xebebebebU, 0xbbbbbbbbU, 0x3c3c3c3cU, + 0x83838383U, 0x53535353U, 0x99999999U, 0x61616161U, + 0x17171717U, 0x2b2b2b2bU, 0x04040404U, 0x7e7e7e7eU, + 0xbabababaU, 0x77777777U, 0xd6d6d6d6U, 0x26262626U, + 0xe1e1e1e1U, 0x69696969U, 0x14141414U, 0x63636363U, + 0x55555555U, 0x21212121U, 0x0c0c0c0cU, 0x7d7d7d7dU, +}; +static const u32 rcon[] = { + 0x01000000, 0x02000000, 0x04000000, 0x08000000, + 0x10000000, 0x20000000, 0x40000000, 0x80000000, + 0x1B000000, 0x36000000, /* for 128-bit blocks, Rijndael never uses more than 10 rcon values */ +}; + +#define SWAP(x) (_lrotl(x, 8) & 0x00ff00ff | _lrotr(x, 8) & 0xff00ff00) + +#ifdef _MSC_VER +#define GETU32(p) SWAP(*((u32 *)(p))) +#define PUTU32(ct, st) { *((u32 *)(ct)) = SWAP((st)); } +#else +#define GETU32(pt) (((u32)(pt)[0] << 24) ^ ((u32)(pt)[1] << 16) ^ ((u32)(pt)[2] << 8) ^ ((u32)(pt)[3])) +#define PUTU32(ct, st) { (ct)[0] = (u8)((st) >> 24); (ct)[1] = (u8)((st) >> 16); (ct)[2] = (u8)((st) >> 8); (ct)[3] = (u8)(st); } +#endif + +/** + * Expand the cipher key into the encryption key schedule. + * + * @return the number of rounds for the given cipher key size. + */ +int rijndaelKeySetupEnc(u32 rk[/*4*(Nr + 1)*/], const u8 cipherKey[], int keyBits) { + int i = 0; + u32 temp; + + rk[0] = GETU32(cipherKey ); + rk[1] = GETU32(cipherKey + 4); + rk[2] = GETU32(cipherKey + 8); + rk[3] = GETU32(cipherKey + 12); + if (keyBits == 128) { + for (;;) { + temp = rk[3]; + rk[4] = rk[0] ^ + (Te4[(temp >> 16) & 0xff] & 0xff000000) ^ + (Te4[(temp >> 8) & 0xff] & 0x00ff0000) ^ + (Te4[(temp ) & 0xff] & 0x0000ff00) ^ + (Te4[(temp >> 24) ] & 0x000000ff) ^ + rcon[i]; + rk[5] = rk[1] ^ rk[4]; + rk[6] = rk[2] ^ rk[5]; + rk[7] = rk[3] ^ rk[6]; + if (++i == 10) { + return 10; + } + rk += 4; + } + } + rk[4] = GETU32(cipherKey + 16); + rk[5] = GETU32(cipherKey + 20); + if (keyBits == 192) { + for (;;) { + temp = rk[ 5]; + rk[ 6] = rk[ 0] ^ + (Te4[(temp >> 16) & 0xff] & 0xff000000) ^ + (Te4[(temp >> 8) & 0xff] & 0x00ff0000) ^ + (Te4[(temp ) & 0xff] & 0x0000ff00) ^ + (Te4[(temp >> 24) ] & 0x000000ff) ^ + rcon[i]; + rk[ 7] = rk[ 1] ^ rk[ 6]; + rk[ 8] = rk[ 2] ^ rk[ 7]; + rk[ 9] = rk[ 3] ^ rk[ 8]; + if (++i == 8) { + return 12; + } + rk[10] = rk[ 4] ^ rk[ 9]; + rk[11] = rk[ 5] ^ rk[10]; + rk += 6; + } + } + rk[6] = GETU32(cipherKey + 24); + rk[7] = GETU32(cipherKey + 28); + if (keyBits == 256) { + for (;;) { + temp = rk[ 7]; + rk[ 8] = rk[ 0] ^ + (Te4[(temp >> 16) & 0xff] & 0xff000000) ^ + (Te4[(temp >> 8) & 0xff] & 0x00ff0000) ^ + (Te4[(temp ) & 0xff] & 0x0000ff00) ^ + (Te4[(temp >> 24) ] & 0x000000ff) ^ + rcon[i]; + rk[ 9] = rk[ 1] ^ rk[ 8]; + rk[10] = rk[ 2] ^ rk[ 9]; + rk[11] = rk[ 3] ^ rk[10]; + if (++i == 7) { + return 14; + } + temp = rk[11]; + rk[12] = rk[ 4] ^ + (Te4[(temp >> 24) ] & 0xff000000) ^ + (Te4[(temp >> 16) & 0xff] & 0x00ff0000) ^ + (Te4[(temp >> 8) & 0xff] & 0x0000ff00) ^ + (Te4[(temp ) & 0xff] & 0x000000ff); + rk[13] = rk[ 5] ^ rk[12]; + rk[14] = rk[ 6] ^ rk[13]; + rk[15] = rk[ 7] ^ rk[14]; + + rk += 8; + } + } + return 0; +} + +/** + * Expand the cipher key into the decryption key schedule. + * + * @return the number of rounds for the given cipher key size. + */ +int rijndaelKeySetupDec(u32 rk[/*4*(Nr + 1)*/], const u8 cipherKey[], int keyBits) { + int Nr, i, j; + u32 temp; + + /* expand the cipher key: */ + Nr = rijndaelKeySetupEnc(rk, cipherKey, keyBits); + /* invert the order of the round keys: */ + for (i = 0, j = 4*Nr; i < j; i += 4, j -= 4) { + temp = rk[i ]; rk[i ] = rk[j ]; rk[j ] = temp; + temp = rk[i + 1]; rk[i + 1] = rk[j + 1]; rk[j + 1] = temp; + temp = rk[i + 2]; rk[i + 2] = rk[j + 2]; rk[j + 2] = temp; + temp = rk[i + 3]; rk[i + 3] = rk[j + 3]; rk[j + 3] = temp; + } + /* apply the inverse MixColumn transform to all round keys but the first and the last: */ + for (i = 1; i < Nr; i++) { + rk += 4; + rk[0] = + Td0[Te4[(rk[0] >> 24) ] & 0xff] ^ + Td1[Te4[(rk[0] >> 16) & 0xff] & 0xff] ^ + Td2[Te4[(rk[0] >> 8) & 0xff] & 0xff] ^ + Td3[Te4[(rk[0] ) & 0xff] & 0xff]; + rk[1] = + Td0[Te4[(rk[1] >> 24) ] & 0xff] ^ + Td1[Te4[(rk[1] >> 16) & 0xff] & 0xff] ^ + Td2[Te4[(rk[1] >> 8) & 0xff] & 0xff] ^ + Td3[Te4[(rk[1] ) & 0xff] & 0xff]; + rk[2] = + Td0[Te4[(rk[2] >> 24) ] & 0xff] ^ + Td1[Te4[(rk[2] >> 16) & 0xff] & 0xff] ^ + Td2[Te4[(rk[2] >> 8) & 0xff] & 0xff] ^ + Td3[Te4[(rk[2] ) & 0xff] & 0xff]; + rk[3] = + Td0[Te4[(rk[3] >> 24) ] & 0xff] ^ + Td1[Te4[(rk[3] >> 16) & 0xff] & 0xff] ^ + Td2[Te4[(rk[3] >> 8) & 0xff] & 0xff] ^ + Td3[Te4[(rk[3] ) & 0xff] & 0xff]; + } + return Nr; +} + +void rijndaelEncrypt(const u32 rk[/*4*(Nr + 1)*/], int Nr, const u8 pt[16], u8 ct[16]) { + u32 s0, s1, s2, s3, t0, t1, t2, t3; +#ifndef FULL_UNROLL + int r; +#endif /* ?FULL_UNROLL */ + + /* + * map byte array block to cipher state + * and add initial round key: + */ + s0 = GETU32(pt ) ^ rk[0]; + s1 = GETU32(pt + 4) ^ rk[1]; + s2 = GETU32(pt + 8) ^ rk[2]; + s3 = GETU32(pt + 12) ^ rk[3]; +#ifdef FULL_UNROLL + /* round 1: */ + t0 = Te0[s0 >> 24] ^ Te1[(s1 >> 16) & 0xff] ^ Te2[(s2 >> 8) & 0xff] ^ Te3[s3 & 0xff] ^ rk[ 4]; + t1 = Te0[s1 >> 24] ^ Te1[(s2 >> 16) & 0xff] ^ Te2[(s3 >> 8) & 0xff] ^ Te3[s0 & 0xff] ^ rk[ 5]; + t2 = Te0[s2 >> 24] ^ Te1[(s3 >> 16) & 0xff] ^ Te2[(s0 >> 8) & 0xff] ^ Te3[s1 & 0xff] ^ rk[ 6]; + t3 = Te0[s3 >> 24] ^ Te1[(s0 >> 16) & 0xff] ^ Te2[(s1 >> 8) & 0xff] ^ Te3[s2 & 0xff] ^ rk[ 7]; + /* round 2: */ + s0 = Te0[t0 >> 24] ^ Te1[(t1 >> 16) & 0xff] ^ Te2[(t2 >> 8) & 0xff] ^ Te3[t3 & 0xff] ^ rk[ 8]; + s1 = Te0[t1 >> 24] ^ Te1[(t2 >> 16) & 0xff] ^ Te2[(t3 >> 8) & 0xff] ^ Te3[t0 & 0xff] ^ rk[ 9]; + s2 = Te0[t2 >> 24] ^ Te1[(t3 >> 16) & 0xff] ^ Te2[(t0 >> 8) & 0xff] ^ Te3[t1 & 0xff] ^ rk[10]; + s3 = Te0[t3 >> 24] ^ Te1[(t0 >> 16) & 0xff] ^ Te2[(t1 >> 8) & 0xff] ^ Te3[t2 & 0xff] ^ rk[11]; + /* round 3: */ + t0 = Te0[s0 >> 24] ^ Te1[(s1 >> 16) & 0xff] ^ Te2[(s2 >> 8) & 0xff] ^ Te3[s3 & 0xff] ^ rk[12]; + t1 = Te0[s1 >> 24] ^ Te1[(s2 >> 16) & 0xff] ^ Te2[(s3 >> 8) & 0xff] ^ Te3[s0 & 0xff] ^ rk[13]; + t2 = Te0[s2 >> 24] ^ Te1[(s3 >> 16) & 0xff] ^ Te2[(s0 >> 8) & 0xff] ^ Te3[s1 & 0xff] ^ rk[14]; + t3 = Te0[s3 >> 24] ^ Te1[(s0 >> 16) & 0xff] ^ Te2[(s1 >> 8) & 0xff] ^ Te3[s2 & 0xff] ^ rk[15]; + /* round 4: */ + s0 = Te0[t0 >> 24] ^ Te1[(t1 >> 16) & 0xff] ^ Te2[(t2 >> 8) & 0xff] ^ Te3[t3 & 0xff] ^ rk[16]; + s1 = Te0[t1 >> 24] ^ Te1[(t2 >> 16) & 0xff] ^ Te2[(t3 >> 8) & 0xff] ^ Te3[t0 & 0xff] ^ rk[17]; + s2 = Te0[t2 >> 24] ^ Te1[(t3 >> 16) & 0xff] ^ Te2[(t0 >> 8) & 0xff] ^ Te3[t1 & 0xff] ^ rk[18]; + s3 = Te0[t3 >> 24] ^ Te1[(t0 >> 16) & 0xff] ^ Te2[(t1 >> 8) & 0xff] ^ Te3[t2 & 0xff] ^ rk[19]; + /* round 5: */ + t0 = Te0[s0 >> 24] ^ Te1[(s1 >> 16) & 0xff] ^ Te2[(s2 >> 8) & 0xff] ^ Te3[s3 & 0xff] ^ rk[20]; + t1 = Te0[s1 >> 24] ^ Te1[(s2 >> 16) & 0xff] ^ Te2[(s3 >> 8) & 0xff] ^ Te3[s0 & 0xff] ^ rk[21]; + t2 = Te0[s2 >> 24] ^ Te1[(s3 >> 16) & 0xff] ^ Te2[(s0 >> 8) & 0xff] ^ Te3[s1 & 0xff] ^ rk[22]; + t3 = Te0[s3 >> 24] ^ Te1[(s0 >> 16) & 0xff] ^ Te2[(s1 >> 8) & 0xff] ^ Te3[s2 & 0xff] ^ rk[23]; + /* round 6: */ + s0 = Te0[t0 >> 24] ^ Te1[(t1 >> 16) & 0xff] ^ Te2[(t2 >> 8) & 0xff] ^ Te3[t3 & 0xff] ^ rk[24]; + s1 = Te0[t1 >> 24] ^ Te1[(t2 >> 16) & 0xff] ^ Te2[(t3 >> 8) & 0xff] ^ Te3[t0 & 0xff] ^ rk[25]; + s2 = Te0[t2 >> 24] ^ Te1[(t3 >> 16) & 0xff] ^ Te2[(t0 >> 8) & 0xff] ^ Te3[t1 & 0xff] ^ rk[26]; + s3 = Te0[t3 >> 24] ^ Te1[(t0 >> 16) & 0xff] ^ Te2[(t1 >> 8) & 0xff] ^ Te3[t2 & 0xff] ^ rk[27]; + /* round 7: */ + t0 = Te0[s0 >> 24] ^ Te1[(s1 >> 16) & 0xff] ^ Te2[(s2 >> 8) & 0xff] ^ Te3[s3 & 0xff] ^ rk[28]; + t1 = Te0[s1 >> 24] ^ Te1[(s2 >> 16) & 0xff] ^ Te2[(s3 >> 8) & 0xff] ^ Te3[s0 & 0xff] ^ rk[29]; + t2 = Te0[s2 >> 24] ^ Te1[(s3 >> 16) & 0xff] ^ Te2[(s0 >> 8) & 0xff] ^ Te3[s1 & 0xff] ^ rk[30]; + t3 = Te0[s3 >> 24] ^ Te1[(s0 >> 16) & 0xff] ^ Te2[(s1 >> 8) & 0xff] ^ Te3[s2 & 0xff] ^ rk[31]; + /* round 8: */ + s0 = Te0[t0 >> 24] ^ Te1[(t1 >> 16) & 0xff] ^ Te2[(t2 >> 8) & 0xff] ^ Te3[t3 & 0xff] ^ rk[32]; + s1 = Te0[t1 >> 24] ^ Te1[(t2 >> 16) & 0xff] ^ Te2[(t3 >> 8) & 0xff] ^ Te3[t0 & 0xff] ^ rk[33]; + s2 = Te0[t2 >> 24] ^ Te1[(t3 >> 16) & 0xff] ^ Te2[(t0 >> 8) & 0xff] ^ Te3[t1 & 0xff] ^ rk[34]; + s3 = Te0[t3 >> 24] ^ Te1[(t0 >> 16) & 0xff] ^ Te2[(t1 >> 8) & 0xff] ^ Te3[t2 & 0xff] ^ rk[35]; + /* round 9: */ + t0 = Te0[s0 >> 24] ^ Te1[(s1 >> 16) & 0xff] ^ Te2[(s2 >> 8) & 0xff] ^ Te3[s3 & 0xff] ^ rk[36]; + t1 = Te0[s1 >> 24] ^ Te1[(s2 >> 16) & 0xff] ^ Te2[(s3 >> 8) & 0xff] ^ Te3[s0 & 0xff] ^ rk[37]; + t2 = Te0[s2 >> 24] ^ Te1[(s3 >> 16) & 0xff] ^ Te2[(s0 >> 8) & 0xff] ^ Te3[s1 & 0xff] ^ rk[38]; + t3 = Te0[s3 >> 24] ^ Te1[(s0 >> 16) & 0xff] ^ Te2[(s1 >> 8) & 0xff] ^ Te3[s2 & 0xff] ^ rk[39]; + if (Nr > 10) { + /* round 10: */ + s0 = Te0[t0 >> 24] ^ Te1[(t1 >> 16) & 0xff] ^ Te2[(t2 >> 8) & 0xff] ^ Te3[t3 & 0xff] ^ rk[40]; + s1 = Te0[t1 >> 24] ^ Te1[(t2 >> 16) & 0xff] ^ Te2[(t3 >> 8) & 0xff] ^ Te3[t0 & 0xff] ^ rk[41]; + s2 = Te0[t2 >> 24] ^ Te1[(t3 >> 16) & 0xff] ^ Te2[(t0 >> 8) & 0xff] ^ Te3[t1 & 0xff] ^ rk[42]; + s3 = Te0[t3 >> 24] ^ Te1[(t0 >> 16) & 0xff] ^ Te2[(t1 >> 8) & 0xff] ^ Te3[t2 & 0xff] ^ rk[43]; + /* round 11: */ + t0 = Te0[s0 >> 24] ^ Te1[(s1 >> 16) & 0xff] ^ Te2[(s2 >> 8) & 0xff] ^ Te3[s3 & 0xff] ^ rk[44]; + t1 = Te0[s1 >> 24] ^ Te1[(s2 >> 16) & 0xff] ^ Te2[(s3 >> 8) & 0xff] ^ Te3[s0 & 0xff] ^ rk[45]; + t2 = Te0[s2 >> 24] ^ Te1[(s3 >> 16) & 0xff] ^ Te2[(s0 >> 8) & 0xff] ^ Te3[s1 & 0xff] ^ rk[46]; + t3 = Te0[s3 >> 24] ^ Te1[(s0 >> 16) & 0xff] ^ Te2[(s1 >> 8) & 0xff] ^ Te3[s2 & 0xff] ^ rk[47]; + if (Nr > 12) { + /* round 12: */ + s0 = Te0[t0 >> 24] ^ Te1[(t1 >> 16) & 0xff] ^ Te2[(t2 >> 8) & 0xff] ^ Te3[t3 & 0xff] ^ rk[48]; + s1 = Te0[t1 >> 24] ^ Te1[(t2 >> 16) & 0xff] ^ Te2[(t3 >> 8) & 0xff] ^ Te3[t0 & 0xff] ^ rk[49]; + s2 = Te0[t2 >> 24] ^ Te1[(t3 >> 16) & 0xff] ^ Te2[(t0 >> 8) & 0xff] ^ Te3[t1 & 0xff] ^ rk[50]; + s3 = Te0[t3 >> 24] ^ Te1[(t0 >> 16) & 0xff] ^ Te2[(t1 >> 8) & 0xff] ^ Te3[t2 & 0xff] ^ rk[51]; + /* round 13: */ + t0 = Te0[s0 >> 24] ^ Te1[(s1 >> 16) & 0xff] ^ Te2[(s2 >> 8) & 0xff] ^ Te3[s3 & 0xff] ^ rk[52]; + t1 = Te0[s1 >> 24] ^ Te1[(s2 >> 16) & 0xff] ^ Te2[(s3 >> 8) & 0xff] ^ Te3[s0 & 0xff] ^ rk[53]; + t2 = Te0[s2 >> 24] ^ Te1[(s3 >> 16) & 0xff] ^ Te2[(s0 >> 8) & 0xff] ^ Te3[s1 & 0xff] ^ rk[54]; + t3 = Te0[s3 >> 24] ^ Te1[(s0 >> 16) & 0xff] ^ Te2[(s1 >> 8) & 0xff] ^ Te3[s2 & 0xff] ^ rk[55]; + } + } + rk += Nr << 2; +#else /* !FULL_UNROLL */ + /* + * Nr - 1 full rounds: + */ + r = Nr >> 1; + for (;;) { + t0 = + Te0[(s0 >> 24) ] ^ + Te1[(s1 >> 16) & 0xff] ^ + Te2[(s2 >> 8) & 0xff] ^ + Te3[(s3 ) & 0xff] ^ + rk[4]; + t1 = + Te0[(s1 >> 24) ] ^ + Te1[(s2 >> 16) & 0xff] ^ + Te2[(s3 >> 8) & 0xff] ^ + Te3[(s0 ) & 0xff] ^ + rk[5]; + t2 = + Te0[(s2 >> 24) ] ^ + Te1[(s3 >> 16) & 0xff] ^ + Te2[(s0 >> 8) & 0xff] ^ + Te3[(s1 ) & 0xff] ^ + rk[6]; + t3 = + Te0[(s3 >> 24) ] ^ + Te1[(s0 >> 16) & 0xff] ^ + Te2[(s1 >> 8) & 0xff] ^ + Te3[(s2 ) & 0xff] ^ + rk[7]; + + rk += 8; + if (--r == 0) { + break; + } + + s0 = + Te0[(t0 >> 24) ] ^ + Te1[(t1 >> 16) & 0xff] ^ + Te2[(t2 >> 8) & 0xff] ^ + Te3[(t3 ) & 0xff] ^ + rk[0]; + s1 = + Te0[(t1 >> 24) ] ^ + Te1[(t2 >> 16) & 0xff] ^ + Te2[(t3 >> 8) & 0xff] ^ + Te3[(t0 ) & 0xff] ^ + rk[1]; + s2 = + Te0[(t2 >> 24) ] ^ + Te1[(t3 >> 16) & 0xff] ^ + Te2[(t0 >> 8) & 0xff] ^ + Te3[(t1 ) & 0xff] ^ + rk[2]; + s3 = + Te0[(t3 >> 24) ] ^ + Te1[(t0 >> 16) & 0xff] ^ + Te2[(t1 >> 8) & 0xff] ^ + Te3[(t2 ) & 0xff] ^ + rk[3]; + } +#endif /* ?FULL_UNROLL */ + /* + * apply last round and + * map cipher state to byte array block: + */ + s0 = + (Te4[(t0 >> 24) ] & 0xff000000) ^ + (Te4[(t1 >> 16) & 0xff] & 0x00ff0000) ^ + (Te4[(t2 >> 8) & 0xff] & 0x0000ff00) ^ + (Te4[(t3 ) & 0xff] & 0x000000ff) ^ + rk[0]; + PUTU32(ct , s0); + s1 = + (Te4[(t1 >> 24) ] & 0xff000000) ^ + (Te4[(t2 >> 16) & 0xff] & 0x00ff0000) ^ + (Te4[(t3 >> 8) & 0xff] & 0x0000ff00) ^ + (Te4[(t0 ) & 0xff] & 0x000000ff) ^ + rk[1]; + PUTU32(ct + 4, s1); + s2 = + (Te4[(t2 >> 24) ] & 0xff000000) ^ + (Te4[(t3 >> 16) & 0xff] & 0x00ff0000) ^ + (Te4[(t0 >> 8) & 0xff] & 0x0000ff00) ^ + (Te4[(t1 ) & 0xff] & 0x000000ff) ^ + rk[2]; + PUTU32(ct + 8, s2); + s3 = + (Te4[(t3 >> 24) ] & 0xff000000) ^ + (Te4[(t0 >> 16) & 0xff] & 0x00ff0000) ^ + (Te4[(t1 >> 8) & 0xff] & 0x0000ff00) ^ + (Te4[(t2 ) & 0xff] & 0x000000ff) ^ + rk[3]; + PUTU32(ct + 12, s3); +} + +void rijndaelDecrypt(const u32 rk[/*4*(Nr + 1)*/], int Nr, const u8 ct[16], u8 pt[16]) { + u32 s0, s1, s2, s3, t0, t1, t2, t3; +#ifndef FULL_UNROLL + int r; +#endif /* ?FULL_UNROLL */ + + /* + * map byte array block to cipher state + * and add initial round key: + */ + s0 = GETU32(ct ) ^ rk[0]; + s1 = GETU32(ct + 4) ^ rk[1]; + s2 = GETU32(ct + 8) ^ rk[2]; + s3 = GETU32(ct + 12) ^ rk[3]; +#ifdef FULL_UNROLL + /* round 1: */ + t0 = Td0[s0 >> 24] ^ Td1[(s3 >> 16) & 0xff] ^ Td2[(s2 >> 8) & 0xff] ^ Td3[s1 & 0xff] ^ rk[ 4]; + t1 = Td0[s1 >> 24] ^ Td1[(s0 >> 16) & 0xff] ^ Td2[(s3 >> 8) & 0xff] ^ Td3[s2 & 0xff] ^ rk[ 5]; + t2 = Td0[s2 >> 24] ^ Td1[(s1 >> 16) & 0xff] ^ Td2[(s0 >> 8) & 0xff] ^ Td3[s3 & 0xff] ^ rk[ 6]; + t3 = Td0[s3 >> 24] ^ Td1[(s2 >> 16) & 0xff] ^ Td2[(s1 >> 8) & 0xff] ^ Td3[s0 & 0xff] ^ rk[ 7]; + /* round 2: */ + s0 = Td0[t0 >> 24] ^ Td1[(t3 >> 16) & 0xff] ^ Td2[(t2 >> 8) & 0xff] ^ Td3[t1 & 0xff] ^ rk[ 8]; + s1 = Td0[t1 >> 24] ^ Td1[(t0 >> 16) & 0xff] ^ Td2[(t3 >> 8) & 0xff] ^ Td3[t2 & 0xff] ^ rk[ 9]; + s2 = Td0[t2 >> 24] ^ Td1[(t1 >> 16) & 0xff] ^ Td2[(t0 >> 8) & 0xff] ^ Td3[t3 & 0xff] ^ rk[10]; + s3 = Td0[t3 >> 24] ^ Td1[(t2 >> 16) & 0xff] ^ Td2[(t1 >> 8) & 0xff] ^ Td3[t0 & 0xff] ^ rk[11]; + /* round 3: */ + t0 = Td0[s0 >> 24] ^ Td1[(s3 >> 16) & 0xff] ^ Td2[(s2 >> 8) & 0xff] ^ Td3[s1 & 0xff] ^ rk[12]; + t1 = Td0[s1 >> 24] ^ Td1[(s0 >> 16) & 0xff] ^ Td2[(s3 >> 8) & 0xff] ^ Td3[s2 & 0xff] ^ rk[13]; + t2 = Td0[s2 >> 24] ^ Td1[(s1 >> 16) & 0xff] ^ Td2[(s0 >> 8) & 0xff] ^ Td3[s3 & 0xff] ^ rk[14]; + t3 = Td0[s3 >> 24] ^ Td1[(s2 >> 16) & 0xff] ^ Td2[(s1 >> 8) & 0xff] ^ Td3[s0 & 0xff] ^ rk[15]; + /* round 4: */ + s0 = Td0[t0 >> 24] ^ Td1[(t3 >> 16) & 0xff] ^ Td2[(t2 >> 8) & 0xff] ^ Td3[t1 & 0xff] ^ rk[16]; + s1 = Td0[t1 >> 24] ^ Td1[(t0 >> 16) & 0xff] ^ Td2[(t3 >> 8) & 0xff] ^ Td3[t2 & 0xff] ^ rk[17]; + s2 = Td0[t2 >> 24] ^ Td1[(t1 >> 16) & 0xff] ^ Td2[(t0 >> 8) & 0xff] ^ Td3[t3 & 0xff] ^ rk[18]; + s3 = Td0[t3 >> 24] ^ Td1[(t2 >> 16) & 0xff] ^ Td2[(t1 >> 8) & 0xff] ^ Td3[t0 & 0xff] ^ rk[19]; + /* round 5: */ + t0 = Td0[s0 >> 24] ^ Td1[(s3 >> 16) & 0xff] ^ Td2[(s2 >> 8) & 0xff] ^ Td3[s1 & 0xff] ^ rk[20]; + t1 = Td0[s1 >> 24] ^ Td1[(s0 >> 16) & 0xff] ^ Td2[(s3 >> 8) & 0xff] ^ Td3[s2 & 0xff] ^ rk[21]; + t2 = Td0[s2 >> 24] ^ Td1[(s1 >> 16) & 0xff] ^ Td2[(s0 >> 8) & 0xff] ^ Td3[s3 & 0xff] ^ rk[22]; + t3 = Td0[s3 >> 24] ^ Td1[(s2 >> 16) & 0xff] ^ Td2[(s1 >> 8) & 0xff] ^ Td3[s0 & 0xff] ^ rk[23]; + /* round 6: */ + s0 = Td0[t0 >> 24] ^ Td1[(t3 >> 16) & 0xff] ^ Td2[(t2 >> 8) & 0xff] ^ Td3[t1 & 0xff] ^ rk[24]; + s1 = Td0[t1 >> 24] ^ Td1[(t0 >> 16) & 0xff] ^ Td2[(t3 >> 8) & 0xff] ^ Td3[t2 & 0xff] ^ rk[25]; + s2 = Td0[t2 >> 24] ^ Td1[(t1 >> 16) & 0xff] ^ Td2[(t0 >> 8) & 0xff] ^ Td3[t3 & 0xff] ^ rk[26]; + s3 = Td0[t3 >> 24] ^ Td1[(t2 >> 16) & 0xff] ^ Td2[(t1 >> 8) & 0xff] ^ Td3[t0 & 0xff] ^ rk[27]; + /* round 7: */ + t0 = Td0[s0 >> 24] ^ Td1[(s3 >> 16) & 0xff] ^ Td2[(s2 >> 8) & 0xff] ^ Td3[s1 & 0xff] ^ rk[28]; + t1 = Td0[s1 >> 24] ^ Td1[(s0 >> 16) & 0xff] ^ Td2[(s3 >> 8) & 0xff] ^ Td3[s2 & 0xff] ^ rk[29]; + t2 = Td0[s2 >> 24] ^ Td1[(s1 >> 16) & 0xff] ^ Td2[(s0 >> 8) & 0xff] ^ Td3[s3 & 0xff] ^ rk[30]; + t3 = Td0[s3 >> 24] ^ Td1[(s2 >> 16) & 0xff] ^ Td2[(s1 >> 8) & 0xff] ^ Td3[s0 & 0xff] ^ rk[31]; + /* round 8: */ + s0 = Td0[t0 >> 24] ^ Td1[(t3 >> 16) & 0xff] ^ Td2[(t2 >> 8) & 0xff] ^ Td3[t1 & 0xff] ^ rk[32]; + s1 = Td0[t1 >> 24] ^ Td1[(t0 >> 16) & 0xff] ^ Td2[(t3 >> 8) & 0xff] ^ Td3[t2 & 0xff] ^ rk[33]; + s2 = Td0[t2 >> 24] ^ Td1[(t1 >> 16) & 0xff] ^ Td2[(t0 >> 8) & 0xff] ^ Td3[t3 & 0xff] ^ rk[34]; + s3 = Td0[t3 >> 24] ^ Td1[(t2 >> 16) & 0xff] ^ Td2[(t1 >> 8) & 0xff] ^ Td3[t0 & 0xff] ^ rk[35]; + /* round 9: */ + t0 = Td0[s0 >> 24] ^ Td1[(s3 >> 16) & 0xff] ^ Td2[(s2 >> 8) & 0xff] ^ Td3[s1 & 0xff] ^ rk[36]; + t1 = Td0[s1 >> 24] ^ Td1[(s0 >> 16) & 0xff] ^ Td2[(s3 >> 8) & 0xff] ^ Td3[s2 & 0xff] ^ rk[37]; + t2 = Td0[s2 >> 24] ^ Td1[(s1 >> 16) & 0xff] ^ Td2[(s0 >> 8) & 0xff] ^ Td3[s3 & 0xff] ^ rk[38]; + t3 = Td0[s3 >> 24] ^ Td1[(s2 >> 16) & 0xff] ^ Td2[(s1 >> 8) & 0xff] ^ Td3[s0 & 0xff] ^ rk[39]; + if (Nr > 10) { + /* round 10: */ + s0 = Td0[t0 >> 24] ^ Td1[(t3 >> 16) & 0xff] ^ Td2[(t2 >> 8) & 0xff] ^ Td3[t1 & 0xff] ^ rk[40]; + s1 = Td0[t1 >> 24] ^ Td1[(t0 >> 16) & 0xff] ^ Td2[(t3 >> 8) & 0xff] ^ Td3[t2 & 0xff] ^ rk[41]; + s2 = Td0[t2 >> 24] ^ Td1[(t1 >> 16) & 0xff] ^ Td2[(t0 >> 8) & 0xff] ^ Td3[t3 & 0xff] ^ rk[42]; + s3 = Td0[t3 >> 24] ^ Td1[(t2 >> 16) & 0xff] ^ Td2[(t1 >> 8) & 0xff] ^ Td3[t0 & 0xff] ^ rk[43]; + /* round 11: */ + t0 = Td0[s0 >> 24] ^ Td1[(s3 >> 16) & 0xff] ^ Td2[(s2 >> 8) & 0xff] ^ Td3[s1 & 0xff] ^ rk[44]; + t1 = Td0[s1 >> 24] ^ Td1[(s0 >> 16) & 0xff] ^ Td2[(s3 >> 8) & 0xff] ^ Td3[s2 & 0xff] ^ rk[45]; + t2 = Td0[s2 >> 24] ^ Td1[(s1 >> 16) & 0xff] ^ Td2[(s0 >> 8) & 0xff] ^ Td3[s3 & 0xff] ^ rk[46]; + t3 = Td0[s3 >> 24] ^ Td1[(s2 >> 16) & 0xff] ^ Td2[(s1 >> 8) & 0xff] ^ Td3[s0 & 0xff] ^ rk[47]; + if (Nr > 12) { + /* round 12: */ + s0 = Td0[t0 >> 24] ^ Td1[(t3 >> 16) & 0xff] ^ Td2[(t2 >> 8) & 0xff] ^ Td3[t1 & 0xff] ^ rk[48]; + s1 = Td0[t1 >> 24] ^ Td1[(t0 >> 16) & 0xff] ^ Td2[(t3 >> 8) & 0xff] ^ Td3[t2 & 0xff] ^ rk[49]; + s2 = Td0[t2 >> 24] ^ Td1[(t1 >> 16) & 0xff] ^ Td2[(t0 >> 8) & 0xff] ^ Td3[t3 & 0xff] ^ rk[50]; + s3 = Td0[t3 >> 24] ^ Td1[(t2 >> 16) & 0xff] ^ Td2[(t1 >> 8) & 0xff] ^ Td3[t0 & 0xff] ^ rk[51]; + /* round 13: */ + t0 = Td0[s0 >> 24] ^ Td1[(s3 >> 16) & 0xff] ^ Td2[(s2 >> 8) & 0xff] ^ Td3[s1 & 0xff] ^ rk[52]; + t1 = Td0[s1 >> 24] ^ Td1[(s0 >> 16) & 0xff] ^ Td2[(s3 >> 8) & 0xff] ^ Td3[s2 & 0xff] ^ rk[53]; + t2 = Td0[s2 >> 24] ^ Td1[(s1 >> 16) & 0xff] ^ Td2[(s0 >> 8) & 0xff] ^ Td3[s3 & 0xff] ^ rk[54]; + t3 = Td0[s3 >> 24] ^ Td1[(s2 >> 16) & 0xff] ^ Td2[(s1 >> 8) & 0xff] ^ Td3[s0 & 0xff] ^ rk[55]; + } + } + rk += Nr << 2; +#else /* !FULL_UNROLL */ + /* + * Nr - 1 full rounds: + */ + r = Nr >> 1; + for (;;) { + t0 = + Td0[(s0 >> 24) ] ^ + Td1[(s3 >> 16) & 0xff] ^ + Td2[(s2 >> 8) & 0xff] ^ + Td3[(s1 ) & 0xff] ^ + rk[4]; + t1 = + Td0[(s1 >> 24) ] ^ + Td1[(s0 >> 16) & 0xff] ^ + Td2[(s3 >> 8) & 0xff] ^ + Td3[(s2 ) & 0xff] ^ + rk[5]; + t2 = + Td0[(s2 >> 24) ] ^ + Td1[(s1 >> 16) & 0xff] ^ + Td2[(s0 >> 8) & 0xff] ^ + Td3[(s3 ) & 0xff] ^ + rk[6]; + t3 = + Td0[(s3 >> 24) ] ^ + Td1[(s2 >> 16) & 0xff] ^ + Td2[(s1 >> 8) & 0xff] ^ + Td3[(s0 ) & 0xff] ^ + rk[7]; + + rk += 8; + if (--r == 0) { + break; + } + + s0 = + Td0[(t0 >> 24) ] ^ + Td1[(t3 >> 16) & 0xff] ^ + Td2[(t2 >> 8) & 0xff] ^ + Td3[(t1 ) & 0xff] ^ + rk[0]; + s1 = + Td0[(t1 >> 24) ] ^ + Td1[(t0 >> 16) & 0xff] ^ + Td2[(t3 >> 8) & 0xff] ^ + Td3[(t2 ) & 0xff] ^ + rk[1]; + s2 = + Td0[(t2 >> 24) ] ^ + Td1[(t1 >> 16) & 0xff] ^ + Td2[(t0 >> 8) & 0xff] ^ + Td3[(t3 ) & 0xff] ^ + rk[2]; + s3 = + Td0[(t3 >> 24) ] ^ + Td1[(t2 >> 16) & 0xff] ^ + Td2[(t1 >> 8) & 0xff] ^ + Td3[(t0 ) & 0xff] ^ + rk[3]; + } +#endif /* ?FULL_UNROLL */ + /* + * apply last round and + * map cipher state to byte array block: + */ + s0 = + (Td4[(t0 >> 24) ] & 0xff000000) ^ + (Td4[(t3 >> 16) & 0xff] & 0x00ff0000) ^ + (Td4[(t2 >> 8) & 0xff] & 0x0000ff00) ^ + (Td4[(t1 ) & 0xff] & 0x000000ff) ^ + rk[0]; + PUTU32(pt , s0); + s1 = + (Td4[(t1 >> 24) ] & 0xff000000) ^ + (Td4[(t0 >> 16) & 0xff] & 0x00ff0000) ^ + (Td4[(t3 >> 8) & 0xff] & 0x0000ff00) ^ + (Td4[(t2 ) & 0xff] & 0x000000ff) ^ + rk[1]; + PUTU32(pt + 4, s1); + s2 = + (Td4[(t2 >> 24) ] & 0xff000000) ^ + (Td4[(t1 >> 16) & 0xff] & 0x00ff0000) ^ + (Td4[(t0 >> 8) & 0xff] & 0x0000ff00) ^ + (Td4[(t3 ) & 0xff] & 0x000000ff) ^ + rk[2]; + PUTU32(pt + 8, s2); + s3 = + (Td4[(t3 >> 24) ] & 0xff000000) ^ + (Td4[(t2 >> 16) & 0xff] & 0x00ff0000) ^ + (Td4[(t1 >> 8) & 0xff] & 0x0000ff00) ^ + (Td4[(t0 ) & 0xff] & 0x000000ff) ^ + rk[3]; + PUTU32(pt + 12, s3); +} + +#ifdef INTERMEDIATE_VALUE_KAT + +void rijndaelEncryptRound(const u32 rk[/*4*(Nr + 1)*/], int Nr, u8 block[16], int rounds) { + int r; + u32 s0, s1, s2, s3, t0, t1, t2, t3; + + /* + * map byte array block to cipher state + * and add initial round key: + */ + s0 = GETU32(block ) ^ rk[0]; + s1 = GETU32(block + 4) ^ rk[1]; + s2 = GETU32(block + 8) ^ rk[2]; + s3 = GETU32(block + 12) ^ rk[3]; + rk += 4; + + /* + * Nr - 1 full rounds: + */ + for (r = (rounds < Nr ? rounds : Nr - 1); r > 0; r--) { + t0 = + Te0[(s0 >> 24) ] ^ + Te1[(s1 >> 16) & 0xff] ^ + Te2[(s2 >> 8) & 0xff] ^ + Te3[(s3 ) & 0xff] ^ + rk[0]; + t1 = + Te0[(s1 >> 24) ] ^ + Te1[(s2 >> 16) & 0xff] ^ + Te2[(s3 >> 8) & 0xff] ^ + Te3[(s0 ) & 0xff] ^ + rk[1]; + t2 = + Te0[(s2 >> 24) ] ^ + Te1[(s3 >> 16) & 0xff] ^ + Te2[(s0 >> 8) & 0xff] ^ + Te3[(s1 ) & 0xff] ^ + rk[2]; + t3 = + Te0[(s3 >> 24) ] ^ + Te1[(s0 >> 16) & 0xff] ^ + Te2[(s1 >> 8) & 0xff] ^ + Te3[(s2 ) & 0xff] ^ + rk[3]; + + s0 = t0; + s1 = t1; + s2 = t2; + s3 = t3; + rk += 4; + + } + + /* + * apply last round and + * map cipher state to byte array block: + */ + if (rounds == Nr) { + t0 = + (Te4[(s0 >> 24) ] & 0xff000000) ^ + (Te4[(s1 >> 16) & 0xff] & 0x00ff0000) ^ + (Te4[(s2 >> 8) & 0xff] & 0x0000ff00) ^ + (Te4[(s3 ) & 0xff] & 0x000000ff) ^ + rk[0]; + t1 = + (Te4[(s1 >> 24) ] & 0xff000000) ^ + (Te4[(s2 >> 16) & 0xff] & 0x00ff0000) ^ + (Te4[(s3 >> 8) & 0xff] & 0x0000ff00) ^ + (Te4[(s0 ) & 0xff] & 0x000000ff) ^ + rk[1]; + t2 = + (Te4[(s2 >> 24) ] & 0xff000000) ^ + (Te4[(s3 >> 16) & 0xff] & 0x00ff0000) ^ + (Te4[(s0 >> 8) & 0xff] & 0x0000ff00) ^ + (Te4[(s1 ) & 0xff] & 0x000000ff) ^ + rk[2]; + t3 = + (Te4[(s3 >> 24) ] & 0xff000000) ^ + (Te4[(s0 >> 16) & 0xff] & 0x00ff0000) ^ + (Te4[(s1 >> 8) & 0xff] & 0x0000ff00) ^ + (Te4[(s2 ) & 0xff] & 0x000000ff) ^ + rk[3]; + + s0 = t0; + s1 = t1; + s2 = t2; + s3 = t3; + } + + PUTU32(block , s0); + PUTU32(block + 4, s1); + PUTU32(block + 8, s2); + PUTU32(block + 12, s3); +} + +void rijndaelDecryptRound(const u32 rk[/*4*(Nr + 1)*/], int Nr, u8 block[16], int rounds) { + int r; + u32 s0, s1, s2, s3, t0, t1, t2, t3; + + /* + * map byte array block to cipher state + * and add initial round key: + */ + s0 = GETU32(block ) ^ rk[0]; + s1 = GETU32(block + 4) ^ rk[1]; + s2 = GETU32(block + 8) ^ rk[2]; + s3 = GETU32(block + 12) ^ rk[3]; + rk += 4; + + /* + * Nr - 1 full rounds: + */ + for (r = (rounds < Nr ? rounds : Nr) - 1; r > 0; r--) { + t0 = + Td0[(s0 >> 24) ] ^ + Td1[(s3 >> 16) & 0xff] ^ + Td2[(s2 >> 8) & 0xff] ^ + Td3[(s1 ) & 0xff] ^ + rk[0]; + t1 = + Td0[(s1 >> 24) ] ^ + Td1[(s0 >> 16) & 0xff] ^ + Td2[(s3 >> 8) & 0xff] ^ + Td3[(s2 ) & 0xff] ^ + rk[1]; + t2 = + Td0[(s2 >> 24) ] ^ + Td1[(s1 >> 16) & 0xff] ^ + Td2[(s0 >> 8) & 0xff] ^ + Td3[(s3 ) & 0xff] ^ + rk[2]; + t3 = + Td0[(s3 >> 24) ] ^ + Td1[(s2 >> 16) & 0xff] ^ + Td2[(s1 >> 8) & 0xff] ^ + Td3[(s0 ) & 0xff] ^ + rk[3]; + + s0 = t0; + s1 = t1; + s2 = t2; + s3 = t3; + rk += 4; + + } + + /* + * complete the last round and + * map cipher state to byte array block: + */ + t0 = + (Td4[(s0 >> 24) ] & 0xff000000) ^ + (Td4[(s3 >> 16) & 0xff] & 0x00ff0000) ^ + (Td4[(s2 >> 8) & 0xff] & 0x0000ff00) ^ + (Td4[(s1 ) & 0xff] & 0x000000ff); + t1 = + (Td4[(s1 >> 24) ] & 0xff000000) ^ + (Td4[(s0 >> 16) & 0xff] & 0x00ff0000) ^ + (Td4[(s3 >> 8) & 0xff] & 0x0000ff00) ^ + (Td4[(s2 ) & 0xff] & 0x000000ff); + t2 = + (Td4[(s2 >> 24) ] & 0xff000000) ^ + (Td4[(s1 >> 16) & 0xff] & 0x00ff0000) ^ + (Td4[(s0 >> 8) & 0xff] & 0x0000ff00) ^ + (Td4[(s3 ) & 0xff] & 0x000000ff); + t3 = + (Td4[(s3 >> 24) ] & 0xff000000) ^ + (Td4[(s2 >> 16) & 0xff] & 0x00ff0000) ^ + (Td4[(s1 >> 8) & 0xff] & 0x0000ff00) ^ + (Td4[(s0 ) & 0xff] & 0x000000ff); + + if (rounds == Nr) { + t0 ^= rk[0]; + t1 ^= rk[1]; + t2 ^= rk[2]; + t3 ^= rk[3]; + } + + PUTU32(block , t0); + PUTU32(block + 4, t1); + PUTU32(block + 8, t2); + PUTU32(block + 12, t3); +} + +#endif /* INTERMEDIATE_VALUE_KAT */ diff --git a/aezref/aezv5/ref/rijndael-alg-fst.h b/aezref/aezv5/ref/rijndael-alg-fst.h new file mode 100644 index 0000000..4b65d95 --- /dev/null +++ b/aezref/aezv5/ref/rijndael-alg-fst.h @@ -0,0 +1,52 @@ +/** + * rijndael-alg-fst.h + * + * @version 3.0 (December 2000) + * + * Optimised ANSI C code for the Rijndael cipher (now AES) + * + * @author Vincent Rijmen <vincent.rijmen@esat.kuleuven.ac.be> + * @author Antoon Bosselaers <antoon.bosselaers@esat.kuleuven.ac.be> + * @author Paulo Barreto <paulo.barreto@terra.com.br> + * + * This code is hereby placed in the public domain. + * + * THIS SOFTWARE IS PROVIDED BY THE AUTHORS ''AS IS'' AND ANY EXPRESS + * OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED + * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE + * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHORS OR CONTRIBUTORS BE + * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR + * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF + * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR + * BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, + * WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE + * OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, + * EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. + */ +#ifndef __RIJNDAEL_ALG_FST_H +#define __RIJNDAEL_ALG_FST_H + +/* INTERMEDIATE_VALUE_KAT must be defined to get AEZ's required AES behavior */ +#ifndef INTERMEDIATE_VALUE_KAT +#define INTERMEDIATE_VALUE_KAT +#endif + +#define MAXKC (256/32) +#define MAXKB (256/8) +#define MAXNR 14 + +typedef unsigned char u8; +typedef unsigned short u16; +typedef unsigned int u32; + +int rijndaelKeySetupEnc(u32 rk[/*4*(Nr + 1)*/], const u8 cipherKey[], int keyBits); +int rijndaelKeySetupDec(u32 rk[/*4*(Nr + 1)*/], const u8 cipherKey[], int keyBits); +void rijndaelEncrypt(const u32 rk[/*4*(Nr + 1)*/], int Nr, const u8 pt[16], u8 ct[16]); +void rijndaelDecrypt(const u32 rk[/*4*(Nr + 1)*/], int Nr, const u8 ct[16], u8 pt[16]); + +#ifdef INTERMEDIATE_VALUE_KAT +void rijndaelEncryptRound(const u32 rk[/*4*(Nr + 1)*/], int Nr, u8 block[16], int rounds); +void rijndaelDecryptRound(const u32 rk[/*4*(Nr + 1)*/], int Nr, u8 block[16], int rounds); +#endif /* INTERMEDIATE_VALUE_KAT */ + +#endif /* __RIJNDAEL_ALG_FST_H */ diff --git a/aezref/build.rs b/aezref/build.rs new file mode 100644 index 0000000..219ef44 --- /dev/null +++ b/aezref/build.rs @@ -0,0 +1,7 @@ +fn main() { + cc::Build::new() + .file("aezv5/ref/blake2b.c") + .file("aezv5/ref/encrypt.c") + .file("aezv5/ref/rijndael-alg-fst.c") + .compile("aezv5"); +} diff --git a/aezref/src/lib.rs b/aezref/src/lib.rs new file mode 100644 index 0000000..8fc902c --- /dev/null +++ b/aezref/src/lib.rs @@ -0,0 +1,155 @@ +//! Bindings to the reference (slow, unsafe) implementation of AEZ. +//! +//! **Warning**: +//! +//! > This version is slow and susceptible to side-channel attacks. +//! > Do not use for any purpose other than to understand AEZ. +//! +//! We only use it to compare outputs between `zears` and `aezref`. +use std::ffi::*; + +unsafe extern "C" { + fn Decrypt( + key: *const c_uchar, + keybytes: c_uint, + nonce: *const c_uchar, + noncebytes: c_uint, + ad: *const *const c_uchar, + adbytes: *const c_uint, + veclen: c_uint, + abytes: c_uint, + ciphertext: *const c_uchar, + cipherbytes: c_uint, + message: *mut c_uchar, + ) -> c_int; + + fn Encrypt( + key: *const c_uchar, + keybytes: c_uint, + nonce: *const c_uchar, + noncebytes: c_uint, + ad: *const *const c_uchar, + adbytes: *const c_uint, + veclen: c_uint, + abytes: c_uint, + message: *const c_uchar, + messagebytes: c_uint, + ciphertext: *mut c_uchar, + ); +} + +/// Encrypt the given `message` using the given `key`, `nonce` and associated data items `ad`. +/// +/// The ciphertext will be written to `ciphertext`. +/// +/// The number of authentication bytes will be inferred from the length difference between +/// `message` and `ciphertext` +/// +/// This function will panic if +/// +/// * any of the given byte strings is longer than [`c_uint::MAX`] (usually 2^32 - 1) +/// * the ciphertext buffer is smaller than the message buffer +pub fn encrypt(key: &[u8], nonce: &[u8], ad: &[&[u8]], message: &[u8], ciphertext: &mut [u8]) { + assert!(ciphertext.len() >= message.len()); + let adlens: Vec<c_uint> = ad + .into_iter() + .map(|x| x.len().try_into().expect("associated data item too long")) + .collect(); + let ad = ad.into_iter().map(|x| x.as_ptr()).collect::<Vec<_>>(); + let abytes = ciphertext.len() - message.len(); + unsafe { + Encrypt( + key.as_ptr(), + key.len().try_into().expect("key too long"), + nonce.as_ptr(), + nonce.len().try_into().expect("nonce too long"), + ad.as_ptr(), + adlens.as_ptr(), + ad.len().try_into().expect("too many associated data items"), + abytes.try_into().expect("too many authentication bytes"), + message.as_ptr(), + message.len().try_into().expect("message too long"), + ciphertext.as_mut_ptr(), + ) + } +} + +/// Decrypt the given `ciphertext` using the given `key`, `nonce` and associated data items `ad`. +/// +/// The plaintext message will be written to `message`. +/// +/// The number of authentication bytes will be inferred from the length difference between +/// `message` and `ciphertext`. +/// +/// If the authentication bytes do not match the expected bytes (i.e. the ciphertext has been +/// modified, or the wrong key/nonce/ad is provided), `Err(())` is returned. Otherwise, `Ok(())` is +/// returned. +/// +/// This function will panic if +/// +/// * any of the given byte strings is longer than [`c_uint::MAX`] (usually 2^32 - 1) +/// * the ciphertext buffer is smaller than the message buffer +pub fn decrypt( + key: &[u8], + nonce: &[u8], + ad: &[&[u8]], + ciphertext: &[u8], + message: &mut [u8], +) -> Result<(), ()> { + assert!(ciphertext.len() >= message.len()); + let adlens: Vec<c_uint> = ad + .into_iter() + .map(|x| x.len().try_into().expect("associated data item too long")) + .collect(); + let ad = ad.into_iter().map(|x| x.as_ptr()).collect::<Vec<_>>(); + let abytes = ciphertext.len() - message.len(); + let result = unsafe { + Decrypt( + key.as_ptr(), + key.len().try_into().expect("key too long"), + nonce.as_ptr(), + nonce.len().try_into().expect("nonce too long"), + ad.as_ptr(), + adlens.as_ptr(), + ad.len().try_into().expect("too many associated data items"), + abytes.try_into().expect("too many authentication bytes"), + ciphertext.as_ptr(), + ciphertext.len().try_into().expect("message too long"), + message.as_mut_ptr(), + ) + }; + match result { + 0 => Ok(()), + -1 => Err(()), + _ => panic!("unexpected return from Decrypt"), + } +} + +#[cfg(test)] +mod tests { + use super::*; + + #[test] + fn it_works() { + let mut ciphertext = [0u8; 10]; + encrypt( + b"foo", + b"bar", + &[b"ad 1", b"ad two"], + b"hey", + &mut ciphertext, + ); + + let mut message = [0u8; 3]; + decrypt( + b"foo", + b"bar", + &[b"ad 1", b"ad two"], + &ciphertext, + &mut message, + ) + .unwrap(); + + assert_eq!(&message, b"hey"); + } +} diff --git a/fuzz/Cargo.lock b/fuzz/Cargo.lock index 2ae96d1..1fce2f7 100644 --- a/fuzz/Cargo.lock +++ b/fuzz/Cargo.lock @@ -14,10 +14,8 @@ dependencies = [ ] [[package]] -name = "aez" -version = "0.0.7" -source = "registry+https://github.com/rust-lang/crates.io-index" -checksum = "4c0763d41b5091ea56fa39dd9390f08355b9e45c0310ec5d9fd99c0d23c9322e" +name = "aezref" +version = "0.1.0" dependencies = [ "cc", ] @@ -283,7 +281,7 @@ dependencies = [ name = "zears-fuzz" version = "0.0.0" dependencies = [ - "aez", + "aezref", "arbitrary", "libfuzzer-sys", "zears", diff --git a/fuzz/Cargo.toml b/fuzz/Cargo.toml index 7c063d6..043fbef 100644 --- a/fuzz/Cargo.toml +++ b/fuzz/Cargo.toml @@ -10,11 +10,13 @@ cargo-fuzz = true [dependencies] arbitrary = { version = "1.4.1", features = ["derive"] } libfuzzer-sys = "0.4" -aez = "0.0.7" [dependencies.zears] path = ".." +[dependencies.aezref] +path = "../aezref" + [[bin]] name = "fuzz_target_1" path = "fuzz_targets/fuzz_target_1.rs" diff --git a/fuzz/fuzz_targets/zears_vs_aez.rs b/fuzz/fuzz_targets/zears_vs_aez.rs index eda644d..39ccac6 100644 --- a/fuzz/fuzz_targets/zears_vs_aez.rs +++ b/fuzz/fuzz_targets/zears_vs_aez.rs @@ -9,28 +9,31 @@ use zears::Aez; struct Parameters<'a> { key: &'a [u8], nonce: &'a [u8], - ad: Option<&'a [u8]>, - tau: u32, + ad: &'a [u8], + adreps: u8, + tau: u8, message: &'a [u8], } +const MAX_LEN: usize = 16 * 1024 * 1024; + fuzz_target!(|data: Parameters| { - // Limitations stem from AEZ's underlying C library - if data.nonce.len() >= 1 - && data.nonce.len() <= 16 - && data.ad.map(|x| x.len()).unwrap_or(0) <= 16 - && data.tau <= 16 - && data.message.len() <= u32::MAX.try_into().unwrap() - && (!data.message.is_empty() || data.tau > 0) - { - let ad = match data.ad { - Some(ad) => &[ad] as &[&[u8]], - None => &[], - }; - let actual = Aez::new(data.key).encrypt(data.nonce, ad, data.tau, data.message); + if data.message.len() + data.tau as usize <= MAX_LEN { + let ad = (0..data.adreps) + .map(|i| { + let i = i.into(); + if i < data.ad.len() { + &data.ad[i..] + } else { + &[] + } + }) + .collect::<Vec<_>>(); + + let actual = Aez::new(data.key).encrypt(data.nonce, &ad, data.tau.into(), data.message); let mut expected = vec![0; data.message.len() + data.tau as usize]; - aez::Aez::new(data.key).encrypt(data.nonce, data.ad, data.message, &mut expected); + aezref::encrypt(data.key, data.nonce, &ad, data.message, &mut expected); assert_eq!(actual, expected); } |