| /* |
| * Copyright (C) 2015 Michael Brown <mbrown@fensystems.co.uk>. |
| * |
| * This program is free software; you can redistribute it and/or |
| * modify it under the terms of the GNU General Public License as |
| * published by the Free Software Foundation; either version 2 of the |
| * License, or any later version. |
| * |
| * This program is distributed in the hope that it will be useful, but |
| * WITHOUT ANY WARRANTY; without even the implied warranty of |
| * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
| * General Public License for more details. |
| * |
| * You should have received a copy of the GNU General Public License |
| * along with this program; if not, write to the Free Software |
| * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA |
| * 02110-1301, USA. |
| * |
| * You can also choose to distribute this program under the terms of |
| * the Unmodified Binary Distribution Licence (as given in the file |
| * COPYING.UBDL), provided that you have satisfied its requirements. |
| */ |
| |
| FILE_LICENCE ( GPL2_OR_LATER_OR_UBDL ); |
| |
| /** @file |
| * |
| * AES algorithm |
| * |
| */ |
| |
| #include <stdint.h> |
| #include <string.h> |
| #include <errno.h> |
| #include <assert.h> |
| #include <byteswap.h> |
| #include <ipxe/rotate.h> |
| #include <ipxe/crypto.h> |
| #include <ipxe/ecb.h> |
| #include <ipxe/cbc.h> |
| #include <ipxe/gcm.h> |
| #include <ipxe/aes.h> |
| |
| /** AES strides |
| * |
| * These are the strides (modulo 16) used to walk through the AES |
| * input state bytes in order of byte position after [Inv]ShiftRows. |
| */ |
| enum aes_stride { |
| /** Input stride for ShiftRows |
| * |
| * 0 4 8 c |
| * \ \ \ |
| * 1 5 9 d |
| * \ \ \ |
| * 2 6 a e |
| * \ \ \ |
| * 3 7 b f |
| */ |
| AES_STRIDE_SHIFTROWS = +5, |
| /** Input stride for InvShiftRows |
| * |
| * 0 4 8 c |
| * / / / |
| * 1 5 9 d |
| * / / / |
| * 2 6 a e |
| * / / / |
| * 3 7 b f |
| */ |
| AES_STRIDE_INVSHIFTROWS = -3, |
| }; |
| |
| /** A single AES lookup table entry |
| * |
| * This represents the product (in the Galois field GF(2^8)) of an |
| * eight-byte vector multiplier with a single scalar multiplicand. |
| * |
| * The vector multipliers used for AES will be {1,1,1,3,2,1,1,3} for |
| * MixColumns and {1,9,13,11,14,9,13,11} for InvMixColumns. This |
| * allows for the result of multiplying any single column of the |
| * [Inv]MixColumns matrix by a scalar value to be obtained simply by |
| * extracting the relevant four-byte subset from the lookup table |
| * entry. |
| * |
| * For example, to find the result of multiplying the second column of |
| * the MixColumns matrix by the scalar value 0x80: |
| * |
| * MixColumns column[0]: { 2, 1, 1, 3 } |
| * MixColumns column[1]: { 3, 2, 1, 1 } |
| * MixColumns column[2]: { 1, 3, 2, 1 } |
| * MixColumns column[3]: { 1, 1, 3, 2 } |
| * Vector multiplier: { 1, 1, 1, 3, 2, 1, 1, 3 } |
| * Scalar multiplicand: 0x80 |
| * Lookup table entry: { 0x80, 0x80, 0x80, 0x9b, 0x1b, 0x80, 0x80, 0x9b } |
| * |
| * The second column of the MixColumns matrix is {3,2,1,1}. The |
| * product of this column with the scalar value 0x80 can be obtained |
| * by extracting the relevant four-byte subset of the lookup table |
| * entry: |
| * |
| * MixColumns column[1]: { 3, 2, 1, 1 } |
| * Vector multiplier: { 1, 1, 1, 3, 2, 1, 1, 3 } |
| * Lookup table entry: { 0x80, 0x80, 0x80, 0x9b, 0x1b, 0x80, 0x80, 0x9b } |
| * Product: { 0x9b, 0x1b, 0x80, 0x80 } |
| * |
| * The column lookups require only seven bytes of the eight-byte |
| * entry: the remaining (first) byte is used to hold the scalar |
| * multiplicand itself (i.e. the first byte of the vector multiplier |
| * is always chosen to be 1). |
| */ |
| union aes_table_entry { |
| /** Viewed as an array of bytes */ |
| uint8_t byte[8]; |
| } __attribute__ (( packed )); |
| |
| /** An AES lookup table |
| * |
| * This represents the products (in the Galois field GF(2^8)) of a |
| * constant eight-byte vector multiplier with all possible 256 scalar |
| * multiplicands. |
| * |
| * The entries are indexed by the AES [Inv]SubBytes S-box output |
| * values (denoted S(N)). This allows for the result of multiplying |
| * any single column of the [Inv]MixColumns matrix by S(N) to be |
| * obtained simply by extracting the relevant four-byte subset from |
| * the Nth table entry. For example: |
| * |
| * Input byte (N): 0x3a |
| * SubBytes output S(N): 0x80 |
| * MixColumns column[1]: { 3, 2, 1, 1 } |
| * Vector multiplier: { 1, 1, 1, 3, 2, 1, 1, 3 } |
| * Table entry[0x3a]: { 0x80, 0x80, 0x80, 0x9b, 0x1b, 0x80, 0x80, 0x9b } |
| * Product: { 0x9b, 0x1b, 0x80, 0x80 } |
| * |
| * Since the first byte of the eight-byte vector multiplier is always |
| * chosen to be 1, the value of S(N) may be lookup up by extracting |
| * the first byte of the Nth table entry. |
| */ |
| struct aes_table { |
| /** Table entries, indexed by S(N) */ |
| union aes_table_entry entry[256]; |
| } __attribute__ (( aligned ( 8 ) )); |
| |
| /** AES MixColumns lookup table */ |
| static struct aes_table aes_mixcolumns; |
| |
| /** AES InvMixColumns lookup table */ |
| static struct aes_table aes_invmixcolumns; |
| |
| /** |
| * Multiply [Inv]MixColumns matrix column by scalar multiplicand |
| * |
| * @v entry AES lookup table entry for scalar multiplicand |
| * @v column [Inv]MixColumns matrix column index |
| * @ret product Product of matrix column with scalar multiplicand |
| */ |
| static inline __attribute__ (( always_inline )) uint32_t |
| aes_entry_column ( const union aes_table_entry *entry, unsigned int column ) { |
| const union { |
| uint8_t byte; |
| uint32_t column; |
| } __attribute__ (( may_alias )) *product; |
| |
| /* Locate relevant four-byte subset */ |
| product = container_of ( &entry->byte[ 4 - column ], |
| typeof ( *product ), byte ); |
| |
| /* Extract this four-byte subset */ |
| return product->column; |
| } |
| |
| /** |
| * Multiply [Inv]MixColumns matrix column by S-boxed input byte |
| * |
| * @v table AES lookup table |
| * @v stride AES row shift stride |
| * @v in AES input state |
| * @v offset Output byte offset (after [Inv]ShiftRows) |
| * @ret product Product of matrix column with S(input byte) |
| * |
| * Note that the specified offset is not the offset of the input byte; |
| * it is the offset of the output byte which corresponds to the input |
| * byte. This output byte offset is used to calculate both the input |
| * byte offset and to select the appropriate matric column. |
| * |
| * With a compile-time constant offset, this function will optimise |
| * down to a single "movzbl" (to extract the input byte) and will |
| * generate a single x86 memory reference expression which can then be |
| * used directly within a single "xorl" instruction. |
| */ |
| static inline __attribute__ (( always_inline )) uint32_t |
| aes_column ( const struct aes_table *table, size_t stride, |
| const union aes_matrix *in, size_t offset ) { |
| const union aes_table_entry *entry; |
| unsigned int byte; |
| |
| /* Extract input byte corresponding to this output byte offset |
| * (i.e. perform [Inv]ShiftRows). |
| */ |
| byte = in->byte[ ( stride * offset ) & 0xf ]; |
| |
| /* Locate lookup table entry for this input byte (i.e. perform |
| * [Inv]SubBytes). |
| */ |
| entry = &table->entry[byte]; |
| |
| /* Multiply appropriate matrix column by this input byte |
| * (i.e. perform [Inv]MixColumns). |
| */ |
| return aes_entry_column ( entry, ( offset & 0x3 ) ); |
| } |
| |
| /** |
| * Calculate intermediate round output column |
| * |
| * @v table AES lookup table |
| * @v stride AES row shift stride |
| * @v in AES input state |
| * @v key AES round key |
| * @v column Column index |
| * @ret output Output column value |
| */ |
| static inline __attribute__ (( always_inline )) uint32_t |
| aes_output ( const struct aes_table *table, size_t stride, |
| const union aes_matrix *in, const union aes_matrix *key, |
| unsigned int column ) { |
| size_t offset = ( column * 4 ); |
| |
| /* Perform [Inv]ShiftRows, [Inv]SubBytes, [Inv]MixColumns, and |
| * AddRoundKey for this column. The loop is unrolled to allow |
| * for the required compile-time constant optimisations. |
| */ |
| return ( aes_column ( table, stride, in, ( offset + 0 ) ) ^ |
| aes_column ( table, stride, in, ( offset + 1 ) ) ^ |
| aes_column ( table, stride, in, ( offset + 2 ) ) ^ |
| aes_column ( table, stride, in, ( offset + 3 ) ) ^ |
| key->column[column] ); |
| } |
| |
| /** |
| * Perform a single intermediate round |
| * |
| * @v table AES lookup table |
| * @v stride AES row shift stride |
| * @v in AES input state |
| * @v out AES output state |
| * @v key AES round key |
| */ |
| static inline __attribute__ (( always_inline )) void |
| aes_round ( const struct aes_table *table, size_t stride, |
| const union aes_matrix *in, union aes_matrix *out, |
| const union aes_matrix *key ) { |
| |
| /* Perform [Inv]ShiftRows, [Inv]SubBytes, [Inv]MixColumns, and |
| * AddRoundKey for all columns. The loop is unrolled to allow |
| * for the required compile-time constant optimisations. |
| */ |
| out->column[0] = aes_output ( table, stride, in, key, 0 ); |
| out->column[1] = aes_output ( table, stride, in, key, 1 ); |
| out->column[2] = aes_output ( table, stride, in, key, 2 ); |
| out->column[3] = aes_output ( table, stride, in, key, 3 ); |
| } |
| |
| /** |
| * Perform encryption intermediate rounds |
| * |
| * @v in AES input state |
| * @v out AES output state |
| * @v key Round keys |
| * @v rounds Number of rounds (must be odd) |
| * |
| * This function is deliberately marked as non-inlinable to ensure |
| * maximal availability of registers for GCC's register allocator, |
| * which has a tendency to otherwise spill performance-critical |
| * registers to the stack. |
| */ |
| static __attribute__ (( noinline )) void |
| aes_encrypt_rounds ( union aes_matrix *in, union aes_matrix *out, |
| const union aes_matrix *key, unsigned int rounds ) { |
| union aes_matrix *tmp; |
| |
| /* Perform intermediate rounds */ |
| do { |
| /* Perform one intermediate round */ |
| aes_round ( &aes_mixcolumns, AES_STRIDE_SHIFTROWS, |
| in, out, key++ ); |
| |
| /* Swap input and output states for next round */ |
| tmp = in; |
| in = out; |
| out = tmp; |
| |
| } while ( --rounds ); |
| } |
| |
| /** |
| * Perform decryption intermediate rounds |
| * |
| * @v in AES input state |
| * @v out AES output state |
| * @v key Round keys |
| * @v rounds Number of rounds (must be odd) |
| * |
| * As with aes_encrypt_rounds(), this function is deliberately marked |
| * as non-inlinable. |
| * |
| * This function could potentially use the same binary code as is used |
| * for encryption. To compensate for the difference between ShiftRows |
| * and InvShiftRows, half of the input byte offsets would have to be |
| * modifiable at runtime (half by an offset of +4/-4, half by an |
| * offset of -4/+4 for ShiftRows/InvShiftRows). This can be |
| * accomplished in x86 assembly within the number of available |
| * registers, but GCC's register allocator struggles to do so, |
| * resulting in a significant performance decrease due to registers |
| * being spilled to the stack. We therefore use two separate but very |
| * similar binary functions based on the same C source. |
| */ |
| static __attribute__ (( noinline )) void |
| aes_decrypt_rounds ( union aes_matrix *in, union aes_matrix *out, |
| const union aes_matrix *key, unsigned int rounds ) { |
| union aes_matrix *tmp; |
| |
| /* Perform intermediate rounds */ |
| do { |
| /* Perform one intermediate round */ |
| aes_round ( &aes_invmixcolumns, AES_STRIDE_INVSHIFTROWS, |
| in, out, key++ ); |
| |
| /* Swap input and output states for next round */ |
| tmp = in; |
| in = out; |
| out = tmp; |
| |
| } while ( --rounds ); |
| } |
| |
| /** |
| * Perform standalone AddRoundKey |
| * |
| * @v state AES state |
| * @v key AES round key |
| */ |
| static inline __attribute__ (( always_inline )) void |
| aes_addroundkey ( union aes_matrix *state, const union aes_matrix *key ) { |
| |
| state->column[0] ^= key->column[0]; |
| state->column[1] ^= key->column[1]; |
| state->column[2] ^= key->column[2]; |
| state->column[3] ^= key->column[3]; |
| } |
| |
| /** |
| * Perform final round |
| * |
| * @v table AES lookup table |
| * @v stride AES row shift stride |
| * @v in AES input state |
| * @v out AES output state |
| * @v key AES round key |
| */ |
| static void aes_final ( const struct aes_table *table, size_t stride, |
| const union aes_matrix *in, union aes_matrix *out, |
| const union aes_matrix *key ) { |
| const union aes_table_entry *entry; |
| unsigned int byte; |
| size_t out_offset; |
| size_t in_offset; |
| |
| /* Perform [Inv]ShiftRows and [Inv]SubBytes */ |
| for ( out_offset = 0, in_offset = 0 ; out_offset < 16 ; |
| out_offset++, in_offset = ( ( in_offset + stride ) & 0xf ) ) { |
| |
| /* Extract input byte (i.e. perform [Inv]ShiftRows) */ |
| byte = in->byte[in_offset]; |
| |
| /* Locate lookup table entry for this input byte |
| * (i.e. perform [Inv]SubBytes). |
| */ |
| entry = &table->entry[byte]; |
| |
| /* Store output byte */ |
| out->byte[out_offset] = entry->byte[0]; |
| } |
| |
| /* Perform AddRoundKey */ |
| aes_addroundkey ( out, key ); |
| } |
| |
| /** |
| * Encrypt data |
| * |
| * @v ctx Context |
| * @v src Data to encrypt |
| * @v dst Buffer for encrypted data |
| * @v len Length of data |
| */ |
| static void aes_encrypt ( void *ctx, const void *src, void *dst, size_t len ) { |
| struct aes_context *aes = ctx; |
| union aes_matrix buffer[2]; |
| union aes_matrix *in = &buffer[0]; |
| union aes_matrix *out = &buffer[1]; |
| unsigned int rounds = aes->rounds; |
| |
| /* Sanity check */ |
| assert ( len == sizeof ( *in ) ); |
| |
| /* Initialise input state */ |
| memcpy ( in, src, sizeof ( *in ) ); |
| |
| /* Perform initial round (AddRoundKey) */ |
| aes_addroundkey ( in, &aes->encrypt.key[0] ); |
| |
| /* Perform intermediate rounds (ShiftRows, SubBytes, |
| * MixColumns, AddRoundKey). |
| */ |
| aes_encrypt_rounds ( in, out, &aes->encrypt.key[1], ( rounds - 2 ) ); |
| in = out; |
| |
| /* Perform final round (ShiftRows, SubBytes, AddRoundKey) */ |
| out = dst; |
| aes_final ( &aes_mixcolumns, AES_STRIDE_SHIFTROWS, in, out, |
| &aes->encrypt.key[ rounds - 1 ] ); |
| } |
| |
| /** |
| * Decrypt data |
| * |
| * @v ctx Context |
| * @v src Data to decrypt |
| * @v dst Buffer for decrypted data |
| * @v len Length of data |
| */ |
| static void aes_decrypt ( void *ctx, const void *src, void *dst, size_t len ) { |
| struct aes_context *aes = ctx; |
| union aes_matrix buffer[2]; |
| union aes_matrix *in = &buffer[0]; |
| union aes_matrix *out = &buffer[1]; |
| unsigned int rounds = aes->rounds; |
| |
| /* Sanity check */ |
| assert ( len == sizeof ( *in ) ); |
| |
| /* Initialise input state */ |
| memcpy ( in, src, sizeof ( *in ) ); |
| |
| /* Perform initial round (AddRoundKey) */ |
| aes_addroundkey ( in, &aes->decrypt.key[0] ); |
| |
| /* Perform intermediate rounds (InvShiftRows, InvSubBytes, |
| * InvMixColumns, AddRoundKey). |
| */ |
| aes_decrypt_rounds ( in, out, &aes->decrypt.key[1], ( rounds - 2 ) ); |
| in = out; |
| |
| /* Perform final round (InvShiftRows, InvSubBytes, AddRoundKey) */ |
| out = dst; |
| aes_final ( &aes_invmixcolumns, AES_STRIDE_INVSHIFTROWS, in, out, |
| &aes->decrypt.key[ rounds - 1 ] ); |
| } |
| |
| /** |
| * Multiply a polynomial by (x) modulo (x^8 + x^4 + x^3 + x^2 + 1) in GF(2^8) |
| * |
| * @v poly Polynomial to be multiplied |
| * @ret result Result |
| */ |
| static __attribute__ (( const )) unsigned int aes_double ( unsigned int poly ) { |
| |
| /* Multiply polynomial by (x), placing the resulting x^8 |
| * coefficient in the LSB (i.e. rotate byte left by one). |
| */ |
| poly = rol8 ( poly, 1 ); |
| |
| /* If coefficient of x^8 (in LSB) is non-zero, then reduce by |
| * subtracting (x^8 + x^4 + x^3 + x^2 + 1) in GF(2^8). |
| */ |
| if ( poly & 0x01 ) { |
| poly ^= 0x01; /* Subtract x^8 (currently in LSB) */ |
| poly ^= 0x1b; /* Subtract (x^4 + x^3 + x^2 + 1) */ |
| } |
| |
| return poly; |
| } |
| |
| /** |
| * Fill in MixColumns lookup table entry |
| * |
| * @v entry AES lookup table entry for scalar multiplicand |
| * |
| * The MixColumns lookup table vector multiplier is {1,1,1,3,2,1,1,3}. |
| */ |
| static void aes_mixcolumns_entry ( union aes_table_entry *entry ) { |
| unsigned int scalar_x_1; |
| unsigned int scalar_x; |
| unsigned int scalar; |
| |
| /* Retrieve scalar multiplicand */ |
| scalar = entry->byte[0]; |
| entry->byte[1] = scalar; |
| entry->byte[2] = scalar; |
| entry->byte[5] = scalar; |
| entry->byte[6] = scalar; |
| |
| /* Calculate scalar multiplied by (x) */ |
| scalar_x = aes_double ( scalar ); |
| entry->byte[4] = scalar_x; |
| |
| /* Calculate scalar multiplied by (x + 1) */ |
| scalar_x_1 = ( scalar_x ^ scalar ); |
| entry->byte[3] = scalar_x_1; |
| entry->byte[7] = scalar_x_1; |
| } |
| |
| /** |
| * Fill in InvMixColumns lookup table entry |
| * |
| * @v entry AES lookup table entry for scalar multiplicand |
| * |
| * The InvMixColumns lookup table vector multiplier is {1,9,13,11,14,9,13,11}. |
| */ |
| static void aes_invmixcolumns_entry ( union aes_table_entry *entry ) { |
| unsigned int scalar_x3_x2_x; |
| unsigned int scalar_x3_x2_1; |
| unsigned int scalar_x3_x2; |
| unsigned int scalar_x3_x_1; |
| unsigned int scalar_x3_1; |
| unsigned int scalar_x3; |
| unsigned int scalar_x2; |
| unsigned int scalar_x; |
| unsigned int scalar; |
| |
| /* Retrieve scalar multiplicand */ |
| scalar = entry->byte[0]; |
| |
| /* Calculate scalar multiplied by (x) */ |
| scalar_x = aes_double ( scalar ); |
| |
| /* Calculate scalar multiplied by (x^2) */ |
| scalar_x2 = aes_double ( scalar_x ); |
| |
| /* Calculate scalar multiplied by (x^3) */ |
| scalar_x3 = aes_double ( scalar_x2 ); |
| |
| /* Calculate scalar multiplied by (x^3 + 1) */ |
| scalar_x3_1 = ( scalar_x3 ^ scalar ); |
| entry->byte[1] = scalar_x3_1; |
| entry->byte[5] = scalar_x3_1; |
| |
| /* Calculate scalar multiplied by (x^3 + x + 1) */ |
| scalar_x3_x_1 = ( scalar_x3_1 ^ scalar_x ); |
| entry->byte[3] = scalar_x3_x_1; |
| entry->byte[7] = scalar_x3_x_1; |
| |
| /* Calculate scalar multiplied by (x^3 + x^2) */ |
| scalar_x3_x2 = ( scalar_x3 ^ scalar_x2 ); |
| |
| /* Calculate scalar multiplied by (x^3 + x^2 + 1) */ |
| scalar_x3_x2_1 = ( scalar_x3_x2 ^ scalar ); |
| entry->byte[2] = scalar_x3_x2_1; |
| entry->byte[6] = scalar_x3_x2_1; |
| |
| /* Calculate scalar multiplied by (x^3 + x^2 + x) */ |
| scalar_x3_x2_x = ( scalar_x3_x2 ^ scalar_x ); |
| entry->byte[4] = scalar_x3_x2_x; |
| } |
| |
| /** |
| * Generate AES lookup tables |
| * |
| */ |
| static void aes_generate ( void ) { |
| union aes_table_entry *entry; |
| union aes_table_entry *inventry; |
| unsigned int poly = 0x01; |
| unsigned int invpoly = 0x01; |
| unsigned int transformed; |
| unsigned int i; |
| |
| /* Iterate over non-zero values of GF(2^8) using generator (x + 1) */ |
| do { |
| |
| /* Multiply polynomial by (x + 1) */ |
| poly ^= aes_double ( poly ); |
| |
| /* Divide inverse polynomial by (x + 1). This code |
| * fragment is taken directly from the Wikipedia page |
| * on the Rijndael S-box. An explanation of why it |
| * works would be greatly appreciated. |
| */ |
| invpoly ^= ( invpoly << 1 ); |
| invpoly ^= ( invpoly << 2 ); |
| invpoly ^= ( invpoly << 4 ); |
| if ( invpoly & 0x80 ) |
| invpoly ^= 0x09; |
| invpoly &= 0xff; |
| |
| /* Apply affine transformation */ |
| transformed = ( 0x63 ^ invpoly ^ rol8 ( invpoly, 1 ) ^ |
| rol8 ( invpoly, 2 ) ^ rol8 ( invpoly, 3 ) ^ |
| rol8 ( invpoly, 4 ) ); |
| |
| /* Populate S-box (within MixColumns lookup table) */ |
| aes_mixcolumns.entry[poly].byte[0] = transformed; |
| |
| } while ( poly != 0x01 ); |
| |
| /* Populate zeroth S-box entry (which has no inverse) */ |
| aes_mixcolumns.entry[0].byte[0] = 0x63; |
| |
| /* Fill in MixColumns and InvMixColumns lookup tables */ |
| for ( i = 0 ; i < 256 ; i++ ) { |
| |
| /* Fill in MixColumns lookup table entry */ |
| entry = &aes_mixcolumns.entry[i]; |
| aes_mixcolumns_entry ( entry ); |
| |
| /* Populate inverse S-box (within InvMixColumns lookup table) */ |
| inventry = &aes_invmixcolumns.entry[ entry->byte[0] ]; |
| inventry->byte[0] = i; |
| |
| /* Fill in InvMixColumns lookup table entry */ |
| aes_invmixcolumns_entry ( inventry ); |
| } |
| } |
| |
| /** |
| * Rotate key column |
| * |
| * @v column Key column |
| * @ret column Updated key column |
| */ |
| static inline __attribute__ (( always_inline )) uint32_t |
| aes_key_rotate ( uint32_t column ) { |
| |
| return ( ( __BYTE_ORDER == __LITTLE_ENDIAN ) ? |
| ror32 ( column, 8 ) : rol32 ( column, 8 ) ); |
| } |
| |
| /** |
| * Apply S-box to key column |
| * |
| * @v column Key column |
| * @ret column Updated key column |
| */ |
| static uint32_t aes_key_sbox ( uint32_t column ) { |
| unsigned int i; |
| uint8_t byte; |
| |
| for ( i = 0 ; i < 4 ; i++ ) { |
| byte = ( column & 0xff ); |
| byte = aes_mixcolumns.entry[byte].byte[0]; |
| column = ( ( column & ~0xff ) | byte ); |
| column = rol32 ( column, 8 ); |
| } |
| return column; |
| } |
| |
| /** |
| * Apply schedule round constant to key column |
| * |
| * @v column Key column |
| * @v rcon Round constant |
| * @ret column Updated key column |
| */ |
| static inline __attribute__ (( always_inline )) uint32_t |
| aes_key_rcon ( uint32_t column, unsigned int rcon ) { |
| |
| return ( ( __BYTE_ORDER == __LITTLE_ENDIAN ) ? |
| ( column ^ rcon ) : ( column ^ ( rcon << 24 ) ) ); |
| } |
| |
| /** |
| * Set key |
| * |
| * @v ctx Context |
| * @v key Key |
| * @v keylen Key length |
| * @ret rc Return status code |
| */ |
| static int aes_setkey ( void *ctx, const void *key, size_t keylen ) { |
| struct aes_context *aes = ctx; |
| union aes_matrix *enc; |
| union aes_matrix *dec; |
| union aes_matrix temp; |
| union aes_matrix zero; |
| unsigned int rcon = 0x01; |
| unsigned int rounds; |
| size_t offset = 0; |
| uint32_t *prev; |
| uint32_t *next; |
| uint32_t *end; |
| uint32_t tmp; |
| |
| /* Generate lookup tables, if not already done */ |
| if ( ! aes_mixcolumns.entry[0].byte[0] ) |
| aes_generate(); |
| |
| /* Validate key length and calculate number of intermediate rounds */ |
| switch ( keylen ) { |
| case ( 128 / 8 ) : |
| rounds = 11; |
| break; |
| case ( 192 / 8 ) : |
| rounds = 13; |
| break; |
| case ( 256 / 8 ) : |
| rounds = 15; |
| break; |
| default: |
| DBGC ( aes, "AES %p unsupported key length (%zd bits)\n", |
| aes, ( keylen * 8 ) ); |
| return -EINVAL; |
| } |
| aes->rounds = rounds; |
| enc = aes->encrypt.key; |
| end = enc[rounds].column; |
| |
| /* Copy raw key */ |
| memcpy ( enc, key, keylen ); |
| prev = enc->column; |
| next = ( ( ( void * ) prev ) + keylen ); |
| tmp = next[-1]; |
| |
| /* Construct expanded key */ |
| while ( next < end ) { |
| |
| /* If this is the first column of an expanded key |
| * block, or the middle column of an AES-256 key |
| * block, then apply the S-box. |
| */ |
| if ( ( offset == 0 ) || ( ( offset | keylen ) == 48 ) ) |
| tmp = aes_key_sbox ( tmp ); |
| |
| /* If this is the first column of an expanded key |
| * block then rotate and apply the round constant. |
| */ |
| if ( offset == 0 ) { |
| tmp = aes_key_rotate ( tmp ); |
| tmp = aes_key_rcon ( tmp, rcon ); |
| rcon = aes_double ( rcon ); |
| } |
| |
| /* XOR with previous key column */ |
| tmp ^= *prev; |
| |
| /* Store column */ |
| *next = tmp; |
| |
| /* Move to next column */ |
| offset += sizeof ( *next ); |
| if ( offset == keylen ) |
| offset = 0; |
| next++; |
| prev++; |
| } |
| DBGC2 ( aes, "AES %p expanded %zd-bit key:\n", aes, ( keylen * 8 ) ); |
| DBGC2_HDA ( aes, 0, &aes->encrypt, ( rounds * sizeof ( *enc ) ) ); |
| |
| /* Convert to decryption key */ |
| memset ( &zero, 0, sizeof ( zero ) ); |
| dec = &aes->decrypt.key[ rounds - 1 ]; |
| memcpy ( dec--, enc++, sizeof ( *dec ) ); |
| while ( dec > aes->decrypt.key ) { |
| /* Perform InvMixColumns (by reusing the encryption |
| * final-round code to perform ShiftRows+SubBytes and |
| * reusing the decryption intermediate-round code to |
| * perform InvShiftRows+InvSubBytes+InvMixColumns, all |
| * with a zero encryption key). |
| */ |
| aes_final ( &aes_mixcolumns, AES_STRIDE_SHIFTROWS, |
| enc++, &temp, &zero ); |
| aes_decrypt_rounds ( &temp, dec--, &zero, 1 ); |
| } |
| memcpy ( dec--, enc++, sizeof ( *dec ) ); |
| DBGC2 ( aes, "AES %p inverted %zd-bit key:\n", aes, ( keylen * 8 ) ); |
| DBGC2_HDA ( aes, 0, &aes->decrypt, ( rounds * sizeof ( *dec ) ) ); |
| |
| return 0; |
| } |
| |
| /** Basic AES algorithm */ |
| struct cipher_algorithm aes_algorithm = { |
| .name = "aes", |
| .ctxsize = sizeof ( struct aes_context ), |
| .blocksize = AES_BLOCKSIZE, |
| .alignsize = 0, |
| .authsize = 0, |
| .setkey = aes_setkey, |
| .setiv = cipher_null_setiv, |
| .encrypt = aes_encrypt, |
| .decrypt = aes_decrypt, |
| .auth = cipher_null_auth, |
| }; |
| |
| /* AES in Electronic Codebook mode */ |
| ECB_CIPHER ( aes_ecb, aes_ecb_algorithm, |
| aes_algorithm, struct aes_context, AES_BLOCKSIZE ); |
| |
| /* AES in Cipher Block Chaining mode */ |
| CBC_CIPHER ( aes_cbc, aes_cbc_algorithm, |
| aes_algorithm, struct aes_context, AES_BLOCKSIZE ); |
| |
| /* AES in Galois/Counter mode */ |
| GCM_CIPHER ( aes_gcm, aes_gcm_algorithm, |
| aes_algorithm, struct aes_context, AES_BLOCKSIZE ); |