summaryrefslogtreecommitdiffstats
path: root/qemu/roms/ipxe/src/crypto/aes.c
diff options
context:
space:
mode:
Diffstat (limited to 'qemu/roms/ipxe/src/crypto/aes.c')
-rw-r--r--qemu/roms/ipxe/src/crypto/aes.c808
1 files changed, 0 insertions, 808 deletions
diff --git a/qemu/roms/ipxe/src/crypto/aes.c b/qemu/roms/ipxe/src/crypto/aes.c
deleted file mode 100644
index b9e206bfb..000000000
--- a/qemu/roms/ipxe/src/crypto/aes.c
+++ /dev/null
@@ -1,808 +0,0 @@
-/*
- * Copyright (C) 2015 Michael Brown <mbrown@fensystems.co.uk>.
- *
- * This program is free software; you can redistribute it and/or
- * modify it under the terms of the GNU General Public License as
- * published by the Free Software Foundation; either version 2 of the
- * License, or any later version.
- *
- * This program is distributed in the hope that it will be useful, but
- * WITHOUT ANY WARRANTY; without even the implied warranty of
- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
- * General Public License for more details.
- *
- * You should have received a copy of the GNU General Public License
- * along with this program; if not, write to the Free Software
- * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA
- * 02110-1301, USA.
- *
- * You can also choose to distribute this program under the terms of
- * the Unmodified Binary Distribution Licence (as given in the file
- * COPYING.UBDL), provided that you have satisfied its requirements.
- */
-
-FILE_LICENCE ( GPL2_OR_LATER_OR_UBDL );
-
-/** @file
- *
- * AES algorithm
- *
- */
-
-#include <stdint.h>
-#include <string.h>
-#include <errno.h>
-#include <assert.h>
-#include <byteswap.h>
-#include <ipxe/rotate.h>
-#include <ipxe/crypto.h>
-#include <ipxe/ecb.h>
-#include <ipxe/cbc.h>
-#include <ipxe/aes.h>
-
-/** AES strides
- *
- * These are the strides (modulo 16) used to walk through the AES
- * input state bytes in order of byte position after [Inv]ShiftRows.
- */
-enum aes_stride {
- /** Input stride for ShiftRows
- *
- * 0 4 8 c
- * \ \ \
- * 1 5 9 d
- * \ \ \
- * 2 6 a e
- * \ \ \
- * 3 7 b f
- */
- AES_STRIDE_SHIFTROWS = +5,
- /** Input stride for InvShiftRows
- *
- * 0 4 8 c
- * / / /
- * 1 5 9 d
- * / / /
- * 2 6 a e
- * / / /
- * 3 7 b f
- */
- AES_STRIDE_INVSHIFTROWS = -3,
-};
-
-/** A single AES lookup table entry
- *
- * This represents the product (in the Galois field GF(2^8)) of an
- * eight-byte vector multiplier with a single scalar multiplicand.
- *
- * The vector multipliers used for AES will be {1,1,1,3,2,1,1,3} for
- * MixColumns and {1,9,13,11,14,9,13,11} for InvMixColumns. This
- * allows for the result of multiplying any single column of the
- * [Inv]MixColumns matrix by a scalar value to be obtained simply by
- * extracting the relevant four-byte subset from the lookup table
- * entry.
- *
- * For example, to find the result of multiplying the second column of
- * the MixColumns matrix by the scalar value 0x80:
- *
- * MixColumns column[0]: { 2, 1, 1, 3 }
- * MixColumns column[1]: { 3, 2, 1, 1 }
- * MixColumns column[2]: { 1, 3, 2, 1 }
- * MixColumns column[3]: { 1, 1, 3, 2 }
- * Vector multiplier: { 1, 1, 1, 3, 2, 1, 1, 3 }
- * Scalar multiplicand: 0x80
- * Lookup table entry: { 0x80, 0x80, 0x80, 0x9b, 0x1b, 0x80, 0x80, 0x9b }
- *
- * The second column of the MixColumns matrix is {3,2,1,1}. The
- * product of this column with the scalar value 0x80 can be obtained
- * by extracting the relevant four-byte subset of the lookup table
- * entry:
- *
- * MixColumns column[1]: { 3, 2, 1, 1 }
- * Vector multiplier: { 1, 1, 1, 3, 2, 1, 1, 3 }
- * Lookup table entry: { 0x80, 0x80, 0x80, 0x9b, 0x1b, 0x80, 0x80, 0x9b }
- * Product: { 0x9b, 0x1b, 0x80, 0x80 }
- *
- * The column lookups require only seven bytes of the eight-byte
- * entry: the remaining (first) byte is used to hold the scalar
- * multiplicand itself (i.e. the first byte of the vector multiplier
- * is always chosen to be 1).
- */
-union aes_table_entry {
- /** Viewed as an array of bytes */
- uint8_t byte[8];
-} __attribute__ (( packed ));
-
-/** An AES lookup table
- *
- * This represents the products (in the Galois field GF(2^8)) of a
- * constant eight-byte vector multiplier with all possible 256 scalar
- * multiplicands.
- *
- * The entries are indexed by the AES [Inv]SubBytes S-box output
- * values (denoted S(N)). This allows for the result of multiplying
- * any single column of the [Inv]MixColumns matrix by S(N) to be
- * obtained simply by extracting the relevant four-byte subset from
- * the Nth table entry. For example:
- *
- * Input byte (N): 0x3a
- * SubBytes output S(N): 0x80
- * MixColumns column[1]: { 3, 2, 1, 1 }
- * Vector multiplier: { 1, 1, 1, 3, 2, 1, 1, 3 }
- * Table entry[0x3a]: { 0x80, 0x80, 0x80, 0x9b, 0x1b, 0x80, 0x80, 0x9b }
- * Product: { 0x9b, 0x1b, 0x80, 0x80 }
- *
- * Since the first byte of the eight-byte vector multiplier is always
- * chosen to be 1, the value of S(N) may be lookup up by extracting
- * the first byte of the Nth table entry.
- */
-struct aes_table {
- /** Table entries, indexed by S(N) */
- union aes_table_entry entry[256];
-} __attribute__ (( aligned ( 8 ) ));
-
-/** AES MixColumns lookup table */
-static struct aes_table aes_mixcolumns;
-
-/** AES InvMixColumns lookup table */
-static struct aes_table aes_invmixcolumns;
-
-/**
- * Multiply [Inv]MixColumns matrix column by scalar multiplicand
- *
- * @v entry AES lookup table entry for scalar multiplicand
- * @v column [Inv]MixColumns matrix column index
- * @ret product Product of matrix column with scalar multiplicand
- */
-static inline __attribute__ (( always_inline )) uint32_t
-aes_entry_column ( const union aes_table_entry *entry, unsigned int column ) {
- const union {
- uint8_t byte;
- uint32_t column;
- } __attribute__ (( may_alias )) *product;
-
- /* Locate relevant four-byte subset */
- product = container_of ( &entry->byte[ 4 - column ],
- typeof ( *product ), byte );
-
- /* Extract this four-byte subset */
- return product->column;
-}
-
-/**
- * Multiply [Inv]MixColumns matrix column by S-boxed input byte
- *
- * @v table AES lookup table
- * @v stride AES row shift stride
- * @v in AES input state
- * @v offset Output byte offset (after [Inv]ShiftRows)
- * @ret product Product of matrix column with S(input byte)
- *
- * Note that the specified offset is not the offset of the input byte;
- * it is the offset of the output byte which corresponds to the input
- * byte. This output byte offset is used to calculate both the input
- * byte offset and to select the appropriate matric column.
- *
- * With a compile-time constant offset, this function will optimise
- * down to a single "movzbl" (to extract the input byte) and will
- * generate a single x86 memory reference expression which can then be
- * used directly within a single "xorl" instruction.
- */
-static inline __attribute__ (( always_inline )) uint32_t
-aes_column ( const struct aes_table *table, size_t stride,
- const union aes_matrix *in, size_t offset ) {
- const union aes_table_entry *entry;
- unsigned int byte;
-
- /* Extract input byte corresponding to this output byte offset
- * (i.e. perform [Inv]ShiftRows).
- */
- byte = in->byte[ ( stride * offset ) & 0xf ];
-
- /* Locate lookup table entry for this input byte (i.e. perform
- * [Inv]SubBytes).
- */
- entry = &table->entry[byte];
-
- /* Multiply appropriate matrix column by this input byte
- * (i.e. perform [Inv]MixColumns).
- */
- return aes_entry_column ( entry, ( offset & 0x3 ) );
-}
-
-/**
- * Calculate intermediate round output column
- *
- * @v table AES lookup table
- * @v stride AES row shift stride
- * @v in AES input state
- * @v key AES round key
- * @v column Column index
- * @ret output Output column value
- */
-static inline __attribute__ (( always_inline )) uint32_t
-aes_output ( const struct aes_table *table, size_t stride,
- const union aes_matrix *in, const union aes_matrix *key,
- unsigned int column ) {
- size_t offset = ( column * 4 );
-
- /* Perform [Inv]ShiftRows, [Inv]SubBytes, [Inv]MixColumns, and
- * AddRoundKey for this column. The loop is unrolled to allow
- * for the required compile-time constant optimisations.
- */
- return ( aes_column ( table, stride, in, ( offset + 0 ) ) ^
- aes_column ( table, stride, in, ( offset + 1 ) ) ^
- aes_column ( table, stride, in, ( offset + 2 ) ) ^
- aes_column ( table, stride, in, ( offset + 3 ) ) ^
- key->column[column] );
-}
-
-/**
- * Perform a single intermediate round
- *
- * @v table AES lookup table
- * @v stride AES row shift stride
- * @v in AES input state
- * @v out AES output state
- * @v key AES round key
- */
-static inline __attribute__ (( always_inline )) void
-aes_round ( const struct aes_table *table, size_t stride,
- const union aes_matrix *in, union aes_matrix *out,
- const union aes_matrix *key ) {
-
- /* Perform [Inv]ShiftRows, [Inv]SubBytes, [Inv]MixColumns, and
- * AddRoundKey for all columns. The loop is unrolled to allow
- * for the required compile-time constant optimisations.
- */
- out->column[0] = aes_output ( table, stride, in, key, 0 );
- out->column[1] = aes_output ( table, stride, in, key, 1 );
- out->column[2] = aes_output ( table, stride, in, key, 2 );
- out->column[3] = aes_output ( table, stride, in, key, 3 );
-}
-
-/**
- * Perform encryption intermediate rounds
- *
- * @v in AES input state
- * @v out AES output state
- * @v key Round keys
- * @v rounds Number of rounds (must be odd)
- *
- * This function is deliberately marked as non-inlinable to ensure
- * maximal availability of registers for GCC's register allocator,
- * which has a tendency to otherwise spill performance-critical
- * registers to the stack.
- */
-static __attribute__ (( noinline )) void
-aes_encrypt_rounds ( union aes_matrix *in, union aes_matrix *out,
- const union aes_matrix *key, unsigned int rounds ) {
- union aes_matrix *tmp;
-
- /* Perform intermediate rounds */
- do {
- /* Perform one intermediate round */
- aes_round ( &aes_mixcolumns, AES_STRIDE_SHIFTROWS,
- in, out, key++ );
-
- /* Swap input and output states for next round */
- tmp = in;
- in = out;
- out = tmp;
-
- } while ( --rounds );
-}
-
-/**
- * Perform decryption intermediate rounds
- *
- * @v in AES input state
- * @v out AES output state
- * @v key Round keys
- * @v rounds Number of rounds (must be odd)
- *
- * As with aes_encrypt_rounds(), this function is deliberately marked
- * as non-inlinable.
- *
- * This function could potentially use the same binary code as is used
- * for encryption. To compensate for the difference between ShiftRows
- * and InvShiftRows, half of the input byte offsets would have to be
- * modifiable at runtime (half by an offset of +4/-4, half by an
- * offset of -4/+4 for ShiftRows/InvShiftRows). This can be
- * accomplished in x86 assembly within the number of available
- * registers, but GCC's register allocator struggles to do so,
- * resulting in a significant performance decrease due to registers
- * being spilled to the stack. We therefore use two separate but very
- * similar binary functions based on the same C source.
- */
-static __attribute__ (( noinline )) void
-aes_decrypt_rounds ( union aes_matrix *in, union aes_matrix *out,
- const union aes_matrix *key, unsigned int rounds ) {
- union aes_matrix *tmp;
-
- /* Perform intermediate rounds */
- do {
- /* Perform one intermediate round */
- aes_round ( &aes_invmixcolumns, AES_STRIDE_INVSHIFTROWS,
- in, out, key++ );
-
- /* Swap input and output states for next round */
- tmp = in;
- in = out;
- out = tmp;
-
- } while ( --rounds );
-}
-
-/**
- * Perform standalone AddRoundKey
- *
- * @v state AES state
- * @v key AES round key
- */
-static inline __attribute__ (( always_inline )) void
-aes_addroundkey ( union aes_matrix *state, const union aes_matrix *key ) {
-
- state->column[0] ^= key->column[0];
- state->column[1] ^= key->column[1];
- state->column[2] ^= key->column[2];
- state->column[3] ^= key->column[3];
-}
-
-/**
- * Perform final round
- *
- * @v table AES lookup table
- * @v stride AES row shift stride
- * @v in AES input state
- * @v out AES output state
- * @v key AES round key
- */
-static void aes_final ( const struct aes_table *table, size_t stride,
- const union aes_matrix *in, union aes_matrix *out,
- const union aes_matrix *key ) {
- const union aes_table_entry *entry;
- unsigned int byte;
- size_t out_offset;
- size_t in_offset;
-
- /* Perform [Inv]ShiftRows and [Inv]SubBytes */
- for ( out_offset = 0, in_offset = 0 ; out_offset < 16 ;
- out_offset++, in_offset = ( ( in_offset + stride ) & 0xf ) ) {
-
- /* Extract input byte (i.e. perform [Inv]ShiftRows) */
- byte = in->byte[in_offset];
-
- /* Locate lookup table entry for this input byte
- * (i.e. perform [Inv]SubBytes).
- */
- entry = &table->entry[byte];
-
- /* Store output byte */
- out->byte[out_offset] = entry->byte[0];
- }
-
- /* Perform AddRoundKey */
- aes_addroundkey ( out, key );
-}
-
-/**
- * Encrypt data
- *
- * @v ctx Context
- * @v src Data to encrypt
- * @v dst Buffer for encrypted data
- * @v len Length of data
- */
-static void aes_encrypt ( void *ctx, const void *src, void *dst, size_t len ) {
- struct aes_context *aes = ctx;
- union aes_matrix buffer[2];
- union aes_matrix *in = &buffer[0];
- union aes_matrix *out = &buffer[1];
- unsigned int rounds = aes->rounds;
-
- /* Sanity check */
- assert ( len == sizeof ( *in ) );
-
- /* Initialise input state */
- memcpy ( in, src, sizeof ( *in ) );
-
- /* Perform initial round (AddRoundKey) */
- aes_addroundkey ( in, &aes->encrypt.key[0] );
-
- /* Perform intermediate rounds (ShiftRows, SubBytes,
- * MixColumns, AddRoundKey).
- */
- aes_encrypt_rounds ( in, out, &aes->encrypt.key[1], ( rounds - 2 ) );
- in = out;
-
- /* Perform final round (ShiftRows, SubBytes, AddRoundKey) */
- out = dst;
- aes_final ( &aes_mixcolumns, AES_STRIDE_SHIFTROWS, in, out,
- &aes->encrypt.key[ rounds - 1 ] );
-}
-
-/**
- * Decrypt data
- *
- * @v ctx Context
- * @v src Data to decrypt
- * @v dst Buffer for decrypted data
- * @v len Length of data
- */
-static void aes_decrypt ( void *ctx, const void *src, void *dst, size_t len ) {
- struct aes_context *aes = ctx;
- union aes_matrix buffer[2];
- union aes_matrix *in = &buffer[0];
- union aes_matrix *out = &buffer[1];
- unsigned int rounds = aes->rounds;
-
- /* Sanity check */
- assert ( len == sizeof ( *in ) );
-
- /* Initialise input state */
- memcpy ( in, src, sizeof ( *in ) );
-
- /* Perform initial round (AddRoundKey) */
- aes_addroundkey ( in, &aes->decrypt.key[0] );
-
- /* Perform intermediate rounds (InvShiftRows, InvSubBytes,
- * InvMixColumns, AddRoundKey).
- */
- aes_decrypt_rounds ( in, out, &aes->decrypt.key[1], ( rounds - 2 ) );
- in = out;
-
- /* Perform final round (InvShiftRows, InvSubBytes, AddRoundKey) */
- out = dst;
- aes_final ( &aes_invmixcolumns, AES_STRIDE_INVSHIFTROWS, in, out,
- &aes->decrypt.key[ rounds - 1 ] );
-}
-
-/**
- * Multiply a polynomial by (x) modulo (x^8 + x^4 + x^3 + x^2 + 1) in GF(2^8)
- *
- * @v poly Polynomial to be multiplied
- * @ret result Result
- */
-static __attribute__ (( const )) unsigned int aes_double ( unsigned int poly ) {
-
- /* Multiply polynomial by (x), placing the resulting x^8
- * coefficient in the LSB (i.e. rotate byte left by one).
- */
- poly = rol8 ( poly, 1 );
-
- /* If coefficient of x^8 (in LSB) is non-zero, then reduce by
- * subtracting (x^8 + x^4 + x^3 + x^2 + 1) in GF(2^8).
- */
- if ( poly & 0x01 ) {
- poly ^= 0x01; /* Subtract x^8 (currently in LSB) */
- poly ^= 0x1b; /* Subtract (x^4 + x^3 + x^2 + 1) */
- }
-
- return poly;
-}
-
-/**
- * Fill in MixColumns lookup table entry
- *
- * @v entry AES lookup table entry for scalar multiplicand
- *
- * The MixColumns lookup table vector multiplier is {1,1,1,3,2,1,1,3}.
- */
-static void aes_mixcolumns_entry ( union aes_table_entry *entry ) {
- unsigned int scalar_x_1;
- unsigned int scalar_x;
- unsigned int scalar;
-
- /* Retrieve scalar multiplicand */
- scalar = entry->byte[0];
- entry->byte[1] = scalar;
- entry->byte[2] = scalar;
- entry->byte[5] = scalar;
- entry->byte[6] = scalar;
-
- /* Calculate scalar multiplied by (x) */
- scalar_x = aes_double ( scalar );
- entry->byte[4] = scalar_x;
-
- /* Calculate scalar multiplied by (x + 1) */
- scalar_x_1 = ( scalar_x ^ scalar );
- entry->byte[3] = scalar_x_1;
- entry->byte[7] = scalar_x_1;
-}
-
-/**
- * Fill in InvMixColumns lookup table entry
- *
- * @v entry AES lookup table entry for scalar multiplicand
- *
- * The InvMixColumns lookup table vector multiplier is {1,9,13,11,14,9,13,11}.
- */
-static void aes_invmixcolumns_entry ( union aes_table_entry *entry ) {
- unsigned int scalar_x3_x2_x;
- unsigned int scalar_x3_x2_1;
- unsigned int scalar_x3_x2;
- unsigned int scalar_x3_x_1;
- unsigned int scalar_x3_1;
- unsigned int scalar_x3;
- unsigned int scalar_x2;
- unsigned int scalar_x;
- unsigned int scalar;
-
- /* Retrieve scalar multiplicand */
- scalar = entry->byte[0];
-
- /* Calculate scalar multiplied by (x) */
- scalar_x = aes_double ( scalar );
-
- /* Calculate scalar multiplied by (x^2) */
- scalar_x2 = aes_double ( scalar_x );
-
- /* Calculate scalar multiplied by (x^3) */
- scalar_x3 = aes_double ( scalar_x2 );
-
- /* Calculate scalar multiplied by (x^3 + 1) */
- scalar_x3_1 = ( scalar_x3 ^ scalar );
- entry->byte[1] = scalar_x3_1;
- entry->byte[5] = scalar_x3_1;
-
- /* Calculate scalar multiplied by (x^3 + x + 1) */
- scalar_x3_x_1 = ( scalar_x3_1 ^ scalar_x );
- entry->byte[3] = scalar_x3_x_1;
- entry->byte[7] = scalar_x3_x_1;
-
- /* Calculate scalar multiplied by (x^3 + x^2) */
- scalar_x3_x2 = ( scalar_x3 ^ scalar_x2 );
-
- /* Calculate scalar multiplied by (x^3 + x^2 + 1) */
- scalar_x3_x2_1 = ( scalar_x3_x2 ^ scalar );
- entry->byte[2] = scalar_x3_x2_1;
- entry->byte[6] = scalar_x3_x2_1;
-
- /* Calculate scalar multiplied by (x^3 + x^2 + x) */
- scalar_x3_x2_x = ( scalar_x3_x2 ^ scalar_x );
- entry->byte[4] = scalar_x3_x2_x;
-}
-
-/**
- * Generate AES lookup tables
- *
- */
-static void aes_generate ( void ) {
- union aes_table_entry *entry;
- union aes_table_entry *inventry;
- unsigned int poly = 0x01;
- unsigned int invpoly = 0x01;
- unsigned int transformed;
- unsigned int i;
-
- /* Iterate over non-zero values of GF(2^8) using generator (x + 1) */
- do {
-
- /* Multiply polynomial by (x + 1) */
- poly ^= aes_double ( poly );
-
- /* Divide inverse polynomial by (x + 1). This code
- * fragment is taken directly from the Wikipedia page
- * on the Rijndael S-box. An explanation of why it
- * works would be greatly appreciated.
- */
- invpoly ^= ( invpoly << 1 );
- invpoly ^= ( invpoly << 2 );
- invpoly ^= ( invpoly << 4 );
- if ( invpoly & 0x80 )
- invpoly ^= 0x09;
- invpoly &= 0xff;
-
- /* Apply affine transformation */
- transformed = ( 0x63 ^ invpoly ^ rol8 ( invpoly, 1 ) ^
- rol8 ( invpoly, 2 ) ^ rol8 ( invpoly, 3 ) ^
- rol8 ( invpoly, 4 ) );
-
- /* Populate S-box (within MixColumns lookup table) */
- aes_mixcolumns.entry[poly].byte[0] = transformed;
-
- } while ( poly != 0x01 );
-
- /* Populate zeroth S-box entry (which has no inverse) */
- aes_mixcolumns.entry[0].byte[0] = 0x63;
-
- /* Fill in MixColumns and InvMixColumns lookup tables */
- for ( i = 0 ; i < 256 ; i++ ) {
-
- /* Fill in MixColumns lookup table entry */
- entry = &aes_mixcolumns.entry[i];
- aes_mixcolumns_entry ( entry );
-
- /* Populate inverse S-box (within InvMixColumns lookup table) */
- inventry = &aes_invmixcolumns.entry[ entry->byte[0] ];
- inventry->byte[0] = i;
-
- /* Fill in InvMixColumns lookup table entry */
- aes_invmixcolumns_entry ( inventry );
- }
-}
-
-/**
- * Rotate key column
- *
- * @v column Key column
- * @ret column Updated key column
- */
-static inline __attribute__ (( always_inline )) uint32_t
-aes_key_rotate ( uint32_t column ) {
-
- return ( ( __BYTE_ORDER == __LITTLE_ENDIAN ) ?
- ror32 ( column, 8 ) : rol32 ( column, 8 ) );
-}
-
-/**
- * Apply S-box to key column
- *
- * @v column Key column
- * @ret column Updated key column
- */
-static uint32_t aes_key_sbox ( uint32_t column ) {
- unsigned int i;
- uint8_t byte;
-
- for ( i = 0 ; i < 4 ; i++ ) {
- byte = ( column & 0xff );
- byte = aes_mixcolumns.entry[byte].byte[0];
- column = ( ( column & ~0xff ) | byte );
- column = rol32 ( column, 8 );
- }
- return column;
-}
-
-/**
- * Apply schedule round constant to key column
- *
- * @v column Key column
- * @v rcon Round constant
- * @ret column Updated key column
- */
-static inline __attribute__ (( always_inline )) uint32_t
-aes_key_rcon ( uint32_t column, unsigned int rcon ) {
-
- return ( ( __BYTE_ORDER == __LITTLE_ENDIAN ) ?
- ( column ^ rcon ) : ( column ^ ( rcon << 24 ) ) );
-}
-
-/**
- * Set key
- *
- * @v ctx Context
- * @v key Key
- * @v keylen Key length
- * @ret rc Return status code
- */
-static int aes_setkey ( void *ctx, const void *key, size_t keylen ) {
- struct aes_context *aes = ctx;
- union aes_matrix *enc;
- union aes_matrix *dec;
- union aes_matrix temp;
- union aes_matrix zero;
- unsigned int rcon = 0x01;
- unsigned int rounds;
- size_t offset = 0;
- uint32_t *prev;
- uint32_t *next;
- uint32_t *end;
- uint32_t tmp;
-
- /* Generate lookup tables, if not already done */
- if ( ! aes_mixcolumns.entry[0].byte[0] )
- aes_generate();
-
- /* Validate key length and calculate number of intermediate rounds */
- switch ( keylen ) {
- case ( 128 / 8 ) :
- rounds = 11;
- break;
- case ( 192 / 8 ) :
- rounds = 13;
- break;
- case ( 256 / 8 ) :
- rounds = 15;
- break;
- default:
- DBGC ( aes, "AES %p unsupported key length (%zd bits)\n",
- aes, ( keylen * 8 ) );
- return -EINVAL;
- }
- aes->rounds = rounds;
- enc = aes->encrypt.key;
- end = enc[rounds].column;
-
- /* Copy raw key */
- memcpy ( enc, key, keylen );
- prev = enc->column;
- next = ( ( ( void * ) prev ) + keylen );
- tmp = next[-1];
-
- /* Construct expanded key */
- while ( next < end ) {
-
- /* If this is the first column of an expanded key
- * block, or the middle column of an AES-256 key
- * block, then apply the S-box.
- */
- if ( ( offset == 0 ) || ( ( offset | keylen ) == 48 ) )
- tmp = aes_key_sbox ( tmp );
-
- /* If this is the first column of an expanded key
- * block then rotate and apply the round constant.
- */
- if ( offset == 0 ) {
- tmp = aes_key_rotate ( tmp );
- tmp = aes_key_rcon ( tmp, rcon );
- rcon = aes_double ( rcon );
- }
-
- /* XOR with previous key column */
- tmp ^= *prev;
-
- /* Store column */
- *next = tmp;
-
- /* Move to next column */
- offset += sizeof ( *next );
- if ( offset == keylen )
- offset = 0;
- next++;
- prev++;
- }
- DBGC2 ( aes, "AES %p expanded %zd-bit key:\n", aes, ( keylen * 8 ) );
- DBGC2_HDA ( aes, 0, &aes->encrypt, ( rounds * sizeof ( *enc ) ) );
-
- /* Convert to decryption key */
- memset ( &zero, 0, sizeof ( zero ) );
- dec = &aes->decrypt.key[ rounds - 1 ];
- memcpy ( dec--, enc++, sizeof ( *dec ) );
- while ( dec > aes->decrypt.key ) {
- /* Perform InvMixColumns (by reusing the encryption
- * final-round code to perform ShiftRows+SubBytes and
- * reusing the decryption intermediate-round code to
- * perform InvShiftRows+InvSubBytes+InvMixColumns, all
- * with a zero encryption key).
- */
- aes_final ( &aes_mixcolumns, AES_STRIDE_SHIFTROWS,
- enc++, &temp, &zero );
- aes_decrypt_rounds ( &temp, dec--, &zero, 1 );
- }
- memcpy ( dec--, enc++, sizeof ( *dec ) );
- DBGC2 ( aes, "AES %p inverted %zd-bit key:\n", aes, ( keylen * 8 ) );
- DBGC2_HDA ( aes, 0, &aes->decrypt, ( rounds * sizeof ( *dec ) ) );
-
- return 0;
-}
-
-/**
- * Set initialisation vector
- *
- * @v ctx Context
- * @v iv Initialisation vector
- */
-static void aes_setiv ( void *ctx __unused, const void *iv __unused ) {
- /* Nothing to do */
-}
-
-/** Basic AES algorithm */
-struct cipher_algorithm aes_algorithm = {
- .name = "aes",
- .ctxsize = sizeof ( struct aes_context ),
- .blocksize = AES_BLOCKSIZE,
- .setkey = aes_setkey,
- .setiv = aes_setiv,
- .encrypt = aes_encrypt,
- .decrypt = aes_decrypt,
-};
-
-/* AES in Electronic Codebook mode */
-ECB_CIPHER ( aes_ecb, aes_ecb_algorithm,
- aes_algorithm, struct aes_context, AES_BLOCKSIZE );
-
-/* AES in Cipher Block Chaining mode */
-CBC_CIPHER ( aes_cbc, aes_cbc_algorithm,
- aes_algorithm, struct aes_context, AES_BLOCKSIZE );