blob: 36ada1be9d1c533b278349fa70ad3cc19eff0bf0 [file] [log] [blame]
/*
* Carry-less multiply operations.
* SPDX-License-Identifier: GPL-2.0-or-later
*
* Copyright (C) 2023 Linaro, Ltd.
*/
#include "qemu/osdep.h"
#include "crypto/clmul.h"
uint64_t clmul_8x8_low(uint64_t n, uint64_t m)
{
uint64_t r = 0;
for (int i = 0; i < 8; ++i) {
uint64_t mask = (n & 0x0101010101010101ull) * 0xff;
r ^= m & mask;
m = (m << 1) & 0xfefefefefefefefeull;
n >>= 1;
}
return r;
}
static uint64_t clmul_8x4_even_int(uint64_t n, uint64_t m)
{
uint64_t r = 0;
for (int i = 0; i < 8; ++i) {
uint64_t mask = (n & 0x0001000100010001ull) * 0xffff;
r ^= m & mask;
n >>= 1;
m <<= 1;
}
return r;
}
uint64_t clmul_8x4_even(uint64_t n, uint64_t m)
{
n &= 0x00ff00ff00ff00ffull;
m &= 0x00ff00ff00ff00ffull;
return clmul_8x4_even_int(n, m);
}
uint64_t clmul_8x4_odd(uint64_t n, uint64_t m)
{
return clmul_8x4_even(n >> 8, m >> 8);
}
static uint64_t unpack_8_to_16(uint64_t x)
{
return (x & 0x000000ff)
| ((x & 0x0000ff00) << 8)
| ((x & 0x00ff0000) << 16)
| ((x & 0xff000000) << 24);
}
uint64_t clmul_8x4_packed(uint32_t n, uint32_t m)
{
return clmul_8x4_even_int(unpack_8_to_16(n), unpack_8_to_16(m));
}
uint64_t clmul_16x2_even(uint64_t n, uint64_t m)
{
uint64_t r = 0;
n &= 0x0000ffff0000ffffull;
m &= 0x0000ffff0000ffffull;
for (int i = 0; i < 16; ++i) {
uint64_t mask = (n & 0x0000000100000001ull) * 0xffffffffull;
r ^= m & mask;
n >>= 1;
m <<= 1;
}
return r;
}
uint64_t clmul_16x2_odd(uint64_t n, uint64_t m)
{
return clmul_16x2_even(n >> 16, m >> 16);
}
uint64_t clmul_32(uint32_t n, uint32_t m32)
{
uint64_t r = 0;
uint64_t m = m32;
for (int i = 0; i < 32; ++i) {
r ^= n & 1 ? m : 0;
n >>= 1;
m <<= 1;
}
return r;
}