blob: 446931fe05ba90623fb504b28b7cc97dec2e06f1 [file] [log] [blame]
/*
* Carry-less multiply operations.
* SPDX-License-Identifier: GPL-2.0-or-later
*
* Copyright (C) 2023 Linaro, Ltd.
*/
#ifndef CRYPTO_CLMUL_H
#define CRYPTO_CLMUL_H
#include "qemu/int128.h"
#include "host/crypto/clmul.h"
/**
* clmul_8x8_low:
*
* Perform eight 8x8->8 carry-less multiplies.
*/
uint64_t clmul_8x8_low(uint64_t, uint64_t);
/**
* clmul_8x4_even:
*
* Perform four 8x8->16 carry-less multiplies.
* The odd bytes of the inputs are ignored.
*/
uint64_t clmul_8x4_even(uint64_t, uint64_t);
/**
* clmul_8x4_odd:
*
* Perform four 8x8->16 carry-less multiplies.
* The even bytes of the inputs are ignored.
*/
uint64_t clmul_8x4_odd(uint64_t, uint64_t);
/**
* clmul_8x4_packed:
*
* Perform four 8x8->16 carry-less multiplies.
*/
uint64_t clmul_8x4_packed(uint32_t, uint32_t);
/**
* clmul_16x2_even:
*
* Perform two 16x16->32 carry-less multiplies.
* The odd words of the inputs are ignored.
*/
uint64_t clmul_16x2_even(uint64_t, uint64_t);
/**
* clmul_16x2_odd:
*
* Perform two 16x16->32 carry-less multiplies.
* The even words of the inputs are ignored.
*/
uint64_t clmul_16x2_odd(uint64_t, uint64_t);
/**
* clmul_32:
*
* Perform a 32x32->64 carry-less multiply.
*/
uint64_t clmul_32(uint32_t, uint32_t);
/**
* clmul_64:
*
* Perform a 64x64->128 carry-less multiply.
*/
Int128 clmul_64_gen(uint64_t, uint64_t);
static inline Int128 clmul_64(uint64_t a, uint64_t b)
{
if (HAVE_CLMUL_ACCEL) {
return clmul_64_accel(a, b);
} else {
return clmul_64_gen(a, b);
}
}
#endif /* CRYPTO_CLMUL_H */