You've already forked linux-apfs
mirror of
https://github.com/linux-apfs/linux-apfs.git
synced 2026-05-01 15:00:59 -07:00
[CRYPTO] all: Pass tfm instead of ctx to algorithms
Up until now algorithms have been happy to get a context pointer since they know everything that's in the tfm already (e.g., alignment, block size). However, once we have parameterised algorithms, such information will be specific to each tfm. So the algorithm API needs to be changed to pass the tfm structure instead of the context pointer. This patch is basically a text substitution. The only tricky bit is the assembly routines that need to get the context pointer offset through asm-offsets.h. Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>
This commit is contained in:
+7
-7
@@ -248,10 +248,10 @@ gen_tabs (void)
|
||||
t ^= E_KEY[8 * i + 7]; E_KEY[8 * i + 15] = t; \
|
||||
}
|
||||
|
||||
static int
|
||||
aes_set_key(void *ctx_arg, const u8 *in_key, unsigned int key_len, u32 *flags)
|
||||
static int aes_set_key(struct crypto_tfm *tfm, const u8 *in_key,
|
||||
unsigned int key_len, u32 *flags)
|
||||
{
|
||||
struct aes_ctx *ctx = ctx_arg;
|
||||
struct aes_ctx *ctx = crypto_tfm_ctx(tfm);
|
||||
const __le32 *key = (const __le32 *)in_key;
|
||||
u32 i, t, u, v, w;
|
||||
|
||||
@@ -318,9 +318,9 @@ aes_set_key(void *ctx_arg, const u8 *in_key, unsigned int key_len, u32 *flags)
|
||||
f_rl(bo, bi, 2, k); \
|
||||
f_rl(bo, bi, 3, k)
|
||||
|
||||
static void aes_encrypt(void *ctx_arg, u8 *out, const u8 *in)
|
||||
static void aes_encrypt(struct crypto_tfm *tfm, u8 *out, const u8 *in)
|
||||
{
|
||||
const struct aes_ctx *ctx = ctx_arg;
|
||||
const struct aes_ctx *ctx = crypto_tfm_ctx(tfm);
|
||||
const __le32 *src = (const __le32 *)in;
|
||||
__le32 *dst = (__le32 *)out;
|
||||
u32 b0[4], b1[4];
|
||||
@@ -373,9 +373,9 @@ static void aes_encrypt(void *ctx_arg, u8 *out, const u8 *in)
|
||||
i_rl(bo, bi, 2, k); \
|
||||
i_rl(bo, bi, 3, k)
|
||||
|
||||
static void aes_decrypt(void *ctx_arg, u8 *out, const u8 *in)
|
||||
static void aes_decrypt(struct crypto_tfm *tfm, u8 *out, const u8 *in)
|
||||
{
|
||||
const struct aes_ctx *ctx = ctx_arg;
|
||||
const struct aes_ctx *ctx = crypto_tfm_ctx(tfm);
|
||||
const __le32 *src = (const __le32 *)in;
|
||||
__le32 *dst = (__le32 *)out;
|
||||
u32 b0[4], b1[4];
|
||||
|
||||
+6
-7
@@ -460,16 +460,15 @@ static const u32 rc[] = {
|
||||
0xf726ffedU, 0xe89d6f8eU, 0x19a0f089U,
|
||||
};
|
||||
|
||||
static int anubis_setkey(void *ctx_arg, const u8 *in_key,
|
||||
static int anubis_setkey(struct crypto_tfm *tfm, const u8 *in_key,
|
||||
unsigned int key_len, u32 *flags)
|
||||
{
|
||||
struct anubis_ctx *ctx = crypto_tfm_ctx(tfm);
|
||||
const __be32 *key = (const __be32 *)in_key;
|
||||
int N, R, i, r;
|
||||
u32 kappa[ANUBIS_MAX_N];
|
||||
u32 inter[ANUBIS_MAX_N];
|
||||
|
||||
struct anubis_ctx *ctx = ctx_arg;
|
||||
|
||||
switch (key_len)
|
||||
{
|
||||
case 16: case 20: case 24: case 28:
|
||||
@@ -660,15 +659,15 @@ static void anubis_crypt(u32 roundKey[ANUBIS_MAX_ROUNDS + 1][4],
|
||||
dst[i] = cpu_to_be32(inter[i]);
|
||||
}
|
||||
|
||||
static void anubis_encrypt(void *ctx_arg, u8 *dst, const u8 *src)
|
||||
static void anubis_encrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src)
|
||||
{
|
||||
struct anubis_ctx *ctx = ctx_arg;
|
||||
struct anubis_ctx *ctx = crypto_tfm_ctx(tfm);
|
||||
anubis_crypt(ctx->E, dst, src, ctx->R);
|
||||
}
|
||||
|
||||
static void anubis_decrypt(void *ctx_arg, u8 *dst, const u8 *src)
|
||||
static void anubis_decrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src)
|
||||
{
|
||||
struct anubis_ctx *ctx = ctx_arg;
|
||||
struct anubis_ctx *ctx = crypto_tfm_ctx(tfm);
|
||||
anubis_crypt(ctx->D, dst, src, ctx->R);
|
||||
}
|
||||
|
||||
|
||||
+5
-4
@@ -24,9 +24,10 @@ struct arc4_ctx {
|
||||
u8 x, y;
|
||||
};
|
||||
|
||||
static int arc4_set_key(void *ctx_arg, const u8 *in_key, unsigned int key_len, u32 *flags)
|
||||
static int arc4_set_key(struct crypto_tfm *tfm, const u8 *in_key,
|
||||
unsigned int key_len, u32 *flags)
|
||||
{
|
||||
struct arc4_ctx *ctx = ctx_arg;
|
||||
struct arc4_ctx *ctx = crypto_tfm_ctx(tfm);
|
||||
int i, j = 0, k = 0;
|
||||
|
||||
ctx->x = 1;
|
||||
@@ -48,9 +49,9 @@ static int arc4_set_key(void *ctx_arg, const u8 *in_key, unsigned int key_len, u
|
||||
return 0;
|
||||
}
|
||||
|
||||
static void arc4_crypt(void *ctx_arg, u8 *out, const u8 *in)
|
||||
static void arc4_crypt(struct crypto_tfm *tfm, u8 *out, const u8 *in)
|
||||
{
|
||||
struct arc4_ctx *ctx = ctx_arg;
|
||||
struct arc4_ctx *ctx = crypto_tfm_ctx(tfm);
|
||||
|
||||
u8 *const S = ctx->S;
|
||||
u8 x = ctx->x;
|
||||
|
||||
+11
-8
@@ -349,7 +349,7 @@ static void encrypt_block(struct bf_ctx *bctx, u32 *dst, u32 *src)
|
||||
dst[1] = yl;
|
||||
}
|
||||
|
||||
static void bf_encrypt(void *ctx, u8 *dst, const u8 *src)
|
||||
static void bf_encrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src)
|
||||
{
|
||||
const __be32 *in_blk = (const __be32 *)src;
|
||||
__be32 *const out_blk = (__be32 *)dst;
|
||||
@@ -357,17 +357,18 @@ static void bf_encrypt(void *ctx, u8 *dst, const u8 *src)
|
||||
|
||||
in32[0] = be32_to_cpu(in_blk[0]);
|
||||
in32[1] = be32_to_cpu(in_blk[1]);
|
||||
encrypt_block(ctx, out32, in32);
|
||||
encrypt_block(crypto_tfm_ctx(tfm), out32, in32);
|
||||
out_blk[0] = cpu_to_be32(out32[0]);
|
||||
out_blk[1] = cpu_to_be32(out32[1]);
|
||||
}
|
||||
|
||||
static void bf_decrypt(void *ctx, u8 *dst, const u8 *src)
|
||||
static void bf_decrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src)
|
||||
{
|
||||
struct bf_ctx *ctx = crypto_tfm_ctx(tfm);
|
||||
const __be32 *in_blk = (const __be32 *)src;
|
||||
__be32 *const out_blk = (__be32 *)dst;
|
||||
const u32 *P = ((struct bf_ctx *)ctx)->p;
|
||||
const u32 *S = ((struct bf_ctx *)ctx)->s;
|
||||
const u32 *P = ctx->p;
|
||||
const u32 *S = ctx->s;
|
||||
u32 yl = be32_to_cpu(in_blk[0]);
|
||||
u32 yr = be32_to_cpu(in_blk[1]);
|
||||
|
||||
@@ -398,12 +399,14 @@ static void bf_decrypt(void *ctx, u8 *dst, const u8 *src)
|
||||
/*
|
||||
* Calculates the blowfish S and P boxes for encryption and decryption.
|
||||
*/
|
||||
static int bf_setkey(void *ctx, const u8 *key, unsigned int keylen, u32 *flags)
|
||||
static int bf_setkey(struct crypto_tfm *tfm, const u8 *key,
|
||||
unsigned int keylen, u32 *flags)
|
||||
{
|
||||
struct bf_ctx *ctx = crypto_tfm_ctx(tfm);
|
||||
u32 *P = ctx->p;
|
||||
u32 *S = ctx->s;
|
||||
short i, j, count;
|
||||
u32 data[2], temp;
|
||||
u32 *P = ((struct bf_ctx *)ctx)->p;
|
||||
u32 *S = ((struct bf_ctx *)ctx)->s;
|
||||
|
||||
/* Copy the initialization s-boxes */
|
||||
for (i = 0, count = 0; i < 256; i++)
|
||||
|
||||
+7
-7
@@ -577,9 +577,9 @@ static const u32 sb8[256] = {
|
||||
(((s1[I >> 24] + s2[(I>>16)&0xff]) ^ s3[(I>>8)&0xff]) - s4[I&0xff]) )
|
||||
|
||||
|
||||
static void cast5_encrypt(void *ctx, u8 * outbuf, const u8 * inbuf)
|
||||
static void cast5_encrypt(struct crypto_tfm *tfm, u8 *outbuf, const u8 *inbuf)
|
||||
{
|
||||
struct cast5_ctx *c = (struct cast5_ctx *) ctx;
|
||||
struct cast5_ctx *c = crypto_tfm_ctx(tfm);
|
||||
const __be32 *src = (const __be32 *)inbuf;
|
||||
__be32 *dst = (__be32 *)outbuf;
|
||||
u32 l, r, t;
|
||||
@@ -642,9 +642,9 @@ static void cast5_encrypt(void *ctx, u8 * outbuf, const u8 * inbuf)
|
||||
dst[1] = cpu_to_be32(l);
|
||||
}
|
||||
|
||||
static void cast5_decrypt(void *ctx, u8 * outbuf, const u8 * inbuf)
|
||||
static void cast5_decrypt(struct crypto_tfm *tfm, u8 *outbuf, const u8 *inbuf)
|
||||
{
|
||||
struct cast5_ctx *c = (struct cast5_ctx *) ctx;
|
||||
struct cast5_ctx *c = crypto_tfm_ctx(tfm);
|
||||
const __be32 *src = (const __be32 *)inbuf;
|
||||
__be32 *dst = (__be32 *)outbuf;
|
||||
u32 l, r, t;
|
||||
@@ -769,15 +769,15 @@ static void key_schedule(u32 * x, u32 * z, u32 * k)
|
||||
}
|
||||
|
||||
|
||||
static int
|
||||
cast5_setkey(void *ctx, const u8 * key, unsigned key_len, u32 * flags)
|
||||
static int cast5_setkey(struct crypto_tfm *tfm, const u8 *key,
|
||||
unsigned key_len, u32 *flags)
|
||||
{
|
||||
struct cast5_ctx *c = crypto_tfm_ctx(tfm);
|
||||
int i;
|
||||
u32 x[4];
|
||||
u32 z[4];
|
||||
u32 k[16];
|
||||
__be32 p_key[4];
|
||||
struct cast5_ctx *c = (struct cast5_ctx *) ctx;
|
||||
|
||||
if (key_len < 5 || key_len > 16) {
|
||||
*flags |= CRYPTO_TFM_RES_BAD_KEY_LEN;
|
||||
|
||||
+8
-7
@@ -381,13 +381,13 @@ static inline void W(u32 *key, unsigned int i) {
|
||||
key[7] ^= F2(key[0], Tr[i % 4][7], Tm[i][7]);
|
||||
}
|
||||
|
||||
static int
|
||||
cast6_setkey(void *ctx, const u8 * in_key, unsigned key_len, u32 * flags)
|
||||
static int cast6_setkey(struct crypto_tfm *tfm, const u8 *in_key,
|
||||
unsigned key_len, u32 *flags)
|
||||
{
|
||||
int i;
|
||||
u32 key[8];
|
||||
__be32 p_key[8]; /* padded key */
|
||||
struct cast6_ctx *c = (struct cast6_ctx *) ctx;
|
||||
struct cast6_ctx *c = crypto_tfm_ctx(tfm);
|
||||
|
||||
if (key_len < 16 || key_len > 32 || key_len % 4 != 0) {
|
||||
*flags |= CRYPTO_TFM_RES_BAD_KEY_LEN;
|
||||
@@ -444,8 +444,9 @@ static inline void QBAR (u32 * block, u8 * Kr, u32 * Km) {
|
||||
block[2] ^= F1(block[3], Kr[0], Km[0]);
|
||||
}
|
||||
|
||||
static void cast6_encrypt (void * ctx, u8 * outbuf, const u8 * inbuf) {
|
||||
struct cast6_ctx * c = (struct cast6_ctx *)ctx;
|
||||
static void cast6_encrypt(struct crypto_tfm *tfm, u8 *outbuf, const u8 *inbuf)
|
||||
{
|
||||
struct cast6_ctx *c = crypto_tfm_ctx(tfm);
|
||||
const __be32 *src = (const __be32 *)inbuf;
|
||||
__be32 *dst = (__be32 *)outbuf;
|
||||
u32 block[4];
|
||||
@@ -476,8 +477,8 @@ static void cast6_encrypt (void * ctx, u8 * outbuf, const u8 * inbuf) {
|
||||
dst[3] = cpu_to_be32(block[3]);
|
||||
}
|
||||
|
||||
static void cast6_decrypt (void * ctx, u8 * outbuf, const u8 * inbuf) {
|
||||
struct cast6_ctx * c = (struct cast6_ctx *)ctx;
|
||||
static void cast6_decrypt(struct crypto_tfm *tfm, u8 *outbuf, const u8 *inbuf) {
|
||||
struct cast6_ctx * c = crypto_tfm_ctx(tfm);
|
||||
const __be32 *src = (const __be32 *)inbuf;
|
||||
__be32 *dst = (__be32 *)outbuf;
|
||||
u32 block[4];
|
||||
|
||||
+7
-7
@@ -187,7 +187,7 @@ static unsigned int cbc_process_encrypt(const struct cipher_desc *desc,
|
||||
void (*xor)(u8 *, const u8 *) = tfm->crt_u.cipher.cit_xor_block;
|
||||
int bsize = crypto_tfm_alg_blocksize(tfm);
|
||||
|
||||
void (*fn)(void *, u8 *, const u8 *) = desc->crfn;
|
||||
void (*fn)(struct crypto_tfm *, u8 *, const u8 *) = desc->crfn;
|
||||
u8 *iv = desc->info;
|
||||
unsigned int done = 0;
|
||||
|
||||
@@ -195,7 +195,7 @@ static unsigned int cbc_process_encrypt(const struct cipher_desc *desc,
|
||||
|
||||
do {
|
||||
xor(iv, src);
|
||||
fn(crypto_tfm_ctx(tfm), dst, iv);
|
||||
fn(tfm, dst, iv);
|
||||
memcpy(iv, dst, bsize);
|
||||
|
||||
src += bsize;
|
||||
@@ -218,7 +218,7 @@ static unsigned int cbc_process_decrypt(const struct cipher_desc *desc,
|
||||
u8 *buf = (u8 *)ALIGN((unsigned long)stack, alignmask + 1);
|
||||
u8 **dst_p = src == dst ? &buf : &dst;
|
||||
|
||||
void (*fn)(void *, u8 *, const u8 *) = desc->crfn;
|
||||
void (*fn)(struct crypto_tfm *, u8 *, const u8 *) = desc->crfn;
|
||||
u8 *iv = desc->info;
|
||||
unsigned int done = 0;
|
||||
|
||||
@@ -227,7 +227,7 @@ static unsigned int cbc_process_decrypt(const struct cipher_desc *desc,
|
||||
do {
|
||||
u8 *tmp_dst = *dst_p;
|
||||
|
||||
fn(crypto_tfm_ctx(tfm), tmp_dst, src);
|
||||
fn(tfm, tmp_dst, src);
|
||||
xor(tmp_dst, iv);
|
||||
memcpy(iv, src, bsize);
|
||||
if (tmp_dst != dst)
|
||||
@@ -245,13 +245,13 @@ static unsigned int ecb_process(const struct cipher_desc *desc, u8 *dst,
|
||||
{
|
||||
struct crypto_tfm *tfm = desc->tfm;
|
||||
int bsize = crypto_tfm_alg_blocksize(tfm);
|
||||
void (*fn)(void *, u8 *, const u8 *) = desc->crfn;
|
||||
void (*fn)(struct crypto_tfm *, u8 *, const u8 *) = desc->crfn;
|
||||
unsigned int done = 0;
|
||||
|
||||
nbytes -= bsize;
|
||||
|
||||
do {
|
||||
fn(crypto_tfm_ctx(tfm), dst, src);
|
||||
fn(tfm, dst, src);
|
||||
|
||||
src += bsize;
|
||||
dst += bsize;
|
||||
@@ -268,7 +268,7 @@ static int setkey(struct crypto_tfm *tfm, const u8 *key, unsigned int keylen)
|
||||
tfm->crt_flags |= CRYPTO_TFM_RES_BAD_KEY_LEN;
|
||||
return -EINVAL;
|
||||
} else
|
||||
return cia->cia_setkey(crypto_tfm_ctx(tfm), key, keylen,
|
||||
return cia->cia_setkey(tfm, key, keylen,
|
||||
&tfm->crt_flags);
|
||||
}
|
||||
|
||||
|
||||
+2
-4
@@ -22,8 +22,7 @@ static int crypto_compress(struct crypto_tfm *tfm,
|
||||
const u8 *src, unsigned int slen,
|
||||
u8 *dst, unsigned int *dlen)
|
||||
{
|
||||
return tfm->__crt_alg->cra_compress.coa_compress(crypto_tfm_ctx(tfm),
|
||||
src, slen, dst,
|
||||
return tfm->__crt_alg->cra_compress.coa_compress(tfm, src, slen, dst,
|
||||
dlen);
|
||||
}
|
||||
|
||||
@@ -31,8 +30,7 @@ static int crypto_decompress(struct crypto_tfm *tfm,
|
||||
const u8 *src, unsigned int slen,
|
||||
u8 *dst, unsigned int *dlen)
|
||||
{
|
||||
return tfm->__crt_alg->cra_compress.coa_decompress(crypto_tfm_ctx(tfm),
|
||||
src, slen, dst,
|
||||
return tfm->__crt_alg->cra_compress.coa_decompress(tfm, src, slen, dst,
|
||||
dlen);
|
||||
}
|
||||
|
||||
|
||||
+10
-9
@@ -31,9 +31,9 @@ struct chksum_ctx {
|
||||
* crc using table.
|
||||
*/
|
||||
|
||||
static void chksum_init(void *ctx)
|
||||
static void chksum_init(struct crypto_tfm *tfm)
|
||||
{
|
||||
struct chksum_ctx *mctx = ctx;
|
||||
struct chksum_ctx *mctx = crypto_tfm_ctx(tfm);
|
||||
|
||||
mctx->crc = ~(u32)0; /* common usage */
|
||||
}
|
||||
@@ -43,10 +43,10 @@ static void chksum_init(void *ctx)
|
||||
* If your algorithm starts with ~0, then XOR with ~0 before you set
|
||||
* the seed.
|
||||
*/
|
||||
static int chksum_setkey(void *ctx, const u8 *key, unsigned int keylen,
|
||||
u32 *flags)
|
||||
static int chksum_setkey(struct crypto_tfm *tfm, const u8 *key,
|
||||
unsigned int keylen, u32 *flags)
|
||||
{
|
||||
struct chksum_ctx *mctx = ctx;
|
||||
struct chksum_ctx *mctx = crypto_tfm_ctx(tfm);
|
||||
|
||||
if (keylen != sizeof(mctx->crc)) {
|
||||
if (flags)
|
||||
@@ -57,9 +57,10 @@ static int chksum_setkey(void *ctx, const u8 *key, unsigned int keylen,
|
||||
return 0;
|
||||
}
|
||||
|
||||
static void chksum_update(void *ctx, const u8 *data, unsigned int length)
|
||||
static void chksum_update(struct crypto_tfm *tfm, const u8 *data,
|
||||
unsigned int length)
|
||||
{
|
||||
struct chksum_ctx *mctx = ctx;
|
||||
struct chksum_ctx *mctx = crypto_tfm_ctx(tfm);
|
||||
u32 mcrc;
|
||||
|
||||
mcrc = crc32c(mctx->crc, data, (size_t)length);
|
||||
@@ -67,9 +68,9 @@ static void chksum_update(void *ctx, const u8 *data, unsigned int length)
|
||||
mctx->crc = mcrc;
|
||||
}
|
||||
|
||||
static void chksum_final(void *ctx, u8 *out)
|
||||
static void chksum_final(struct crypto_tfm *tfm, u8 *out)
|
||||
{
|
||||
struct chksum_ctx *mctx = ctx;
|
||||
struct chksum_ctx *mctx = crypto_tfm_ctx(tfm);
|
||||
u32 mcrc = (mctx->crc ^ ~(u32)0);
|
||||
|
||||
*(u32 *)out = __le32_to_cpu(mcrc);
|
||||
|
||||
@@ -27,8 +27,8 @@
|
||||
#define NULL_BLOCK_SIZE 1
|
||||
#define NULL_DIGEST_SIZE 0
|
||||
|
||||
static int null_compress(void *ctx, const u8 *src, unsigned int slen,
|
||||
u8 *dst, unsigned int *dlen)
|
||||
static int null_compress(struct crypto_tfm *tfm, const u8 *src,
|
||||
unsigned int slen, u8 *dst, unsigned int *dlen)
|
||||
{
|
||||
if (slen > *dlen)
|
||||
return -EINVAL;
|
||||
@@ -37,20 +37,21 @@ static int null_compress(void *ctx, const u8 *src, unsigned int slen,
|
||||
return 0;
|
||||
}
|
||||
|
||||
static void null_init(void *ctx)
|
||||
static void null_init(struct crypto_tfm *tfm)
|
||||
{ }
|
||||
|
||||
static void null_update(void *ctx, const u8 *data, unsigned int len)
|
||||
static void null_update(struct crypto_tfm *tfm, const u8 *data,
|
||||
unsigned int len)
|
||||
{ }
|
||||
|
||||
static void null_final(void *ctx, u8 *out)
|
||||
static void null_final(struct crypto_tfm *tfm, u8 *out)
|
||||
{ }
|
||||
|
||||
static int null_setkey(void *ctx, const u8 *key,
|
||||
unsigned int keylen, u32 *flags)
|
||||
static int null_setkey(struct crypto_tfm *tfm, const u8 *key,
|
||||
unsigned int keylen, u32 *flags)
|
||||
{ return 0; }
|
||||
|
||||
static void null_crypt(void *ctx, u8 *dst, const u8 *src)
|
||||
static void null_crypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src)
|
||||
{
|
||||
memcpy(dst, src, NULL_BLOCK_SIZE);
|
||||
}
|
||||
|
||||
+11
-8
@@ -102,8 +102,9 @@ static void deflate_decomp_exit(struct deflate_ctx *ctx)
|
||||
kfree(ctx->decomp_stream.workspace);
|
||||
}
|
||||
|
||||
static int deflate_init(void *ctx)
|
||||
static int deflate_init(struct crypto_tfm *tfm)
|
||||
{
|
||||
struct deflate_ctx *ctx = crypto_tfm_ctx(tfm);
|
||||
int ret;
|
||||
|
||||
ret = deflate_comp_init(ctx);
|
||||
@@ -116,17 +117,19 @@ out:
|
||||
return ret;
|
||||
}
|
||||
|
||||
static void deflate_exit(void *ctx)
|
||||
static void deflate_exit(struct crypto_tfm *tfm)
|
||||
{
|
||||
struct deflate_ctx *ctx = crypto_tfm_ctx(tfm);
|
||||
|
||||
deflate_comp_exit(ctx);
|
||||
deflate_decomp_exit(ctx);
|
||||
}
|
||||
|
||||
static int deflate_compress(void *ctx, const u8 *src, unsigned int slen,
|
||||
u8 *dst, unsigned int *dlen)
|
||||
static int deflate_compress(struct crypto_tfm *tfm, const u8 *src,
|
||||
unsigned int slen, u8 *dst, unsigned int *dlen)
|
||||
{
|
||||
int ret = 0;
|
||||
struct deflate_ctx *dctx = ctx;
|
||||
struct deflate_ctx *dctx = crypto_tfm_ctx(tfm);
|
||||
struct z_stream_s *stream = &dctx->comp_stream;
|
||||
|
||||
ret = zlib_deflateReset(stream);
|
||||
@@ -151,12 +154,12 @@ out:
|
||||
return ret;
|
||||
}
|
||||
|
||||
static int deflate_decompress(void *ctx, const u8 *src, unsigned int slen,
|
||||
u8 *dst, unsigned int *dlen)
|
||||
static int deflate_decompress(struct crypto_tfm *tfm, const u8 *src,
|
||||
unsigned int slen, u8 *dst, unsigned int *dlen)
|
||||
{
|
||||
|
||||
int ret = 0;
|
||||
struct deflate_ctx *dctx = ctx;
|
||||
struct deflate_ctx *dctx = crypto_tfm_ctx(tfm);
|
||||
struct z_stream_s *stream = &dctx->decomp_stream;
|
||||
|
||||
ret = zlib_inflateReset(stream);
|
||||
|
||||
+15
-12
@@ -783,9 +783,10 @@ static void dkey(u32 *pe, const u8 *k)
|
||||
}
|
||||
}
|
||||
|
||||
static int des_setkey(void *ctx, const u8 *key, unsigned int keylen, u32 *flags)
|
||||
static int des_setkey(struct crypto_tfm *tfm, const u8 *key,
|
||||
unsigned int keylen, u32 *flags)
|
||||
{
|
||||
struct des_ctx *dctx = ctx;
|
||||
struct des_ctx *dctx = crypto_tfm_ctx(tfm);
|
||||
u32 tmp[DES_EXPKEY_WORDS];
|
||||
int ret;
|
||||
|
||||
@@ -803,9 +804,10 @@ static int des_setkey(void *ctx, const u8 *key, unsigned int keylen, u32 *flags)
|
||||
return 0;
|
||||
}
|
||||
|
||||
static void des_encrypt(void *ctx, u8 *dst, const u8 *src)
|
||||
static void des_encrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src)
|
||||
{
|
||||
const u32 *K = ((struct des_ctx *)ctx)->expkey;
|
||||
struct des_ctx *ctx = crypto_tfm_ctx(tfm);
|
||||
const u32 *K = ctx->expkey;
|
||||
const __le32 *s = (const __le32 *)src;
|
||||
__le32 *d = (__le32 *)dst;
|
||||
u32 L, R, A, B;
|
||||
@@ -825,9 +827,10 @@ static void des_encrypt(void *ctx, u8 *dst, const u8 *src)
|
||||
d[1] = cpu_to_le32(L);
|
||||
}
|
||||
|
||||
static void des_decrypt(void *ctx, u8 *dst, const u8 *src)
|
||||
static void des_decrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src)
|
||||
{
|
||||
const u32 *K = ((struct des_ctx *)ctx)->expkey + DES_EXPKEY_WORDS - 2;
|
||||
struct des_ctx *ctx = crypto_tfm_ctx(tfm);
|
||||
const u32 *K = ctx->expkey + DES_EXPKEY_WORDS - 2;
|
||||
const __le32 *s = (const __le32 *)src;
|
||||
__le32 *d = (__le32 *)dst;
|
||||
u32 L, R, A, B;
|
||||
@@ -860,11 +863,11 @@ static void des_decrypt(void *ctx, u8 *dst, const u8 *src)
|
||||
* property.
|
||||
*
|
||||
*/
|
||||
static int des3_ede_setkey(void *ctx, const u8 *key,
|
||||
static int des3_ede_setkey(struct crypto_tfm *tfm, const u8 *key,
|
||||
unsigned int keylen, u32 *flags)
|
||||
{
|
||||
const u32 *K = (const u32 *)key;
|
||||
struct des3_ede_ctx *dctx = ctx;
|
||||
struct des3_ede_ctx *dctx = crypto_tfm_ctx(tfm);
|
||||
u32 *expkey = dctx->expkey;
|
||||
|
||||
if (unlikely(!((K[0] ^ K[2]) | (K[1] ^ K[3])) ||
|
||||
@@ -881,9 +884,9 @@ static int des3_ede_setkey(void *ctx, const u8 *key,
|
||||
return 0;
|
||||
}
|
||||
|
||||
static void des3_ede_encrypt(void *ctx, u8 *dst, const u8 *src)
|
||||
static void des3_ede_encrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src)
|
||||
{
|
||||
struct des3_ede_ctx *dctx = ctx;
|
||||
struct des3_ede_ctx *dctx = crypto_tfm_ctx(tfm);
|
||||
const u32 *K = dctx->expkey;
|
||||
const __le32 *s = (const __le32 *)src;
|
||||
__le32 *d = (__le32 *)dst;
|
||||
@@ -912,9 +915,9 @@ static void des3_ede_encrypt(void *ctx, u8 *dst, const u8 *src)
|
||||
d[1] = cpu_to_le32(L);
|
||||
}
|
||||
|
||||
static void des3_ede_decrypt(void *ctx, u8 *dst, const u8 *src)
|
||||
static void des3_ede_decrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src)
|
||||
{
|
||||
struct des3_ede_ctx *dctx = ctx;
|
||||
struct des3_ede_ctx *dctx = crypto_tfm_ctx(tfm);
|
||||
const u32 *K = dctx->expkey + DES3_EDE_EXPKEY_WORDS - 2;
|
||||
const __le32 *s = (const __le32 *)src;
|
||||
__le32 *d = (__le32 *)dst;
|
||||
|
||||
+6
-9
@@ -20,7 +20,7 @@
|
||||
|
||||
static void init(struct crypto_tfm *tfm)
|
||||
{
|
||||
tfm->__crt_alg->cra_digest.dia_init(crypto_tfm_ctx(tfm));
|
||||
tfm->__crt_alg->cra_digest.dia_init(tfm);
|
||||
}
|
||||
|
||||
static void update(struct crypto_tfm *tfm,
|
||||
@@ -46,16 +46,14 @@ static void update(struct crypto_tfm *tfm,
|
||||
unsigned int bytes =
|
||||
alignmask + 1 - (offset & alignmask);
|
||||
bytes = min(bytes, bytes_from_page);
|
||||
tfm->__crt_alg->cra_digest.dia_update
|
||||
(crypto_tfm_ctx(tfm), p,
|
||||
bytes);
|
||||
tfm->__crt_alg->cra_digest.dia_update(tfm, p,
|
||||
bytes);
|
||||
p += bytes;
|
||||
bytes_from_page -= bytes;
|
||||
l -= bytes;
|
||||
}
|
||||
tfm->__crt_alg->cra_digest.dia_update
|
||||
(crypto_tfm_ctx(tfm), p,
|
||||
bytes_from_page);
|
||||
tfm->__crt_alg->cra_digest.dia_update(tfm, p,
|
||||
bytes_from_page);
|
||||
crypto_kunmap(src, 0);
|
||||
crypto_yield(tfm);
|
||||
offset = 0;
|
||||
@@ -83,8 +81,7 @@ static int setkey(struct crypto_tfm *tfm, const u8 *key, unsigned int keylen)
|
||||
u32 flags;
|
||||
if (tfm->__crt_alg->cra_digest.dia_setkey == NULL)
|
||||
return -ENOSYS;
|
||||
return tfm->__crt_alg->cra_digest.dia_setkey(crypto_tfm_ctx(tfm),
|
||||
key, keylen, &flags);
|
||||
return tfm->__crt_alg->cra_digest.dia_setkey(tfm, key, keylen, &flags);
|
||||
}
|
||||
|
||||
static void digest(struct crypto_tfm *tfm,
|
||||
|
||||
+7
-7
@@ -754,10 +754,10 @@ static const u64 c[KHAZAD_ROUNDS + 1] = {
|
||||
0xccc41d14c363da5dULL, 0x5fdc7dcd7f5a6c5cULL, 0xf726ffede89d6f8eULL
|
||||
};
|
||||
|
||||
static int khazad_setkey(void *ctx_arg, const u8 *in_key,
|
||||
unsigned int key_len, u32 *flags)
|
||||
static int khazad_setkey(struct crypto_tfm *tfm, const u8 *in_key,
|
||||
unsigned int key_len, u32 *flags)
|
||||
{
|
||||
struct khazad_ctx *ctx = ctx_arg;
|
||||
struct khazad_ctx *ctx = crypto_tfm_ctx(tfm);
|
||||
const __be32 *key = (const __be32 *)in_key;
|
||||
int r;
|
||||
const u64 *S = T7;
|
||||
@@ -841,15 +841,15 @@ static void khazad_crypt(const u64 roundKey[KHAZAD_ROUNDS + 1],
|
||||
*dst = cpu_to_be64(state);
|
||||
}
|
||||
|
||||
static void khazad_encrypt(void *ctx_arg, u8 *dst, const u8 *src)
|
||||
static void khazad_encrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src)
|
||||
{
|
||||
struct khazad_ctx *ctx = ctx_arg;
|
||||
struct khazad_ctx *ctx = crypto_tfm_ctx(tfm);
|
||||
khazad_crypt(ctx->E, dst, src);
|
||||
}
|
||||
|
||||
static void khazad_decrypt(void *ctx_arg, u8 *dst, const u8 *src)
|
||||
static void khazad_decrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src)
|
||||
{
|
||||
struct khazad_ctx *ctx = ctx_arg;
|
||||
struct khazad_ctx *ctx = crypto_tfm_ctx(tfm);
|
||||
khazad_crypt(ctx->D, dst, src);
|
||||
}
|
||||
|
||||
|
||||
+6
-6
@@ -152,9 +152,9 @@ static inline void md4_transform_helper(struct md4_ctx *ctx)
|
||||
md4_transform(ctx->hash, ctx->block);
|
||||
}
|
||||
|
||||
static void md4_init(void *ctx)
|
||||
static void md4_init(struct crypto_tfm *tfm)
|
||||
{
|
||||
struct md4_ctx *mctx = ctx;
|
||||
struct md4_ctx *mctx = crypto_tfm_ctx(tfm);
|
||||
|
||||
mctx->hash[0] = 0x67452301;
|
||||
mctx->hash[1] = 0xefcdab89;
|
||||
@@ -163,9 +163,9 @@ static void md4_init(void *ctx)
|
||||
mctx->byte_count = 0;
|
||||
}
|
||||
|
||||
static void md4_update(void *ctx, const u8 *data, unsigned int len)
|
||||
static void md4_update(struct crypto_tfm *tfm, const u8 *data, unsigned int len)
|
||||
{
|
||||
struct md4_ctx *mctx = ctx;
|
||||
struct md4_ctx *mctx = crypto_tfm_ctx(tfm);
|
||||
const u32 avail = sizeof(mctx->block) - (mctx->byte_count & 0x3f);
|
||||
|
||||
mctx->byte_count += len;
|
||||
@@ -193,9 +193,9 @@ static void md4_update(void *ctx, const u8 *data, unsigned int len)
|
||||
memcpy(mctx->block, data, len);
|
||||
}
|
||||
|
||||
static void md4_final(void *ctx, u8 *out)
|
||||
static void md4_final(struct crypto_tfm *tfm, u8 *out)
|
||||
{
|
||||
struct md4_ctx *mctx = ctx;
|
||||
struct md4_ctx *mctx = crypto_tfm_ctx(tfm);
|
||||
const unsigned int offset = mctx->byte_count & 0x3f;
|
||||
char *p = (char *)mctx->block + offset;
|
||||
int padding = 56 - (offset + 1);
|
||||
|
||||
+6
-6
@@ -147,9 +147,9 @@ static inline void md5_transform_helper(struct md5_ctx *ctx)
|
||||
md5_transform(ctx->hash, ctx->block);
|
||||
}
|
||||
|
||||
static void md5_init(void *ctx)
|
||||
static void md5_init(struct crypto_tfm *tfm)
|
||||
{
|
||||
struct md5_ctx *mctx = ctx;
|
||||
struct md5_ctx *mctx = crypto_tfm_ctx(tfm);
|
||||
|
||||
mctx->hash[0] = 0x67452301;
|
||||
mctx->hash[1] = 0xefcdab89;
|
||||
@@ -158,9 +158,9 @@ static void md5_init(void *ctx)
|
||||
mctx->byte_count = 0;
|
||||
}
|
||||
|
||||
static void md5_update(void *ctx, const u8 *data, unsigned int len)
|
||||
static void md5_update(struct crypto_tfm *tfm, const u8 *data, unsigned int len)
|
||||
{
|
||||
struct md5_ctx *mctx = ctx;
|
||||
struct md5_ctx *mctx = crypto_tfm_ctx(tfm);
|
||||
const u32 avail = sizeof(mctx->block) - (mctx->byte_count & 0x3f);
|
||||
|
||||
mctx->byte_count += len;
|
||||
@@ -188,9 +188,9 @@ static void md5_update(void *ctx, const u8 *data, unsigned int len)
|
||||
memcpy(mctx->block, data, len);
|
||||
}
|
||||
|
||||
static void md5_final(void *ctx, u8 *out)
|
||||
static void md5_final(struct crypto_tfm *tfm, u8 *out)
|
||||
{
|
||||
struct md5_ctx *mctx = ctx;
|
||||
struct md5_ctx *mctx = crypto_tfm_ctx(tfm);
|
||||
const unsigned int offset = mctx->byte_count & 0x3f;
|
||||
char *p = (char *)mctx->block + offset;
|
||||
int padding = 56 - (offset + 1);
|
||||
|
||||
+10
-9
@@ -45,16 +45,17 @@ do { \
|
||||
} while (0)
|
||||
|
||||
|
||||
static void michael_init(void *ctx)
|
||||
static void michael_init(struct crypto_tfm *tfm)
|
||||
{
|
||||
struct michael_mic_ctx *mctx = ctx;
|
||||
struct michael_mic_ctx *mctx = crypto_tfm_ctx(tfm);
|
||||
mctx->pending_len = 0;
|
||||
}
|
||||
|
||||
|
||||
static void michael_update(void *ctx, const u8 *data, unsigned int len)
|
||||
static void michael_update(struct crypto_tfm *tfm, const u8 *data,
|
||||
unsigned int len)
|
||||
{
|
||||
struct michael_mic_ctx *mctx = ctx;
|
||||
struct michael_mic_ctx *mctx = crypto_tfm_ctx(tfm);
|
||||
const __le32 *src;
|
||||
|
||||
if (mctx->pending_len) {
|
||||
@@ -90,9 +91,9 @@ static void michael_update(void *ctx, const u8 *data, unsigned int len)
|
||||
}
|
||||
|
||||
|
||||
static void michael_final(void *ctx, u8 *out)
|
||||
static void michael_final(struct crypto_tfm *tfm, u8 *out)
|
||||
{
|
||||
struct michael_mic_ctx *mctx = ctx;
|
||||
struct michael_mic_ctx *mctx = crypto_tfm_ctx(tfm);
|
||||
u8 *data = mctx->pending;
|
||||
__le32 *dst = (__le32 *)out;
|
||||
|
||||
@@ -121,10 +122,10 @@ static void michael_final(void *ctx, u8 *out)
|
||||
}
|
||||
|
||||
|
||||
static int michael_setkey(void *ctx, const u8 *key, unsigned int keylen,
|
||||
u32 *flags)
|
||||
static int michael_setkey(struct crypto_tfm *tfm, const u8 *key,
|
||||
unsigned int keylen, u32 *flags)
|
||||
{
|
||||
struct michael_mic_ctx *mctx = ctx;
|
||||
struct michael_mic_ctx *mctx = crypto_tfm_ctx(tfm);
|
||||
const __le32 *data = (const __le32 *)key;
|
||||
|
||||
if (keylen != 8) {
|
||||
|
||||
+16
-11
@@ -215,9 +215,11 @@ struct serpent_ctx {
|
||||
};
|
||||
|
||||
|
||||
static int serpent_setkey(void *ctx, const u8 *key, unsigned int keylen, u32 *flags)
|
||||
static int serpent_setkey(struct crypto_tfm *tfm, const u8 *key,
|
||||
unsigned int keylen, u32 *flags)
|
||||
{
|
||||
u32 *k = ((struct serpent_ctx *)ctx)->expkey;
|
||||
struct serpent_ctx *ctx = crypto_tfm_ctx(tfm);
|
||||
u32 *k = ctx->expkey;
|
||||
u8 *k8 = (u8 *)k;
|
||||
u32 r0,r1,r2,r3,r4;
|
||||
int i;
|
||||
@@ -365,10 +367,11 @@ static int serpent_setkey(void *ctx, const u8 *key, unsigned int keylen, u32 *fl
|
||||
return 0;
|
||||
}
|
||||
|
||||
static void serpent_encrypt(void *ctx, u8 *dst, const u8 *src)
|
||||
static void serpent_encrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src)
|
||||
{
|
||||
struct serpent_ctx *ctx = crypto_tfm_ctx(tfm);
|
||||
const u32
|
||||
*k = ((struct serpent_ctx *)ctx)->expkey,
|
||||
*k = ctx->expkey,
|
||||
*s = (const u32 *)src;
|
||||
u32 *d = (u32 *)dst,
|
||||
r0, r1, r2, r3, r4;
|
||||
@@ -423,8 +426,9 @@ static void serpent_encrypt(void *ctx, u8 *dst, const u8 *src)
|
||||
d[3] = cpu_to_le32(r3);
|
||||
}
|
||||
|
||||
static void serpent_decrypt(void *ctx, u8 *dst, const u8 *src)
|
||||
static void serpent_decrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src)
|
||||
{
|
||||
struct serpent_ctx *ctx = crypto_tfm_ctx(tfm);
|
||||
const u32
|
||||
*k = ((struct serpent_ctx *)ctx)->expkey,
|
||||
*s = (const u32 *)src;
|
||||
@@ -492,7 +496,8 @@ static struct crypto_alg serpent_alg = {
|
||||
.cia_decrypt = serpent_decrypt } }
|
||||
};
|
||||
|
||||
static int tnepres_setkey(void *ctx, const u8 *key, unsigned int keylen, u32 *flags)
|
||||
static int tnepres_setkey(struct crypto_tfm *tfm, const u8 *key,
|
||||
unsigned int keylen, u32 *flags)
|
||||
{
|
||||
u8 rev_key[SERPENT_MAX_KEY_SIZE];
|
||||
int i;
|
||||
@@ -506,10 +511,10 @@ static int tnepres_setkey(void *ctx, const u8 *key, unsigned int keylen, u32 *fl
|
||||
for (i = 0; i < keylen; ++i)
|
||||
rev_key[keylen - i - 1] = key[i];
|
||||
|
||||
return serpent_setkey(ctx, rev_key, keylen, flags);
|
||||
return serpent_setkey(tfm, rev_key, keylen, flags);
|
||||
}
|
||||
|
||||
static void tnepres_encrypt(void *ctx, u8 *dst, const u8 *src)
|
||||
static void tnepres_encrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src)
|
||||
{
|
||||
const u32 * const s = (const u32 * const)src;
|
||||
u32 * const d = (u32 * const)dst;
|
||||
@@ -521,7 +526,7 @@ static void tnepres_encrypt(void *ctx, u8 *dst, const u8 *src)
|
||||
rs[2] = swab32(s[1]);
|
||||
rs[3] = swab32(s[0]);
|
||||
|
||||
serpent_encrypt(ctx, (u8 *)rd, (u8 *)rs);
|
||||
serpent_encrypt(tfm, (u8 *)rd, (u8 *)rs);
|
||||
|
||||
d[0] = swab32(rd[3]);
|
||||
d[1] = swab32(rd[2]);
|
||||
@@ -529,7 +534,7 @@ static void tnepres_encrypt(void *ctx, u8 *dst, const u8 *src)
|
||||
d[3] = swab32(rd[0]);
|
||||
}
|
||||
|
||||
static void tnepres_decrypt(void *ctx, u8 *dst, const u8 *src)
|
||||
static void tnepres_decrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src)
|
||||
{
|
||||
const u32 * const s = (const u32 * const)src;
|
||||
u32 * const d = (u32 * const)dst;
|
||||
@@ -541,7 +546,7 @@ static void tnepres_decrypt(void *ctx, u8 *dst, const u8 *src)
|
||||
rs[2] = swab32(s[1]);
|
||||
rs[3] = swab32(s[0]);
|
||||
|
||||
serpent_decrypt(ctx, (u8 *)rd, (u8 *)rs);
|
||||
serpent_decrypt(tfm, (u8 *)rd, (u8 *)rs);
|
||||
|
||||
d[0] = swab32(rd[3]);
|
||||
d[1] = swab32(rd[2]);
|
||||
|
||||
+9
-8
@@ -34,9 +34,9 @@ struct sha1_ctx {
|
||||
u8 buffer[64];
|
||||
};
|
||||
|
||||
static void sha1_init(void *ctx)
|
||||
static void sha1_init(struct crypto_tfm *tfm)
|
||||
{
|
||||
struct sha1_ctx *sctx = ctx;
|
||||
struct sha1_ctx *sctx = crypto_tfm_ctx(tfm);
|
||||
static const struct sha1_ctx initstate = {
|
||||
0,
|
||||
{ 0x67452301, 0xEFCDAB89, 0x98BADCFE, 0x10325476, 0xC3D2E1F0 },
|
||||
@@ -46,9 +46,10 @@ static void sha1_init(void *ctx)
|
||||
*sctx = initstate;
|
||||
}
|
||||
|
||||
static void sha1_update(void *ctx, const u8 *data, unsigned int len)
|
||||
static void sha1_update(struct crypto_tfm *tfm, const u8 *data,
|
||||
unsigned int len)
|
||||
{
|
||||
struct sha1_ctx *sctx = ctx;
|
||||
struct sha1_ctx *sctx = crypto_tfm_ctx(tfm);
|
||||
unsigned int partial, done;
|
||||
const u8 *src;
|
||||
|
||||
@@ -80,9 +81,9 @@ static void sha1_update(void *ctx, const u8 *data, unsigned int len)
|
||||
|
||||
|
||||
/* Add padding and return the message digest. */
|
||||
static void sha1_final(void* ctx, u8 *out)
|
||||
static void sha1_final(struct crypto_tfm *tfm, u8 *out)
|
||||
{
|
||||
struct sha1_ctx *sctx = ctx;
|
||||
struct sha1_ctx *sctx = crypto_tfm_ctx(tfm);
|
||||
__be32 *dst = (__be32 *)out;
|
||||
u32 i, index, padlen;
|
||||
__be64 bits;
|
||||
@@ -93,10 +94,10 @@ static void sha1_final(void* ctx, u8 *out)
|
||||
/* Pad out to 56 mod 64 */
|
||||
index = sctx->count & 0x3f;
|
||||
padlen = (index < 56) ? (56 - index) : ((64+56) - index);
|
||||
sha1_update(sctx, padding, padlen);
|
||||
sha1_update(tfm, padding, padlen);
|
||||
|
||||
/* Append length */
|
||||
sha1_update(sctx, (const u8 *)&bits, sizeof(bits));
|
||||
sha1_update(tfm, (const u8 *)&bits, sizeof(bits));
|
||||
|
||||
/* Store state in digest */
|
||||
for (i = 0; i < 5; i++)
|
||||
|
||||
+9
-8
@@ -230,9 +230,9 @@ static void sha256_transform(u32 *state, const u8 *input)
|
||||
memset(W, 0, 64 * sizeof(u32));
|
||||
}
|
||||
|
||||
static void sha256_init(void *ctx)
|
||||
static void sha256_init(struct crypto_tfm *tfm)
|
||||
{
|
||||
struct sha256_ctx *sctx = ctx;
|
||||
struct sha256_ctx *sctx = crypto_tfm_ctx(tfm);
|
||||
sctx->state[0] = H0;
|
||||
sctx->state[1] = H1;
|
||||
sctx->state[2] = H2;
|
||||
@@ -244,9 +244,10 @@ static void sha256_init(void *ctx)
|
||||
sctx->count[0] = sctx->count[1] = 0;
|
||||
}
|
||||
|
||||
static void sha256_update(void *ctx, const u8 *data, unsigned int len)
|
||||
static void sha256_update(struct crypto_tfm *tfm, const u8 *data,
|
||||
unsigned int len)
|
||||
{
|
||||
struct sha256_ctx *sctx = ctx;
|
||||
struct sha256_ctx *sctx = crypto_tfm_ctx(tfm);
|
||||
unsigned int i, index, part_len;
|
||||
|
||||
/* Compute number of bytes mod 128 */
|
||||
@@ -276,9 +277,9 @@ static void sha256_update(void *ctx, const u8 *data, unsigned int len)
|
||||
memcpy(&sctx->buf[index], &data[i], len-i);
|
||||
}
|
||||
|
||||
static void sha256_final(void* ctx, u8 *out)
|
||||
static void sha256_final(struct crypto_tfm *tfm, u8 *out)
|
||||
{
|
||||
struct sha256_ctx *sctx = ctx;
|
||||
struct sha256_ctx *sctx = crypto_tfm_ctx(tfm);
|
||||
__be32 *dst = (__be32 *)out;
|
||||
__be32 bits[2];
|
||||
unsigned int index, pad_len;
|
||||
@@ -292,10 +293,10 @@ static void sha256_final(void* ctx, u8 *out)
|
||||
/* Pad out to 56 mod 64. */
|
||||
index = (sctx->count[0] >> 3) & 0x3f;
|
||||
pad_len = (index < 56) ? (56 - index) : ((64+56) - index);
|
||||
sha256_update(sctx, padding, pad_len);
|
||||
sha256_update(tfm, padding, pad_len);
|
||||
|
||||
/* Append length (before padding) */
|
||||
sha256_update(sctx, (const u8 *)bits, sizeof(bits));
|
||||
sha256_update(tfm, (const u8 *)bits, sizeof(bits));
|
||||
|
||||
/* Store state in digest */
|
||||
for (i = 0; i < 8; i++)
|
||||
|
||||
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user