/linux-4.1.27/crypto/ |
H A D | seqiv.c | 85 unsigned int ivsize) seqiv_geniv() 87 unsigned int len = ivsize; seqiv_geniv() 89 if (ivsize > sizeof(u64)) { seqiv_geniv() 90 memset(info, 0, ivsize - sizeof(u64)); seqiv_geniv() 94 memcpy(info + ivsize - len, &seq, len); seqiv_geniv() 95 crypto_xor(info, ctx->salt, ivsize); seqiv_geniv() 106 unsigned int ivsize; seqiv_givencrypt() local 115 ivsize = crypto_ablkcipher_ivsize(geniv); seqiv_givencrypt() 119 info = kmalloc(ivsize, req->creq.base.flags & seqiv_givencrypt() 134 seqiv_geniv(ctx, info, req->seq, ivsize); seqiv_givencrypt() 135 memcpy(req->giv, info, ivsize); seqiv_givencrypt() 152 unsigned int ivsize; seqiv_aead_givencrypt() local 161 ivsize = crypto_aead_ivsize(geniv); seqiv_aead_givencrypt() 165 info = kmalloc(ivsize, areq->base.flags & seqiv_aead_givencrypt() 180 seqiv_geniv(ctx, info, req->seq, ivsize); seqiv_aead_givencrypt() 181 memcpy(req->giv, info, ivsize); seqiv_aead_givencrypt() 270 if (inst->alg.cra_ablkcipher.ivsize < sizeof(u64)) { seqiv_ablkcipher_alloc() 281 inst->alg.cra_ctxsize += inst->alg.cra_ablkcipher.ivsize; seqiv_ablkcipher_alloc() 296 if (inst->alg.cra_aead.ivsize < sizeof(u64)) { seqiv_aead_alloc() 307 inst->alg.cra_ctxsize = inst->alg.cra_aead.ivsize; seqiv_aead_alloc() 84 seqiv_geniv(struct seqiv_ctx *ctx, u8 *info, u64 seq, unsigned int ivsize) seqiv_geniv() argument
|
H A D | eseqiv.c | 81 unsigned int ivsize; eseqiv_givencrypt() local 99 ivsize = crypto_ablkcipher_ivsize(geniv); eseqiv_givencrypt() 101 if (vsrc != giv + ivsize && vdst != giv + ivsize) { eseqiv_givencrypt() 112 sg_set_buf(reqctx->src, giv, ivsize); eseqiv_givencrypt() 113 scatterwalk_crypto_chain(reqctx->src, osrc, vsrc == giv + ivsize, 2); eseqiv_givencrypt() 118 sg_set_buf(reqctx->dst, giv, ivsize); eseqiv_givencrypt() 119 scatterwalk_crypto_chain(reqctx->dst, odst, vdst == giv + ivsize, 2); eseqiv_givencrypt() 125 req->creq.nbytes + ivsize, eseqiv_givencrypt() 128 memcpy(req->creq.info, ctx->salt, ivsize); eseqiv_givencrypt() 130 len = ivsize; eseqiv_givencrypt() 131 if (ivsize > sizeof(u64)) { eseqiv_givencrypt() 132 memset(req->giv, 0, ivsize - sizeof(u64)); eseqiv_givencrypt() 136 memcpy(req->giv + ivsize - len, &seq, len); eseqiv_givencrypt() 220 if (inst->alg.cra_ablkcipher.ivsize != inst->alg.cra_blocksize) eseqiv_alloc() 229 inst->alg.cra_ctxsize += inst->alg.cra_ablkcipher.ivsize; eseqiv_alloc()
|
H A D | ablkcipher.c | 142 memcpy(req->info, walk->iv, tfm->crt_ablkcipher.ivsize); ablkcipher_walk_done() 194 unsigned int ivsize = tfm->crt_ablkcipher.ivsize; ablkcipher_copy_iv() local 196 unsigned int size = aligned_bs * 2 + ivsize + max(aligned_bs, ivsize) - ablkcipher_copy_iv() 208 iv = ablkcipher_get_spot(iv, ivsize); ablkcipher_copy_iv() 210 walk->iv = memcpy(iv, walk->iv, ivsize); ablkcipher_copy_iv() 368 if (alg->ivsize > PAGE_SIZE / 8) crypto_init_ablkcipher_ops() 374 if (!alg->ivsize) { crypto_init_ablkcipher_ops() 379 crt->ivsize = alg->ivsize; crypto_init_ablkcipher_ops() 396 rblkcipher.ivsize = alg->cra_ablkcipher.ivsize; crypto_ablkcipher_report() 425 seq_printf(m, "ivsize : %u\n", ablkcipher->ivsize); crypto_ablkcipher_show() 450 if (alg->ivsize > PAGE_SIZE / 8) crypto_init_givcipher_ops() 460 crt->ivsize = alg->ivsize; crypto_init_givcipher_ops() 477 rblkcipher.ivsize = alg->cra_ablkcipher.ivsize; crypto_givcipher_report() 506 seq_printf(m, "ivsize : %u\n", ablkcipher->ivsize); crypto_givcipher_show() 523 CRYPTO_ALG_TYPE_BLKCIPHER ? alg->cra_blkcipher.ivsize : crypto_default_geniv() 524 alg->cra_ablkcipher.ivsize) != crypto_default_geniv() 627 CRYPTO_ALG_TYPE_BLKCIPHER ? alg->cra_blkcipher.ivsize : crypto_lookup_skcipher() 628 alg->cra_ablkcipher.ivsize)) crypto_lookup_skcipher() 647 CRYPTO_ALG_TYPE_BLKCIPHER ? alg->cra_blkcipher.ivsize : crypto_lookup_skcipher() 648 alg->cra_ablkcipher.ivsize)); crypto_lookup_skcipher()
|
H A D | aead.c | 98 if (max(alg->maxauthsize, alg->ivsize) > PAGE_SIZE / 8) crypto_init_aead_ops() 108 crt->ivsize = alg->ivsize; crypto_init_aead_ops() 125 raead.ivsize = aead->ivsize; crypto_aead_report() 152 seq_printf(m, "ivsize : %u\n", aead->ivsize); crypto_aead_show() 182 if (max(alg->maxauthsize, alg->ivsize) > PAGE_SIZE / 8) crypto_init_nivaead_ops() 188 if (!alg->ivsize) { crypto_init_nivaead_ops() 193 crt->ivsize = alg->ivsize; crypto_init_nivaead_ops() 210 raead.ivsize = aead->ivsize; crypto_nivaead_report() 238 seq_printf(m, "ivsize : %u\n", aead->ivsize); crypto_nivaead_show() 312 if (!alg->cra_aead.ivsize) aead_geniv_alloc() 346 inst->alg.cra_aead.ivsize = alg->cra_aead.ivsize; aead_geniv_alloc() 482 if (!alg->cra_aead.ivsize) crypto_lookup_aead() 499 BUG_ON(!alg->cra_aead.ivsize); crypto_lookup_aead()
|
H A D | blkcipher.c | 136 memcpy(desc->info, walk->iv, walk->ivsize); blkcipher_walk_done() 280 walk->ivsize + max(aligned_bs, walk->ivsize) - blkcipher_copy_iv() 292 iv = blkcipher_get_spot(iv, walk->ivsize); blkcipher_copy_iv() 294 walk->iv = memcpy(iv, walk->iv, walk->ivsize); blkcipher_copy_iv() 304 walk->ivsize = crypto_blkcipher_ivsize(desc->tfm); blkcipher_walk_virt() 316 walk->ivsize = crypto_blkcipher_ivsize(desc->tfm); blkcipher_walk_phys() 354 walk->ivsize = crypto_blkcipher_ivsize(desc->tfm); blkcipher_walk_virt_block() 368 walk->ivsize = crypto_aead_ivsize(tfm); blkcipher_aead_walk_virt_block() 452 cipher->ivsize) { crypto_blkcipher_ctxsize() 454 len += cipher->ivsize; crypto_blkcipher_ctxsize() 468 if (!alg->ivsize) { crypto_init_blkcipher_ops_async() 473 crt->ivsize = alg->ivsize; crypto_init_blkcipher_ops_async() 501 if (alg->ivsize > PAGE_SIZE / 8) crypto_init_blkcipher_ops() 522 rblkcipher.ivsize = alg->cra_blkcipher.ivsize; crypto_blkcipher_report() 547 seq_printf(m, "ivsize : %u\n", alg->cra_blkcipher.ivsize); crypto_blkcipher_show() 592 unsigned int ivsize; skcipher_geniv_alloc() member in struct:__anon3250 633 balg.ivsize = alg->cra_blkcipher.ivsize; skcipher_geniv_alloc() 643 balg.ivsize = alg->cra_ablkcipher.ivsize; skcipher_geniv_alloc() 655 if (!balg.ivsize) skcipher_geniv_alloc() 692 inst->alg.cra_ablkcipher.ivsize = balg.ivsize; skcipher_geniv_alloc()
|
H A D | chainiv.c | 53 unsigned int ivsize; chainiv_givencrypt() local 66 ivsize = crypto_ablkcipher_ivsize(geniv); chainiv_givencrypt() 68 memcpy(req->giv, ctx->iv, ivsize); chainiv_givencrypt() 69 memcpy(subreq->info, ctx->iv, ivsize); chainiv_givencrypt() 75 memcpy(ctx->iv, subreq->info, ivsize); chainiv_givencrypt() 166 unsigned int ivsize = crypto_ablkcipher_ivsize(geniv); async_chainiv_givencrypt_tail() local 168 memcpy(req->giv, ctx->iv, ivsize); async_chainiv_givencrypt_tail() 169 memcpy(subreq->info, ctx->iv, ivsize); async_chainiv_givencrypt_tail() 175 memcpy(ctx->iv, subreq->info, ivsize); async_chainiv_givencrypt_tail() 324 inst->alg.cra_ctxsize += inst->alg.cra_ablkcipher.ivsize; chainiv_alloc()
|
H A D | authenc.c | 333 unsigned int ivsize = crypto_aead_ivsize(authenc); crypto_authenc_genicv() local 343 if (ivsize) { crypto_authenc_genicv() 345 sg_set_buf(cipher, iv, ivsize); crypto_authenc_genicv() 346 scatterwalk_crypto_chain(cipher, dst, vdst == iv + ivsize, 2); crypto_authenc_genicv() 348 cryptlen += ivsize; crypto_authenc_genicv() 490 unsigned int ivsize = crypto_aead_ivsize(authenc); crypto_authenc_iverify() local 498 if (ivsize) { crypto_authenc_iverify() 500 sg_set_buf(cipher, iv, ivsize); crypto_authenc_iverify() 501 scatterwalk_crypto_chain(cipher, src, vsrc == iv + ivsize, 2); crypto_authenc_iverify() 503 cryptlen += ivsize; crypto_authenc_iverify() 665 inst->alg.cra_aead.ivsize = enc->cra_ablkcipher.ivsize; crypto_authenc_alloc()
|
H A D | authencesn.c | 398 unsigned int ivsize = crypto_aead_ivsize(authenc_esn); crypto_authenc_esn_genicv() local 407 if (ivsize) { crypto_authenc_esn_genicv() 409 sg_set_buf(cipher, iv, ivsize); crypto_authenc_esn_genicv() 410 scatterwalk_crypto_chain(cipher, dst, vdst == iv + ivsize, 2); crypto_authenc_esn_genicv() 412 cryptlen += ivsize; crypto_authenc_esn_genicv() 568 unsigned int ivsize = crypto_aead_ivsize(authenc_esn); crypto_authenc_esn_iverify() local 575 if (ivsize) { crypto_authenc_esn_iverify() 577 sg_set_buf(cipher, iv, ivsize); crypto_authenc_esn_iverify() 578 scatterwalk_crypto_chain(cipher, src, vsrc == iv + ivsize, 2); crypto_authenc_esn_iverify() 580 cryptlen += ivsize; crypto_authenc_esn_iverify() 757 inst->alg.cra_aead.ivsize = enc->cra_ablkcipher.ivsize; crypto_authenc_esn_alloc()
|
H A D | algboss.c | 254 CRYPTO_ALG_TYPE_BLKCIPHER ? alg->cra_blkcipher.ivsize : cryptomgr_schedule_test() 255 alg->cra_ablkcipher.ivsize)) || cryptomgr_schedule_test() 257 alg->cra_type == &crypto_nivaead_type && alg->cra_aead.ivsize)) cryptomgr_schedule_test()
|
H A D | ctr.c | 215 inst->alg.cra_blkcipher.ivsize = alg->cra_blocksize; crypto_ctr_alloc() 373 if (alg->cra_ablkcipher.ivsize != CTR_RFC3686_BLOCK_SIZE) crypto_rfc3686_alloc() 397 inst->alg.cra_ablkcipher.ivsize = CTR_RFC3686_IV_SIZE; crypto_rfc3686_alloc()
|
H A D | crypto_null.c | 133 .ivsize = NULL_IV_SIZE,
|
H A D | algif_aead.c | 171 unsigned ivsize = aead_sendmsg() local 197 if (con.iv && con.iv->ivlen != ivsize) aead_sendmsg() 208 memcpy(ctx->iv, con.iv->iv, ivsize); aead_sendmsg()
|
H A D | ccm.c | 530 if (ctr->cra_ablkcipher.ivsize != 16) crypto_ccm_alloc_common() 548 inst->alg.cra_aead.ivsize = 16; crypto_ccm_alloc_common() 783 if (alg->cra_aead.ivsize != 16) crypto_rfc4309_alloc() 805 inst->alg.cra_aead.ivsize = 8; crypto_rfc4309_alloc()
|
H A D | cryptd.c | 367 inst->alg.cra_ablkcipher.ivsize = alg->cra_blkcipher.ivsize; cryptd_create_blkcipher() 783 inst->alg.cra_aead.ivsize = alg->cra_aead.ivsize; cryptd_create_aead()
|
H A D | salsa20_generic.c | 231 .ivsize = SALSA20_IV_SIZE,
|
H A D | gcm.c | 745 if (ctr->cra_ablkcipher.ivsize != 16) crypto_gcm_alloc_common() 768 inst->alg.cra_aead.ivsize = 16; crypto_gcm_alloc_common() 1000 if (alg->cra_aead.ivsize != 16) crypto_rfc4106_alloc() 1022 inst->alg.cra_aead.ivsize = 8; crypto_rfc4106_alloc() 1325 if (alg->cra_aead.ivsize != 16) crypto_rfc4543_alloc() 1347 inst->alg.cra_aead.ivsize = 8; crypto_rfc4543_alloc()
|
H A D | algif_skcipher.c | 311 unsigned ivsize = crypto_ablkcipher_ivsize(tfm); skcipher_sendmsg() local 337 if (con.iv && con.iv->ivlen != ivsize) skcipher_sendmsg() 350 memcpy(ctx->iv, con.iv->iv, ivsize); skcipher_sendmsg()
|
H A D | pcrypt.c | 364 inst->alg.cra_aead.ivsize = alg->cra_aead.ivsize; pcrypt_alloc_aead()
|
H A D | cbc.c | 246 inst->alg.cra_blkcipher.ivsize = alg->cra_blocksize; crypto_cbc_alloc()
|
H A D | lrw.c | 355 inst->alg.cra_blkcipher.ivsize = alg->cra_blocksize; alloc()
|
H A D | pcbc.c | 252 inst->alg.cra_blkcipher.ivsize = alg->cra_blocksize; crypto_pcbc_alloc()
|
H A D | xts.c | 317 inst->alg.cra_blkcipher.ivsize = alg->cra_blocksize; alloc()
|
H A D | cts.c | 309 inst->alg.cra_blkcipher.ivsize = alg->cra_blocksize; crypto_cts_alloc()
|
/linux-4.1.27/drivers/crypto/caam/ |
H A D | caamalg.c | 149 static inline void aead_append_ld_iv(u32 *desc, int ivsize, int ivoffset) aead_append_ld_iv() argument 151 append_seq_load(desc, ivsize, LDST_CLASS_1_CCB | aead_append_ld_iv() 155 (ivoffset << MOVE_OFFSET_SHIFT) | ivsize); aead_append_ld_iv() 384 /* assoclen + cryptlen = seqinlen - ivsize - authsize */ aead_null_set_sh_desc() 386 ctx->authsize + tfm->ivsize); aead_null_set_sh_desc() 512 /* assoclen + cryptlen = seqinlen - ivsize */ aead_set_sh_desc() 513 append_math_sub_imm_u32(desc, REG2, SEQINLEN, IMM, tfm->ivsize); aead_set_sh_desc() 521 aead_append_ld_iv(desc, tfm->ivsize, ctx1_iv_off); aead_set_sh_desc() 578 /* assoclen + cryptlen = seqinlen - ivsize - authsize */ aead_set_sh_desc() 580 ctx->authsize + tfm->ivsize); aead_set_sh_desc() 589 aead_append_ld_iv(desc, tfm->ivsize, ctx1_iv_off); aead_set_sh_desc() 648 NFIFOENTRY_PTYPE_RND | (tfm->ivsize << NFIFOENTRY_DLEN_SHIFT); aead_set_sh_desc() 655 (tfm->ivsize << MOVE_LEN_SHIFT)); aead_set_sh_desc() 661 (tfm->ivsize << MOVE_LEN_SHIFT)); aead_set_sh_desc() 667 /* ivsize + cryptlen = seqoutlen - authsize */ aead_set_sh_desc() 670 /* assoclen = seqinlen - (ivsize + cryptlen) */ aead_set_sh_desc() 679 NFIFOENTRY_DTYPE_MSG | (tfm->ivsize << NFIFOENTRY_DLEN_SHIFT); aead_set_sh_desc() 682 append_load_imm_u32(desc, tfm->ivsize, LDST_CLASS_2_CCB | aead_set_sh_desc() 697 /* Will write ivsize + cryptlen */ aead_set_sh_desc() 701 append_seq_fifo_load(desc, tfm->ivsize, aead_set_sh_desc() 783 /* assoclen + cryptlen = seqinlen - ivsize */ gcm_set_sh_desc() 784 append_math_sub_imm_u32(desc, REG2, SEQINLEN, IMM, tfm->ivsize); gcm_set_sh_desc() 794 append_seq_fifo_load(desc, tfm->ivsize, FIFOLD_CLASS_CLASS1 | gcm_set_sh_desc() 827 append_seq_fifo_load(desc, tfm->ivsize, FIFOLD_CLASS_CLASS1 | gcm_set_sh_desc() 839 append_seq_fifo_load(desc, tfm->ivsize, FIFOLD_CLASS_CLASS1 | gcm_set_sh_desc() 889 /* assoclen + cryptlen = seqinlen - ivsize - icvsize */ gcm_set_sh_desc() 891 ctx->authsize + tfm->ivsize); gcm_set_sh_desc() 898 append_seq_fifo_load(desc, tfm->ivsize, FIFOLD_CLASS_CLASS1 | gcm_set_sh_desc() 1014 /* assoclen + cryptlen = seqinlen - ivsize */ rfc4106_set_sh_desc() 1015 append_math_sub_imm_u32(desc, REG2, SEQINLEN, IMM, tfm->ivsize); rfc4106_set_sh_desc() 1024 append_seq_fifo_load(desc, tfm->ivsize, FIFOLD_CLASS_CLASS1 | rfc4106_set_sh_desc() 1086 /* assoclen + cryptlen = seqinlen - ivsize - icvsize */ rfc4106_set_sh_desc() 1088 ctx->authsize + tfm->ivsize); rfc4106_set_sh_desc() 1101 append_seq_fifo_load(desc, tfm->ivsize, FIFOLD_CLASS_CLASS1 | rfc4106_set_sh_desc() 1164 NFIFOENTRY_PTYPE_RND | (tfm->ivsize << NFIFOENTRY_DLEN_SHIFT); rfc4106_set_sh_desc() 1169 (tfm->ivsize << MOVE_LEN_SHIFT)); rfc4106_set_sh_desc() 1174 (tfm->ivsize << MOVE_LEN_SHIFT)); rfc4106_set_sh_desc() 1180 /* ivsize + cryptlen = seqoutlen - authsize */ rfc4106_set_sh_desc() 1183 /* assoclen = seqinlen - (ivsize + cryptlen) */ rfc4106_set_sh_desc() 1186 /* Will write ivsize + cryptlen */ rfc4106_set_sh_desc() 1202 append_seq_fifo_load(desc, tfm->ivsize, FIFOLD_CLASS_SKIP); rfc4106_set_sh_desc() 1294 LDST_CLASS_DECO | tfm->ivsize); rfc4543_set_sh_desc() 1302 (tfm->ivsize << MOVE_LEN_SHIFT)); rfc4543_set_sh_desc() 1306 (tfm->ivsize << MOVE_LEN_SHIFT)); rfc4543_set_sh_desc() 1311 /* assoclen = (seqinlen - ivsize) - cryptlen */ rfc4543_set_sh_desc() 1316 FIFOLD_TYPE_IV | FIFOLD_TYPE_FLUSH1 | (4 + tfm->ivsize)); rfc4543_set_sh_desc() 1347 FIFOLD_TYPE_AAD | tfm->ivsize); rfc4543_set_sh_desc() 1410 LDST_CLASS_DECO | tfm->ivsize); rfc4543_set_sh_desc() 1416 /* assoclen + cryptlen = (seqinlen - ivsize) - icvsize */ rfc4543_set_sh_desc() 1421 (tfm->ivsize << MOVE_LEN_SHIFT)); rfc4543_set_sh_desc() 1425 (tfm->ivsize << MOVE_LEN_SHIFT)); rfc4543_set_sh_desc() 1443 FIFOLD_TYPE_IV | FIFOLD_TYPE_FLUSH1 | (4 + tfm->ivsize)); rfc4543_set_sh_desc() 1464 FIFOLD_TYPE_AAD | tfm->ivsize); rfc4543_set_sh_desc() 1530 NFIFOENTRY_PTYPE_RND | (tfm->ivsize << NFIFOENTRY_DLEN_SHIFT); rfc4543_set_sh_desc() 1536 (tfm->ivsize << MOVE_LEN_SHIFT)); rfc4543_set_sh_desc() 1541 (tfm->ivsize << MOVE_LEN_SHIFT)); rfc4543_set_sh_desc() 1545 (tfm->ivsize << MOVE_LEN_SHIFT)); rfc4543_set_sh_desc() 1549 (tfm->ivsize << MOVE_LEN_SHIFT)); rfc4543_set_sh_desc() 1555 /* ivsize + cryptlen = seqoutlen - authsize */ rfc4543_set_sh_desc() 1558 /* assoclen = seqinlen - (ivsize + cryptlen) */ rfc4543_set_sh_desc() 1561 /* Will write ivsize + cryptlen */ rfc4543_set_sh_desc() 1576 FIFOLD_TYPE_IV | FIFOLD_TYPE_FLUSH1 | (4 + tfm->ivsize)); rfc4543_set_sh_desc() 1586 append_seq_fifo_load(desc, tfm->ivsize, FIFOLD_CLASS_SKIP); rfc4543_set_sh_desc() 1597 FIFOLD_TYPE_AAD | tfm->ivsize); rfc4543_set_sh_desc() 1903 append_seq_load(desc, crt->ivsize, LDST_SRCDST_BYTE_CONTEXT | ablkcipher_setkey() 1962 append_seq_load(desc, crt->ivsize, LDST_SRCDST_BYTE_CONTEXT | ablkcipher_setkey() 2026 NFIFOENTRY_PTYPE_RND | (crt->ivsize << NFIFOENTRY_DLEN_SHIFT); ablkcipher_setkey() 2033 (crt->ivsize << MOVE_LEN_SHIFT) | ablkcipher_setkey() 2038 append_seq_store(desc, crt->ivsize, ablkcipher_setkey() 2133 dma_addr_t iv_dma, int ivsize, dma_addr_t sec4_sg_dma, caam_unmap() 2147 dma_unmap_single(dev, iv_dma, ivsize, DMA_TO_DEVICE); caam_unmap() 2158 int ivsize = crypto_aead_ivsize(aead); aead_unmap() local 2165 edesc->dst_chained, edesc->iv_dma, ivsize, aead_unmap() 2174 int ivsize = crypto_ablkcipher_ivsize(ablkcipher); ablkcipher_unmap() local 2178 edesc->dst_chained, edesc->iv_dma, ivsize, ablkcipher_unmap() 2190 int ivsize = crypto_aead_ivsize(aead); aead_encrypt_done() local 2208 DUMP_PREFIX_ADDRESS, 16, 4, sg_virt(req->src) - ivsize, aead_encrypt_done() 2209 edesc->src_nents ? 100 : ivsize, 1); aead_encrypt_done() 2229 int ivsize = crypto_aead_ivsize(aead); aead_decrypt_done() local 2240 ivsize, 1); aead_decrypt_done() 2284 int ivsize = crypto_ablkcipher_ivsize(ablkcipher); ablkcipher_encrypt_done() local 2298 edesc->src_nents > 1 ? 100 : ivsize, 1); ablkcipher_encrypt_done() 2317 int ivsize = crypto_ablkcipher_ivsize(ablkcipher); ablkcipher_decrypt_done() local 2330 ivsize, 1); ablkcipher_decrypt_done() 2352 int ivsize = crypto_aead_ivsize(aead); init_aead_job() local 2368 edesc->src_nents ? 100 : ivsize, 1); init_aead_job() 2398 append_seq_in_ptr(desc, src_dma, req->assoclen + ivsize + req->cryptlen, init_aead_job() 2437 int ivsize = crypto_aead_ivsize(aead); init_aead_giv_job() local 2452 DUMP_PREFIX_ADDRESS, 16, 4, req->iv, ivsize, 1); init_aead_giv_job() 2480 append_seq_in_ptr(desc, src_dma, req->assoclen + ivsize + req->cryptlen, init_aead_giv_job() 2499 append_seq_out_ptr(desc, dst_dma, ivsize + req->cryptlen + authsize, init_aead_giv_job() 2512 int ivsize = crypto_ablkcipher_ivsize(ablkcipher); init_ablkcipher_job() local 2521 ivsize, 1); init_ablkcipher_job() 2538 append_seq_in_ptr(desc, src_dma, req->nbytes + ivsize, in_options); init_ablkcipher_job() 2569 int ivsize = crypto_ablkcipher_ivsize(ablkcipher); init_ablkcipher_giv_job() local 2578 ivsize, 1); init_ablkcipher_giv_job() 2605 append_seq_out_ptr(desc, dst_dma, req->nbytes + ivsize, out_options); init_ablkcipher_giv_job() 2626 int ivsize = crypto_aead_ivsize(aead); aead_edesc_alloc() local 2658 iv_dma = dma_map_single(jrdev, req->iv, ivsize, DMA_TO_DEVICE); aead_edesc_alloc() 2676 iv_dma + ivsize == sg_dma_address(req->assoc) && aead_edesc_alloc() 2682 iv_dma + ivsize == sg_dma_address(req->src)); aead_edesc_alloc() 2724 iv_dma, ivsize, 0); aead_edesc_alloc() 2853 int ivsize = crypto_aead_ivsize(aead); aead_giv_edesc_alloc() local 2877 iv_dma = dma_map_single(jrdev, greq->giv, ivsize, DMA_TO_DEVICE); aead_giv_edesc_alloc() 2895 if (assoc_nents || iv_dma + ivsize != aead_giv_edesc_alloc() 2903 src_nents || iv_dma + ivsize != sg_dma_address(req->src)) aead_giv_edesc_alloc() 2907 if (dst_nents || iv_dma + ivsize != sg_dma_address(req->dst)) aead_giv_edesc_alloc() 2915 (src_nents || iv_dma + ivsize != sg_dma_address(req->src))) aead_giv_edesc_alloc() 2962 iv_dma, ivsize, 0); aead_giv_edesc_alloc() 2979 iv_dma, ivsize, 0); aead_giv_edesc_alloc() 2987 iv_dma, ivsize, 0); aead_giv_edesc_alloc() 3070 int ivsize = crypto_ablkcipher_ivsize(ablkcipher); ablkcipher_edesc_alloc() local 3089 iv_dma = dma_map_single(jrdev, req->info, ivsize, DMA_TO_DEVICE); ablkcipher_edesc_alloc() 3099 if (!src_nents && iv_dma + ivsize == sg_dma_address(req->src)) ablkcipher_edesc_alloc() 3124 dma_to_sec4_sg_one(edesc->sec4_sg, iv_dma, ivsize, 0); ablkcipher_edesc_alloc() 3249 int ivsize = crypto_ablkcipher_ivsize(ablkcipher); ablkcipher_giv_edesc_alloc() local 3272 iv_dma = dma_map_single(jrdev, greq->giv, ivsize, DMA_TO_DEVICE); ablkcipher_giv_edesc_alloc() 3278 if (!dst_nents && iv_dma + ivsize == sg_dma_address(req->dst)) ablkcipher_giv_edesc_alloc() 3309 iv_dma, ivsize, 0); ablkcipher_giv_edesc_alloc() 3407 .ivsize = NULL_IV_SIZE, 3426 .ivsize = NULL_IV_SIZE, 3445 .ivsize = NULL_IV_SIZE, 3465 .ivsize = NULL_IV_SIZE, 3485 .ivsize = NULL_IV_SIZE, 3505 .ivsize = NULL_IV_SIZE, 3525 .ivsize = AES_BLOCK_SIZE, 3544 .ivsize = AES_BLOCK_SIZE, 3563 .ivsize = AES_BLOCK_SIZE, 3583 .ivsize = AES_BLOCK_SIZE, 3603 .ivsize = AES_BLOCK_SIZE, 3624 .ivsize = AES_BLOCK_SIZE, 3644 .ivsize = DES3_EDE_BLOCK_SIZE, 3663 .ivsize = DES3_EDE_BLOCK_SIZE, 3682 .ivsize = DES3_EDE_BLOCK_SIZE, 3702 .ivsize = DES3_EDE_BLOCK_SIZE, 3722 .ivsize = DES3_EDE_BLOCK_SIZE, 3742 .ivsize = DES3_EDE_BLOCK_SIZE, 3762 .ivsize = DES_BLOCK_SIZE, 3781 .ivsize = DES_BLOCK_SIZE, 3800 .ivsize = DES_BLOCK_SIZE, 3820 .ivsize = DES_BLOCK_SIZE, 3840 .ivsize = DES_BLOCK_SIZE, 3860 .ivsize = DES_BLOCK_SIZE, 3880 .ivsize = CTR_RFC3686_IV_SIZE, 3899 .ivsize = CTR_RFC3686_IV_SIZE, 3918 .ivsize = CTR_RFC3686_IV_SIZE, 3938 .ivsize = CTR_RFC3686_IV_SIZE, 3958 .ivsize = CTR_RFC3686_IV_SIZE, 3978 .ivsize = CTR_RFC3686_IV_SIZE, 3998 .ivsize = 8, 4015 .ivsize = 8, 4033 .ivsize = 12, 4052 .ivsize = AES_BLOCK_SIZE, 4069 .ivsize = DES3_EDE_BLOCK_SIZE, 4086 .ivsize = DES_BLOCK_SIZE, 4102 .ivsize = AES_BLOCK_SIZE, 4121 .ivsize = CTR_RFC3686_IV_SIZE, 2130 caam_unmap(struct device *dev, struct scatterlist *src, struct scatterlist *dst, int src_nents, bool src_chained, int dst_nents, bool dst_chained, dma_addr_t iv_dma, int ivsize, dma_addr_t sec4_sg_dma, int sec4_sg_bytes) caam_unmap() argument
|
/linux-4.1.27/drivers/crypto/qce/ |
H A D | ablkcipher.c | 76 rctx->ivsize = crypto_ablkcipher_ivsize(ablkcipher); qce_ablkcipher_async_req_handle() 266 unsigned int ivsize; member in struct:qce_ablkcipher_def 277 .ivsize = AES_BLOCK_SIZE, 286 .ivsize = AES_BLOCK_SIZE, 295 .ivsize = AES_BLOCK_SIZE, 304 .ivsize = AES_BLOCK_SIZE, 313 .ivsize = 0, 322 .ivsize = DES_BLOCK_SIZE, 331 .ivsize = 0, 340 .ivsize = DES3_EDE_BLOCK_SIZE, 364 alg->cra_ablkcipher.ivsize = def->ivsize; qce_ablkcipher_register_one()
|
H A D | cipher.h | 32 * @ivsize: IV size 47 unsigned int ivsize; member in struct:qce_cipher_reqctx
|
H A D | common.c | 180 static void qce_xts_swapiv(__be32 *dst, const u8 *src, unsigned int ivsize) qce_xts_swapiv() argument 185 if (ivsize > QCE_AES_IV_LENGTH) qce_xts_swapiv() 190 for (i = (QCE_AES_IV_LENGTH - ivsize), j = ivsize - 1; qce_xts_swapiv() 328 unsigned int ivsize = rctx->ivsize; qce_setup_regs_ablkcipher() local 364 qce_xts_swapiv(enciv, rctx->iv, ivsize); qce_setup_regs_ablkcipher() 366 qce_cpu_to_be32p_array(enciv, rctx->iv, ivsize); qce_setup_regs_ablkcipher()
|
/linux-4.1.27/drivers/crypto/ccp/ |
H A D | ccp-crypto-aes.c | 265 unsigned int ivsize; member in struct:ccp_aes_def 275 .ivsize = 0, 283 .ivsize = AES_BLOCK_SIZE, 291 .ivsize = AES_BLOCK_SIZE, 299 .ivsize = AES_BLOCK_SIZE, 307 .ivsize = AES_BLOCK_SIZE, 315 .ivsize = CTR_RFC3686_IV_SIZE, 342 alg->cra_ablkcipher.ivsize = def->ivsize; ccp_register_aes_alg()
|
H A D | ccp-crypto-aes-xts.c | 255 alg->cra_ablkcipher.ivsize = AES_BLOCK_SIZE; ccp_register_aes_xts_alg()
|
/linux-4.1.27/drivers/crypto/ |
H A D | ixp4xx_crypto.c | 870 unsigned ivsize = crypto_ablkcipher_ivsize(tfm); ablk_perform() local 900 BUG_ON(ivsize && !req->info); ablk_perform() 901 memcpy(crypt->iv, req->info, ivsize); ablk_perform() 994 unsigned ivsize = crypto_aead_ivsize(tfm); aead_perform() local 1032 crypt->auth_len = req->assoclen + ivsize + cryptlen; aead_perform() 1033 BUG_ON(ivsize && !req->iv); aead_perform() 1034 memcpy(crypt->iv, req->iv, ivsize); aead_perform() 1049 sg_set_buf(&req_ctx->ivlist, iv, ivsize); aead_perform() 1050 buf = chainup_buffers(dev, &req_ctx->ivlist, ivsize, buf, flags, aead_perform() 1175 unsigned ivsize = crypto_aead_ivsize(crypto_aead_reqtfm(req)); aead_encrypt() local 1176 return aead_perform(req, 1, req->assoclen + ivsize, aead_encrypt() 1182 unsigned ivsize = crypto_aead_ivsize(crypto_aead_reqtfm(req)); aead_decrypt() local 1183 return aead_perform(req, 0, req->assoclen + ivsize, aead_decrypt() 1191 unsigned len, ivsize = crypto_aead_ivsize(tfm); aead_givencrypt() local 1196 get_random_bytes(ctx->salt, ivsize); aead_givencrypt() 1199 memcpy(req->areq.iv, ctx->salt, ivsize); aead_givencrypt() 1200 len = ivsize; aead_givencrypt() 1201 if (ivsize > sizeof(u64)) { aead_givencrypt() 1202 memset(req->giv, 0, ivsize - sizeof(u64)); aead_givencrypt() 1206 memcpy(req->giv + ivsize - len, &seq, len); aead_givencrypt() 1208 req->areq.cryptlen +ivsize, req->giv); aead_givencrypt() 1219 .ivsize = DES_BLOCK_SIZE, 1246 .ivsize = DES3_EDE_BLOCK_SIZE, 1272 .ivsize = AES_BLOCK_SIZE, 1298 .ivsize = AES_BLOCK_SIZE, 1312 .ivsize = AES_BLOCK_SIZE, 1326 .ivsize = DES_BLOCK_SIZE, 1339 .ivsize = DES3_EDE_BLOCK_SIZE, 1352 .ivsize = DES_BLOCK_SIZE, 1365 .ivsize = DES3_EDE_BLOCK_SIZE, 1378 .ivsize = AES_BLOCK_SIZE, 1391 .ivsize = AES_BLOCK_SIZE,
|
H A D | picoxcell_crypto.c | 327 unsigned ivsize = crypto_aead_ivsize(crypto_aead_reqtfm(areq)); spacc_aead_make_ddts() local 364 iv_addr = dma_map_single(engine->dev, iv, ivsize, spacc_aead_make_ddts() 378 ddt_set(src_ddt++, iv_addr, ivsize); spacc_aead_make_ddts() 381 ddt_set(dst_ddt++, iv_addr, ivsize); spacc_aead_make_ddts() 411 unsigned ivsize = alg->alg.cra_aead.ivsize; spacc_aead_free_ddts() local 425 dma_unmap_single(engine->dev, req->giv_pa, ivsize, DMA_BIDIRECTIONAL); spacc_aead_free_ddts() 608 ctx->cipher_key_len, aead_req->iv, alg->cra_aead.ivsize, spacc_aead_submit() 689 unsigned ivsize = crypto_aead_ivsize(crypto_aead_reqtfm(req)); spacc_aead_setup() local 692 dev_req->giv_len = ivsize; spacc_aead_setup() 741 size_t ivsize = crypto_aead_ivsize(tfm); spacc_aead_givencrypt() local 746 memcpy(req->areq.iv, ctx->salt, ivsize); spacc_aead_givencrypt() 747 len = ivsize; spacc_aead_givencrypt() 748 if (ivsize > sizeof(u64)) { spacc_aead_givencrypt() 749 memset(req->giv, 0, ivsize - sizeof(u64)); spacc_aead_givencrypt() 753 memcpy(req->giv + ivsize - len, &seq, len); spacc_aead_givencrypt() 953 ctx->key_len, ablk_req->info, alg->cra_ablkcipher.ivsize, spacc_ablk_submit() 1322 .ivsize = AES_BLOCK_SIZE, 1375 .ivsize = DES_BLOCK_SIZE, 1428 .ivsize = DES3_EDE_BLOCK_SIZE, 1482 .ivsize = AES_BLOCK_SIZE, 1512 .ivsize = AES_BLOCK_SIZE, 1541 .ivsize = AES_BLOCK_SIZE, 1570 .ivsize = DES3_EDE_BLOCK_SIZE, 1600 .ivsize = DES3_EDE_BLOCK_SIZE, 1629 .ivsize = DES3_EDE_BLOCK_SIZE, 1661 .ivsize = 8,
|
H A D | talitos.c | 951 unsigned int ivsize = crypto_aead_ivsize(aead); ipsec_esp() local 960 desc->ptr[1].len = cpu_to_be16(areq->assoclen + ivsize); ipsec_esp() 978 tbl_ptr->len = cpu_to_be16(ivsize); ipsec_esp() 994 desc->ptr[2].len = cpu_to_be16(ivsize); ipsec_esp() 997 dma_sync_single_for_device(dev, edesc->iv_dma, ivsize, DMA_TO_DEVICE); ipsec_esp() 1079 map_single_talitos_ptr(dev, &desc->ptr[6], ivsize, ctx->iv, 0, ipsec_esp() 1121 unsigned int ivsize, talitos_edesc_alloc() 1138 if (ivsize) talitos_edesc_alloc() 1139 iv_dma = dma_map_single(dev, iv, ivsize, DMA_TO_DEVICE); talitos_edesc_alloc() 1195 dma_unmap_single(dev, iv_dma, ivsize, DMA_TO_DEVICE); talitos_edesc_alloc() 1222 unsigned int ivsize = crypto_aead_ivsize(authenc); aead_edesc_alloc() local 1226 ctx->authsize, ivsize, icv_stashing, aead_edesc_alloc() 1373 unsigned int ivsize = crypto_ablkcipher_ivsize(cipher); common_nonsnoop() local 1383 desc->ptr[1].len = cpu_to_be16(ivsize); common_nonsnoop() 1445 map_single_talitos_ptr(dev, &desc->ptr[5], ivsize, ctx->iv, 0, common_nonsnoop() 1466 unsigned int ivsize = crypto_ablkcipher_ivsize(cipher); ablkcipher_edesc_alloc() local 1469 areq->info, 0, areq->nbytes, 0, ivsize, 0, ablkcipher_edesc_alloc() 1927 .ivsize = AES_BLOCK_SIZE, 1946 .ivsize = DES3_EDE_BLOCK_SIZE, 1966 .ivsize = AES_BLOCK_SIZE, 1985 .ivsize = DES3_EDE_BLOCK_SIZE, 2005 .ivsize = AES_BLOCK_SIZE, 2024 .ivsize = DES3_EDE_BLOCK_SIZE, 2044 .ivsize = AES_BLOCK_SIZE, 2063 .ivsize = DES3_EDE_BLOCK_SIZE, 2083 .ivsize = AES_BLOCK_SIZE, 2102 .ivsize = DES3_EDE_BLOCK_SIZE, 2122 .ivsize = AES_BLOCK_SIZE, 2141 .ivsize = DES3_EDE_BLOCK_SIZE, 2165 .ivsize = AES_BLOCK_SIZE, 2182 .ivsize = DES3_EDE_BLOCK_SIZE, 1113 talitos_edesc_alloc(struct device *dev, struct scatterlist *assoc, struct scatterlist *src, struct scatterlist *dst, u8 *iv, unsigned int assoclen, unsigned int cryptlen, unsigned int authsize, unsigned int ivsize, int icv_stashing, u32 cryptoflags, bool encrypt) talitos_edesc_alloc() argument
|
H A D | atmel-tdes.c | 955 .ivsize = DES_BLOCK_SIZE, 976 .ivsize = DES_BLOCK_SIZE, 997 .ivsize = DES_BLOCK_SIZE, 1018 .ivsize = DES_BLOCK_SIZE, 1039 .ivsize = DES_BLOCK_SIZE, 1060 .ivsize = DES_BLOCK_SIZE, 1101 .ivsize = DES_BLOCK_SIZE, 1122 .ivsize = DES_BLOCK_SIZE, 1143 .ivsize = DES_BLOCK_SIZE, 1164 .ivsize = DES_BLOCK_SIZE, 1185 .ivsize = DES_BLOCK_SIZE, 1206 .ivsize = DES_BLOCK_SIZE,
|
H A D | atmel-aes.c | 995 .ivsize = AES_BLOCK_SIZE, 1016 .ivsize = AES_BLOCK_SIZE, 1037 .ivsize = AES_BLOCK_SIZE, 1058 .ivsize = AES_BLOCK_SIZE, 1079 .ivsize = AES_BLOCK_SIZE, 1100 .ivsize = AES_BLOCK_SIZE, 1121 .ivsize = AES_BLOCK_SIZE, 1144 .ivsize = AES_BLOCK_SIZE,
|
H A D | hifn_795x.c | 658 unsigned int ivsize; member in struct:hifn_request_context 1126 u8 *key, int keylen, u8 *iv, int ivsize, u16 mode) hifn_setup_crypto_command() 1155 if (ivsize) { hifn_setup_crypto_command() 1156 memcpy(buf_pos, iv, ivsize); hifn_setup_crypto_command() 1157 buf_pos += ivsize; hifn_setup_crypto_command() 1255 rctx->iv, rctx->ivsize, md); hifn_setup_cmd_desc() 1585 if (rctx->iv && !rctx->ivsize && rctx->mode != ACRYPTO_MODE_ECB) hifn_setup_session() 1637 dev->name, rctx->iv, rctx->ivsize, hifn_setup_session() 1663 rctx.ivsize = 0; hifn_test() 2102 unsigned ivsize; hifn_setup_crypto_req() local 2104 ivsize = crypto_ablkcipher_ivsize(crypto_ablkcipher_reqtfm(req)); hifn_setup_crypto_req() 2108 ivsize = HIFN_AES_IV_LENGTH; hifn_setup_crypto_req() 2110 ivsize = HIFN_DES_KEY_LENGTH; hifn_setup_crypto_req() 2112 ivsize = HIFN_3DES_KEY_LENGTH; hifn_setup_crypto_req() 2126 rctx->ivsize = ivsize; hifn_setup_crypto_req() 2362 .ivsize = HIFN_IV_LENGTH, 2407 .ivsize = HIFN_IV_LENGTH, 2442 .ivsize = HIFN_AES_IV_LENGTH, 1124 hifn_setup_crypto_command(struct hifn_device *dev, u8 *buf, unsigned dlen, unsigned slen, u8 *key, int keylen, u8 *iv, int ivsize, u16 mode) hifn_setup_crypto_command() argument
|
H A D | n2_core.c | 1157 .ivsize = DES_BLOCK_SIZE, 1199 .ivsize = DES_BLOCK_SIZE, 1240 .ivsize = AES_BLOCK_SIZE, 1254 .ivsize = AES_BLOCK_SIZE,
|
H A D | geode-aes.c | 411 .ivsize = AES_BLOCK_SIZE,
|
H A D | padlock-aes.c | 496 .ivsize = AES_BLOCK_SIZE,
|
H A D | s5p-sss.c | 633 .ivsize = AES_BLOCK_SIZE,
|
H A D | omap-aes.c | 872 .ivsize = AES_BLOCK_SIZE, 899 .ivsize = AES_BLOCK_SIZE,
|
H A D | omap-des.c | 819 .ivsize = DES_BLOCK_SIZE, 864 .ivsize = DES_BLOCK_SIZE,
|
H A D | mxs-dcp.c | 824 .ivsize = AES_BLOCK_SIZE,
|
H A D | mv_cesa.c | 963 .ivsize = AES_BLOCK_SIZE,
|
H A D | sahara.c | 1309 .ivsize = AES_BLOCK_SIZE,
|
/linux-4.1.27/net/ceph/ |
H A D | crypto.c | 173 int ivsize; ceph_aes_encrypt() local 193 ivsize = crypto_blkcipher_ivsize(tfm); ceph_aes_encrypt() 194 memcpy(iv, aes_iv, ivsize); ceph_aes_encrypt() 233 int ivsize; ceph_aes_encrypt2() local 254 ivsize = crypto_blkcipher_ivsize(tfm); ceph_aes_encrypt2() 255 memcpy(iv, aes_iv, ivsize); ceph_aes_encrypt2() 295 int ivsize; ceph_aes_decrypt() local 311 ivsize = crypto_blkcipher_ivsize(tfm); ceph_aes_decrypt() 312 memcpy(iv, aes_iv, ivsize); ceph_aes_decrypt() 360 int ivsize; ceph_aes_decrypt2() local 377 ivsize = crypto_blkcipher_ivsize(tfm); ceph_aes_decrypt2() 378 memcpy(iv, aes_iv, ivsize); ceph_aes_decrypt2()
|
/linux-4.1.27/include/linux/ |
H A D | cryptouser.h | 84 unsigned int ivsize; member in struct:crypto_report_blkcipher 92 unsigned int ivsize; member in struct:crypto_report_aead
|
H A D | crypto.h | 275 * @ivsize: IV size applicable for transformation. The consumer must provide an 278 * All fields except @givencrypt , @givdecrypt , @geniv and @ivsize are 293 unsigned int ivsize; member in struct:ablkcipher_alg 317 * @ivsize: see struct ablkcipher_alg 319 * All fields except @givencrypt , @givdecrypt , @geniv and @ivsize are 333 unsigned int ivsize; member in struct:aead_alg 345 * @ivsize: see struct ablkcipher_alg 347 * All fields except @geniv and @ivsize are mandatory and must be filled. 363 unsigned int ivsize; member in struct:blkcipher_alg 601 unsigned int ivsize; member in struct:ablkcipher_tfm 615 unsigned int ivsize; member in struct:aead_tfm 940 return crypto_ablkcipher_crt(tfm)->ivsize; crypto_ablkcipher_ivsize() 1277 return crypto_aead_crt(tfm)->ivsize; crypto_aead_ivsize() 1725 return crypto_blkcipher_alg(tfm)->ivsize; crypto_blkcipher_ivsize()
|
/linux-4.1.27/arch/x86/crypto/ |
H A D | camellia_aesni_avx_glue.c | 389 .ivsize = CAMELLIA_BLOCK_SIZE, 413 .ivsize = CAMELLIA_BLOCK_SIZE, 434 .ivsize = CAMELLIA_BLOCK_SIZE, 477 .ivsize = CAMELLIA_BLOCK_SIZE, 499 .ivsize = CAMELLIA_BLOCK_SIZE, 524 .ivsize = CAMELLIA_BLOCK_SIZE, 546 .ivsize = CAMELLIA_BLOCK_SIZE,
|
H A D | serpent_avx2_glue.c | 366 .ivsize = SERPENT_BLOCK_SIZE, 391 .ivsize = SERPENT_BLOCK_SIZE, 413 .ivsize = SERPENT_BLOCK_SIZE, 458 .ivsize = SERPENT_BLOCK_SIZE, 481 .ivsize = SERPENT_BLOCK_SIZE, 507 .ivsize = SERPENT_BLOCK_SIZE, 530 .ivsize = SERPENT_BLOCK_SIZE,
|
H A D | fpu.c | 126 inst->alg.cra_blkcipher.ivsize = alg->cra_blkcipher.ivsize; crypto_fpu_alloc()
|
H A D | salsa20_glue.c | 35 u32 keysize, u32 ivsize); 102 .ivsize = SALSA20_IV_SIZE,
|
H A D | camellia_aesni_avx2_glue.c | 397 .ivsize = CAMELLIA_BLOCK_SIZE, 421 .ivsize = CAMELLIA_BLOCK_SIZE, 442 .ivsize = CAMELLIA_BLOCK_SIZE, 485 .ivsize = CAMELLIA_BLOCK_SIZE, 507 .ivsize = CAMELLIA_BLOCK_SIZE, 532 .ivsize = CAMELLIA_BLOCK_SIZE, 554 .ivsize = CAMELLIA_BLOCK_SIZE,
|
H A D | cast6_avx_glue.c | 426 .ivsize = CAST6_BLOCK_SIZE, 450 .ivsize = CAST6_BLOCK_SIZE, 471 .ivsize = CAST6_BLOCK_SIZE, 514 .ivsize = CAST6_BLOCK_SIZE, 536 .ivsize = CAST6_BLOCK_SIZE, 561 .ivsize = CAST6_BLOCK_SIZE, 583 .ivsize = CAST6_BLOCK_SIZE,
|
H A D | serpent_avx_glue.c | 432 .ivsize = SERPENT_BLOCK_SIZE, 456 .ivsize = SERPENT_BLOCK_SIZE, 477 .ivsize = SERPENT_BLOCK_SIZE, 520 .ivsize = SERPENT_BLOCK_SIZE, 542 .ivsize = SERPENT_BLOCK_SIZE, 567 .ivsize = SERPENT_BLOCK_SIZE, 589 .ivsize = SERPENT_BLOCK_SIZE,
|
H A D | serpent_sse2_glue.c | 441 .ivsize = SERPENT_BLOCK_SIZE, 465 .ivsize = SERPENT_BLOCK_SIZE, 486 .ivsize = SERPENT_BLOCK_SIZE, 529 .ivsize = SERPENT_BLOCK_SIZE, 551 .ivsize = SERPENT_BLOCK_SIZE, 576 .ivsize = SERPENT_BLOCK_SIZE, 598 .ivsize = SERPENT_BLOCK_SIZE,
|
H A D | twofish_avx_glue.c | 394 .ivsize = TF_BLOCK_SIZE, 418 .ivsize = TF_BLOCK_SIZE, 439 .ivsize = TF_BLOCK_SIZE, 482 .ivsize = TF_BLOCK_SIZE, 504 .ivsize = TF_BLOCK_SIZE, 529 .ivsize = TF_BLOCK_SIZE, 551 .ivsize = TF_BLOCK_SIZE,
|
H A D | twofish_glue_3way.c | 368 .ivsize = TF_BLOCK_SIZE, 388 .ivsize = TF_BLOCK_SIZE, 409 .ivsize = TF_BLOCK_SIZE, 429 .ivsize = TF_BLOCK_SIZE,
|
H A D | cast5_avx_glue.c | 395 .ivsize = CAST5_BLOCK_SIZE, 438 .ivsize = CAST5_BLOCK_SIZE, 460 .ivsize = CAST5_BLOCK_SIZE,
|
H A D | aesni-intel_glue.c | 1361 .ivsize = AES_BLOCK_SIZE, 1384 .ivsize = AES_BLOCK_SIZE, 1406 .ivsize = AES_BLOCK_SIZE, 1430 .ivsize = 8, 1454 .ivsize = 8, 1476 .ivsize = AES_BLOCK_SIZE, 1499 .ivsize = AES_BLOCK_SIZE, 1520 .ivsize = AES_BLOCK_SIZE, 1542 .ivsize = AES_BLOCK_SIZE, 1564 .ivsize = AES_BLOCK_SIZE,
|
H A D | blowfish_glue.c | 409 .ivsize = BF_BLOCK_SIZE, 429 .ivsize = BF_BLOCK_SIZE,
|
H A D | des3_ede_glue.c | 436 .ivsize = DES3_EDE_BLOCK_SIZE, 456 .ivsize = DES3_EDE_BLOCK_SIZE,
|
H A D | camellia_glue.c | 1612 .ivsize = CAMELLIA_BLOCK_SIZE, 1632 .ivsize = CAMELLIA_BLOCK_SIZE, 1655 .ivsize = CAMELLIA_BLOCK_SIZE, 1675 .ivsize = CAMELLIA_BLOCK_SIZE,
|
/linux-4.1.27/arch/arm/crypto/ |
H A D | aesbs-glue.c | 314 .ivsize = AES_BLOCK_SIZE, 333 .ivsize = AES_BLOCK_SIZE, 352 .ivsize = AES_BLOCK_SIZE, 372 .ivsize = AES_BLOCK_SIZE, 392 .ivsize = AES_BLOCK_SIZE, 412 .ivsize = AES_BLOCK_SIZE,
|
H A D | aes-ce-glue.c | 367 .ivsize = AES_BLOCK_SIZE, 386 .ivsize = AES_BLOCK_SIZE, 405 .ivsize = AES_BLOCK_SIZE, 424 .ivsize = AES_BLOCK_SIZE, 444 .ivsize = AES_BLOCK_SIZE, 464 .ivsize = AES_BLOCK_SIZE, 484 .ivsize = AES_BLOCK_SIZE, 504 .ivsize = AES_BLOCK_SIZE,
|
/linux-4.1.27/arch/arm64/crypto/ |
H A D | aes-glue.c | 297 .ivsize = AES_BLOCK_SIZE, 316 .ivsize = AES_BLOCK_SIZE, 335 .ivsize = AES_BLOCK_SIZE, 354 .ivsize = AES_BLOCK_SIZE, 374 .ivsize = AES_BLOCK_SIZE, 394 .ivsize = AES_BLOCK_SIZE, 414 .ivsize = AES_BLOCK_SIZE, 434 .ivsize = AES_BLOCK_SIZE,
|
H A D | aes-ce-ccm-glue.c | 272 .ivsize = AES_BLOCK_SIZE,
|
/linux-4.1.27/security/keys/encrypted-keys/ |
H A D | encrypted.c | 47 static unsigned int ivsize; variable 96 ivsize = crypto_blkcipher_ivsize(tfm); aes_get_sizes() 406 unsigned int ivsize) init_blkcipher_desc() 424 crypto_blkcipher_set_iv(desc->tfm, iv, ivsize); init_blkcipher_desc() 480 epayload->iv, ivsize); derived_key_encrypt() 575 epayload->iv, ivsize); derived_key_decrypt() 633 + strlen(datalen) + 1 + ivsize + 1 + encrypted_datalen; encrypted_key_alloc() 665 asciilen = (ivsize + 1 + encrypted_datalen + HASH_SIZE) * 2; encrypted_key_decrypt() 669 hex_encoded_data = hex_encoded_iv + (2 * ivsize) + 2; encrypted_key_decrypt() 670 ret = hex2bin(epayload->iv, hex_encoded_iv, ivsize); encrypted_key_decrypt() 718 epayload->encrypted_data = epayload->iv + ivsize + 1; __ekey_init() 759 get_random_bytes(epayload->iv, ivsize); encrypted_init() 877 memcpy(new_epayload->iv, epayload->iv, ivsize); encrypted_update() 911 asciiblob_len = epayload->datablob_len + ivsize + 1 encrypted_read() 404 init_blkcipher_desc(struct blkcipher_desc *desc, const u8 *key, unsigned int key_len, const u8 *iv, unsigned int ivsize) init_blkcipher_desc() argument
|
/linux-4.1.27/drivers/crypto/nx/ |
H A D | nx-aes-ctr.c | 161 .ivsize = AES_BLOCK_SIZE, 182 .ivsize = CTR_RFC3686_IV_SIZE,
|
H A D | nx-aes-cbc.c | 145 .ivsize = AES_BLOCK_SIZE,
|
H A D | nx-aes-ccm.c | 576 .ivsize = AES_BLOCK_SIZE, 598 .ivsize = 8,
|
H A D | nx-aes-gcm.c | 499 .ivsize = AES_BLOCK_SIZE, 520 .ivsize = 8,
|
/linux-4.1.27/arch/s390/crypto/ |
H A D | des_s390.c | 218 .ivsize = DES_BLOCK_SIZE, 364 .ivsize = DES_BLOCK_SIZE, 487 .ivsize = DES_BLOCK_SIZE, 530 .ivsize = DES_BLOCK_SIZE,
|
H A D | aes_s390.c | 523 .ivsize = AES_BLOCK_SIZE, 730 .ivsize = AES_BLOCK_SIZE, 881 .ivsize = AES_BLOCK_SIZE,
|
/linux-4.1.27/arch/powerpc/crypto/ |
H A D | aes-spe-glue.c | 423 .ivsize = AES_BLOCK_SIZE, 443 .ivsize = AES_BLOCK_SIZE, 463 .ivsize = AES_BLOCK_SIZE, 483 .ivsize = AES_BLOCK_SIZE,
|
/linux-4.1.27/drivers/crypto/vmx/ |
H A D | aes_cbc.c | 179 .ivsize = 0,
|
H A D | aes_ctr.c | 163 .ivsize = 0,
|
/linux-4.1.27/drivers/usb/wusbcore/ |
H A D | crypto.c | 211 size_t ivsize, dst_size; wusb_ccm_mac() local 236 ivsize = crypto_blkcipher_ivsize(tfm_cbc); wusb_ccm_mac() 237 memset(iv, 0, ivsize); wusb_ccm_mac()
|
/linux-4.1.27/arch/sparc/crypto/ |
H A D | aes_glue.c | 436 .ivsize = AES_BLOCK_SIZE, 456 .ivsize = AES_BLOCK_SIZE,
|
H A D | des_glue.c | 432 .ivsize = DES_BLOCK_SIZE, 489 .ivsize = DES3_EDE_BLOCK_SIZE,
|
H A D | camellia_glue.c | 277 .ivsize = CAMELLIA_BLOCK_SIZE,
|
/linux-4.1.27/drivers/crypto/qat/qat_common/ |
H A D | qat_algs.c | 1201 .ivsize = AES_BLOCK_SIZE, 1223 .ivsize = AES_BLOCK_SIZE, 1245 .ivsize = AES_BLOCK_SIZE, 1268 .ivsize = AES_BLOCK_SIZE,
|
/linux-4.1.27/include/crypto/ |
H A D | algapi.h | 103 unsigned int ivsize; member in struct:blkcipher_walk
|
/linux-4.1.27/drivers/crypto/ux500/cryp/ |
H A D | cryp_core.c | 1179 .ivsize = AES_BLOCK_SIZE, 1205 .ivsize = AES_BLOCK_SIZE, 1357 .ivsize = DES3_EDE_BLOCK_SIZE,
|
/linux-4.1.27/drivers/crypto/amcc/ |
H A D | crypto4xx_core.c | 1146 .ivsize = AES_IV_SIZE,
|