crypto: crypto4xx - pointer arithmetic overhaul
This patch improves the readability of various functions, by replacing various void* pointers declarations with their respective structs *. This makes it possible to go for the eye-friendly array-indexing methods. Signed-off-by: Christian Lamparter <chunkeey@googlemail.com> Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>
This commit is contained in:
parent
5d59ad6eea
commit
9e0a0b3a19
@ -134,7 +134,7 @@ static int crypto4xx_setkey_aes(struct crypto_ablkcipher *cipher,
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
/* Setup SA */
|
/* Setup SA */
|
||||||
sa = (struct dynamic_sa_ctl *) ctx->sa_in;
|
sa = ctx->sa_in;
|
||||||
|
|
||||||
set_dynamic_sa_command_0(sa, SA_NOT_SAVE_HASH, SA_NOT_SAVE_IV,
|
set_dynamic_sa_command_0(sa, SA_NOT_SAVE_HASH, SA_NOT_SAVE_IV,
|
||||||
SA_LOAD_HASH_FROM_SA, SA_LOAD_IV_FROM_STATE,
|
SA_LOAD_HASH_FROM_SA, SA_LOAD_IV_FROM_STATE,
|
||||||
@ -159,7 +159,7 @@ static int crypto4xx_setkey_aes(struct crypto_ablkcipher *cipher,
|
|||||||
ctx->offset_to_sr_ptr = get_dynamic_sa_offset_state_ptr_field(sa);
|
ctx->offset_to_sr_ptr = get_dynamic_sa_offset_state_ptr_field(sa);
|
||||||
|
|
||||||
memcpy(ctx->sa_out, ctx->sa_in, ctx->sa_len * 4);
|
memcpy(ctx->sa_out, ctx->sa_in, ctx->sa_len * 4);
|
||||||
sa = (struct dynamic_sa_ctl *) ctx->sa_out;
|
sa = ctx->sa_out;
|
||||||
sa->sa_command_0.bf.dir = DIR_OUTBOUND;
|
sa->sa_command_0.bf.dir = DIR_OUTBOUND;
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
@ -248,8 +248,7 @@ static int crypto4xx_hash_alg_init(struct crypto_tfm *tfm,
|
|||||||
struct crypto_alg *alg = tfm->__crt_alg;
|
struct crypto_alg *alg = tfm->__crt_alg;
|
||||||
struct crypto4xx_alg *my_alg = crypto_alg_to_crypto4xx_alg(alg);
|
struct crypto4xx_alg *my_alg = crypto_alg_to_crypto4xx_alg(alg);
|
||||||
struct crypto4xx_ctx *ctx = crypto_tfm_ctx(tfm);
|
struct crypto4xx_ctx *ctx = crypto_tfm_ctx(tfm);
|
||||||
struct dynamic_sa_ctl *sa;
|
struct dynamic_sa_hash160 *sa;
|
||||||
struct dynamic_sa_hash160 *sa_in;
|
|
||||||
int rc;
|
int rc;
|
||||||
|
|
||||||
ctx->dev = my_alg->dev;
|
ctx->dev = my_alg->dev;
|
||||||
@ -273,25 +272,24 @@ static int crypto4xx_hash_alg_init(struct crypto_tfm *tfm,
|
|||||||
|
|
||||||
crypto_ahash_set_reqsize(__crypto_ahash_cast(tfm),
|
crypto_ahash_set_reqsize(__crypto_ahash_cast(tfm),
|
||||||
sizeof(struct crypto4xx_ctx));
|
sizeof(struct crypto4xx_ctx));
|
||||||
sa = (struct dynamic_sa_ctl *) ctx->sa_in;
|
sa = (struct dynamic_sa_hash160 *)ctx->sa_in;
|
||||||
set_dynamic_sa_command_0(sa, SA_SAVE_HASH, SA_NOT_SAVE_IV,
|
set_dynamic_sa_command_0(&sa->ctrl, SA_SAVE_HASH, SA_NOT_SAVE_IV,
|
||||||
SA_NOT_LOAD_HASH, SA_LOAD_IV_FROM_SA,
|
SA_NOT_LOAD_HASH, SA_LOAD_IV_FROM_SA,
|
||||||
SA_NO_HEADER_PROC, ha, SA_CIPHER_ALG_NULL,
|
SA_NO_HEADER_PROC, ha, SA_CIPHER_ALG_NULL,
|
||||||
SA_PAD_TYPE_ZERO, SA_OP_GROUP_BASIC,
|
SA_PAD_TYPE_ZERO, SA_OP_GROUP_BASIC,
|
||||||
SA_OPCODE_HASH, DIR_INBOUND);
|
SA_OPCODE_HASH, DIR_INBOUND);
|
||||||
set_dynamic_sa_command_1(sa, 0, SA_HASH_MODE_HASH,
|
set_dynamic_sa_command_1(&sa->ctrl, 0, SA_HASH_MODE_HASH,
|
||||||
CRYPTO_FEEDBACK_MODE_NO_FB, SA_EXTENDED_SN_OFF,
|
CRYPTO_FEEDBACK_MODE_NO_FB, SA_EXTENDED_SN_OFF,
|
||||||
SA_SEQ_MASK_OFF, SA_MC_ENABLE,
|
SA_SEQ_MASK_OFF, SA_MC_ENABLE,
|
||||||
SA_NOT_COPY_PAD, SA_NOT_COPY_PAYLOAD,
|
SA_NOT_COPY_PAD, SA_NOT_COPY_PAYLOAD,
|
||||||
SA_NOT_COPY_HDR);
|
SA_NOT_COPY_HDR);
|
||||||
ctx->direction = DIR_INBOUND;
|
ctx->direction = DIR_INBOUND;
|
||||||
sa->sa_contents.w = SA_HASH160_CONTENTS;
|
|
||||||
sa_in = (struct dynamic_sa_hash160 *) ctx->sa_in;
|
|
||||||
/* Need to zero hash digest in SA */
|
/* Need to zero hash digest in SA */
|
||||||
memset(sa_in->inner_digest, 0, sizeof(sa_in->inner_digest));
|
memset(sa->inner_digest, 0, sizeof(sa->inner_digest));
|
||||||
memset(sa_in->outer_digest, 0, sizeof(sa_in->outer_digest));
|
memset(sa->outer_digest, 0, sizeof(sa->outer_digest));
|
||||||
sa_in->state_ptr = ctx->state_record_dma_addr;
|
sa->state_ptr = ctx->state_record_dma_addr;
|
||||||
ctx->offset_to_sr_ptr = get_dynamic_sa_offset_state_ptr_field(sa);
|
ctx->offset_to_sr_ptr =
|
||||||
|
get_dynamic_sa_offset_state_ptr_field(&sa->ctrl);
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
@ -302,7 +300,7 @@ int crypto4xx_hash_init(struct ahash_request *req)
|
|||||||
int ds;
|
int ds;
|
||||||
struct dynamic_sa_ctl *sa;
|
struct dynamic_sa_ctl *sa;
|
||||||
|
|
||||||
sa = (struct dynamic_sa_ctl *) ctx->sa_in;
|
sa = ctx->sa_in;
|
||||||
ds = crypto_ahash_digestsize(
|
ds = crypto_ahash_digestsize(
|
||||||
__crypto_ahash_cast(req->base.tfm));
|
__crypto_ahash_cast(req->base.tfm));
|
||||||
sa->sa_command_0.bf.digest_len = ds >> 2;
|
sa->sa_command_0.bf.digest_len = ds >> 2;
|
||||||
|
@ -211,7 +211,7 @@ static u32 crypto4xx_build_pdr(struct crypto4xx_device *dev)
|
|||||||
}
|
}
|
||||||
memset(dev->pdr, 0, sizeof(struct ce_pd) * PPC4XX_NUM_PD);
|
memset(dev->pdr, 0, sizeof(struct ce_pd) * PPC4XX_NUM_PD);
|
||||||
dev->shadow_sa_pool = dma_alloc_coherent(dev->core_dev->device,
|
dev->shadow_sa_pool = dma_alloc_coherent(dev->core_dev->device,
|
||||||
256 * PPC4XX_NUM_PD,
|
sizeof(union shadow_sa_buf) * PPC4XX_NUM_PD,
|
||||||
&dev->shadow_sa_pool_pa,
|
&dev->shadow_sa_pool_pa,
|
||||||
GFP_ATOMIC);
|
GFP_ATOMIC);
|
||||||
if (!dev->shadow_sa_pool)
|
if (!dev->shadow_sa_pool)
|
||||||
@ -223,16 +223,14 @@ static u32 crypto4xx_build_pdr(struct crypto4xx_device *dev)
|
|||||||
if (!dev->shadow_sr_pool)
|
if (!dev->shadow_sr_pool)
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
for (i = 0; i < PPC4XX_NUM_PD; i++) {
|
for (i = 0; i < PPC4XX_NUM_PD; i++) {
|
||||||
pd_uinfo = (struct pd_uinfo *) (dev->pdr_uinfo +
|
pd_uinfo = &dev->pdr_uinfo[i];
|
||||||
sizeof(struct pd_uinfo) * i);
|
|
||||||
|
|
||||||
/* alloc 256 bytes which is enough for any kind of dynamic sa */
|
/* alloc 256 bytes which is enough for any kind of dynamic sa */
|
||||||
pd_uinfo->sa_va = dev->shadow_sa_pool + 256 * i;
|
pd_uinfo->sa_va = &dev->shadow_sa_pool[i].sa;
|
||||||
pd_uinfo->sa_pa = dev->shadow_sa_pool_pa + 256 * i;
|
pd_uinfo->sa_pa = dev->shadow_sa_pool_pa + 256 * i;
|
||||||
|
|
||||||
/* alloc state record */
|
/* alloc state record */
|
||||||
pd_uinfo->sr_va = dev->shadow_sr_pool +
|
pd_uinfo->sr_va = &dev->shadow_sr_pool[i];
|
||||||
sizeof(struct sa_state_record) * i;
|
|
||||||
pd_uinfo->sr_pa = dev->shadow_sr_pool_pa +
|
pd_uinfo->sr_pa = dev->shadow_sr_pool_pa +
|
||||||
sizeof(struct sa_state_record) * i;
|
sizeof(struct sa_state_record) * i;
|
||||||
}
|
}
|
||||||
@ -248,7 +246,8 @@ static void crypto4xx_destroy_pdr(struct crypto4xx_device *dev)
|
|||||||
dev->pdr, dev->pdr_pa);
|
dev->pdr, dev->pdr_pa);
|
||||||
|
|
||||||
if (dev->shadow_sa_pool)
|
if (dev->shadow_sa_pool)
|
||||||
dma_free_coherent(dev->core_dev->device, 256 * PPC4XX_NUM_PD,
|
dma_free_coherent(dev->core_dev->device,
|
||||||
|
sizeof(union shadow_sa_buf) * PPC4XX_NUM_PD,
|
||||||
dev->shadow_sa_pool, dev->shadow_sa_pool_pa);
|
dev->shadow_sa_pool, dev->shadow_sa_pool_pa);
|
||||||
|
|
||||||
if (dev->shadow_sr_pool)
|
if (dev->shadow_sr_pool)
|
||||||
@ -277,11 +276,9 @@ static u32 crypto4xx_get_pd_from_pdr_nolock(struct crypto4xx_device *dev)
|
|||||||
|
|
||||||
static u32 crypto4xx_put_pd_to_pdr(struct crypto4xx_device *dev, u32 idx)
|
static u32 crypto4xx_put_pd_to_pdr(struct crypto4xx_device *dev, u32 idx)
|
||||||
{
|
{
|
||||||
struct pd_uinfo *pd_uinfo;
|
struct pd_uinfo *pd_uinfo = &dev->pdr_uinfo[idx];
|
||||||
unsigned long flags;
|
unsigned long flags;
|
||||||
|
|
||||||
pd_uinfo = (struct pd_uinfo *)(dev->pdr_uinfo +
|
|
||||||
sizeof(struct pd_uinfo) * idx);
|
|
||||||
spin_lock_irqsave(&dev->core_dev->lock, flags);
|
spin_lock_irqsave(&dev->core_dev->lock, flags);
|
||||||
if (dev->pdr_tail != PPC4XX_LAST_PD)
|
if (dev->pdr_tail != PPC4XX_LAST_PD)
|
||||||
dev->pdr_tail++;
|
dev->pdr_tail++;
|
||||||
@ -298,7 +295,7 @@ static struct ce_pd *crypto4xx_get_pdp(struct crypto4xx_device *dev,
|
|||||||
{
|
{
|
||||||
*pd_dma = dev->pdr_pa + sizeof(struct ce_pd) * idx;
|
*pd_dma = dev->pdr_pa + sizeof(struct ce_pd) * idx;
|
||||||
|
|
||||||
return dev->pdr + sizeof(struct ce_pd) * idx;
|
return &dev->pdr[idx];
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@ -376,7 +373,7 @@ static inline struct ce_gd *crypto4xx_get_gdp(struct crypto4xx_device *dev,
|
|||||||
{
|
{
|
||||||
*gd_dma = dev->gdr_pa + sizeof(struct ce_gd) * idx;
|
*gd_dma = dev->gdr_pa + sizeof(struct ce_gd) * idx;
|
||||||
|
|
||||||
return (struct ce_gd *) (dev->gdr + sizeof(struct ce_gd) * idx);
|
return &dev->gdr[idx];
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@ -387,7 +384,6 @@ static inline struct ce_gd *crypto4xx_get_gdp(struct crypto4xx_device *dev,
|
|||||||
static u32 crypto4xx_build_sdr(struct crypto4xx_device *dev)
|
static u32 crypto4xx_build_sdr(struct crypto4xx_device *dev)
|
||||||
{
|
{
|
||||||
int i;
|
int i;
|
||||||
struct ce_sd *sd_array;
|
|
||||||
|
|
||||||
/* alloc memory for scatter descriptor ring */
|
/* alloc memory for scatter descriptor ring */
|
||||||
dev->sdr = dma_alloc_coherent(dev->core_dev->device,
|
dev->sdr = dma_alloc_coherent(dev->core_dev->device,
|
||||||
@ -407,10 +403,8 @@ static u32 crypto4xx_build_sdr(struct crypto4xx_device *dev)
|
|||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
}
|
}
|
||||||
|
|
||||||
sd_array = dev->sdr;
|
|
||||||
|
|
||||||
for (i = 0; i < PPC4XX_NUM_SD; i++) {
|
for (i = 0; i < PPC4XX_NUM_SD; i++) {
|
||||||
sd_array[i].ptr = dev->scatter_buffer_pa +
|
dev->sdr[i].ptr = dev->scatter_buffer_pa +
|
||||||
PPC4XX_SD_BUFFER_SIZE * i;
|
PPC4XX_SD_BUFFER_SIZE * i;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -480,7 +474,7 @@ static inline struct ce_sd *crypto4xx_get_sdp(struct crypto4xx_device *dev,
|
|||||||
{
|
{
|
||||||
*sd_dma = dev->sdr_pa + sizeof(struct ce_sd) * idx;
|
*sd_dma = dev->sdr_pa + sizeof(struct ce_sd) * idx;
|
||||||
|
|
||||||
return (struct ce_sd *)(dev->sdr + sizeof(struct ce_sd) * idx);
|
return &dev->sdr[idx];
|
||||||
}
|
}
|
||||||
|
|
||||||
static void crypto4xx_copy_pkt_to_dst(struct crypto4xx_device *dev,
|
static void crypto4xx_copy_pkt_to_dst(struct crypto4xx_device *dev,
|
||||||
@ -529,11 +523,10 @@ static u32 crypto4xx_copy_digest_to_dst(struct pd_uinfo *pd_uinfo,
|
|||||||
struct crypto4xx_ctx *ctx)
|
struct crypto4xx_ctx *ctx)
|
||||||
{
|
{
|
||||||
struct dynamic_sa_ctl *sa = (struct dynamic_sa_ctl *) ctx->sa_in;
|
struct dynamic_sa_ctl *sa = (struct dynamic_sa_ctl *) ctx->sa_in;
|
||||||
struct sa_state_record *state_record =
|
|
||||||
(struct sa_state_record *) pd_uinfo->sr_va;
|
|
||||||
|
|
||||||
if (sa->sa_command_0.bf.hash_alg == SA_HASH_ALG_SHA1) {
|
if (sa->sa_command_0.bf.hash_alg == SA_HASH_ALG_SHA1) {
|
||||||
memcpy((void *) pd_uinfo->dest_va, state_record->save_digest,
|
memcpy((void *) pd_uinfo->dest_va,
|
||||||
|
pd_uinfo->sr_va->save_digest,
|
||||||
SA_HASH_ALG_SHA1_DIGEST_SIZE);
|
SA_HASH_ALG_SHA1_DIGEST_SIZE);
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -607,11 +600,9 @@ static u32 crypto4xx_ahash_done(struct crypto4xx_device *dev,
|
|||||||
|
|
||||||
static u32 crypto4xx_pd_done(struct crypto4xx_device *dev, u32 idx)
|
static u32 crypto4xx_pd_done(struct crypto4xx_device *dev, u32 idx)
|
||||||
{
|
{
|
||||||
struct ce_pd *pd;
|
struct ce_pd *pd = &dev->pdr[idx];
|
||||||
struct pd_uinfo *pd_uinfo;
|
struct pd_uinfo *pd_uinfo = &dev->pdr_uinfo[idx];
|
||||||
|
|
||||||
pd = dev->pdr + sizeof(struct ce_pd)*idx;
|
|
||||||
pd_uinfo = dev->pdr_uinfo + sizeof(struct pd_uinfo)*idx;
|
|
||||||
if (crypto_tfm_alg_type(pd_uinfo->async_req->tfm) ==
|
if (crypto_tfm_alg_type(pd_uinfo->async_req->tfm) ==
|
||||||
CRYPTO_ALG_TYPE_ABLKCIPHER)
|
CRYPTO_ALG_TYPE_ABLKCIPHER)
|
||||||
return crypto4xx_ablkcipher_done(dev, pd_uinfo, pd);
|
return crypto4xx_ablkcipher_done(dev, pd_uinfo, pd);
|
||||||
@ -712,7 +703,6 @@ u32 crypto4xx_build_pd(struct crypto_async_request *req,
|
|||||||
unsigned long flags;
|
unsigned long flags;
|
||||||
struct pd_uinfo *pd_uinfo = NULL;
|
struct pd_uinfo *pd_uinfo = NULL;
|
||||||
unsigned int nbytes = datalen, idx;
|
unsigned int nbytes = datalen, idx;
|
||||||
unsigned int ivlen = 0;
|
|
||||||
u32 gd_idx = 0;
|
u32 gd_idx = 0;
|
||||||
|
|
||||||
/* figure how many gd is needed */
|
/* figure how many gd is needed */
|
||||||
@ -771,17 +761,15 @@ u32 crypto4xx_build_pd(struct crypto_async_request *req,
|
|||||||
}
|
}
|
||||||
spin_unlock_irqrestore(&dev->core_dev->lock, flags);
|
spin_unlock_irqrestore(&dev->core_dev->lock, flags);
|
||||||
|
|
||||||
pd_uinfo = (struct pd_uinfo *)(dev->pdr_uinfo +
|
pd_uinfo = &dev->pdr_uinfo[pd_entry];
|
||||||
sizeof(struct pd_uinfo) * pd_entry);
|
|
||||||
pd = crypto4xx_get_pdp(dev, &pd_dma, pd_entry);
|
pd = crypto4xx_get_pdp(dev, &pd_dma, pd_entry);
|
||||||
pd_uinfo->async_req = req;
|
pd_uinfo->async_req = req;
|
||||||
pd_uinfo->num_gd = num_gd;
|
pd_uinfo->num_gd = num_gd;
|
||||||
pd_uinfo->num_sd = num_sd;
|
pd_uinfo->num_sd = num_sd;
|
||||||
|
|
||||||
if (iv_len || ctx->is_hash) {
|
if (iv_len || ctx->is_hash) {
|
||||||
ivlen = iv_len;
|
|
||||||
pd->sa = pd_uinfo->sa_pa;
|
pd->sa = pd_uinfo->sa_pa;
|
||||||
sa = (struct dynamic_sa_ctl *) pd_uinfo->sa_va;
|
sa = pd_uinfo->sa_va;
|
||||||
if (ctx->direction == DIR_INBOUND)
|
if (ctx->direction == DIR_INBOUND)
|
||||||
memcpy(sa, ctx->sa_in, ctx->sa_len * 4);
|
memcpy(sa, ctx->sa_in, ctx->sa_len * 4);
|
||||||
else
|
else
|
||||||
@ -791,14 +779,15 @@ u32 crypto4xx_build_pd(struct crypto_async_request *req,
|
|||||||
&pd_uinfo->sr_pa, 4);
|
&pd_uinfo->sr_pa, 4);
|
||||||
|
|
||||||
if (iv_len)
|
if (iv_len)
|
||||||
crypto4xx_memcpy_le(pd_uinfo->sr_va, iv, iv_len);
|
crypto4xx_memcpy_le(pd_uinfo->sr_va->save_iv,
|
||||||
|
iv, iv_len);
|
||||||
} else {
|
} else {
|
||||||
if (ctx->direction == DIR_INBOUND) {
|
if (ctx->direction == DIR_INBOUND) {
|
||||||
pd->sa = ctx->sa_in_dma_addr;
|
pd->sa = ctx->sa_in_dma_addr;
|
||||||
sa = (struct dynamic_sa_ctl *) ctx->sa_in;
|
sa = ctx->sa_in;
|
||||||
} else {
|
} else {
|
||||||
pd->sa = ctx->sa_out_dma_addr;
|
pd->sa = ctx->sa_out_dma_addr;
|
||||||
sa = (struct dynamic_sa_ctl *) ctx->sa_out;
|
sa = ctx->sa_out;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
pd->sa_len = ctx->sa_len;
|
pd->sa_len = ctx->sa_len;
|
||||||
@ -1006,9 +995,8 @@ static void crypto4xx_bh_tasklet_cb(unsigned long data)
|
|||||||
|
|
||||||
while (core_dev->dev->pdr_head != core_dev->dev->pdr_tail) {
|
while (core_dev->dev->pdr_head != core_dev->dev->pdr_tail) {
|
||||||
tail = core_dev->dev->pdr_tail;
|
tail = core_dev->dev->pdr_tail;
|
||||||
pd_uinfo = core_dev->dev->pdr_uinfo +
|
pd_uinfo = &core_dev->dev->pdr_uinfo[tail];
|
||||||
sizeof(struct pd_uinfo)*tail;
|
pd = &core_dev->dev->pdr[tail];
|
||||||
pd = core_dev->dev->pdr + sizeof(struct ce_pd) * tail;
|
|
||||||
if ((pd_uinfo->state == PD_ENTRY_INUSE) &&
|
if ((pd_uinfo->state == PD_ENTRY_INUSE) &&
|
||||||
pd->pd_ctl.bf.pe_done &&
|
pd->pd_ctl.bf.pe_done &&
|
||||||
!pd->pd_ctl.bf.host_ready) {
|
!pd->pd_ctl.bf.host_ready) {
|
||||||
|
@ -23,6 +23,8 @@
|
|||||||
#define __CRYPTO4XX_CORE_H__
|
#define __CRYPTO4XX_CORE_H__
|
||||||
|
|
||||||
#include <crypto/internal/hash.h>
|
#include <crypto/internal/hash.h>
|
||||||
|
#include "crypto4xx_reg_def.h"
|
||||||
|
#include "crypto4xx_sa.h"
|
||||||
|
|
||||||
#define MODULE_NAME "crypto4xx"
|
#define MODULE_NAME "crypto4xx"
|
||||||
|
|
||||||
@ -48,6 +50,13 @@
|
|||||||
|
|
||||||
struct crypto4xx_device;
|
struct crypto4xx_device;
|
||||||
|
|
||||||
|
union shadow_sa_buf {
|
||||||
|
struct dynamic_sa_ctl sa;
|
||||||
|
|
||||||
|
/* alloc 256 bytes which is enough for any kind of dynamic sa */
|
||||||
|
u8 buf[256];
|
||||||
|
} __packed;
|
||||||
|
|
||||||
struct pd_uinfo {
|
struct pd_uinfo {
|
||||||
struct crypto4xx_device *dev;
|
struct crypto4xx_device *dev;
|
||||||
u32 state;
|
u32 state;
|
||||||
@ -60,9 +69,9 @@ struct pd_uinfo {
|
|||||||
used by this packet */
|
used by this packet */
|
||||||
u32 num_sd; /* number of scatter discriptors
|
u32 num_sd; /* number of scatter discriptors
|
||||||
used by this packet */
|
used by this packet */
|
||||||
void *sa_va; /* shadow sa, when using cp from ctx->sa */
|
struct dynamic_sa_ctl *sa_va; /* shadow sa */
|
||||||
u32 sa_pa;
|
u32 sa_pa;
|
||||||
void *sr_va; /* state record for shadow sa */
|
struct sa_state_record *sr_va; /* state record for shadow sa */
|
||||||
u32 sr_pa;
|
u32 sr_pa;
|
||||||
struct scatterlist *dest_va;
|
struct scatterlist *dest_va;
|
||||||
struct crypto_async_request *async_req; /* base crypto request
|
struct crypto_async_request *async_req; /* base crypto request
|
||||||
@ -75,22 +84,18 @@ struct crypto4xx_device {
|
|||||||
void __iomem *ce_base;
|
void __iomem *ce_base;
|
||||||
void __iomem *trng_base;
|
void __iomem *trng_base;
|
||||||
|
|
||||||
void *pdr; /* base address of packet
|
struct ce_pd *pdr; /* base address of packet descriptor ring */
|
||||||
descriptor ring */
|
dma_addr_t pdr_pa; /* physical address of pdr_base_register */
|
||||||
dma_addr_t pdr_pa; /* physical address used to
|
struct ce_gd *gdr; /* gather descriptor ring */
|
||||||
program ce pdr_base_register */
|
dma_addr_t gdr_pa; /* physical address of gdr_base_register */
|
||||||
void *gdr; /* gather descriptor ring */
|
struct ce_sd *sdr; /* scatter descriptor ring */
|
||||||
dma_addr_t gdr_pa; /* physical address used to
|
dma_addr_t sdr_pa; /* physical address of sdr_base_register */
|
||||||
program ce gdr_base_register */
|
|
||||||
void *sdr; /* scatter descriptor ring */
|
|
||||||
dma_addr_t sdr_pa; /* physical address used to
|
|
||||||
program ce sdr_base_register */
|
|
||||||
void *scatter_buffer_va;
|
void *scatter_buffer_va;
|
||||||
dma_addr_t scatter_buffer_pa;
|
dma_addr_t scatter_buffer_pa;
|
||||||
|
|
||||||
void *shadow_sa_pool; /* pool of memory for sa in pd_uinfo */
|
union shadow_sa_buf *shadow_sa_pool;
|
||||||
dma_addr_t shadow_sa_pool_pa;
|
dma_addr_t shadow_sa_pool_pa;
|
||||||
void *shadow_sr_pool; /* pool of memory for sr in pd_uinfo */
|
struct sa_state_record *shadow_sr_pool;
|
||||||
dma_addr_t shadow_sr_pool_pa;
|
dma_addr_t shadow_sr_pool_pa;
|
||||||
u32 pdr_tail;
|
u32 pdr_tail;
|
||||||
u32 pdr_head;
|
u32 pdr_head;
|
||||||
@ -98,7 +103,7 @@ struct crypto4xx_device {
|
|||||||
u32 gdr_head;
|
u32 gdr_head;
|
||||||
u32 sdr_tail;
|
u32 sdr_tail;
|
||||||
u32 sdr_head;
|
u32 sdr_head;
|
||||||
void *pdr_uinfo;
|
struct pd_uinfo *pdr_uinfo;
|
||||||
struct list_head alg_list; /* List of algorithm supported
|
struct list_head alg_list; /* List of algorithm supported
|
||||||
by this device */
|
by this device */
|
||||||
};
|
};
|
||||||
@ -116,11 +121,11 @@ struct crypto4xx_core_device {
|
|||||||
|
|
||||||
struct crypto4xx_ctx {
|
struct crypto4xx_ctx {
|
||||||
struct crypto4xx_device *dev;
|
struct crypto4xx_device *dev;
|
||||||
void *sa_in;
|
struct dynamic_sa_ctl *sa_in;
|
||||||
dma_addr_t sa_in_dma_addr;
|
dma_addr_t sa_in_dma_addr;
|
||||||
void *sa_out;
|
struct dynamic_sa_ctl *sa_out;
|
||||||
dma_addr_t sa_out_dma_addr;
|
dma_addr_t sa_out_dma_addr;
|
||||||
void *state_record;
|
struct sa_state_record *state_record;
|
||||||
dma_addr_t state_record_dma_addr;
|
dma_addr_t state_record_dma_addr;
|
||||||
u32 sa_len;
|
u32 sa_len;
|
||||||
u32 offset_to_sr_ptr; /* offset to state ptr, in dynamic sa */
|
u32 offset_to_sr_ptr; /* offset to state ptr, in dynamic sa */
|
||||||
|
Loading…
Reference in New Issue
Block a user