429 lines
12 KiB
C
429 lines
12 KiB
C
/* SPDX-License-Identifier: GPL-2.0 */
|
|
/* Copyright (c) 2016-2017 HiSilicon Limited. */
|
|
|
|
#ifndef _SEC_DRV_H_
|
|
#define _SEC_DRV_H_
|
|
|
|
#include <crypto/algapi.h>
|
|
#include <linux/kfifo.h>
|
|
|
|
#define SEC_MAX_SGE_NUM 64
|
|
#define SEC_HW_RING_NUM 3
|
|
|
|
#define SEC_CMD_RING 0
|
|
#define SEC_OUTORDER_RING 1
|
|
#define SEC_DBG_RING 2
|
|
|
|
/* A reasonable length to balance memory use against flexibility */
|
|
#define SEC_QUEUE_LEN 512
|
|
|
|
#define SEC_MAX_SGE_NUM 64
|
|
|
|
struct sec_bd_info {
|
|
#define SEC_BD_W0_T_LEN_M GENMASK(4, 0)
|
|
#define SEC_BD_W0_T_LEN_S 0
|
|
|
|
#define SEC_BD_W0_C_WIDTH_M GENMASK(6, 5)
|
|
#define SEC_BD_W0_C_WIDTH_S 5
|
|
#define SEC_C_WIDTH_AES_128BIT 0
|
|
#define SEC_C_WIDTH_AES_8BIT 1
|
|
#define SEC_C_WIDTH_AES_1BIT 2
|
|
#define SEC_C_WIDTH_DES_64BIT 0
|
|
#define SEC_C_WIDTH_DES_8BIT 1
|
|
#define SEC_C_WIDTH_DES_1BIT 2
|
|
|
|
#define SEC_BD_W0_C_MODE_M GENMASK(9, 7)
|
|
#define SEC_BD_W0_C_MODE_S 7
|
|
#define SEC_C_MODE_ECB 0
|
|
#define SEC_C_MODE_CBC 1
|
|
#define SEC_C_MODE_CTR 4
|
|
#define SEC_C_MODE_CCM 5
|
|
#define SEC_C_MODE_GCM 6
|
|
#define SEC_C_MODE_XTS 7
|
|
|
|
#define SEC_BD_W0_SEQ BIT(10)
|
|
#define SEC_BD_W0_DE BIT(11)
|
|
#define SEC_BD_W0_DAT_SKIP_M GENMASK(13, 12)
|
|
#define SEC_BD_W0_DAT_SKIP_S 12
|
|
#define SEC_BD_W0_C_GRAN_SIZE_19_16_M GENMASK(17, 14)
|
|
#define SEC_BD_W0_C_GRAN_SIZE_19_16_S 14
|
|
|
|
#define SEC_BD_W0_CIPHER_M GENMASK(19, 18)
|
|
#define SEC_BD_W0_CIPHER_S 18
|
|
#define SEC_CIPHER_NULL 0
|
|
#define SEC_CIPHER_ENCRYPT 1
|
|
#define SEC_CIPHER_DECRYPT 2
|
|
|
|
#define SEC_BD_W0_AUTH_M GENMASK(21, 20)
|
|
#define SEC_BD_W0_AUTH_S 20
|
|
#define SEC_AUTH_NULL 0
|
|
#define SEC_AUTH_MAC 1
|
|
#define SEC_AUTH_VERIF 2
|
|
|
|
#define SEC_BD_W0_AI_GEN BIT(22)
|
|
#define SEC_BD_W0_CI_GEN BIT(23)
|
|
#define SEC_BD_W0_NO_HPAD BIT(24)
|
|
#define SEC_BD_W0_HM_M GENMASK(26, 25)
|
|
#define SEC_BD_W0_HM_S 25
|
|
#define SEC_BD_W0_ICV_OR_SKEY_EN_M GENMASK(28, 27)
|
|
#define SEC_BD_W0_ICV_OR_SKEY_EN_S 27
|
|
|
|
/* Multi purpose field - gran size bits for send, flag for recv */
|
|
#define SEC_BD_W0_FLAG_M GENMASK(30, 29)
|
|
#define SEC_BD_W0_C_GRAN_SIZE_21_20_M GENMASK(30, 29)
|
|
#define SEC_BD_W0_FLAG_S 29
|
|
#define SEC_BD_W0_C_GRAN_SIZE_21_20_S 29
|
|
|
|
#define SEC_BD_W0_DONE BIT(31)
|
|
u32 w0;
|
|
|
|
#define SEC_BD_W1_AUTH_GRAN_SIZE_M GENMASK(21, 0)
|
|
#define SEC_BD_W1_AUTH_GRAN_SIZE_S 0
|
|
#define SEC_BD_W1_M_KEY_EN BIT(22)
|
|
#define SEC_BD_W1_BD_INVALID BIT(23)
|
|
#define SEC_BD_W1_ADDR_TYPE BIT(24)
|
|
|
|
#define SEC_BD_W1_A_ALG_M GENMASK(28, 25)
|
|
#define SEC_BD_W1_A_ALG_S 25
|
|
#define SEC_A_ALG_SHA1 0
|
|
#define SEC_A_ALG_SHA256 1
|
|
#define SEC_A_ALG_MD5 2
|
|
#define SEC_A_ALG_SHA224 3
|
|
#define SEC_A_ALG_HMAC_SHA1 8
|
|
#define SEC_A_ALG_HMAC_SHA224 10
|
|
#define SEC_A_ALG_HMAC_SHA256 11
|
|
#define SEC_A_ALG_HMAC_MD5 12
|
|
#define SEC_A_ALG_AES_XCBC 13
|
|
#define SEC_A_ALG_AES_CMAC 14
|
|
|
|
#define SEC_BD_W1_C_ALG_M GENMASK(31, 29)
|
|
#define SEC_BD_W1_C_ALG_S 29
|
|
#define SEC_C_ALG_DES 0
|
|
#define SEC_C_ALG_3DES 1
|
|
#define SEC_C_ALG_AES 2
|
|
|
|
u32 w1;
|
|
|
|
#define SEC_BD_W2_C_GRAN_SIZE_15_0_M GENMASK(15, 0)
|
|
#define SEC_BD_W2_C_GRAN_SIZE_15_0_S 0
|
|
#define SEC_BD_W2_GRAN_NUM_M GENMASK(31, 16)
|
|
#define SEC_BD_W2_GRAN_NUM_S 16
|
|
u32 w2;
|
|
|
|
#define SEC_BD_W3_AUTH_LEN_OFFSET_M GENMASK(9, 0)
|
|
#define SEC_BD_W3_AUTH_LEN_OFFSET_S 0
|
|
#define SEC_BD_W3_CIPHER_LEN_OFFSET_M GENMASK(19, 10)
|
|
#define SEC_BD_W3_CIPHER_LEN_OFFSET_S 10
|
|
#define SEC_BD_W3_MAC_LEN_M GENMASK(24, 20)
|
|
#define SEC_BD_W3_MAC_LEN_S 20
|
|
#define SEC_BD_W3_A_KEY_LEN_M GENMASK(29, 25)
|
|
#define SEC_BD_W3_A_KEY_LEN_S 25
|
|
#define SEC_BD_W3_C_KEY_LEN_M GENMASK(31, 30)
|
|
#define SEC_BD_W3_C_KEY_LEN_S 30
|
|
#define SEC_KEY_LEN_AES_128 0
|
|
#define SEC_KEY_LEN_AES_192 1
|
|
#define SEC_KEY_LEN_AES_256 2
|
|
#define SEC_KEY_LEN_DES 1
|
|
#define SEC_KEY_LEN_3DES_3_KEY 1
|
|
#define SEC_KEY_LEN_3DES_2_KEY 3
|
|
u32 w3;
|
|
|
|
/* W4,5 */
|
|
union {
|
|
u32 authkey_addr_lo;
|
|
u32 authiv_addr_lo;
|
|
};
|
|
union {
|
|
u32 authkey_addr_hi;
|
|
u32 authiv_addr_hi;
|
|
};
|
|
|
|
/* W6,7 */
|
|
u32 cipher_key_addr_lo;
|
|
u32 cipher_key_addr_hi;
|
|
|
|
/* W8,9 */
|
|
u32 cipher_iv_addr_lo;
|
|
u32 cipher_iv_addr_hi;
|
|
|
|
/* W10,11 */
|
|
u32 data_addr_lo;
|
|
u32 data_addr_hi;
|
|
|
|
/* W12,13 */
|
|
u32 mac_addr_lo;
|
|
u32 mac_addr_hi;
|
|
|
|
/* W14,15 */
|
|
u32 cipher_destin_addr_lo;
|
|
u32 cipher_destin_addr_hi;
|
|
};
|
|
|
|
enum sec_mem_region {
|
|
SEC_COMMON = 0,
|
|
SEC_SAA,
|
|
SEC_NUM_ADDR_REGIONS
|
|
};
|
|
|
|
#define SEC_NAME_SIZE 64
|
|
#define SEC_Q_NUM 16
|
|
|
|
|
|
/**
|
|
* struct sec_queue_ring_cmd - store information about a SEC HW cmd ring
|
|
* @used: Local counter used to cheaply establish if the ring is empty.
|
|
* @lock: Protect against simultaneous adjusting of the read and write pointers.
|
|
* @vaddr: Virtual address for the ram pages used for the ring.
|
|
* @paddr: Physical address of the dma mapped region of ram used for the ring.
|
|
* @callback: Callback function called on a ring element completing.
|
|
*/
|
|
struct sec_queue_ring_cmd {
|
|
atomic_t used;
|
|
struct mutex lock;
|
|
struct sec_bd_info *vaddr;
|
|
dma_addr_t paddr;
|
|
void (*callback)(struct sec_bd_info *resp, void *ctx);
|
|
};
|
|
|
|
struct sec_debug_bd_info;
|
|
struct sec_queue_ring_db {
|
|
struct sec_debug_bd_info *vaddr;
|
|
dma_addr_t paddr;
|
|
};
|
|
|
|
struct sec_out_bd_info;
|
|
struct sec_queue_ring_cq {
|
|
struct sec_out_bd_info *vaddr;
|
|
dma_addr_t paddr;
|
|
};
|
|
|
|
struct sec_dev_info;
|
|
|
|
enum sec_cipher_alg {
|
|
SEC_C_DES_ECB_64,
|
|
SEC_C_DES_CBC_64,
|
|
|
|
SEC_C_3DES_ECB_192_3KEY,
|
|
SEC_C_3DES_ECB_192_2KEY,
|
|
|
|
SEC_C_3DES_CBC_192_3KEY,
|
|
SEC_C_3DES_CBC_192_2KEY,
|
|
|
|
SEC_C_AES_ECB_128,
|
|
SEC_C_AES_ECB_192,
|
|
SEC_C_AES_ECB_256,
|
|
|
|
SEC_C_AES_CBC_128,
|
|
SEC_C_AES_CBC_192,
|
|
SEC_C_AES_CBC_256,
|
|
|
|
SEC_C_AES_CTR_128,
|
|
SEC_C_AES_CTR_192,
|
|
SEC_C_AES_CTR_256,
|
|
|
|
SEC_C_AES_XTS_128,
|
|
SEC_C_AES_XTS_256,
|
|
|
|
SEC_C_NULL,
|
|
};
|
|
|
|
/**
|
|
* struct sec_alg_tfm_ctx - hardware specific tranformation context
|
|
* @cipher_alg: Cipher algorithm enabled include encryption mode.
|
|
* @key: Key storage if required.
|
|
* @pkey: DMA address for the key storage.
|
|
* @req_template: Request template to save time on setup.
|
|
* @queue: The hardware queue associated with this tfm context.
|
|
* @lock: Protect key and pkey to ensure they are consistent
|
|
* @auth_buf: Current context buffer for auth operations.
|
|
* @backlog: The backlog queue used for cases where our buffers aren't
|
|
* large enough.
|
|
*/
|
|
struct sec_alg_tfm_ctx {
|
|
enum sec_cipher_alg cipher_alg;
|
|
u8 *key;
|
|
dma_addr_t pkey;
|
|
struct sec_bd_info req_template;
|
|
struct sec_queue *queue;
|
|
struct mutex lock;
|
|
u8 *auth_buf;
|
|
struct list_head backlog;
|
|
};
|
|
|
|
/**
|
|
* struct sec_request - data associate with a single crypto request
|
|
* @elements: List of subparts of this request (hardware size restriction)
|
|
* @num_elements: The number of subparts (used as an optimization)
|
|
* @lock: Protect elements of this structure against concurrent change.
|
|
* @tfm_ctx: hardware specific context.
|
|
* @len_in: length of in sgl from upper layers
|
|
* @len_out: length of out sgl from upper layers
|
|
* @dma_iv: initialization vector - phsyical address
|
|
* @err: store used to track errors across subelements of this request.
|
|
* @req_base: pointer to base element of associate crypto context.
|
|
* This is needed to allow shared handling skcipher, ahash etc.
|
|
* @cb: completion callback.
|
|
* @backlog_head: list head to allow backlog maintenance.
|
|
*
|
|
* The hardware is limited in the maximum size of data that it can
|
|
* process from a single BD. Typically this is fairly large (32MB)
|
|
* but still requires the complexity of splitting the incoming
|
|
* skreq up into a number of elements complete with appropriate
|
|
* iv chaining.
|
|
*/
|
|
struct sec_request {
|
|
struct list_head elements;
|
|
int num_elements;
|
|
struct mutex lock;
|
|
struct sec_alg_tfm_ctx *tfm_ctx;
|
|
int len_in;
|
|
int len_out;
|
|
dma_addr_t dma_iv;
|
|
int err;
|
|
struct crypto_async_request *req_base;
|
|
void (*cb)(struct sec_bd_info *resp, struct crypto_async_request *req);
|
|
struct list_head backlog_head;
|
|
};
|
|
|
|
/**
|
|
* struct sec_request_el - A subpart of a request.
|
|
* @head: allow us to attach this to the list in the sec_request
|
|
* @req: hardware block descriptor corresponding to this request subpart
|
|
* @in: hardware sgl for input - virtual address
|
|
* @dma_in: hardware sgl for input - physical address
|
|
* @sgl_in: scatterlist for this request subpart
|
|
* @out: hardware sgl for output - virtual address
|
|
* @dma_out: hardware sgl for output - physical address
|
|
* @sgl_out: scatterlist for this request subpart
|
|
* @sec_req: The request which this subpart forms a part of
|
|
* @el_length: Number of bytes in this subpart. Needed to locate
|
|
* last ivsize chunk for iv chaining.
|
|
*/
|
|
struct sec_request_el {
|
|
struct list_head head;
|
|
struct sec_bd_info req;
|
|
struct sec_hw_sgl *in;
|
|
dma_addr_t dma_in;
|
|
struct scatterlist *sgl_in;
|
|
struct sec_hw_sgl *out;
|
|
dma_addr_t dma_out;
|
|
struct scatterlist *sgl_out;
|
|
struct sec_request *sec_req;
|
|
size_t el_length;
|
|
};
|
|
|
|
/**
|
|
* struct sec_queue - All the information about a HW queue
|
|
* @dev_info: The parent SEC device to which this queue belongs.
|
|
* @task_irq: Completion interrupt for the queue.
|
|
* @name: Human readable queue description also used as irq name.
|
|
* @ring: The several HW rings associated with one queue.
|
|
* @regs: The iomapped device registers
|
|
* @queue_id: Index of the queue used for naming and resource selection.
|
|
* @in_use: Flag to say if the queue is in use.
|
|
* @expected: The next expected element to finish assuming we were in order.
|
|
* @uprocessed: A bitmap to track which OoO elements are done but not handled.
|
|
* @softqueue: A software queue used when chaining requirements prevent direct
|
|
* use of the hardware queues.
|
|
* @havesoftqueue: A flag to say we have a queues - as we may need one for the
|
|
* current mode.
|
|
* @queuelock: Protect the soft queue from concurrent changes to avoid some
|
|
* potential loss of data races.
|
|
* @shadow: Pointers back to the shadow copy of the hardware ring element
|
|
* need because we can't store any context reference in the bd element.
|
|
*/
|
|
struct sec_queue {
|
|
struct sec_dev_info *dev_info;
|
|
int task_irq;
|
|
char name[SEC_NAME_SIZE];
|
|
struct sec_queue_ring_cmd ring_cmd;
|
|
struct sec_queue_ring_cq ring_cq;
|
|
struct sec_queue_ring_db ring_db;
|
|
void __iomem *regs;
|
|
u32 queue_id;
|
|
bool in_use;
|
|
int expected;
|
|
|
|
DECLARE_BITMAP(unprocessed, SEC_QUEUE_LEN);
|
|
DECLARE_KFIFO_PTR(softqueue, typeof(struct sec_request_el *));
|
|
bool havesoftqueue;
|
|
spinlock_t queuelock;
|
|
void *shadow[SEC_QUEUE_LEN];
|
|
};
|
|
|
|
/**
|
|
* struct sec_hw_sge: Track each of the 64 element SEC HW SGL entries
|
|
* @buf: The IOV dma address for this entry.
|
|
* @len: Length of this IOV.
|
|
* @pad: Reserved space.
|
|
*/
|
|
struct sec_hw_sge {
|
|
dma_addr_t buf;
|
|
unsigned int len;
|
|
unsigned int pad;
|
|
};
|
|
|
|
/**
|
|
* struct sec_hw_sgl: One hardware SGL entry.
|
|
* @next_sgl: The next entry if we need to chain dma address. Null if last.
|
|
* @entry_sum_in_chain: The full count of SGEs - only matters for first SGL.
|
|
* @entry_sum_in_sgl: The number of SGEs in this SGL element.
|
|
* @flag: Unused in skciphers.
|
|
* @serial_num: Unsued in skciphers.
|
|
* @cpuid: Currently unused.
|
|
* @data_bytes_in_sgl: Count of bytes from all SGEs in this SGL.
|
|
* @next: Virtual address used to stash the next sgl - useful in completion.
|
|
* @reserved: A reserved field not currently used.
|
|
* @sge_entries: The (up to) 64 Scatter Gather Entries, representing IOVs.
|
|
* @node: Currently unused.
|
|
*/
|
|
struct sec_hw_sgl {
|
|
dma_addr_t next_sgl;
|
|
u16 entry_sum_in_chain;
|
|
u16 entry_sum_in_sgl;
|
|
u32 flag;
|
|
u64 serial_num;
|
|
u32 cpuid;
|
|
u32 data_bytes_in_sgl;
|
|
struct sec_hw_sgl *next;
|
|
u64 reserved;
|
|
struct sec_hw_sge sge_entries[SEC_MAX_SGE_NUM];
|
|
u8 node[16];
|
|
};
|
|
|
|
struct dma_pool;
|
|
|
|
/**
|
|
* struct sec_dev_info: The full SEC unit comprising queues and processors.
|
|
* @sec_id: Index used to track which SEC this is when more than one is present.
|
|
* @num_saas: The number of backed processors enabled.
|
|
* @regs: iomapped register regions shared by whole SEC unit.
|
|
* @dev_lock: Protects concurrent queue allocation / freeing for the SEC.
|
|
* @queues: The 16 queues that this SEC instance provides.
|
|
* @dev: Device pointer.
|
|
* @hw_sgl_pool: DMA pool used to mimise mapping for the scatter gather lists.
|
|
*/
|
|
struct sec_dev_info {
|
|
int sec_id;
|
|
int num_saas;
|
|
void __iomem *regs[SEC_NUM_ADDR_REGIONS];
|
|
struct mutex dev_lock;
|
|
int queues_in_use;
|
|
struct sec_queue queues[SEC_Q_NUM];
|
|
struct device *dev;
|
|
struct dma_pool *hw_sgl_pool;
|
|
};
|
|
|
|
int sec_queue_send(struct sec_queue *queue, struct sec_bd_info *msg, void *ctx);
|
|
bool sec_queue_can_enqueue(struct sec_queue *queue, int num);
|
|
int sec_queue_stop_release(struct sec_queue *queue);
|
|
struct sec_queue *sec_queue_alloc_start_safe(void);
|
|
bool sec_queue_empty(struct sec_queue *queue);
|
|
|
|
/* Algorithm specific elements from sec_algs.c */
|
|
void sec_alg_callback(struct sec_bd_info *resp, void *ctx);
|
|
int sec_algs_register(void);
|
|
void sec_algs_unregister(void);
|
|
|
|
#endif /* _SEC_DRV_H_ */
|