123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840 |
- // SPDX-License-Identifier: GPL-2.0-only
- /*
- * Intel Keem Bay OCS HCU Crypto Driver.
- *
- * Copyright (C) 2018-2020 Intel Corporation
- */
- #include <linux/delay.h>
- #include <linux/device.h>
- #include <linux/iopoll.h>
- #include <linux/irq.h>
- #include <linux/module.h>
- #include <crypto/sha2.h>
- #include "ocs-hcu.h"
- /* Registers. */
- #define OCS_HCU_MODE 0x00
- #define OCS_HCU_CHAIN 0x04
- #define OCS_HCU_OPERATION 0x08
- #define OCS_HCU_KEY_0 0x0C
- #define OCS_HCU_ISR 0x50
- #define OCS_HCU_IER 0x54
- #define OCS_HCU_STATUS 0x58
- #define OCS_HCU_MSG_LEN_LO 0x60
- #define OCS_HCU_MSG_LEN_HI 0x64
- #define OCS_HCU_KEY_BYTE_ORDER_CFG 0x80
- #define OCS_HCU_DMA_SRC_ADDR 0x400
- #define OCS_HCU_DMA_SRC_SIZE 0x408
- #define OCS_HCU_DMA_DST_SIZE 0x40C
- #define OCS_HCU_DMA_DMA_MODE 0x410
- #define OCS_HCU_DMA_NEXT_SRC_DESCR 0x418
- #define OCS_HCU_DMA_MSI_ISR 0x480
- #define OCS_HCU_DMA_MSI_IER 0x484
- #define OCS_HCU_DMA_MSI_MASK 0x488
- /* Register bit definitions. */
- #define HCU_MODE_ALGO_SHIFT 16
- #define HCU_MODE_HMAC_SHIFT 22
- #define HCU_STATUS_BUSY BIT(0)
- #define HCU_BYTE_ORDER_SWAP BIT(0)
- #define HCU_IRQ_HASH_DONE BIT(2)
- #define HCU_IRQ_HASH_ERR_MASK (BIT(3) | BIT(1) | BIT(0))
- #define HCU_DMA_IRQ_SRC_DONE BIT(0)
- #define HCU_DMA_IRQ_SAI_ERR BIT(2)
- #define HCU_DMA_IRQ_BAD_COMP_ERR BIT(3)
- #define HCU_DMA_IRQ_INBUF_RD_ERR BIT(4)
- #define HCU_DMA_IRQ_INBUF_WD_ERR BIT(5)
- #define HCU_DMA_IRQ_OUTBUF_WR_ERR BIT(6)
- #define HCU_DMA_IRQ_OUTBUF_RD_ERR BIT(7)
- #define HCU_DMA_IRQ_CRD_ERR BIT(8)
- #define HCU_DMA_IRQ_ERR_MASK (HCU_DMA_IRQ_SAI_ERR | \
- HCU_DMA_IRQ_BAD_COMP_ERR | \
- HCU_DMA_IRQ_INBUF_RD_ERR | \
- HCU_DMA_IRQ_INBUF_WD_ERR | \
- HCU_DMA_IRQ_OUTBUF_WR_ERR | \
- HCU_DMA_IRQ_OUTBUF_RD_ERR | \
- HCU_DMA_IRQ_CRD_ERR)
- #define HCU_DMA_SNOOP_MASK (0x7 << 28)
- #define HCU_DMA_SRC_LL_EN BIT(25)
- #define HCU_DMA_EN BIT(31)
- #define OCS_HCU_ENDIANNESS_VALUE 0x2A
- #define HCU_DMA_MSI_UNMASK BIT(0)
- #define HCU_DMA_MSI_DISABLE 0
- #define HCU_IRQ_DISABLE 0
- #define OCS_HCU_START BIT(0)
- #define OCS_HCU_TERMINATE BIT(1)
- #define OCS_LL_DMA_FLAG_TERMINATE BIT(31)
- #define OCS_HCU_HW_KEY_LEN_U32 (OCS_HCU_HW_KEY_LEN / sizeof(u32))
- #define HCU_DATA_WRITE_ENDIANNESS_OFFSET 26
- #define OCS_HCU_NUM_CHAINS_SHA256_224_SM3 (SHA256_DIGEST_SIZE / sizeof(u32))
- #define OCS_HCU_NUM_CHAINS_SHA384_512 (SHA512_DIGEST_SIZE / sizeof(u32))
- /*
- * While polling on a busy HCU, wait maximum 200us between one check and the
- * other.
- */
- #define OCS_HCU_WAIT_BUSY_RETRY_DELAY_US 200
- /* Wait on a busy HCU for maximum 1 second. */
- #define OCS_HCU_WAIT_BUSY_TIMEOUT_US 1000000
- /**
- * struct ocs_hcu_dma_entry - An entry in an OCS DMA linked list.
- * @src_addr: Source address of the data.
- * @src_len: Length of data to be fetched.
- * @nxt_desc: Next descriptor to fetch.
- * @ll_flags: Flags (Freeze @ terminate) for the DMA engine.
- */
- struct ocs_hcu_dma_entry {
- u32 src_addr;
- u32 src_len;
- u32 nxt_desc;
- u32 ll_flags;
- };
- /**
- * struct ocs_hcu_dma_list - OCS-specific DMA linked list.
- * @head: The head of the list (points to the array backing the list).
- * @tail: The current tail of the list; NULL if the list is empty.
- * @dma_addr: The DMA address of @head (i.e., the DMA address of the backing
- * array).
- * @max_nents: Maximum number of entries in the list (i.e., number of elements
- * in the backing array).
- *
- * The OCS DMA list is an array-backed list of OCS DMA descriptors. The array
- * backing the list is allocated with dma_alloc_coherent() and pointed by
- * @head.
- */
- struct ocs_hcu_dma_list {
- struct ocs_hcu_dma_entry *head;
- struct ocs_hcu_dma_entry *tail;
- dma_addr_t dma_addr;
- size_t max_nents;
- };
- static inline u32 ocs_hcu_num_chains(enum ocs_hcu_algo algo)
- {
- switch (algo) {
- case OCS_HCU_ALGO_SHA224:
- case OCS_HCU_ALGO_SHA256:
- case OCS_HCU_ALGO_SM3:
- return OCS_HCU_NUM_CHAINS_SHA256_224_SM3;
- case OCS_HCU_ALGO_SHA384:
- case OCS_HCU_ALGO_SHA512:
- return OCS_HCU_NUM_CHAINS_SHA384_512;
- default:
- return 0;
- };
- }
- static inline u32 ocs_hcu_digest_size(enum ocs_hcu_algo algo)
- {
- switch (algo) {
- case OCS_HCU_ALGO_SHA224:
- return SHA224_DIGEST_SIZE;
- case OCS_HCU_ALGO_SHA256:
- case OCS_HCU_ALGO_SM3:
- /* SM3 shares the same block size. */
- return SHA256_DIGEST_SIZE;
- case OCS_HCU_ALGO_SHA384:
- return SHA384_DIGEST_SIZE;
- case OCS_HCU_ALGO_SHA512:
- return SHA512_DIGEST_SIZE;
- default:
- return 0;
- }
- }
- /**
- * ocs_hcu_wait_busy() - Wait for HCU OCS hardware to became usable.
- * @hcu_dev: OCS HCU device to wait for.
- *
- * Return: 0 if device free, -ETIMEOUT if device busy and internal timeout has
- * expired.
- */
- static int ocs_hcu_wait_busy(struct ocs_hcu_dev *hcu_dev)
- {
- long val;
- return readl_poll_timeout(hcu_dev->io_base + OCS_HCU_STATUS, val,
- !(val & HCU_STATUS_BUSY),
- OCS_HCU_WAIT_BUSY_RETRY_DELAY_US,
- OCS_HCU_WAIT_BUSY_TIMEOUT_US);
- }
- static void ocs_hcu_done_irq_en(struct ocs_hcu_dev *hcu_dev)
- {
- /* Clear any pending interrupts. */
- writel(0xFFFFFFFF, hcu_dev->io_base + OCS_HCU_ISR);
- hcu_dev->irq_err = false;
- /* Enable error and HCU done interrupts. */
- writel(HCU_IRQ_HASH_DONE | HCU_IRQ_HASH_ERR_MASK,
- hcu_dev->io_base + OCS_HCU_IER);
- }
- static void ocs_hcu_dma_irq_en(struct ocs_hcu_dev *hcu_dev)
- {
- /* Clear any pending interrupts. */
- writel(0xFFFFFFFF, hcu_dev->io_base + OCS_HCU_DMA_MSI_ISR);
- hcu_dev->irq_err = false;
- /* Only operating on DMA source completion and error interrupts. */
- writel(HCU_DMA_IRQ_ERR_MASK | HCU_DMA_IRQ_SRC_DONE,
- hcu_dev->io_base + OCS_HCU_DMA_MSI_IER);
- /* Unmask */
- writel(HCU_DMA_MSI_UNMASK, hcu_dev->io_base + OCS_HCU_DMA_MSI_MASK);
- }
- static void ocs_hcu_irq_dis(struct ocs_hcu_dev *hcu_dev)
- {
- writel(HCU_IRQ_DISABLE, hcu_dev->io_base + OCS_HCU_IER);
- writel(HCU_DMA_MSI_DISABLE, hcu_dev->io_base + OCS_HCU_DMA_MSI_IER);
- }
- static int ocs_hcu_wait_and_disable_irq(struct ocs_hcu_dev *hcu_dev)
- {
- int rc;
- rc = wait_for_completion_interruptible(&hcu_dev->irq_done);
- if (rc)
- goto exit;
- if (hcu_dev->irq_err) {
- /* Unset flag and return error. */
- hcu_dev->irq_err = false;
- rc = -EIO;
- goto exit;
- }
- exit:
- ocs_hcu_irq_dis(hcu_dev);
- return rc;
- }
- /**
- * ocs_hcu_get_intermediate_data() - Get intermediate data.
- * @hcu_dev: The target HCU device.
- * @data: Where to store the intermediate.
- * @algo: The algorithm being used.
- *
- * This function is used to save the current hashing process state in order to
- * continue it in the future.
- *
- * Note: once all data has been processed, the intermediate data actually
- * contains the hashing result. So this function is also used to retrieve the
- * final result of a hashing process.
- *
- * Return: 0 on success, negative error code otherwise.
- */
- static int ocs_hcu_get_intermediate_data(struct ocs_hcu_dev *hcu_dev,
- struct ocs_hcu_idata *data,
- enum ocs_hcu_algo algo)
- {
- const int n = ocs_hcu_num_chains(algo);
- u32 *chain;
- int rc;
- int i;
- /* Data not requested. */
- if (!data)
- return -EINVAL;
- chain = (u32 *)data->digest;
- /* Ensure that the OCS is no longer busy before reading the chains. */
- rc = ocs_hcu_wait_busy(hcu_dev);
- if (rc)
- return rc;
- /*
- * This loops is safe because data->digest is an array of
- * SHA512_DIGEST_SIZE bytes and the maximum value returned by
- * ocs_hcu_num_chains() is OCS_HCU_NUM_CHAINS_SHA384_512 which is equal
- * to SHA512_DIGEST_SIZE / sizeof(u32).
- */
- for (i = 0; i < n; i++)
- chain[i] = readl(hcu_dev->io_base + OCS_HCU_CHAIN);
- data->msg_len_lo = readl(hcu_dev->io_base + OCS_HCU_MSG_LEN_LO);
- data->msg_len_hi = readl(hcu_dev->io_base + OCS_HCU_MSG_LEN_HI);
- return 0;
- }
- /**
- * ocs_hcu_set_intermediate_data() - Set intermediate data.
- * @hcu_dev: The target HCU device.
- * @data: The intermediate data to be set.
- * @algo: The algorithm being used.
- *
- * This function is used to continue a previous hashing process.
- */
- static void ocs_hcu_set_intermediate_data(struct ocs_hcu_dev *hcu_dev,
- const struct ocs_hcu_idata *data,
- enum ocs_hcu_algo algo)
- {
- const int n = ocs_hcu_num_chains(algo);
- u32 *chain = (u32 *)data->digest;
- int i;
- /*
- * This loops is safe because data->digest is an array of
- * SHA512_DIGEST_SIZE bytes and the maximum value returned by
- * ocs_hcu_num_chains() is OCS_HCU_NUM_CHAINS_SHA384_512 which is equal
- * to SHA512_DIGEST_SIZE / sizeof(u32).
- */
- for (i = 0; i < n; i++)
- writel(chain[i], hcu_dev->io_base + OCS_HCU_CHAIN);
- writel(data->msg_len_lo, hcu_dev->io_base + OCS_HCU_MSG_LEN_LO);
- writel(data->msg_len_hi, hcu_dev->io_base + OCS_HCU_MSG_LEN_HI);
- }
- static int ocs_hcu_get_digest(struct ocs_hcu_dev *hcu_dev,
- enum ocs_hcu_algo algo, u8 *dgst, size_t dgst_len)
- {
- u32 *chain;
- int rc;
- int i;
- if (!dgst)
- return -EINVAL;
- /* Length of the output buffer must match the algo digest size. */
- if (dgst_len != ocs_hcu_digest_size(algo))
- return -EINVAL;
- /* Ensure that the OCS is no longer busy before reading the chains. */
- rc = ocs_hcu_wait_busy(hcu_dev);
- if (rc)
- return rc;
- chain = (u32 *)dgst;
- for (i = 0; i < dgst_len / sizeof(u32); i++)
- chain[i] = readl(hcu_dev->io_base + OCS_HCU_CHAIN);
- return 0;
- }
- /**
- * ocs_hcu_hw_cfg() - Configure the HCU hardware.
- * @hcu_dev: The HCU device to configure.
- * @algo: The algorithm to be used by the HCU device.
- * @use_hmac: Whether or not HW HMAC should be used.
- *
- * Return: 0 on success, negative error code otherwise.
- */
- static int ocs_hcu_hw_cfg(struct ocs_hcu_dev *hcu_dev, enum ocs_hcu_algo algo,
- bool use_hmac)
- {
- u32 cfg;
- int rc;
- if (algo != OCS_HCU_ALGO_SHA256 && algo != OCS_HCU_ALGO_SHA224 &&
- algo != OCS_HCU_ALGO_SHA384 && algo != OCS_HCU_ALGO_SHA512 &&
- algo != OCS_HCU_ALGO_SM3)
- return -EINVAL;
- rc = ocs_hcu_wait_busy(hcu_dev);
- if (rc)
- return rc;
- /* Ensure interrupts are disabled. */
- ocs_hcu_irq_dis(hcu_dev);
- /* Configure endianness, hashing algorithm and HW HMAC (if needed) */
- cfg = OCS_HCU_ENDIANNESS_VALUE << HCU_DATA_WRITE_ENDIANNESS_OFFSET;
- cfg |= algo << HCU_MODE_ALGO_SHIFT;
- if (use_hmac)
- cfg |= BIT(HCU_MODE_HMAC_SHIFT);
- writel(cfg, hcu_dev->io_base + OCS_HCU_MODE);
- return 0;
- }
- /**
- * ocs_hcu_clear_key() - Clear key stored in OCS HMAC KEY registers.
- * @hcu_dev: The OCS HCU device whose key registers should be cleared.
- */
- static void ocs_hcu_clear_key(struct ocs_hcu_dev *hcu_dev)
- {
- int reg_off;
- /* Clear OCS_HCU_KEY_[0..15] */
- for (reg_off = 0; reg_off < OCS_HCU_HW_KEY_LEN; reg_off += sizeof(u32))
- writel(0, hcu_dev->io_base + OCS_HCU_KEY_0 + reg_off);
- }
- /**
- * ocs_hcu_write_key() - Write key to OCS HMAC KEY registers.
- * @hcu_dev: The OCS HCU device the key should be written to.
- * @key: The key to be written.
- * @len: The size of the key to write. It must be OCS_HCU_HW_KEY_LEN.
- *
- * Return: 0 on success, negative error code otherwise.
- */
- static int ocs_hcu_write_key(struct ocs_hcu_dev *hcu_dev, const u8 *key, size_t len)
- {
- u32 key_u32[OCS_HCU_HW_KEY_LEN_U32];
- int i;
- if (len > OCS_HCU_HW_KEY_LEN)
- return -EINVAL;
- /* Copy key into temporary u32 array. */
- memcpy(key_u32, key, len);
- /*
- * Hardware requires all the bytes of the HW Key vector to be
- * written. So pad with zero until we reach OCS_HCU_HW_KEY_LEN.
- */
- memzero_explicit((u8 *)key_u32 + len, OCS_HCU_HW_KEY_LEN - len);
- /*
- * OCS hardware expects the MSB of the key to be written at the highest
- * address of the HCU Key vector; in other word, the key must be
- * written in reverse order.
- *
- * Therefore, we first enable byte swapping for the HCU key vector;
- * so that bytes of 32-bit word written to OCS_HCU_KEY_[0..15] will be
- * swapped:
- * 3 <---> 0, 2 <---> 1.
- */
- writel(HCU_BYTE_ORDER_SWAP,
- hcu_dev->io_base + OCS_HCU_KEY_BYTE_ORDER_CFG);
- /*
- * And then we write the 32-bit words composing the key starting from
- * the end of the key.
- */
- for (i = 0; i < OCS_HCU_HW_KEY_LEN_U32; i++)
- writel(key_u32[OCS_HCU_HW_KEY_LEN_U32 - 1 - i],
- hcu_dev->io_base + OCS_HCU_KEY_0 + (sizeof(u32) * i));
- memzero_explicit(key_u32, OCS_HCU_HW_KEY_LEN);
- return 0;
- }
- /**
- * ocs_hcu_ll_dma_start() - Start OCS HCU hashing via DMA
- * @hcu_dev: The OCS HCU device to use.
- * @dma_list: The OCS DMA list mapping the data to hash.
- * @finalize: Whether or not this is the last hashing operation and therefore
- * the final hash should be compute even if data is not
- * block-aligned.
- *
- * Return: 0 on success, negative error code otherwise.
- */
- static int ocs_hcu_ll_dma_start(struct ocs_hcu_dev *hcu_dev,
- const struct ocs_hcu_dma_list *dma_list,
- bool finalize)
- {
- u32 cfg = HCU_DMA_SNOOP_MASK | HCU_DMA_SRC_LL_EN | HCU_DMA_EN;
- int rc;
- if (!dma_list)
- return -EINVAL;
- /*
- * For final requests we use HCU_DONE IRQ to be notified when all input
- * data has been processed by the HCU; however, we cannot do so for
- * non-final requests, because we don't get a HCU_DONE IRQ when we
- * don't terminate the operation.
- *
- * Therefore, for non-final requests, we use the DMA IRQ, which
- * triggers when DMA has finishing feeding all the input data to the
- * HCU, but the HCU may still be processing it. This is fine, since we
- * will wait for the HCU processing to be completed when we try to read
- * intermediate results, in ocs_hcu_get_intermediate_data().
- */
- if (finalize)
- ocs_hcu_done_irq_en(hcu_dev);
- else
- ocs_hcu_dma_irq_en(hcu_dev);
- reinit_completion(&hcu_dev->irq_done);
- writel(dma_list->dma_addr, hcu_dev->io_base + OCS_HCU_DMA_NEXT_SRC_DESCR);
- writel(0, hcu_dev->io_base + OCS_HCU_DMA_SRC_SIZE);
- writel(0, hcu_dev->io_base + OCS_HCU_DMA_DST_SIZE);
- writel(OCS_HCU_START, hcu_dev->io_base + OCS_HCU_OPERATION);
- writel(cfg, hcu_dev->io_base + OCS_HCU_DMA_DMA_MODE);
- if (finalize)
- writel(OCS_HCU_TERMINATE, hcu_dev->io_base + OCS_HCU_OPERATION);
- rc = ocs_hcu_wait_and_disable_irq(hcu_dev);
- if (rc)
- return rc;
- return 0;
- }
- struct ocs_hcu_dma_list *ocs_hcu_dma_list_alloc(struct ocs_hcu_dev *hcu_dev,
- int max_nents)
- {
- struct ocs_hcu_dma_list *dma_list;
- dma_list = kmalloc(sizeof(*dma_list), GFP_KERNEL);
- if (!dma_list)
- return NULL;
- /* Total size of the DMA list to allocate. */
- dma_list->head = dma_alloc_coherent(hcu_dev->dev,
- sizeof(*dma_list->head) * max_nents,
- &dma_list->dma_addr, GFP_KERNEL);
- if (!dma_list->head) {
- kfree(dma_list);
- return NULL;
- }
- dma_list->max_nents = max_nents;
- dma_list->tail = NULL;
- return dma_list;
- }
- void ocs_hcu_dma_list_free(struct ocs_hcu_dev *hcu_dev,
- struct ocs_hcu_dma_list *dma_list)
- {
- if (!dma_list)
- return;
- dma_free_coherent(hcu_dev->dev,
- sizeof(*dma_list->head) * dma_list->max_nents,
- dma_list->head, dma_list->dma_addr);
- kfree(dma_list);
- }
- /* Add a new DMA entry at the end of the OCS DMA list. */
- int ocs_hcu_dma_list_add_tail(struct ocs_hcu_dev *hcu_dev,
- struct ocs_hcu_dma_list *dma_list,
- dma_addr_t addr, u32 len)
- {
- struct device *dev = hcu_dev->dev;
- struct ocs_hcu_dma_entry *old_tail;
- struct ocs_hcu_dma_entry *new_tail;
- if (!len)
- return 0;
- if (!dma_list)
- return -EINVAL;
- if (addr & ~OCS_HCU_DMA_BIT_MASK) {
- dev_err(dev,
- "Unexpected error: Invalid DMA address for OCS HCU\n");
- return -EINVAL;
- }
- old_tail = dma_list->tail;
- new_tail = old_tail ? old_tail + 1 : dma_list->head;
- /* Check if list is full. */
- if (new_tail - dma_list->head >= dma_list->max_nents)
- return -ENOMEM;
- /*
- * If there was an old tail (i.e., this is not the first element we are
- * adding), un-terminate the old tail and make it point to the new one.
- */
- if (old_tail) {
- old_tail->ll_flags &= ~OCS_LL_DMA_FLAG_TERMINATE;
- /*
- * The old tail 'nxt_desc' must point to the DMA address of the
- * new tail.
- */
- old_tail->nxt_desc = dma_list->dma_addr +
- sizeof(*dma_list->tail) * (new_tail -
- dma_list->head);
- }
- new_tail->src_addr = (u32)addr;
- new_tail->src_len = (u32)len;
- new_tail->ll_flags = OCS_LL_DMA_FLAG_TERMINATE;
- new_tail->nxt_desc = 0;
- /* Update list tail with new tail. */
- dma_list->tail = new_tail;
- return 0;
- }
- /**
- * ocs_hcu_hash_init() - Initialize hash operation context.
- * @ctx: The context to initialize.
- * @algo: The hashing algorithm to use.
- *
- * Return: 0 on success, negative error code otherwise.
- */
- int ocs_hcu_hash_init(struct ocs_hcu_hash_ctx *ctx, enum ocs_hcu_algo algo)
- {
- if (!ctx)
- return -EINVAL;
- ctx->algo = algo;
- ctx->idata.msg_len_lo = 0;
- ctx->idata.msg_len_hi = 0;
- /* No need to set idata.digest to 0. */
- return 0;
- }
- /**
- * ocs_hcu_hash_update() - Perform a hashing iteration.
- * @hcu_dev: The OCS HCU device to use.
- * @ctx: The OCS HCU hashing context.
- * @dma_list: The OCS DMA list mapping the input data to process.
- *
- * Return: 0 on success; negative error code otherwise.
- */
- int ocs_hcu_hash_update(struct ocs_hcu_dev *hcu_dev,
- struct ocs_hcu_hash_ctx *ctx,
- const struct ocs_hcu_dma_list *dma_list)
- {
- int rc;
- if (!hcu_dev || !ctx)
- return -EINVAL;
- /* Configure the hardware for the current request. */
- rc = ocs_hcu_hw_cfg(hcu_dev, ctx->algo, false);
- if (rc)
- return rc;
- /* If we already processed some data, idata needs to be set. */
- if (ctx->idata.msg_len_lo || ctx->idata.msg_len_hi)
- ocs_hcu_set_intermediate_data(hcu_dev, &ctx->idata, ctx->algo);
- /* Start linked-list DMA hashing. */
- rc = ocs_hcu_ll_dma_start(hcu_dev, dma_list, false);
- if (rc)
- return rc;
- /* Update idata and return. */
- return ocs_hcu_get_intermediate_data(hcu_dev, &ctx->idata, ctx->algo);
- }
- /**
- * ocs_hcu_hash_finup() - Update and finalize hash computation.
- * @hcu_dev: The OCS HCU device to use.
- * @ctx: The OCS HCU hashing context.
- * @dma_list: The OCS DMA list mapping the input data to process.
- * @dgst: The buffer where to save the computed digest.
- * @dgst_len: The length of @dgst.
- *
- * Return: 0 on success; negative error code otherwise.
- */
- int ocs_hcu_hash_finup(struct ocs_hcu_dev *hcu_dev,
- const struct ocs_hcu_hash_ctx *ctx,
- const struct ocs_hcu_dma_list *dma_list,
- u8 *dgst, size_t dgst_len)
- {
- int rc;
- if (!hcu_dev || !ctx)
- return -EINVAL;
- /* Configure the hardware for the current request. */
- rc = ocs_hcu_hw_cfg(hcu_dev, ctx->algo, false);
- if (rc)
- return rc;
- /* If we already processed some data, idata needs to be set. */
- if (ctx->idata.msg_len_lo || ctx->idata.msg_len_hi)
- ocs_hcu_set_intermediate_data(hcu_dev, &ctx->idata, ctx->algo);
- /* Start linked-list DMA hashing. */
- rc = ocs_hcu_ll_dma_start(hcu_dev, dma_list, true);
- if (rc)
- return rc;
- /* Get digest and return. */
- return ocs_hcu_get_digest(hcu_dev, ctx->algo, dgst, dgst_len);
- }
- /**
- * ocs_hcu_hash_final() - Finalize hash computation.
- * @hcu_dev: The OCS HCU device to use.
- * @ctx: The OCS HCU hashing context.
- * @dgst: The buffer where to save the computed digest.
- * @dgst_len: The length of @dgst.
- *
- * Return: 0 on success; negative error code otherwise.
- */
- int ocs_hcu_hash_final(struct ocs_hcu_dev *hcu_dev,
- const struct ocs_hcu_hash_ctx *ctx, u8 *dgst,
- size_t dgst_len)
- {
- int rc;
- if (!hcu_dev || !ctx)
- return -EINVAL;
- /* Configure the hardware for the current request. */
- rc = ocs_hcu_hw_cfg(hcu_dev, ctx->algo, false);
- if (rc)
- return rc;
- /* If we already processed some data, idata needs to be set. */
- if (ctx->idata.msg_len_lo || ctx->idata.msg_len_hi)
- ocs_hcu_set_intermediate_data(hcu_dev, &ctx->idata, ctx->algo);
- /*
- * Enable HCU interrupts, so that HCU_DONE will be triggered once the
- * final hash is computed.
- */
- ocs_hcu_done_irq_en(hcu_dev);
- reinit_completion(&hcu_dev->irq_done);
- writel(OCS_HCU_TERMINATE, hcu_dev->io_base + OCS_HCU_OPERATION);
- rc = ocs_hcu_wait_and_disable_irq(hcu_dev);
- if (rc)
- return rc;
- /* Get digest and return. */
- return ocs_hcu_get_digest(hcu_dev, ctx->algo, dgst, dgst_len);
- }
- /**
- * ocs_hcu_digest() - Compute hash digest.
- * @hcu_dev: The OCS HCU device to use.
- * @algo: The hash algorithm to use.
- * @data: The input data to process.
- * @data_len: The length of @data.
- * @dgst: The buffer where to save the computed digest.
- * @dgst_len: The length of @dgst.
- *
- * Return: 0 on success; negative error code otherwise.
- */
- int ocs_hcu_digest(struct ocs_hcu_dev *hcu_dev, enum ocs_hcu_algo algo,
- void *data, size_t data_len, u8 *dgst, size_t dgst_len)
- {
- struct device *dev = hcu_dev->dev;
- dma_addr_t dma_handle;
- u32 reg;
- int rc;
- /* Configure the hardware for the current request. */
- rc = ocs_hcu_hw_cfg(hcu_dev, algo, false);
- if (rc)
- return rc;
- dma_handle = dma_map_single(dev, data, data_len, DMA_TO_DEVICE);
- if (dma_mapping_error(dev, dma_handle))
- return -EIO;
- reg = HCU_DMA_SNOOP_MASK | HCU_DMA_EN;
- ocs_hcu_done_irq_en(hcu_dev);
- reinit_completion(&hcu_dev->irq_done);
- writel(dma_handle, hcu_dev->io_base + OCS_HCU_DMA_SRC_ADDR);
- writel(data_len, hcu_dev->io_base + OCS_HCU_DMA_SRC_SIZE);
- writel(OCS_HCU_START, hcu_dev->io_base + OCS_HCU_OPERATION);
- writel(reg, hcu_dev->io_base + OCS_HCU_DMA_DMA_MODE);
- writel(OCS_HCU_TERMINATE, hcu_dev->io_base + OCS_HCU_OPERATION);
- rc = ocs_hcu_wait_and_disable_irq(hcu_dev);
- if (rc)
- return rc;
- dma_unmap_single(dev, dma_handle, data_len, DMA_TO_DEVICE);
- return ocs_hcu_get_digest(hcu_dev, algo, dgst, dgst_len);
- }
- /**
- * ocs_hcu_hmac() - Compute HMAC.
- * @hcu_dev: The OCS HCU device to use.
- * @algo: The hash algorithm to use with HMAC.
- * @key: The key to use.
- * @dma_list: The OCS DMA list mapping the input data to process.
- * @key_len: The length of @key.
- * @dgst: The buffer where to save the computed HMAC.
- * @dgst_len: The length of @dgst.
- *
- * Return: 0 on success; negative error code otherwise.
- */
- int ocs_hcu_hmac(struct ocs_hcu_dev *hcu_dev, enum ocs_hcu_algo algo,
- const u8 *key, size_t key_len,
- const struct ocs_hcu_dma_list *dma_list,
- u8 *dgst, size_t dgst_len)
- {
- int rc;
- /* Ensure 'key' is not NULL. */
- if (!key || key_len == 0)
- return -EINVAL;
- /* Configure the hardware for the current request. */
- rc = ocs_hcu_hw_cfg(hcu_dev, algo, true);
- if (rc)
- return rc;
- rc = ocs_hcu_write_key(hcu_dev, key, key_len);
- if (rc)
- return rc;
- rc = ocs_hcu_ll_dma_start(hcu_dev, dma_list, true);
- /* Clear HW key before processing return code. */
- ocs_hcu_clear_key(hcu_dev);
- if (rc)
- return rc;
- return ocs_hcu_get_digest(hcu_dev, algo, dgst, dgst_len);
- }
- irqreturn_t ocs_hcu_irq_handler(int irq, void *dev_id)
- {
- struct ocs_hcu_dev *hcu_dev = dev_id;
- u32 hcu_irq;
- u32 dma_irq;
- /* Read and clear the HCU interrupt. */
- hcu_irq = readl(hcu_dev->io_base + OCS_HCU_ISR);
- writel(hcu_irq, hcu_dev->io_base + OCS_HCU_ISR);
- /* Read and clear the HCU DMA interrupt. */
- dma_irq = readl(hcu_dev->io_base + OCS_HCU_DMA_MSI_ISR);
- writel(dma_irq, hcu_dev->io_base + OCS_HCU_DMA_MSI_ISR);
- /* Check for errors. */
- if (hcu_irq & HCU_IRQ_HASH_ERR_MASK || dma_irq & HCU_DMA_IRQ_ERR_MASK) {
- hcu_dev->irq_err = true;
- goto complete;
- }
- /* Check for DONE IRQs. */
- if (hcu_irq & HCU_IRQ_HASH_DONE || dma_irq & HCU_DMA_IRQ_SRC_DONE)
- goto complete;
- return IRQ_NONE;
- complete:
- complete(&hcu_dev->irq_done);
- return IRQ_HANDLED;
- }
- MODULE_LICENSE("GPL");
|