1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677787980818283848586878889909192939495969798991001011021031041051061071081091101111121131141151161171181191201211221231241251261271281291301311321331341351361371381391401411421431441451461471481491501511521531541551561571581591601611621631641651661671681691701711721731741751761771781791801811821831841851861871881891901911921931941951961971981992002012022032042052062072082092102112122132142152162172182192202212222232242252262272282292302312322332342352362372382392402412422432442452462472482492502512522532542552562572582592602612622632642652662672682692702712722732742752762772782792802812822832842852862872882892902912922932942952962972982993003013023033043053063073083093103113123133143153163173183193203213223233243253263273283293303313323333343353363373383393403413423433443453463473483493503513523533543553563573583593603613623633643653663673683693703713723733743753763773783793803813823833843853863873883893903913923933943953963973983994004014024034044054064074084094104114124134144154164174184194204214224234244254264274284294304314324334344354364374384394404414424434444454464474484494504514524534544554564574584594604614624634644654664674684694704714724734744754764774784794804814824834844854864874884894904914924934944954964974984995005015025035045055065075085095105115125135145155165175185195205215225235245255265275285295305315325335345355365375385395405415425435445455465475485495505515525535545555565575585595605615625635645655665675685695705715725735745755765775785795805815825835845855865875885895905915925935945955965975985996006016026036046056066076086096106116126136146156166176186196206216226236246256266276286296306316326336346356366376386396406416426436446456466476486496506516526536546556566576586596606616626636646656666676686696706716726736746756766776786796806816826836846856866876886896906916926936946956966976986997007017027037047057067077087097107117127137147157167177187197207217227237247257267277287297307317327337347357367377387397407417427437447457467477487497507517527537547557567577587597607617627637647657667677687697707717727737747757767777787797807817827837847857867877887897907917927937947957967977987998008018028038048058068078088098108118128138148158168178188198208218228238248258268278288298308318328338348358368378388398408418428438448458468478488498508518528538548558568578588598608618628638648658668678688698708718728738748758768778788798808818828838848858868878888898908918928938948958968978988999009019029039049059069079089099109119129139149159169179189199209219229239249259269279289299309319329339349359369379389399409419429439449459469479489499509519529539549559569579589599609619629639649659669679689699709719729739749759769779789799809819829839849859869879889899909919929939949959969979989991000100110021003100410051006100710081009101010111012101310141015101610171018101910201021102210231024102510261027102810291030103110321033103410351036103710381039104010411042104310441045104610471048104910501051105210531054105510561057105810591060106110621063106410651066106710681069107010711072107310741075107610771078107910801081108210831084108510861087108810891090109110921093109410951096109710981099110011011102110311041105110611071108110911101111111211131114111511161117111811191120112111221123112411251126112711281129113011311132113311341135113611371138113911401141114211431144114511461147114811491150115111521153115411551156115711581159116011611162116311641165116611671168116911701171117211731174117511761177117811791180118111821183118411851186118711881189119011911192119311941195119611971198119912001201120212031204120512061207120812091210121112121213121412151216121712181219122012211222122312241225122612271228122912301231123212331234123512361237123812391240124112421243124412451246124712481249125012511252125312541255125612571258125912601261126212631264126512661267126812691270127112721273127412751276127712781279128012811282128312841285128612871288128912901291129212931294129512961297129812991300130113021303130413051306130713081309131013111312131313141315131613171318131913201321132213231324132513261327132813291330133113321333133413351336133713381339134013411342134313441345134613471348134913501351135213531354135513561357135813591360136113621363136413651366136713681369137013711372137313741375137613771378137913801381138213831384138513861387138813891390139113921393139413951396139713981399140014011402140314041405140614071408140914101411141214131414141514161417141814191420142114221423142414251426142714281429143014311432143314341435143614371438143914401441144214431444144514461447144814491450145114521453145414551456145714581459146014611462146314641465146614671468146914701471147214731474147514761477147814791480148114821483148414851486148714881489 |
- // SPDX-License-Identifier: GPL-2.0-only
- /*
- * Intel Keem Bay OCS AES Crypto Driver.
- *
- * Copyright (C) 2018-2020 Intel Corporation
- */
- #include <linux/dma-mapping.h>
- #include <linux/interrupt.h>
- #include <linux/platform_device.h>
- #include <linux/slab.h>
- #include <linux/swab.h>
- #include <asm/byteorder.h>
- #include <asm/errno.h>
- #include <crypto/aes.h>
- #include <crypto/gcm.h>
- #include "ocs-aes.h"
- #define AES_COMMAND_OFFSET 0x0000
- #define AES_KEY_0_OFFSET 0x0004
- #define AES_KEY_1_OFFSET 0x0008
- #define AES_KEY_2_OFFSET 0x000C
- #define AES_KEY_3_OFFSET 0x0010
- #define AES_KEY_4_OFFSET 0x0014
- #define AES_KEY_5_OFFSET 0x0018
- #define AES_KEY_6_OFFSET 0x001C
- #define AES_KEY_7_OFFSET 0x0020
- #define AES_IV_0_OFFSET 0x0024
- #define AES_IV_1_OFFSET 0x0028
- #define AES_IV_2_OFFSET 0x002C
- #define AES_IV_3_OFFSET 0x0030
- #define AES_ACTIVE_OFFSET 0x0034
- #define AES_STATUS_OFFSET 0x0038
- #define AES_KEY_SIZE_OFFSET 0x0044
- #define AES_IER_OFFSET 0x0048
- #define AES_ISR_OFFSET 0x005C
- #define AES_MULTIPURPOSE1_0_OFFSET 0x0200
- #define AES_MULTIPURPOSE1_1_OFFSET 0x0204
- #define AES_MULTIPURPOSE1_2_OFFSET 0x0208
- #define AES_MULTIPURPOSE1_3_OFFSET 0x020C
- #define AES_MULTIPURPOSE2_0_OFFSET 0x0220
- #define AES_MULTIPURPOSE2_1_OFFSET 0x0224
- #define AES_MULTIPURPOSE2_2_OFFSET 0x0228
- #define AES_MULTIPURPOSE2_3_OFFSET 0x022C
- #define AES_BYTE_ORDER_CFG_OFFSET 0x02C0
- #define AES_TLEN_OFFSET 0x0300
- #define AES_T_MAC_0_OFFSET 0x0304
- #define AES_T_MAC_1_OFFSET 0x0308
- #define AES_T_MAC_2_OFFSET 0x030C
- #define AES_T_MAC_3_OFFSET 0x0310
- #define AES_PLEN_OFFSET 0x0314
- #define AES_A_DMA_SRC_ADDR_OFFSET 0x0400
- #define AES_A_DMA_DST_ADDR_OFFSET 0x0404
- #define AES_A_DMA_SRC_SIZE_OFFSET 0x0408
- #define AES_A_DMA_DST_SIZE_OFFSET 0x040C
- #define AES_A_DMA_DMA_MODE_OFFSET 0x0410
- #define AES_A_DMA_NEXT_SRC_DESCR_OFFSET 0x0418
- #define AES_A_DMA_NEXT_DST_DESCR_OFFSET 0x041C
- #define AES_A_DMA_WHILE_ACTIVE_MODE_OFFSET 0x0420
- #define AES_A_DMA_LOG_OFFSET 0x0424
- #define AES_A_DMA_STATUS_OFFSET 0x0428
- #define AES_A_DMA_PERF_CNTR_OFFSET 0x042C
- #define AES_A_DMA_MSI_ISR_OFFSET 0x0480
- #define AES_A_DMA_MSI_IER_OFFSET 0x0484
- #define AES_A_DMA_MSI_MASK_OFFSET 0x0488
- #define AES_A_DMA_INBUFFER_WRITE_FIFO_OFFSET 0x0600
- #define AES_A_DMA_OUTBUFFER_READ_FIFO_OFFSET 0x0700
- /*
- * AES_A_DMA_DMA_MODE register.
- * Default: 0x00000000.
- * bit[31] ACTIVE
- * This bit activates the DMA. When the DMA finishes, it resets
- * this bit to zero.
- * bit[30:26] Unused by this driver.
- * bit[25] SRC_LINK_LIST_EN
- * Source link list enable bit. When the linked list is terminated
- * this bit is reset by the DMA.
- * bit[24] DST_LINK_LIST_EN
- * Destination link list enable bit. When the linked list is
- * terminated this bit is reset by the DMA.
- * bit[23:0] Unused by this driver.
- */
- #define AES_A_DMA_DMA_MODE_ACTIVE BIT(31)
- #define AES_A_DMA_DMA_MODE_SRC_LINK_LIST_EN BIT(25)
- #define AES_A_DMA_DMA_MODE_DST_LINK_LIST_EN BIT(24)
- /*
- * AES_ACTIVE register
- * default 0x00000000
- * bit[31:10] Reserved
- * bit[9] LAST_ADATA
- * bit[8] LAST_GCX
- * bit[7:2] Reserved
- * bit[1] TERMINATION
- * bit[0] TRIGGER
- */
- #define AES_ACTIVE_LAST_ADATA BIT(9)
- #define AES_ACTIVE_LAST_CCM_GCM BIT(8)
- #define AES_ACTIVE_TERMINATION BIT(1)
- #define AES_ACTIVE_TRIGGER BIT(0)
- #define AES_DISABLE_INT 0x00000000
- #define AES_DMA_CPD_ERR_INT BIT(8)
- #define AES_DMA_OUTBUF_RD_ERR_INT BIT(7)
- #define AES_DMA_OUTBUF_WR_ERR_INT BIT(6)
- #define AES_DMA_INBUF_RD_ERR_INT BIT(5)
- #define AES_DMA_INBUF_WR_ERR_INT BIT(4)
- #define AES_DMA_BAD_COMP_INT BIT(3)
- #define AES_DMA_SAI_INT BIT(2)
- #define AES_DMA_SRC_DONE_INT BIT(0)
- #define AES_COMPLETE_INT BIT(1)
- #define AES_DMA_MSI_MASK_CLEAR BIT(0)
- #define AES_128_BIT_KEY 0x00000000
- #define AES_256_BIT_KEY BIT(0)
- #define AES_DEACTIVATE_PERF_CNTR 0x00000000
- #define AES_ACTIVATE_PERF_CNTR BIT(0)
- #define AES_MAX_TAG_SIZE_U32 4
- #define OCS_LL_DMA_FLAG_TERMINATE BIT(31)
- /*
- * There is an inconsistency in the documentation. This is documented as a
- * 11-bit value, but it is actually 10-bits.
- */
- #define AES_DMA_STATUS_INPUT_BUFFER_OCCUPANCY_MASK 0x3FF
- /*
- * During CCM decrypt, the OCS block needs to finish processing the ciphertext
- * before the tag is written. For 128-bit mode this required delay is 28 OCS
- * clock cycles. For 256-bit mode it is 36 OCS clock cycles.
- */
- #define CCM_DECRYPT_DELAY_TAG_CLK_COUNT 36UL
- /*
- * During CCM decrypt there must be a delay of at least 42 OCS clock cycles
- * between setting the TRIGGER bit in AES_ACTIVE and setting the LAST_CCM_GCM
- * bit in the same register (as stated in the OCS databook)
- */
- #define CCM_DECRYPT_DELAY_LAST_GCX_CLK_COUNT 42UL
- /* See RFC3610 section 2.2 */
- #define L_PRIME_MIN (1)
- #define L_PRIME_MAX (7)
- /*
- * CCM IV format from RFC 3610 section 2.3
- *
- * Octet Number Contents
- * ------------ ---------
- * 0 Flags
- * 1 ... 15-L Nonce N
- * 16-L ... 15 Counter i
- *
- * Flags = L' = L - 1
- */
- #define L_PRIME_IDX 0
- #define COUNTER_START(lprime) (16 - ((lprime) + 1))
- #define COUNTER_LEN(lprime) ((lprime) + 1)
- enum aes_counter_mode {
- AES_CTR_M_NO_INC = 0,
- AES_CTR_M_32_INC = 1,
- AES_CTR_M_64_INC = 2,
- AES_CTR_M_128_INC = 3,
- };
- /**
- * struct ocs_dma_linked_list - OCS DMA linked list entry.
- * @src_addr: Source address of the data.
- * @src_len: Length of data to be fetched.
- * @next: Next dma_list to fetch.
- * @ll_flags: Flags (Freeze @ terminate) for the DMA engine.
- */
- struct ocs_dma_linked_list {
- u32 src_addr;
- u32 src_len;
- u32 next;
- u32 ll_flags;
- } __packed;
- /*
- * Set endianness of inputs and outputs
- * AES_BYTE_ORDER_CFG
- * default 0x00000000
- * bit [10] - KEY_HI_LO_SWAP
- * bit [9] - KEY_HI_SWAP_DWORDS_IN_OCTWORD
- * bit [8] - KEY_HI_SWAP_BYTES_IN_DWORD
- * bit [7] - KEY_LO_SWAP_DWORDS_IN_OCTWORD
- * bit [6] - KEY_LO_SWAP_BYTES_IN_DWORD
- * bit [5] - IV_SWAP_DWORDS_IN_OCTWORD
- * bit [4] - IV_SWAP_BYTES_IN_DWORD
- * bit [3] - DOUT_SWAP_DWORDS_IN_OCTWORD
- * bit [2] - DOUT_SWAP_BYTES_IN_DWORD
- * bit [1] - DOUT_SWAP_DWORDS_IN_OCTWORD
- * bit [0] - DOUT_SWAP_BYTES_IN_DWORD
- */
- static inline void aes_a_set_endianness(const struct ocs_aes_dev *aes_dev)
- {
- iowrite32(0x7FF, aes_dev->base_reg + AES_BYTE_ORDER_CFG_OFFSET);
- }
- /* Trigger AES process start. */
- static inline void aes_a_op_trigger(const struct ocs_aes_dev *aes_dev)
- {
- iowrite32(AES_ACTIVE_TRIGGER, aes_dev->base_reg + AES_ACTIVE_OFFSET);
- }
- /* Indicate last bulk of data. */
- static inline void aes_a_op_termination(const struct ocs_aes_dev *aes_dev)
- {
- iowrite32(AES_ACTIVE_TERMINATION,
- aes_dev->base_reg + AES_ACTIVE_OFFSET);
- }
- /*
- * Set LAST_CCM_GCM in AES_ACTIVE register and clear all other bits.
- *
- * Called when DMA is programmed to fetch the last batch of data.
- * - For AES-CCM it is called for the last batch of Payload data and Ciphertext
- * data.
- * - For AES-GCM, it is called for the last batch of Plaintext data and
- * Ciphertext data.
- */
- static inline void aes_a_set_last_gcx(const struct ocs_aes_dev *aes_dev)
- {
- iowrite32(AES_ACTIVE_LAST_CCM_GCM,
- aes_dev->base_reg + AES_ACTIVE_OFFSET);
- }
- /* Wait for LAST_CCM_GCM bit to be unset. */
- static inline void aes_a_wait_last_gcx(const struct ocs_aes_dev *aes_dev)
- {
- u32 aes_active_reg;
- do {
- aes_active_reg = ioread32(aes_dev->base_reg +
- AES_ACTIVE_OFFSET);
- } while (aes_active_reg & AES_ACTIVE_LAST_CCM_GCM);
- }
- /* Wait for 10 bits of input occupancy. */
- static void aes_a_dma_wait_input_buffer_occupancy(const struct ocs_aes_dev *aes_dev)
- {
- u32 reg;
- do {
- reg = ioread32(aes_dev->base_reg + AES_A_DMA_STATUS_OFFSET);
- } while (reg & AES_DMA_STATUS_INPUT_BUFFER_OCCUPANCY_MASK);
- }
- /*
- * Set LAST_CCM_GCM and LAST_ADATA bits in AES_ACTIVE register (and clear all
- * other bits).
- *
- * Called when DMA is programmed to fetch the last batch of Associated Data
- * (CCM case) or Additional Authenticated Data (GCM case).
- */
- static inline void aes_a_set_last_gcx_and_adata(const struct ocs_aes_dev *aes_dev)
- {
- iowrite32(AES_ACTIVE_LAST_ADATA | AES_ACTIVE_LAST_CCM_GCM,
- aes_dev->base_reg + AES_ACTIVE_OFFSET);
- }
- /* Set DMA src and dst transfer size to 0 */
- static inline void aes_a_dma_set_xfer_size_zero(const struct ocs_aes_dev *aes_dev)
- {
- iowrite32(0, aes_dev->base_reg + AES_A_DMA_SRC_SIZE_OFFSET);
- iowrite32(0, aes_dev->base_reg + AES_A_DMA_DST_SIZE_OFFSET);
- }
- /* Activate DMA for zero-byte transfer case. */
- static inline void aes_a_dma_active(const struct ocs_aes_dev *aes_dev)
- {
- iowrite32(AES_A_DMA_DMA_MODE_ACTIVE,
- aes_dev->base_reg + AES_A_DMA_DMA_MODE_OFFSET);
- }
- /* Activate DMA and enable src linked list */
- static inline void aes_a_dma_active_src_ll_en(const struct ocs_aes_dev *aes_dev)
- {
- iowrite32(AES_A_DMA_DMA_MODE_ACTIVE |
- AES_A_DMA_DMA_MODE_SRC_LINK_LIST_EN,
- aes_dev->base_reg + AES_A_DMA_DMA_MODE_OFFSET);
- }
- /* Activate DMA and enable dst linked list */
- static inline void aes_a_dma_active_dst_ll_en(const struct ocs_aes_dev *aes_dev)
- {
- iowrite32(AES_A_DMA_DMA_MODE_ACTIVE |
- AES_A_DMA_DMA_MODE_DST_LINK_LIST_EN,
- aes_dev->base_reg + AES_A_DMA_DMA_MODE_OFFSET);
- }
- /* Activate DMA and enable src and dst linked lists */
- static inline void aes_a_dma_active_src_dst_ll_en(const struct ocs_aes_dev *aes_dev)
- {
- iowrite32(AES_A_DMA_DMA_MODE_ACTIVE |
- AES_A_DMA_DMA_MODE_SRC_LINK_LIST_EN |
- AES_A_DMA_DMA_MODE_DST_LINK_LIST_EN,
- aes_dev->base_reg + AES_A_DMA_DMA_MODE_OFFSET);
- }
- /* Reset PERF_CNTR to 0 and activate it */
- static inline void aes_a_dma_reset_and_activate_perf_cntr(const struct ocs_aes_dev *aes_dev)
- {
- iowrite32(0x00000000, aes_dev->base_reg + AES_A_DMA_PERF_CNTR_OFFSET);
- iowrite32(AES_ACTIVATE_PERF_CNTR,
- aes_dev->base_reg + AES_A_DMA_WHILE_ACTIVE_MODE_OFFSET);
- }
- /* Wait until PERF_CNTR is > delay, then deactivate it */
- static inline void aes_a_dma_wait_and_deactivate_perf_cntr(const struct ocs_aes_dev *aes_dev,
- int delay)
- {
- while (ioread32(aes_dev->base_reg + AES_A_DMA_PERF_CNTR_OFFSET) < delay)
- ;
- iowrite32(AES_DEACTIVATE_PERF_CNTR,
- aes_dev->base_reg + AES_A_DMA_WHILE_ACTIVE_MODE_OFFSET);
- }
- /* Disable AES and DMA IRQ. */
- static void aes_irq_disable(struct ocs_aes_dev *aes_dev)
- {
- u32 isr_val = 0;
- /* Disable interrupts */
- iowrite32(AES_DISABLE_INT,
- aes_dev->base_reg + AES_A_DMA_MSI_IER_OFFSET);
- iowrite32(AES_DISABLE_INT, aes_dev->base_reg + AES_IER_OFFSET);
- /* Clear any pending interrupt */
- isr_val = ioread32(aes_dev->base_reg + AES_A_DMA_MSI_ISR_OFFSET);
- if (isr_val)
- iowrite32(isr_val,
- aes_dev->base_reg + AES_A_DMA_MSI_ISR_OFFSET);
- isr_val = ioread32(aes_dev->base_reg + AES_A_DMA_MSI_MASK_OFFSET);
- if (isr_val)
- iowrite32(isr_val,
- aes_dev->base_reg + AES_A_DMA_MSI_MASK_OFFSET);
- isr_val = ioread32(aes_dev->base_reg + AES_ISR_OFFSET);
- if (isr_val)
- iowrite32(isr_val, aes_dev->base_reg + AES_ISR_OFFSET);
- }
- /* Enable AES or DMA IRQ. IRQ is disabled once fired. */
- static void aes_irq_enable(struct ocs_aes_dev *aes_dev, u8 irq)
- {
- if (irq == AES_COMPLETE_INT) {
- /* Ensure DMA error interrupts are enabled */
- iowrite32(AES_DMA_CPD_ERR_INT |
- AES_DMA_OUTBUF_RD_ERR_INT |
- AES_DMA_OUTBUF_WR_ERR_INT |
- AES_DMA_INBUF_RD_ERR_INT |
- AES_DMA_INBUF_WR_ERR_INT |
- AES_DMA_BAD_COMP_INT |
- AES_DMA_SAI_INT,
- aes_dev->base_reg + AES_A_DMA_MSI_IER_OFFSET);
- /*
- * AES_IER
- * default 0x00000000
- * bits [31:3] - reserved
- * bit [2] - EN_SKS_ERR
- * bit [1] - EN_AES_COMPLETE
- * bit [0] - reserved
- */
- iowrite32(AES_COMPLETE_INT, aes_dev->base_reg + AES_IER_OFFSET);
- return;
- }
- if (irq == AES_DMA_SRC_DONE_INT) {
- /* Ensure AES interrupts are disabled */
- iowrite32(AES_DISABLE_INT, aes_dev->base_reg + AES_IER_OFFSET);
- /*
- * DMA_MSI_IER
- * default 0x00000000
- * bits [31:9] - reserved
- * bit [8] - CPD_ERR_INT_EN
- * bit [7] - OUTBUF_RD_ERR_INT_EN
- * bit [6] - OUTBUF_WR_ERR_INT_EN
- * bit [5] - INBUF_RD_ERR_INT_EN
- * bit [4] - INBUF_WR_ERR_INT_EN
- * bit [3] - BAD_COMP_INT_EN
- * bit [2] - SAI_INT_EN
- * bit [1] - DST_DONE_INT_EN
- * bit [0] - SRC_DONE_INT_EN
- */
- iowrite32(AES_DMA_CPD_ERR_INT |
- AES_DMA_OUTBUF_RD_ERR_INT |
- AES_DMA_OUTBUF_WR_ERR_INT |
- AES_DMA_INBUF_RD_ERR_INT |
- AES_DMA_INBUF_WR_ERR_INT |
- AES_DMA_BAD_COMP_INT |
- AES_DMA_SAI_INT |
- AES_DMA_SRC_DONE_INT,
- aes_dev->base_reg + AES_A_DMA_MSI_IER_OFFSET);
- }
- }
- /* Enable and wait for IRQ (either from OCS AES engine or DMA) */
- static int ocs_aes_irq_enable_and_wait(struct ocs_aes_dev *aes_dev, u8 irq)
- {
- int rc;
- reinit_completion(&aes_dev->irq_completion);
- aes_irq_enable(aes_dev, irq);
- rc = wait_for_completion_interruptible(&aes_dev->irq_completion);
- if (rc)
- return rc;
- return aes_dev->dma_err_mask ? -EIO : 0;
- }
- /* Configure DMA to OCS, linked list mode */
- static inline void dma_to_ocs_aes_ll(struct ocs_aes_dev *aes_dev,
- dma_addr_t dma_list)
- {
- iowrite32(0, aes_dev->base_reg + AES_A_DMA_SRC_SIZE_OFFSET);
- iowrite32(dma_list,
- aes_dev->base_reg + AES_A_DMA_NEXT_SRC_DESCR_OFFSET);
- }
- /* Configure DMA from OCS, linked list mode */
- static inline void dma_from_ocs_aes_ll(struct ocs_aes_dev *aes_dev,
- dma_addr_t dma_list)
- {
- iowrite32(0, aes_dev->base_reg + AES_A_DMA_DST_SIZE_OFFSET);
- iowrite32(dma_list,
- aes_dev->base_reg + AES_A_DMA_NEXT_DST_DESCR_OFFSET);
- }
- irqreturn_t ocs_aes_irq_handler(int irq, void *dev_id)
- {
- struct ocs_aes_dev *aes_dev = dev_id;
- u32 aes_dma_isr;
- /* Read DMA ISR status. */
- aes_dma_isr = ioread32(aes_dev->base_reg + AES_A_DMA_MSI_ISR_OFFSET);
- /* Disable and clear interrupts. */
- aes_irq_disable(aes_dev);
- /* Save DMA error status. */
- aes_dev->dma_err_mask = aes_dma_isr &
- (AES_DMA_CPD_ERR_INT |
- AES_DMA_OUTBUF_RD_ERR_INT |
- AES_DMA_OUTBUF_WR_ERR_INT |
- AES_DMA_INBUF_RD_ERR_INT |
- AES_DMA_INBUF_WR_ERR_INT |
- AES_DMA_BAD_COMP_INT |
- AES_DMA_SAI_INT);
- /* Signal IRQ completion. */
- complete(&aes_dev->irq_completion);
- return IRQ_HANDLED;
- }
- /**
- * ocs_aes_set_key() - Write key into OCS AES hardware.
- * @aes_dev: The OCS AES device to write the key to.
- * @key_size: The size of the key (in bytes).
- * @key: The key to write.
- * @cipher: The cipher the key is for.
- *
- * For AES @key_size must be either 16 or 32. For SM4 @key_size must be 16.
- *
- * Return: 0 on success, negative error code otherwise.
- */
- int ocs_aes_set_key(struct ocs_aes_dev *aes_dev, u32 key_size, const u8 *key,
- enum ocs_cipher cipher)
- {
- const u32 *key_u32;
- u32 val;
- int i;
- /* OCS AES supports 128-bit and 256-bit keys only. */
- if (cipher == OCS_AES && !(key_size == 32 || key_size == 16)) {
- dev_err(aes_dev->dev,
- "%d-bit keys not supported by AES cipher\n",
- key_size * 8);
- return -EINVAL;
- }
- /* OCS SM4 supports 128-bit keys only. */
- if (cipher == OCS_SM4 && key_size != 16) {
- dev_err(aes_dev->dev,
- "%d-bit keys not supported for SM4 cipher\n",
- key_size * 8);
- return -EINVAL;
- }
- if (!key)
- return -EINVAL;
- key_u32 = (const u32 *)key;
- /* Write key to AES_KEY[0-7] registers */
- for (i = 0; i < (key_size / sizeof(u32)); i++) {
- iowrite32(key_u32[i],
- aes_dev->base_reg + AES_KEY_0_OFFSET +
- (i * sizeof(u32)));
- }
- /*
- * Write key size
- * bits [31:1] - reserved
- * bit [0] - AES_KEY_SIZE
- * 0 - 128 bit key
- * 1 - 256 bit key
- */
- val = (key_size == 16) ? AES_128_BIT_KEY : AES_256_BIT_KEY;
- iowrite32(val, aes_dev->base_reg + AES_KEY_SIZE_OFFSET);
- return 0;
- }
- /* Write AES_COMMAND */
- static inline void set_ocs_aes_command(struct ocs_aes_dev *aes_dev,
- enum ocs_cipher cipher,
- enum ocs_mode mode,
- enum ocs_instruction instruction)
- {
- u32 val;
- /* AES_COMMAND
- * default 0x000000CC
- * bit [14] - CIPHER_SELECT
- * 0 - AES
- * 1 - SM4
- * bits [11:8] - OCS_AES_MODE
- * 0000 - ECB
- * 0001 - CBC
- * 0010 - CTR
- * 0110 - CCM
- * 0111 - GCM
- * 1001 - CTS
- * bits [7:6] - AES_INSTRUCTION
- * 00 - ENCRYPT
- * 01 - DECRYPT
- * 10 - EXPAND
- * 11 - BYPASS
- * bits [3:2] - CTR_M_BITS
- * 00 - No increment
- * 01 - Least significant 32 bits are incremented
- * 10 - Least significant 64 bits are incremented
- * 11 - Full 128 bits are incremented
- */
- val = (cipher << 14) | (mode << 8) | (instruction << 6) |
- (AES_CTR_M_128_INC << 2);
- iowrite32(val, aes_dev->base_reg + AES_COMMAND_OFFSET);
- }
- static void ocs_aes_init(struct ocs_aes_dev *aes_dev,
- enum ocs_mode mode,
- enum ocs_cipher cipher,
- enum ocs_instruction instruction)
- {
- /* Ensure interrupts are disabled and pending interrupts cleared. */
- aes_irq_disable(aes_dev);
- /* Set endianness recommended by data-sheet. */
- aes_a_set_endianness(aes_dev);
- /* Set AES_COMMAND register. */
- set_ocs_aes_command(aes_dev, cipher, mode, instruction);
- }
- /*
- * Write the byte length of the last AES/SM4 block of Payload data (without
- * zero padding and without the length of the MAC) in register AES_PLEN.
- */
- static inline void ocs_aes_write_last_data_blk_len(struct ocs_aes_dev *aes_dev,
- u32 size)
- {
- u32 val;
- if (size == 0) {
- val = 0;
- goto exit;
- }
- val = size % AES_BLOCK_SIZE;
- if (val == 0)
- val = AES_BLOCK_SIZE;
- exit:
- iowrite32(val, aes_dev->base_reg + AES_PLEN_OFFSET);
- }
- /*
- * Validate inputs according to mode.
- * If OK return 0; else return -EINVAL.
- */
- static int ocs_aes_validate_inputs(dma_addr_t src_dma_list, u32 src_size,
- const u8 *iv, u32 iv_size,
- dma_addr_t aad_dma_list, u32 aad_size,
- const u8 *tag, u32 tag_size,
- enum ocs_cipher cipher, enum ocs_mode mode,
- enum ocs_instruction instruction,
- dma_addr_t dst_dma_list)
- {
- /* Ensure cipher, mode and instruction are valid. */
- if (!(cipher == OCS_AES || cipher == OCS_SM4))
- return -EINVAL;
- if (mode != OCS_MODE_ECB && mode != OCS_MODE_CBC &&
- mode != OCS_MODE_CTR && mode != OCS_MODE_CCM &&
- mode != OCS_MODE_GCM && mode != OCS_MODE_CTS)
- return -EINVAL;
- if (instruction != OCS_ENCRYPT && instruction != OCS_DECRYPT &&
- instruction != OCS_EXPAND && instruction != OCS_BYPASS)
- return -EINVAL;
- /*
- * When instruction is OCS_BYPASS, OCS simply copies data from source
- * to destination using DMA.
- *
- * AES mode is irrelevant, but both source and destination DMA
- * linked-list must be defined.
- */
- if (instruction == OCS_BYPASS) {
- if (src_dma_list == DMA_MAPPING_ERROR ||
- dst_dma_list == DMA_MAPPING_ERROR)
- return -EINVAL;
- return 0;
- }
- /*
- * For performance reasons switch based on mode to limit unnecessary
- * conditionals for each mode
- */
- switch (mode) {
- case OCS_MODE_ECB:
- /* Ensure input length is multiple of block size */
- if (src_size % AES_BLOCK_SIZE != 0)
- return -EINVAL;
- /* Ensure source and destination linked lists are created */
- if (src_dma_list == DMA_MAPPING_ERROR ||
- dst_dma_list == DMA_MAPPING_ERROR)
- return -EINVAL;
- return 0;
- case OCS_MODE_CBC:
- /* Ensure input length is multiple of block size */
- if (src_size % AES_BLOCK_SIZE != 0)
- return -EINVAL;
- /* Ensure source and destination linked lists are created */
- if (src_dma_list == DMA_MAPPING_ERROR ||
- dst_dma_list == DMA_MAPPING_ERROR)
- return -EINVAL;
- /* Ensure IV is present and block size in length */
- if (!iv || iv_size != AES_BLOCK_SIZE)
- return -EINVAL;
- return 0;
- case OCS_MODE_CTR:
- /* Ensure input length of 1 byte or greater */
- if (src_size == 0)
- return -EINVAL;
- /* Ensure source and destination linked lists are created */
- if (src_dma_list == DMA_MAPPING_ERROR ||
- dst_dma_list == DMA_MAPPING_ERROR)
- return -EINVAL;
- /* Ensure IV is present and block size in length */
- if (!iv || iv_size != AES_BLOCK_SIZE)
- return -EINVAL;
- return 0;
- case OCS_MODE_CTS:
- /* Ensure input length >= block size */
- if (src_size < AES_BLOCK_SIZE)
- return -EINVAL;
- /* Ensure source and destination linked lists are created */
- if (src_dma_list == DMA_MAPPING_ERROR ||
- dst_dma_list == DMA_MAPPING_ERROR)
- return -EINVAL;
- /* Ensure IV is present and block size in length */
- if (!iv || iv_size != AES_BLOCK_SIZE)
- return -EINVAL;
- return 0;
- case OCS_MODE_GCM:
- /* Ensure IV is present and GCM_AES_IV_SIZE in length */
- if (!iv || iv_size != GCM_AES_IV_SIZE)
- return -EINVAL;
- /*
- * If input data present ensure source and destination linked
- * lists are created
- */
- if (src_size && (src_dma_list == DMA_MAPPING_ERROR ||
- dst_dma_list == DMA_MAPPING_ERROR))
- return -EINVAL;
- /* If aad present ensure aad linked list is created */
- if (aad_size && aad_dma_list == DMA_MAPPING_ERROR)
- return -EINVAL;
- /* Ensure tag destination is set */
- if (!tag)
- return -EINVAL;
- /* Just ensure that tag_size doesn't cause overflows. */
- if (tag_size > (AES_MAX_TAG_SIZE_U32 * sizeof(u32)))
- return -EINVAL;
- return 0;
- case OCS_MODE_CCM:
- /* Ensure IV is present and block size in length */
- if (!iv || iv_size != AES_BLOCK_SIZE)
- return -EINVAL;
- /* 2 <= L <= 8, so 1 <= L' <= 7 */
- if (iv[L_PRIME_IDX] < L_PRIME_MIN ||
- iv[L_PRIME_IDX] > L_PRIME_MAX)
- return -EINVAL;
- /* If aad present ensure aad linked list is created */
- if (aad_size && aad_dma_list == DMA_MAPPING_ERROR)
- return -EINVAL;
- /* Just ensure that tag_size doesn't cause overflows. */
- if (tag_size > (AES_MAX_TAG_SIZE_U32 * sizeof(u32)))
- return -EINVAL;
- if (instruction == OCS_DECRYPT) {
- /*
- * If input data present ensure source and destination
- * linked lists are created
- */
- if (src_size && (src_dma_list == DMA_MAPPING_ERROR ||
- dst_dma_list == DMA_MAPPING_ERROR))
- return -EINVAL;
- /* Ensure input tag is present */
- if (!tag)
- return -EINVAL;
- return 0;
- }
- /* Instruction == OCS_ENCRYPT */
- /*
- * Destination linked list always required (for tag even if no
- * input data)
- */
- if (dst_dma_list == DMA_MAPPING_ERROR)
- return -EINVAL;
- /* If input data present ensure src linked list is created */
- if (src_size && src_dma_list == DMA_MAPPING_ERROR)
- return -EINVAL;
- return 0;
- default:
- return -EINVAL;
- }
- }
- /**
- * ocs_aes_op() - Perform AES/SM4 operation.
- * @aes_dev: The OCS AES device to use.
- * @mode: The mode to use (ECB, CBC, CTR, or CTS).
- * @cipher: The cipher to use (AES or SM4).
- * @instruction: The instruction to perform (encrypt or decrypt).
- * @dst_dma_list: The OCS DMA list mapping output memory.
- * @src_dma_list: The OCS DMA list mapping input payload data.
- * @src_size: The amount of data mapped by @src_dma_list.
- * @iv: The IV vector.
- * @iv_size: The size (in bytes) of @iv.
- *
- * Return: 0 on success, negative error code otherwise.
- */
- int ocs_aes_op(struct ocs_aes_dev *aes_dev,
- enum ocs_mode mode,
- enum ocs_cipher cipher,
- enum ocs_instruction instruction,
- dma_addr_t dst_dma_list,
- dma_addr_t src_dma_list,
- u32 src_size,
- u8 *iv,
- u32 iv_size)
- {
- u32 *iv32;
- int rc;
- rc = ocs_aes_validate_inputs(src_dma_list, src_size, iv, iv_size, 0, 0,
- NULL, 0, cipher, mode, instruction,
- dst_dma_list);
- if (rc)
- return rc;
- /*
- * ocs_aes_validate_inputs() is a generic check, now ensure mode is not
- * GCM or CCM.
- */
- if (mode == OCS_MODE_GCM || mode == OCS_MODE_CCM)
- return -EINVAL;
- /* Cast IV to u32 array. */
- iv32 = (u32 *)iv;
- ocs_aes_init(aes_dev, mode, cipher, instruction);
- if (mode == OCS_MODE_CTS) {
- /* Write the byte length of the last data block to engine. */
- ocs_aes_write_last_data_blk_len(aes_dev, src_size);
- }
- /* ECB is the only mode that doesn't use IV. */
- if (mode != OCS_MODE_ECB) {
- iowrite32(iv32[0], aes_dev->base_reg + AES_IV_0_OFFSET);
- iowrite32(iv32[1], aes_dev->base_reg + AES_IV_1_OFFSET);
- iowrite32(iv32[2], aes_dev->base_reg + AES_IV_2_OFFSET);
- iowrite32(iv32[3], aes_dev->base_reg + AES_IV_3_OFFSET);
- }
- /* Set AES_ACTIVE.TRIGGER to start the operation. */
- aes_a_op_trigger(aes_dev);
- /* Configure and activate input / output DMA. */
- dma_to_ocs_aes_ll(aes_dev, src_dma_list);
- dma_from_ocs_aes_ll(aes_dev, dst_dma_list);
- aes_a_dma_active_src_dst_ll_en(aes_dev);
- if (mode == OCS_MODE_CTS) {
- /*
- * For CTS mode, instruct engine to activate ciphertext
- * stealing if last block of data is incomplete.
- */
- aes_a_set_last_gcx(aes_dev);
- } else {
- /* For all other modes, just write the 'termination' bit. */
- aes_a_op_termination(aes_dev);
- }
- /* Wait for engine to complete processing. */
- rc = ocs_aes_irq_enable_and_wait(aes_dev, AES_COMPLETE_INT);
- if (rc)
- return rc;
- if (mode == OCS_MODE_CTR) {
- /* Read back IV for streaming mode */
- iv32[0] = ioread32(aes_dev->base_reg + AES_IV_0_OFFSET);
- iv32[1] = ioread32(aes_dev->base_reg + AES_IV_1_OFFSET);
- iv32[2] = ioread32(aes_dev->base_reg + AES_IV_2_OFFSET);
- iv32[3] = ioread32(aes_dev->base_reg + AES_IV_3_OFFSET);
- }
- return 0;
- }
- /* Compute and write J0 to engine registers. */
- static void ocs_aes_gcm_write_j0(const struct ocs_aes_dev *aes_dev,
- const u8 *iv)
- {
- const u32 *j0 = (u32 *)iv;
- /*
- * IV must be 12 bytes; Other sizes not supported as Linux crypto API
- * does only expects/allows 12 byte IV for GCM
- */
- iowrite32(0x00000001, aes_dev->base_reg + AES_IV_0_OFFSET);
- iowrite32(__swab32(j0[2]), aes_dev->base_reg + AES_IV_1_OFFSET);
- iowrite32(__swab32(j0[1]), aes_dev->base_reg + AES_IV_2_OFFSET);
- iowrite32(__swab32(j0[0]), aes_dev->base_reg + AES_IV_3_OFFSET);
- }
- /* Read GCM tag from engine registers. */
- static inline void ocs_aes_gcm_read_tag(struct ocs_aes_dev *aes_dev,
- u8 *tag, u32 tag_size)
- {
- u32 tag_u32[AES_MAX_TAG_SIZE_U32];
- /*
- * The Authentication Tag T is stored in Little Endian order in the
- * registers with the most significant bytes stored from AES_T_MAC[3]
- * downward.
- */
- tag_u32[0] = __swab32(ioread32(aes_dev->base_reg + AES_T_MAC_3_OFFSET));
- tag_u32[1] = __swab32(ioread32(aes_dev->base_reg + AES_T_MAC_2_OFFSET));
- tag_u32[2] = __swab32(ioread32(aes_dev->base_reg + AES_T_MAC_1_OFFSET));
- tag_u32[3] = __swab32(ioread32(aes_dev->base_reg + AES_T_MAC_0_OFFSET));
- memcpy(tag, tag_u32, tag_size);
- }
- /**
- * ocs_aes_gcm_op() - Perform GCM operation.
- * @aes_dev: The OCS AES device to use.
- * @cipher: The Cipher to use (AES or SM4).
- * @instruction: The instruction to perform (encrypt or decrypt).
- * @dst_dma_list: The OCS DMA list mapping output memory.
- * @src_dma_list: The OCS DMA list mapping input payload data.
- * @src_size: The amount of data mapped by @src_dma_list.
- * @iv: The input IV vector.
- * @aad_dma_list: The OCS DMA list mapping input AAD data.
- * @aad_size: The amount of data mapped by @aad_dma_list.
- * @out_tag: Where to store computed tag.
- * @tag_size: The size (in bytes) of @out_tag.
- *
- * Return: 0 on success, negative error code otherwise.
- */
- int ocs_aes_gcm_op(struct ocs_aes_dev *aes_dev,
- enum ocs_cipher cipher,
- enum ocs_instruction instruction,
- dma_addr_t dst_dma_list,
- dma_addr_t src_dma_list,
- u32 src_size,
- const u8 *iv,
- dma_addr_t aad_dma_list,
- u32 aad_size,
- u8 *out_tag,
- u32 tag_size)
- {
- u64 bit_len;
- u32 val;
- int rc;
- rc = ocs_aes_validate_inputs(src_dma_list, src_size, iv,
- GCM_AES_IV_SIZE, aad_dma_list,
- aad_size, out_tag, tag_size, cipher,
- OCS_MODE_GCM, instruction,
- dst_dma_list);
- if (rc)
- return rc;
- ocs_aes_init(aes_dev, OCS_MODE_GCM, cipher, instruction);
- /* Compute and write J0 to OCS HW. */
- ocs_aes_gcm_write_j0(aes_dev, iv);
- /* Write out_tag byte length */
- iowrite32(tag_size, aes_dev->base_reg + AES_TLEN_OFFSET);
- /* Write the byte length of the last plaintext / ciphertext block. */
- ocs_aes_write_last_data_blk_len(aes_dev, src_size);
- /* Write ciphertext bit length */
- bit_len = (u64)src_size * 8;
- val = bit_len & 0xFFFFFFFF;
- iowrite32(val, aes_dev->base_reg + AES_MULTIPURPOSE2_0_OFFSET);
- val = bit_len >> 32;
- iowrite32(val, aes_dev->base_reg + AES_MULTIPURPOSE2_1_OFFSET);
- /* Write aad bit length */
- bit_len = (u64)aad_size * 8;
- val = bit_len & 0xFFFFFFFF;
- iowrite32(val, aes_dev->base_reg + AES_MULTIPURPOSE2_2_OFFSET);
- val = bit_len >> 32;
- iowrite32(val, aes_dev->base_reg + AES_MULTIPURPOSE2_3_OFFSET);
- /* Set AES_ACTIVE.TRIGGER to start the operation. */
- aes_a_op_trigger(aes_dev);
- /* Process AAD. */
- if (aad_size) {
- /* If aad present, configure DMA to feed it to the engine. */
- dma_to_ocs_aes_ll(aes_dev, aad_dma_list);
- aes_a_dma_active_src_ll_en(aes_dev);
- /* Instructs engine to pad last block of aad, if needed. */
- aes_a_set_last_gcx_and_adata(aes_dev);
- /* Wait for DMA transfer to complete. */
- rc = ocs_aes_irq_enable_and_wait(aes_dev, AES_DMA_SRC_DONE_INT);
- if (rc)
- return rc;
- } else {
- aes_a_set_last_gcx_and_adata(aes_dev);
- }
- /* Wait until adata (if present) has been processed. */
- aes_a_wait_last_gcx(aes_dev);
- aes_a_dma_wait_input_buffer_occupancy(aes_dev);
- /* Now process payload. */
- if (src_size) {
- /* Configure and activate DMA for both input and output data. */
- dma_to_ocs_aes_ll(aes_dev, src_dma_list);
- dma_from_ocs_aes_ll(aes_dev, dst_dma_list);
- aes_a_dma_active_src_dst_ll_en(aes_dev);
- } else {
- aes_a_dma_set_xfer_size_zero(aes_dev);
- aes_a_dma_active(aes_dev);
- }
- /* Instruct AES/SMA4 engine payload processing is over. */
- aes_a_set_last_gcx(aes_dev);
- /* Wait for OCS AES engine to complete processing. */
- rc = ocs_aes_irq_enable_and_wait(aes_dev, AES_COMPLETE_INT);
- if (rc)
- return rc;
- ocs_aes_gcm_read_tag(aes_dev, out_tag, tag_size);
- return 0;
- }
- /* Write encrypted tag to AES/SM4 engine. */
- static void ocs_aes_ccm_write_encrypted_tag(struct ocs_aes_dev *aes_dev,
- const u8 *in_tag, u32 tag_size)
- {
- int i;
- /* Ensure DMA input buffer is empty */
- aes_a_dma_wait_input_buffer_occupancy(aes_dev);
- /*
- * During CCM decrypt, the OCS block needs to finish processing the
- * ciphertext before the tag is written. So delay needed after DMA has
- * completed writing the ciphertext
- */
- aes_a_dma_reset_and_activate_perf_cntr(aes_dev);
- aes_a_dma_wait_and_deactivate_perf_cntr(aes_dev,
- CCM_DECRYPT_DELAY_TAG_CLK_COUNT);
- /* Write encrypted tag to AES/SM4 engine. */
- for (i = 0; i < tag_size; i++) {
- iowrite8(in_tag[i], aes_dev->base_reg +
- AES_A_DMA_INBUFFER_WRITE_FIFO_OFFSET);
- }
- }
- /*
- * Write B0 CCM block to OCS AES HW.
- *
- * Note: B0 format is documented in NIST Special Publication 800-38C
- * https://nvlpubs.nist.gov/nistpubs/Legacy/SP/nistspecialpublication800-38c.pdf
- * (see Section A.2.1)
- */
- static int ocs_aes_ccm_write_b0(const struct ocs_aes_dev *aes_dev,
- const u8 *iv, u32 adata_size, u32 tag_size,
- u32 cryptlen)
- {
- u8 b0[16]; /* CCM B0 block is 16 bytes long. */
- int i, q;
- /* Initialize B0 to 0. */
- memset(b0, 0, sizeof(b0));
- /*
- * B0[0] is the 'Flags Octet' and has the following structure:
- * bit 7: Reserved
- * bit 6: Adata flag
- * bit 5-3: t value encoded as (t-2)/2
- * bit 2-0: q value encoded as q - 1
- */
- /* If there is AAD data, set the Adata flag. */
- if (adata_size)
- b0[0] |= BIT(6);
- /*
- * t denotes the octet length of T.
- * t can only be an element of { 4, 6, 8, 10, 12, 14, 16} and is
- * encoded as (t - 2) / 2
- */
- b0[0] |= (((tag_size - 2) / 2) & 0x7) << 3;
- /*
- * q is the octet length of Q.
- * q can only be an element of {2, 3, 4, 5, 6, 7, 8} and is encoded as
- * q - 1 == iv[0] & 0x7;
- */
- b0[0] |= iv[0] & 0x7;
- /*
- * Copy the Nonce N from IV to B0; N is located in iv[1]..iv[15 - q]
- * and must be copied to b0[1]..b0[15-q].
- * q == (iv[0] & 0x7) + 1
- */
- q = (iv[0] & 0x7) + 1;
- for (i = 1; i <= 15 - q; i++)
- b0[i] = iv[i];
- /*
- * The rest of B0 must contain Q, i.e., the message length.
- * Q is encoded in q octets, in big-endian order, so to write it, we
- * start from the end of B0 and we move backward.
- */
- i = sizeof(b0) - 1;
- while (q) {
- b0[i] = cryptlen & 0xff;
- cryptlen >>= 8;
- i--;
- q--;
- }
- /*
- * If cryptlen is not zero at this point, it means that its original
- * value was too big.
- */
- if (cryptlen)
- return -EOVERFLOW;
- /* Now write B0 to OCS AES input buffer. */
- for (i = 0; i < sizeof(b0); i++)
- iowrite8(b0[i], aes_dev->base_reg +
- AES_A_DMA_INBUFFER_WRITE_FIFO_OFFSET);
- return 0;
- }
- /*
- * Write adata length to OCS AES HW.
- *
- * Note: adata len encoding is documented in NIST Special Publication 800-38C
- * https://nvlpubs.nist.gov/nistpubs/Legacy/SP/nistspecialpublication800-38c.pdf
- * (see Section A.2.2)
- */
- static void ocs_aes_ccm_write_adata_len(const struct ocs_aes_dev *aes_dev,
- u64 adata_len)
- {
- u8 enc_a[10]; /* Maximum encoded size: 10 octets. */
- int i, len;
- /*
- * adata_len ('a') is encoded as follows:
- * If 0 < a < 2^16 - 2^8 ==> 'a' encoded as [a]16, i.e., two octets
- * (big endian).
- * If 2^16 - 2^8 ≤ a < 2^32 ==> 'a' encoded as 0xff || 0xfe || [a]32,
- * i.e., six octets (big endian).
- * If 2^32 ≤ a < 2^64 ==> 'a' encoded as 0xff || 0xff || [a]64,
- * i.e., ten octets (big endian).
- */
- if (adata_len < 65280) {
- len = 2;
- *(__be16 *)enc_a = cpu_to_be16(adata_len);
- } else if (adata_len <= 0xFFFFFFFF) {
- len = 6;
- *(__be16 *)enc_a = cpu_to_be16(0xfffe);
- *(__be32 *)&enc_a[2] = cpu_to_be32(adata_len);
- } else { /* adata_len >= 2^32 */
- len = 10;
- *(__be16 *)enc_a = cpu_to_be16(0xffff);
- *(__be64 *)&enc_a[2] = cpu_to_be64(adata_len);
- }
- for (i = 0; i < len; i++)
- iowrite8(enc_a[i],
- aes_dev->base_reg +
- AES_A_DMA_INBUFFER_WRITE_FIFO_OFFSET);
- }
- static int ocs_aes_ccm_do_adata(struct ocs_aes_dev *aes_dev,
- dma_addr_t adata_dma_list, u32 adata_size)
- {
- int rc;
- if (!adata_size) {
- /* Since no aad the LAST_GCX bit can be set now */
- aes_a_set_last_gcx_and_adata(aes_dev);
- goto exit;
- }
- /* Adata case. */
- /*
- * Form the encoding of the Associated data length and write it
- * to the AES/SM4 input buffer.
- */
- ocs_aes_ccm_write_adata_len(aes_dev, adata_size);
- /* Configure the AES/SM4 DMA to fetch the Associated Data */
- dma_to_ocs_aes_ll(aes_dev, adata_dma_list);
- /* Activate DMA to fetch Associated data. */
- aes_a_dma_active_src_ll_en(aes_dev);
- /* Set LAST_GCX and LAST_ADATA in AES ACTIVE register. */
- aes_a_set_last_gcx_and_adata(aes_dev);
- /* Wait for DMA transfer to complete. */
- rc = ocs_aes_irq_enable_and_wait(aes_dev, AES_DMA_SRC_DONE_INT);
- if (rc)
- return rc;
- exit:
- /* Wait until adata (if present) has been processed. */
- aes_a_wait_last_gcx(aes_dev);
- aes_a_dma_wait_input_buffer_occupancy(aes_dev);
- return 0;
- }
- static int ocs_aes_ccm_encrypt_do_payload(struct ocs_aes_dev *aes_dev,
- dma_addr_t dst_dma_list,
- dma_addr_t src_dma_list,
- u32 src_size)
- {
- if (src_size) {
- /*
- * Configure and activate DMA for both input and output
- * data.
- */
- dma_to_ocs_aes_ll(aes_dev, src_dma_list);
- dma_from_ocs_aes_ll(aes_dev, dst_dma_list);
- aes_a_dma_active_src_dst_ll_en(aes_dev);
- } else {
- /* Configure and activate DMA for output data only. */
- dma_from_ocs_aes_ll(aes_dev, dst_dma_list);
- aes_a_dma_active_dst_ll_en(aes_dev);
- }
- /*
- * Set the LAST GCX bit in AES_ACTIVE Register to instruct
- * AES/SM4 engine to pad the last block of data.
- */
- aes_a_set_last_gcx(aes_dev);
- /* We are done, wait for IRQ and return. */
- return ocs_aes_irq_enable_and_wait(aes_dev, AES_COMPLETE_INT);
- }
- static int ocs_aes_ccm_decrypt_do_payload(struct ocs_aes_dev *aes_dev,
- dma_addr_t dst_dma_list,
- dma_addr_t src_dma_list,
- u32 src_size)
- {
- if (!src_size) {
- /* Let engine process 0-length input. */
- aes_a_dma_set_xfer_size_zero(aes_dev);
- aes_a_dma_active(aes_dev);
- aes_a_set_last_gcx(aes_dev);
- return 0;
- }
- /*
- * Configure and activate DMA for both input and output
- * data.
- */
- dma_to_ocs_aes_ll(aes_dev, src_dma_list);
- dma_from_ocs_aes_ll(aes_dev, dst_dma_list);
- aes_a_dma_active_src_dst_ll_en(aes_dev);
- /*
- * Set the LAST GCX bit in AES_ACTIVE Register; this allows the
- * AES/SM4 engine to differentiate between encrypted data and
- * encrypted MAC.
- */
- aes_a_set_last_gcx(aes_dev);
- /*
- * Enable DMA DONE interrupt; once DMA transfer is over,
- * interrupt handler will process the MAC/tag.
- */
- return ocs_aes_irq_enable_and_wait(aes_dev, AES_DMA_SRC_DONE_INT);
- }
- /*
- * Compare Tag to Yr.
- *
- * Only used at the end of CCM decrypt. If tag == yr, message authentication
- * has succeeded.
- */
- static inline int ccm_compare_tag_to_yr(struct ocs_aes_dev *aes_dev,
- u8 tag_size_bytes)
- {
- u32 tag[AES_MAX_TAG_SIZE_U32];
- u32 yr[AES_MAX_TAG_SIZE_U32];
- u8 i;
- /* Read Tag and Yr from AES registers. */
- for (i = 0; i < AES_MAX_TAG_SIZE_U32; i++) {
- tag[i] = ioread32(aes_dev->base_reg +
- AES_T_MAC_0_OFFSET + (i * sizeof(u32)));
- yr[i] = ioread32(aes_dev->base_reg +
- AES_MULTIPURPOSE2_0_OFFSET +
- (i * sizeof(u32)));
- }
- return memcmp(tag, yr, tag_size_bytes) ? -EBADMSG : 0;
- }
- /**
- * ocs_aes_ccm_op() - Perform CCM operation.
- * @aes_dev: The OCS AES device to use.
- * @cipher: The Cipher to use (AES or SM4).
- * @instruction: The instruction to perform (encrypt or decrypt).
- * @dst_dma_list: The OCS DMA list mapping output memory.
- * @src_dma_list: The OCS DMA list mapping input payload data.
- * @src_size: The amount of data mapped by @src_dma_list.
- * @iv: The input IV vector.
- * @adata_dma_list: The OCS DMA list mapping input A-data.
- * @adata_size: The amount of data mapped by @adata_dma_list.
- * @in_tag: Input tag.
- * @tag_size: The size (in bytes) of @in_tag.
- *
- * Note: for encrypt the tag is appended to the ciphertext (in the memory
- * mapped by @dst_dma_list).
- *
- * Return: 0 on success, negative error code otherwise.
- */
- int ocs_aes_ccm_op(struct ocs_aes_dev *aes_dev,
- enum ocs_cipher cipher,
- enum ocs_instruction instruction,
- dma_addr_t dst_dma_list,
- dma_addr_t src_dma_list,
- u32 src_size,
- u8 *iv,
- dma_addr_t adata_dma_list,
- u32 adata_size,
- u8 *in_tag,
- u32 tag_size)
- {
- u32 *iv_32;
- u8 lprime;
- int rc;
- rc = ocs_aes_validate_inputs(src_dma_list, src_size, iv,
- AES_BLOCK_SIZE, adata_dma_list, adata_size,
- in_tag, tag_size, cipher, OCS_MODE_CCM,
- instruction, dst_dma_list);
- if (rc)
- return rc;
- ocs_aes_init(aes_dev, OCS_MODE_CCM, cipher, instruction);
- /*
- * Note: rfc 3610 and NIST 800-38C require counter of zero to encrypt
- * auth tag so ensure this is the case
- */
- lprime = iv[L_PRIME_IDX];
- memset(&iv[COUNTER_START(lprime)], 0, COUNTER_LEN(lprime));
- /*
- * Nonce is already converted to ctr0 before being passed into this
- * function as iv.
- */
- iv_32 = (u32 *)iv;
- iowrite32(__swab32(iv_32[0]),
- aes_dev->base_reg + AES_MULTIPURPOSE1_3_OFFSET);
- iowrite32(__swab32(iv_32[1]),
- aes_dev->base_reg + AES_MULTIPURPOSE1_2_OFFSET);
- iowrite32(__swab32(iv_32[2]),
- aes_dev->base_reg + AES_MULTIPURPOSE1_1_OFFSET);
- iowrite32(__swab32(iv_32[3]),
- aes_dev->base_reg + AES_MULTIPURPOSE1_0_OFFSET);
- /* Write MAC/tag length in register AES_TLEN */
- iowrite32(tag_size, aes_dev->base_reg + AES_TLEN_OFFSET);
- /*
- * Write the byte length of the last AES/SM4 block of Payload data
- * (without zero padding and without the length of the MAC) in register
- * AES_PLEN.
- */
- ocs_aes_write_last_data_blk_len(aes_dev, src_size);
- /* Set AES_ACTIVE.TRIGGER to start the operation. */
- aes_a_op_trigger(aes_dev);
- aes_a_dma_reset_and_activate_perf_cntr(aes_dev);
- /* Form block B0 and write it to the AES/SM4 input buffer. */
- rc = ocs_aes_ccm_write_b0(aes_dev, iv, adata_size, tag_size, src_size);
- if (rc)
- return rc;
- /*
- * Ensure there has been at least CCM_DECRYPT_DELAY_LAST_GCX_CLK_COUNT
- * clock cycles since TRIGGER bit was set
- */
- aes_a_dma_wait_and_deactivate_perf_cntr(aes_dev,
- CCM_DECRYPT_DELAY_LAST_GCX_CLK_COUNT);
- /* Process Adata. */
- ocs_aes_ccm_do_adata(aes_dev, adata_dma_list, adata_size);
- /* For Encrypt case we just process the payload and return. */
- if (instruction == OCS_ENCRYPT) {
- return ocs_aes_ccm_encrypt_do_payload(aes_dev, dst_dma_list,
- src_dma_list, src_size);
- }
- /* For Decypt we need to process the payload and then the tag. */
- rc = ocs_aes_ccm_decrypt_do_payload(aes_dev, dst_dma_list,
- src_dma_list, src_size);
- if (rc)
- return rc;
- /* Process MAC/tag directly: feed tag to engine and wait for IRQ. */
- ocs_aes_ccm_write_encrypted_tag(aes_dev, in_tag, tag_size);
- rc = ocs_aes_irq_enable_and_wait(aes_dev, AES_COMPLETE_INT);
- if (rc)
- return rc;
- return ccm_compare_tag_to_yr(aes_dev, tag_size);
- }
- /**
- * ocs_create_linked_list_from_sg() - Create OCS DMA linked list from SG list.
- * @aes_dev: The OCS AES device the list will be created for.
- * @sg: The SG list OCS DMA linked list will be created from. When
- * passed to this function, @sg must have been already mapped
- * with dma_map_sg().
- * @sg_dma_count: The number of DMA-mapped entries in @sg. This must be the
- * value returned by dma_map_sg() when @sg was mapped.
- * @dll_desc: The OCS DMA dma_list to use to store information about the
- * created linked list.
- * @data_size: The size of the data (from the SG list) to be mapped into the
- * OCS DMA linked list.
- * @data_offset: The offset (within the SG list) of the data to be mapped.
- *
- * Return: 0 on success, negative error code otherwise.
- */
- int ocs_create_linked_list_from_sg(const struct ocs_aes_dev *aes_dev,
- struct scatterlist *sg,
- int sg_dma_count,
- struct ocs_dll_desc *dll_desc,
- size_t data_size, size_t data_offset)
- {
- struct ocs_dma_linked_list *ll = NULL;
- struct scatterlist *sg_tmp;
- unsigned int tmp;
- int dma_nents;
- int i;
- if (!dll_desc || !sg || !aes_dev)
- return -EINVAL;
- /* Default values for when no ddl_desc is created. */
- dll_desc->vaddr = NULL;
- dll_desc->dma_addr = DMA_MAPPING_ERROR;
- dll_desc->size = 0;
- if (data_size == 0)
- return 0;
- /* Loop over sg_list until we reach entry at specified offset. */
- while (data_offset >= sg_dma_len(sg)) {
- data_offset -= sg_dma_len(sg);
- sg_dma_count--;
- sg = sg_next(sg);
- /* If we reach the end of the list, offset was invalid. */
- if (!sg || sg_dma_count == 0)
- return -EINVAL;
- }
- /* Compute number of DMA-mapped SG entries to add into OCS DMA list. */
- dma_nents = 0;
- tmp = 0;
- sg_tmp = sg;
- while (tmp < data_offset + data_size) {
- /* If we reach the end of the list, data_size was invalid. */
- if (!sg_tmp)
- return -EINVAL;
- tmp += sg_dma_len(sg_tmp);
- dma_nents++;
- sg_tmp = sg_next(sg_tmp);
- }
- if (dma_nents > sg_dma_count)
- return -EINVAL;
- /* Allocate the DMA list, one entry for each SG entry. */
- dll_desc->size = sizeof(struct ocs_dma_linked_list) * dma_nents;
- dll_desc->vaddr = dma_alloc_coherent(aes_dev->dev, dll_desc->size,
- &dll_desc->dma_addr, GFP_KERNEL);
- if (!dll_desc->vaddr)
- return -ENOMEM;
- /* Populate DMA linked list entries. */
- ll = dll_desc->vaddr;
- for (i = 0; i < dma_nents; i++, sg = sg_next(sg)) {
- ll[i].src_addr = sg_dma_address(sg) + data_offset;
- ll[i].src_len = (sg_dma_len(sg) - data_offset) < data_size ?
- (sg_dma_len(sg) - data_offset) : data_size;
- data_offset = 0;
- data_size -= ll[i].src_len;
- /* Current element points to the DMA address of the next one. */
- ll[i].next = dll_desc->dma_addr + (sizeof(*ll) * (i + 1));
- ll[i].ll_flags = 0;
- }
- /* Terminate last element. */
- ll[i - 1].next = 0;
- ll[i - 1].ll_flags = OCS_LL_DMA_FLAG_TERMINATE;
- return 0;
- }
|