1 // SPDX-License-Identifier: GPL-2.0 2 /* 3 * Cryptographic API. 4 * 5 * Support for StarFive hardware cryptographic engine. 6 * Copyright (c) 2022 StarFive Technology 7 * 8 */ 9 10 #include <crypto/engine.h> 11 #include "jh7110-cryp.h" 12 #include <linux/clk.h> 13 #include <linux/completion.h> 14 #include <linux/err.h> 15 #include <linux/interrupt.h> 16 #include <linux/iopoll.h> 17 #include <linux/kernel.h> 18 #include <linux/mod_devicetable.h> 19 #include <linux/module.h> 20 #include <linux/platform_device.h> 21 #include <linux/pm_runtime.h> 22 #include <linux/reset.h> 23 #include <linux/spinlock.h> 24 25 #define DRIVER_NAME "jh7110-crypto" 26 27 struct starfive_dev_list { 28 struct list_head dev_list; 29 spinlock_t lock; /* protect dev_list */ 30 }; 31 32 static struct starfive_dev_list dev_list = { 33 .dev_list = LIST_HEAD_INIT(dev_list.dev_list), 34 .lock = __SPIN_LOCK_UNLOCKED(dev_list.lock), 35 }; 36 37 struct starfive_cryp_dev *starfive_cryp_find_dev(struct starfive_cryp_ctx *ctx) 38 { 39 struct starfive_cryp_dev *cryp = NULL, *tmp; 40 41 spin_lock_bh(&dev_list.lock); 42 if (!ctx->cryp) { 43 list_for_each_entry(tmp, &dev_list.dev_list, list) { 44 cryp = tmp; 45 break; 46 } 47 ctx->cryp = cryp; 48 } else { 49 cryp = ctx->cryp; 50 } 51 52 spin_unlock_bh(&dev_list.lock); 53 54 return cryp; 55 } 56 57 static u16 side_chan; 58 module_param(side_chan, ushort, 0); 59 MODULE_PARM_DESC(side_chan, "Enable side channel mitigation for AES module.\n" 60 "Enabling this feature will reduce speed performance.\n" 61 " 0 - Disabled\n" 62 " other - Enabled"); 63 64 static int starfive_dma_init(struct starfive_cryp_dev *cryp) 65 { 66 dma_cap_mask_t mask; 67 68 dma_cap_zero(mask); 69 dma_cap_set(DMA_SLAVE, mask); 70 71 cryp->tx = dma_request_chan(cryp->dev, "tx"); 72 if (IS_ERR(cryp->tx)) 73 return dev_err_probe(cryp->dev, PTR_ERR(cryp->tx), 74 "Error requesting tx dma channel.\n"); 75 76 cryp->rx = dma_request_chan(cryp->dev, "rx"); 77 if (IS_ERR(cryp->rx)) { 78 dma_release_channel(cryp->tx); 79 return dev_err_probe(cryp->dev, PTR_ERR(cryp->rx), 80 "Error requesting rx dma channel.\n"); 81 } 82 83 return 0; 84 } 85 86 static void starfive_dma_cleanup(struct starfive_cryp_dev *cryp) 87 { 88 dma_release_channel(cryp->tx); 89 dma_release_channel(cryp->rx); 90 } 91 92 static irqreturn_t starfive_cryp_irq(int irq, void *priv) 93 { 94 u32 status; 95 u32 mask; 96 struct starfive_cryp_dev *cryp = (struct starfive_cryp_dev *)priv; 97 98 mask = readl(cryp->base + STARFIVE_IE_MASK_OFFSET); 99 status = readl(cryp->base + STARFIVE_IE_FLAG_OFFSET); 100 if (status & STARFIVE_IE_FLAG_AES_DONE) { 101 mask |= STARFIVE_IE_MASK_AES_DONE; 102 writel(mask, cryp->base + STARFIVE_IE_MASK_OFFSET); 103 tasklet_schedule(&cryp->aes_done); 104 } 105 106 if (status & STARFIVE_IE_FLAG_HASH_DONE) { 107 mask |= STARFIVE_IE_MASK_HASH_DONE; 108 writel(mask, cryp->base + STARFIVE_IE_MASK_OFFSET); 109 tasklet_schedule(&cryp->hash_done); 110 } 111 112 if (status & STARFIVE_IE_FLAG_PKA_DONE) { 113 mask |= STARFIVE_IE_MASK_PKA_DONE; 114 writel(mask, cryp->base + STARFIVE_IE_MASK_OFFSET); 115 complete(&cryp->pka_done); 116 } 117 118 return IRQ_HANDLED; 119 } 120 121 static int starfive_cryp_probe(struct platform_device *pdev) 122 { 123 struct starfive_cryp_dev *cryp; 124 struct resource *res; 125 int irq; 126 int ret; 127 128 cryp = devm_kzalloc(&pdev->dev, sizeof(*cryp), GFP_KERNEL); 129 if (!cryp) 130 return -ENOMEM; 131 132 platform_set_drvdata(pdev, cryp); 133 cryp->dev = &pdev->dev; 134 135 cryp->base = devm_platform_get_and_ioremap_resource(pdev, 0, &res); 136 if (IS_ERR(cryp->base)) 137 return dev_err_probe(&pdev->dev, PTR_ERR(cryp->base), 138 "Error remapping memory for platform device\n"); 139 140 tasklet_init(&cryp->aes_done, starfive_aes_done_task, (unsigned long)cryp); 141 tasklet_init(&cryp->hash_done, starfive_hash_done_task, (unsigned long)cryp); 142 143 cryp->phys_base = res->start; 144 cryp->dma_maxburst = 32; 145 cryp->side_chan = side_chan; 146 147 cryp->hclk = devm_clk_get(&pdev->dev, "hclk"); 148 if (IS_ERR(cryp->hclk)) 149 return dev_err_probe(&pdev->dev, PTR_ERR(cryp->hclk), 150 "Error getting hardware reference clock\n"); 151 152 cryp->ahb = devm_clk_get(&pdev->dev, "ahb"); 153 if (IS_ERR(cryp->ahb)) 154 return dev_err_probe(&pdev->dev, PTR_ERR(cryp->ahb), 155 "Error getting ahb reference clock\n"); 156 157 cryp->rst = devm_reset_control_get_shared(cryp->dev, NULL); 158 if (IS_ERR(cryp->rst)) 159 return dev_err_probe(&pdev->dev, PTR_ERR(cryp->rst), 160 "Error getting hardware reset line\n"); 161 162 init_completion(&cryp->pka_done); 163 164 irq = platform_get_irq(pdev, 0); 165 if (irq < 0) 166 return irq; 167 168 ret = devm_request_irq(&pdev->dev, irq, starfive_cryp_irq, 0, pdev->name, 169 (void *)cryp); 170 if (ret) 171 return dev_err_probe(&pdev->dev, irq, 172 "Failed to register interrupt handler\n"); 173 174 clk_prepare_enable(cryp->hclk); 175 clk_prepare_enable(cryp->ahb); 176 reset_control_deassert(cryp->rst); 177 178 spin_lock(&dev_list.lock); 179 list_add(&cryp->list, &dev_list.dev_list); 180 spin_unlock(&dev_list.lock); 181 182 ret = starfive_dma_init(cryp); 183 if (ret) { 184 if (ret == -EPROBE_DEFER) 185 goto err_probe_defer; 186 else 187 goto err_dma_init; 188 } 189 190 /* Initialize crypto engine */ 191 cryp->engine = crypto_engine_alloc_init(&pdev->dev, 1); 192 if (!cryp->engine) { 193 ret = -ENOMEM; 194 goto err_engine; 195 } 196 197 ret = crypto_engine_start(cryp->engine); 198 if (ret) 199 goto err_engine_start; 200 201 ret = starfive_aes_register_algs(); 202 if (ret) 203 goto err_algs_aes; 204 205 ret = starfive_hash_register_algs(); 206 if (ret) 207 goto err_algs_hash; 208 209 ret = starfive_rsa_register_algs(); 210 if (ret) 211 goto err_algs_rsa; 212 213 return 0; 214 215 err_algs_rsa: 216 starfive_hash_unregister_algs(); 217 err_algs_hash: 218 starfive_aes_unregister_algs(); 219 err_algs_aes: 220 crypto_engine_stop(cryp->engine); 221 err_engine_start: 222 crypto_engine_exit(cryp->engine); 223 err_engine: 224 starfive_dma_cleanup(cryp); 225 err_dma_init: 226 spin_lock(&dev_list.lock); 227 list_del(&cryp->list); 228 spin_unlock(&dev_list.lock); 229 230 clk_disable_unprepare(cryp->hclk); 231 clk_disable_unprepare(cryp->ahb); 232 reset_control_assert(cryp->rst); 233 234 tasklet_kill(&cryp->aes_done); 235 tasklet_kill(&cryp->hash_done); 236 err_probe_defer: 237 return ret; 238 } 239 240 static void starfive_cryp_remove(struct platform_device *pdev) 241 { 242 struct starfive_cryp_dev *cryp = platform_get_drvdata(pdev); 243 244 starfive_aes_unregister_algs(); 245 starfive_hash_unregister_algs(); 246 starfive_rsa_unregister_algs(); 247 248 tasklet_kill(&cryp->aes_done); 249 tasklet_kill(&cryp->hash_done); 250 251 crypto_engine_stop(cryp->engine); 252 crypto_engine_exit(cryp->engine); 253 254 starfive_dma_cleanup(cryp); 255 256 spin_lock(&dev_list.lock); 257 list_del(&cryp->list); 258 spin_unlock(&dev_list.lock); 259 260 clk_disable_unprepare(cryp->hclk); 261 clk_disable_unprepare(cryp->ahb); 262 reset_control_assert(cryp->rst); 263 } 264 265 static const struct of_device_id starfive_dt_ids[] __maybe_unused = { 266 { .compatible = "starfive,jh7110-crypto", .data = NULL}, 267 {}, 268 }; 269 MODULE_DEVICE_TABLE(of, starfive_dt_ids); 270 271 static struct platform_driver starfive_cryp_driver = { 272 .probe = starfive_cryp_probe, 273 .remove_new = starfive_cryp_remove, 274 .driver = { 275 .name = DRIVER_NAME, 276 .of_match_table = starfive_dt_ids, 277 }, 278 }; 279 280 module_platform_driver(starfive_cryp_driver); 281 282 MODULE_LICENSE("GPL"); 283 MODULE_DESCRIPTION("StarFive JH7110 Cryptographic Module"); 284