1*2afbf43aSAbel Vesa // SPDX-License-Identifier: GPL-2.0 2*2afbf43aSAbel Vesa /* 3*2afbf43aSAbel Vesa * Qualcomm ICE (Inline Crypto Engine) support. 4*2afbf43aSAbel Vesa * 5*2afbf43aSAbel Vesa * Copyright (c) 2013-2019, The Linux Foundation. All rights reserved. 6*2afbf43aSAbel Vesa * Copyright (c) 2019, Google LLC 7*2afbf43aSAbel Vesa * Copyright (c) 2023, Linaro Limited 8*2afbf43aSAbel Vesa */ 9*2afbf43aSAbel Vesa 10*2afbf43aSAbel Vesa #include <linux/bitfield.h> 11*2afbf43aSAbel Vesa #include <linux/clk.h> 12*2afbf43aSAbel Vesa #include <linux/delay.h> 13*2afbf43aSAbel Vesa #include <linux/iopoll.h> 14*2afbf43aSAbel Vesa #include <linux/of_platform.h> 15*2afbf43aSAbel Vesa 16*2afbf43aSAbel Vesa #include <linux/firmware/qcom/qcom_scm.h> 17*2afbf43aSAbel Vesa 18*2afbf43aSAbel Vesa #include <soc/qcom/ice.h> 19*2afbf43aSAbel Vesa 20*2afbf43aSAbel Vesa #define AES_256_XTS_KEY_SIZE 64 21*2afbf43aSAbel Vesa 22*2afbf43aSAbel Vesa /* QCOM ICE registers */ 23*2afbf43aSAbel Vesa #define QCOM_ICE_REG_VERSION 0x0008 24*2afbf43aSAbel Vesa #define QCOM_ICE_REG_FUSE_SETTING 0x0010 25*2afbf43aSAbel Vesa #define QCOM_ICE_REG_BIST_STATUS 0x0070 26*2afbf43aSAbel Vesa #define QCOM_ICE_REG_ADVANCED_CONTROL 0x1000 27*2afbf43aSAbel Vesa 28*2afbf43aSAbel Vesa /* BIST ("built-in self-test") status flags */ 29*2afbf43aSAbel Vesa #define QCOM_ICE_BIST_STATUS_MASK GENMASK(31, 28) 30*2afbf43aSAbel Vesa 31*2afbf43aSAbel Vesa #define QCOM_ICE_FUSE_SETTING_MASK 0x1 32*2afbf43aSAbel Vesa #define QCOM_ICE_FORCE_HW_KEY0_SETTING_MASK 0x2 33*2afbf43aSAbel Vesa #define QCOM_ICE_FORCE_HW_KEY1_SETTING_MASK 0x4 34*2afbf43aSAbel Vesa 35*2afbf43aSAbel Vesa #define qcom_ice_writel(engine, val, reg) \ 36*2afbf43aSAbel Vesa writel((val), (engine)->base + (reg)) 37*2afbf43aSAbel Vesa 38*2afbf43aSAbel Vesa #define qcom_ice_readl(engine, reg) \ 39*2afbf43aSAbel Vesa readl((engine)->base + (reg)) 40*2afbf43aSAbel Vesa 41*2afbf43aSAbel Vesa struct qcom_ice { 42*2afbf43aSAbel Vesa struct device *dev; 43*2afbf43aSAbel Vesa void __iomem *base; 44*2afbf43aSAbel Vesa struct device_link *link; 45*2afbf43aSAbel Vesa 46*2afbf43aSAbel Vesa struct clk *core_clk; 47*2afbf43aSAbel Vesa }; 48*2afbf43aSAbel Vesa 49*2afbf43aSAbel Vesa static bool qcom_ice_check_supported(struct qcom_ice *ice) 50*2afbf43aSAbel Vesa { 51*2afbf43aSAbel Vesa u32 regval = qcom_ice_readl(ice, QCOM_ICE_REG_VERSION); 52*2afbf43aSAbel Vesa struct device *dev = ice->dev; 53*2afbf43aSAbel Vesa int major = FIELD_GET(GENMASK(31, 24), regval); 54*2afbf43aSAbel Vesa int minor = FIELD_GET(GENMASK(23, 16), regval); 55*2afbf43aSAbel Vesa int step = FIELD_GET(GENMASK(15, 0), regval); 56*2afbf43aSAbel Vesa 57*2afbf43aSAbel Vesa /* For now this driver only supports ICE version 3 and 4. */ 58*2afbf43aSAbel Vesa if (major != 3 && major != 4) { 59*2afbf43aSAbel Vesa dev_warn(dev, "Unsupported ICE version: v%d.%d.%d\n", 60*2afbf43aSAbel Vesa major, minor, step); 61*2afbf43aSAbel Vesa return false; 62*2afbf43aSAbel Vesa } 63*2afbf43aSAbel Vesa 64*2afbf43aSAbel Vesa dev_info(dev, "Found QC Inline Crypto Engine (ICE) v%d.%d.%d\n", 65*2afbf43aSAbel Vesa major, minor, step); 66*2afbf43aSAbel Vesa 67*2afbf43aSAbel Vesa /* If fuses are blown, ICE might not work in the standard way. */ 68*2afbf43aSAbel Vesa regval = qcom_ice_readl(ice, QCOM_ICE_REG_FUSE_SETTING); 69*2afbf43aSAbel Vesa if (regval & (QCOM_ICE_FUSE_SETTING_MASK | 70*2afbf43aSAbel Vesa QCOM_ICE_FORCE_HW_KEY0_SETTING_MASK | 71*2afbf43aSAbel Vesa QCOM_ICE_FORCE_HW_KEY1_SETTING_MASK)) { 72*2afbf43aSAbel Vesa dev_warn(dev, "Fuses are blown; ICE is unusable!\n"); 73*2afbf43aSAbel Vesa return false; 74*2afbf43aSAbel Vesa } 75*2afbf43aSAbel Vesa 76*2afbf43aSAbel Vesa return true; 77*2afbf43aSAbel Vesa } 78*2afbf43aSAbel Vesa 79*2afbf43aSAbel Vesa static void qcom_ice_low_power_mode_enable(struct qcom_ice *ice) 80*2afbf43aSAbel Vesa { 81*2afbf43aSAbel Vesa u32 regval; 82*2afbf43aSAbel Vesa 83*2afbf43aSAbel Vesa regval = qcom_ice_readl(ice, QCOM_ICE_REG_ADVANCED_CONTROL); 84*2afbf43aSAbel Vesa 85*2afbf43aSAbel Vesa /* Enable low power mode sequence */ 86*2afbf43aSAbel Vesa regval |= 0x7000; 87*2afbf43aSAbel Vesa qcom_ice_writel(ice, regval, QCOM_ICE_REG_ADVANCED_CONTROL); 88*2afbf43aSAbel Vesa } 89*2afbf43aSAbel Vesa 90*2afbf43aSAbel Vesa static void qcom_ice_optimization_enable(struct qcom_ice *ice) 91*2afbf43aSAbel Vesa { 92*2afbf43aSAbel Vesa u32 regval; 93*2afbf43aSAbel Vesa 94*2afbf43aSAbel Vesa /* ICE Optimizations Enable Sequence */ 95*2afbf43aSAbel Vesa regval = qcom_ice_readl(ice, QCOM_ICE_REG_ADVANCED_CONTROL); 96*2afbf43aSAbel Vesa regval |= 0xd807100; 97*2afbf43aSAbel Vesa /* ICE HPG requires delay before writing */ 98*2afbf43aSAbel Vesa udelay(5); 99*2afbf43aSAbel Vesa qcom_ice_writel(ice, regval, QCOM_ICE_REG_ADVANCED_CONTROL); 100*2afbf43aSAbel Vesa udelay(5); 101*2afbf43aSAbel Vesa } 102*2afbf43aSAbel Vesa 103*2afbf43aSAbel Vesa /* 104*2afbf43aSAbel Vesa * Wait until the ICE BIST (built-in self-test) has completed. 105*2afbf43aSAbel Vesa * 106*2afbf43aSAbel Vesa * This may be necessary before ICE can be used. 107*2afbf43aSAbel Vesa * Note that we don't really care whether the BIST passed or failed; 108*2afbf43aSAbel Vesa * we really just want to make sure that it isn't still running. This is 109*2afbf43aSAbel Vesa * because (a) the BIST is a FIPS compliance thing that never fails in 110*2afbf43aSAbel Vesa * practice, (b) ICE is documented to reject crypto requests if the BIST 111*2afbf43aSAbel Vesa * fails, so we needn't do it in software too, and (c) properly testing 112*2afbf43aSAbel Vesa * storage encryption requires testing the full storage stack anyway, 113*2afbf43aSAbel Vesa * and not relying on hardware-level self-tests. 114*2afbf43aSAbel Vesa */ 115*2afbf43aSAbel Vesa static int qcom_ice_wait_bist_status(struct qcom_ice *ice) 116*2afbf43aSAbel Vesa { 117*2afbf43aSAbel Vesa u32 regval; 118*2afbf43aSAbel Vesa int err; 119*2afbf43aSAbel Vesa 120*2afbf43aSAbel Vesa err = readl_poll_timeout(ice->base + QCOM_ICE_REG_BIST_STATUS, 121*2afbf43aSAbel Vesa regval, !(regval & QCOM_ICE_BIST_STATUS_MASK), 122*2afbf43aSAbel Vesa 50, 5000); 123*2afbf43aSAbel Vesa if (err) 124*2afbf43aSAbel Vesa dev_err(ice->dev, "Timed out waiting for ICE self-test to complete\n"); 125*2afbf43aSAbel Vesa 126*2afbf43aSAbel Vesa return err; 127*2afbf43aSAbel Vesa } 128*2afbf43aSAbel Vesa 129*2afbf43aSAbel Vesa int qcom_ice_enable(struct qcom_ice *ice) 130*2afbf43aSAbel Vesa { 131*2afbf43aSAbel Vesa qcom_ice_low_power_mode_enable(ice); 132*2afbf43aSAbel Vesa qcom_ice_optimization_enable(ice); 133*2afbf43aSAbel Vesa 134*2afbf43aSAbel Vesa return qcom_ice_wait_bist_status(ice); 135*2afbf43aSAbel Vesa } 136*2afbf43aSAbel Vesa EXPORT_SYMBOL_GPL(qcom_ice_enable); 137*2afbf43aSAbel Vesa 138*2afbf43aSAbel Vesa int qcom_ice_resume(struct qcom_ice *ice) 139*2afbf43aSAbel Vesa { 140*2afbf43aSAbel Vesa struct device *dev = ice->dev; 141*2afbf43aSAbel Vesa int err; 142*2afbf43aSAbel Vesa 143*2afbf43aSAbel Vesa err = clk_prepare_enable(ice->core_clk); 144*2afbf43aSAbel Vesa if (err) { 145*2afbf43aSAbel Vesa dev_err(dev, "failed to enable core clock (%d)\n", 146*2afbf43aSAbel Vesa err); 147*2afbf43aSAbel Vesa return err; 148*2afbf43aSAbel Vesa } 149*2afbf43aSAbel Vesa 150*2afbf43aSAbel Vesa return qcom_ice_wait_bist_status(ice); 151*2afbf43aSAbel Vesa } 152*2afbf43aSAbel Vesa EXPORT_SYMBOL_GPL(qcom_ice_resume); 153*2afbf43aSAbel Vesa 154*2afbf43aSAbel Vesa int qcom_ice_suspend(struct qcom_ice *ice) 155*2afbf43aSAbel Vesa { 156*2afbf43aSAbel Vesa clk_disable_unprepare(ice->core_clk); 157*2afbf43aSAbel Vesa 158*2afbf43aSAbel Vesa return 0; 159*2afbf43aSAbel Vesa } 160*2afbf43aSAbel Vesa EXPORT_SYMBOL_GPL(qcom_ice_suspend); 161*2afbf43aSAbel Vesa 162*2afbf43aSAbel Vesa int qcom_ice_program_key(struct qcom_ice *ice, 163*2afbf43aSAbel Vesa u8 algorithm_id, u8 key_size, 164*2afbf43aSAbel Vesa const u8 crypto_key[], u8 data_unit_size, 165*2afbf43aSAbel Vesa int slot) 166*2afbf43aSAbel Vesa { 167*2afbf43aSAbel Vesa struct device *dev = ice->dev; 168*2afbf43aSAbel Vesa union { 169*2afbf43aSAbel Vesa u8 bytes[AES_256_XTS_KEY_SIZE]; 170*2afbf43aSAbel Vesa u32 words[AES_256_XTS_KEY_SIZE / sizeof(u32)]; 171*2afbf43aSAbel Vesa } key; 172*2afbf43aSAbel Vesa int i; 173*2afbf43aSAbel Vesa int err; 174*2afbf43aSAbel Vesa 175*2afbf43aSAbel Vesa /* Only AES-256-XTS has been tested so far. */ 176*2afbf43aSAbel Vesa if (algorithm_id != QCOM_ICE_CRYPTO_ALG_AES_XTS || 177*2afbf43aSAbel Vesa key_size != QCOM_ICE_CRYPTO_KEY_SIZE_256) { 178*2afbf43aSAbel Vesa dev_err_ratelimited(dev, 179*2afbf43aSAbel Vesa "Unhandled crypto capability; algorithm_id=%d, key_size=%d\n", 180*2afbf43aSAbel Vesa algorithm_id, key_size); 181*2afbf43aSAbel Vesa return -EINVAL; 182*2afbf43aSAbel Vesa } 183*2afbf43aSAbel Vesa 184*2afbf43aSAbel Vesa memcpy(key.bytes, crypto_key, AES_256_XTS_KEY_SIZE); 185*2afbf43aSAbel Vesa 186*2afbf43aSAbel Vesa /* The SCM call requires that the key words are encoded in big endian */ 187*2afbf43aSAbel Vesa for (i = 0; i < ARRAY_SIZE(key.words); i++) 188*2afbf43aSAbel Vesa __cpu_to_be32s(&key.words[i]); 189*2afbf43aSAbel Vesa 190*2afbf43aSAbel Vesa err = qcom_scm_ice_set_key(slot, key.bytes, AES_256_XTS_KEY_SIZE, 191*2afbf43aSAbel Vesa QCOM_SCM_ICE_CIPHER_AES_256_XTS, 192*2afbf43aSAbel Vesa data_unit_size); 193*2afbf43aSAbel Vesa 194*2afbf43aSAbel Vesa memzero_explicit(&key, sizeof(key)); 195*2afbf43aSAbel Vesa 196*2afbf43aSAbel Vesa return err; 197*2afbf43aSAbel Vesa } 198*2afbf43aSAbel Vesa EXPORT_SYMBOL_GPL(qcom_ice_program_key); 199*2afbf43aSAbel Vesa 200*2afbf43aSAbel Vesa int qcom_ice_evict_key(struct qcom_ice *ice, int slot) 201*2afbf43aSAbel Vesa { 202*2afbf43aSAbel Vesa return qcom_scm_ice_invalidate_key(slot); 203*2afbf43aSAbel Vesa } 204*2afbf43aSAbel Vesa EXPORT_SYMBOL_GPL(qcom_ice_evict_key); 205*2afbf43aSAbel Vesa 206*2afbf43aSAbel Vesa static struct qcom_ice *qcom_ice_create(struct device *dev, 207*2afbf43aSAbel Vesa void __iomem *base) 208*2afbf43aSAbel Vesa { 209*2afbf43aSAbel Vesa struct qcom_ice *engine; 210*2afbf43aSAbel Vesa 211*2afbf43aSAbel Vesa if (!qcom_scm_is_available()) 212*2afbf43aSAbel Vesa return ERR_PTR(-EPROBE_DEFER); 213*2afbf43aSAbel Vesa 214*2afbf43aSAbel Vesa if (!qcom_scm_ice_available()) { 215*2afbf43aSAbel Vesa dev_warn(dev, "ICE SCM interface not found\n"); 216*2afbf43aSAbel Vesa return NULL; 217*2afbf43aSAbel Vesa } 218*2afbf43aSAbel Vesa 219*2afbf43aSAbel Vesa engine = devm_kzalloc(dev, sizeof(*engine), GFP_KERNEL); 220*2afbf43aSAbel Vesa if (!engine) 221*2afbf43aSAbel Vesa return ERR_PTR(-ENOMEM); 222*2afbf43aSAbel Vesa 223*2afbf43aSAbel Vesa engine->dev = dev; 224*2afbf43aSAbel Vesa engine->base = base; 225*2afbf43aSAbel Vesa 226*2afbf43aSAbel Vesa /* 227*2afbf43aSAbel Vesa * Legacy DT binding uses different clk names for each consumer, 228*2afbf43aSAbel Vesa * so lets try those first. If none of those are a match, it means 229*2afbf43aSAbel Vesa * the we only have one clock and it is part of the dedicated DT node. 230*2afbf43aSAbel Vesa * Also, enable the clock before we check what HW version the driver 231*2afbf43aSAbel Vesa * supports. 232*2afbf43aSAbel Vesa */ 233*2afbf43aSAbel Vesa engine->core_clk = devm_clk_get_optional_enabled(dev, "ice_core_clk"); 234*2afbf43aSAbel Vesa if (!engine->core_clk) 235*2afbf43aSAbel Vesa engine->core_clk = devm_clk_get_optional_enabled(dev, "ice"); 236*2afbf43aSAbel Vesa if (!engine->core_clk) 237*2afbf43aSAbel Vesa engine->core_clk = devm_clk_get_enabled(dev, NULL); 238*2afbf43aSAbel Vesa if (IS_ERR(engine->core_clk)) 239*2afbf43aSAbel Vesa return ERR_CAST(engine->core_clk); 240*2afbf43aSAbel Vesa 241*2afbf43aSAbel Vesa if (!qcom_ice_check_supported(engine)) 242*2afbf43aSAbel Vesa return ERR_PTR(-EOPNOTSUPP); 243*2afbf43aSAbel Vesa 244*2afbf43aSAbel Vesa dev_dbg(dev, "Registered Qualcomm Inline Crypto Engine\n"); 245*2afbf43aSAbel Vesa 246*2afbf43aSAbel Vesa return engine; 247*2afbf43aSAbel Vesa } 248*2afbf43aSAbel Vesa 249*2afbf43aSAbel Vesa /** 250*2afbf43aSAbel Vesa * of_qcom_ice_get() - get an ICE instance from a DT node 251*2afbf43aSAbel Vesa * @dev: device pointer for the consumer device 252*2afbf43aSAbel Vesa * 253*2afbf43aSAbel Vesa * This function will provide an ICE instance either by creating one for the 254*2afbf43aSAbel Vesa * consumer device if its DT node provides the 'ice' reg range and the 'ice' 255*2afbf43aSAbel Vesa * clock (for legacy DT style). On the other hand, if consumer provides a 256*2afbf43aSAbel Vesa * phandle via 'qcom,ice' property to an ICE DT, the ICE instance will already 257*2afbf43aSAbel Vesa * be created and so this function will return that instead. 258*2afbf43aSAbel Vesa * 259*2afbf43aSAbel Vesa * Return: ICE pointer on success, NULL if there is no ICE data provided by the 260*2afbf43aSAbel Vesa * consumer or ERR_PTR() on error. 261*2afbf43aSAbel Vesa */ 262*2afbf43aSAbel Vesa struct qcom_ice *of_qcom_ice_get(struct device *dev) 263*2afbf43aSAbel Vesa { 264*2afbf43aSAbel Vesa struct platform_device *pdev = to_platform_device(dev); 265*2afbf43aSAbel Vesa struct qcom_ice *ice; 266*2afbf43aSAbel Vesa struct device_node *node; 267*2afbf43aSAbel Vesa struct resource *res; 268*2afbf43aSAbel Vesa void __iomem *base; 269*2afbf43aSAbel Vesa 270*2afbf43aSAbel Vesa if (!dev || !dev->of_node) 271*2afbf43aSAbel Vesa return ERR_PTR(-ENODEV); 272*2afbf43aSAbel Vesa 273*2afbf43aSAbel Vesa /* 274*2afbf43aSAbel Vesa * In order to support legacy style devicetree bindings, we need 275*2afbf43aSAbel Vesa * to create the ICE instance using the consumer device and the reg 276*2afbf43aSAbel Vesa * range called 'ice' it provides. 277*2afbf43aSAbel Vesa */ 278*2afbf43aSAbel Vesa res = platform_get_resource_byname(pdev, IORESOURCE_MEM, "ice"); 279*2afbf43aSAbel Vesa if (res) { 280*2afbf43aSAbel Vesa base = devm_ioremap_resource(&pdev->dev, res); 281*2afbf43aSAbel Vesa if (IS_ERR(base)) 282*2afbf43aSAbel Vesa return ERR_CAST(base); 283*2afbf43aSAbel Vesa 284*2afbf43aSAbel Vesa /* create ICE instance using consumer dev */ 285*2afbf43aSAbel Vesa return qcom_ice_create(&pdev->dev, base); 286*2afbf43aSAbel Vesa } 287*2afbf43aSAbel Vesa 288*2afbf43aSAbel Vesa /* 289*2afbf43aSAbel Vesa * If the consumer node does not provider an 'ice' reg range 290*2afbf43aSAbel Vesa * (legacy DT binding), then it must at least provide a phandle 291*2afbf43aSAbel Vesa * to the ICE devicetree node, otherwise ICE is not supported. 292*2afbf43aSAbel Vesa */ 293*2afbf43aSAbel Vesa node = of_parse_phandle(dev->of_node, "qcom,ice", 0); 294*2afbf43aSAbel Vesa if (!node) 295*2afbf43aSAbel Vesa return NULL; 296*2afbf43aSAbel Vesa 297*2afbf43aSAbel Vesa pdev = of_find_device_by_node(node); 298*2afbf43aSAbel Vesa if (!pdev) { 299*2afbf43aSAbel Vesa dev_err(dev, "Cannot find device node %s\n", node->name); 300*2afbf43aSAbel Vesa ice = ERR_PTR(-EPROBE_DEFER); 301*2afbf43aSAbel Vesa goto out; 302*2afbf43aSAbel Vesa } 303*2afbf43aSAbel Vesa 304*2afbf43aSAbel Vesa ice = platform_get_drvdata(pdev); 305*2afbf43aSAbel Vesa if (!ice) { 306*2afbf43aSAbel Vesa dev_err(dev, "Cannot get ice instance from %s\n", 307*2afbf43aSAbel Vesa dev_name(&pdev->dev)); 308*2afbf43aSAbel Vesa platform_device_put(pdev); 309*2afbf43aSAbel Vesa ice = ERR_PTR(-EPROBE_DEFER); 310*2afbf43aSAbel Vesa goto out; 311*2afbf43aSAbel Vesa } 312*2afbf43aSAbel Vesa 313*2afbf43aSAbel Vesa ice->link = device_link_add(dev, &pdev->dev, DL_FLAG_AUTOREMOVE_SUPPLIER); 314*2afbf43aSAbel Vesa if (!ice->link) { 315*2afbf43aSAbel Vesa dev_err(&pdev->dev, 316*2afbf43aSAbel Vesa "Failed to create device link to consumer %s\n", 317*2afbf43aSAbel Vesa dev_name(dev)); 318*2afbf43aSAbel Vesa platform_device_put(pdev); 319*2afbf43aSAbel Vesa ice = ERR_PTR(-EINVAL); 320*2afbf43aSAbel Vesa } 321*2afbf43aSAbel Vesa 322*2afbf43aSAbel Vesa out: 323*2afbf43aSAbel Vesa of_node_put(node); 324*2afbf43aSAbel Vesa 325*2afbf43aSAbel Vesa return ice; 326*2afbf43aSAbel Vesa } 327*2afbf43aSAbel Vesa EXPORT_SYMBOL_GPL(of_qcom_ice_get); 328*2afbf43aSAbel Vesa 329*2afbf43aSAbel Vesa static int qcom_ice_probe(struct platform_device *pdev) 330*2afbf43aSAbel Vesa { 331*2afbf43aSAbel Vesa struct qcom_ice *engine; 332*2afbf43aSAbel Vesa void __iomem *base; 333*2afbf43aSAbel Vesa 334*2afbf43aSAbel Vesa base = devm_platform_ioremap_resource(pdev, 0); 335*2afbf43aSAbel Vesa if (IS_ERR(base)) { 336*2afbf43aSAbel Vesa dev_warn(&pdev->dev, "ICE registers not found\n"); 337*2afbf43aSAbel Vesa return PTR_ERR(base); 338*2afbf43aSAbel Vesa } 339*2afbf43aSAbel Vesa 340*2afbf43aSAbel Vesa engine = qcom_ice_create(&pdev->dev, base); 341*2afbf43aSAbel Vesa if (IS_ERR(engine)) 342*2afbf43aSAbel Vesa return PTR_ERR(engine); 343*2afbf43aSAbel Vesa 344*2afbf43aSAbel Vesa platform_set_drvdata(pdev, engine); 345*2afbf43aSAbel Vesa 346*2afbf43aSAbel Vesa return 0; 347*2afbf43aSAbel Vesa } 348*2afbf43aSAbel Vesa 349*2afbf43aSAbel Vesa static const struct of_device_id qcom_ice_of_match_table[] = { 350*2afbf43aSAbel Vesa { .compatible = "qcom,inline-crypto-engine" }, 351*2afbf43aSAbel Vesa { }, 352*2afbf43aSAbel Vesa }; 353*2afbf43aSAbel Vesa MODULE_DEVICE_TABLE(of, qcom_ice_of_match_table); 354*2afbf43aSAbel Vesa 355*2afbf43aSAbel Vesa static struct platform_driver qcom_ice_driver = { 356*2afbf43aSAbel Vesa .probe = qcom_ice_probe, 357*2afbf43aSAbel Vesa .driver = { 358*2afbf43aSAbel Vesa .name = "qcom-ice", 359*2afbf43aSAbel Vesa .of_match_table = qcom_ice_of_match_table, 360*2afbf43aSAbel Vesa }, 361*2afbf43aSAbel Vesa }; 362*2afbf43aSAbel Vesa 363*2afbf43aSAbel Vesa module_platform_driver(qcom_ice_driver); 364*2afbf43aSAbel Vesa 365*2afbf43aSAbel Vesa MODULE_DESCRIPTION("Qualcomm Inline Crypto Engine driver"); 366*2afbf43aSAbel Vesa MODULE_LICENSE("GPL"); 367