xref: /openbmc/linux/drivers/soc/qcom/ice.c (revision 2afbf43a4aec6e31dac7835e65d52c867f2be400)
1*2afbf43aSAbel Vesa // SPDX-License-Identifier: GPL-2.0
2*2afbf43aSAbel Vesa /*
3*2afbf43aSAbel Vesa  * Qualcomm ICE (Inline Crypto Engine) support.
4*2afbf43aSAbel Vesa  *
5*2afbf43aSAbel Vesa  * Copyright (c) 2013-2019, The Linux Foundation. All rights reserved.
6*2afbf43aSAbel Vesa  * Copyright (c) 2019, Google LLC
7*2afbf43aSAbel Vesa  * Copyright (c) 2023, Linaro Limited
8*2afbf43aSAbel Vesa  */
9*2afbf43aSAbel Vesa 
10*2afbf43aSAbel Vesa #include <linux/bitfield.h>
11*2afbf43aSAbel Vesa #include <linux/clk.h>
12*2afbf43aSAbel Vesa #include <linux/delay.h>
13*2afbf43aSAbel Vesa #include <linux/iopoll.h>
14*2afbf43aSAbel Vesa #include <linux/of_platform.h>
15*2afbf43aSAbel Vesa 
16*2afbf43aSAbel Vesa #include <linux/firmware/qcom/qcom_scm.h>
17*2afbf43aSAbel Vesa 
18*2afbf43aSAbel Vesa #include <soc/qcom/ice.h>
19*2afbf43aSAbel Vesa 
20*2afbf43aSAbel Vesa #define AES_256_XTS_KEY_SIZE			64
21*2afbf43aSAbel Vesa 
22*2afbf43aSAbel Vesa /* QCOM ICE registers */
23*2afbf43aSAbel Vesa #define QCOM_ICE_REG_VERSION			0x0008
24*2afbf43aSAbel Vesa #define QCOM_ICE_REG_FUSE_SETTING		0x0010
25*2afbf43aSAbel Vesa #define QCOM_ICE_REG_BIST_STATUS		0x0070
26*2afbf43aSAbel Vesa #define QCOM_ICE_REG_ADVANCED_CONTROL		0x1000
27*2afbf43aSAbel Vesa 
28*2afbf43aSAbel Vesa /* BIST ("built-in self-test") status flags */
29*2afbf43aSAbel Vesa #define QCOM_ICE_BIST_STATUS_MASK		GENMASK(31, 28)
30*2afbf43aSAbel Vesa 
31*2afbf43aSAbel Vesa #define QCOM_ICE_FUSE_SETTING_MASK		0x1
32*2afbf43aSAbel Vesa #define QCOM_ICE_FORCE_HW_KEY0_SETTING_MASK	0x2
33*2afbf43aSAbel Vesa #define QCOM_ICE_FORCE_HW_KEY1_SETTING_MASK	0x4
34*2afbf43aSAbel Vesa 
35*2afbf43aSAbel Vesa #define qcom_ice_writel(engine, val, reg)	\
36*2afbf43aSAbel Vesa 	writel((val), (engine)->base + (reg))
37*2afbf43aSAbel Vesa 
38*2afbf43aSAbel Vesa #define qcom_ice_readl(engine, reg)	\
39*2afbf43aSAbel Vesa 	readl((engine)->base + (reg))
40*2afbf43aSAbel Vesa 
41*2afbf43aSAbel Vesa struct qcom_ice {
42*2afbf43aSAbel Vesa 	struct device *dev;
43*2afbf43aSAbel Vesa 	void __iomem *base;
44*2afbf43aSAbel Vesa 	struct device_link *link;
45*2afbf43aSAbel Vesa 
46*2afbf43aSAbel Vesa 	struct clk *core_clk;
47*2afbf43aSAbel Vesa };
48*2afbf43aSAbel Vesa 
49*2afbf43aSAbel Vesa static bool qcom_ice_check_supported(struct qcom_ice *ice)
50*2afbf43aSAbel Vesa {
51*2afbf43aSAbel Vesa 	u32 regval = qcom_ice_readl(ice, QCOM_ICE_REG_VERSION);
52*2afbf43aSAbel Vesa 	struct device *dev = ice->dev;
53*2afbf43aSAbel Vesa 	int major = FIELD_GET(GENMASK(31, 24), regval);
54*2afbf43aSAbel Vesa 	int minor = FIELD_GET(GENMASK(23, 16), regval);
55*2afbf43aSAbel Vesa 	int step = FIELD_GET(GENMASK(15, 0), regval);
56*2afbf43aSAbel Vesa 
57*2afbf43aSAbel Vesa 	/* For now this driver only supports ICE version 3 and 4. */
58*2afbf43aSAbel Vesa 	if (major != 3 && major != 4) {
59*2afbf43aSAbel Vesa 		dev_warn(dev, "Unsupported ICE version: v%d.%d.%d\n",
60*2afbf43aSAbel Vesa 			 major, minor, step);
61*2afbf43aSAbel Vesa 		return false;
62*2afbf43aSAbel Vesa 	}
63*2afbf43aSAbel Vesa 
64*2afbf43aSAbel Vesa 	dev_info(dev, "Found QC Inline Crypto Engine (ICE) v%d.%d.%d\n",
65*2afbf43aSAbel Vesa 		 major, minor, step);
66*2afbf43aSAbel Vesa 
67*2afbf43aSAbel Vesa 	/* If fuses are blown, ICE might not work in the standard way. */
68*2afbf43aSAbel Vesa 	regval = qcom_ice_readl(ice, QCOM_ICE_REG_FUSE_SETTING);
69*2afbf43aSAbel Vesa 	if (regval & (QCOM_ICE_FUSE_SETTING_MASK |
70*2afbf43aSAbel Vesa 		      QCOM_ICE_FORCE_HW_KEY0_SETTING_MASK |
71*2afbf43aSAbel Vesa 		      QCOM_ICE_FORCE_HW_KEY1_SETTING_MASK)) {
72*2afbf43aSAbel Vesa 		dev_warn(dev, "Fuses are blown; ICE is unusable!\n");
73*2afbf43aSAbel Vesa 		return false;
74*2afbf43aSAbel Vesa 	}
75*2afbf43aSAbel Vesa 
76*2afbf43aSAbel Vesa 	return true;
77*2afbf43aSAbel Vesa }
78*2afbf43aSAbel Vesa 
79*2afbf43aSAbel Vesa static void qcom_ice_low_power_mode_enable(struct qcom_ice *ice)
80*2afbf43aSAbel Vesa {
81*2afbf43aSAbel Vesa 	u32 regval;
82*2afbf43aSAbel Vesa 
83*2afbf43aSAbel Vesa 	regval = qcom_ice_readl(ice, QCOM_ICE_REG_ADVANCED_CONTROL);
84*2afbf43aSAbel Vesa 
85*2afbf43aSAbel Vesa 	/* Enable low power mode sequence */
86*2afbf43aSAbel Vesa 	regval |= 0x7000;
87*2afbf43aSAbel Vesa 	qcom_ice_writel(ice, regval, QCOM_ICE_REG_ADVANCED_CONTROL);
88*2afbf43aSAbel Vesa }
89*2afbf43aSAbel Vesa 
90*2afbf43aSAbel Vesa static void qcom_ice_optimization_enable(struct qcom_ice *ice)
91*2afbf43aSAbel Vesa {
92*2afbf43aSAbel Vesa 	u32 regval;
93*2afbf43aSAbel Vesa 
94*2afbf43aSAbel Vesa 	/* ICE Optimizations Enable Sequence */
95*2afbf43aSAbel Vesa 	regval = qcom_ice_readl(ice, QCOM_ICE_REG_ADVANCED_CONTROL);
96*2afbf43aSAbel Vesa 	regval |= 0xd807100;
97*2afbf43aSAbel Vesa 	/* ICE HPG requires delay before writing */
98*2afbf43aSAbel Vesa 	udelay(5);
99*2afbf43aSAbel Vesa 	qcom_ice_writel(ice, regval, QCOM_ICE_REG_ADVANCED_CONTROL);
100*2afbf43aSAbel Vesa 	udelay(5);
101*2afbf43aSAbel Vesa }
102*2afbf43aSAbel Vesa 
103*2afbf43aSAbel Vesa /*
104*2afbf43aSAbel Vesa  * Wait until the ICE BIST (built-in self-test) has completed.
105*2afbf43aSAbel Vesa  *
106*2afbf43aSAbel Vesa  * This may be necessary before ICE can be used.
107*2afbf43aSAbel Vesa  * Note that we don't really care whether the BIST passed or failed;
108*2afbf43aSAbel Vesa  * we really just want to make sure that it isn't still running. This is
109*2afbf43aSAbel Vesa  * because (a) the BIST is a FIPS compliance thing that never fails in
110*2afbf43aSAbel Vesa  * practice, (b) ICE is documented to reject crypto requests if the BIST
111*2afbf43aSAbel Vesa  * fails, so we needn't do it in software too, and (c) properly testing
112*2afbf43aSAbel Vesa  * storage encryption requires testing the full storage stack anyway,
113*2afbf43aSAbel Vesa  * and not relying on hardware-level self-tests.
114*2afbf43aSAbel Vesa  */
115*2afbf43aSAbel Vesa static int qcom_ice_wait_bist_status(struct qcom_ice *ice)
116*2afbf43aSAbel Vesa {
117*2afbf43aSAbel Vesa 	u32 regval;
118*2afbf43aSAbel Vesa 	int err;
119*2afbf43aSAbel Vesa 
120*2afbf43aSAbel Vesa 	err = readl_poll_timeout(ice->base + QCOM_ICE_REG_BIST_STATUS,
121*2afbf43aSAbel Vesa 				 regval, !(regval & QCOM_ICE_BIST_STATUS_MASK),
122*2afbf43aSAbel Vesa 				 50, 5000);
123*2afbf43aSAbel Vesa 	if (err)
124*2afbf43aSAbel Vesa 		dev_err(ice->dev, "Timed out waiting for ICE self-test to complete\n");
125*2afbf43aSAbel Vesa 
126*2afbf43aSAbel Vesa 	return err;
127*2afbf43aSAbel Vesa }
128*2afbf43aSAbel Vesa 
129*2afbf43aSAbel Vesa int qcom_ice_enable(struct qcom_ice *ice)
130*2afbf43aSAbel Vesa {
131*2afbf43aSAbel Vesa 	qcom_ice_low_power_mode_enable(ice);
132*2afbf43aSAbel Vesa 	qcom_ice_optimization_enable(ice);
133*2afbf43aSAbel Vesa 
134*2afbf43aSAbel Vesa 	return qcom_ice_wait_bist_status(ice);
135*2afbf43aSAbel Vesa }
136*2afbf43aSAbel Vesa EXPORT_SYMBOL_GPL(qcom_ice_enable);
137*2afbf43aSAbel Vesa 
138*2afbf43aSAbel Vesa int qcom_ice_resume(struct qcom_ice *ice)
139*2afbf43aSAbel Vesa {
140*2afbf43aSAbel Vesa 	struct device *dev = ice->dev;
141*2afbf43aSAbel Vesa 	int err;
142*2afbf43aSAbel Vesa 
143*2afbf43aSAbel Vesa 	err = clk_prepare_enable(ice->core_clk);
144*2afbf43aSAbel Vesa 	if (err) {
145*2afbf43aSAbel Vesa 		dev_err(dev, "failed to enable core clock (%d)\n",
146*2afbf43aSAbel Vesa 			err);
147*2afbf43aSAbel Vesa 		return err;
148*2afbf43aSAbel Vesa 	}
149*2afbf43aSAbel Vesa 
150*2afbf43aSAbel Vesa 	return qcom_ice_wait_bist_status(ice);
151*2afbf43aSAbel Vesa }
152*2afbf43aSAbel Vesa EXPORT_SYMBOL_GPL(qcom_ice_resume);
153*2afbf43aSAbel Vesa 
154*2afbf43aSAbel Vesa int qcom_ice_suspend(struct qcom_ice *ice)
155*2afbf43aSAbel Vesa {
156*2afbf43aSAbel Vesa 	clk_disable_unprepare(ice->core_clk);
157*2afbf43aSAbel Vesa 
158*2afbf43aSAbel Vesa 	return 0;
159*2afbf43aSAbel Vesa }
160*2afbf43aSAbel Vesa EXPORT_SYMBOL_GPL(qcom_ice_suspend);
161*2afbf43aSAbel Vesa 
162*2afbf43aSAbel Vesa int qcom_ice_program_key(struct qcom_ice *ice,
163*2afbf43aSAbel Vesa 			 u8 algorithm_id, u8 key_size,
164*2afbf43aSAbel Vesa 			 const u8 crypto_key[], u8 data_unit_size,
165*2afbf43aSAbel Vesa 			 int slot)
166*2afbf43aSAbel Vesa {
167*2afbf43aSAbel Vesa 	struct device *dev = ice->dev;
168*2afbf43aSAbel Vesa 	union {
169*2afbf43aSAbel Vesa 		u8 bytes[AES_256_XTS_KEY_SIZE];
170*2afbf43aSAbel Vesa 		u32 words[AES_256_XTS_KEY_SIZE / sizeof(u32)];
171*2afbf43aSAbel Vesa 	} key;
172*2afbf43aSAbel Vesa 	int i;
173*2afbf43aSAbel Vesa 	int err;
174*2afbf43aSAbel Vesa 
175*2afbf43aSAbel Vesa 	/* Only AES-256-XTS has been tested so far. */
176*2afbf43aSAbel Vesa 	if (algorithm_id != QCOM_ICE_CRYPTO_ALG_AES_XTS ||
177*2afbf43aSAbel Vesa 	    key_size != QCOM_ICE_CRYPTO_KEY_SIZE_256) {
178*2afbf43aSAbel Vesa 		dev_err_ratelimited(dev,
179*2afbf43aSAbel Vesa 				    "Unhandled crypto capability; algorithm_id=%d, key_size=%d\n",
180*2afbf43aSAbel Vesa 				    algorithm_id, key_size);
181*2afbf43aSAbel Vesa 		return -EINVAL;
182*2afbf43aSAbel Vesa 	}
183*2afbf43aSAbel Vesa 
184*2afbf43aSAbel Vesa 	memcpy(key.bytes, crypto_key, AES_256_XTS_KEY_SIZE);
185*2afbf43aSAbel Vesa 
186*2afbf43aSAbel Vesa 	/* The SCM call requires that the key words are encoded in big endian */
187*2afbf43aSAbel Vesa 	for (i = 0; i < ARRAY_SIZE(key.words); i++)
188*2afbf43aSAbel Vesa 		__cpu_to_be32s(&key.words[i]);
189*2afbf43aSAbel Vesa 
190*2afbf43aSAbel Vesa 	err = qcom_scm_ice_set_key(slot, key.bytes, AES_256_XTS_KEY_SIZE,
191*2afbf43aSAbel Vesa 				   QCOM_SCM_ICE_CIPHER_AES_256_XTS,
192*2afbf43aSAbel Vesa 				   data_unit_size);
193*2afbf43aSAbel Vesa 
194*2afbf43aSAbel Vesa 	memzero_explicit(&key, sizeof(key));
195*2afbf43aSAbel Vesa 
196*2afbf43aSAbel Vesa 	return err;
197*2afbf43aSAbel Vesa }
198*2afbf43aSAbel Vesa EXPORT_SYMBOL_GPL(qcom_ice_program_key);
199*2afbf43aSAbel Vesa 
200*2afbf43aSAbel Vesa int qcom_ice_evict_key(struct qcom_ice *ice, int slot)
201*2afbf43aSAbel Vesa {
202*2afbf43aSAbel Vesa 	return qcom_scm_ice_invalidate_key(slot);
203*2afbf43aSAbel Vesa }
204*2afbf43aSAbel Vesa EXPORT_SYMBOL_GPL(qcom_ice_evict_key);
205*2afbf43aSAbel Vesa 
206*2afbf43aSAbel Vesa static struct qcom_ice *qcom_ice_create(struct device *dev,
207*2afbf43aSAbel Vesa 					void __iomem *base)
208*2afbf43aSAbel Vesa {
209*2afbf43aSAbel Vesa 	struct qcom_ice *engine;
210*2afbf43aSAbel Vesa 
211*2afbf43aSAbel Vesa 	if (!qcom_scm_is_available())
212*2afbf43aSAbel Vesa 		return ERR_PTR(-EPROBE_DEFER);
213*2afbf43aSAbel Vesa 
214*2afbf43aSAbel Vesa 	if (!qcom_scm_ice_available()) {
215*2afbf43aSAbel Vesa 		dev_warn(dev, "ICE SCM interface not found\n");
216*2afbf43aSAbel Vesa 		return NULL;
217*2afbf43aSAbel Vesa 	}
218*2afbf43aSAbel Vesa 
219*2afbf43aSAbel Vesa 	engine = devm_kzalloc(dev, sizeof(*engine), GFP_KERNEL);
220*2afbf43aSAbel Vesa 	if (!engine)
221*2afbf43aSAbel Vesa 		return ERR_PTR(-ENOMEM);
222*2afbf43aSAbel Vesa 
223*2afbf43aSAbel Vesa 	engine->dev = dev;
224*2afbf43aSAbel Vesa 	engine->base = base;
225*2afbf43aSAbel Vesa 
226*2afbf43aSAbel Vesa 	/*
227*2afbf43aSAbel Vesa 	 * Legacy DT binding uses different clk names for each consumer,
228*2afbf43aSAbel Vesa 	 * so lets try those first. If none of those are a match, it means
229*2afbf43aSAbel Vesa 	 * the we only have one clock and it is part of the dedicated DT node.
230*2afbf43aSAbel Vesa 	 * Also, enable the clock before we check what HW version the driver
231*2afbf43aSAbel Vesa 	 * supports.
232*2afbf43aSAbel Vesa 	 */
233*2afbf43aSAbel Vesa 	engine->core_clk = devm_clk_get_optional_enabled(dev, "ice_core_clk");
234*2afbf43aSAbel Vesa 	if (!engine->core_clk)
235*2afbf43aSAbel Vesa 		engine->core_clk = devm_clk_get_optional_enabled(dev, "ice");
236*2afbf43aSAbel Vesa 	if (!engine->core_clk)
237*2afbf43aSAbel Vesa 		engine->core_clk = devm_clk_get_enabled(dev, NULL);
238*2afbf43aSAbel Vesa 	if (IS_ERR(engine->core_clk))
239*2afbf43aSAbel Vesa 		return ERR_CAST(engine->core_clk);
240*2afbf43aSAbel Vesa 
241*2afbf43aSAbel Vesa 	if (!qcom_ice_check_supported(engine))
242*2afbf43aSAbel Vesa 		return ERR_PTR(-EOPNOTSUPP);
243*2afbf43aSAbel Vesa 
244*2afbf43aSAbel Vesa 	dev_dbg(dev, "Registered Qualcomm Inline Crypto Engine\n");
245*2afbf43aSAbel Vesa 
246*2afbf43aSAbel Vesa 	return engine;
247*2afbf43aSAbel Vesa }
248*2afbf43aSAbel Vesa 
249*2afbf43aSAbel Vesa /**
250*2afbf43aSAbel Vesa  * of_qcom_ice_get() - get an ICE instance from a DT node
251*2afbf43aSAbel Vesa  * @dev: device pointer for the consumer device
252*2afbf43aSAbel Vesa  *
253*2afbf43aSAbel Vesa  * This function will provide an ICE instance either by creating one for the
254*2afbf43aSAbel Vesa  * consumer device if its DT node provides the 'ice' reg range and the 'ice'
255*2afbf43aSAbel Vesa  * clock (for legacy DT style). On the other hand, if consumer provides a
256*2afbf43aSAbel Vesa  * phandle via 'qcom,ice' property to an ICE DT, the ICE instance will already
257*2afbf43aSAbel Vesa  * be created and so this function will return that instead.
258*2afbf43aSAbel Vesa  *
259*2afbf43aSAbel Vesa  * Return: ICE pointer on success, NULL if there is no ICE data provided by the
260*2afbf43aSAbel Vesa  * consumer or ERR_PTR() on error.
261*2afbf43aSAbel Vesa  */
262*2afbf43aSAbel Vesa struct qcom_ice *of_qcom_ice_get(struct device *dev)
263*2afbf43aSAbel Vesa {
264*2afbf43aSAbel Vesa 	struct platform_device *pdev = to_platform_device(dev);
265*2afbf43aSAbel Vesa 	struct qcom_ice *ice;
266*2afbf43aSAbel Vesa 	struct device_node *node;
267*2afbf43aSAbel Vesa 	struct resource *res;
268*2afbf43aSAbel Vesa 	void __iomem *base;
269*2afbf43aSAbel Vesa 
270*2afbf43aSAbel Vesa 	if (!dev || !dev->of_node)
271*2afbf43aSAbel Vesa 		return ERR_PTR(-ENODEV);
272*2afbf43aSAbel Vesa 
273*2afbf43aSAbel Vesa 	/*
274*2afbf43aSAbel Vesa 	 * In order to support legacy style devicetree bindings, we need
275*2afbf43aSAbel Vesa 	 * to create the ICE instance using the consumer device and the reg
276*2afbf43aSAbel Vesa 	 * range called 'ice' it provides.
277*2afbf43aSAbel Vesa 	 */
278*2afbf43aSAbel Vesa 	res = platform_get_resource_byname(pdev, IORESOURCE_MEM, "ice");
279*2afbf43aSAbel Vesa 	if (res) {
280*2afbf43aSAbel Vesa 		base = devm_ioremap_resource(&pdev->dev, res);
281*2afbf43aSAbel Vesa 		if (IS_ERR(base))
282*2afbf43aSAbel Vesa 			return ERR_CAST(base);
283*2afbf43aSAbel Vesa 
284*2afbf43aSAbel Vesa 		/* create ICE instance using consumer dev */
285*2afbf43aSAbel Vesa 		return qcom_ice_create(&pdev->dev, base);
286*2afbf43aSAbel Vesa 	}
287*2afbf43aSAbel Vesa 
288*2afbf43aSAbel Vesa 	/*
289*2afbf43aSAbel Vesa 	 * If the consumer node does not provider an 'ice' reg range
290*2afbf43aSAbel Vesa 	 * (legacy DT binding), then it must at least provide a phandle
291*2afbf43aSAbel Vesa 	 * to the ICE devicetree node, otherwise ICE is not supported.
292*2afbf43aSAbel Vesa 	 */
293*2afbf43aSAbel Vesa 	node = of_parse_phandle(dev->of_node, "qcom,ice", 0);
294*2afbf43aSAbel Vesa 	if (!node)
295*2afbf43aSAbel Vesa 		return NULL;
296*2afbf43aSAbel Vesa 
297*2afbf43aSAbel Vesa 	pdev = of_find_device_by_node(node);
298*2afbf43aSAbel Vesa 	if (!pdev) {
299*2afbf43aSAbel Vesa 		dev_err(dev, "Cannot find device node %s\n", node->name);
300*2afbf43aSAbel Vesa 		ice = ERR_PTR(-EPROBE_DEFER);
301*2afbf43aSAbel Vesa 		goto out;
302*2afbf43aSAbel Vesa 	}
303*2afbf43aSAbel Vesa 
304*2afbf43aSAbel Vesa 	ice = platform_get_drvdata(pdev);
305*2afbf43aSAbel Vesa 	if (!ice) {
306*2afbf43aSAbel Vesa 		dev_err(dev, "Cannot get ice instance from %s\n",
307*2afbf43aSAbel Vesa 			dev_name(&pdev->dev));
308*2afbf43aSAbel Vesa 		platform_device_put(pdev);
309*2afbf43aSAbel Vesa 		ice = ERR_PTR(-EPROBE_DEFER);
310*2afbf43aSAbel Vesa 		goto out;
311*2afbf43aSAbel Vesa 	}
312*2afbf43aSAbel Vesa 
313*2afbf43aSAbel Vesa 	ice->link = device_link_add(dev, &pdev->dev, DL_FLAG_AUTOREMOVE_SUPPLIER);
314*2afbf43aSAbel Vesa 	if (!ice->link) {
315*2afbf43aSAbel Vesa 		dev_err(&pdev->dev,
316*2afbf43aSAbel Vesa 			"Failed to create device link to consumer %s\n",
317*2afbf43aSAbel Vesa 			dev_name(dev));
318*2afbf43aSAbel Vesa 		platform_device_put(pdev);
319*2afbf43aSAbel Vesa 		ice = ERR_PTR(-EINVAL);
320*2afbf43aSAbel Vesa 	}
321*2afbf43aSAbel Vesa 
322*2afbf43aSAbel Vesa out:
323*2afbf43aSAbel Vesa 	of_node_put(node);
324*2afbf43aSAbel Vesa 
325*2afbf43aSAbel Vesa 	return ice;
326*2afbf43aSAbel Vesa }
327*2afbf43aSAbel Vesa EXPORT_SYMBOL_GPL(of_qcom_ice_get);
328*2afbf43aSAbel Vesa 
329*2afbf43aSAbel Vesa static int qcom_ice_probe(struct platform_device *pdev)
330*2afbf43aSAbel Vesa {
331*2afbf43aSAbel Vesa 	struct qcom_ice *engine;
332*2afbf43aSAbel Vesa 	void __iomem *base;
333*2afbf43aSAbel Vesa 
334*2afbf43aSAbel Vesa 	base = devm_platform_ioremap_resource(pdev, 0);
335*2afbf43aSAbel Vesa 	if (IS_ERR(base)) {
336*2afbf43aSAbel Vesa 		dev_warn(&pdev->dev, "ICE registers not found\n");
337*2afbf43aSAbel Vesa 		return PTR_ERR(base);
338*2afbf43aSAbel Vesa 	}
339*2afbf43aSAbel Vesa 
340*2afbf43aSAbel Vesa 	engine = qcom_ice_create(&pdev->dev, base);
341*2afbf43aSAbel Vesa 	if (IS_ERR(engine))
342*2afbf43aSAbel Vesa 		return PTR_ERR(engine);
343*2afbf43aSAbel Vesa 
344*2afbf43aSAbel Vesa 	platform_set_drvdata(pdev, engine);
345*2afbf43aSAbel Vesa 
346*2afbf43aSAbel Vesa 	return 0;
347*2afbf43aSAbel Vesa }
348*2afbf43aSAbel Vesa 
349*2afbf43aSAbel Vesa static const struct of_device_id qcom_ice_of_match_table[] = {
350*2afbf43aSAbel Vesa 	{ .compatible = "qcom,inline-crypto-engine" },
351*2afbf43aSAbel Vesa 	{ },
352*2afbf43aSAbel Vesa };
353*2afbf43aSAbel Vesa MODULE_DEVICE_TABLE(of, qcom_ice_of_match_table);
354*2afbf43aSAbel Vesa 
355*2afbf43aSAbel Vesa static struct platform_driver qcom_ice_driver = {
356*2afbf43aSAbel Vesa 	.probe	= qcom_ice_probe,
357*2afbf43aSAbel Vesa 	.driver = {
358*2afbf43aSAbel Vesa 		.name = "qcom-ice",
359*2afbf43aSAbel Vesa 		.of_match_table = qcom_ice_of_match_table,
360*2afbf43aSAbel Vesa 	},
361*2afbf43aSAbel Vesa };
362*2afbf43aSAbel Vesa 
363*2afbf43aSAbel Vesa module_platform_driver(qcom_ice_driver);
364*2afbf43aSAbel Vesa 
365*2afbf43aSAbel Vesa MODULE_DESCRIPTION("Qualcomm Inline Crypto Engine driver");
366*2afbf43aSAbel Vesa MODULE_LICENSE("GPL");
367