xref: /openbmc/linux/drivers/crypto/intel/qat/qat_common/qat_compression.c (revision c900529f3d9161bfde5cca0754f83b4d3c3e0220)
1  // SPDX-License-Identifier: GPL-2.0-only
2  /* Copyright(c) 2022 Intel Corporation */
3  #include <linux/module.h>
4  #include <linux/slab.h>
5  #include "adf_accel_devices.h"
6  #include "adf_common_drv.h"
7  #include "adf_transport.h"
8  #include "adf_transport_access_macros.h"
9  #include "adf_cfg.h"
10  #include "adf_cfg_strings.h"
11  #include "qat_compression.h"
12  #include "icp_qat_fw.h"
13  
14  #define SEC ADF_KERNEL_SEC
15  
16  static struct service_hndl qat_compression;
17  
qat_compression_put_instance(struct qat_compression_instance * inst)18  void qat_compression_put_instance(struct qat_compression_instance *inst)
19  {
20  	atomic_dec(&inst->refctr);
21  	adf_dev_put(inst->accel_dev);
22  }
23  
qat_compression_free_instances(struct adf_accel_dev * accel_dev)24  static int qat_compression_free_instances(struct adf_accel_dev *accel_dev)
25  {
26  	struct qat_compression_instance *inst;
27  	struct list_head *list_ptr, *tmp;
28  	int i;
29  
30  	list_for_each_safe(list_ptr, tmp, &accel_dev->compression_list) {
31  		inst = list_entry(list_ptr,
32  				  struct qat_compression_instance, list);
33  
34  		for (i = 0; i < atomic_read(&inst->refctr); i++)
35  			qat_compression_put_instance(inst);
36  
37  		if (inst->dc_tx)
38  			adf_remove_ring(inst->dc_tx);
39  
40  		if (inst->dc_rx)
41  			adf_remove_ring(inst->dc_rx);
42  
43  		list_del(list_ptr);
44  		kfree(inst);
45  	}
46  	return 0;
47  }
48  
qat_compression_get_instance_node(int node)49  struct qat_compression_instance *qat_compression_get_instance_node(int node)
50  {
51  	struct qat_compression_instance *inst = NULL;
52  	struct adf_accel_dev *accel_dev = NULL;
53  	unsigned long best = ~0;
54  	struct list_head *itr;
55  
56  	list_for_each(itr, adf_devmgr_get_head()) {
57  		struct adf_accel_dev *tmp_dev;
58  		unsigned long ctr;
59  		int tmp_dev_node;
60  
61  		tmp_dev = list_entry(itr, struct adf_accel_dev, list);
62  		tmp_dev_node = dev_to_node(&GET_DEV(tmp_dev));
63  
64  		if ((node == tmp_dev_node || tmp_dev_node < 0) &&
65  		    adf_dev_started(tmp_dev) && !list_empty(&tmp_dev->compression_list)) {
66  			ctr = atomic_read(&tmp_dev->ref_count);
67  			if (best > ctr) {
68  				accel_dev = tmp_dev;
69  				best = ctr;
70  			}
71  		}
72  	}
73  
74  	if (!accel_dev) {
75  		pr_debug_ratelimited("QAT: Could not find a device on node %d\n", node);
76  		/* Get any started device */
77  		list_for_each(itr, adf_devmgr_get_head()) {
78  			struct adf_accel_dev *tmp_dev;
79  
80  			tmp_dev = list_entry(itr, struct adf_accel_dev, list);
81  			if (adf_dev_started(tmp_dev) &&
82  			    !list_empty(&tmp_dev->compression_list)) {
83  				accel_dev = tmp_dev;
84  				break;
85  			}
86  		}
87  	}
88  
89  	if (!accel_dev)
90  		return NULL;
91  
92  	best = ~0;
93  	list_for_each(itr, &accel_dev->compression_list) {
94  		struct qat_compression_instance *tmp_inst;
95  		unsigned long ctr;
96  
97  		tmp_inst = list_entry(itr, struct qat_compression_instance, list);
98  		ctr = atomic_read(&tmp_inst->refctr);
99  		if (best > ctr) {
100  			inst = tmp_inst;
101  			best = ctr;
102  		}
103  	}
104  	if (inst) {
105  		if (adf_dev_get(accel_dev)) {
106  			dev_err(&GET_DEV(accel_dev), "Could not increment dev refctr\n");
107  			return NULL;
108  		}
109  		atomic_inc(&inst->refctr);
110  	}
111  	return inst;
112  }
113  
qat_compression_create_instances(struct adf_accel_dev * accel_dev)114  static int qat_compression_create_instances(struct adf_accel_dev *accel_dev)
115  {
116  	struct qat_compression_instance *inst;
117  	char key[ADF_CFG_MAX_KEY_LEN_IN_BYTES];
118  	char val[ADF_CFG_MAX_VAL_LEN_IN_BYTES];
119  	unsigned long num_inst, num_msg_dc;
120  	unsigned long bank;
121  	int msg_size;
122  	int ret;
123  	int i;
124  
125  	INIT_LIST_HEAD(&accel_dev->compression_list);
126  	strscpy(key, ADF_NUM_DC, sizeof(key));
127  	ret = adf_cfg_get_param_value(accel_dev, SEC, key, val);
128  	if (ret)
129  		return ret;
130  
131  	ret = kstrtoul(val, 10, &num_inst);
132  	if (ret)
133  		return ret;
134  
135  	for (i = 0; i < num_inst; i++) {
136  		inst = kzalloc_node(sizeof(*inst), GFP_KERNEL,
137  				    dev_to_node(&GET_DEV(accel_dev)));
138  		if (!inst) {
139  			ret = -ENOMEM;
140  			goto err;
141  		}
142  
143  		list_add_tail(&inst->list, &accel_dev->compression_list);
144  		inst->id = i;
145  		atomic_set(&inst->refctr, 0);
146  		inst->accel_dev = accel_dev;
147  		inst->build_deflate_ctx = GET_DC_OPS(accel_dev)->build_deflate_ctx;
148  
149  		snprintf(key, sizeof(key), ADF_DC "%d" ADF_RING_DC_BANK_NUM, i);
150  		ret = adf_cfg_get_param_value(accel_dev, SEC, key, val);
151  		if (ret)
152  			return ret;
153  
154  		ret = kstrtoul(val, 10, &bank);
155  		if (ret)
156  			return ret;
157  
158  		snprintf(key, sizeof(key), ADF_DC "%d" ADF_RING_DC_SIZE, i);
159  		ret = adf_cfg_get_param_value(accel_dev, SEC, key, val);
160  		if (ret)
161  			return ret;
162  
163  		ret = kstrtoul(val, 10, &num_msg_dc);
164  		if (ret)
165  			return ret;
166  
167  		msg_size = ICP_QAT_FW_REQ_DEFAULT_SZ;
168  		snprintf(key, sizeof(key), ADF_DC "%d" ADF_RING_DC_TX, i);
169  		ret = adf_create_ring(accel_dev, SEC, bank, num_msg_dc,
170  				      msg_size, key, NULL, 0, &inst->dc_tx);
171  		if (ret)
172  			return ret;
173  
174  		msg_size = ICP_QAT_FW_RESP_DEFAULT_SZ;
175  		snprintf(key, sizeof(key), ADF_DC "%d" ADF_RING_DC_RX, i);
176  		ret = adf_create_ring(accel_dev, SEC, bank, num_msg_dc,
177  				      msg_size, key, qat_comp_alg_callback, 0,
178  				      &inst->dc_rx);
179  		if (ret)
180  			return ret;
181  
182  		inst->dc_data = accel_dev->dc_data;
183  		INIT_LIST_HEAD(&inst->backlog.list);
184  		spin_lock_init(&inst->backlog.lock);
185  	}
186  	return 0;
187  err:
188  	qat_compression_free_instances(accel_dev);
189  	return ret;
190  }
191  
qat_compression_alloc_dc_data(struct adf_accel_dev * accel_dev)192  static int qat_compression_alloc_dc_data(struct adf_accel_dev *accel_dev)
193  {
194  	struct device *dev = &GET_DEV(accel_dev);
195  	dma_addr_t obuff_p = DMA_MAPPING_ERROR;
196  	size_t ovf_buff_sz = QAT_COMP_MAX_SKID;
197  	struct adf_dc_data *dc_data = NULL;
198  	u8 *obuff = NULL;
199  
200  	dc_data = devm_kzalloc(dev, sizeof(*dc_data), GFP_KERNEL);
201  	if (!dc_data)
202  		goto err;
203  
204  	obuff = kzalloc_node(ovf_buff_sz, GFP_KERNEL, dev_to_node(dev));
205  	if (!obuff)
206  		goto err;
207  
208  	obuff_p = dma_map_single(dev, obuff, ovf_buff_sz, DMA_FROM_DEVICE);
209  	if (unlikely(dma_mapping_error(dev, obuff_p)))
210  		goto err;
211  
212  	dc_data->ovf_buff = obuff;
213  	dc_data->ovf_buff_p = obuff_p;
214  	dc_data->ovf_buff_sz = ovf_buff_sz;
215  
216  	accel_dev->dc_data = dc_data;
217  
218  	return 0;
219  
220  err:
221  	accel_dev->dc_data = NULL;
222  	kfree(obuff);
223  	devm_kfree(dev, dc_data);
224  	return -ENOMEM;
225  }
226  
qat_free_dc_data(struct adf_accel_dev * accel_dev)227  static void qat_free_dc_data(struct adf_accel_dev *accel_dev)
228  {
229  	struct adf_dc_data *dc_data = accel_dev->dc_data;
230  	struct device *dev = &GET_DEV(accel_dev);
231  
232  	if (!dc_data)
233  		return;
234  
235  	dma_unmap_single(dev, dc_data->ovf_buff_p, dc_data->ovf_buff_sz,
236  			 DMA_FROM_DEVICE);
237  	kfree_sensitive(dc_data->ovf_buff);
238  	devm_kfree(dev, dc_data);
239  	accel_dev->dc_data = NULL;
240  }
241  
qat_compression_init(struct adf_accel_dev * accel_dev)242  static int qat_compression_init(struct adf_accel_dev *accel_dev)
243  {
244  	int ret;
245  
246  	ret = qat_compression_alloc_dc_data(accel_dev);
247  	if (ret)
248  		return ret;
249  
250  	ret = qat_compression_create_instances(accel_dev);
251  	if (ret)
252  		qat_free_dc_data(accel_dev);
253  
254  	return ret;
255  }
256  
qat_compression_shutdown(struct adf_accel_dev * accel_dev)257  static int qat_compression_shutdown(struct adf_accel_dev *accel_dev)
258  {
259  	qat_free_dc_data(accel_dev);
260  	return qat_compression_free_instances(accel_dev);
261  }
262  
qat_compression_event_handler(struct adf_accel_dev * accel_dev,enum adf_event event)263  static int qat_compression_event_handler(struct adf_accel_dev *accel_dev,
264  					 enum adf_event event)
265  {
266  	int ret;
267  
268  	switch (event) {
269  	case ADF_EVENT_INIT:
270  		ret = qat_compression_init(accel_dev);
271  		break;
272  	case ADF_EVENT_SHUTDOWN:
273  		ret = qat_compression_shutdown(accel_dev);
274  		break;
275  	case ADF_EVENT_RESTARTING:
276  	case ADF_EVENT_RESTARTED:
277  	case ADF_EVENT_START:
278  	case ADF_EVENT_STOP:
279  	default:
280  		ret = 0;
281  	}
282  	return ret;
283  }
284  
qat_compression_register(void)285  int qat_compression_register(void)
286  {
287  	memset(&qat_compression, 0, sizeof(qat_compression));
288  	qat_compression.event_hld = qat_compression_event_handler;
289  	qat_compression.name = "qat_compression";
290  	return adf_service_register(&qat_compression);
291  }
292  
qat_compression_unregister(void)293  int qat_compression_unregister(void)
294  {
295  	return adf_service_unregister(&qat_compression);
296  }
297