xref: /openbmc/linux/drivers/accel/habanalabs/common/habanalabs_ioctl.c (revision 1188f7f111c61394ec56beb8e30322305a8220b6)
1  // SPDX-License-Identifier: GPL-2.0
2  
3  /*
4   * Copyright 2016-2022 HabanaLabs, Ltd.
5   * All Rights Reserved.
6   */
7  
8  #define pr_fmt(fmt)	"habanalabs: " fmt
9  
10  #include <uapi/drm/habanalabs_accel.h>
11  #include "habanalabs.h"
12  
13  #include <linux/fs.h>
14  #include <linux/kernel.h>
15  #include <linux/pci.h>
16  #include <linux/slab.h>
17  #include <linux/uaccess.h>
18  #include <linux/vmalloc.h>
19  
20  static u32 hl_debug_struct_size[HL_DEBUG_OP_TIMESTAMP + 1] = {
21  	[HL_DEBUG_OP_ETR] = sizeof(struct hl_debug_params_etr),
22  	[HL_DEBUG_OP_ETF] = sizeof(struct hl_debug_params_etf),
23  	[HL_DEBUG_OP_STM] = sizeof(struct hl_debug_params_stm),
24  	[HL_DEBUG_OP_FUNNEL] = 0,
25  	[HL_DEBUG_OP_BMON] = sizeof(struct hl_debug_params_bmon),
26  	[HL_DEBUG_OP_SPMU] = sizeof(struct hl_debug_params_spmu),
27  	[HL_DEBUG_OP_TIMESTAMP] = 0
28  
29  };
30  
device_status_info(struct hl_device * hdev,struct hl_info_args * args)31  static int device_status_info(struct hl_device *hdev, struct hl_info_args *args)
32  {
33  	struct hl_info_device_status dev_stat = {0};
34  	u32 size = args->return_size;
35  	void __user *out = (void __user *) (uintptr_t) args->return_pointer;
36  
37  	if ((!size) || (!out))
38  		return -EINVAL;
39  
40  	dev_stat.status = hl_device_status(hdev);
41  
42  	return copy_to_user(out, &dev_stat,
43  			min((size_t)size, sizeof(dev_stat))) ? -EFAULT : 0;
44  }
45  
hw_ip_info(struct hl_device * hdev,struct hl_info_args * args)46  static int hw_ip_info(struct hl_device *hdev, struct hl_info_args *args)
47  {
48  	struct hl_info_hw_ip_info hw_ip = {0};
49  	u32 size = args->return_size;
50  	void __user *out = (void __user *) (uintptr_t) args->return_pointer;
51  	struct asic_fixed_properties *prop = &hdev->asic_prop;
52  	u64 sram_kmd_size, dram_kmd_size, dram_available_size;
53  
54  	if ((!size) || (!out))
55  		return -EINVAL;
56  
57  	sram_kmd_size = (prop->sram_user_base_address -
58  				prop->sram_base_address);
59  	dram_kmd_size = (prop->dram_user_base_address -
60  				prop->dram_base_address);
61  
62  	hw_ip.device_id = hdev->asic_funcs->get_pci_id(hdev);
63  	hw_ip.sram_base_address = prop->sram_user_base_address;
64  	hw_ip.dram_base_address =
65  			prop->dram_supports_virtual_memory ?
66  			prop->dmmu.start_addr : prop->dram_user_base_address;
67  	hw_ip.tpc_enabled_mask = prop->tpc_enabled_mask & 0xFF;
68  	hw_ip.tpc_enabled_mask_ext = prop->tpc_enabled_mask;
69  
70  	hw_ip.sram_size = prop->sram_size - sram_kmd_size;
71  
72  	dram_available_size = prop->dram_size - dram_kmd_size;
73  
74  	hw_ip.dram_size = DIV_ROUND_DOWN_ULL(dram_available_size, prop->dram_page_size) *
75  				prop->dram_page_size;
76  
77  	if (hw_ip.dram_size > PAGE_SIZE)
78  		hw_ip.dram_enabled = 1;
79  
80  	hw_ip.dram_page_size = prop->dram_page_size;
81  	hw_ip.device_mem_alloc_default_page_size = prop->device_mem_alloc_default_page_size;
82  	hw_ip.num_of_events = prop->num_of_events;
83  
84  	memcpy(hw_ip.cpucp_version, prop->cpucp_info.cpucp_version,
85  		min(VERSION_MAX_LEN, HL_INFO_VERSION_MAX_LEN));
86  
87  	memcpy(hw_ip.card_name, prop->cpucp_info.card_name,
88  		min(CARD_NAME_MAX_LEN, HL_INFO_CARD_NAME_MAX_LEN));
89  
90  	hw_ip.cpld_version = le32_to_cpu(prop->cpucp_info.cpld_version);
91  	hw_ip.module_id = le32_to_cpu(prop->cpucp_info.card_location);
92  
93  	hw_ip.psoc_pci_pll_nr = prop->psoc_pci_pll_nr;
94  	hw_ip.psoc_pci_pll_nf = prop->psoc_pci_pll_nf;
95  	hw_ip.psoc_pci_pll_od = prop->psoc_pci_pll_od;
96  	hw_ip.psoc_pci_pll_div_factor = prop->psoc_pci_pll_div_factor;
97  
98  	hw_ip.decoder_enabled_mask = prop->decoder_enabled_mask;
99  	hw_ip.mme_master_slave_mode = prop->mme_master_slave_mode;
100  	hw_ip.first_available_interrupt_id = prop->first_available_user_interrupt;
101  	hw_ip.number_of_user_interrupts = prop->user_interrupt_count;
102  	hw_ip.tpc_interrupt_id = prop->tpc_interrupt_id;
103  
104  	hw_ip.edma_enabled_mask = prop->edma_enabled_mask;
105  	hw_ip.server_type = prop->server_type;
106  	hw_ip.security_enabled = prop->fw_security_enabled;
107  	hw_ip.revision_id = hdev->pdev->revision;
108  	hw_ip.rotator_enabled_mask = prop->rotator_enabled_mask;
109  	hw_ip.engine_core_interrupt_reg_addr = prop->engine_core_interrupt_reg_addr;
110  	hw_ip.reserved_dram_size = dram_kmd_size;
111  
112  	return copy_to_user(out, &hw_ip,
113  		min((size_t) size, sizeof(hw_ip))) ? -EFAULT : 0;
114  }
115  
hw_events_info(struct hl_device * hdev,bool aggregate,struct hl_info_args * args)116  static int hw_events_info(struct hl_device *hdev, bool aggregate,
117  			struct hl_info_args *args)
118  {
119  	u32 size, max_size = args->return_size;
120  	void __user *out = (void __user *) (uintptr_t) args->return_pointer;
121  	void *arr;
122  
123  	if ((!max_size) || (!out))
124  		return -EINVAL;
125  
126  	arr = hdev->asic_funcs->get_events_stat(hdev, aggregate, &size);
127  	if (!arr) {
128  		dev_err(hdev->dev, "Events info not supported\n");
129  		return -EOPNOTSUPP;
130  	}
131  
132  	return copy_to_user(out, arr, min(max_size, size)) ? -EFAULT : 0;
133  }
134  
events_info(struct hl_fpriv * hpriv,struct hl_info_args * args)135  static int events_info(struct hl_fpriv *hpriv, struct hl_info_args *args)
136  {
137  	u32 max_size = args->return_size;
138  	u64 events_mask;
139  	void __user *out = (void __user *) (uintptr_t) args->return_pointer;
140  
141  	if ((max_size < sizeof(u64)) || (!out))
142  		return -EINVAL;
143  
144  	mutex_lock(&hpriv->notifier_event.lock);
145  	events_mask = hpriv->notifier_event.events_mask;
146  	hpriv->notifier_event.events_mask = 0;
147  	mutex_unlock(&hpriv->notifier_event.lock);
148  
149  	return copy_to_user(out, &events_mask, sizeof(u64)) ? -EFAULT : 0;
150  }
151  
dram_usage_info(struct hl_fpriv * hpriv,struct hl_info_args * args)152  static int dram_usage_info(struct hl_fpriv *hpriv, struct hl_info_args *args)
153  {
154  	struct hl_device *hdev = hpriv->hdev;
155  	struct hl_info_dram_usage dram_usage = {0};
156  	u32 max_size = args->return_size;
157  	void __user *out = (void __user *) (uintptr_t) args->return_pointer;
158  	struct asic_fixed_properties *prop = &hdev->asic_prop;
159  	u64 dram_kmd_size;
160  
161  	if ((!max_size) || (!out))
162  		return -EINVAL;
163  
164  	dram_kmd_size = (prop->dram_user_base_address -
165  				prop->dram_base_address);
166  	dram_usage.dram_free_mem = (prop->dram_size - dram_kmd_size) -
167  					atomic64_read(&hdev->dram_used_mem);
168  	if (hpriv->ctx)
169  		dram_usage.ctx_dram_mem =
170  			atomic64_read(&hpriv->ctx->dram_phys_mem);
171  
172  	return copy_to_user(out, &dram_usage,
173  		min((size_t) max_size, sizeof(dram_usage))) ? -EFAULT : 0;
174  }
175  
hw_idle(struct hl_device * hdev,struct hl_info_args * args)176  static int hw_idle(struct hl_device *hdev, struct hl_info_args *args)
177  {
178  	struct hl_info_hw_idle hw_idle = {0};
179  	u32 max_size = args->return_size;
180  	void __user *out = (void __user *) (uintptr_t) args->return_pointer;
181  
182  	if ((!max_size) || (!out))
183  		return -EINVAL;
184  
185  	hw_idle.is_idle = hdev->asic_funcs->is_device_idle(hdev,
186  					hw_idle.busy_engines_mask_ext,
187  					HL_BUSY_ENGINES_MASK_EXT_SIZE, NULL);
188  	hw_idle.busy_engines_mask =
189  			lower_32_bits(hw_idle.busy_engines_mask_ext[0]);
190  
191  	return copy_to_user(out, &hw_idle,
192  		min((size_t) max_size, sizeof(hw_idle))) ? -EFAULT : 0;
193  }
194  
debug_coresight(struct hl_device * hdev,struct hl_ctx * ctx,struct hl_debug_args * args)195  static int debug_coresight(struct hl_device *hdev, struct hl_ctx *ctx, struct hl_debug_args *args)
196  {
197  	struct hl_debug_params *params;
198  	void *input = NULL, *output = NULL;
199  	int rc;
200  
201  	params = kzalloc(sizeof(*params), GFP_KERNEL);
202  	if (!params)
203  		return -ENOMEM;
204  
205  	params->reg_idx = args->reg_idx;
206  	params->enable = args->enable;
207  	params->op = args->op;
208  
209  	if (args->input_ptr && args->input_size) {
210  		input = kzalloc(hl_debug_struct_size[args->op], GFP_KERNEL);
211  		if (!input) {
212  			rc = -ENOMEM;
213  			goto out;
214  		}
215  
216  		if (copy_from_user(input, u64_to_user_ptr(args->input_ptr),
217  					args->input_size)) {
218  			rc = -EFAULT;
219  			dev_err(hdev->dev, "failed to copy input debug data\n");
220  			goto out;
221  		}
222  
223  		params->input = input;
224  	}
225  
226  	if (args->output_ptr && args->output_size) {
227  		output = kzalloc(args->output_size, GFP_KERNEL);
228  		if (!output) {
229  			rc = -ENOMEM;
230  			goto out;
231  		}
232  
233  		params->output = output;
234  		params->output_size = args->output_size;
235  	}
236  
237  	rc = hdev->asic_funcs->debug_coresight(hdev, ctx, params);
238  	if (rc) {
239  		dev_err(hdev->dev,
240  			"debug coresight operation failed %d\n", rc);
241  		goto out;
242  	}
243  
244  	if (output && copy_to_user((void __user *) (uintptr_t) args->output_ptr,
245  					output, args->output_size)) {
246  		dev_err(hdev->dev, "copy to user failed in debug ioctl\n");
247  		rc = -EFAULT;
248  		goto out;
249  	}
250  
251  
252  out:
253  	kfree(params);
254  	kfree(output);
255  	kfree(input);
256  
257  	return rc;
258  }
259  
device_utilization(struct hl_device * hdev,struct hl_info_args * args)260  static int device_utilization(struct hl_device *hdev, struct hl_info_args *args)
261  {
262  	struct hl_info_device_utilization device_util = {0};
263  	u32 max_size = args->return_size;
264  	void __user *out = (void __user *) (uintptr_t) args->return_pointer;
265  	int rc;
266  
267  	if ((!max_size) || (!out))
268  		return -EINVAL;
269  
270  	rc = hl_device_utilization(hdev, &device_util.utilization);
271  	if (rc)
272  		return -EINVAL;
273  
274  	return copy_to_user(out, &device_util,
275  		min((size_t) max_size, sizeof(device_util))) ? -EFAULT : 0;
276  }
277  
get_clk_rate(struct hl_device * hdev,struct hl_info_args * args)278  static int get_clk_rate(struct hl_device *hdev, struct hl_info_args *args)
279  {
280  	struct hl_info_clk_rate clk_rate = {0};
281  	u32 max_size = args->return_size;
282  	void __user *out = (void __user *) (uintptr_t) args->return_pointer;
283  	int rc;
284  
285  	if ((!max_size) || (!out))
286  		return -EINVAL;
287  
288  	rc = hl_fw_get_clk_rate(hdev, &clk_rate.cur_clk_rate_mhz, &clk_rate.max_clk_rate_mhz);
289  	if (rc)
290  		return rc;
291  
292  	return copy_to_user(out, &clk_rate, min_t(size_t, max_size, sizeof(clk_rate)))
293  										? -EFAULT : 0;
294  }
295  
get_reset_count(struct hl_device * hdev,struct hl_info_args * args)296  static int get_reset_count(struct hl_device *hdev, struct hl_info_args *args)
297  {
298  	struct hl_info_reset_count reset_count = {0};
299  	u32 max_size = args->return_size;
300  	void __user *out = (void __user *) (uintptr_t) args->return_pointer;
301  
302  	if ((!max_size) || (!out))
303  		return -EINVAL;
304  
305  	reset_count.hard_reset_cnt = hdev->reset_info.hard_reset_cnt;
306  	reset_count.soft_reset_cnt = hdev->reset_info.compute_reset_cnt;
307  
308  	return copy_to_user(out, &reset_count,
309  		min((size_t) max_size, sizeof(reset_count))) ? -EFAULT : 0;
310  }
311  
time_sync_info(struct hl_device * hdev,struct hl_info_args * args)312  static int time_sync_info(struct hl_device *hdev, struct hl_info_args *args)
313  {
314  	struct hl_info_time_sync time_sync = {0};
315  	u32 max_size = args->return_size;
316  	void __user *out = (void __user *) (uintptr_t) args->return_pointer;
317  
318  	if ((!max_size) || (!out))
319  		return -EINVAL;
320  
321  	time_sync.device_time = hdev->asic_funcs->get_device_time(hdev);
322  	time_sync.host_time = ktime_get_raw_ns();
323  
324  	return copy_to_user(out, &time_sync,
325  		min((size_t) max_size, sizeof(time_sync))) ? -EFAULT : 0;
326  }
327  
pci_counters_info(struct hl_fpriv * hpriv,struct hl_info_args * args)328  static int pci_counters_info(struct hl_fpriv *hpriv, struct hl_info_args *args)
329  {
330  	struct hl_device *hdev = hpriv->hdev;
331  	struct hl_info_pci_counters pci_counters = {0};
332  	u32 max_size = args->return_size;
333  	void __user *out = (void __user *) (uintptr_t) args->return_pointer;
334  	int rc;
335  
336  	if ((!max_size) || (!out))
337  		return -EINVAL;
338  
339  	rc = hl_fw_cpucp_pci_counters_get(hdev, &pci_counters);
340  	if (rc)
341  		return rc;
342  
343  	return copy_to_user(out, &pci_counters,
344  		min((size_t) max_size, sizeof(pci_counters))) ? -EFAULT : 0;
345  }
346  
clk_throttle_info(struct hl_fpriv * hpriv,struct hl_info_args * args)347  static int clk_throttle_info(struct hl_fpriv *hpriv, struct hl_info_args *args)
348  {
349  	void __user *out = (void __user *) (uintptr_t) args->return_pointer;
350  	struct hl_device *hdev = hpriv->hdev;
351  	struct hl_info_clk_throttle clk_throttle = {0};
352  	ktime_t end_time, zero_time = ktime_set(0, 0);
353  	u32 max_size = args->return_size;
354  	int i;
355  
356  	if ((!max_size) || (!out))
357  		return -EINVAL;
358  
359  	mutex_lock(&hdev->clk_throttling.lock);
360  
361  	clk_throttle.clk_throttling_reason = hdev->clk_throttling.current_reason;
362  
363  	for (i = 0 ; i < HL_CLK_THROTTLE_TYPE_MAX ; i++) {
364  		if (!(hdev->clk_throttling.aggregated_reason & BIT(i)))
365  			continue;
366  
367  		clk_throttle.clk_throttling_timestamp_us[i] =
368  			ktime_to_us(hdev->clk_throttling.timestamp[i].start);
369  
370  		if (ktime_compare(hdev->clk_throttling.timestamp[i].end, zero_time))
371  			end_time = hdev->clk_throttling.timestamp[i].end;
372  		else
373  			end_time = ktime_get();
374  
375  		clk_throttle.clk_throttling_duration_ns[i] =
376  			ktime_to_ns(ktime_sub(end_time,
377  				hdev->clk_throttling.timestamp[i].start));
378  
379  	}
380  	mutex_unlock(&hdev->clk_throttling.lock);
381  
382  	return copy_to_user(out, &clk_throttle,
383  		min((size_t) max_size, sizeof(clk_throttle))) ? -EFAULT : 0;
384  }
385  
cs_counters_info(struct hl_fpriv * hpriv,struct hl_info_args * args)386  static int cs_counters_info(struct hl_fpriv *hpriv, struct hl_info_args *args)
387  {
388  	void __user *out = (void __user *) (uintptr_t) args->return_pointer;
389  	struct hl_info_cs_counters cs_counters = {0};
390  	struct hl_device *hdev = hpriv->hdev;
391  	struct hl_cs_counters_atomic *cntr;
392  	u32 max_size = args->return_size;
393  
394  	cntr = &hdev->aggregated_cs_counters;
395  
396  	if ((!max_size) || (!out))
397  		return -EINVAL;
398  
399  	cs_counters.total_out_of_mem_drop_cnt =
400  			atomic64_read(&cntr->out_of_mem_drop_cnt);
401  	cs_counters.total_parsing_drop_cnt =
402  			atomic64_read(&cntr->parsing_drop_cnt);
403  	cs_counters.total_queue_full_drop_cnt =
404  			atomic64_read(&cntr->queue_full_drop_cnt);
405  	cs_counters.total_device_in_reset_drop_cnt =
406  			atomic64_read(&cntr->device_in_reset_drop_cnt);
407  	cs_counters.total_max_cs_in_flight_drop_cnt =
408  			atomic64_read(&cntr->max_cs_in_flight_drop_cnt);
409  	cs_counters.total_validation_drop_cnt =
410  			atomic64_read(&cntr->validation_drop_cnt);
411  
412  	if (hpriv->ctx) {
413  		cs_counters.ctx_out_of_mem_drop_cnt =
414  				atomic64_read(
415  				&hpriv->ctx->cs_counters.out_of_mem_drop_cnt);
416  		cs_counters.ctx_parsing_drop_cnt =
417  				atomic64_read(
418  				&hpriv->ctx->cs_counters.parsing_drop_cnt);
419  		cs_counters.ctx_queue_full_drop_cnt =
420  				atomic64_read(
421  				&hpriv->ctx->cs_counters.queue_full_drop_cnt);
422  		cs_counters.ctx_device_in_reset_drop_cnt =
423  				atomic64_read(
424  			&hpriv->ctx->cs_counters.device_in_reset_drop_cnt);
425  		cs_counters.ctx_max_cs_in_flight_drop_cnt =
426  				atomic64_read(
427  			&hpriv->ctx->cs_counters.max_cs_in_flight_drop_cnt);
428  		cs_counters.ctx_validation_drop_cnt =
429  				atomic64_read(
430  				&hpriv->ctx->cs_counters.validation_drop_cnt);
431  	}
432  
433  	return copy_to_user(out, &cs_counters,
434  		min((size_t) max_size, sizeof(cs_counters))) ? -EFAULT : 0;
435  }
436  
sync_manager_info(struct hl_fpriv * hpriv,struct hl_info_args * args)437  static int sync_manager_info(struct hl_fpriv *hpriv, struct hl_info_args *args)
438  {
439  	struct hl_device *hdev = hpriv->hdev;
440  	struct asic_fixed_properties *prop = &hdev->asic_prop;
441  	struct hl_info_sync_manager sm_info = {0};
442  	u32 max_size = args->return_size;
443  	void __user *out = (void __user *) (uintptr_t) args->return_pointer;
444  
445  	if ((!max_size) || (!out))
446  		return -EINVAL;
447  
448  	if (args->dcore_id >= HL_MAX_DCORES)
449  		return -EINVAL;
450  
451  	sm_info.first_available_sync_object =
452  			prop->first_available_user_sob[args->dcore_id];
453  	sm_info.first_available_monitor =
454  			prop->first_available_user_mon[args->dcore_id];
455  	sm_info.first_available_cq =
456  			prop->first_available_cq[args->dcore_id];
457  
458  	return copy_to_user(out, &sm_info, min_t(size_t, (size_t) max_size,
459  			sizeof(sm_info))) ? -EFAULT : 0;
460  }
461  
total_energy_consumption_info(struct hl_fpriv * hpriv,struct hl_info_args * args)462  static int total_energy_consumption_info(struct hl_fpriv *hpriv,
463  			struct hl_info_args *args)
464  {
465  	struct hl_device *hdev = hpriv->hdev;
466  	struct hl_info_energy total_energy = {0};
467  	u32 max_size = args->return_size;
468  	void __user *out = (void __user *) (uintptr_t) args->return_pointer;
469  	int rc;
470  
471  	if ((!max_size) || (!out))
472  		return -EINVAL;
473  
474  	rc = hl_fw_cpucp_total_energy_get(hdev,
475  			&total_energy.total_energy_consumption);
476  	if (rc)
477  		return rc;
478  
479  	return copy_to_user(out, &total_energy,
480  		min((size_t) max_size, sizeof(total_energy))) ? -EFAULT : 0;
481  }
482  
pll_frequency_info(struct hl_fpriv * hpriv,struct hl_info_args * args)483  static int pll_frequency_info(struct hl_fpriv *hpriv, struct hl_info_args *args)
484  {
485  	struct hl_device *hdev = hpriv->hdev;
486  	struct hl_pll_frequency_info freq_info = { {0} };
487  	u32 max_size = args->return_size;
488  	void __user *out = (void __user *) (uintptr_t) args->return_pointer;
489  	int rc;
490  
491  	if ((!max_size) || (!out))
492  		return -EINVAL;
493  
494  	rc = hl_fw_cpucp_pll_info_get(hdev, args->pll_index, freq_info.output);
495  	if (rc)
496  		return rc;
497  
498  	return copy_to_user(out, &freq_info,
499  		min((size_t) max_size, sizeof(freq_info))) ? -EFAULT : 0;
500  }
501  
power_info(struct hl_fpriv * hpriv,struct hl_info_args * args)502  static int power_info(struct hl_fpriv *hpriv, struct hl_info_args *args)
503  {
504  	struct hl_device *hdev = hpriv->hdev;
505  	u32 max_size = args->return_size;
506  	struct hl_power_info power_info = {0};
507  	void __user *out = (void __user *) (uintptr_t) args->return_pointer;
508  	int rc;
509  
510  	if ((!max_size) || (!out))
511  		return -EINVAL;
512  
513  	rc = hl_fw_cpucp_power_get(hdev, &power_info.power);
514  	if (rc)
515  		return rc;
516  
517  	return copy_to_user(out, &power_info,
518  		min((size_t) max_size, sizeof(power_info))) ? -EFAULT : 0;
519  }
520  
open_stats_info(struct hl_fpriv * hpriv,struct hl_info_args * args)521  static int open_stats_info(struct hl_fpriv *hpriv, struct hl_info_args *args)
522  {
523  	struct hl_device *hdev = hpriv->hdev;
524  	u32 max_size = args->return_size;
525  	struct hl_open_stats_info open_stats_info = {0};
526  	void __user *out = (void __user *) (uintptr_t) args->return_pointer;
527  
528  	if ((!max_size) || (!out))
529  		return -EINVAL;
530  
531  	open_stats_info.last_open_period_ms = jiffies64_to_msecs(
532  		hdev->last_open_session_duration_jif);
533  	open_stats_info.open_counter = hdev->open_counter;
534  	open_stats_info.is_compute_ctx_active = hdev->is_compute_ctx_active;
535  	open_stats_info.compute_ctx_in_release = hdev->compute_ctx_in_release;
536  
537  	return copy_to_user(out, &open_stats_info,
538  		min((size_t) max_size, sizeof(open_stats_info))) ? -EFAULT : 0;
539  }
540  
dram_pending_rows_info(struct hl_fpriv * hpriv,struct hl_info_args * args)541  static int dram_pending_rows_info(struct hl_fpriv *hpriv, struct hl_info_args *args)
542  {
543  	struct hl_device *hdev = hpriv->hdev;
544  	u32 max_size = args->return_size;
545  	u32 pend_rows_num = 0;
546  	void __user *out = (void __user *) (uintptr_t) args->return_pointer;
547  	int rc;
548  
549  	if ((!max_size) || (!out))
550  		return -EINVAL;
551  
552  	rc = hl_fw_dram_pending_row_get(hdev, &pend_rows_num);
553  	if (rc)
554  		return rc;
555  
556  	return copy_to_user(out, &pend_rows_num,
557  			min_t(size_t, max_size, sizeof(pend_rows_num))) ? -EFAULT : 0;
558  }
559  
dram_replaced_rows_info(struct hl_fpriv * hpriv,struct hl_info_args * args)560  static int dram_replaced_rows_info(struct hl_fpriv *hpriv, struct hl_info_args *args)
561  {
562  	struct hl_device *hdev = hpriv->hdev;
563  	u32 max_size = args->return_size;
564  	struct cpucp_hbm_row_info info = {0};
565  	void __user *out = (void __user *) (uintptr_t) args->return_pointer;
566  	int rc;
567  
568  	if ((!max_size) || (!out))
569  		return -EINVAL;
570  
571  	rc = hl_fw_dram_replaced_row_get(hdev, &info);
572  	if (rc)
573  		return rc;
574  
575  	return copy_to_user(out, &info, min_t(size_t, max_size, sizeof(info))) ? -EFAULT : 0;
576  }
577  
last_err_open_dev_info(struct hl_fpriv * hpriv,struct hl_info_args * args)578  static int last_err_open_dev_info(struct hl_fpriv *hpriv, struct hl_info_args *args)
579  {
580  	struct hl_info_last_err_open_dev_time info = {0};
581  	struct hl_device *hdev = hpriv->hdev;
582  	u32 max_size = args->return_size;
583  	void __user *out = (void __user *) (uintptr_t) args->return_pointer;
584  
585  	if ((!max_size) || (!out))
586  		return -EINVAL;
587  
588  	info.timestamp = ktime_to_ns(hdev->last_successful_open_ktime);
589  
590  	return copy_to_user(out, &info, min_t(size_t, max_size, sizeof(info))) ? -EFAULT : 0;
591  }
592  
cs_timeout_info(struct hl_fpriv * hpriv,struct hl_info_args * args)593  static int cs_timeout_info(struct hl_fpriv *hpriv, struct hl_info_args *args)
594  {
595  	struct hl_info_cs_timeout_event info = {0};
596  	struct hl_device *hdev = hpriv->hdev;
597  	u32 max_size = args->return_size;
598  	void __user *out = (void __user *) (uintptr_t) args->return_pointer;
599  
600  	if ((!max_size) || (!out))
601  		return -EINVAL;
602  
603  	info.seq = hdev->captured_err_info.cs_timeout.seq;
604  	info.timestamp = ktime_to_ns(hdev->captured_err_info.cs_timeout.timestamp);
605  
606  	return copy_to_user(out, &info, min_t(size_t, max_size, sizeof(info))) ? -EFAULT : 0;
607  }
608  
razwi_info(struct hl_fpriv * hpriv,struct hl_info_args * args)609  static int razwi_info(struct hl_fpriv *hpriv, struct hl_info_args *args)
610  {
611  	void __user *out = (void __user *) (uintptr_t) args->return_pointer;
612  	struct hl_device *hdev = hpriv->hdev;
613  	u32 max_size = args->return_size;
614  	struct razwi_info *razwi_info;
615  
616  	if ((!max_size) || (!out))
617  		return -EINVAL;
618  
619  	razwi_info = &hdev->captured_err_info.razwi_info;
620  	if (!razwi_info->razwi_info_available)
621  		return 0;
622  
623  	return copy_to_user(out, &razwi_info->razwi,
624  			min_t(size_t, max_size, sizeof(struct hl_info_razwi_event))) ? -EFAULT : 0;
625  }
626  
undefined_opcode_info(struct hl_fpriv * hpriv,struct hl_info_args * args)627  static int undefined_opcode_info(struct hl_fpriv *hpriv, struct hl_info_args *args)
628  {
629  	struct hl_device *hdev = hpriv->hdev;
630  	u32 max_size = args->return_size;
631  	struct hl_info_undefined_opcode_event info = {0};
632  	void __user *out = (void __user *) (uintptr_t) args->return_pointer;
633  
634  	if ((!max_size) || (!out))
635  		return -EINVAL;
636  
637  	info.timestamp = ktime_to_ns(hdev->captured_err_info.undef_opcode.timestamp);
638  	info.engine_id = hdev->captured_err_info.undef_opcode.engine_id;
639  	info.cq_addr = hdev->captured_err_info.undef_opcode.cq_addr;
640  	info.cq_size = hdev->captured_err_info.undef_opcode.cq_size;
641  	info.stream_id = hdev->captured_err_info.undef_opcode.stream_id;
642  	info.cb_addr_streams_len = hdev->captured_err_info.undef_opcode.cb_addr_streams_len;
643  	memcpy(info.cb_addr_streams, hdev->captured_err_info.undef_opcode.cb_addr_streams,
644  			sizeof(info.cb_addr_streams));
645  
646  	return copy_to_user(out, &info, min_t(size_t, max_size, sizeof(info))) ? -EFAULT : 0;
647  }
648  
dev_mem_alloc_page_sizes_info(struct hl_fpriv * hpriv,struct hl_info_args * args)649  static int dev_mem_alloc_page_sizes_info(struct hl_fpriv *hpriv, struct hl_info_args *args)
650  {
651  	void __user *out = (void __user *) (uintptr_t) args->return_pointer;
652  	struct hl_info_dev_memalloc_page_sizes info = {0};
653  	struct hl_device *hdev = hpriv->hdev;
654  	u32 max_size = args->return_size;
655  
656  	if ((!max_size) || (!out))
657  		return -EINVAL;
658  
659  	/*
660  	 * Future ASICs that will support multiple DRAM page sizes will support only "powers of 2"
661  	 * pages (unlike some of the ASICs before supporting multiple page sizes).
662  	 * For this reason for all ASICs that not support multiple page size the function will
663  	 * return an empty bitmask indicating that multiple page sizes is not supported.
664  	 */
665  	info.page_order_bitmask = hdev->asic_prop.dmmu.supported_pages_mask;
666  
667  	return copy_to_user(out, &info, min_t(size_t, max_size, sizeof(info))) ? -EFAULT : 0;
668  }
669  
sec_attest_info(struct hl_fpriv * hpriv,struct hl_info_args * args)670  static int sec_attest_info(struct hl_fpriv *hpriv, struct hl_info_args *args)
671  {
672  	void __user *out = (void __user *) (uintptr_t) args->return_pointer;
673  	struct cpucp_sec_attest_info *sec_attest_info;
674  	struct hl_info_sec_attest *info;
675  	u32 max_size = args->return_size;
676  	int rc;
677  
678  	if ((!max_size) || (!out))
679  		return -EINVAL;
680  
681  	sec_attest_info = kmalloc(sizeof(*sec_attest_info), GFP_KERNEL);
682  	if (!sec_attest_info)
683  		return -ENOMEM;
684  
685  	info = kzalloc(sizeof(*info), GFP_KERNEL);
686  	if (!info) {
687  		rc = -ENOMEM;
688  		goto free_sec_attest_info;
689  	}
690  
691  	rc = hl_fw_get_sec_attest_info(hpriv->hdev, sec_attest_info, args->sec_attest_nonce);
692  	if (rc)
693  		goto free_info;
694  
695  	info->nonce = le32_to_cpu(sec_attest_info->nonce);
696  	info->pcr_quote_len = le16_to_cpu(sec_attest_info->pcr_quote_len);
697  	info->pub_data_len = le16_to_cpu(sec_attest_info->pub_data_len);
698  	info->certificate_len = le16_to_cpu(sec_attest_info->certificate_len);
699  	info->pcr_num_reg = sec_attest_info->pcr_num_reg;
700  	info->pcr_reg_len = sec_attest_info->pcr_reg_len;
701  	info->quote_sig_len = sec_attest_info->quote_sig_len;
702  	memcpy(&info->pcr_data, &sec_attest_info->pcr_data, sizeof(info->pcr_data));
703  	memcpy(&info->pcr_quote, &sec_attest_info->pcr_quote, sizeof(info->pcr_quote));
704  	memcpy(&info->public_data, &sec_attest_info->public_data, sizeof(info->public_data));
705  	memcpy(&info->certificate, &sec_attest_info->certificate, sizeof(info->certificate));
706  	memcpy(&info->quote_sig, &sec_attest_info->quote_sig, sizeof(info->quote_sig));
707  
708  	rc = copy_to_user(out, info,
709  				min_t(size_t, max_size, sizeof(*info))) ? -EFAULT : 0;
710  
711  free_info:
712  	kfree(info);
713  free_sec_attest_info:
714  	kfree(sec_attest_info);
715  
716  	return rc;
717  }
718  
eventfd_register(struct hl_fpriv * hpriv,struct hl_info_args * args)719  static int eventfd_register(struct hl_fpriv *hpriv, struct hl_info_args *args)
720  {
721  	int rc;
722  
723  	/* check if there is already a registered on that process */
724  	mutex_lock(&hpriv->notifier_event.lock);
725  	if (hpriv->notifier_event.eventfd) {
726  		mutex_unlock(&hpriv->notifier_event.lock);
727  		return -EINVAL;
728  	}
729  
730  	hpriv->notifier_event.eventfd = eventfd_ctx_fdget(args->eventfd);
731  	if (IS_ERR(hpriv->notifier_event.eventfd)) {
732  		rc = PTR_ERR(hpriv->notifier_event.eventfd);
733  		hpriv->notifier_event.eventfd = NULL;
734  		mutex_unlock(&hpriv->notifier_event.lock);
735  		return rc;
736  	}
737  
738  	mutex_unlock(&hpriv->notifier_event.lock);
739  	return 0;
740  }
741  
eventfd_unregister(struct hl_fpriv * hpriv,struct hl_info_args * args)742  static int eventfd_unregister(struct hl_fpriv *hpriv, struct hl_info_args *args)
743  {
744  	mutex_lock(&hpriv->notifier_event.lock);
745  	if (!hpriv->notifier_event.eventfd) {
746  		mutex_unlock(&hpriv->notifier_event.lock);
747  		return -EINVAL;
748  	}
749  
750  	eventfd_ctx_put(hpriv->notifier_event.eventfd);
751  	hpriv->notifier_event.eventfd = NULL;
752  	mutex_unlock(&hpriv->notifier_event.lock);
753  	return 0;
754  }
755  
engine_status_info(struct hl_fpriv * hpriv,struct hl_info_args * args)756  static int engine_status_info(struct hl_fpriv *hpriv, struct hl_info_args *args)
757  {
758  	void __user *out = (void __user *) (uintptr_t) args->return_pointer;
759  	u32 status_buf_size = args->return_size;
760  	struct hl_device *hdev = hpriv->hdev;
761  	struct engines_data eng_data;
762  	int rc;
763  
764  	if ((status_buf_size < SZ_1K) || (status_buf_size > HL_ENGINES_DATA_MAX_SIZE) || (!out))
765  		return -EINVAL;
766  
767  	eng_data.actual_size = 0;
768  	eng_data.allocated_buf_size = status_buf_size;
769  	eng_data.buf = vmalloc(status_buf_size);
770  	if (!eng_data.buf)
771  		return -ENOMEM;
772  
773  	hdev->asic_funcs->is_device_idle(hdev, NULL, 0, &eng_data);
774  
775  	if (eng_data.actual_size > eng_data.allocated_buf_size) {
776  		dev_err(hdev->dev,
777  			"Engines data size (%d Bytes) is bigger than allocated size (%u Bytes)\n",
778  			eng_data.actual_size, status_buf_size);
779  		vfree(eng_data.buf);
780  		return -ENOMEM;
781  	}
782  
783  	args->user_buffer_actual_size = eng_data.actual_size;
784  	rc = copy_to_user(out, eng_data.buf, min_t(size_t, status_buf_size, eng_data.actual_size)) ?
785  				-EFAULT : 0;
786  
787  	vfree(eng_data.buf);
788  
789  	return rc;
790  }
791  
page_fault_info(struct hl_fpriv * hpriv,struct hl_info_args * args)792  static int page_fault_info(struct hl_fpriv *hpriv, struct hl_info_args *args)
793  {
794  	void __user *out = (void __user *) (uintptr_t) args->return_pointer;
795  	struct hl_device *hdev = hpriv->hdev;
796  	u32 max_size = args->return_size;
797  	struct page_fault_info *pgf_info;
798  
799  	if ((!max_size) || (!out))
800  		return -EINVAL;
801  
802  	pgf_info = &hdev->captured_err_info.page_fault_info;
803  	if (!pgf_info->page_fault_info_available)
804  		return 0;
805  
806  	return copy_to_user(out, &pgf_info->page_fault,
807  			min_t(size_t, max_size, sizeof(struct hl_page_fault_info))) ? -EFAULT : 0;
808  }
809  
user_mappings_info(struct hl_fpriv * hpriv,struct hl_info_args * args)810  static int user_mappings_info(struct hl_fpriv *hpriv, struct hl_info_args *args)
811  {
812  	void __user *out = (void __user *) (uintptr_t) args->return_pointer;
813  	u32 user_buf_size = args->return_size;
814  	struct hl_device *hdev = hpriv->hdev;
815  	struct page_fault_info *pgf_info;
816  	u64 actual_size;
817  
818  	if (!out)
819  		return -EINVAL;
820  
821  	pgf_info = &hdev->captured_err_info.page_fault_info;
822  	if (!pgf_info->page_fault_info_available)
823  		return 0;
824  
825  	args->array_size = pgf_info->num_of_user_mappings;
826  
827  	actual_size = pgf_info->num_of_user_mappings * sizeof(struct hl_user_mapping);
828  	if (user_buf_size < actual_size)
829  		return -ENOMEM;
830  
831  	return copy_to_user(out, pgf_info->user_mappings, actual_size) ? -EFAULT : 0;
832  }
833  
hw_err_info(struct hl_fpriv * hpriv,struct hl_info_args * args)834  static int hw_err_info(struct hl_fpriv *hpriv, struct hl_info_args *args)
835  {
836  	void __user *user_buf = (void __user *) (uintptr_t) args->return_pointer;
837  	struct hl_device *hdev = hpriv->hdev;
838  	u32 user_buf_size = args->return_size;
839  	struct hw_err_info *info;
840  	int rc;
841  
842  	if (!user_buf)
843  		return -EINVAL;
844  
845  	info = &hdev->captured_err_info.hw_err;
846  	if (!info->event_info_available)
847  		return 0;
848  
849  	if (user_buf_size < sizeof(struct hl_info_hw_err_event))
850  		return -ENOMEM;
851  
852  	rc = copy_to_user(user_buf, &info->event, sizeof(struct hl_info_hw_err_event));
853  	return rc ? -EFAULT : 0;
854  }
855  
fw_err_info(struct hl_fpriv * hpriv,struct hl_info_args * args)856  static int fw_err_info(struct hl_fpriv *hpriv, struct hl_info_args *args)
857  {
858  	void __user *user_buf = (void __user *) (uintptr_t) args->return_pointer;
859  	struct hl_device *hdev = hpriv->hdev;
860  	u32 user_buf_size = args->return_size;
861  	struct fw_err_info *info;
862  	int rc;
863  
864  	if (!user_buf)
865  		return -EINVAL;
866  
867  	info = &hdev->captured_err_info.fw_err;
868  	if (!info->event_info_available)
869  		return 0;
870  
871  	if (user_buf_size < sizeof(struct hl_info_fw_err_event))
872  		return -ENOMEM;
873  
874  	rc = copy_to_user(user_buf, &info->event, sizeof(struct hl_info_fw_err_event));
875  	return rc ? -EFAULT : 0;
876  }
877  
send_fw_generic_request(struct hl_device * hdev,struct hl_info_args * info_args)878  static int send_fw_generic_request(struct hl_device *hdev, struct hl_info_args *info_args)
879  {
880  	void __user *buff = (void __user *) (uintptr_t) info_args->return_pointer;
881  	u32 size = info_args->return_size;
882  	dma_addr_t dma_handle;
883  	bool need_input_buff;
884  	void *fw_buff;
885  	int rc = 0;
886  
887  	switch (info_args->fw_sub_opcode) {
888  	case HL_PASSTHROUGH_VERSIONS:
889  		need_input_buff = false;
890  		break;
891  	default:
892  		return -EINVAL;
893  	}
894  
895  	if (size > SZ_1M) {
896  		dev_err(hdev->dev, "buffer size cannot exceed 1MB\n");
897  		return -EINVAL;
898  	}
899  
900  	fw_buff = hl_cpu_accessible_dma_pool_alloc(hdev, size, &dma_handle);
901  	if (!fw_buff)
902  		return -ENOMEM;
903  
904  
905  	if (need_input_buff && copy_from_user(fw_buff, buff, size)) {
906  		dev_dbg(hdev->dev, "Failed to copy from user FW buff\n");
907  		rc = -EFAULT;
908  		goto free_buff;
909  	}
910  
911  	rc = hl_fw_send_generic_request(hdev, info_args->fw_sub_opcode, dma_handle, &size);
912  	if (rc)
913  		goto free_buff;
914  
915  	if (copy_to_user(buff, fw_buff, min(size, info_args->return_size))) {
916  		dev_dbg(hdev->dev, "Failed to copy to user FW generic req output\n");
917  		rc = -EFAULT;
918  	}
919  
920  free_buff:
921  	hl_cpu_accessible_dma_pool_free(hdev, info_args->return_size, fw_buff);
922  
923  	return rc;
924  }
925  
_hl_info_ioctl(struct hl_fpriv * hpriv,void * data,struct device * dev)926  static int _hl_info_ioctl(struct hl_fpriv *hpriv, void *data,
927  				struct device *dev)
928  {
929  	enum hl_device_status status;
930  	struct hl_info_args *args = data;
931  	struct hl_device *hdev = hpriv->hdev;
932  	int rc;
933  
934  	if (args->pad) {
935  		dev_dbg(hdev->dev, "Padding bytes must be 0\n");
936  		return -EINVAL;
937  	}
938  
939  	/*
940  	 * Information is returned for the following opcodes even if the device
941  	 * is disabled or in reset.
942  	 */
943  	switch (args->op) {
944  	case HL_INFO_HW_IP_INFO:
945  		return hw_ip_info(hdev, args);
946  
947  	case HL_INFO_DEVICE_STATUS:
948  		return device_status_info(hdev, args);
949  
950  	case HL_INFO_RESET_COUNT:
951  		return get_reset_count(hdev, args);
952  
953  	case HL_INFO_HW_EVENTS:
954  		return hw_events_info(hdev, false, args);
955  
956  	case HL_INFO_HW_EVENTS_AGGREGATE:
957  		return hw_events_info(hdev, true, args);
958  
959  	case HL_INFO_CS_COUNTERS:
960  		return cs_counters_info(hpriv, args);
961  
962  	case HL_INFO_CLK_THROTTLE_REASON:
963  		return clk_throttle_info(hpriv, args);
964  
965  	case HL_INFO_SYNC_MANAGER:
966  		return sync_manager_info(hpriv, args);
967  
968  	case HL_INFO_OPEN_STATS:
969  		return open_stats_info(hpriv, args);
970  
971  	case HL_INFO_LAST_ERR_OPEN_DEV_TIME:
972  		return last_err_open_dev_info(hpriv, args);
973  
974  	case HL_INFO_CS_TIMEOUT_EVENT:
975  		return cs_timeout_info(hpriv, args);
976  
977  	case HL_INFO_RAZWI_EVENT:
978  		return razwi_info(hpriv, args);
979  
980  	case HL_INFO_UNDEFINED_OPCODE_EVENT:
981  		return undefined_opcode_info(hpriv, args);
982  
983  	case HL_INFO_DEV_MEM_ALLOC_PAGE_SIZES:
984  		return dev_mem_alloc_page_sizes_info(hpriv, args);
985  
986  	case HL_INFO_GET_EVENTS:
987  		return events_info(hpriv, args);
988  
989  	case HL_INFO_PAGE_FAULT_EVENT:
990  		return page_fault_info(hpriv, args);
991  
992  	case HL_INFO_USER_MAPPINGS:
993  		return user_mappings_info(hpriv, args);
994  
995  	case HL_INFO_UNREGISTER_EVENTFD:
996  		return eventfd_unregister(hpriv, args);
997  
998  	case HL_INFO_HW_ERR_EVENT:
999  		return hw_err_info(hpriv, args);
1000  
1001  	case HL_INFO_FW_ERR_EVENT:
1002  		return fw_err_info(hpriv, args);
1003  
1004  	case HL_INFO_DRAM_USAGE:
1005  		return dram_usage_info(hpriv, args);
1006  	default:
1007  		break;
1008  	}
1009  
1010  	if (!hl_device_operational(hdev, &status)) {
1011  		dev_dbg_ratelimited(dev,
1012  			"Device is %s. Can't execute INFO IOCTL\n",
1013  			hdev->status[status]);
1014  		return -EBUSY;
1015  	}
1016  
1017  	switch (args->op) {
1018  	case HL_INFO_HW_IDLE:
1019  		rc = hw_idle(hdev, args);
1020  		break;
1021  
1022  	case HL_INFO_DEVICE_UTILIZATION:
1023  		rc = device_utilization(hdev, args);
1024  		break;
1025  
1026  	case HL_INFO_CLK_RATE:
1027  		rc = get_clk_rate(hdev, args);
1028  		break;
1029  
1030  	case HL_INFO_TIME_SYNC:
1031  		return time_sync_info(hdev, args);
1032  
1033  	case HL_INFO_PCI_COUNTERS:
1034  		return pci_counters_info(hpriv, args);
1035  
1036  	case HL_INFO_TOTAL_ENERGY:
1037  		return total_energy_consumption_info(hpriv, args);
1038  
1039  	case HL_INFO_PLL_FREQUENCY:
1040  		return pll_frequency_info(hpriv, args);
1041  
1042  	case HL_INFO_POWER:
1043  		return power_info(hpriv, args);
1044  
1045  
1046  	case HL_INFO_DRAM_REPLACED_ROWS:
1047  		return dram_replaced_rows_info(hpriv, args);
1048  
1049  	case HL_INFO_DRAM_PENDING_ROWS:
1050  		return dram_pending_rows_info(hpriv, args);
1051  
1052  	case HL_INFO_SECURED_ATTESTATION:
1053  		return sec_attest_info(hpriv, args);
1054  
1055  	case HL_INFO_REGISTER_EVENTFD:
1056  		return eventfd_register(hpriv, args);
1057  
1058  	case HL_INFO_ENGINE_STATUS:
1059  		return engine_status_info(hpriv, args);
1060  
1061  	case HL_INFO_FW_GENERIC_REQ:
1062  		return send_fw_generic_request(hdev, args);
1063  
1064  	default:
1065  		dev_err(dev, "Invalid request %d\n", args->op);
1066  		rc = -EINVAL;
1067  		break;
1068  	}
1069  
1070  	return rc;
1071  }
1072  
hl_info_ioctl(struct hl_fpriv * hpriv,void * data)1073  static int hl_info_ioctl(struct hl_fpriv *hpriv, void *data)
1074  {
1075  	return _hl_info_ioctl(hpriv, data, hpriv->hdev->dev);
1076  }
1077  
hl_info_ioctl_control(struct hl_fpriv * hpriv,void * data)1078  static int hl_info_ioctl_control(struct hl_fpriv *hpriv, void *data)
1079  {
1080  	return _hl_info_ioctl(hpriv, data, hpriv->hdev->dev_ctrl);
1081  }
1082  
hl_debug_ioctl(struct hl_fpriv * hpriv,void * data)1083  static int hl_debug_ioctl(struct hl_fpriv *hpriv, void *data)
1084  {
1085  	struct hl_debug_args *args = data;
1086  	struct hl_device *hdev = hpriv->hdev;
1087  	enum hl_device_status status;
1088  
1089  	int rc = 0;
1090  
1091  	if (!hl_device_operational(hdev, &status)) {
1092  		dev_dbg_ratelimited(hdev->dev,
1093  			"Device is %s. Can't execute DEBUG IOCTL\n",
1094  			hdev->status[status]);
1095  		return -EBUSY;
1096  	}
1097  
1098  	switch (args->op) {
1099  	case HL_DEBUG_OP_ETR:
1100  	case HL_DEBUG_OP_ETF:
1101  	case HL_DEBUG_OP_STM:
1102  	case HL_DEBUG_OP_FUNNEL:
1103  	case HL_DEBUG_OP_BMON:
1104  	case HL_DEBUG_OP_SPMU:
1105  	case HL_DEBUG_OP_TIMESTAMP:
1106  		if (!hdev->in_debug) {
1107  			dev_err_ratelimited(hdev->dev,
1108  				"Rejecting debug configuration request because device not in debug mode\n");
1109  			return -EFAULT;
1110  		}
1111  		args->input_size = min(args->input_size, hl_debug_struct_size[args->op]);
1112  		rc = debug_coresight(hdev, hpriv->ctx, args);
1113  		break;
1114  
1115  	case HL_DEBUG_OP_SET_MODE:
1116  		rc = hl_device_set_debug_mode(hdev, hpriv->ctx, (bool) args->enable);
1117  		break;
1118  
1119  	default:
1120  		dev_err(hdev->dev, "Invalid request %d\n", args->op);
1121  		rc = -EINVAL;
1122  		break;
1123  	}
1124  
1125  	return rc;
1126  }
1127  
1128  #define HL_IOCTL_DEF(ioctl, _func) \
1129  	[_IOC_NR(ioctl)] = {.cmd = ioctl, .func = _func}
1130  
1131  static const struct hl_ioctl_desc hl_ioctls[] = {
1132  	HL_IOCTL_DEF(HL_IOCTL_INFO, hl_info_ioctl),
1133  	HL_IOCTL_DEF(HL_IOCTL_CB, hl_cb_ioctl),
1134  	HL_IOCTL_DEF(HL_IOCTL_CS, hl_cs_ioctl),
1135  	HL_IOCTL_DEF(HL_IOCTL_WAIT_CS, hl_wait_ioctl),
1136  	HL_IOCTL_DEF(HL_IOCTL_MEMORY, hl_mem_ioctl),
1137  	HL_IOCTL_DEF(HL_IOCTL_DEBUG, hl_debug_ioctl)
1138  };
1139  
1140  static const struct hl_ioctl_desc hl_ioctls_control[] = {
1141  	HL_IOCTL_DEF(HL_IOCTL_INFO, hl_info_ioctl_control)
1142  };
1143  
_hl_ioctl(struct file * filep,unsigned int cmd,unsigned long arg,const struct hl_ioctl_desc * ioctl,struct device * dev)1144  static long _hl_ioctl(struct file *filep, unsigned int cmd, unsigned long arg,
1145  		const struct hl_ioctl_desc *ioctl, struct device *dev)
1146  {
1147  	struct hl_fpriv *hpriv = filep->private_data;
1148  	unsigned int nr = _IOC_NR(cmd);
1149  	char stack_kdata[128] = {0};
1150  	char *kdata = NULL;
1151  	unsigned int usize, asize;
1152  	hl_ioctl_t *func;
1153  	u32 hl_size;
1154  	int retcode;
1155  
1156  	/* Do not trust userspace, use our own definition */
1157  	func = ioctl->func;
1158  
1159  	if (unlikely(!func)) {
1160  		dev_dbg(dev, "no function\n");
1161  		retcode = -ENOTTY;
1162  		goto out_err;
1163  	}
1164  
1165  	hl_size = _IOC_SIZE(ioctl->cmd);
1166  	usize = asize = _IOC_SIZE(cmd);
1167  	if (hl_size > asize)
1168  		asize = hl_size;
1169  
1170  	cmd = ioctl->cmd;
1171  
1172  	if (cmd & (IOC_IN | IOC_OUT)) {
1173  		if (asize <= sizeof(stack_kdata)) {
1174  			kdata = stack_kdata;
1175  		} else {
1176  			kdata = kzalloc(asize, GFP_KERNEL);
1177  			if (!kdata) {
1178  				retcode = -ENOMEM;
1179  				goto out_err;
1180  			}
1181  		}
1182  	}
1183  
1184  	if (cmd & IOC_IN) {
1185  		if (copy_from_user(kdata, (void __user *)arg, usize)) {
1186  			retcode = -EFAULT;
1187  			goto out_err;
1188  		}
1189  	}
1190  
1191  	retcode = func(hpriv, kdata);
1192  
1193  	if ((cmd & IOC_OUT) && copy_to_user((void __user *)arg, kdata, usize))
1194  		retcode = -EFAULT;
1195  
1196  out_err:
1197  	if (retcode)
1198  		dev_dbg_ratelimited(dev, "error in ioctl: pid=%d, cmd=0x%02x, nr=0x%02x\n",
1199  			  task_pid_nr(current), cmd, nr);
1200  
1201  	if (kdata != stack_kdata)
1202  		kfree(kdata);
1203  
1204  	return retcode;
1205  }
1206  
hl_ioctl(struct file * filep,unsigned int cmd,unsigned long arg)1207  long hl_ioctl(struct file *filep, unsigned int cmd, unsigned long arg)
1208  {
1209  	struct hl_fpriv *hpriv = filep->private_data;
1210  	struct hl_device *hdev = hpriv->hdev;
1211  	const struct hl_ioctl_desc *ioctl = NULL;
1212  	unsigned int nr = _IOC_NR(cmd);
1213  
1214  	if (!hdev) {
1215  		pr_err_ratelimited("Sending ioctl after device was removed! Please close FD\n");
1216  		return -ENODEV;
1217  	}
1218  
1219  	if ((nr >= HL_COMMAND_START) && (nr < HL_COMMAND_END)) {
1220  		ioctl = &hl_ioctls[nr];
1221  	} else {
1222  		dev_dbg_ratelimited(hdev->dev, "invalid ioctl: pid=%d, nr=0x%02x\n",
1223  			task_pid_nr(current), nr);
1224  		return -ENOTTY;
1225  	}
1226  
1227  	return _hl_ioctl(filep, cmd, arg, ioctl, hdev->dev);
1228  }
1229  
hl_ioctl_control(struct file * filep,unsigned int cmd,unsigned long arg)1230  long hl_ioctl_control(struct file *filep, unsigned int cmd, unsigned long arg)
1231  {
1232  	struct hl_fpriv *hpriv = filep->private_data;
1233  	struct hl_device *hdev = hpriv->hdev;
1234  	const struct hl_ioctl_desc *ioctl = NULL;
1235  	unsigned int nr = _IOC_NR(cmd);
1236  
1237  	if (!hdev) {
1238  		pr_err_ratelimited("Sending ioctl after device was removed! Please close FD\n");
1239  		return -ENODEV;
1240  	}
1241  
1242  	if (nr == _IOC_NR(HL_IOCTL_INFO)) {
1243  		ioctl = &hl_ioctls_control[nr];
1244  	} else {
1245  		dev_dbg_ratelimited(hdev->dev_ctrl, "invalid ioctl: pid=%d, nr=0x%02x\n",
1246  			task_pid_nr(current), nr);
1247  		return -ENOTTY;
1248  	}
1249  
1250  	return _hl_ioctl(filep, cmd, arg, ioctl, hdev->dev_ctrl);
1251  }
1252