1 // SPDX-License-Identifier: GPL-2.0
2
3 /*
4 * Copyright 2016-2022 HabanaLabs, Ltd.
5 * All Rights Reserved.
6 */
7
8 #define pr_fmt(fmt) "habanalabs: " fmt
9
10 #include <uapi/drm/habanalabs_accel.h>
11 #include "habanalabs.h"
12
13 #include <linux/fs.h>
14 #include <linux/kernel.h>
15 #include <linux/pci.h>
16 #include <linux/slab.h>
17 #include <linux/uaccess.h>
18 #include <linux/vmalloc.h>
19
20 static u32 hl_debug_struct_size[HL_DEBUG_OP_TIMESTAMP + 1] = {
21 [HL_DEBUG_OP_ETR] = sizeof(struct hl_debug_params_etr),
22 [HL_DEBUG_OP_ETF] = sizeof(struct hl_debug_params_etf),
23 [HL_DEBUG_OP_STM] = sizeof(struct hl_debug_params_stm),
24 [HL_DEBUG_OP_FUNNEL] = 0,
25 [HL_DEBUG_OP_BMON] = sizeof(struct hl_debug_params_bmon),
26 [HL_DEBUG_OP_SPMU] = sizeof(struct hl_debug_params_spmu),
27 [HL_DEBUG_OP_TIMESTAMP] = 0
28
29 };
30
device_status_info(struct hl_device * hdev,struct hl_info_args * args)31 static int device_status_info(struct hl_device *hdev, struct hl_info_args *args)
32 {
33 struct hl_info_device_status dev_stat = {0};
34 u32 size = args->return_size;
35 void __user *out = (void __user *) (uintptr_t) args->return_pointer;
36
37 if ((!size) || (!out))
38 return -EINVAL;
39
40 dev_stat.status = hl_device_status(hdev);
41
42 return copy_to_user(out, &dev_stat,
43 min((size_t)size, sizeof(dev_stat))) ? -EFAULT : 0;
44 }
45
hw_ip_info(struct hl_device * hdev,struct hl_info_args * args)46 static int hw_ip_info(struct hl_device *hdev, struct hl_info_args *args)
47 {
48 struct hl_info_hw_ip_info hw_ip = {0};
49 u32 size = args->return_size;
50 void __user *out = (void __user *) (uintptr_t) args->return_pointer;
51 struct asic_fixed_properties *prop = &hdev->asic_prop;
52 u64 sram_kmd_size, dram_kmd_size, dram_available_size;
53
54 if ((!size) || (!out))
55 return -EINVAL;
56
57 sram_kmd_size = (prop->sram_user_base_address -
58 prop->sram_base_address);
59 dram_kmd_size = (prop->dram_user_base_address -
60 prop->dram_base_address);
61
62 hw_ip.device_id = hdev->asic_funcs->get_pci_id(hdev);
63 hw_ip.sram_base_address = prop->sram_user_base_address;
64 hw_ip.dram_base_address =
65 prop->dram_supports_virtual_memory ?
66 prop->dmmu.start_addr : prop->dram_user_base_address;
67 hw_ip.tpc_enabled_mask = prop->tpc_enabled_mask & 0xFF;
68 hw_ip.tpc_enabled_mask_ext = prop->tpc_enabled_mask;
69
70 hw_ip.sram_size = prop->sram_size - sram_kmd_size;
71
72 dram_available_size = prop->dram_size - dram_kmd_size;
73
74 hw_ip.dram_size = DIV_ROUND_DOWN_ULL(dram_available_size, prop->dram_page_size) *
75 prop->dram_page_size;
76
77 if (hw_ip.dram_size > PAGE_SIZE)
78 hw_ip.dram_enabled = 1;
79
80 hw_ip.dram_page_size = prop->dram_page_size;
81 hw_ip.device_mem_alloc_default_page_size = prop->device_mem_alloc_default_page_size;
82 hw_ip.num_of_events = prop->num_of_events;
83
84 memcpy(hw_ip.cpucp_version, prop->cpucp_info.cpucp_version,
85 min(VERSION_MAX_LEN, HL_INFO_VERSION_MAX_LEN));
86
87 memcpy(hw_ip.card_name, prop->cpucp_info.card_name,
88 min(CARD_NAME_MAX_LEN, HL_INFO_CARD_NAME_MAX_LEN));
89
90 hw_ip.cpld_version = le32_to_cpu(prop->cpucp_info.cpld_version);
91 hw_ip.module_id = le32_to_cpu(prop->cpucp_info.card_location);
92
93 hw_ip.psoc_pci_pll_nr = prop->psoc_pci_pll_nr;
94 hw_ip.psoc_pci_pll_nf = prop->psoc_pci_pll_nf;
95 hw_ip.psoc_pci_pll_od = prop->psoc_pci_pll_od;
96 hw_ip.psoc_pci_pll_div_factor = prop->psoc_pci_pll_div_factor;
97
98 hw_ip.decoder_enabled_mask = prop->decoder_enabled_mask;
99 hw_ip.mme_master_slave_mode = prop->mme_master_slave_mode;
100 hw_ip.first_available_interrupt_id = prop->first_available_user_interrupt;
101 hw_ip.number_of_user_interrupts = prop->user_interrupt_count;
102 hw_ip.tpc_interrupt_id = prop->tpc_interrupt_id;
103
104 hw_ip.edma_enabled_mask = prop->edma_enabled_mask;
105 hw_ip.server_type = prop->server_type;
106 hw_ip.security_enabled = prop->fw_security_enabled;
107 hw_ip.revision_id = hdev->pdev->revision;
108 hw_ip.rotator_enabled_mask = prop->rotator_enabled_mask;
109 hw_ip.engine_core_interrupt_reg_addr = prop->engine_core_interrupt_reg_addr;
110 hw_ip.reserved_dram_size = dram_kmd_size;
111
112 return copy_to_user(out, &hw_ip,
113 min((size_t) size, sizeof(hw_ip))) ? -EFAULT : 0;
114 }
115
hw_events_info(struct hl_device * hdev,bool aggregate,struct hl_info_args * args)116 static int hw_events_info(struct hl_device *hdev, bool aggregate,
117 struct hl_info_args *args)
118 {
119 u32 size, max_size = args->return_size;
120 void __user *out = (void __user *) (uintptr_t) args->return_pointer;
121 void *arr;
122
123 if ((!max_size) || (!out))
124 return -EINVAL;
125
126 arr = hdev->asic_funcs->get_events_stat(hdev, aggregate, &size);
127 if (!arr) {
128 dev_err(hdev->dev, "Events info not supported\n");
129 return -EOPNOTSUPP;
130 }
131
132 return copy_to_user(out, arr, min(max_size, size)) ? -EFAULT : 0;
133 }
134
events_info(struct hl_fpriv * hpriv,struct hl_info_args * args)135 static int events_info(struct hl_fpriv *hpriv, struct hl_info_args *args)
136 {
137 u32 max_size = args->return_size;
138 u64 events_mask;
139 void __user *out = (void __user *) (uintptr_t) args->return_pointer;
140
141 if ((max_size < sizeof(u64)) || (!out))
142 return -EINVAL;
143
144 mutex_lock(&hpriv->notifier_event.lock);
145 events_mask = hpriv->notifier_event.events_mask;
146 hpriv->notifier_event.events_mask = 0;
147 mutex_unlock(&hpriv->notifier_event.lock);
148
149 return copy_to_user(out, &events_mask, sizeof(u64)) ? -EFAULT : 0;
150 }
151
dram_usage_info(struct hl_fpriv * hpriv,struct hl_info_args * args)152 static int dram_usage_info(struct hl_fpriv *hpriv, struct hl_info_args *args)
153 {
154 struct hl_device *hdev = hpriv->hdev;
155 struct hl_info_dram_usage dram_usage = {0};
156 u32 max_size = args->return_size;
157 void __user *out = (void __user *) (uintptr_t) args->return_pointer;
158 struct asic_fixed_properties *prop = &hdev->asic_prop;
159 u64 dram_kmd_size;
160
161 if ((!max_size) || (!out))
162 return -EINVAL;
163
164 dram_kmd_size = (prop->dram_user_base_address -
165 prop->dram_base_address);
166 dram_usage.dram_free_mem = (prop->dram_size - dram_kmd_size) -
167 atomic64_read(&hdev->dram_used_mem);
168 if (hpriv->ctx)
169 dram_usage.ctx_dram_mem =
170 atomic64_read(&hpriv->ctx->dram_phys_mem);
171
172 return copy_to_user(out, &dram_usage,
173 min((size_t) max_size, sizeof(dram_usage))) ? -EFAULT : 0;
174 }
175
hw_idle(struct hl_device * hdev,struct hl_info_args * args)176 static int hw_idle(struct hl_device *hdev, struct hl_info_args *args)
177 {
178 struct hl_info_hw_idle hw_idle = {0};
179 u32 max_size = args->return_size;
180 void __user *out = (void __user *) (uintptr_t) args->return_pointer;
181
182 if ((!max_size) || (!out))
183 return -EINVAL;
184
185 hw_idle.is_idle = hdev->asic_funcs->is_device_idle(hdev,
186 hw_idle.busy_engines_mask_ext,
187 HL_BUSY_ENGINES_MASK_EXT_SIZE, NULL);
188 hw_idle.busy_engines_mask =
189 lower_32_bits(hw_idle.busy_engines_mask_ext[0]);
190
191 return copy_to_user(out, &hw_idle,
192 min((size_t) max_size, sizeof(hw_idle))) ? -EFAULT : 0;
193 }
194
debug_coresight(struct hl_device * hdev,struct hl_ctx * ctx,struct hl_debug_args * args)195 static int debug_coresight(struct hl_device *hdev, struct hl_ctx *ctx, struct hl_debug_args *args)
196 {
197 struct hl_debug_params *params;
198 void *input = NULL, *output = NULL;
199 int rc;
200
201 params = kzalloc(sizeof(*params), GFP_KERNEL);
202 if (!params)
203 return -ENOMEM;
204
205 params->reg_idx = args->reg_idx;
206 params->enable = args->enable;
207 params->op = args->op;
208
209 if (args->input_ptr && args->input_size) {
210 input = kzalloc(hl_debug_struct_size[args->op], GFP_KERNEL);
211 if (!input) {
212 rc = -ENOMEM;
213 goto out;
214 }
215
216 if (copy_from_user(input, u64_to_user_ptr(args->input_ptr),
217 args->input_size)) {
218 rc = -EFAULT;
219 dev_err(hdev->dev, "failed to copy input debug data\n");
220 goto out;
221 }
222
223 params->input = input;
224 }
225
226 if (args->output_ptr && args->output_size) {
227 output = kzalloc(args->output_size, GFP_KERNEL);
228 if (!output) {
229 rc = -ENOMEM;
230 goto out;
231 }
232
233 params->output = output;
234 params->output_size = args->output_size;
235 }
236
237 rc = hdev->asic_funcs->debug_coresight(hdev, ctx, params);
238 if (rc) {
239 dev_err(hdev->dev,
240 "debug coresight operation failed %d\n", rc);
241 goto out;
242 }
243
244 if (output && copy_to_user((void __user *) (uintptr_t) args->output_ptr,
245 output, args->output_size)) {
246 dev_err(hdev->dev, "copy to user failed in debug ioctl\n");
247 rc = -EFAULT;
248 goto out;
249 }
250
251
252 out:
253 kfree(params);
254 kfree(output);
255 kfree(input);
256
257 return rc;
258 }
259
device_utilization(struct hl_device * hdev,struct hl_info_args * args)260 static int device_utilization(struct hl_device *hdev, struct hl_info_args *args)
261 {
262 struct hl_info_device_utilization device_util = {0};
263 u32 max_size = args->return_size;
264 void __user *out = (void __user *) (uintptr_t) args->return_pointer;
265 int rc;
266
267 if ((!max_size) || (!out))
268 return -EINVAL;
269
270 rc = hl_device_utilization(hdev, &device_util.utilization);
271 if (rc)
272 return -EINVAL;
273
274 return copy_to_user(out, &device_util,
275 min((size_t) max_size, sizeof(device_util))) ? -EFAULT : 0;
276 }
277
get_clk_rate(struct hl_device * hdev,struct hl_info_args * args)278 static int get_clk_rate(struct hl_device *hdev, struct hl_info_args *args)
279 {
280 struct hl_info_clk_rate clk_rate = {0};
281 u32 max_size = args->return_size;
282 void __user *out = (void __user *) (uintptr_t) args->return_pointer;
283 int rc;
284
285 if ((!max_size) || (!out))
286 return -EINVAL;
287
288 rc = hl_fw_get_clk_rate(hdev, &clk_rate.cur_clk_rate_mhz, &clk_rate.max_clk_rate_mhz);
289 if (rc)
290 return rc;
291
292 return copy_to_user(out, &clk_rate, min_t(size_t, max_size, sizeof(clk_rate)))
293 ? -EFAULT : 0;
294 }
295
get_reset_count(struct hl_device * hdev,struct hl_info_args * args)296 static int get_reset_count(struct hl_device *hdev, struct hl_info_args *args)
297 {
298 struct hl_info_reset_count reset_count = {0};
299 u32 max_size = args->return_size;
300 void __user *out = (void __user *) (uintptr_t) args->return_pointer;
301
302 if ((!max_size) || (!out))
303 return -EINVAL;
304
305 reset_count.hard_reset_cnt = hdev->reset_info.hard_reset_cnt;
306 reset_count.soft_reset_cnt = hdev->reset_info.compute_reset_cnt;
307
308 return copy_to_user(out, &reset_count,
309 min((size_t) max_size, sizeof(reset_count))) ? -EFAULT : 0;
310 }
311
time_sync_info(struct hl_device * hdev,struct hl_info_args * args)312 static int time_sync_info(struct hl_device *hdev, struct hl_info_args *args)
313 {
314 struct hl_info_time_sync time_sync = {0};
315 u32 max_size = args->return_size;
316 void __user *out = (void __user *) (uintptr_t) args->return_pointer;
317
318 if ((!max_size) || (!out))
319 return -EINVAL;
320
321 time_sync.device_time = hdev->asic_funcs->get_device_time(hdev);
322 time_sync.host_time = ktime_get_raw_ns();
323
324 return copy_to_user(out, &time_sync,
325 min((size_t) max_size, sizeof(time_sync))) ? -EFAULT : 0;
326 }
327
pci_counters_info(struct hl_fpriv * hpriv,struct hl_info_args * args)328 static int pci_counters_info(struct hl_fpriv *hpriv, struct hl_info_args *args)
329 {
330 struct hl_device *hdev = hpriv->hdev;
331 struct hl_info_pci_counters pci_counters = {0};
332 u32 max_size = args->return_size;
333 void __user *out = (void __user *) (uintptr_t) args->return_pointer;
334 int rc;
335
336 if ((!max_size) || (!out))
337 return -EINVAL;
338
339 rc = hl_fw_cpucp_pci_counters_get(hdev, &pci_counters);
340 if (rc)
341 return rc;
342
343 return copy_to_user(out, &pci_counters,
344 min((size_t) max_size, sizeof(pci_counters))) ? -EFAULT : 0;
345 }
346
clk_throttle_info(struct hl_fpriv * hpriv,struct hl_info_args * args)347 static int clk_throttle_info(struct hl_fpriv *hpriv, struct hl_info_args *args)
348 {
349 void __user *out = (void __user *) (uintptr_t) args->return_pointer;
350 struct hl_device *hdev = hpriv->hdev;
351 struct hl_info_clk_throttle clk_throttle = {0};
352 ktime_t end_time, zero_time = ktime_set(0, 0);
353 u32 max_size = args->return_size;
354 int i;
355
356 if ((!max_size) || (!out))
357 return -EINVAL;
358
359 mutex_lock(&hdev->clk_throttling.lock);
360
361 clk_throttle.clk_throttling_reason = hdev->clk_throttling.current_reason;
362
363 for (i = 0 ; i < HL_CLK_THROTTLE_TYPE_MAX ; i++) {
364 if (!(hdev->clk_throttling.aggregated_reason & BIT(i)))
365 continue;
366
367 clk_throttle.clk_throttling_timestamp_us[i] =
368 ktime_to_us(hdev->clk_throttling.timestamp[i].start);
369
370 if (ktime_compare(hdev->clk_throttling.timestamp[i].end, zero_time))
371 end_time = hdev->clk_throttling.timestamp[i].end;
372 else
373 end_time = ktime_get();
374
375 clk_throttle.clk_throttling_duration_ns[i] =
376 ktime_to_ns(ktime_sub(end_time,
377 hdev->clk_throttling.timestamp[i].start));
378
379 }
380 mutex_unlock(&hdev->clk_throttling.lock);
381
382 return copy_to_user(out, &clk_throttle,
383 min((size_t) max_size, sizeof(clk_throttle))) ? -EFAULT : 0;
384 }
385
cs_counters_info(struct hl_fpriv * hpriv,struct hl_info_args * args)386 static int cs_counters_info(struct hl_fpriv *hpriv, struct hl_info_args *args)
387 {
388 void __user *out = (void __user *) (uintptr_t) args->return_pointer;
389 struct hl_info_cs_counters cs_counters = {0};
390 struct hl_device *hdev = hpriv->hdev;
391 struct hl_cs_counters_atomic *cntr;
392 u32 max_size = args->return_size;
393
394 cntr = &hdev->aggregated_cs_counters;
395
396 if ((!max_size) || (!out))
397 return -EINVAL;
398
399 cs_counters.total_out_of_mem_drop_cnt =
400 atomic64_read(&cntr->out_of_mem_drop_cnt);
401 cs_counters.total_parsing_drop_cnt =
402 atomic64_read(&cntr->parsing_drop_cnt);
403 cs_counters.total_queue_full_drop_cnt =
404 atomic64_read(&cntr->queue_full_drop_cnt);
405 cs_counters.total_device_in_reset_drop_cnt =
406 atomic64_read(&cntr->device_in_reset_drop_cnt);
407 cs_counters.total_max_cs_in_flight_drop_cnt =
408 atomic64_read(&cntr->max_cs_in_flight_drop_cnt);
409 cs_counters.total_validation_drop_cnt =
410 atomic64_read(&cntr->validation_drop_cnt);
411
412 if (hpriv->ctx) {
413 cs_counters.ctx_out_of_mem_drop_cnt =
414 atomic64_read(
415 &hpriv->ctx->cs_counters.out_of_mem_drop_cnt);
416 cs_counters.ctx_parsing_drop_cnt =
417 atomic64_read(
418 &hpriv->ctx->cs_counters.parsing_drop_cnt);
419 cs_counters.ctx_queue_full_drop_cnt =
420 atomic64_read(
421 &hpriv->ctx->cs_counters.queue_full_drop_cnt);
422 cs_counters.ctx_device_in_reset_drop_cnt =
423 atomic64_read(
424 &hpriv->ctx->cs_counters.device_in_reset_drop_cnt);
425 cs_counters.ctx_max_cs_in_flight_drop_cnt =
426 atomic64_read(
427 &hpriv->ctx->cs_counters.max_cs_in_flight_drop_cnt);
428 cs_counters.ctx_validation_drop_cnt =
429 atomic64_read(
430 &hpriv->ctx->cs_counters.validation_drop_cnt);
431 }
432
433 return copy_to_user(out, &cs_counters,
434 min((size_t) max_size, sizeof(cs_counters))) ? -EFAULT : 0;
435 }
436
sync_manager_info(struct hl_fpriv * hpriv,struct hl_info_args * args)437 static int sync_manager_info(struct hl_fpriv *hpriv, struct hl_info_args *args)
438 {
439 struct hl_device *hdev = hpriv->hdev;
440 struct asic_fixed_properties *prop = &hdev->asic_prop;
441 struct hl_info_sync_manager sm_info = {0};
442 u32 max_size = args->return_size;
443 void __user *out = (void __user *) (uintptr_t) args->return_pointer;
444
445 if ((!max_size) || (!out))
446 return -EINVAL;
447
448 if (args->dcore_id >= HL_MAX_DCORES)
449 return -EINVAL;
450
451 sm_info.first_available_sync_object =
452 prop->first_available_user_sob[args->dcore_id];
453 sm_info.first_available_monitor =
454 prop->first_available_user_mon[args->dcore_id];
455 sm_info.first_available_cq =
456 prop->first_available_cq[args->dcore_id];
457
458 return copy_to_user(out, &sm_info, min_t(size_t, (size_t) max_size,
459 sizeof(sm_info))) ? -EFAULT : 0;
460 }
461
total_energy_consumption_info(struct hl_fpriv * hpriv,struct hl_info_args * args)462 static int total_energy_consumption_info(struct hl_fpriv *hpriv,
463 struct hl_info_args *args)
464 {
465 struct hl_device *hdev = hpriv->hdev;
466 struct hl_info_energy total_energy = {0};
467 u32 max_size = args->return_size;
468 void __user *out = (void __user *) (uintptr_t) args->return_pointer;
469 int rc;
470
471 if ((!max_size) || (!out))
472 return -EINVAL;
473
474 rc = hl_fw_cpucp_total_energy_get(hdev,
475 &total_energy.total_energy_consumption);
476 if (rc)
477 return rc;
478
479 return copy_to_user(out, &total_energy,
480 min((size_t) max_size, sizeof(total_energy))) ? -EFAULT : 0;
481 }
482
pll_frequency_info(struct hl_fpriv * hpriv,struct hl_info_args * args)483 static int pll_frequency_info(struct hl_fpriv *hpriv, struct hl_info_args *args)
484 {
485 struct hl_device *hdev = hpriv->hdev;
486 struct hl_pll_frequency_info freq_info = { {0} };
487 u32 max_size = args->return_size;
488 void __user *out = (void __user *) (uintptr_t) args->return_pointer;
489 int rc;
490
491 if ((!max_size) || (!out))
492 return -EINVAL;
493
494 rc = hl_fw_cpucp_pll_info_get(hdev, args->pll_index, freq_info.output);
495 if (rc)
496 return rc;
497
498 return copy_to_user(out, &freq_info,
499 min((size_t) max_size, sizeof(freq_info))) ? -EFAULT : 0;
500 }
501
power_info(struct hl_fpriv * hpriv,struct hl_info_args * args)502 static int power_info(struct hl_fpriv *hpriv, struct hl_info_args *args)
503 {
504 struct hl_device *hdev = hpriv->hdev;
505 u32 max_size = args->return_size;
506 struct hl_power_info power_info = {0};
507 void __user *out = (void __user *) (uintptr_t) args->return_pointer;
508 int rc;
509
510 if ((!max_size) || (!out))
511 return -EINVAL;
512
513 rc = hl_fw_cpucp_power_get(hdev, &power_info.power);
514 if (rc)
515 return rc;
516
517 return copy_to_user(out, &power_info,
518 min((size_t) max_size, sizeof(power_info))) ? -EFAULT : 0;
519 }
520
open_stats_info(struct hl_fpriv * hpriv,struct hl_info_args * args)521 static int open_stats_info(struct hl_fpriv *hpriv, struct hl_info_args *args)
522 {
523 struct hl_device *hdev = hpriv->hdev;
524 u32 max_size = args->return_size;
525 struct hl_open_stats_info open_stats_info = {0};
526 void __user *out = (void __user *) (uintptr_t) args->return_pointer;
527
528 if ((!max_size) || (!out))
529 return -EINVAL;
530
531 open_stats_info.last_open_period_ms = jiffies64_to_msecs(
532 hdev->last_open_session_duration_jif);
533 open_stats_info.open_counter = hdev->open_counter;
534 open_stats_info.is_compute_ctx_active = hdev->is_compute_ctx_active;
535 open_stats_info.compute_ctx_in_release = hdev->compute_ctx_in_release;
536
537 return copy_to_user(out, &open_stats_info,
538 min((size_t) max_size, sizeof(open_stats_info))) ? -EFAULT : 0;
539 }
540
dram_pending_rows_info(struct hl_fpriv * hpriv,struct hl_info_args * args)541 static int dram_pending_rows_info(struct hl_fpriv *hpriv, struct hl_info_args *args)
542 {
543 struct hl_device *hdev = hpriv->hdev;
544 u32 max_size = args->return_size;
545 u32 pend_rows_num = 0;
546 void __user *out = (void __user *) (uintptr_t) args->return_pointer;
547 int rc;
548
549 if ((!max_size) || (!out))
550 return -EINVAL;
551
552 rc = hl_fw_dram_pending_row_get(hdev, &pend_rows_num);
553 if (rc)
554 return rc;
555
556 return copy_to_user(out, &pend_rows_num,
557 min_t(size_t, max_size, sizeof(pend_rows_num))) ? -EFAULT : 0;
558 }
559
dram_replaced_rows_info(struct hl_fpriv * hpriv,struct hl_info_args * args)560 static int dram_replaced_rows_info(struct hl_fpriv *hpriv, struct hl_info_args *args)
561 {
562 struct hl_device *hdev = hpriv->hdev;
563 u32 max_size = args->return_size;
564 struct cpucp_hbm_row_info info = {0};
565 void __user *out = (void __user *) (uintptr_t) args->return_pointer;
566 int rc;
567
568 if ((!max_size) || (!out))
569 return -EINVAL;
570
571 rc = hl_fw_dram_replaced_row_get(hdev, &info);
572 if (rc)
573 return rc;
574
575 return copy_to_user(out, &info, min_t(size_t, max_size, sizeof(info))) ? -EFAULT : 0;
576 }
577
last_err_open_dev_info(struct hl_fpriv * hpriv,struct hl_info_args * args)578 static int last_err_open_dev_info(struct hl_fpriv *hpriv, struct hl_info_args *args)
579 {
580 struct hl_info_last_err_open_dev_time info = {0};
581 struct hl_device *hdev = hpriv->hdev;
582 u32 max_size = args->return_size;
583 void __user *out = (void __user *) (uintptr_t) args->return_pointer;
584
585 if ((!max_size) || (!out))
586 return -EINVAL;
587
588 info.timestamp = ktime_to_ns(hdev->last_successful_open_ktime);
589
590 return copy_to_user(out, &info, min_t(size_t, max_size, sizeof(info))) ? -EFAULT : 0;
591 }
592
cs_timeout_info(struct hl_fpriv * hpriv,struct hl_info_args * args)593 static int cs_timeout_info(struct hl_fpriv *hpriv, struct hl_info_args *args)
594 {
595 struct hl_info_cs_timeout_event info = {0};
596 struct hl_device *hdev = hpriv->hdev;
597 u32 max_size = args->return_size;
598 void __user *out = (void __user *) (uintptr_t) args->return_pointer;
599
600 if ((!max_size) || (!out))
601 return -EINVAL;
602
603 info.seq = hdev->captured_err_info.cs_timeout.seq;
604 info.timestamp = ktime_to_ns(hdev->captured_err_info.cs_timeout.timestamp);
605
606 return copy_to_user(out, &info, min_t(size_t, max_size, sizeof(info))) ? -EFAULT : 0;
607 }
608
razwi_info(struct hl_fpriv * hpriv,struct hl_info_args * args)609 static int razwi_info(struct hl_fpriv *hpriv, struct hl_info_args *args)
610 {
611 void __user *out = (void __user *) (uintptr_t) args->return_pointer;
612 struct hl_device *hdev = hpriv->hdev;
613 u32 max_size = args->return_size;
614 struct razwi_info *razwi_info;
615
616 if ((!max_size) || (!out))
617 return -EINVAL;
618
619 razwi_info = &hdev->captured_err_info.razwi_info;
620 if (!razwi_info->razwi_info_available)
621 return 0;
622
623 return copy_to_user(out, &razwi_info->razwi,
624 min_t(size_t, max_size, sizeof(struct hl_info_razwi_event))) ? -EFAULT : 0;
625 }
626
undefined_opcode_info(struct hl_fpriv * hpriv,struct hl_info_args * args)627 static int undefined_opcode_info(struct hl_fpriv *hpriv, struct hl_info_args *args)
628 {
629 struct hl_device *hdev = hpriv->hdev;
630 u32 max_size = args->return_size;
631 struct hl_info_undefined_opcode_event info = {0};
632 void __user *out = (void __user *) (uintptr_t) args->return_pointer;
633
634 if ((!max_size) || (!out))
635 return -EINVAL;
636
637 info.timestamp = ktime_to_ns(hdev->captured_err_info.undef_opcode.timestamp);
638 info.engine_id = hdev->captured_err_info.undef_opcode.engine_id;
639 info.cq_addr = hdev->captured_err_info.undef_opcode.cq_addr;
640 info.cq_size = hdev->captured_err_info.undef_opcode.cq_size;
641 info.stream_id = hdev->captured_err_info.undef_opcode.stream_id;
642 info.cb_addr_streams_len = hdev->captured_err_info.undef_opcode.cb_addr_streams_len;
643 memcpy(info.cb_addr_streams, hdev->captured_err_info.undef_opcode.cb_addr_streams,
644 sizeof(info.cb_addr_streams));
645
646 return copy_to_user(out, &info, min_t(size_t, max_size, sizeof(info))) ? -EFAULT : 0;
647 }
648
dev_mem_alloc_page_sizes_info(struct hl_fpriv * hpriv,struct hl_info_args * args)649 static int dev_mem_alloc_page_sizes_info(struct hl_fpriv *hpriv, struct hl_info_args *args)
650 {
651 void __user *out = (void __user *) (uintptr_t) args->return_pointer;
652 struct hl_info_dev_memalloc_page_sizes info = {0};
653 struct hl_device *hdev = hpriv->hdev;
654 u32 max_size = args->return_size;
655
656 if ((!max_size) || (!out))
657 return -EINVAL;
658
659 /*
660 * Future ASICs that will support multiple DRAM page sizes will support only "powers of 2"
661 * pages (unlike some of the ASICs before supporting multiple page sizes).
662 * For this reason for all ASICs that not support multiple page size the function will
663 * return an empty bitmask indicating that multiple page sizes is not supported.
664 */
665 info.page_order_bitmask = hdev->asic_prop.dmmu.supported_pages_mask;
666
667 return copy_to_user(out, &info, min_t(size_t, max_size, sizeof(info))) ? -EFAULT : 0;
668 }
669
sec_attest_info(struct hl_fpriv * hpriv,struct hl_info_args * args)670 static int sec_attest_info(struct hl_fpriv *hpriv, struct hl_info_args *args)
671 {
672 void __user *out = (void __user *) (uintptr_t) args->return_pointer;
673 struct cpucp_sec_attest_info *sec_attest_info;
674 struct hl_info_sec_attest *info;
675 u32 max_size = args->return_size;
676 int rc;
677
678 if ((!max_size) || (!out))
679 return -EINVAL;
680
681 sec_attest_info = kmalloc(sizeof(*sec_attest_info), GFP_KERNEL);
682 if (!sec_attest_info)
683 return -ENOMEM;
684
685 info = kzalloc(sizeof(*info), GFP_KERNEL);
686 if (!info) {
687 rc = -ENOMEM;
688 goto free_sec_attest_info;
689 }
690
691 rc = hl_fw_get_sec_attest_info(hpriv->hdev, sec_attest_info, args->sec_attest_nonce);
692 if (rc)
693 goto free_info;
694
695 info->nonce = le32_to_cpu(sec_attest_info->nonce);
696 info->pcr_quote_len = le16_to_cpu(sec_attest_info->pcr_quote_len);
697 info->pub_data_len = le16_to_cpu(sec_attest_info->pub_data_len);
698 info->certificate_len = le16_to_cpu(sec_attest_info->certificate_len);
699 info->pcr_num_reg = sec_attest_info->pcr_num_reg;
700 info->pcr_reg_len = sec_attest_info->pcr_reg_len;
701 info->quote_sig_len = sec_attest_info->quote_sig_len;
702 memcpy(&info->pcr_data, &sec_attest_info->pcr_data, sizeof(info->pcr_data));
703 memcpy(&info->pcr_quote, &sec_attest_info->pcr_quote, sizeof(info->pcr_quote));
704 memcpy(&info->public_data, &sec_attest_info->public_data, sizeof(info->public_data));
705 memcpy(&info->certificate, &sec_attest_info->certificate, sizeof(info->certificate));
706 memcpy(&info->quote_sig, &sec_attest_info->quote_sig, sizeof(info->quote_sig));
707
708 rc = copy_to_user(out, info,
709 min_t(size_t, max_size, sizeof(*info))) ? -EFAULT : 0;
710
711 free_info:
712 kfree(info);
713 free_sec_attest_info:
714 kfree(sec_attest_info);
715
716 return rc;
717 }
718
eventfd_register(struct hl_fpriv * hpriv,struct hl_info_args * args)719 static int eventfd_register(struct hl_fpriv *hpriv, struct hl_info_args *args)
720 {
721 int rc;
722
723 /* check if there is already a registered on that process */
724 mutex_lock(&hpriv->notifier_event.lock);
725 if (hpriv->notifier_event.eventfd) {
726 mutex_unlock(&hpriv->notifier_event.lock);
727 return -EINVAL;
728 }
729
730 hpriv->notifier_event.eventfd = eventfd_ctx_fdget(args->eventfd);
731 if (IS_ERR(hpriv->notifier_event.eventfd)) {
732 rc = PTR_ERR(hpriv->notifier_event.eventfd);
733 hpriv->notifier_event.eventfd = NULL;
734 mutex_unlock(&hpriv->notifier_event.lock);
735 return rc;
736 }
737
738 mutex_unlock(&hpriv->notifier_event.lock);
739 return 0;
740 }
741
eventfd_unregister(struct hl_fpriv * hpriv,struct hl_info_args * args)742 static int eventfd_unregister(struct hl_fpriv *hpriv, struct hl_info_args *args)
743 {
744 mutex_lock(&hpriv->notifier_event.lock);
745 if (!hpriv->notifier_event.eventfd) {
746 mutex_unlock(&hpriv->notifier_event.lock);
747 return -EINVAL;
748 }
749
750 eventfd_ctx_put(hpriv->notifier_event.eventfd);
751 hpriv->notifier_event.eventfd = NULL;
752 mutex_unlock(&hpriv->notifier_event.lock);
753 return 0;
754 }
755
engine_status_info(struct hl_fpriv * hpriv,struct hl_info_args * args)756 static int engine_status_info(struct hl_fpriv *hpriv, struct hl_info_args *args)
757 {
758 void __user *out = (void __user *) (uintptr_t) args->return_pointer;
759 u32 status_buf_size = args->return_size;
760 struct hl_device *hdev = hpriv->hdev;
761 struct engines_data eng_data;
762 int rc;
763
764 if ((status_buf_size < SZ_1K) || (status_buf_size > HL_ENGINES_DATA_MAX_SIZE) || (!out))
765 return -EINVAL;
766
767 eng_data.actual_size = 0;
768 eng_data.allocated_buf_size = status_buf_size;
769 eng_data.buf = vmalloc(status_buf_size);
770 if (!eng_data.buf)
771 return -ENOMEM;
772
773 hdev->asic_funcs->is_device_idle(hdev, NULL, 0, &eng_data);
774
775 if (eng_data.actual_size > eng_data.allocated_buf_size) {
776 dev_err(hdev->dev,
777 "Engines data size (%d Bytes) is bigger than allocated size (%u Bytes)\n",
778 eng_data.actual_size, status_buf_size);
779 vfree(eng_data.buf);
780 return -ENOMEM;
781 }
782
783 args->user_buffer_actual_size = eng_data.actual_size;
784 rc = copy_to_user(out, eng_data.buf, min_t(size_t, status_buf_size, eng_data.actual_size)) ?
785 -EFAULT : 0;
786
787 vfree(eng_data.buf);
788
789 return rc;
790 }
791
page_fault_info(struct hl_fpriv * hpriv,struct hl_info_args * args)792 static int page_fault_info(struct hl_fpriv *hpriv, struct hl_info_args *args)
793 {
794 void __user *out = (void __user *) (uintptr_t) args->return_pointer;
795 struct hl_device *hdev = hpriv->hdev;
796 u32 max_size = args->return_size;
797 struct page_fault_info *pgf_info;
798
799 if ((!max_size) || (!out))
800 return -EINVAL;
801
802 pgf_info = &hdev->captured_err_info.page_fault_info;
803 if (!pgf_info->page_fault_info_available)
804 return 0;
805
806 return copy_to_user(out, &pgf_info->page_fault,
807 min_t(size_t, max_size, sizeof(struct hl_page_fault_info))) ? -EFAULT : 0;
808 }
809
user_mappings_info(struct hl_fpriv * hpriv,struct hl_info_args * args)810 static int user_mappings_info(struct hl_fpriv *hpriv, struct hl_info_args *args)
811 {
812 void __user *out = (void __user *) (uintptr_t) args->return_pointer;
813 u32 user_buf_size = args->return_size;
814 struct hl_device *hdev = hpriv->hdev;
815 struct page_fault_info *pgf_info;
816 u64 actual_size;
817
818 if (!out)
819 return -EINVAL;
820
821 pgf_info = &hdev->captured_err_info.page_fault_info;
822 if (!pgf_info->page_fault_info_available)
823 return 0;
824
825 args->array_size = pgf_info->num_of_user_mappings;
826
827 actual_size = pgf_info->num_of_user_mappings * sizeof(struct hl_user_mapping);
828 if (user_buf_size < actual_size)
829 return -ENOMEM;
830
831 return copy_to_user(out, pgf_info->user_mappings, actual_size) ? -EFAULT : 0;
832 }
833
hw_err_info(struct hl_fpriv * hpriv,struct hl_info_args * args)834 static int hw_err_info(struct hl_fpriv *hpriv, struct hl_info_args *args)
835 {
836 void __user *user_buf = (void __user *) (uintptr_t) args->return_pointer;
837 struct hl_device *hdev = hpriv->hdev;
838 u32 user_buf_size = args->return_size;
839 struct hw_err_info *info;
840 int rc;
841
842 if (!user_buf)
843 return -EINVAL;
844
845 info = &hdev->captured_err_info.hw_err;
846 if (!info->event_info_available)
847 return 0;
848
849 if (user_buf_size < sizeof(struct hl_info_hw_err_event))
850 return -ENOMEM;
851
852 rc = copy_to_user(user_buf, &info->event, sizeof(struct hl_info_hw_err_event));
853 return rc ? -EFAULT : 0;
854 }
855
fw_err_info(struct hl_fpriv * hpriv,struct hl_info_args * args)856 static int fw_err_info(struct hl_fpriv *hpriv, struct hl_info_args *args)
857 {
858 void __user *user_buf = (void __user *) (uintptr_t) args->return_pointer;
859 struct hl_device *hdev = hpriv->hdev;
860 u32 user_buf_size = args->return_size;
861 struct fw_err_info *info;
862 int rc;
863
864 if (!user_buf)
865 return -EINVAL;
866
867 info = &hdev->captured_err_info.fw_err;
868 if (!info->event_info_available)
869 return 0;
870
871 if (user_buf_size < sizeof(struct hl_info_fw_err_event))
872 return -ENOMEM;
873
874 rc = copy_to_user(user_buf, &info->event, sizeof(struct hl_info_fw_err_event));
875 return rc ? -EFAULT : 0;
876 }
877
send_fw_generic_request(struct hl_device * hdev,struct hl_info_args * info_args)878 static int send_fw_generic_request(struct hl_device *hdev, struct hl_info_args *info_args)
879 {
880 void __user *buff = (void __user *) (uintptr_t) info_args->return_pointer;
881 u32 size = info_args->return_size;
882 dma_addr_t dma_handle;
883 bool need_input_buff;
884 void *fw_buff;
885 int rc = 0;
886
887 switch (info_args->fw_sub_opcode) {
888 case HL_PASSTHROUGH_VERSIONS:
889 need_input_buff = false;
890 break;
891 default:
892 return -EINVAL;
893 }
894
895 if (size > SZ_1M) {
896 dev_err(hdev->dev, "buffer size cannot exceed 1MB\n");
897 return -EINVAL;
898 }
899
900 fw_buff = hl_cpu_accessible_dma_pool_alloc(hdev, size, &dma_handle);
901 if (!fw_buff)
902 return -ENOMEM;
903
904
905 if (need_input_buff && copy_from_user(fw_buff, buff, size)) {
906 dev_dbg(hdev->dev, "Failed to copy from user FW buff\n");
907 rc = -EFAULT;
908 goto free_buff;
909 }
910
911 rc = hl_fw_send_generic_request(hdev, info_args->fw_sub_opcode, dma_handle, &size);
912 if (rc)
913 goto free_buff;
914
915 if (copy_to_user(buff, fw_buff, min(size, info_args->return_size))) {
916 dev_dbg(hdev->dev, "Failed to copy to user FW generic req output\n");
917 rc = -EFAULT;
918 }
919
920 free_buff:
921 hl_cpu_accessible_dma_pool_free(hdev, info_args->return_size, fw_buff);
922
923 return rc;
924 }
925
_hl_info_ioctl(struct hl_fpriv * hpriv,void * data,struct device * dev)926 static int _hl_info_ioctl(struct hl_fpriv *hpriv, void *data,
927 struct device *dev)
928 {
929 enum hl_device_status status;
930 struct hl_info_args *args = data;
931 struct hl_device *hdev = hpriv->hdev;
932 int rc;
933
934 if (args->pad) {
935 dev_dbg(hdev->dev, "Padding bytes must be 0\n");
936 return -EINVAL;
937 }
938
939 /*
940 * Information is returned for the following opcodes even if the device
941 * is disabled or in reset.
942 */
943 switch (args->op) {
944 case HL_INFO_HW_IP_INFO:
945 return hw_ip_info(hdev, args);
946
947 case HL_INFO_DEVICE_STATUS:
948 return device_status_info(hdev, args);
949
950 case HL_INFO_RESET_COUNT:
951 return get_reset_count(hdev, args);
952
953 case HL_INFO_HW_EVENTS:
954 return hw_events_info(hdev, false, args);
955
956 case HL_INFO_HW_EVENTS_AGGREGATE:
957 return hw_events_info(hdev, true, args);
958
959 case HL_INFO_CS_COUNTERS:
960 return cs_counters_info(hpriv, args);
961
962 case HL_INFO_CLK_THROTTLE_REASON:
963 return clk_throttle_info(hpriv, args);
964
965 case HL_INFO_SYNC_MANAGER:
966 return sync_manager_info(hpriv, args);
967
968 case HL_INFO_OPEN_STATS:
969 return open_stats_info(hpriv, args);
970
971 case HL_INFO_LAST_ERR_OPEN_DEV_TIME:
972 return last_err_open_dev_info(hpriv, args);
973
974 case HL_INFO_CS_TIMEOUT_EVENT:
975 return cs_timeout_info(hpriv, args);
976
977 case HL_INFO_RAZWI_EVENT:
978 return razwi_info(hpriv, args);
979
980 case HL_INFO_UNDEFINED_OPCODE_EVENT:
981 return undefined_opcode_info(hpriv, args);
982
983 case HL_INFO_DEV_MEM_ALLOC_PAGE_SIZES:
984 return dev_mem_alloc_page_sizes_info(hpriv, args);
985
986 case HL_INFO_GET_EVENTS:
987 return events_info(hpriv, args);
988
989 case HL_INFO_PAGE_FAULT_EVENT:
990 return page_fault_info(hpriv, args);
991
992 case HL_INFO_USER_MAPPINGS:
993 return user_mappings_info(hpriv, args);
994
995 case HL_INFO_UNREGISTER_EVENTFD:
996 return eventfd_unregister(hpriv, args);
997
998 case HL_INFO_HW_ERR_EVENT:
999 return hw_err_info(hpriv, args);
1000
1001 case HL_INFO_FW_ERR_EVENT:
1002 return fw_err_info(hpriv, args);
1003
1004 case HL_INFO_DRAM_USAGE:
1005 return dram_usage_info(hpriv, args);
1006 default:
1007 break;
1008 }
1009
1010 if (!hl_device_operational(hdev, &status)) {
1011 dev_dbg_ratelimited(dev,
1012 "Device is %s. Can't execute INFO IOCTL\n",
1013 hdev->status[status]);
1014 return -EBUSY;
1015 }
1016
1017 switch (args->op) {
1018 case HL_INFO_HW_IDLE:
1019 rc = hw_idle(hdev, args);
1020 break;
1021
1022 case HL_INFO_DEVICE_UTILIZATION:
1023 rc = device_utilization(hdev, args);
1024 break;
1025
1026 case HL_INFO_CLK_RATE:
1027 rc = get_clk_rate(hdev, args);
1028 break;
1029
1030 case HL_INFO_TIME_SYNC:
1031 return time_sync_info(hdev, args);
1032
1033 case HL_INFO_PCI_COUNTERS:
1034 return pci_counters_info(hpriv, args);
1035
1036 case HL_INFO_TOTAL_ENERGY:
1037 return total_energy_consumption_info(hpriv, args);
1038
1039 case HL_INFO_PLL_FREQUENCY:
1040 return pll_frequency_info(hpriv, args);
1041
1042 case HL_INFO_POWER:
1043 return power_info(hpriv, args);
1044
1045
1046 case HL_INFO_DRAM_REPLACED_ROWS:
1047 return dram_replaced_rows_info(hpriv, args);
1048
1049 case HL_INFO_DRAM_PENDING_ROWS:
1050 return dram_pending_rows_info(hpriv, args);
1051
1052 case HL_INFO_SECURED_ATTESTATION:
1053 return sec_attest_info(hpriv, args);
1054
1055 case HL_INFO_REGISTER_EVENTFD:
1056 return eventfd_register(hpriv, args);
1057
1058 case HL_INFO_ENGINE_STATUS:
1059 return engine_status_info(hpriv, args);
1060
1061 case HL_INFO_FW_GENERIC_REQ:
1062 return send_fw_generic_request(hdev, args);
1063
1064 default:
1065 dev_err(dev, "Invalid request %d\n", args->op);
1066 rc = -EINVAL;
1067 break;
1068 }
1069
1070 return rc;
1071 }
1072
hl_info_ioctl(struct hl_fpriv * hpriv,void * data)1073 static int hl_info_ioctl(struct hl_fpriv *hpriv, void *data)
1074 {
1075 return _hl_info_ioctl(hpriv, data, hpriv->hdev->dev);
1076 }
1077
hl_info_ioctl_control(struct hl_fpriv * hpriv,void * data)1078 static int hl_info_ioctl_control(struct hl_fpriv *hpriv, void *data)
1079 {
1080 return _hl_info_ioctl(hpriv, data, hpriv->hdev->dev_ctrl);
1081 }
1082
hl_debug_ioctl(struct hl_fpriv * hpriv,void * data)1083 static int hl_debug_ioctl(struct hl_fpriv *hpriv, void *data)
1084 {
1085 struct hl_debug_args *args = data;
1086 struct hl_device *hdev = hpriv->hdev;
1087 enum hl_device_status status;
1088
1089 int rc = 0;
1090
1091 if (!hl_device_operational(hdev, &status)) {
1092 dev_dbg_ratelimited(hdev->dev,
1093 "Device is %s. Can't execute DEBUG IOCTL\n",
1094 hdev->status[status]);
1095 return -EBUSY;
1096 }
1097
1098 switch (args->op) {
1099 case HL_DEBUG_OP_ETR:
1100 case HL_DEBUG_OP_ETF:
1101 case HL_DEBUG_OP_STM:
1102 case HL_DEBUG_OP_FUNNEL:
1103 case HL_DEBUG_OP_BMON:
1104 case HL_DEBUG_OP_SPMU:
1105 case HL_DEBUG_OP_TIMESTAMP:
1106 if (!hdev->in_debug) {
1107 dev_err_ratelimited(hdev->dev,
1108 "Rejecting debug configuration request because device not in debug mode\n");
1109 return -EFAULT;
1110 }
1111 args->input_size = min(args->input_size, hl_debug_struct_size[args->op]);
1112 rc = debug_coresight(hdev, hpriv->ctx, args);
1113 break;
1114
1115 case HL_DEBUG_OP_SET_MODE:
1116 rc = hl_device_set_debug_mode(hdev, hpriv->ctx, (bool) args->enable);
1117 break;
1118
1119 default:
1120 dev_err(hdev->dev, "Invalid request %d\n", args->op);
1121 rc = -EINVAL;
1122 break;
1123 }
1124
1125 return rc;
1126 }
1127
1128 #define HL_IOCTL_DEF(ioctl, _func) \
1129 [_IOC_NR(ioctl)] = {.cmd = ioctl, .func = _func}
1130
1131 static const struct hl_ioctl_desc hl_ioctls[] = {
1132 HL_IOCTL_DEF(HL_IOCTL_INFO, hl_info_ioctl),
1133 HL_IOCTL_DEF(HL_IOCTL_CB, hl_cb_ioctl),
1134 HL_IOCTL_DEF(HL_IOCTL_CS, hl_cs_ioctl),
1135 HL_IOCTL_DEF(HL_IOCTL_WAIT_CS, hl_wait_ioctl),
1136 HL_IOCTL_DEF(HL_IOCTL_MEMORY, hl_mem_ioctl),
1137 HL_IOCTL_DEF(HL_IOCTL_DEBUG, hl_debug_ioctl)
1138 };
1139
1140 static const struct hl_ioctl_desc hl_ioctls_control[] = {
1141 HL_IOCTL_DEF(HL_IOCTL_INFO, hl_info_ioctl_control)
1142 };
1143
_hl_ioctl(struct file * filep,unsigned int cmd,unsigned long arg,const struct hl_ioctl_desc * ioctl,struct device * dev)1144 static long _hl_ioctl(struct file *filep, unsigned int cmd, unsigned long arg,
1145 const struct hl_ioctl_desc *ioctl, struct device *dev)
1146 {
1147 struct hl_fpriv *hpriv = filep->private_data;
1148 unsigned int nr = _IOC_NR(cmd);
1149 char stack_kdata[128] = {0};
1150 char *kdata = NULL;
1151 unsigned int usize, asize;
1152 hl_ioctl_t *func;
1153 u32 hl_size;
1154 int retcode;
1155
1156 /* Do not trust userspace, use our own definition */
1157 func = ioctl->func;
1158
1159 if (unlikely(!func)) {
1160 dev_dbg(dev, "no function\n");
1161 retcode = -ENOTTY;
1162 goto out_err;
1163 }
1164
1165 hl_size = _IOC_SIZE(ioctl->cmd);
1166 usize = asize = _IOC_SIZE(cmd);
1167 if (hl_size > asize)
1168 asize = hl_size;
1169
1170 cmd = ioctl->cmd;
1171
1172 if (cmd & (IOC_IN | IOC_OUT)) {
1173 if (asize <= sizeof(stack_kdata)) {
1174 kdata = stack_kdata;
1175 } else {
1176 kdata = kzalloc(asize, GFP_KERNEL);
1177 if (!kdata) {
1178 retcode = -ENOMEM;
1179 goto out_err;
1180 }
1181 }
1182 }
1183
1184 if (cmd & IOC_IN) {
1185 if (copy_from_user(kdata, (void __user *)arg, usize)) {
1186 retcode = -EFAULT;
1187 goto out_err;
1188 }
1189 }
1190
1191 retcode = func(hpriv, kdata);
1192
1193 if ((cmd & IOC_OUT) && copy_to_user((void __user *)arg, kdata, usize))
1194 retcode = -EFAULT;
1195
1196 out_err:
1197 if (retcode)
1198 dev_dbg_ratelimited(dev, "error in ioctl: pid=%d, cmd=0x%02x, nr=0x%02x\n",
1199 task_pid_nr(current), cmd, nr);
1200
1201 if (kdata != stack_kdata)
1202 kfree(kdata);
1203
1204 return retcode;
1205 }
1206
hl_ioctl(struct file * filep,unsigned int cmd,unsigned long arg)1207 long hl_ioctl(struct file *filep, unsigned int cmd, unsigned long arg)
1208 {
1209 struct hl_fpriv *hpriv = filep->private_data;
1210 struct hl_device *hdev = hpriv->hdev;
1211 const struct hl_ioctl_desc *ioctl = NULL;
1212 unsigned int nr = _IOC_NR(cmd);
1213
1214 if (!hdev) {
1215 pr_err_ratelimited("Sending ioctl after device was removed! Please close FD\n");
1216 return -ENODEV;
1217 }
1218
1219 if ((nr >= HL_COMMAND_START) && (nr < HL_COMMAND_END)) {
1220 ioctl = &hl_ioctls[nr];
1221 } else {
1222 dev_dbg_ratelimited(hdev->dev, "invalid ioctl: pid=%d, nr=0x%02x\n",
1223 task_pid_nr(current), nr);
1224 return -ENOTTY;
1225 }
1226
1227 return _hl_ioctl(filep, cmd, arg, ioctl, hdev->dev);
1228 }
1229
hl_ioctl_control(struct file * filep,unsigned int cmd,unsigned long arg)1230 long hl_ioctl_control(struct file *filep, unsigned int cmd, unsigned long arg)
1231 {
1232 struct hl_fpriv *hpriv = filep->private_data;
1233 struct hl_device *hdev = hpriv->hdev;
1234 const struct hl_ioctl_desc *ioctl = NULL;
1235 unsigned int nr = _IOC_NR(cmd);
1236
1237 if (!hdev) {
1238 pr_err_ratelimited("Sending ioctl after device was removed! Please close FD\n");
1239 return -ENODEV;
1240 }
1241
1242 if (nr == _IOC_NR(HL_IOCTL_INFO)) {
1243 ioctl = &hl_ioctls_control[nr];
1244 } else {
1245 dev_dbg_ratelimited(hdev->dev_ctrl, "invalid ioctl: pid=%d, nr=0x%02x\n",
1246 task_pid_nr(current), nr);
1247 return -ENOTTY;
1248 }
1249
1250 return _hl_ioctl(filep, cmd, arg, ioctl, hdev->dev_ctrl);
1251 }
1252