1 // SPDX-License-Identifier: GPL-2.0-only
2 /* Atlantic Network Driver
3  *
4  * Copyright (C) 2014-2019 aQuantia Corporation
5  * Copyright (C) 2019-2020 Marvell International Ltd.
6  */
7 
8 /* File aq_pci_func.c: Definition of PCI functions. */
9 
10 #include <linux/interrupt.h>
11 #include <linux/module.h>
12 
13 #include "aq_main.h"
14 #include "aq_nic.h"
15 #include "aq_vec.h"
16 #include "aq_hw.h"
17 #include "aq_pci_func.h"
18 #include "hw_atl/hw_atl_a0.h"
19 #include "hw_atl/hw_atl_b0.h"
20 #include "hw_atl2/hw_atl2.h"
21 #include "aq_filters.h"
22 #include "aq_drvinfo.h"
23 #include "aq_macsec.h"
24 
25 static const struct pci_device_id aq_pci_tbl[] = {
26 	{ PCI_VDEVICE(AQUANTIA, AQ_DEVICE_ID_0001), },
27 	{ PCI_VDEVICE(AQUANTIA, AQ_DEVICE_ID_D100), },
28 	{ PCI_VDEVICE(AQUANTIA, AQ_DEVICE_ID_D107), },
29 	{ PCI_VDEVICE(AQUANTIA, AQ_DEVICE_ID_D108), },
30 	{ PCI_VDEVICE(AQUANTIA, AQ_DEVICE_ID_D109), },
31 
32 	{ PCI_VDEVICE(AQUANTIA, AQ_DEVICE_ID_AQC100), },
33 	{ PCI_VDEVICE(AQUANTIA, AQ_DEVICE_ID_AQC107), },
34 	{ PCI_VDEVICE(AQUANTIA, AQ_DEVICE_ID_AQC108), },
35 	{ PCI_VDEVICE(AQUANTIA, AQ_DEVICE_ID_AQC109), },
36 	{ PCI_VDEVICE(AQUANTIA, AQ_DEVICE_ID_AQC111), },
37 	{ PCI_VDEVICE(AQUANTIA, AQ_DEVICE_ID_AQC112), },
38 
39 	{ PCI_VDEVICE(AQUANTIA, AQ_DEVICE_ID_AQC100S), },
40 	{ PCI_VDEVICE(AQUANTIA, AQ_DEVICE_ID_AQC107S), },
41 	{ PCI_VDEVICE(AQUANTIA, AQ_DEVICE_ID_AQC108S), },
42 	{ PCI_VDEVICE(AQUANTIA, AQ_DEVICE_ID_AQC109S), },
43 	{ PCI_VDEVICE(AQUANTIA, AQ_DEVICE_ID_AQC111S), },
44 	{ PCI_VDEVICE(AQUANTIA, AQ_DEVICE_ID_AQC112S), },
45 
46 	{ PCI_VDEVICE(AQUANTIA, AQ_DEVICE_ID_AQC113DEV), },
47 	{ PCI_VDEVICE(AQUANTIA, AQ_DEVICE_ID_AQC113CS), },
48 	{ PCI_VDEVICE(AQUANTIA, AQ_DEVICE_ID_AQC114CS), },
49 	{ PCI_VDEVICE(AQUANTIA, AQ_DEVICE_ID_AQC113), },
50 	{ PCI_VDEVICE(AQUANTIA, AQ_DEVICE_ID_AQC113C), },
51 	{ PCI_VDEVICE(AQUANTIA, AQ_DEVICE_ID_AQC115C), },
52 	{ PCI_VDEVICE(AQUANTIA, AQ_DEVICE_ID_AQC113CA), },
53 	{ PCI_VDEVICE(AQUANTIA, AQ_DEVICE_ID_AQC116C), },
54 
55 	{}
56 };
57 
58 static const struct aq_board_revision_s hw_atl_boards[] = {
59 	{ AQ_DEVICE_ID_0001,	AQ_HWREV_1,	&hw_atl_ops_a0, &hw_atl_a0_caps_aqc107, },
60 	{ AQ_DEVICE_ID_D100,	AQ_HWREV_1,	&hw_atl_ops_a0, &hw_atl_a0_caps_aqc100, },
61 	{ AQ_DEVICE_ID_D107,	AQ_HWREV_1,	&hw_atl_ops_a0, &hw_atl_a0_caps_aqc107, },
62 	{ AQ_DEVICE_ID_D108,	AQ_HWREV_1,	&hw_atl_ops_a0, &hw_atl_a0_caps_aqc108, },
63 	{ AQ_DEVICE_ID_D109,	AQ_HWREV_1,	&hw_atl_ops_a0, &hw_atl_a0_caps_aqc109, },
64 
65 	{ AQ_DEVICE_ID_0001,	AQ_HWREV_2,	&hw_atl_ops_b0, &hw_atl_b0_caps_aqc107, },
66 	{ AQ_DEVICE_ID_D100,	AQ_HWREV_2,	&hw_atl_ops_b0, &hw_atl_b0_caps_aqc100, },
67 	{ AQ_DEVICE_ID_D107,	AQ_HWREV_2,	&hw_atl_ops_b0, &hw_atl_b0_caps_aqc107, },
68 	{ AQ_DEVICE_ID_D108,	AQ_HWREV_2,	&hw_atl_ops_b0, &hw_atl_b0_caps_aqc108, },
69 	{ AQ_DEVICE_ID_D109,	AQ_HWREV_2,	&hw_atl_ops_b0, &hw_atl_b0_caps_aqc109, },
70 
71 	{ AQ_DEVICE_ID_AQC100,	AQ_HWREV_ANY,	&hw_atl_ops_b1, &hw_atl_b0_caps_aqc100, },
72 	{ AQ_DEVICE_ID_AQC107,	AQ_HWREV_ANY,	&hw_atl_ops_b1, &hw_atl_b0_caps_aqc107, },
73 	{ AQ_DEVICE_ID_AQC108,	AQ_HWREV_ANY,	&hw_atl_ops_b1, &hw_atl_b0_caps_aqc108, },
74 	{ AQ_DEVICE_ID_AQC109,	AQ_HWREV_ANY,	&hw_atl_ops_b1, &hw_atl_b0_caps_aqc109, },
75 	{ AQ_DEVICE_ID_AQC111,	AQ_HWREV_ANY,	&hw_atl_ops_b1, &hw_atl_b0_caps_aqc111, },
76 	{ AQ_DEVICE_ID_AQC112,	AQ_HWREV_ANY,	&hw_atl_ops_b1, &hw_atl_b0_caps_aqc112, },
77 
78 	{ AQ_DEVICE_ID_AQC100S,	AQ_HWREV_ANY,	&hw_atl_ops_b1, &hw_atl_b0_caps_aqc100s, },
79 	{ AQ_DEVICE_ID_AQC107S,	AQ_HWREV_ANY,	&hw_atl_ops_b1, &hw_atl_b0_caps_aqc107s, },
80 	{ AQ_DEVICE_ID_AQC108S,	AQ_HWREV_ANY,	&hw_atl_ops_b1, &hw_atl_b0_caps_aqc108s, },
81 	{ AQ_DEVICE_ID_AQC109S,	AQ_HWREV_ANY,	&hw_atl_ops_b1, &hw_atl_b0_caps_aqc109s, },
82 	{ AQ_DEVICE_ID_AQC111S,	AQ_HWREV_ANY,	&hw_atl_ops_b1, &hw_atl_b0_caps_aqc111s, },
83 	{ AQ_DEVICE_ID_AQC112S,	AQ_HWREV_ANY,	&hw_atl_ops_b1, &hw_atl_b0_caps_aqc112s, },
84 
85 	{ AQ_DEVICE_ID_AQC113DEV,	AQ_HWREV_ANY,	&hw_atl2_ops, &hw_atl2_caps_aqc113, },
86 	{ AQ_DEVICE_ID_AQC113,		AQ_HWREV_ANY,	&hw_atl2_ops, &hw_atl2_caps_aqc113, },
87 	{ AQ_DEVICE_ID_AQC113CS,	AQ_HWREV_ANY,	&hw_atl2_ops, &hw_atl2_caps_aqc113, },
88 	{ AQ_DEVICE_ID_AQC114CS,	AQ_HWREV_ANY,	&hw_atl2_ops, &hw_atl2_caps_aqc113, },
89 	{ AQ_DEVICE_ID_AQC113C,		AQ_HWREV_ANY,	&hw_atl2_ops, &hw_atl2_caps_aqc113, },
90 	{ AQ_DEVICE_ID_AQC115C,		AQ_HWREV_ANY,	&hw_atl2_ops, &hw_atl2_caps_aqc115c, },
91 	{ AQ_DEVICE_ID_AQC113CA,	AQ_HWREV_ANY,	&hw_atl2_ops, &hw_atl2_caps_aqc113, },
92 	{ AQ_DEVICE_ID_AQC116C,		AQ_HWREV_ANY,	&hw_atl2_ops, &hw_atl2_caps_aqc116c, },
93 
94 };
95 
96 MODULE_DEVICE_TABLE(pci, aq_pci_tbl);
97 
98 static int aq_pci_probe_get_hw_by_id(struct pci_dev *pdev,
99 				     const struct aq_hw_ops **ops,
100 				     const struct aq_hw_caps_s **caps)
101 {
102 	int i;
103 
104 	if (pdev->vendor != PCI_VENDOR_ID_AQUANTIA)
105 		return -EINVAL;
106 
107 	for (i = 0; i < ARRAY_SIZE(hw_atl_boards); i++) {
108 		if (hw_atl_boards[i].devid == pdev->device &&
109 		    (hw_atl_boards[i].revision == AQ_HWREV_ANY ||
110 		     hw_atl_boards[i].revision == pdev->revision)) {
111 			*ops = hw_atl_boards[i].ops;
112 			*caps = hw_atl_boards[i].caps;
113 			break;
114 		}
115 	}
116 
117 	if (i == ARRAY_SIZE(hw_atl_boards))
118 		return -EINVAL;
119 
120 	return 0;
121 }
122 
123 static int aq_pci_func_init(struct pci_dev *pdev)
124 {
125 	int err;
126 
127 	err = dma_set_mask_and_coherent(&pdev->dev, DMA_BIT_MASK(64));
128 	if (err)
129 		err = dma_set_mask_and_coherent(&pdev->dev, DMA_BIT_MASK(32));
130 	if (err) {
131 		err = -ENOSR;
132 		goto err_exit;
133 	}
134 
135 	err = pci_request_regions(pdev, AQ_CFG_DRV_NAME "_mmio");
136 	if (err < 0)
137 		goto err_exit;
138 
139 	pci_set_master(pdev);
140 
141 	return 0;
142 
143 err_exit:
144 	return err;
145 }
146 
147 int aq_pci_func_alloc_irq(struct aq_nic_s *self, unsigned int i,
148 			  char *name, irq_handler_t irq_handler,
149 			  void *irq_arg, cpumask_t *affinity_mask)
150 {
151 	struct pci_dev *pdev = self->pdev;
152 	int err;
153 
154 	if (pdev->msix_enabled || pdev->msi_enabled)
155 		err = request_irq(pci_irq_vector(pdev, i), irq_handler, 0,
156 				  name, irq_arg);
157 	else
158 		err = request_irq(pci_irq_vector(pdev, i), aq_vec_isr_legacy,
159 				  IRQF_SHARED, name, irq_arg);
160 
161 	if (err >= 0) {
162 		self->msix_entry_mask |= (1 << i);
163 
164 		if (pdev->msix_enabled && affinity_mask)
165 			irq_set_affinity_hint(pci_irq_vector(pdev, i),
166 					      affinity_mask);
167 	}
168 
169 	return err;
170 }
171 
172 void aq_pci_func_free_irqs(struct aq_nic_s *self)
173 {
174 	struct pci_dev *pdev = self->pdev;
175 	unsigned int i;
176 	void *irq_data;
177 
178 	for (i = 32U; i--;) {
179 		if (!((1U << i) & self->msix_entry_mask))
180 			continue;
181 		if (self->aq_nic_cfg.link_irq_vec &&
182 		    i == self->aq_nic_cfg.link_irq_vec)
183 			irq_data = self;
184 		else if (i < AQ_CFG_VECS_MAX)
185 			irq_data = self->aq_vec[i];
186 		else
187 			continue;
188 
189 		if (pdev->msix_enabled)
190 			irq_set_affinity_hint(pci_irq_vector(pdev, i), NULL);
191 		free_irq(pci_irq_vector(pdev, i), irq_data);
192 		self->msix_entry_mask &= ~(1U << i);
193 	}
194 }
195 
196 unsigned int aq_pci_func_get_irq_type(struct aq_nic_s *self)
197 {
198 	if (self->pdev->msix_enabled)
199 		return AQ_HW_IRQ_MSIX;
200 	if (self->pdev->msi_enabled)
201 		return AQ_HW_IRQ_MSI;
202 
203 	return AQ_HW_IRQ_LEGACY;
204 }
205 
206 static void aq_pci_free_irq_vectors(struct aq_nic_s *self)
207 {
208 	pci_free_irq_vectors(self->pdev);
209 }
210 
211 static int aq_pci_probe(struct pci_dev *pdev,
212 			const struct pci_device_id *pci_id)
213 {
214 	struct net_device *ndev;
215 	resource_size_t mmio_pa;
216 	struct aq_nic_s *self;
217 	u32 numvecs;
218 	u32 bar;
219 	int err;
220 
221 	err = pci_enable_device(pdev);
222 	if (err)
223 		return err;
224 
225 	err = aq_pci_func_init(pdev);
226 	if (err)
227 		goto err_pci_func;
228 
229 	ndev = aq_ndev_alloc();
230 	if (!ndev) {
231 		err = -ENOMEM;
232 		goto err_ndev;
233 	}
234 
235 	self = netdev_priv(ndev);
236 	self->pdev = pdev;
237 	SET_NETDEV_DEV(ndev, &pdev->dev);
238 	pci_set_drvdata(pdev, self);
239 
240 	mutex_init(&self->fwreq_mutex);
241 
242 	err = aq_pci_probe_get_hw_by_id(pdev, &self->aq_hw_ops,
243 					&aq_nic_get_cfg(self)->aq_hw_caps);
244 	if (err)
245 		goto err_ioremap;
246 
247 	self->aq_hw = kzalloc(sizeof(*self->aq_hw), GFP_KERNEL);
248 	if (!self->aq_hw) {
249 		err = -ENOMEM;
250 		goto err_ioremap;
251 	}
252 	self->aq_hw->aq_nic_cfg = aq_nic_get_cfg(self);
253 	if (self->aq_hw->aq_nic_cfg->aq_hw_caps->priv_data_len) {
254 		int len = self->aq_hw->aq_nic_cfg->aq_hw_caps->priv_data_len;
255 
256 		self->aq_hw->priv = kzalloc(len, GFP_KERNEL);
257 		if (!self->aq_hw->priv) {
258 			err = -ENOMEM;
259 			goto err_free_aq_hw;
260 		}
261 	}
262 
263 	for (bar = 0; bar < 4; ++bar) {
264 		if (IORESOURCE_MEM & pci_resource_flags(pdev, bar)) {
265 			resource_size_t reg_sz;
266 
267 			mmio_pa = pci_resource_start(pdev, bar);
268 			if (mmio_pa == 0U) {
269 				err = -EIO;
270 				goto err_free_aq_hw_priv;
271 			}
272 
273 			reg_sz = pci_resource_len(pdev, bar);
274 			if ((reg_sz <= 24 /*ATL_REGS_SIZE*/)) {
275 				err = -EIO;
276 				goto err_free_aq_hw_priv;
277 			}
278 
279 			self->aq_hw->mmio = ioremap(mmio_pa, reg_sz);
280 			if (!self->aq_hw->mmio) {
281 				err = -EIO;
282 				goto err_free_aq_hw_priv;
283 			}
284 			break;
285 		}
286 	}
287 
288 	if (bar == 4) {
289 		err = -EIO;
290 		goto err_free_aq_hw_priv;
291 	}
292 
293 	numvecs = min((u8)AQ_CFG_VECS_DEF,
294 		      aq_nic_get_cfg(self)->aq_hw_caps->msix_irqs);
295 	numvecs = min(numvecs, num_online_cpus());
296 	/* Request IRQ vector for PTP */
297 	numvecs += 1;
298 
299 	numvecs += AQ_HW_SERVICE_IRQS;
300 	/*enable interrupts */
301 #if !AQ_CFG_FORCE_LEGACY_INT
302 	err = pci_alloc_irq_vectors(self->pdev, 1, numvecs,
303 				    PCI_IRQ_MSIX | PCI_IRQ_MSI |
304 				    PCI_IRQ_LEGACY);
305 
306 	if (err < 0)
307 		goto err_hwinit;
308 	numvecs = err;
309 #endif
310 	self->irqvecs = numvecs;
311 
312 	/* net device init */
313 	aq_nic_cfg_start(self);
314 
315 	aq_nic_ndev_init(self);
316 
317 	err = aq_nic_ndev_register(self);
318 	if (err < 0)
319 		goto err_register;
320 
321 	aq_drvinfo_init(ndev);
322 
323 	return 0;
324 
325 err_register:
326 	aq_nic_free_vectors(self);
327 	aq_pci_free_irq_vectors(self);
328 err_hwinit:
329 	iounmap(self->aq_hw->mmio);
330 err_free_aq_hw_priv:
331 	kfree(self->aq_hw->priv);
332 err_free_aq_hw:
333 	kfree(self->aq_hw);
334 err_ioremap:
335 	free_netdev(ndev);
336 err_ndev:
337 	pci_release_regions(pdev);
338 err_pci_func:
339 	pci_disable_device(pdev);
340 
341 	return err;
342 }
343 
344 static void aq_pci_remove(struct pci_dev *pdev)
345 {
346 	struct aq_nic_s *self = pci_get_drvdata(pdev);
347 
348 	if (self->ndev) {
349 		aq_clear_rxnfc_all_rules(self);
350 		if (self->ndev->reg_state == NETREG_REGISTERED)
351 			unregister_netdev(self->ndev);
352 
353 #if IS_ENABLED(CONFIG_MACSEC)
354 		aq_macsec_free(self);
355 #endif
356 		aq_nic_free_vectors(self);
357 		aq_pci_free_irq_vectors(self);
358 		iounmap(self->aq_hw->mmio);
359 		kfree(self->aq_hw->priv);
360 		kfree(self->aq_hw);
361 		pci_release_regions(pdev);
362 		free_netdev(self->ndev);
363 	}
364 
365 	pci_disable_device(pdev);
366 }
367 
368 static void aq_pci_shutdown(struct pci_dev *pdev)
369 {
370 	struct aq_nic_s *self = pci_get_drvdata(pdev);
371 
372 	aq_nic_shutdown(self);
373 
374 	pci_disable_device(pdev);
375 
376 	if (system_state == SYSTEM_POWER_OFF) {
377 		pci_wake_from_d3(pdev, false);
378 		pci_set_power_state(pdev, PCI_D3hot);
379 	}
380 }
381 
382 static int aq_suspend_common(struct device *dev, bool deep)
383 {
384 	struct aq_nic_s *nic = pci_get_drvdata(to_pci_dev(dev));
385 
386 	rtnl_lock();
387 
388 	nic->power_state = AQ_HW_POWER_STATE_D3;
389 	netif_device_detach(nic->ndev);
390 	netif_tx_stop_all_queues(nic->ndev);
391 
392 	if (netif_running(nic->ndev))
393 		aq_nic_stop(nic);
394 
395 	if (deep) {
396 		aq_nic_deinit(nic, !nic->aq_hw->aq_nic_cfg->wol);
397 		aq_nic_set_power(nic);
398 	}
399 
400 	rtnl_unlock();
401 
402 	return 0;
403 }
404 
405 static int atl_resume_common(struct device *dev, bool deep)
406 {
407 	struct pci_dev *pdev = to_pci_dev(dev);
408 	struct aq_nic_s *nic;
409 	int ret = 0;
410 
411 	nic = pci_get_drvdata(pdev);
412 
413 	rtnl_lock();
414 
415 	pci_set_power_state(pdev, PCI_D0);
416 	pci_restore_state(pdev);
417 
418 	if (deep) {
419 		/* Reinitialize Nic/Vecs objects */
420 		aq_nic_deinit(nic, !nic->aq_hw->aq_nic_cfg->wol);
421 	}
422 
423 	if (netif_running(nic->ndev)) {
424 		ret = aq_nic_init(nic);
425 		if (ret)
426 			goto err_exit;
427 
428 		ret = aq_nic_start(nic);
429 		if (ret)
430 			goto err_exit;
431 	}
432 
433 	netif_device_attach(nic->ndev);
434 	netif_tx_start_all_queues(nic->ndev);
435 
436 err_exit:
437 	if (ret < 0)
438 		aq_nic_deinit(nic, true);
439 
440 	rtnl_unlock();
441 
442 	return ret;
443 }
444 
445 static int aq_pm_freeze(struct device *dev)
446 {
447 	return aq_suspend_common(dev, true);
448 }
449 
450 static int aq_pm_suspend_poweroff(struct device *dev)
451 {
452 	return aq_suspend_common(dev, false);
453 }
454 
455 static int aq_pm_thaw(struct device *dev)
456 {
457 	return atl_resume_common(dev, true);
458 }
459 
460 static int aq_pm_resume_restore(struct device *dev)
461 {
462 	return atl_resume_common(dev, false);
463 }
464 
465 static const struct dev_pm_ops aq_pm_ops = {
466 	.suspend = aq_pm_suspend_poweroff,
467 	.poweroff = aq_pm_suspend_poweroff,
468 	.freeze = aq_pm_freeze,
469 	.resume = aq_pm_resume_restore,
470 	.restore = aq_pm_resume_restore,
471 	.thaw = aq_pm_thaw,
472 };
473 
474 static struct pci_driver aq_pci_ops = {
475 	.name = AQ_CFG_DRV_NAME,
476 	.id_table = aq_pci_tbl,
477 	.probe = aq_pci_probe,
478 	.remove = aq_pci_remove,
479 	.shutdown = aq_pci_shutdown,
480 #ifdef CONFIG_PM
481 	.driver.pm = &aq_pm_ops,
482 #endif
483 };
484 
485 int aq_pci_func_register_driver(void)
486 {
487 	return pci_register_driver(&aq_pci_ops);
488 }
489 
490 void aq_pci_func_unregister_driver(void)
491 {
492 	pci_unregister_driver(&aq_pci_ops);
493 }
494 
495