1 // SPDX-License-Identifier: GPL-2.0-only 2 /* 3 * Intel IXP4xx Queue Manager driver for Linux 4 * 5 * Copyright (C) 2007 Krzysztof Halasa <khc@pm.waw.pl> 6 */ 7 8 #include <linux/ioport.h> 9 #include <linux/interrupt.h> 10 #include <linux/kernel.h> 11 #include <linux/module.h> 12 #include <linux/of.h> 13 #include <linux/platform_device.h> 14 #include <linux/soc/ixp4xx/qmgr.h> 15 #include <linux/soc/ixp4xx/cpu.h> 16 17 static struct qmgr_regs __iomem *qmgr_regs; 18 static int qmgr_irq_1; 19 static int qmgr_irq_2; 20 static spinlock_t qmgr_lock; 21 static u32 used_sram_bitmap[4]; /* 128 16-dword pages */ 22 static void (*irq_handlers[QUEUES])(void *pdev); 23 static void *irq_pdevs[QUEUES]; 24 25 #if DEBUG_QMGR 26 char qmgr_queue_descs[QUEUES][32]; 27 #endif 28 29 void qmgr_put_entry(unsigned int queue, u32 val) 30 { 31 #if DEBUG_QMGR 32 BUG_ON(!qmgr_queue_descs[queue]); /* not yet requested */ 33 34 printk(KERN_DEBUG "Queue %s(%i) put %X\n", 35 qmgr_queue_descs[queue], queue, val); 36 #endif 37 __raw_writel(val, &qmgr_regs->acc[queue][0]); 38 } 39 40 u32 qmgr_get_entry(unsigned int queue) 41 { 42 u32 val; 43 val = __raw_readl(&qmgr_regs->acc[queue][0]); 44 #if DEBUG_QMGR 45 BUG_ON(!qmgr_queue_descs[queue]); /* not yet requested */ 46 47 printk(KERN_DEBUG "Queue %s(%i) get %X\n", 48 qmgr_queue_descs[queue], queue, val); 49 #endif 50 return val; 51 } 52 53 static int __qmgr_get_stat1(unsigned int queue) 54 { 55 return (__raw_readl(&qmgr_regs->stat1[queue >> 3]) 56 >> ((queue & 7) << 2)) & 0xF; 57 } 58 59 static int __qmgr_get_stat2(unsigned int queue) 60 { 61 BUG_ON(queue >= HALF_QUEUES); 62 return (__raw_readl(&qmgr_regs->stat2[queue >> 4]) 63 >> ((queue & 0xF) << 1)) & 0x3; 64 } 65 66 /** 67 * qmgr_stat_empty() - checks if a hardware queue is empty 68 * @queue: queue number 69 * 70 * Returns non-zero value if the queue is empty. 71 */ 72 int qmgr_stat_empty(unsigned int queue) 73 { 74 BUG_ON(queue >= HALF_QUEUES); 75 return __qmgr_get_stat1(queue) & QUEUE_STAT1_EMPTY; 76 } 77 78 /** 79 * qmgr_stat_below_low_watermark() - checks if a queue is below low watermark 80 * @queue: queue number 81 * 82 * Returns non-zero value if the queue is below low watermark. 83 */ 84 int qmgr_stat_below_low_watermark(unsigned int queue) 85 { 86 if (queue >= HALF_QUEUES) 87 return (__raw_readl(&qmgr_regs->statne_h) >> 88 (queue - HALF_QUEUES)) & 0x01; 89 return __qmgr_get_stat1(queue) & QUEUE_STAT1_NEARLY_EMPTY; 90 } 91 92 /** 93 * qmgr_stat_full() - checks if a hardware queue is full 94 * @queue: queue number 95 * 96 * Returns non-zero value if the queue is full. 97 */ 98 int qmgr_stat_full(unsigned int queue) 99 { 100 if (queue >= HALF_QUEUES) 101 return (__raw_readl(&qmgr_regs->statf_h) >> 102 (queue - HALF_QUEUES)) & 0x01; 103 return __qmgr_get_stat1(queue) & QUEUE_STAT1_FULL; 104 } 105 106 /** 107 * qmgr_stat_overflow() - checks if a hardware queue experienced overflow 108 * @queue: queue number 109 * 110 * Returns non-zero value if the queue experienced overflow. 111 */ 112 int qmgr_stat_overflow(unsigned int queue) 113 { 114 return __qmgr_get_stat2(queue) & QUEUE_STAT2_OVERFLOW; 115 } 116 117 void qmgr_set_irq(unsigned int queue, int src, 118 void (*handler)(void *pdev), void *pdev) 119 { 120 unsigned long flags; 121 122 spin_lock_irqsave(&qmgr_lock, flags); 123 if (queue < HALF_QUEUES) { 124 u32 __iomem *reg; 125 int bit; 126 BUG_ON(src > QUEUE_IRQ_SRC_NOT_FULL); 127 reg = &qmgr_regs->irqsrc[queue >> 3]; /* 8 queues per u32 */ 128 bit = (queue % 8) * 4; /* 3 bits + 1 reserved bit per queue */ 129 __raw_writel((__raw_readl(reg) & ~(7 << bit)) | (src << bit), 130 reg); 131 } else 132 /* IRQ source for queues 32-63 is fixed */ 133 BUG_ON(src != QUEUE_IRQ_SRC_NOT_NEARLY_EMPTY); 134 135 irq_handlers[queue] = handler; 136 irq_pdevs[queue] = pdev; 137 spin_unlock_irqrestore(&qmgr_lock, flags); 138 } 139 140 141 static irqreturn_t qmgr_irq1_a0(int irq, void *pdev) 142 { 143 int i, ret = 0; 144 u32 en_bitmap, src, stat; 145 146 /* ACK - it may clear any bits so don't rely on it */ 147 __raw_writel(0xFFFFFFFF, &qmgr_regs->irqstat[0]); 148 149 en_bitmap = __raw_readl(&qmgr_regs->irqen[0]); 150 while (en_bitmap) { 151 i = __fls(en_bitmap); /* number of the last "low" queue */ 152 en_bitmap &= ~BIT(i); 153 src = __raw_readl(&qmgr_regs->irqsrc[i >> 3]); 154 stat = __raw_readl(&qmgr_regs->stat1[i >> 3]); 155 if (src & 4) /* the IRQ condition is inverted */ 156 stat = ~stat; 157 if (stat & BIT(src & 3)) { 158 irq_handlers[i](irq_pdevs[i]); 159 ret = IRQ_HANDLED; 160 } 161 } 162 return ret; 163 } 164 165 166 static irqreturn_t qmgr_irq2_a0(int irq, void *pdev) 167 { 168 int i, ret = 0; 169 u32 req_bitmap; 170 171 /* ACK - it may clear any bits so don't rely on it */ 172 __raw_writel(0xFFFFFFFF, &qmgr_regs->irqstat[1]); 173 174 req_bitmap = __raw_readl(&qmgr_regs->irqen[1]) & 175 __raw_readl(&qmgr_regs->statne_h); 176 while (req_bitmap) { 177 i = __fls(req_bitmap); /* number of the last "high" queue */ 178 req_bitmap &= ~BIT(i); 179 irq_handlers[HALF_QUEUES + i](irq_pdevs[HALF_QUEUES + i]); 180 ret = IRQ_HANDLED; 181 } 182 return ret; 183 } 184 185 186 static irqreturn_t qmgr_irq(int irq, void *pdev) 187 { 188 int i, half = (irq == qmgr_irq_1 ? 0 : 1); 189 u32 req_bitmap = __raw_readl(&qmgr_regs->irqstat[half]); 190 191 if (!req_bitmap) 192 return 0; 193 __raw_writel(req_bitmap, &qmgr_regs->irqstat[half]); /* ACK */ 194 195 while (req_bitmap) { 196 i = __fls(req_bitmap); /* number of the last queue */ 197 req_bitmap &= ~BIT(i); 198 i += half * HALF_QUEUES; 199 irq_handlers[i](irq_pdevs[i]); 200 } 201 return IRQ_HANDLED; 202 } 203 204 205 void qmgr_enable_irq(unsigned int queue) 206 { 207 unsigned long flags; 208 int half = queue / 32; 209 u32 mask = 1 << (queue & (HALF_QUEUES - 1)); 210 211 spin_lock_irqsave(&qmgr_lock, flags); 212 __raw_writel(__raw_readl(&qmgr_regs->irqen[half]) | mask, 213 &qmgr_regs->irqen[half]); 214 spin_unlock_irqrestore(&qmgr_lock, flags); 215 } 216 217 void qmgr_disable_irq(unsigned int queue) 218 { 219 unsigned long flags; 220 int half = queue / 32; 221 u32 mask = 1 << (queue & (HALF_QUEUES - 1)); 222 223 spin_lock_irqsave(&qmgr_lock, flags); 224 __raw_writel(__raw_readl(&qmgr_regs->irqen[half]) & ~mask, 225 &qmgr_regs->irqen[half]); 226 __raw_writel(mask, &qmgr_regs->irqstat[half]); /* clear */ 227 spin_unlock_irqrestore(&qmgr_lock, flags); 228 } 229 230 static inline void shift_mask(u32 *mask) 231 { 232 mask[3] = mask[3] << 1 | mask[2] >> 31; 233 mask[2] = mask[2] << 1 | mask[1] >> 31; 234 mask[1] = mask[1] << 1 | mask[0] >> 31; 235 mask[0] <<= 1; 236 } 237 238 #if DEBUG_QMGR 239 int qmgr_request_queue(unsigned int queue, unsigned int len /* dwords */, 240 unsigned int nearly_empty_watermark, 241 unsigned int nearly_full_watermark, 242 const char *desc_format, const char* name) 243 #else 244 int __qmgr_request_queue(unsigned int queue, unsigned int len /* dwords */, 245 unsigned int nearly_empty_watermark, 246 unsigned int nearly_full_watermark) 247 #endif 248 { 249 u32 cfg, addr = 0, mask[4]; /* in 16-dwords */ 250 int err; 251 252 BUG_ON(queue >= QUEUES); 253 254 if ((nearly_empty_watermark | nearly_full_watermark) & ~7) 255 return -EINVAL; 256 257 switch (len) { 258 case 16: 259 cfg = 0 << 24; 260 mask[0] = 0x1; 261 break; 262 case 32: 263 cfg = 1 << 24; 264 mask[0] = 0x3; 265 break; 266 case 64: 267 cfg = 2 << 24; 268 mask[0] = 0xF; 269 break; 270 case 128: 271 cfg = 3 << 24; 272 mask[0] = 0xFF; 273 break; 274 default: 275 return -EINVAL; 276 } 277 278 cfg |= nearly_empty_watermark << 26; 279 cfg |= nearly_full_watermark << 29; 280 len /= 16; /* in 16-dwords: 1, 2, 4 or 8 */ 281 mask[1] = mask[2] = mask[3] = 0; 282 283 if (!try_module_get(THIS_MODULE)) 284 return -ENODEV; 285 286 spin_lock_irq(&qmgr_lock); 287 if (__raw_readl(&qmgr_regs->sram[queue])) { 288 err = -EBUSY; 289 goto err; 290 } 291 292 while (1) { 293 if (!(used_sram_bitmap[0] & mask[0]) && 294 !(used_sram_bitmap[1] & mask[1]) && 295 !(used_sram_bitmap[2] & mask[2]) && 296 !(used_sram_bitmap[3] & mask[3])) 297 break; /* found free space */ 298 299 addr++; 300 shift_mask(mask); 301 if (addr + len > ARRAY_SIZE(qmgr_regs->sram)) { 302 printk(KERN_ERR "qmgr: no free SRAM space for" 303 " queue %i\n", queue); 304 err = -ENOMEM; 305 goto err; 306 } 307 } 308 309 used_sram_bitmap[0] |= mask[0]; 310 used_sram_bitmap[1] |= mask[1]; 311 used_sram_bitmap[2] |= mask[2]; 312 used_sram_bitmap[3] |= mask[3]; 313 __raw_writel(cfg | (addr << 14), &qmgr_regs->sram[queue]); 314 #if DEBUG_QMGR 315 snprintf(qmgr_queue_descs[queue], sizeof(qmgr_queue_descs[0]), 316 desc_format, name); 317 printk(KERN_DEBUG "qmgr: requested queue %s(%i) addr = 0x%02X\n", 318 qmgr_queue_descs[queue], queue, addr); 319 #endif 320 spin_unlock_irq(&qmgr_lock); 321 return 0; 322 323 err: 324 spin_unlock_irq(&qmgr_lock); 325 module_put(THIS_MODULE); 326 return err; 327 } 328 329 void qmgr_release_queue(unsigned int queue) 330 { 331 u32 cfg, addr, mask[4]; 332 333 BUG_ON(queue >= QUEUES); /* not in valid range */ 334 335 spin_lock_irq(&qmgr_lock); 336 cfg = __raw_readl(&qmgr_regs->sram[queue]); 337 addr = (cfg >> 14) & 0xFF; 338 339 BUG_ON(!addr); /* not requested */ 340 341 switch ((cfg >> 24) & 3) { 342 case 0: mask[0] = 0x1; break; 343 case 1: mask[0] = 0x3; break; 344 case 2: mask[0] = 0xF; break; 345 case 3: mask[0] = 0xFF; break; 346 } 347 348 mask[1] = mask[2] = mask[3] = 0; 349 350 while (addr--) 351 shift_mask(mask); 352 353 #if DEBUG_QMGR 354 printk(KERN_DEBUG "qmgr: releasing queue %s(%i)\n", 355 qmgr_queue_descs[queue], queue); 356 qmgr_queue_descs[queue][0] = '\x0'; 357 #endif 358 359 while ((addr = qmgr_get_entry(queue))) 360 printk(KERN_ERR "qmgr: released queue %i not empty: 0x%08X\n", 361 queue, addr); 362 363 __raw_writel(0, &qmgr_regs->sram[queue]); 364 365 used_sram_bitmap[0] &= ~mask[0]; 366 used_sram_bitmap[1] &= ~mask[1]; 367 used_sram_bitmap[2] &= ~mask[2]; 368 used_sram_bitmap[3] &= ~mask[3]; 369 irq_handlers[queue] = NULL; /* catch IRQ bugs */ 370 spin_unlock_irq(&qmgr_lock); 371 372 module_put(THIS_MODULE); 373 } 374 375 static int ixp4xx_qmgr_probe(struct platform_device *pdev) 376 { 377 int i, err; 378 irq_handler_t handler1, handler2; 379 struct device *dev = &pdev->dev; 380 struct resource *res; 381 int irq1, irq2; 382 383 res = platform_get_resource(pdev, IORESOURCE_MEM, 0); 384 if (!res) 385 return -ENODEV; 386 qmgr_regs = devm_ioremap_resource(dev, res); 387 if (IS_ERR(qmgr_regs)) 388 return PTR_ERR(qmgr_regs); 389 390 irq1 = platform_get_irq(pdev, 0); 391 if (irq1 <= 0) 392 return irq1 ? irq1 : -EINVAL; 393 qmgr_irq_1 = irq1; 394 irq2 = platform_get_irq(pdev, 1); 395 if (irq2 <= 0) 396 return irq2 ? irq2 : -EINVAL; 397 qmgr_irq_2 = irq2; 398 399 /* reset qmgr registers */ 400 for (i = 0; i < 4; i++) { 401 __raw_writel(0x33333333, &qmgr_regs->stat1[i]); 402 __raw_writel(0, &qmgr_regs->irqsrc[i]); 403 } 404 for (i = 0; i < 2; i++) { 405 __raw_writel(0, &qmgr_regs->stat2[i]); 406 __raw_writel(0xFFFFFFFF, &qmgr_regs->irqstat[i]); /* clear */ 407 __raw_writel(0, &qmgr_regs->irqen[i]); 408 } 409 410 __raw_writel(0xFFFFFFFF, &qmgr_regs->statne_h); 411 __raw_writel(0, &qmgr_regs->statf_h); 412 413 for (i = 0; i < QUEUES; i++) 414 __raw_writel(0, &qmgr_regs->sram[i]); 415 416 if (cpu_is_ixp42x_rev_a0()) { 417 handler1 = qmgr_irq1_a0; 418 handler2 = qmgr_irq2_a0; 419 } else 420 handler1 = handler2 = qmgr_irq; 421 422 err = devm_request_irq(dev, irq1, handler1, 0, "IXP4xx Queue Manager", 423 NULL); 424 if (err) { 425 dev_err(dev, "failed to request IRQ%i (%i)\n", 426 irq1, err); 427 return err; 428 } 429 430 err = devm_request_irq(dev, irq2, handler2, 0, "IXP4xx Queue Manager", 431 NULL); 432 if (err) { 433 dev_err(dev, "failed to request IRQ%i (%i)\n", 434 irq2, err); 435 return err; 436 } 437 438 used_sram_bitmap[0] = 0xF; /* 4 first pages reserved for config */ 439 spin_lock_init(&qmgr_lock); 440 441 dev_info(dev, "IXP4xx Queue Manager initialized.\n"); 442 return 0; 443 } 444 445 static int ixp4xx_qmgr_remove(struct platform_device *pdev) 446 { 447 synchronize_irq(qmgr_irq_1); 448 synchronize_irq(qmgr_irq_2); 449 return 0; 450 } 451 452 static const struct of_device_id ixp4xx_qmgr_of_match[] = { 453 { 454 .compatible = "intel,ixp4xx-ahb-queue-manager", 455 }, 456 {}, 457 }; 458 459 static struct platform_driver ixp4xx_qmgr_driver = { 460 .driver = { 461 .name = "ixp4xx-qmgr", 462 .of_match_table = of_match_ptr(ixp4xx_qmgr_of_match), 463 }, 464 .probe = ixp4xx_qmgr_probe, 465 .remove = ixp4xx_qmgr_remove, 466 }; 467 module_platform_driver(ixp4xx_qmgr_driver); 468 469 MODULE_LICENSE("GPL v2"); 470 MODULE_AUTHOR("Krzysztof Halasa"); 471 472 EXPORT_SYMBOL(qmgr_put_entry); 473 EXPORT_SYMBOL(qmgr_get_entry); 474 EXPORT_SYMBOL(qmgr_stat_empty); 475 EXPORT_SYMBOL(qmgr_stat_below_low_watermark); 476 EXPORT_SYMBOL(qmgr_stat_full); 477 EXPORT_SYMBOL(qmgr_stat_overflow); 478 EXPORT_SYMBOL(qmgr_set_irq); 479 EXPORT_SYMBOL(qmgr_enable_irq); 480 EXPORT_SYMBOL(qmgr_disable_irq); 481 #if DEBUG_QMGR 482 EXPORT_SYMBOL(qmgr_queue_descs); 483 EXPORT_SYMBOL(qmgr_request_queue); 484 #else 485 EXPORT_SYMBOL(__qmgr_request_queue); 486 #endif 487 EXPORT_SYMBOL(qmgr_release_queue); 488