1 /* SPDX-License-Identifier: GPL-2.0 */ 2 #ifndef __SPARC64_IO_H 3 #define __SPARC64_IO_H 4 5 #include <linux/kernel.h> 6 #include <linux/compiler.h> 7 #include <linux/types.h> 8 9 #include <asm/page.h> /* IO address mapping routines need this */ 10 #include <asm/asi.h> 11 #include <asm-generic/pci_iomap.h> 12 13 /* BIO layer definitions. */ 14 extern unsigned long kern_base, kern_size; 15 16 /* __raw_{read,write}{b,w,l,q} uses direct access. 17 * Access the memory as big endian bypassing the cache 18 * by using ASI_PHYS_BYPASS_EC_E 19 */ 20 #define __raw_readb __raw_readb 21 static inline u8 __raw_readb(const volatile void __iomem *addr) 22 { 23 u8 ret; 24 25 __asm__ __volatile__("lduba\t[%1] %2, %0\t/* pci_raw_readb */" 26 : "=r" (ret) 27 : "r" (addr), "i" (ASI_PHYS_BYPASS_EC_E)); 28 29 return ret; 30 } 31 32 #define __raw_readw __raw_readw 33 static inline u16 __raw_readw(const volatile void __iomem *addr) 34 { 35 u16 ret; 36 37 __asm__ __volatile__("lduha\t[%1] %2, %0\t/* pci_raw_readw */" 38 : "=r" (ret) 39 : "r" (addr), "i" (ASI_PHYS_BYPASS_EC_E)); 40 41 return ret; 42 } 43 44 #define __raw_readl __raw_readl 45 static inline u32 __raw_readl(const volatile void __iomem *addr) 46 { 47 u32 ret; 48 49 __asm__ __volatile__("lduwa\t[%1] %2, %0\t/* pci_raw_readl */" 50 : "=r" (ret) 51 : "r" (addr), "i" (ASI_PHYS_BYPASS_EC_E)); 52 53 return ret; 54 } 55 56 #define __raw_readq __raw_readq 57 static inline u64 __raw_readq(const volatile void __iomem *addr) 58 { 59 u64 ret; 60 61 __asm__ __volatile__("ldxa\t[%1] %2, %0\t/* pci_raw_readq */" 62 : "=r" (ret) 63 : "r" (addr), "i" (ASI_PHYS_BYPASS_EC_E)); 64 65 return ret; 66 } 67 68 #define __raw_writeb __raw_writeb 69 static inline void __raw_writeb(u8 b, const volatile void __iomem *addr) 70 { 71 __asm__ __volatile__("stba\t%r0, [%1] %2\t/* pci_raw_writeb */" 72 : /* no outputs */ 73 : "Jr" (b), "r" (addr), "i" (ASI_PHYS_BYPASS_EC_E)); 74 } 75 76 #define __raw_writew __raw_writew 77 static inline void __raw_writew(u16 w, const volatile void __iomem *addr) 78 { 79 __asm__ __volatile__("stha\t%r0, [%1] %2\t/* pci_raw_writew */" 80 : /* no outputs */ 81 : "Jr" (w), "r" (addr), "i" (ASI_PHYS_BYPASS_EC_E)); 82 } 83 84 #define __raw_writel __raw_writel 85 static inline void __raw_writel(u32 l, const volatile void __iomem *addr) 86 { 87 __asm__ __volatile__("stwa\t%r0, [%1] %2\t/* pci_raw_writel */" 88 : /* no outputs */ 89 : "Jr" (l), "r" (addr), "i" (ASI_PHYS_BYPASS_EC_E)); 90 } 91 92 #define __raw_writeq __raw_writeq 93 static inline void __raw_writeq(u64 q, const volatile void __iomem *addr) 94 { 95 __asm__ __volatile__("stxa\t%r0, [%1] %2\t/* pci_raw_writeq */" 96 : /* no outputs */ 97 : "Jr" (q), "r" (addr), "i" (ASI_PHYS_BYPASS_EC_E)); 98 } 99 100 /* Memory functions, same as I/O accesses on Ultra. 101 * Access memory as little endian bypassing 102 * the cache by using ASI_PHYS_BYPASS_EC_E_L 103 */ 104 #define readb readb 105 #define readb_relaxed readb 106 static inline u8 readb(const volatile void __iomem *addr) 107 { u8 ret; 108 109 __asm__ __volatile__("lduba\t[%1] %2, %0\t/* pci_readb */" 110 : "=r" (ret) 111 : "r" (addr), "i" (ASI_PHYS_BYPASS_EC_E_L) 112 : "memory"); 113 return ret; 114 } 115 116 #define readw readw 117 #define readw_relaxed readw 118 static inline u16 readw(const volatile void __iomem *addr) 119 { u16 ret; 120 121 __asm__ __volatile__("lduha\t[%1] %2, %0\t/* pci_readw */" 122 : "=r" (ret) 123 : "r" (addr), "i" (ASI_PHYS_BYPASS_EC_E_L) 124 : "memory"); 125 126 return ret; 127 } 128 129 #define readl readl 130 #define readl_relaxed readl 131 static inline u32 readl(const volatile void __iomem *addr) 132 { u32 ret; 133 134 __asm__ __volatile__("lduwa\t[%1] %2, %0\t/* pci_readl */" 135 : "=r" (ret) 136 : "r" (addr), "i" (ASI_PHYS_BYPASS_EC_E_L) 137 : "memory"); 138 139 return ret; 140 } 141 142 #define readq readq 143 #define readq_relaxed readq 144 static inline u64 readq(const volatile void __iomem *addr) 145 { u64 ret; 146 147 __asm__ __volatile__("ldxa\t[%1] %2, %0\t/* pci_readq */" 148 : "=r" (ret) 149 : "r" (addr), "i" (ASI_PHYS_BYPASS_EC_E_L) 150 : "memory"); 151 152 return ret; 153 } 154 155 #define writeb writeb 156 #define writeb_relaxed writeb 157 static inline void writeb(u8 b, volatile void __iomem *addr) 158 { 159 __asm__ __volatile__("stba\t%r0, [%1] %2\t/* pci_writeb */" 160 : /* no outputs */ 161 : "Jr" (b), "r" (addr), "i" (ASI_PHYS_BYPASS_EC_E_L) 162 : "memory"); 163 } 164 165 #define writew writew 166 #define writew_relaxed writew 167 static inline void writew(u16 w, volatile void __iomem *addr) 168 { 169 __asm__ __volatile__("stha\t%r0, [%1] %2\t/* pci_writew */" 170 : /* no outputs */ 171 : "Jr" (w), "r" (addr), "i" (ASI_PHYS_BYPASS_EC_E_L) 172 : "memory"); 173 } 174 175 #define writel writel 176 #define writel_relaxed writel 177 static inline void writel(u32 l, volatile void __iomem *addr) 178 { 179 __asm__ __volatile__("stwa\t%r0, [%1] %2\t/* pci_writel */" 180 : /* no outputs */ 181 : "Jr" (l), "r" (addr), "i" (ASI_PHYS_BYPASS_EC_E_L) 182 : "memory"); 183 } 184 185 #define writeq writeq 186 #define writeq_relaxed writeq 187 static inline void writeq(u64 q, volatile void __iomem *addr) 188 { 189 __asm__ __volatile__("stxa\t%r0, [%1] %2\t/* pci_writeq */" 190 : /* no outputs */ 191 : "Jr" (q), "r" (addr), "i" (ASI_PHYS_BYPASS_EC_E_L) 192 : "memory"); 193 } 194 195 #define inb inb 196 static inline u8 inb(unsigned long addr) 197 { 198 return readb((volatile void __iomem *)addr); 199 } 200 201 #define inw inw 202 static inline u16 inw(unsigned long addr) 203 { 204 return readw((volatile void __iomem *)addr); 205 } 206 207 #define inl inl 208 static inline u32 inl(unsigned long addr) 209 { 210 return readl((volatile void __iomem *)addr); 211 } 212 213 #define outb outb 214 static inline void outb(u8 b, unsigned long addr) 215 { 216 writeb(b, (volatile void __iomem *)addr); 217 } 218 219 #define outw outw 220 static inline void outw(u16 w, unsigned long addr) 221 { 222 writew(w, (volatile void __iomem *)addr); 223 } 224 225 #define outl outl 226 static inline void outl(u32 l, unsigned long addr) 227 { 228 writel(l, (volatile void __iomem *)addr); 229 } 230 231 232 #define inb_p(__addr) inb(__addr) 233 #define outb_p(__b, __addr) outb(__b, __addr) 234 #define inw_p(__addr) inw(__addr) 235 #define outw_p(__w, __addr) outw(__w, __addr) 236 #define inl_p(__addr) inl(__addr) 237 #define outl_p(__l, __addr) outl(__l, __addr) 238 239 void outsb(unsigned long, const void *, unsigned long); 240 void outsw(unsigned long, const void *, unsigned long); 241 void outsl(unsigned long, const void *, unsigned long); 242 void insb(unsigned long, void *, unsigned long); 243 void insw(unsigned long, void *, unsigned long); 244 void insl(unsigned long, void *, unsigned long); 245 246 static inline void readsb(void __iomem *port, void *buf, unsigned long count) 247 { 248 insb((unsigned long __force)port, buf, count); 249 } 250 static inline void readsw(void __iomem *port, void *buf, unsigned long count) 251 { 252 insw((unsigned long __force)port, buf, count); 253 } 254 255 static inline void readsl(void __iomem *port, void *buf, unsigned long count) 256 { 257 insl((unsigned long __force)port, buf, count); 258 } 259 260 static inline void writesb(void __iomem *port, const void *buf, unsigned long count) 261 { 262 outsb((unsigned long __force)port, buf, count); 263 } 264 265 static inline void writesw(void __iomem *port, const void *buf, unsigned long count) 266 { 267 outsw((unsigned long __force)port, buf, count); 268 } 269 270 static inline void writesl(void __iomem *port, const void *buf, unsigned long count) 271 { 272 outsl((unsigned long __force)port, buf, count); 273 } 274 275 #define ioread8_rep(p,d,l) readsb(p,d,l) 276 #define ioread16_rep(p,d,l) readsw(p,d,l) 277 #define ioread32_rep(p,d,l) readsl(p,d,l) 278 #define iowrite8_rep(p,d,l) writesb(p,d,l) 279 #define iowrite16_rep(p,d,l) writesw(p,d,l) 280 #define iowrite32_rep(p,d,l) writesl(p,d,l) 281 282 /* Valid I/O Space regions are anywhere, because each PCI bus supported 283 * can live in an arbitrary area of the physical address range. 284 */ 285 #define IO_SPACE_LIMIT 0xffffffffffffffffUL 286 287 /* Now, SBUS variants, only difference from PCI is that we do 288 * not use little-endian ASIs. 289 */ 290 static inline u8 sbus_readb(const volatile void __iomem *addr) 291 { 292 return __raw_readb(addr); 293 } 294 295 static inline u16 sbus_readw(const volatile void __iomem *addr) 296 { 297 return __raw_readw(addr); 298 } 299 300 static inline u32 sbus_readl(const volatile void __iomem *addr) 301 { 302 return __raw_readl(addr); 303 } 304 305 static inline u64 sbus_readq(const volatile void __iomem *addr) 306 { 307 return __raw_readq(addr); 308 } 309 310 static inline void sbus_writeb(u8 b, volatile void __iomem *addr) 311 { 312 __raw_writeb(b, addr); 313 } 314 315 static inline void sbus_writew(u16 w, volatile void __iomem *addr) 316 { 317 __raw_writew(w, addr); 318 } 319 320 static inline void sbus_writel(u32 l, volatile void __iomem *addr) 321 { 322 __raw_writel(l, addr); 323 } 324 325 static inline void sbus_writeq(u64 q, volatile void __iomem *addr) 326 { 327 __raw_writeq(q, addr); 328 } 329 330 static inline void sbus_memset_io(volatile void __iomem *dst, int c, __kernel_size_t n) 331 { 332 while(n--) { 333 sbus_writeb(c, dst); 334 dst++; 335 } 336 } 337 338 static inline void memset_io(volatile void __iomem *dst, int c, __kernel_size_t n) 339 { 340 volatile void __iomem *d = dst; 341 342 while (n--) { 343 writeb(c, d); 344 d++; 345 } 346 } 347 348 static inline void sbus_memcpy_fromio(void *dst, const volatile void __iomem *src, 349 __kernel_size_t n) 350 { 351 char *d = dst; 352 353 while (n--) { 354 char tmp = sbus_readb(src); 355 *d++ = tmp; 356 src++; 357 } 358 } 359 360 361 static inline void memcpy_fromio(void *dst, const volatile void __iomem *src, 362 __kernel_size_t n) 363 { 364 char *d = dst; 365 366 while (n--) { 367 char tmp = readb(src); 368 *d++ = tmp; 369 src++; 370 } 371 } 372 373 static inline void sbus_memcpy_toio(volatile void __iomem *dst, const void *src, 374 __kernel_size_t n) 375 { 376 const char *s = src; 377 volatile void __iomem *d = dst; 378 379 while (n--) { 380 char tmp = *s++; 381 sbus_writeb(tmp, d); 382 d++; 383 } 384 } 385 386 static inline void memcpy_toio(volatile void __iomem *dst, const void *src, 387 __kernel_size_t n) 388 { 389 const char *s = src; 390 volatile void __iomem *d = dst; 391 392 while (n--) { 393 char tmp = *s++; 394 writeb(tmp, d); 395 d++; 396 } 397 } 398 399 #ifdef __KERNEL__ 400 401 /* On sparc64 we have the whole physical IO address space accessible 402 * using physically addressed loads and stores, so this does nothing. 403 */ 404 static inline void __iomem *ioremap(unsigned long offset, unsigned long size) 405 { 406 return (void __iomem *)offset; 407 } 408 409 #define ioremap_nocache(X,Y) ioremap((X),(Y)) 410 #define ioremap_uc(X,Y) ioremap((X),(Y)) 411 #define ioremap_wc(X,Y) ioremap((X),(Y)) 412 #define ioremap_wt(X,Y) ioremap((X),(Y)) 413 414 static inline void iounmap(volatile void __iomem *addr) 415 { 416 } 417 418 #define ioread8 readb 419 #define ioread16 readw 420 #define ioread16be __raw_readw 421 #define ioread32 readl 422 #define ioread32be __raw_readl 423 #define iowrite8 writeb 424 #define iowrite16 writew 425 #define iowrite16be __raw_writew 426 #define iowrite32 writel 427 #define iowrite32be __raw_writel 428 429 /* Create a virtual mapping cookie for an IO port range */ 430 void __iomem *ioport_map(unsigned long port, unsigned int nr); 431 void ioport_unmap(void __iomem *); 432 433 /* Create a virtual mapping cookie for a PCI BAR (memory or IO) */ 434 struct pci_dev; 435 void pci_iounmap(struct pci_dev *dev, void __iomem *); 436 437 static inline int sbus_can_dma_64bit(void) 438 { 439 return 1; 440 } 441 static inline int sbus_can_burst64(void) 442 { 443 return 1; 444 } 445 struct device; 446 void sbus_set_sbus64(struct device *, int); 447 448 /* 449 * Convert a physical pointer to a virtual kernel pointer for /dev/mem 450 * access 451 */ 452 #define xlate_dev_mem_ptr(p) __va(p) 453 454 /* 455 * Convert a virtual cached pointer to an uncached pointer 456 */ 457 #define xlate_dev_kmem_ptr(p) p 458 459 #endif 460 461 #endif /* !(__SPARC64_IO_H) */ 462