xref: /openbmc/linux/arch/xtensa/include/asm/io.h (revision 1ab142d4)
1 /*
2  * include/asm-xtensa/io.h
3  *
4  * This file is subject to the terms and conditions of the GNU General Public
5  * License.  See the file "COPYING" in the main directory of this archive
6  * for more details.
7  *
8  * Copyright (C) 2001 - 2005 Tensilica Inc.
9  */
10 
11 #ifndef _XTENSA_IO_H
12 #define _XTENSA_IO_H
13 
14 #ifdef __KERNEL__
15 #include <asm/byteorder.h>
16 #include <asm/page.h>
17 #include <linux/kernel.h>
18 
19 #include <linux/types.h>
20 
21 #define XCHAL_KIO_CACHED_VADDR	0xe0000000
22 #define XCHAL_KIO_BYPASS_VADDR	0xf0000000
23 #define XCHAL_KIO_PADDR		0xf0000000
24 #define XCHAL_KIO_SIZE		0x10000000
25 
26 #define IOADDR(x)		(XCHAL_KIO_BYPASS_VADDR + (x))
27 
28 /*
29  * swap functions to change byte order from little-endian to big-endian and
30  * vice versa.
31  */
32 
33 static inline unsigned short _swapw (unsigned short v)
34 {
35 	return (v << 8) | (v >> 8);
36 }
37 
38 static inline unsigned int _swapl (unsigned int v)
39 {
40 	return (v << 24) | ((v & 0xff00) << 8) | ((v >> 8) & 0xff00) | (v >> 24);
41 }
42 
43 /*
44  * Change virtual addresses to physical addresses and vv.
45  * These are trivial on the 1:1 Linux/Xtensa mapping
46  */
47 
48 static inline unsigned long virt_to_phys(volatile void * address)
49 {
50 	return __pa(address);
51 }
52 
53 static inline void * phys_to_virt(unsigned long address)
54 {
55 	return __va(address);
56 }
57 
58 /*
59  * virt_to_bus and bus_to_virt are deprecated.
60  */
61 
62 #define virt_to_bus(x)	virt_to_phys(x)
63 #define bus_to_virt(x)	phys_to_virt(x)
64 
65 /*
66  * Return the virtual (cached) address for the specified bus memory.
67  * Note that we currently don't support any address outside the KIO segment.
68  */
69 
70 static inline void *ioremap(unsigned long offset, unsigned long size)
71 {
72 #ifdef CONFIG_MMU
73 	if (offset >= XCHAL_KIO_PADDR
74 	    && offset < XCHAL_KIO_PADDR + XCHAL_KIO_SIZE)
75 		return (void*)(offset-XCHAL_KIO_PADDR+XCHAL_KIO_BYPASS_VADDR);
76 	else
77 		BUG();
78 #else
79 	return (void *)offset;
80 #endif
81 }
82 
83 static inline void *ioremap_nocache(unsigned long offset, unsigned long size)
84 {
85 #ifdef CONFIG_MMU
86 	if (offset >= XCHAL_KIO_PADDR
87 	    && offset < XCHAL_KIO_PADDR + XCHAL_KIO_SIZE)
88 		return (void*)(offset-XCHAL_KIO_PADDR+XCHAL_KIO_CACHED_VADDR);
89 	else
90 		BUG();
91 #else
92 	return (void *)offset;
93 #endif
94 }
95 
96 static inline void iounmap(void *addr)
97 {
98 }
99 
100 /*
101  * Generic I/O
102  */
103 
104 #define readb(addr) \
105 	({ unsigned char __v = (*(volatile unsigned char *)(addr)); __v; })
106 #define readw(addr) \
107 	({ unsigned short __v = (*(volatile unsigned short *)(addr)); __v; })
108 #define readl(addr) \
109 	({ unsigned int __v = (*(volatile unsigned int *)(addr)); __v; })
110 #define writeb(b, addr) (void)((*(volatile unsigned char *)(addr)) = (b))
111 #define writew(b, addr) (void)((*(volatile unsigned short *)(addr)) = (b))
112 #define writel(b, addr) (void)((*(volatile unsigned int *)(addr)) = (b))
113 
114 static inline __u8 __raw_readb(const volatile void __iomem *addr)
115 {
116           return *(__force volatile __u8 *)(addr);
117 }
118 static inline __u16 __raw_readw(const volatile void __iomem *addr)
119 {
120           return *(__force volatile __u16 *)(addr);
121 }
122 static inline __u32 __raw_readl(const volatile void __iomem *addr)
123 {
124           return *(__force volatile __u32 *)(addr);
125 }
126 static inline void __raw_writeb(__u8 b, volatile void __iomem *addr)
127 {
128           *(__force volatile __u8 *)(addr) = b;
129 }
130 static inline void __raw_writew(__u16 b, volatile void __iomem *addr)
131 {
132           *(__force volatile __u16 *)(addr) = b;
133 }
134 static inline void __raw_writel(__u32 b, volatile void __iomem *addr)
135 {
136           *(__force volatile __u32 *)(addr) = b;
137 }
138 
139 /* These are the definitions for the x86 IO instructions
140  * inb/inw/inl/outb/outw/outl, the "string" versions
141  * insb/insw/insl/outsb/outsw/outsl, and the "pausing" versions
142  * inb_p/inw_p/...
143  * The macros don't do byte-swapping.
144  */
145 
146 #define inb(port)		readb((u8 *)((port)))
147 #define outb(val, port)		writeb((val),(u8 *)((unsigned long)(port)))
148 #define inw(port)		readw((u16 *)((port)))
149 #define outw(val, port)		writew((val),(u16 *)((unsigned long)(port)))
150 #define inl(port)		readl((u32 *)((port)))
151 #define outl(val, port)		writel((val),(u32 *)((unsigned long)(port)))
152 
153 #define inb_p(port)		inb((port))
154 #define outb_p(val, port)	outb((val), (port))
155 #define inw_p(port)		inw((port))
156 #define outw_p(val, port)	outw((val), (port))
157 #define inl_p(port)		inl((port))
158 #define outl_p(val, port)	outl((val), (port))
159 
160 extern void insb (unsigned long port, void *dst, unsigned long count);
161 extern void insw (unsigned long port, void *dst, unsigned long count);
162 extern void insl (unsigned long port, void *dst, unsigned long count);
163 extern void outsb (unsigned long port, const void *src, unsigned long count);
164 extern void outsw (unsigned long port, const void *src, unsigned long count);
165 extern void outsl (unsigned long port, const void *src, unsigned long count);
166 
167 #define IO_SPACE_LIMIT ~0
168 
169 #define memset_io(a,b,c)       memset((void *)(a),(b),(c))
170 #define memcpy_fromio(a,b,c)   memcpy((a),(void *)(b),(c))
171 #define memcpy_toio(a,b,c)      memcpy((void *)(a),(b),(c))
172 
173 /* At this point the Xtensa doesn't provide byte swap instructions */
174 
175 #ifdef __XTENSA_EB__
176 # define in_8(addr) (*(u8*)(addr))
177 # define in_le16(addr) _swapw(*(u16*)(addr))
178 # define in_le32(addr) _swapl(*(u32*)(addr))
179 # define out_8(b, addr) *(u8*)(addr) = (b)
180 # define out_le16(b, addr) *(u16*)(addr) = _swapw(b)
181 # define out_le32(b, addr) *(u32*)(addr) = _swapl(b)
182 #elif defined(__XTENSA_EL__)
183 # define in_8(addr)  (*(u8*)(addr))
184 # define in_le16(addr) (*(u16*)(addr))
185 # define in_le32(addr) (*(u32*)(addr))
186 # define out_8(b, addr) *(u8*)(addr) = (b)
187 # define out_le16(b, addr) *(u16*)(addr) = (b)
188 # define out_le32(b, addr) *(u32*)(addr) = (b)
189 #else
190 # error processor byte order undefined!
191 #endif
192 
193 
194 /*
195  * Convert a physical pointer to a virtual kernel pointer for /dev/mem access
196  */
197 #define xlate_dev_mem_ptr(p)    __va(p)
198 
199 /*
200  * Convert a virtual cached pointer to an uncached pointer
201  */
202 #define xlate_dev_kmem_ptr(p)   p
203 
204 
205 #endif	/* __KERNEL__ */
206 
207 #endif	/* _XTENSA_IO_H */
208