blob: 103270d5a9fc630d4f06f64f5173ebfe6e76c48c [file] [log] [blame]
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001/* SPDX-License-Identifier: GPL-2.0 */
2#ifndef __ALPHA_IO_H
3#define __ALPHA_IO_H
4
5#ifdef __KERNEL__
6
7#include <linux/kernel.h>
8#include <linux/mm.h>
9#include <asm/compiler.h>
10#include <asm/pgtable.h>
11#include <asm/machvec.h>
12#include <asm/hwrpb.h>
13
14/* The generic header contains only prototypes. Including it ensures that
15 the implementation we have here matches that interface. */
16#include <asm-generic/iomap.h>
17
18/* We don't use IO slowdowns on the Alpha, but.. */
19#define __SLOW_DOWN_IO do { } while (0)
20#define SLOW_DOWN_IO do { } while (0)
21
22/*
23 * Virtual -> physical identity mapping starts at this offset
24 */
25#ifdef USE_48_BIT_KSEG
26#define IDENT_ADDR 0xffff800000000000UL
27#else
28#define IDENT_ADDR 0xfffffc0000000000UL
29#endif
30
31/*
32 * We try to avoid hae updates (thus the cache), but when we
33 * do need to update the hae, we need to do it atomically, so
34 * that any interrupts wouldn't get confused with the hae
35 * register not being up-to-date with respect to the hardware
36 * value.
37 */
38extern inline void __set_hae(unsigned long new_hae)
39{
40 unsigned long flags = swpipl(IPL_MAX);
41
42 barrier();
43
44 alpha_mv.hae_cache = new_hae;
45 *alpha_mv.hae_register = new_hae;
46 mb();
47 /* Re-read to make sure it was written. */
48 new_hae = *alpha_mv.hae_register;
49
50 setipl(flags);
51 barrier();
52}
53
54extern inline void set_hae(unsigned long new_hae)
55{
56 if (new_hae != alpha_mv.hae_cache)
57 __set_hae(new_hae);
58}
59
60/*
61 * Change virtual addresses to physical addresses and vv.
62 */
63#ifdef USE_48_BIT_KSEG
64static inline unsigned long virt_to_phys(void *address)
65{
66 return (unsigned long)address - IDENT_ADDR;
67}
68
69static inline void * phys_to_virt(unsigned long address)
70{
71 return (void *) (address + IDENT_ADDR);
72}
73#else
74static inline unsigned long virt_to_phys(void *address)
75{
76 unsigned long phys = (unsigned long)address;
77
78 /* Sign-extend from bit 41. */
79 phys <<= (64 - 41);
80 phys = (long)phys >> (64 - 41);
81
82 /* Crop to the physical address width of the processor. */
83 phys &= (1ul << hwrpb->pa_bits) - 1;
84
85 return phys;
86}
87
88static inline void * phys_to_virt(unsigned long address)
89{
90 return (void *)(IDENT_ADDR + (address & ((1ul << 41) - 1)));
91}
92#endif
93
94#define page_to_phys(page) page_to_pa(page)
95
Andrew Scullb4b6d4a2019-01-02 15:54:55 +000096/* Maximum PIO space address supported? */
97#define IO_SPACE_LIMIT 0xffff
98
99/*
100 * Change addresses as seen by the kernel (virtual) to addresses as
101 * seen by a device (bus), and vice versa.
102 *
103 * Note that this only works for a limited range of kernel addresses,
104 * and very well may not span all memory. Consider this interface
105 * deprecated in favour of the DMA-mapping API.
106 */
107extern unsigned long __direct_map_base;
108extern unsigned long __direct_map_size;
109
110static inline unsigned long __deprecated virt_to_bus(void *address)
111{
112 unsigned long phys = virt_to_phys(address);
113 unsigned long bus = phys + __direct_map_base;
114 return phys <= __direct_map_size ? bus : 0;
115}
116#define isa_virt_to_bus virt_to_bus
117
118static inline void * __deprecated bus_to_virt(unsigned long address)
119{
120 void *virt;
121
122 /* This check is a sanity check but also ensures that bus address 0
123 maps to virtual address 0 which is useful to detect null pointers
124 (the NCR driver is much simpler if NULL pointers are preserved). */
125 address -= __direct_map_base;
126 virt = phys_to_virt(address);
127 return (long)address <= 0 ? NULL : virt;
128}
129#define isa_bus_to_virt bus_to_virt
130
131/*
132 * There are different chipsets to interface the Alpha CPUs to the world.
133 */
134
135#define IO_CONCAT(a,b) _IO_CONCAT(a,b)
136#define _IO_CONCAT(a,b) a ## _ ## b
137
138#ifdef CONFIG_ALPHA_GENERIC
139
140/* In a generic kernel, we always go through the machine vector. */
141
142#define REMAP1(TYPE, NAME, QUAL) \
143static inline TYPE generic_##NAME(QUAL void __iomem *addr) \
144{ \
145 return alpha_mv.mv_##NAME(addr); \
146}
147
148#define REMAP2(TYPE, NAME, QUAL) \
149static inline void generic_##NAME(TYPE b, QUAL void __iomem *addr) \
150{ \
151 alpha_mv.mv_##NAME(b, addr); \
152}
153
154REMAP1(unsigned int, ioread8, /**/)
155REMAP1(unsigned int, ioread16, /**/)
156REMAP1(unsigned int, ioread32, /**/)
157REMAP1(u8, readb, const volatile)
158REMAP1(u16, readw, const volatile)
159REMAP1(u32, readl, const volatile)
160REMAP1(u64, readq, const volatile)
161
162REMAP2(u8, iowrite8, /**/)
163REMAP2(u16, iowrite16, /**/)
164REMAP2(u32, iowrite32, /**/)
165REMAP2(u8, writeb, volatile)
166REMAP2(u16, writew, volatile)
167REMAP2(u32, writel, volatile)
168REMAP2(u64, writeq, volatile)
169
170#undef REMAP1
171#undef REMAP2
172
173extern inline void __iomem *generic_ioportmap(unsigned long a)
174{
175 return alpha_mv.mv_ioportmap(a);
176}
177
178static inline void __iomem *generic_ioremap(unsigned long a, unsigned long s)
179{
180 return alpha_mv.mv_ioremap(a, s);
181}
182
183static inline void generic_iounmap(volatile void __iomem *a)
184{
185 return alpha_mv.mv_iounmap(a);
186}
187
188static inline int generic_is_ioaddr(unsigned long a)
189{
190 return alpha_mv.mv_is_ioaddr(a);
191}
192
193static inline int generic_is_mmio(const volatile void __iomem *a)
194{
195 return alpha_mv.mv_is_mmio(a);
196}
197
198#define __IO_PREFIX generic
199#define generic_trivial_rw_bw 0
200#define generic_trivial_rw_lq 0
201#define generic_trivial_io_bw 0
202#define generic_trivial_io_lq 0
203#define generic_trivial_iounmap 0
204
205#else
206
207#if defined(CONFIG_ALPHA_APECS)
208# include <asm/core_apecs.h>
209#elif defined(CONFIG_ALPHA_CIA)
210# include <asm/core_cia.h>
211#elif defined(CONFIG_ALPHA_IRONGATE)
212# include <asm/core_irongate.h>
213#elif defined(CONFIG_ALPHA_JENSEN)
214# include <asm/jensen.h>
215#elif defined(CONFIG_ALPHA_LCA)
216# include <asm/core_lca.h>
217#elif defined(CONFIG_ALPHA_MARVEL)
218# include <asm/core_marvel.h>
219#elif defined(CONFIG_ALPHA_MCPCIA)
220# include <asm/core_mcpcia.h>
221#elif defined(CONFIG_ALPHA_POLARIS)
222# include <asm/core_polaris.h>
223#elif defined(CONFIG_ALPHA_T2)
224# include <asm/core_t2.h>
225#elif defined(CONFIG_ALPHA_TSUNAMI)
226# include <asm/core_tsunami.h>
227#elif defined(CONFIG_ALPHA_TITAN)
228# include <asm/core_titan.h>
229#elif defined(CONFIG_ALPHA_WILDFIRE)
230# include <asm/core_wildfire.h>
231#else
232#error "What system is this?"
233#endif
234
235#endif /* GENERIC */
236
237/*
238 * We always have external versions of these routines.
239 */
240extern u8 inb(unsigned long port);
241extern u16 inw(unsigned long port);
242extern u32 inl(unsigned long port);
243extern void outb(u8 b, unsigned long port);
244extern void outw(u16 b, unsigned long port);
245extern void outl(u32 b, unsigned long port);
246
247extern u8 readb(const volatile void __iomem *addr);
248extern u16 readw(const volatile void __iomem *addr);
249extern u32 readl(const volatile void __iomem *addr);
250extern u64 readq(const volatile void __iomem *addr);
251extern void writeb(u8 b, volatile void __iomem *addr);
252extern void writew(u16 b, volatile void __iomem *addr);
253extern void writel(u32 b, volatile void __iomem *addr);
254extern void writeq(u64 b, volatile void __iomem *addr);
255
256extern u8 __raw_readb(const volatile void __iomem *addr);
257extern u16 __raw_readw(const volatile void __iomem *addr);
258extern u32 __raw_readl(const volatile void __iomem *addr);
259extern u64 __raw_readq(const volatile void __iomem *addr);
260extern void __raw_writeb(u8 b, volatile void __iomem *addr);
261extern void __raw_writew(u16 b, volatile void __iomem *addr);
262extern void __raw_writel(u32 b, volatile void __iomem *addr);
263extern void __raw_writeq(u64 b, volatile void __iomem *addr);
264
265/*
266 * Mapping from port numbers to __iomem space is pretty easy.
267 */
268
269/* These two have to be extern inline because of the extern prototype from
270 <asm-generic/iomap.h>. It is not legal to mix "extern" and "static" for
271 the same declaration. */
272extern inline void __iomem *ioport_map(unsigned long port, unsigned int size)
273{
274 return IO_CONCAT(__IO_PREFIX,ioportmap) (port);
275}
276
277extern inline void ioport_unmap(void __iomem *addr)
278{
279}
280
281static inline void __iomem *ioremap(unsigned long port, unsigned long size)
282{
283 return IO_CONCAT(__IO_PREFIX,ioremap) (port, size);
284}
285
286static inline void __iomem *__ioremap(unsigned long port, unsigned long size,
287 unsigned long flags)
288{
289 return ioremap(port, size);
290}
291
292static inline void __iomem * ioremap_nocache(unsigned long offset,
293 unsigned long size)
294{
295 return ioremap(offset, size);
296}
297
298#define ioremap_wc ioremap_nocache
299#define ioremap_uc ioremap_nocache
300
301static inline void iounmap(volatile void __iomem *addr)
302{
303 IO_CONCAT(__IO_PREFIX,iounmap)(addr);
304}
305
306static inline int __is_ioaddr(unsigned long addr)
307{
308 return IO_CONCAT(__IO_PREFIX,is_ioaddr)(addr);
309}
310#define __is_ioaddr(a) __is_ioaddr((unsigned long)(a))
311
312static inline int __is_mmio(const volatile void __iomem *addr)
313{
314 return IO_CONCAT(__IO_PREFIX,is_mmio)(addr);
315}
316
317
318/*
319 * If the actual I/O bits are sufficiently trivial, then expand inline.
320 */
321
322#if IO_CONCAT(__IO_PREFIX,trivial_io_bw)
323extern inline unsigned int ioread8(void __iomem *addr)
324{
Olivier Deprez0e641232021-09-23 10:07:05 +0200325 unsigned int ret;
326 mb();
327 ret = IO_CONCAT(__IO_PREFIX,ioread8)(addr);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000328 mb();
329 return ret;
330}
331
332extern inline unsigned int ioread16(void __iomem *addr)
333{
Olivier Deprez0e641232021-09-23 10:07:05 +0200334 unsigned int ret;
335 mb();
336 ret = IO_CONCAT(__IO_PREFIX,ioread16)(addr);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000337 mb();
338 return ret;
339}
340
341extern inline void iowrite8(u8 b, void __iomem *addr)
342{
343 mb();
344 IO_CONCAT(__IO_PREFIX, iowrite8)(b, addr);
345}
346
347extern inline void iowrite16(u16 b, void __iomem *addr)
348{
349 mb();
350 IO_CONCAT(__IO_PREFIX, iowrite16)(b, addr);
351}
352
353extern inline u8 inb(unsigned long port)
354{
355 return ioread8(ioport_map(port, 1));
356}
357
358extern inline u16 inw(unsigned long port)
359{
360 return ioread16(ioport_map(port, 2));
361}
362
363extern inline void outb(u8 b, unsigned long port)
364{
365 iowrite8(b, ioport_map(port, 1));
366}
367
368extern inline void outw(u16 b, unsigned long port)
369{
370 iowrite16(b, ioport_map(port, 2));
371}
372#endif
373
374#if IO_CONCAT(__IO_PREFIX,trivial_io_lq)
375extern inline unsigned int ioread32(void __iomem *addr)
376{
Olivier Deprez0e641232021-09-23 10:07:05 +0200377 unsigned int ret;
378 mb();
379 ret = IO_CONCAT(__IO_PREFIX,ioread32)(addr);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000380 mb();
381 return ret;
382}
383
384extern inline void iowrite32(u32 b, void __iomem *addr)
385{
386 mb();
387 IO_CONCAT(__IO_PREFIX, iowrite32)(b, addr);
388}
389
390extern inline u32 inl(unsigned long port)
391{
392 return ioread32(ioport_map(port, 4));
393}
394
395extern inline void outl(u32 b, unsigned long port)
396{
397 iowrite32(b, ioport_map(port, 4));
398}
399#endif
400
401#if IO_CONCAT(__IO_PREFIX,trivial_rw_bw) == 1
402extern inline u8 __raw_readb(const volatile void __iomem *addr)
403{
404 return IO_CONCAT(__IO_PREFIX,readb)(addr);
405}
406
407extern inline u16 __raw_readw(const volatile void __iomem *addr)
408{
409 return IO_CONCAT(__IO_PREFIX,readw)(addr);
410}
411
412extern inline void __raw_writeb(u8 b, volatile void __iomem *addr)
413{
414 IO_CONCAT(__IO_PREFIX,writeb)(b, addr);
415}
416
417extern inline void __raw_writew(u16 b, volatile void __iomem *addr)
418{
419 IO_CONCAT(__IO_PREFIX,writew)(b, addr);
420}
421
422extern inline u8 readb(const volatile void __iomem *addr)
423{
Olivier Deprez0e641232021-09-23 10:07:05 +0200424 u8 ret;
425 mb();
426 ret = __raw_readb(addr);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000427 mb();
428 return ret;
429}
430
431extern inline u16 readw(const volatile void __iomem *addr)
432{
Olivier Deprez0e641232021-09-23 10:07:05 +0200433 u16 ret;
434 mb();
435 ret = __raw_readw(addr);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000436 mb();
437 return ret;
438}
439
440extern inline void writeb(u8 b, volatile void __iomem *addr)
441{
442 mb();
443 __raw_writeb(b, addr);
444}
445
446extern inline void writew(u16 b, volatile void __iomem *addr)
447{
448 mb();
449 __raw_writew(b, addr);
450}
451#endif
452
453#if IO_CONCAT(__IO_PREFIX,trivial_rw_lq) == 1
454extern inline u32 __raw_readl(const volatile void __iomem *addr)
455{
456 return IO_CONCAT(__IO_PREFIX,readl)(addr);
457}
458
459extern inline u64 __raw_readq(const volatile void __iomem *addr)
460{
461 return IO_CONCAT(__IO_PREFIX,readq)(addr);
462}
463
464extern inline void __raw_writel(u32 b, volatile void __iomem *addr)
465{
466 IO_CONCAT(__IO_PREFIX,writel)(b, addr);
467}
468
469extern inline void __raw_writeq(u64 b, volatile void __iomem *addr)
470{
471 IO_CONCAT(__IO_PREFIX,writeq)(b, addr);
472}
473
474extern inline u32 readl(const volatile void __iomem *addr)
475{
Olivier Deprez0e641232021-09-23 10:07:05 +0200476 u32 ret;
477 mb();
478 ret = __raw_readl(addr);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000479 mb();
480 return ret;
481}
482
483extern inline u64 readq(const volatile void __iomem *addr)
484{
Olivier Deprez0e641232021-09-23 10:07:05 +0200485 u64 ret;
486 mb();
487 ret = __raw_readq(addr);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000488 mb();
489 return ret;
490}
491
492extern inline void writel(u32 b, volatile void __iomem *addr)
493{
494 mb();
495 __raw_writel(b, addr);
496}
497
498extern inline void writeq(u64 b, volatile void __iomem *addr)
499{
500 mb();
501 __raw_writeq(b, addr);
502}
503#endif
504
Olivier Deprez0e641232021-09-23 10:07:05 +0200505#define ioread16be(p) swab16(ioread16(p))
506#define ioread32be(p) swab32(ioread32(p))
507#define iowrite16be(v,p) iowrite16(swab16(v), (p))
508#define iowrite32be(v,p) iowrite32(swab32(v), (p))
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000509
510#define inb_p inb
511#define inw_p inw
512#define inl_p inl
513#define outb_p outb
514#define outw_p outw
515#define outl_p outl
Olivier Deprez0e641232021-09-23 10:07:05 +0200516
517extern u8 readb_relaxed(const volatile void __iomem *addr);
518extern u16 readw_relaxed(const volatile void __iomem *addr);
519extern u32 readl_relaxed(const volatile void __iomem *addr);
520extern u64 readq_relaxed(const volatile void __iomem *addr);
521
522#if IO_CONCAT(__IO_PREFIX,trivial_io_bw)
523extern inline u8 readb_relaxed(const volatile void __iomem *addr)
524{
525 mb();
526 return __raw_readb(addr);
527}
528
529extern inline u16 readw_relaxed(const volatile void __iomem *addr)
530{
531 mb();
532 return __raw_readw(addr);
533}
534#endif
535
536#if IO_CONCAT(__IO_PREFIX,trivial_io_lq)
537extern inline u32 readl_relaxed(const volatile void __iomem *addr)
538{
539 mb();
540 return __raw_readl(addr);
541}
542
543extern inline u64 readq_relaxed(const volatile void __iomem *addr)
544{
545 mb();
546 return __raw_readq(addr);
547}
548#endif
549
550#define writeb_relaxed writeb
551#define writew_relaxed writew
552#define writel_relaxed writel
553#define writeq_relaxed writeq
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000554
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000555/*
556 * String version of IO memory access ops:
557 */
558extern void memcpy_fromio(void *, const volatile void __iomem *, long);
559extern void memcpy_toio(volatile void __iomem *, const void *, long);
560extern void _memset_c_io(volatile void __iomem *, unsigned long, long);
561
562static inline void memset_io(volatile void __iomem *addr, u8 c, long len)
563{
564 _memset_c_io(addr, 0x0101010101010101UL * c, len);
565}
566
567#define __HAVE_ARCH_MEMSETW_IO
568static inline void memsetw_io(volatile void __iomem *addr, u16 c, long len)
569{
570 _memset_c_io(addr, 0x0001000100010001UL * c, len);
571}
572
573/*
574 * String versions of in/out ops:
575 */
576extern void insb (unsigned long port, void *dst, unsigned long count);
577extern void insw (unsigned long port, void *dst, unsigned long count);
578extern void insl (unsigned long port, void *dst, unsigned long count);
579extern void outsb (unsigned long port, const void *src, unsigned long count);
580extern void outsw (unsigned long port, const void *src, unsigned long count);
581extern void outsl (unsigned long port, const void *src, unsigned long count);
582
583/*
584 * The Alpha Jensen hardware for some rather strange reason puts
585 * the RTC clock at 0x170 instead of 0x70. Probably due to some
586 * misguided idea about using 0x70 for NMI stuff.
587 *
588 * These defines will override the defaults when doing RTC queries
589 */
590
591#ifdef CONFIG_ALPHA_GENERIC
592# define RTC_PORT(x) ((x) + alpha_mv.rtc_port)
593#else
594# ifdef CONFIG_ALPHA_JENSEN
595# define RTC_PORT(x) (0x170+(x))
596# else
597# define RTC_PORT(x) (0x70 + (x))
598# endif
599#endif
600#define RTC_ALWAYS_BCD 0
601
602/*
603 * Some mucking forons use if[n]def writeq to check if platform has it.
604 * It's a bloody bad idea and we probably want ARCH_HAS_WRITEQ for them
605 * to play with; for now just use cpp anti-recursion logics and make sure
606 * that damn thing is defined and expands to itself.
607 */
608
609#define writeq writeq
610#define readq readq
611
612/*
613 * Convert a physical pointer to a virtual kernel pointer for /dev/mem
614 * access
615 */
616#define xlate_dev_mem_ptr(p) __va(p)
617
618/*
619 * Convert a virtual cached pointer to an uncached pointer
620 */
621#define xlate_dev_kmem_ptr(p) p
622
623#endif /* __KERNEL__ */
624
625#endif /* __ALPHA_IO_H */