1c6557e7fSMartin Schwidefsky /* 2c6557e7fSMartin Schwidefsky * S390 version 3a53c8fabSHeiko Carstens * Copyright IBM Corp. 1999 4c6557e7fSMartin Schwidefsky * Author(s): Hartmut Penner (hp@de.ibm.com), 5c6557e7fSMartin Schwidefsky * Martin Schwidefsky (schwidefsky@de.ibm.com) 6c6557e7fSMartin Schwidefsky * 7c6557e7fSMartin Schwidefsky * Derived from "include/asm-i386/processor.h" 8c6557e7fSMartin Schwidefsky * Copyright (C) 1994, Linus Torvalds 9c6557e7fSMartin Schwidefsky */ 10c6557e7fSMartin Schwidefsky 11c6557e7fSMartin Schwidefsky #ifndef __ASM_S390_PROCESSOR_H 12c6557e7fSMartin Schwidefsky #define __ASM_S390_PROCESSOR_H 13c6557e7fSMartin Schwidefsky 1492778b99SHeiko Carstens #include <linux/const.h> 1592778b99SHeiko Carstens 16d3a73acbSMartin Schwidefsky #define CIF_MCCK_PENDING 0 /* machine check handling is pending */ 17d3a73acbSMartin Schwidefsky #define CIF_ASCE 1 /* user asce needs fixup / uaccess */ 18fe0f4976SMartin Schwidefsky #define CIF_NOHZ_DELAY 2 /* delay HZ disable for a tick */ 19b0753902SHendrik Brueckner #define CIF_FPU 3 /* restore FPU registers */ 20db7e007fSHeiko Carstens #define CIF_IGNORE_IRQ 4 /* ignore interrupt (for udelay) */ 21419123f9SMartin Schwidefsky #define CIF_ENABLED_WAIT 5 /* in enabled wait state */ 22d3a73acbSMartin Schwidefsky 2392778b99SHeiko Carstens #define _CIF_MCCK_PENDING _BITUL(CIF_MCCK_PENDING) 2492778b99SHeiko Carstens #define _CIF_ASCE _BITUL(CIF_ASCE) 2592778b99SHeiko Carstens #define _CIF_NOHZ_DELAY _BITUL(CIF_NOHZ_DELAY) 2692778b99SHeiko Carstens #define _CIF_FPU _BITUL(CIF_FPU) 2792778b99SHeiko Carstens #define _CIF_IGNORE_IRQ _BITUL(CIF_IGNORE_IRQ) 28419123f9SMartin Schwidefsky #define _CIF_ENABLED_WAIT _BITUL(CIF_ENABLED_WAIT) 29d3a73acbSMartin Schwidefsky 30eb608fb3SHeiko Carstens #ifndef __ASSEMBLY__ 31eb608fb3SHeiko Carstens 32edd53787SHeiko Carstens #include <linux/linkage.h> 33a0616cdeSDavid Howells #include <linux/irqflags.h> 34e86a6ed6SHeiko Carstens #include <asm/cpu.h> 3525097bf1SChristian Ehrhardt #include <asm/page.h> 36c6557e7fSMartin Schwidefsky #include <asm/ptrace.h> 3725097bf1SChristian Ehrhardt #include <asm/setup.h> 38e4b8b3f3SJan Glauber #include <asm/runtime_instr.h> 39b0753902SHendrik Brueckner #include <asm/fpu/types.h> 40b0753902SHendrik Brueckner #include <asm/fpu/internal.h> 41c6557e7fSMartin Schwidefsky 42d3a73acbSMartin Schwidefsky static inline void set_cpu_flag(int flag) 43d3a73acbSMartin Schwidefsky { 44ac25e790SHeiko Carstens S390_lowcore.cpu_flags |= (1UL << flag); 45d3a73acbSMartin Schwidefsky } 46d3a73acbSMartin Schwidefsky 47d3a73acbSMartin Schwidefsky static inline void clear_cpu_flag(int flag) 48d3a73acbSMartin Schwidefsky { 49ac25e790SHeiko Carstens S390_lowcore.cpu_flags &= ~(1UL << flag); 50d3a73acbSMartin Schwidefsky } 51d3a73acbSMartin Schwidefsky 52d3a73acbSMartin Schwidefsky static inline int test_cpu_flag(int flag) 53d3a73acbSMartin Schwidefsky { 54ac25e790SHeiko Carstens return !!(S390_lowcore.cpu_flags & (1UL << flag)); 55d3a73acbSMartin Schwidefsky } 56d3a73acbSMartin Schwidefsky 57419123f9SMartin Schwidefsky /* 58419123f9SMartin Schwidefsky * Test CIF flag of another CPU. The caller needs to ensure that 59419123f9SMartin Schwidefsky * CPU hotplug can not happen, e.g. by disabling preemption. 60419123f9SMartin Schwidefsky */ 61419123f9SMartin Schwidefsky static inline int test_cpu_flag_of(int flag, int cpu) 62419123f9SMartin Schwidefsky { 63c667aeacSHeiko Carstens struct lowcore *lc = lowcore_ptr[cpu]; 64419123f9SMartin Schwidefsky return !!(lc->cpu_flags & (1UL << flag)); 65419123f9SMartin Schwidefsky } 66419123f9SMartin Schwidefsky 67fe0f4976SMartin Schwidefsky #define arch_needs_cpu() test_cpu_flag(CIF_NOHZ_DELAY) 68fe0f4976SMartin Schwidefsky 69c6557e7fSMartin Schwidefsky /* 70c6557e7fSMartin Schwidefsky * Default implementation of macro that returns current 71c6557e7fSMartin Schwidefsky * instruction pointer ("program counter"). 72c6557e7fSMartin Schwidefsky */ 73c6557e7fSMartin Schwidefsky #define current_text_addr() ({ void *pc; asm("basr %0,0" : "=a" (pc)); pc; }) 74c6557e7fSMartin Schwidefsky 75e86a6ed6SHeiko Carstens static inline void get_cpu_id(struct cpuid *ptr) 76c6557e7fSMartin Schwidefsky { 77987bcdacSMartin Schwidefsky asm volatile("stidp %0" : "=Q" (*ptr)); 78c6557e7fSMartin Schwidefsky } 79c6557e7fSMartin Schwidefsky 80097a116cSHeiko Carstens void s390_adjust_jiffies(void); 81097a116cSHeiko Carstens void s390_update_cpu_mhz(void); 82097a116cSHeiko Carstens void cpu_detect_mhz_feature(void); 83097a116cSHeiko Carstens 84638ad34aSMartin Schwidefsky extern const struct seq_operations cpuinfo_op; 85638ad34aSMartin Schwidefsky extern int sysctl_ieee_emulation_warnings; 8665f22a90SAl Viro extern void execve_tail(void); 87c6557e7fSMartin Schwidefsky 88c6557e7fSMartin Schwidefsky /* 89f481bfafSMartin Schwidefsky * User space process size: 2GB for 31 bit, 4TB or 8PT for 64 bit. 90c6557e7fSMartin Schwidefsky */ 91c6557e7fSMartin Schwidefsky 92f481bfafSMartin Schwidefsky #define TASK_SIZE_OF(tsk) ((tsk)->mm->context.asce_limit) 93c6557e7fSMartin Schwidefsky #define TASK_UNMAPPED_BASE (test_thread_flag(TIF_31BIT) ? \ 94c6557e7fSMartin Schwidefsky (1UL << 30) : (1UL << 41)) 95c6557e7fSMartin Schwidefsky #define TASK_SIZE TASK_SIZE_OF(current) 96ee6ee55bSMartin Schwidefsky #define TASK_MAX_SIZE (1UL << 53) 97c6557e7fSMartin Schwidefsky 98c6557e7fSMartin Schwidefsky #define STACK_TOP (1UL << (test_thread_flag(TIF_31BIT) ? 31:42)) 99c6557e7fSMartin Schwidefsky #define STACK_TOP_MAX (1UL << 42) 100c6557e7fSMartin Schwidefsky 101c6557e7fSMartin Schwidefsky #define HAVE_ARCH_PICK_MMAP_LAYOUT 102c6557e7fSMartin Schwidefsky 103c6557e7fSMartin Schwidefsky typedef struct { 104c6557e7fSMartin Schwidefsky __u32 ar4; 105c6557e7fSMartin Schwidefsky } mm_segment_t; 106c6557e7fSMartin Schwidefsky 107c6557e7fSMartin Schwidefsky /* 108c6557e7fSMartin Schwidefsky * Thread structure 109c6557e7fSMartin Schwidefsky */ 110c6557e7fSMartin Schwidefsky struct thread_struct { 111c6557e7fSMartin Schwidefsky unsigned int acrs[NUM_ACRS]; 112c6557e7fSMartin Schwidefsky unsigned long ksp; /* kernel stack pointer */ 113c6557e7fSMartin Schwidefsky mm_segment_t mm_segment; 114e5992f2eSMartin Schwidefsky unsigned long gmap_addr; /* address of last gmap fault. */ 1154be130a0SMartin Schwidefsky unsigned int gmap_write_flag; /* gmap fault write indication */ 1164a494439SDavid Hildenbrand unsigned int gmap_int_code; /* int code of last gmap fault */ 11724eb3a82SDominik Dingel unsigned int gmap_pfault; /* signal of a pending guest pfault */ 1185e9a2692SMartin Schwidefsky struct per_regs per_user; /* User specified PER registers */ 1195e9a2692SMartin Schwidefsky struct per_event per_event; /* Cause of the last PER trap */ 120d35339a4SMartin Schwidefsky unsigned long per_flags; /* Flags to control debug behavior */ 121c6557e7fSMartin Schwidefsky /* pfault_wait is used to block the process on a pfault event */ 122c6557e7fSMartin Schwidefsky unsigned long pfault_wait; 123f2db2e6cSHeiko Carstens struct list_head list; 124e4b8b3f3SJan Glauber /* cpu runtime instrumentation */ 125e4b8b3f3SJan Glauber struct runtime_instr_cb *ri_cb; 126d35339a4SMartin Schwidefsky unsigned char trap_tdb[256]; /* Transaction abort diagnose block */ 1273f6813b9SMartin Schwidefsky /* 1283f6813b9SMartin Schwidefsky * Warning: 'fpu' is dynamically-sized. It *MUST* be at 1293f6813b9SMartin Schwidefsky * the end. 1303f6813b9SMartin Schwidefsky */ 1313f6813b9SMartin Schwidefsky struct fpu fpu; /* FP and VX register save area */ 132c6557e7fSMartin Schwidefsky }; 133c6557e7fSMartin Schwidefsky 13464597f9dSMichael Mueller /* Flag to disable transactions. */ 13564597f9dSMichael Mueller #define PER_FLAG_NO_TE 1UL 13664597f9dSMichael Mueller /* Flag to enable random transaction aborts. */ 13764597f9dSMichael Mueller #define PER_FLAG_TE_ABORT_RAND 2UL 13864597f9dSMichael Mueller /* Flag to specify random transaction abort mode: 13964597f9dSMichael Mueller * - abort each transaction at a random instruction before TEND if set. 14064597f9dSMichael Mueller * - abort random transactions at a random instruction if cleared. 14164597f9dSMichael Mueller */ 14264597f9dSMichael Mueller #define PER_FLAG_TE_ABORT_RAND_TEND 4UL 143d35339a4SMartin Schwidefsky 144c6557e7fSMartin Schwidefsky typedef struct thread_struct thread_struct; 145c6557e7fSMartin Schwidefsky 146c6557e7fSMartin Schwidefsky /* 147c6557e7fSMartin Schwidefsky * Stack layout of a C stack frame. 148c6557e7fSMartin Schwidefsky */ 149c6557e7fSMartin Schwidefsky #ifndef __PACK_STACK 150c6557e7fSMartin Schwidefsky struct stack_frame { 151c6557e7fSMartin Schwidefsky unsigned long back_chain; 152c6557e7fSMartin Schwidefsky unsigned long empty1[5]; 153c6557e7fSMartin Schwidefsky unsigned long gprs[10]; 154c6557e7fSMartin Schwidefsky unsigned int empty2[8]; 155c6557e7fSMartin Schwidefsky }; 156c6557e7fSMartin Schwidefsky #else 157c6557e7fSMartin Schwidefsky struct stack_frame { 158c6557e7fSMartin Schwidefsky unsigned long empty1[5]; 159c6557e7fSMartin Schwidefsky unsigned int empty2[8]; 160c6557e7fSMartin Schwidefsky unsigned long gprs[10]; 161c6557e7fSMartin Schwidefsky unsigned long back_chain; 162c6557e7fSMartin Schwidefsky }; 163c6557e7fSMartin Schwidefsky #endif 164c6557e7fSMartin Schwidefsky 165c6557e7fSMartin Schwidefsky #define ARCH_MIN_TASKALIGN 8 166c6557e7fSMartin Schwidefsky 167c6557e7fSMartin Schwidefsky #define INIT_THREAD { \ 168c6557e7fSMartin Schwidefsky .ksp = sizeof(init_stack) + (unsigned long) &init_stack, \ 1693f6813b9SMartin Schwidefsky .fpu.regs = (void *) init_task.thread.fpu.fprs, \ 170c6557e7fSMartin Schwidefsky } 171c6557e7fSMartin Schwidefsky 172c6557e7fSMartin Schwidefsky /* 173c6557e7fSMartin Schwidefsky * Do necessary setup to start up a new thread. 174c6557e7fSMartin Schwidefsky */ 175c6557e7fSMartin Schwidefsky #define start_thread(regs, new_psw, new_stackp) do { \ 176e258d719SMartin Schwidefsky regs->psw.mask = PSW_USER_BITS | PSW_MASK_EA | PSW_MASK_BA; \ 177fecc868aSHeiko Carstens regs->psw.addr = new_psw; \ 178c6557e7fSMartin Schwidefsky regs->gprs[15] = new_stackp; \ 17965f22a90SAl Viro execve_tail(); \ 180c6557e7fSMartin Schwidefsky } while (0) 181c6557e7fSMartin Schwidefsky 182c6557e7fSMartin Schwidefsky #define start_thread31(regs, new_psw, new_stackp) do { \ 183e258d719SMartin Schwidefsky regs->psw.mask = PSW_USER_BITS | PSW_MASK_BA; \ 184fecc868aSHeiko Carstens regs->psw.addr = new_psw; \ 185c6557e7fSMartin Schwidefsky regs->gprs[15] = new_stackp; \ 186723cacbdSGerald Schaefer crst_table_downgrade(current->mm); \ 18765f22a90SAl Viro execve_tail(); \ 188c6557e7fSMartin Schwidefsky } while (0) 189c6557e7fSMartin Schwidefsky 190c6557e7fSMartin Schwidefsky /* Forward declaration, a strange C thing */ 191c6557e7fSMartin Schwidefsky struct task_struct; 192c6557e7fSMartin Schwidefsky struct mm_struct; 193c6557e7fSMartin Schwidefsky struct seq_file; 194c6557e7fSMartin Schwidefsky 195d0208639SHeiko Carstens typedef int (*dump_trace_func_t)(void *data, unsigned long address, int reliable); 196758d39ebSHeiko Carstens void dump_trace(dump_trace_func_t func, void *data, 197758d39ebSHeiko Carstens struct task_struct *task, unsigned long sp); 198758d39ebSHeiko Carstens 1995a79859aSHeiko Carstens void show_cacheinfo(struct seq_file *m); 2006668022cSHeiko Carstens 201c6557e7fSMartin Schwidefsky /* Free all resources held by a thread. */ 202c6557e7fSMartin Schwidefsky extern void release_thread(struct task_struct *); 203c6557e7fSMartin Schwidefsky 204c6557e7fSMartin Schwidefsky /* 205c6557e7fSMartin Schwidefsky * Return saved PC of a blocked thread. 206c6557e7fSMartin Schwidefsky */ 207c6557e7fSMartin Schwidefsky extern unsigned long thread_saved_pc(struct task_struct *t); 208c6557e7fSMartin Schwidefsky 209c6557e7fSMartin Schwidefsky unsigned long get_wchan(struct task_struct *p); 210c6557e7fSMartin Schwidefsky #define task_pt_regs(tsk) ((struct pt_regs *) \ 211c6557e7fSMartin Schwidefsky (task_stack_page(tsk) + THREAD_SIZE) - 1) 212c6557e7fSMartin Schwidefsky #define KSTK_EIP(tsk) (task_pt_regs(tsk)->psw.addr) 213c6557e7fSMartin Schwidefsky #define KSTK_ESP(tsk) (task_pt_regs(tsk)->gprs[15]) 214c6557e7fSMartin Schwidefsky 2155ebf250dSHeiko Carstens /* Has task runtime instrumentation enabled ? */ 2165ebf250dSHeiko Carstens #define is_ri_task(tsk) (!!(tsk)->thread.ri_cb) 2175ebf250dSHeiko Carstens 21876737ce1SHeiko Carstens static inline unsigned long current_stack_pointer(void) 21976737ce1SHeiko Carstens { 22076737ce1SHeiko Carstens unsigned long sp; 22176737ce1SHeiko Carstens 22276737ce1SHeiko Carstens asm volatile("la %0,0(15)" : "=a" (sp)); 22376737ce1SHeiko Carstens return sp; 22476737ce1SHeiko Carstens } 22576737ce1SHeiko Carstens 226a0616cdeSDavid Howells static inline unsigned short stap(void) 227a0616cdeSDavid Howells { 228a0616cdeSDavid Howells unsigned short cpu_address; 229a0616cdeSDavid Howells 230a0616cdeSDavid Howells asm volatile("stap %0" : "=m" (cpu_address)); 231a0616cdeSDavid Howells return cpu_address; 232a0616cdeSDavid Howells } 233a0616cdeSDavid Howells 234c6557e7fSMartin Schwidefsky /* 235c6557e7fSMartin Schwidefsky * Give up the time slice of the virtual PU. 236c6557e7fSMartin Schwidefsky */ 23779ab11cdSChristian Borntraeger void cpu_relax_yield(void); 238c6557e7fSMartin Schwidefsky 23922b6430dSChristian Borntraeger #define cpu_relax() barrier() 2403a6bfbc9SDavidlohr Bueso #define cpu_relax_lowlatency() barrier() 241083986e8SHeiko Carstens 242097a116cSHeiko Carstens #define ECAG_CACHE_ATTRIBUTE 0 243097a116cSHeiko Carstens #define ECAG_CPU_ATTRIBUTE 1 244097a116cSHeiko Carstens 245097a116cSHeiko Carstens static inline unsigned long __ecag(unsigned int asi, unsigned char parm) 246097a116cSHeiko Carstens { 247097a116cSHeiko Carstens unsigned long val; 248097a116cSHeiko Carstens 249097a116cSHeiko Carstens asm volatile(".insn rsy,0xeb000000004c,%0,0,0(%1)" /* ecag */ 250097a116cSHeiko Carstens : "=d" (val) : "a" (asi << 8 | parm)); 251097a116cSHeiko Carstens return val; 252097a116cSHeiko Carstens } 253097a116cSHeiko Carstens 254c6557e7fSMartin Schwidefsky static inline void psw_set_key(unsigned int key) 255c6557e7fSMartin Schwidefsky { 256c6557e7fSMartin Schwidefsky asm volatile("spka 0(%0)" : : "d" (key)); 257c6557e7fSMartin Schwidefsky } 258c6557e7fSMartin Schwidefsky 259c6557e7fSMartin Schwidefsky /* 260c6557e7fSMartin Schwidefsky * Set PSW to specified value. 261c6557e7fSMartin Schwidefsky */ 262c6557e7fSMartin Schwidefsky static inline void __load_psw(psw_t psw) 263c6557e7fSMartin Schwidefsky { 264987bcdacSMartin Schwidefsky asm volatile("lpswe %0" : : "Q" (psw) : "cc"); 265c6557e7fSMartin Schwidefsky } 266c6557e7fSMartin Schwidefsky 267c6557e7fSMartin Schwidefsky /* 268c6557e7fSMartin Schwidefsky * Set PSW mask to specified value, while leaving the 269c6557e7fSMartin Schwidefsky * PSW addr pointing to the next instruction. 270c6557e7fSMartin Schwidefsky */ 271c6557e7fSMartin Schwidefsky static inline void __load_psw_mask(unsigned long mask) 272c6557e7fSMartin Schwidefsky { 273c6557e7fSMartin Schwidefsky unsigned long addr; 274c6557e7fSMartin Schwidefsky psw_t psw; 275c6557e7fSMartin Schwidefsky 276c6557e7fSMartin Schwidefsky psw.mask = mask; 277c6557e7fSMartin Schwidefsky 278c6557e7fSMartin Schwidefsky asm volatile( 279c6557e7fSMartin Schwidefsky " larl %0,1f\n" 280987bcdacSMartin Schwidefsky " stg %0,%O1+8(%R1)\n" 281987bcdacSMartin Schwidefsky " lpswe %1\n" 282c6557e7fSMartin Schwidefsky "1:" 283987bcdacSMartin Schwidefsky : "=&d" (addr), "=Q" (psw) : "Q" (psw) : "memory", "cc"); 284c6557e7fSMartin Schwidefsky } 285c6557e7fSMartin Schwidefsky 286c6557e7fSMartin Schwidefsky /* 28722362a0eSMartin Schwidefsky * Extract current PSW mask 28822362a0eSMartin Schwidefsky */ 28922362a0eSMartin Schwidefsky static inline unsigned long __extract_psw(void) 29022362a0eSMartin Schwidefsky { 29122362a0eSMartin Schwidefsky unsigned int reg1, reg2; 29222362a0eSMartin Schwidefsky 29322362a0eSMartin Schwidefsky asm volatile("epsw %0,%1" : "=d" (reg1), "=a" (reg2)); 29422362a0eSMartin Schwidefsky return (((unsigned long) reg1) << 32) | ((unsigned long) reg2); 29522362a0eSMartin Schwidefsky } 29622362a0eSMartin Schwidefsky 297ecbafda8SHeiko Carstens static inline void local_mcck_enable(void) 298ecbafda8SHeiko Carstens { 299ecbafda8SHeiko Carstens __load_psw_mask(__extract_psw() | PSW_MASK_MCHECK); 300ecbafda8SHeiko Carstens } 301ecbafda8SHeiko Carstens 302ecbafda8SHeiko Carstens static inline void local_mcck_disable(void) 303ecbafda8SHeiko Carstens { 304ecbafda8SHeiko Carstens __load_psw_mask(__extract_psw() & ~PSW_MASK_MCHECK); 305ecbafda8SHeiko Carstens } 306ecbafda8SHeiko Carstens 30722362a0eSMartin Schwidefsky /* 308ccf45cafSMartin Schwidefsky * Rewind PSW instruction address by specified number of bytes. 309ccf45cafSMartin Schwidefsky */ 310ccf45cafSMartin Schwidefsky static inline unsigned long __rewind_psw(psw_t psw, unsigned long ilc) 311ccf45cafSMartin Schwidefsky { 312ccf45cafSMartin Schwidefsky unsigned long mask; 313ccf45cafSMartin Schwidefsky 314ccf45cafSMartin Schwidefsky mask = (psw.mask & PSW_MASK_EA) ? -1UL : 315ccf45cafSMartin Schwidefsky (psw.mask & PSW_MASK_BA) ? (1UL << 31) - 1 : 316ccf45cafSMartin Schwidefsky (1UL << 24) - 1; 317ccf45cafSMartin Schwidefsky return (psw.addr - ilc) & mask; 318ccf45cafSMartin Schwidefsky } 319ccf45cafSMartin Schwidefsky 320b5f87f15SMartin Schwidefsky /* 321b5f87f15SMartin Schwidefsky * Function to stop a processor until the next interrupt occurs 322b5f87f15SMartin Schwidefsky */ 323b5f87f15SMartin Schwidefsky void enabled_wait(void); 324b5f87f15SMartin Schwidefsky 325ccf45cafSMartin Schwidefsky /* 326c6557e7fSMartin Schwidefsky * Function to drop a processor into disabled wait state 327c6557e7fSMartin Schwidefsky */ 328ff2d8b19SJoe Perches static inline void __noreturn disabled_wait(unsigned long code) 329c6557e7fSMartin Schwidefsky { 330f9e6edfbSHeiko Carstens psw_t psw; 331c6557e7fSMartin Schwidefsky 332f9e6edfbSHeiko Carstens psw.mask = PSW_MASK_BASE | PSW_MASK_WAIT | PSW_MASK_BA | PSW_MASK_EA; 333f9e6edfbSHeiko Carstens psw.addr = code; 334f9e6edfbSHeiko Carstens __load_psw(psw); 335edd53787SHeiko Carstens while (1); 336c6557e7fSMartin Schwidefsky } 337c6557e7fSMartin Schwidefsky 338c6557e7fSMartin Schwidefsky /* 339c6557e7fSMartin Schwidefsky * Basic Machine Check/Program Check Handler. 340c6557e7fSMartin Schwidefsky */ 341c6557e7fSMartin Schwidefsky 342c6557e7fSMartin Schwidefsky extern void s390_base_mcck_handler(void); 343c6557e7fSMartin Schwidefsky extern void s390_base_pgm_handler(void); 344c6557e7fSMartin Schwidefsky extern void s390_base_ext_handler(void); 345c6557e7fSMartin Schwidefsky 346c6557e7fSMartin Schwidefsky extern void (*s390_base_mcck_handler_fn)(void); 347c6557e7fSMartin Schwidefsky extern void (*s390_base_pgm_handler_fn)(void); 348c6557e7fSMartin Schwidefsky extern void (*s390_base_ext_handler_fn)(void); 349c6557e7fSMartin Schwidefsky 350c6557e7fSMartin Schwidefsky #define ARCH_LOW_ADDRESS_LIMIT 0x7fffffffUL 351c6557e7fSMartin Schwidefsky 352fbe76568SHeiko Carstens extern int memcpy_real(void *, void *, size_t); 353fbe76568SHeiko Carstens extern void memcpy_absolute(void *, void *, size_t); 354fbe76568SHeiko Carstens 355fbe76568SHeiko Carstens #define mem_assign_absolute(dest, val) { \ 356fbe76568SHeiko Carstens __typeof__(dest) __tmp = (val); \ 357fbe76568SHeiko Carstens \ 358fbe76568SHeiko Carstens BUILD_BUG_ON(sizeof(__tmp) != sizeof(val)); \ 359fbe76568SHeiko Carstens memcpy_absolute(&(dest), &__tmp, sizeof(__tmp)); \ 360fbe76568SHeiko Carstens } 361fbe76568SHeiko Carstens 362eb608fb3SHeiko Carstens #endif /* __ASSEMBLY__ */ 363eb608fb3SHeiko Carstens 364c6557e7fSMartin Schwidefsky #endif /* __ASM_S390_PROCESSOR_H */ 365