1 /* SPDX-License-Identifier: BSD-3-Clause OR GPL-2.0 */
2 /******************************************************************************
3  *
4  * Name: aclinuxex.h - Extra OS specific defines, etc. for Linux
5  *
6  * Copyright (C) 2000 - 2023, Intel Corp.
7  *
8  *****************************************************************************/
9 
10 #ifndef __ACLINUXEX_H__
11 #define __ACLINUXEX_H__
12 
13 #ifdef __KERNEL__
14 
15 #ifndef ACPI_USE_NATIVE_DIVIDE
16 
17 #ifndef ACPI_DIV_64_BY_32
18 #define ACPI_DIV_64_BY_32(n_hi, n_lo, d32, q32, r32) \
19 	do { \
20 		u64 (__n) = ((u64) n_hi) << 32 | (n_lo); \
21 		(r32) = do_div ((__n), (d32)); \
22 		(q32) = (u32) (__n); \
23 	} while (0)
24 #endif
25 
26 #ifndef ACPI_SHIFT_RIGHT_64
27 #define ACPI_SHIFT_RIGHT_64(n_hi, n_lo) \
28 	do { \
29 		(n_lo) >>= 1; \
30 		(n_lo) |= (((n_hi) & 1) << 31); \
31 		(n_hi) >>= 1; \
32 	} while (0)
33 #endif
34 
35 #endif
36 
37 /*
38  * Overrides for in-kernel ACPICA
39  */
40 acpi_status ACPI_INIT_FUNCTION acpi_os_initialize(void);
41 
42 acpi_status acpi_os_terminate(void);
43 
44 /*
45  * The irqs_disabled() check is for resume from RAM.
46  * Interrupts are off during resume, just like they are for boot.
47  * However, boot has  (system_state != SYSTEM_RUNNING)
48  * to quiet __might_sleep() in kmalloc() and resume does not.
49  */
50 static inline void *acpi_os_allocate(acpi_size size)
51 {
52 	return kmalloc(size, irqs_disabled()? GFP_ATOMIC : GFP_KERNEL);
53 }
54 
55 static inline void *acpi_os_allocate_zeroed(acpi_size size)
56 {
57 	return kzalloc(size, irqs_disabled()? GFP_ATOMIC : GFP_KERNEL);
58 }
59 
60 static inline void acpi_os_free(void *memory)
61 {
62 	kfree(memory);
63 }
64 
65 static inline void *acpi_os_acquire_object(acpi_cache_t * cache)
66 {
67 	return kmem_cache_zalloc(cache,
68 				 irqs_disabled()? GFP_ATOMIC : GFP_KERNEL);
69 }
70 
71 static inline acpi_thread_id acpi_os_get_thread_id(void)
72 {
73 	return (acpi_thread_id) (unsigned long)current;
74 }
75 
76 /*
77  * When lockdep is enabled, the spin_lock_init() macro stringifies it's
78  * argument and uses that as a name for the lock in debugging.
79  * By executing spin_lock_init() in a macro the key changes from "lock" for
80  * all locks to the name of the argument of acpi_os_create_lock(), which
81  * prevents lockdep from reporting false positives for ACPICA locks.
82  */
83 #define acpi_os_create_lock(__handle) \
84 	({ \
85 		spinlock_t *lock = ACPI_ALLOCATE(sizeof(*lock)); \
86 		if (lock) { \
87 			*(__handle) = lock; \
88 			spin_lock_init(*(__handle)); \
89 		} \
90 		lock ? AE_OK : AE_NO_MEMORY; \
91 	})
92 
93 
94 #define acpi_os_create_raw_lock(__handle) \
95 	({ \
96 		raw_spinlock_t *lock = ACPI_ALLOCATE(sizeof(*lock)); \
97 		if (lock) { \
98 			*(__handle) = lock; \
99 			raw_spin_lock_init(*(__handle)); \
100 		} \
101 		lock ? AE_OK : AE_NO_MEMORY; \
102 	})
103 
104 static inline acpi_cpu_flags acpi_os_acquire_raw_lock(acpi_raw_spinlock lockp)
105 {
106 	acpi_cpu_flags flags;
107 
108 	raw_spin_lock_irqsave(lockp, flags);
109 	return flags;
110 }
111 
112 static inline void acpi_os_release_raw_lock(acpi_raw_spinlock lockp,
113 					    acpi_cpu_flags flags)
114 {
115 	raw_spin_unlock_irqrestore(lockp, flags);
116 }
117 
118 static inline void acpi_os_delete_raw_lock(acpi_raw_spinlock handle)
119 {
120 	ACPI_FREE(handle);
121 }
122 
123 static inline u8 acpi_os_readable(void *pointer, acpi_size length)
124 {
125 	return TRUE;
126 }
127 
128 static inline acpi_status acpi_os_initialize_debugger(void)
129 {
130 	return AE_OK;
131 }
132 
133 static inline void acpi_os_terminate_debugger(void)
134 {
135 	return;
136 }
137 
138 /*
139  * OSL interfaces added by Linux
140  */
141 
142 #endif				/* __KERNEL__ */
143 
144 #endif				/* __ACLINUXEX_H__ */
145