xref: /openbmc/linux/drivers/acpi/sleep.c (revision c34a8052)
1 // SPDX-License-Identifier: GPL-2.0-only
2 /*
3  * sleep.c - ACPI sleep support.
4  *
5  * Copyright (c) 2005 Alexey Starikovskiy <alexey.y.starikovskiy@intel.com>
6  * Copyright (c) 2004 David Shaohua Li <shaohua.li@intel.com>
7  * Copyright (c) 2000-2003 Patrick Mochel
8  * Copyright (c) 2003 Open Source Development Lab
9  */
10 
11 #define pr_fmt(fmt) "ACPI: PM: " fmt
12 
13 #include <linux/delay.h>
14 #include <linux/irq.h>
15 #include <linux/dmi.h>
16 #include <linux/device.h>
17 #include <linux/interrupt.h>
18 #include <linux/suspend.h>
19 #include <linux/reboot.h>
20 #include <linux/acpi.h>
21 #include <linux/module.h>
22 #include <linux/syscore_ops.h>
23 #include <asm/io.h>
24 #include <trace/events/power.h>
25 
26 #include "internal.h"
27 #include "sleep.h"
28 
29 /*
30  * Some HW-full platforms do not have _S5, so they may need
31  * to leverage efi power off for a shutdown.
32  */
33 bool acpi_no_s5;
34 static u8 sleep_states[ACPI_S_STATE_COUNT];
35 
36 static void acpi_sleep_tts_switch(u32 acpi_state)
37 {
38 	acpi_status status;
39 
40 	status = acpi_execute_simple_method(NULL, "\\_TTS", acpi_state);
41 	if (ACPI_FAILURE(status) && status != AE_NOT_FOUND) {
42 		/*
43 		 * OS can't evaluate the _TTS object correctly. Some warning
44 		 * message will be printed. But it won't break anything.
45 		 */
46 		pr_notice("Failure in evaluating _TTS object\n");
47 	}
48 }
49 
50 static int tts_notify_reboot(struct notifier_block *this,
51 			unsigned long code, void *x)
52 {
53 	acpi_sleep_tts_switch(ACPI_STATE_S5);
54 	return NOTIFY_DONE;
55 }
56 
57 static struct notifier_block tts_notifier = {
58 	.notifier_call	= tts_notify_reboot,
59 	.next		= NULL,
60 	.priority	= 0,
61 };
62 
63 #ifndef acpi_skip_set_wakeup_address
64 #define acpi_skip_set_wakeup_address() false
65 #endif
66 
67 static int acpi_sleep_prepare(u32 acpi_state)
68 {
69 #ifdef CONFIG_ACPI_SLEEP
70 	unsigned long acpi_wakeup_address;
71 
72 	/* do we have a wakeup address for S2 and S3? */
73 	if (acpi_state == ACPI_STATE_S3 && !acpi_skip_set_wakeup_address()) {
74 		acpi_wakeup_address = acpi_get_wakeup_address();
75 		if (!acpi_wakeup_address)
76 			return -EFAULT;
77 		acpi_set_waking_vector(acpi_wakeup_address);
78 
79 	}
80 #endif
81 	pr_info("Preparing to enter system sleep state S%d\n", acpi_state);
82 	acpi_enable_wakeup_devices(acpi_state);
83 	acpi_enter_sleep_state_prep(acpi_state);
84 	return 0;
85 }
86 
87 bool acpi_sleep_state_supported(u8 sleep_state)
88 {
89 	acpi_status status;
90 	u8 type_a, type_b;
91 
92 	status = acpi_get_sleep_type_data(sleep_state, &type_a, &type_b);
93 	return ACPI_SUCCESS(status) && (!acpi_gbl_reduced_hardware
94 		|| (acpi_gbl_FADT.sleep_control.address
95 			&& acpi_gbl_FADT.sleep_status.address));
96 }
97 
98 #ifdef CONFIG_ACPI_SLEEP
99 static u32 acpi_target_sleep_state = ACPI_STATE_S0;
100 
101 u32 acpi_target_system_state(void)
102 {
103 	return acpi_target_sleep_state;
104 }
105 EXPORT_SYMBOL_GPL(acpi_target_system_state);
106 
107 static bool pwr_btn_event_pending;
108 
109 /*
110  * The ACPI specification wants us to save NVS memory regions during hibernation
111  * and to restore them during the subsequent resume.  Windows does that also for
112  * suspend to RAM.  However, it is known that this mechanism does not work on
113  * all machines, so we allow the user to disable it with the help of the
114  * 'acpi_sleep=nonvs' kernel command line option.
115  */
116 static bool nvs_nosave;
117 
118 void __init acpi_nvs_nosave(void)
119 {
120 	nvs_nosave = true;
121 }
122 
123 /*
124  * The ACPI specification wants us to save NVS memory regions during hibernation
125  * but says nothing about saving NVS during S3.  Not all versions of Windows
126  * save NVS on S3 suspend either, and it is clear that not all systems need
127  * NVS to be saved at S3 time.  To improve suspend/resume time, allow the
128  * user to disable saving NVS on S3 if their system does not require it, but
129  * continue to save/restore NVS for S4 as specified.
130  */
131 static bool nvs_nosave_s3;
132 
133 void __init acpi_nvs_nosave_s3(void)
134 {
135 	nvs_nosave_s3 = true;
136 }
137 
138 static int __init init_nvs_save_s3(const struct dmi_system_id *d)
139 {
140 	nvs_nosave_s3 = false;
141 	return 0;
142 }
143 
144 /*
145  * ACPI 1.0 wants us to execute _PTS before suspending devices, so we allow the
146  * user to request that behavior by using the 'acpi_old_suspend_ordering'
147  * kernel command line option that causes the following variable to be set.
148  */
149 static bool old_suspend_ordering;
150 
151 void __init acpi_old_suspend_ordering(void)
152 {
153 	old_suspend_ordering = true;
154 }
155 
156 static int __init init_old_suspend_ordering(const struct dmi_system_id *d)
157 {
158 	acpi_old_suspend_ordering();
159 	return 0;
160 }
161 
162 static int __init init_nvs_nosave(const struct dmi_system_id *d)
163 {
164 	acpi_nvs_nosave();
165 	return 0;
166 }
167 
168 bool acpi_sleep_default_s3;
169 
170 static int __init init_default_s3(const struct dmi_system_id *d)
171 {
172 	acpi_sleep_default_s3 = true;
173 	return 0;
174 }
175 
176 static const struct dmi_system_id acpisleep_dmi_table[] __initconst = {
177 	{
178 	.callback = init_old_suspend_ordering,
179 	.ident = "Abit KN9 (nForce4 variant)",
180 	.matches = {
181 		DMI_MATCH(DMI_BOARD_VENDOR, "http://www.abit.com.tw/"),
182 		DMI_MATCH(DMI_BOARD_NAME, "KN9 Series(NF-CK804)"),
183 		},
184 	},
185 	{
186 	.callback = init_old_suspend_ordering,
187 	.ident = "HP xw4600 Workstation",
188 	.matches = {
189 		DMI_MATCH(DMI_SYS_VENDOR, "Hewlett-Packard"),
190 		DMI_MATCH(DMI_PRODUCT_NAME, "HP xw4600 Workstation"),
191 		},
192 	},
193 	{
194 	.callback = init_old_suspend_ordering,
195 	.ident = "Asus Pundit P1-AH2 (M2N8L motherboard)",
196 	.matches = {
197 		DMI_MATCH(DMI_BOARD_VENDOR, "ASUSTek Computer INC."),
198 		DMI_MATCH(DMI_BOARD_NAME, "M2N8L"),
199 		},
200 	},
201 	{
202 	.callback = init_old_suspend_ordering,
203 	.ident = "Panasonic CF51-2L",
204 	.matches = {
205 		DMI_MATCH(DMI_BOARD_VENDOR,
206 				"Matsushita Electric Industrial Co.,Ltd."),
207 		DMI_MATCH(DMI_BOARD_NAME, "CF51-2L"),
208 		},
209 	},
210 	{
211 	.callback = init_nvs_nosave,
212 	.ident = "Sony Vaio VGN-FW41E_H",
213 	.matches = {
214 		DMI_MATCH(DMI_SYS_VENDOR, "Sony Corporation"),
215 		DMI_MATCH(DMI_PRODUCT_NAME, "VGN-FW41E_H"),
216 		},
217 	},
218 	{
219 	.callback = init_nvs_nosave,
220 	.ident = "Sony Vaio VGN-FW21E",
221 	.matches = {
222 		DMI_MATCH(DMI_SYS_VENDOR, "Sony Corporation"),
223 		DMI_MATCH(DMI_PRODUCT_NAME, "VGN-FW21E"),
224 		},
225 	},
226 	{
227 	.callback = init_nvs_nosave,
228 	.ident = "Sony Vaio VGN-FW21M",
229 	.matches = {
230 		DMI_MATCH(DMI_SYS_VENDOR, "Sony Corporation"),
231 		DMI_MATCH(DMI_PRODUCT_NAME, "VGN-FW21M"),
232 		},
233 	},
234 	{
235 	.callback = init_nvs_nosave,
236 	.ident = "Sony Vaio VPCEB17FX",
237 	.matches = {
238 		DMI_MATCH(DMI_SYS_VENDOR, "Sony Corporation"),
239 		DMI_MATCH(DMI_PRODUCT_NAME, "VPCEB17FX"),
240 		},
241 	},
242 	{
243 	.callback = init_nvs_nosave,
244 	.ident = "Sony Vaio VGN-SR11M",
245 	.matches = {
246 		DMI_MATCH(DMI_SYS_VENDOR, "Sony Corporation"),
247 		DMI_MATCH(DMI_PRODUCT_NAME, "VGN-SR11M"),
248 		},
249 	},
250 	{
251 	.callback = init_nvs_nosave,
252 	.ident = "Everex StepNote Series",
253 	.matches = {
254 		DMI_MATCH(DMI_SYS_VENDOR, "Everex Systems, Inc."),
255 		DMI_MATCH(DMI_PRODUCT_NAME, "Everex StepNote Series"),
256 		},
257 	},
258 	{
259 	.callback = init_nvs_nosave,
260 	.ident = "Sony Vaio VPCEB1Z1E",
261 	.matches = {
262 		DMI_MATCH(DMI_SYS_VENDOR, "Sony Corporation"),
263 		DMI_MATCH(DMI_PRODUCT_NAME, "VPCEB1Z1E"),
264 		},
265 	},
266 	{
267 	.callback = init_nvs_nosave,
268 	.ident = "Sony Vaio VGN-NW130D",
269 	.matches = {
270 		DMI_MATCH(DMI_SYS_VENDOR, "Sony Corporation"),
271 		DMI_MATCH(DMI_PRODUCT_NAME, "VGN-NW130D"),
272 		},
273 	},
274 	{
275 	.callback = init_nvs_nosave,
276 	.ident = "Sony Vaio VPCCW29FX",
277 	.matches = {
278 		DMI_MATCH(DMI_SYS_VENDOR, "Sony Corporation"),
279 		DMI_MATCH(DMI_PRODUCT_NAME, "VPCCW29FX"),
280 		},
281 	},
282 	{
283 	.callback = init_nvs_nosave,
284 	.ident = "Averatec AV1020-ED2",
285 	.matches = {
286 		DMI_MATCH(DMI_SYS_VENDOR, "AVERATEC"),
287 		DMI_MATCH(DMI_PRODUCT_NAME, "1000 Series"),
288 		},
289 	},
290 	{
291 	.callback = init_old_suspend_ordering,
292 	.ident = "Asus A8N-SLI DELUXE",
293 	.matches = {
294 		DMI_MATCH(DMI_BOARD_VENDOR, "ASUSTeK Computer INC."),
295 		DMI_MATCH(DMI_BOARD_NAME, "A8N-SLI DELUXE"),
296 		},
297 	},
298 	{
299 	.callback = init_old_suspend_ordering,
300 	.ident = "Asus A8N-SLI Premium",
301 	.matches = {
302 		DMI_MATCH(DMI_BOARD_VENDOR, "ASUSTeK Computer INC."),
303 		DMI_MATCH(DMI_BOARD_NAME, "A8N-SLI Premium"),
304 		},
305 	},
306 	{
307 	.callback = init_nvs_nosave,
308 	.ident = "Sony Vaio VGN-SR26GN_P",
309 	.matches = {
310 		DMI_MATCH(DMI_SYS_VENDOR, "Sony Corporation"),
311 		DMI_MATCH(DMI_PRODUCT_NAME, "VGN-SR26GN_P"),
312 		},
313 	},
314 	{
315 	.callback = init_nvs_nosave,
316 	.ident = "Sony Vaio VPCEB1S1E",
317 	.matches = {
318 		DMI_MATCH(DMI_SYS_VENDOR, "Sony Corporation"),
319 		DMI_MATCH(DMI_PRODUCT_NAME, "VPCEB1S1E"),
320 		},
321 	},
322 	{
323 	.callback = init_nvs_nosave,
324 	.ident = "Sony Vaio VGN-FW520F",
325 	.matches = {
326 		DMI_MATCH(DMI_SYS_VENDOR, "Sony Corporation"),
327 		DMI_MATCH(DMI_PRODUCT_NAME, "VGN-FW520F"),
328 		},
329 	},
330 	{
331 	.callback = init_nvs_nosave,
332 	.ident = "Asus K54C",
333 	.matches = {
334 		DMI_MATCH(DMI_SYS_VENDOR, "ASUSTeK Computer Inc."),
335 		DMI_MATCH(DMI_PRODUCT_NAME, "K54C"),
336 		},
337 	},
338 	{
339 	.callback = init_nvs_nosave,
340 	.ident = "Asus K54HR",
341 	.matches = {
342 		DMI_MATCH(DMI_SYS_VENDOR, "ASUSTeK Computer Inc."),
343 		DMI_MATCH(DMI_PRODUCT_NAME, "K54HR"),
344 		},
345 	},
346 	{
347 	.callback = init_nvs_save_s3,
348 	.ident = "Asus 1025C",
349 	.matches = {
350 		DMI_MATCH(DMI_SYS_VENDOR, "ASUSTeK COMPUTER INC."),
351 		DMI_MATCH(DMI_PRODUCT_NAME, "1025C"),
352 		},
353 	},
354 	/*
355 	 * https://bugzilla.kernel.org/show_bug.cgi?id=189431
356 	 * Lenovo G50-45 is a platform later than 2012, but needs nvs memory
357 	 * saving during S3.
358 	 */
359 	{
360 	.callback = init_nvs_save_s3,
361 	.ident = "Lenovo G50-45",
362 	.matches = {
363 		DMI_MATCH(DMI_SYS_VENDOR, "LENOVO"),
364 		DMI_MATCH(DMI_PRODUCT_NAME, "80E3"),
365 		},
366 	},
367 	{
368 	.callback = init_nvs_save_s3,
369 	.ident = "Lenovo G40-45",
370 	.matches = {
371 		DMI_MATCH(DMI_SYS_VENDOR, "LENOVO"),
372 		DMI_MATCH(DMI_PRODUCT_NAME, "80E1"),
373 		},
374 	},
375 	/*
376 	 * ThinkPad X1 Tablet(2016) cannot do suspend-to-idle using
377 	 * the Low Power S0 Idle firmware interface (see
378 	 * https://bugzilla.kernel.org/show_bug.cgi?id=199057).
379 	 */
380 	{
381 	.callback = init_default_s3,
382 	.ident = "ThinkPad X1 Tablet(2016)",
383 	.matches = {
384 		DMI_MATCH(DMI_SYS_VENDOR, "LENOVO"),
385 		DMI_MATCH(DMI_PRODUCT_NAME, "20GGA00L00"),
386 		},
387 	},
388 	{},
389 };
390 
391 static bool ignore_blacklist;
392 
393 void __init acpi_sleep_no_blacklist(void)
394 {
395 	ignore_blacklist = true;
396 }
397 
398 static void __init acpi_sleep_dmi_check(void)
399 {
400 	if (ignore_blacklist)
401 		return;
402 
403 	if (dmi_get_bios_year() >= 2012)
404 		acpi_nvs_nosave_s3();
405 
406 	dmi_check_system(acpisleep_dmi_table);
407 }
408 
409 /**
410  * acpi_pm_freeze - Disable the GPEs and suspend EC transactions.
411  */
412 static int acpi_pm_freeze(void)
413 {
414 	acpi_disable_all_gpes();
415 	acpi_os_wait_events_complete();
416 	acpi_ec_block_transactions();
417 	return 0;
418 }
419 
420 /**
421  * acpi_pm_pre_suspend - Enable wakeup devices, "freeze" EC and save NVS.
422  */
423 static int acpi_pm_pre_suspend(void)
424 {
425 	acpi_pm_freeze();
426 	return suspend_nvs_save();
427 }
428 
429 /**
430  *	__acpi_pm_prepare - Prepare the platform to enter the target state.
431  *
432  *	If necessary, set the firmware waking vector and do arch-specific
433  *	nastiness to get the wakeup code to the waking vector.
434  */
435 static int __acpi_pm_prepare(void)
436 {
437 	int error = acpi_sleep_prepare(acpi_target_sleep_state);
438 	if (error)
439 		acpi_target_sleep_state = ACPI_STATE_S0;
440 
441 	return error;
442 }
443 
444 /**
445  *	acpi_pm_prepare - Prepare the platform to enter the target sleep
446  *		state and disable the GPEs.
447  */
448 static int acpi_pm_prepare(void)
449 {
450 	int error = __acpi_pm_prepare();
451 	if (!error)
452 		error = acpi_pm_pre_suspend();
453 
454 	return error;
455 }
456 
457 /**
458  *	acpi_pm_finish - Instruct the platform to leave a sleep state.
459  *
460  *	This is called after we wake back up (or if entering the sleep state
461  *	failed).
462  */
463 static void acpi_pm_finish(void)
464 {
465 	struct acpi_device *pwr_btn_adev;
466 	u32 acpi_state = acpi_target_sleep_state;
467 
468 	acpi_ec_unblock_transactions();
469 	suspend_nvs_free();
470 
471 	if (acpi_state == ACPI_STATE_S0)
472 		return;
473 
474 	pr_info("Waking up from system sleep state S%d\n", acpi_state);
475 	acpi_disable_wakeup_devices(acpi_state);
476 	acpi_leave_sleep_state(acpi_state);
477 
478 	/* reset firmware waking vector */
479 	acpi_set_waking_vector(0);
480 
481 	acpi_target_sleep_state = ACPI_STATE_S0;
482 
483 	acpi_resume_power_resources();
484 
485 	/* If we were woken with the fixed power button, provide a small
486 	 * hint to userspace in the form of a wakeup event on the fixed power
487 	 * button device (if it can be found).
488 	 *
489 	 * We delay the event generation til now, as the PM layer requires
490 	 * timekeeping to be running before we generate events. */
491 	if (!pwr_btn_event_pending)
492 		return;
493 
494 	pwr_btn_event_pending = false;
495 	pwr_btn_adev = acpi_dev_get_first_match_dev(ACPI_BUTTON_HID_POWERF,
496 						    NULL, -1);
497 	if (pwr_btn_adev) {
498 		pm_wakeup_event(&pwr_btn_adev->dev, 0);
499 		acpi_dev_put(pwr_btn_adev);
500 	}
501 }
502 
503 /**
504  * acpi_pm_start - Start system PM transition.
505  */
506 static void acpi_pm_start(u32 acpi_state)
507 {
508 	acpi_target_sleep_state = acpi_state;
509 	acpi_sleep_tts_switch(acpi_target_sleep_state);
510 	acpi_scan_lock_acquire();
511 }
512 
513 /**
514  * acpi_pm_end - Finish up system PM transition.
515  */
516 static void acpi_pm_end(void)
517 {
518 	acpi_turn_off_unused_power_resources();
519 	acpi_scan_lock_release();
520 	/*
521 	 * This is necessary in case acpi_pm_finish() is not called during a
522 	 * failing transition to a sleep state.
523 	 */
524 	acpi_target_sleep_state = ACPI_STATE_S0;
525 	acpi_sleep_tts_switch(acpi_target_sleep_state);
526 }
527 #else /* !CONFIG_ACPI_SLEEP */
528 #define sleep_no_lps0	(1)
529 #define acpi_target_sleep_state	ACPI_STATE_S0
530 #define acpi_sleep_default_s3	(1)
531 static inline void acpi_sleep_dmi_check(void) {}
532 #endif /* CONFIG_ACPI_SLEEP */
533 
534 #ifdef CONFIG_SUSPEND
535 static u32 acpi_suspend_states[] = {
536 	[PM_SUSPEND_ON] = ACPI_STATE_S0,
537 	[PM_SUSPEND_STANDBY] = ACPI_STATE_S1,
538 	[PM_SUSPEND_MEM] = ACPI_STATE_S3,
539 	[PM_SUSPEND_MAX] = ACPI_STATE_S5
540 };
541 
542 /**
543  *	acpi_suspend_begin - Set the target system sleep state to the state
544  *		associated with given @pm_state, if supported.
545  */
546 static int acpi_suspend_begin(suspend_state_t pm_state)
547 {
548 	u32 acpi_state = acpi_suspend_states[pm_state];
549 	int error;
550 
551 	error = (nvs_nosave || nvs_nosave_s3) ? 0 : suspend_nvs_alloc();
552 	if (error)
553 		return error;
554 
555 	if (!sleep_states[acpi_state]) {
556 		pr_err("ACPI does not support sleep state S%u\n", acpi_state);
557 		return -ENOSYS;
558 	}
559 	if (acpi_state > ACPI_STATE_S1)
560 		pm_set_suspend_via_firmware();
561 
562 	acpi_pm_start(acpi_state);
563 	return 0;
564 }
565 
566 /**
567  *	acpi_suspend_enter - Actually enter a sleep state.
568  *	@pm_state: ignored
569  *
570  *	Flush caches and go to sleep. For STR we have to call arch-specific
571  *	assembly, which in turn call acpi_enter_sleep_state().
572  *	It's unfortunate, but it works. Please fix if you're feeling frisky.
573  */
574 static int acpi_suspend_enter(suspend_state_t pm_state)
575 {
576 	acpi_status status = AE_OK;
577 	u32 acpi_state = acpi_target_sleep_state;
578 	int error;
579 
580 	trace_suspend_resume(TPS("acpi_suspend"), acpi_state, true);
581 	switch (acpi_state) {
582 	case ACPI_STATE_S1:
583 		barrier();
584 		status = acpi_enter_sleep_state(acpi_state);
585 		break;
586 
587 	case ACPI_STATE_S3:
588 		if (!acpi_suspend_lowlevel)
589 			return -ENOSYS;
590 		error = acpi_suspend_lowlevel();
591 		if (error)
592 			return error;
593 		pr_info("Low-level resume complete\n");
594 		pm_set_resume_via_firmware();
595 		break;
596 	}
597 	trace_suspend_resume(TPS("acpi_suspend"), acpi_state, false);
598 
599 	/* This violates the spec but is required for bug compatibility. */
600 	acpi_write_bit_register(ACPI_BITREG_SCI_ENABLE, 1);
601 
602 	/* Reprogram control registers */
603 	acpi_leave_sleep_state_prep(acpi_state);
604 
605 	/* ACPI 3.0 specs (P62) says that it's the responsibility
606 	 * of the OSPM to clear the status bit [ implying that the
607 	 * POWER_BUTTON event should not reach userspace ]
608 	 *
609 	 * However, we do generate a small hint for userspace in the form of
610 	 * a wakeup event. We flag this condition for now and generate the
611 	 * event later, as we're currently too early in resume to be able to
612 	 * generate wakeup events.
613 	 */
614 	if (ACPI_SUCCESS(status) && (acpi_state == ACPI_STATE_S3)) {
615 		acpi_event_status pwr_btn_status = ACPI_EVENT_FLAG_DISABLED;
616 
617 		acpi_get_event_status(ACPI_EVENT_POWER_BUTTON, &pwr_btn_status);
618 
619 		if (pwr_btn_status & ACPI_EVENT_FLAG_STATUS_SET) {
620 			acpi_clear_event(ACPI_EVENT_POWER_BUTTON);
621 			/* Flag for later */
622 			pwr_btn_event_pending = true;
623 		}
624 	}
625 
626 	/*
627 	 * Disable all GPE and clear their status bits before interrupts are
628 	 * enabled. Some GPEs (like wakeup GPEs) have no handlers and this can
629 	 * prevent them from producing spurious interrups.
630 	 *
631 	 * acpi_leave_sleep_state() will reenable specific GPEs later.
632 	 *
633 	 * Because this code runs on one CPU with disabled interrupts (all of
634 	 * the other CPUs are offline at this time), it need not acquire any
635 	 * sleeping locks which may trigger an implicit preemption point even
636 	 * if there is no contention, so avoid doing that by using a low-level
637 	 * library routine here.
638 	 */
639 	acpi_hw_disable_all_gpes();
640 	/* Allow EC transactions to happen. */
641 	acpi_ec_unblock_transactions();
642 
643 	suspend_nvs_restore();
644 
645 	return ACPI_SUCCESS(status) ? 0 : -EFAULT;
646 }
647 
648 static int acpi_suspend_state_valid(suspend_state_t pm_state)
649 {
650 	u32 acpi_state;
651 
652 	switch (pm_state) {
653 	case PM_SUSPEND_ON:
654 	case PM_SUSPEND_STANDBY:
655 	case PM_SUSPEND_MEM:
656 		acpi_state = acpi_suspend_states[pm_state];
657 
658 		return sleep_states[acpi_state];
659 	default:
660 		return 0;
661 	}
662 }
663 
664 static const struct platform_suspend_ops acpi_suspend_ops = {
665 	.valid = acpi_suspend_state_valid,
666 	.begin = acpi_suspend_begin,
667 	.prepare_late = acpi_pm_prepare,
668 	.enter = acpi_suspend_enter,
669 	.wake = acpi_pm_finish,
670 	.end = acpi_pm_end,
671 };
672 
673 /**
674  *	acpi_suspend_begin_old - Set the target system sleep state to the
675  *		state associated with given @pm_state, if supported, and
676  *		execute the _PTS control method.  This function is used if the
677  *		pre-ACPI 2.0 suspend ordering has been requested.
678  */
679 static int acpi_suspend_begin_old(suspend_state_t pm_state)
680 {
681 	int error = acpi_suspend_begin(pm_state);
682 	if (!error)
683 		error = __acpi_pm_prepare();
684 
685 	return error;
686 }
687 
688 /*
689  * The following callbacks are used if the pre-ACPI 2.0 suspend ordering has
690  * been requested.
691  */
692 static const struct platform_suspend_ops acpi_suspend_ops_old = {
693 	.valid = acpi_suspend_state_valid,
694 	.begin = acpi_suspend_begin_old,
695 	.prepare_late = acpi_pm_pre_suspend,
696 	.enter = acpi_suspend_enter,
697 	.wake = acpi_pm_finish,
698 	.end = acpi_pm_end,
699 	.recover = acpi_pm_finish,
700 };
701 
702 static bool s2idle_wakeup;
703 
704 int acpi_s2idle_begin(void)
705 {
706 	acpi_scan_lock_acquire();
707 	return 0;
708 }
709 
710 int acpi_s2idle_prepare(void)
711 {
712 	if (acpi_sci_irq_valid()) {
713 		int error;
714 
715 		error = enable_irq_wake(acpi_sci_irq);
716 		if (error)
717 			pr_warn("Warning: Failed to enable wakeup from IRQ %d: %d\n",
718 				acpi_sci_irq, error);
719 
720 		acpi_ec_set_gpe_wake_mask(ACPI_GPE_ENABLE);
721 	}
722 
723 	acpi_enable_wakeup_devices(ACPI_STATE_S0);
724 
725 	/* Change the configuration of GPEs to avoid spurious wakeup. */
726 	acpi_enable_all_wakeup_gpes();
727 	acpi_os_wait_events_complete();
728 
729 	s2idle_wakeup = true;
730 	return 0;
731 }
732 
733 bool acpi_s2idle_wake(void)
734 {
735 	if (!acpi_sci_irq_valid())
736 		return pm_wakeup_pending();
737 
738 	while (pm_wakeup_pending()) {
739 		/*
740 		 * If IRQD_WAKEUP_ARMED is set for the SCI at this point, the
741 		 * SCI has not triggered while suspended, so bail out (the
742 		 * wakeup is pending anyway and the SCI is not the source of
743 		 * it).
744 		 */
745 		if (irqd_is_wakeup_armed(irq_get_irq_data(acpi_sci_irq))) {
746 			pm_pr_dbg("Wakeup unrelated to ACPI SCI\n");
747 			return true;
748 		}
749 
750 		/*
751 		 * If the status bit of any enabled fixed event is set, the
752 		 * wakeup is regarded as valid.
753 		 */
754 		if (acpi_any_fixed_event_status_set()) {
755 			pm_pr_dbg("ACPI fixed event wakeup\n");
756 			return true;
757 		}
758 
759 		/* Check wakeups from drivers sharing the SCI. */
760 		if (acpi_check_wakeup_handlers()) {
761 			pm_pr_dbg("ACPI custom handler wakeup\n");
762 			return true;
763 		}
764 
765 		/*
766 		 * Check non-EC GPE wakeups and if there are none, cancel the
767 		 * SCI-related wakeup and dispatch the EC GPE.
768 		 */
769 		if (acpi_ec_dispatch_gpe()) {
770 			pm_pr_dbg("ACPI non-EC GPE wakeup\n");
771 			return true;
772 		}
773 
774 		acpi_os_wait_events_complete();
775 
776 		/*
777 		 * The SCI is in the "suspended" state now and it cannot produce
778 		 * new wakeup events till the rearming below, so if any of them
779 		 * are pending here, they must be resulting from the processing
780 		 * of EC events above or coming from somewhere else.
781 		 */
782 		if (pm_wakeup_pending()) {
783 			pm_pr_dbg("Wakeup after ACPI Notify sync\n");
784 			return true;
785 		}
786 
787 		pm_pr_dbg("Rearming ACPI SCI for wakeup\n");
788 
789 		pm_wakeup_clear(acpi_sci_irq);
790 		rearm_wake_irq(acpi_sci_irq);
791 	}
792 
793 	return false;
794 }
795 
796 void acpi_s2idle_restore(void)
797 {
798 	/*
799 	 * Drain pending events before restoring the working-state configuration
800 	 * of GPEs.
801 	 */
802 	acpi_os_wait_events_complete(); /* synchronize GPE processing */
803 	acpi_ec_flush_work(); /* flush the EC driver's workqueues */
804 	acpi_os_wait_events_complete(); /* synchronize Notify handling */
805 
806 	s2idle_wakeup = false;
807 
808 	acpi_enable_all_runtime_gpes();
809 
810 	acpi_disable_wakeup_devices(ACPI_STATE_S0);
811 
812 	if (acpi_sci_irq_valid()) {
813 		acpi_ec_set_gpe_wake_mask(ACPI_GPE_DISABLE);
814 		disable_irq_wake(acpi_sci_irq);
815 	}
816 }
817 
818 void acpi_s2idle_end(void)
819 {
820 	acpi_scan_lock_release();
821 }
822 
823 static const struct platform_s2idle_ops acpi_s2idle_ops = {
824 	.begin = acpi_s2idle_begin,
825 	.prepare = acpi_s2idle_prepare,
826 	.wake = acpi_s2idle_wake,
827 	.restore = acpi_s2idle_restore,
828 	.end = acpi_s2idle_end,
829 };
830 
831 void __weak acpi_s2idle_setup(void)
832 {
833 	if (acpi_gbl_FADT.flags & ACPI_FADT_LOW_POWER_S0)
834 		pr_info("Efficient low-power S0 idle declared\n");
835 
836 	s2idle_set_ops(&acpi_s2idle_ops);
837 }
838 
839 static void __init acpi_sleep_suspend_setup(void)
840 {
841 	bool suspend_ops_needed = false;
842 	int i;
843 
844 	for (i = ACPI_STATE_S1; i < ACPI_STATE_S4; i++)
845 		if (acpi_sleep_state_supported(i)) {
846 			sleep_states[i] = 1;
847 			suspend_ops_needed = true;
848 		}
849 
850 	if (suspend_ops_needed)
851 		suspend_set_ops(old_suspend_ordering ?
852 				&acpi_suspend_ops_old : &acpi_suspend_ops);
853 
854 	acpi_s2idle_setup();
855 }
856 
857 #else /* !CONFIG_SUSPEND */
858 #define s2idle_wakeup		(false)
859 static inline void acpi_sleep_suspend_setup(void) {}
860 #endif /* !CONFIG_SUSPEND */
861 
862 bool acpi_s2idle_wakeup(void)
863 {
864 	return s2idle_wakeup;
865 }
866 
867 #ifdef CONFIG_PM_SLEEP
868 static u32 saved_bm_rld;
869 
870 static int  acpi_save_bm_rld(void)
871 {
872 	acpi_read_bit_register(ACPI_BITREG_BUS_MASTER_RLD, &saved_bm_rld);
873 	return 0;
874 }
875 
876 static void  acpi_restore_bm_rld(void)
877 {
878 	u32 resumed_bm_rld = 0;
879 
880 	acpi_read_bit_register(ACPI_BITREG_BUS_MASTER_RLD, &resumed_bm_rld);
881 	if (resumed_bm_rld == saved_bm_rld)
882 		return;
883 
884 	acpi_write_bit_register(ACPI_BITREG_BUS_MASTER_RLD, saved_bm_rld);
885 }
886 
887 static struct syscore_ops acpi_sleep_syscore_ops = {
888 	.suspend = acpi_save_bm_rld,
889 	.resume = acpi_restore_bm_rld,
890 };
891 
892 static void acpi_sleep_syscore_init(void)
893 {
894 	register_syscore_ops(&acpi_sleep_syscore_ops);
895 }
896 #else
897 static inline void acpi_sleep_syscore_init(void) {}
898 #endif /* CONFIG_PM_SLEEP */
899 
900 #ifdef CONFIG_HIBERNATION
901 static unsigned long s4_hardware_signature;
902 static struct acpi_table_facs *facs;
903 int acpi_check_s4_hw_signature = -1; /* Default behaviour is just to warn */
904 
905 static int acpi_hibernation_begin(pm_message_t stage)
906 {
907 	if (!nvs_nosave) {
908 		int error = suspend_nvs_alloc();
909 		if (error)
910 			return error;
911 	}
912 
913 	if (stage.event == PM_EVENT_HIBERNATE)
914 		pm_set_suspend_via_firmware();
915 
916 	acpi_pm_start(ACPI_STATE_S4);
917 	return 0;
918 }
919 
920 static int acpi_hibernation_enter(void)
921 {
922 	acpi_status status = AE_OK;
923 
924 	/* This shouldn't return.  If it returns, we have a problem */
925 	status = acpi_enter_sleep_state(ACPI_STATE_S4);
926 	/* Reprogram control registers */
927 	acpi_leave_sleep_state_prep(ACPI_STATE_S4);
928 
929 	return ACPI_SUCCESS(status) ? 0 : -EFAULT;
930 }
931 
932 static void acpi_hibernation_leave(void)
933 {
934 	pm_set_resume_via_firmware();
935 	/*
936 	 * If ACPI is not enabled by the BIOS and the boot kernel, we need to
937 	 * enable it here.
938 	 */
939 	acpi_enable();
940 	/* Reprogram control registers */
941 	acpi_leave_sleep_state_prep(ACPI_STATE_S4);
942 	/* Check the hardware signature */
943 	if (facs && s4_hardware_signature != facs->hardware_signature)
944 		pr_crit("Hardware changed while hibernated, success doubtful!\n");
945 	/* Restore the NVS memory area */
946 	suspend_nvs_restore();
947 	/* Allow EC transactions to happen. */
948 	acpi_ec_unblock_transactions();
949 }
950 
951 static void acpi_pm_thaw(void)
952 {
953 	acpi_ec_unblock_transactions();
954 	acpi_enable_all_runtime_gpes();
955 }
956 
957 static const struct platform_hibernation_ops acpi_hibernation_ops = {
958 	.begin = acpi_hibernation_begin,
959 	.end = acpi_pm_end,
960 	.pre_snapshot = acpi_pm_prepare,
961 	.finish = acpi_pm_finish,
962 	.prepare = acpi_pm_prepare,
963 	.enter = acpi_hibernation_enter,
964 	.leave = acpi_hibernation_leave,
965 	.pre_restore = acpi_pm_freeze,
966 	.restore_cleanup = acpi_pm_thaw,
967 };
968 
969 /**
970  *	acpi_hibernation_begin_old - Set the target system sleep state to
971  *		ACPI_STATE_S4 and execute the _PTS control method.  This
972  *		function is used if the pre-ACPI 2.0 suspend ordering has been
973  *		requested.
974  */
975 static int acpi_hibernation_begin_old(pm_message_t stage)
976 {
977 	int error;
978 	/*
979 	 * The _TTS object should always be evaluated before the _PTS object.
980 	 * When the old_suspended_ordering is true, the _PTS object is
981 	 * evaluated in the acpi_sleep_prepare.
982 	 */
983 	acpi_sleep_tts_switch(ACPI_STATE_S4);
984 
985 	error = acpi_sleep_prepare(ACPI_STATE_S4);
986 	if (error)
987 		return error;
988 
989 	if (!nvs_nosave) {
990 		error = suspend_nvs_alloc();
991 		if (error)
992 			return error;
993 	}
994 
995 	if (stage.event == PM_EVENT_HIBERNATE)
996 		pm_set_suspend_via_firmware();
997 
998 	acpi_target_sleep_state = ACPI_STATE_S4;
999 	acpi_scan_lock_acquire();
1000 	return 0;
1001 }
1002 
1003 /*
1004  * The following callbacks are used if the pre-ACPI 2.0 suspend ordering has
1005  * been requested.
1006  */
1007 static const struct platform_hibernation_ops acpi_hibernation_ops_old = {
1008 	.begin = acpi_hibernation_begin_old,
1009 	.end = acpi_pm_end,
1010 	.pre_snapshot = acpi_pm_pre_suspend,
1011 	.prepare = acpi_pm_freeze,
1012 	.finish = acpi_pm_finish,
1013 	.enter = acpi_hibernation_enter,
1014 	.leave = acpi_hibernation_leave,
1015 	.pre_restore = acpi_pm_freeze,
1016 	.restore_cleanup = acpi_pm_thaw,
1017 	.recover = acpi_pm_finish,
1018 };
1019 
1020 static void acpi_sleep_hibernate_setup(void)
1021 {
1022 	if (!acpi_sleep_state_supported(ACPI_STATE_S4))
1023 		return;
1024 
1025 	hibernation_set_ops(old_suspend_ordering ?
1026 			&acpi_hibernation_ops_old : &acpi_hibernation_ops);
1027 	sleep_states[ACPI_STATE_S4] = 1;
1028 	if (!acpi_check_s4_hw_signature)
1029 		return;
1030 
1031 	acpi_get_table(ACPI_SIG_FACS, 1, (struct acpi_table_header **)&facs);
1032 	if (facs) {
1033 		/*
1034 		 * s4_hardware_signature is the local variable which is just
1035 		 * used to warn about mismatch after we're attempting to
1036 		 * resume (in violation of the ACPI specification.)
1037 		 */
1038 		s4_hardware_signature = facs->hardware_signature;
1039 
1040 		if (acpi_check_s4_hw_signature > 0) {
1041 			/*
1042 			 * If we're actually obeying the ACPI specification
1043 			 * then the signature is written out as part of the
1044 			 * swsusp header, in order to allow the boot kernel
1045 			 * to gracefully decline to resume.
1046 			 */
1047 			swsusp_hardware_signature = facs->hardware_signature;
1048 		}
1049 	}
1050 }
1051 #else /* !CONFIG_HIBERNATION */
1052 static inline void acpi_sleep_hibernate_setup(void) {}
1053 #endif /* !CONFIG_HIBERNATION */
1054 
1055 static int acpi_power_off_prepare(struct sys_off_data *data)
1056 {
1057 	/* Prepare to power off the system */
1058 	acpi_sleep_prepare(ACPI_STATE_S5);
1059 	acpi_disable_all_gpes();
1060 	acpi_os_wait_events_complete();
1061 	return NOTIFY_DONE;
1062 }
1063 
1064 static int acpi_power_off(struct sys_off_data *data)
1065 {
1066 	/* acpi_sleep_prepare(ACPI_STATE_S5) should have already been called */
1067 	pr_debug("%s called\n", __func__);
1068 	local_irq_disable();
1069 	acpi_enter_sleep_state(ACPI_STATE_S5);
1070 	return NOTIFY_DONE;
1071 }
1072 
1073 int __init acpi_sleep_init(void)
1074 {
1075 	char supported[ACPI_S_STATE_COUNT * 3 + 1];
1076 	char *pos = supported;
1077 	int i;
1078 
1079 	acpi_sleep_dmi_check();
1080 
1081 	sleep_states[ACPI_STATE_S0] = 1;
1082 
1083 	acpi_sleep_syscore_init();
1084 	acpi_sleep_suspend_setup();
1085 	acpi_sleep_hibernate_setup();
1086 
1087 	if (acpi_sleep_state_supported(ACPI_STATE_S5)) {
1088 		sleep_states[ACPI_STATE_S5] = 1;
1089 
1090 		register_sys_off_handler(SYS_OFF_MODE_POWER_OFF_PREPARE,
1091 					 SYS_OFF_PRIO_FIRMWARE,
1092 					 acpi_power_off_prepare, NULL);
1093 
1094 		register_sys_off_handler(SYS_OFF_MODE_POWER_OFF,
1095 					 SYS_OFF_PRIO_FIRMWARE,
1096 					 acpi_power_off, NULL);
1097 
1098 		/*
1099 		 * Windows uses S5 for reboot, so some BIOSes depend on it to
1100 		 * perform proper reboot.
1101 		 */
1102 		register_sys_off_handler(SYS_OFF_MODE_RESTART_PREPARE,
1103 					 SYS_OFF_PRIO_FIRMWARE,
1104 					 acpi_power_off_prepare, NULL);
1105 	} else {
1106 		acpi_no_s5 = true;
1107 	}
1108 
1109 	supported[0] = 0;
1110 	for (i = 0; i < ACPI_S_STATE_COUNT; i++) {
1111 		if (sleep_states[i])
1112 			pos += sprintf(pos, " S%d", i);
1113 	}
1114 	pr_info("(supports%s)\n", supported);
1115 
1116 	/*
1117 	 * Register the tts_notifier to reboot notifier list so that the _TTS
1118 	 * object can also be evaluated when the system enters S5.
1119 	 */
1120 	register_reboot_notifier(&tts_notifier);
1121 	return 0;
1122 }
1123