1 /*
2  * ladder.c - the residency ladder algorithm
3  *
4  *  Copyright (C) 2001, 2002 Andy Grover <andrew.grover@intel.com>
5  *  Copyright (C) 2001, 2002 Paul Diefenbaugh <paul.s.diefenbaugh@intel.com>
6  *  Copyright (C) 2004, 2005 Dominik Brodowski <linux@brodo.de>
7  *
8  * (C) 2006-2007 Venkatesh Pallipadi <venkatesh.pallipadi@intel.com>
9  *               Shaohua Li <shaohua.li@intel.com>
10  *               Adam Belay <abelay@novell.com>
11  *
12  * This code is licenced under the GPL.
13  */
14 
15 #include <linux/kernel.h>
16 #include <linux/cpuidle.h>
17 #include <linux/pm_qos.h>
18 #include <linux/module.h>
19 #include <linux/jiffies.h>
20 
21 #include <asm/io.h>
22 #include <asm/uaccess.h>
23 
24 #define PROMOTION_COUNT 4
25 #define DEMOTION_COUNT 1
26 
27 struct ladder_device_state {
28 	struct {
29 		u32 promotion_count;
30 		u32 demotion_count;
31 		u32 promotion_time;
32 		u32 demotion_time;
33 	} threshold;
34 	struct {
35 		int promotion_count;
36 		int demotion_count;
37 	} stats;
38 };
39 
40 struct ladder_device {
41 	struct ladder_device_state states[CPUIDLE_STATE_MAX];
42 	int last_state_idx;
43 };
44 
45 static DEFINE_PER_CPU(struct ladder_device, ladder_devices);
46 
47 /**
48  * ladder_do_selection - prepares private data for a state change
49  * @ldev: the ladder device
50  * @old_idx: the current state index
51  * @new_idx: the new target state index
52  */
53 static inline void ladder_do_selection(struct ladder_device *ldev,
54 				       int old_idx, int new_idx)
55 {
56 	ldev->states[old_idx].stats.promotion_count = 0;
57 	ldev->states[old_idx].stats.demotion_count = 0;
58 	ldev->last_state_idx = new_idx;
59 }
60 
61 /**
62  * ladder_select_state - selects the next state to enter
63  * @drv: cpuidle driver
64  * @dev: the CPU
65  */
66 static int ladder_select_state(struct cpuidle_driver *drv,
67 				struct cpuidle_device *dev)
68 {
69 	struct ladder_device *ldev = this_cpu_ptr(&ladder_devices);
70 	struct ladder_device_state *last_state;
71 	int last_residency, last_idx = ldev->last_state_idx;
72 	int latency_req = pm_qos_request(PM_QOS_CPU_DMA_LATENCY);
73 
74 	/* Special case when user has set very strict latency requirement */
75 	if (unlikely(latency_req == 0)) {
76 		ladder_do_selection(ldev, last_idx, 0);
77 		return 0;
78 	}
79 
80 	last_state = &ldev->states[last_idx];
81 
82 	if (drv->states[last_idx].flags & CPUIDLE_FLAG_TIME_VALID) {
83 		last_residency = cpuidle_get_last_residency(dev) - \
84 					 drv->states[last_idx].exit_latency;
85 	}
86 	else
87 		last_residency = last_state->threshold.promotion_time + 1;
88 
89 	/* consider promotion */
90 	if (last_idx < drv->state_count - 1 &&
91 	    !drv->states[last_idx + 1].disabled &&
92 	    !dev->states_usage[last_idx + 1].disable &&
93 	    last_residency > last_state->threshold.promotion_time &&
94 	    drv->states[last_idx + 1].exit_latency <= latency_req) {
95 		last_state->stats.promotion_count++;
96 		last_state->stats.demotion_count = 0;
97 		if (last_state->stats.promotion_count >= last_state->threshold.promotion_count) {
98 			ladder_do_selection(ldev, last_idx, last_idx + 1);
99 			return last_idx + 1;
100 		}
101 	}
102 
103 	/* consider demotion */
104 	if (last_idx > CPUIDLE_DRIVER_STATE_START &&
105 	    (drv->states[last_idx].disabled ||
106 	    dev->states_usage[last_idx].disable ||
107 	    drv->states[last_idx].exit_latency > latency_req)) {
108 		int i;
109 
110 		for (i = last_idx - 1; i > CPUIDLE_DRIVER_STATE_START; i--) {
111 			if (drv->states[i].exit_latency <= latency_req)
112 				break;
113 		}
114 		ladder_do_selection(ldev, last_idx, i);
115 		return i;
116 	}
117 
118 	if (last_idx > CPUIDLE_DRIVER_STATE_START &&
119 	    last_residency < last_state->threshold.demotion_time) {
120 		last_state->stats.demotion_count++;
121 		last_state->stats.promotion_count = 0;
122 		if (last_state->stats.demotion_count >= last_state->threshold.demotion_count) {
123 			ladder_do_selection(ldev, last_idx, last_idx - 1);
124 			return last_idx - 1;
125 		}
126 	}
127 
128 	/* otherwise remain at the current state */
129 	return last_idx;
130 }
131 
132 /**
133  * ladder_enable_device - setup for the governor
134  * @drv: cpuidle driver
135  * @dev: the CPU
136  */
137 static int ladder_enable_device(struct cpuidle_driver *drv,
138 				struct cpuidle_device *dev)
139 {
140 	int i;
141 	struct ladder_device *ldev = &per_cpu(ladder_devices, dev->cpu);
142 	struct ladder_device_state *lstate;
143 	struct cpuidle_state *state;
144 
145 	ldev->last_state_idx = CPUIDLE_DRIVER_STATE_START;
146 
147 	for (i = CPUIDLE_DRIVER_STATE_START; i < drv->state_count; i++) {
148 		state = &drv->states[i];
149 		lstate = &ldev->states[i];
150 
151 		lstate->stats.promotion_count = 0;
152 		lstate->stats.demotion_count = 0;
153 
154 		lstate->threshold.promotion_count = PROMOTION_COUNT;
155 		lstate->threshold.demotion_count = DEMOTION_COUNT;
156 
157 		if (i < drv->state_count - 1)
158 			lstate->threshold.promotion_time = state->exit_latency;
159 		if (i > CPUIDLE_DRIVER_STATE_START)
160 			lstate->threshold.demotion_time = state->exit_latency;
161 	}
162 
163 	return 0;
164 }
165 
166 /**
167  * ladder_reflect - update the correct last_state_idx
168  * @dev: the CPU
169  * @index: the index of actual state entered
170  */
171 static void ladder_reflect(struct cpuidle_device *dev, int index)
172 {
173 	struct ladder_device *ldev = this_cpu_ptr(&ladder_devices);
174 	if (index > 0)
175 		ldev->last_state_idx = index;
176 }
177 
178 static struct cpuidle_governor ladder_governor = {
179 	.name =		"ladder",
180 	.rating =	10,
181 	.enable =	ladder_enable_device,
182 	.select =	ladder_select_state,
183 	.reflect =	ladder_reflect,
184 	.owner =	THIS_MODULE,
185 };
186 
187 /**
188  * init_ladder - initializes the governor
189  */
190 static int __init init_ladder(void)
191 {
192 	return cpuidle_register_governor(&ladder_governor);
193 }
194 
195 postcore_initcall(init_ladder);
196