H A D | fair.c | 1054 static int select_idle_sibling(struct task_struct *p, int prev_cpu, int cpu); 7043 wake_affine_idle(int this_cpu, int prev_cpu, int sync) in wake_affine_idle() 7051 * If the prev_cpu is idle and cache affine then avoid a migration. in wake_affine_idle() 7053 * is more important than cache hot data on the prev_cpu and from in wake_affine_idle() 7057 if (available_idle_cpu(this_cpu) && cpus_share_cache(this_cpu, prev_cpu)) in wake_affine_idle() 7058 return available_idle_cpu(prev_cpu) ? prev_cpu : this_cpu; in wake_affine_idle() 7063 if (available_idle_cpu(prev_cpu)) in wake_affine_idle() 7064 return prev_cpu; in wake_affine_idle() 7071 int this_cpu, int prev_cpu, in in wake_affine_weight() 7041 wake_affine_idle(int this_cpu,int prev_cpu,int sync) wake_affine_idle() argument 7069 wake_affine_weight(struct sched_domain * sd,struct task_struct * p,int this_cpu,int prev_cpu,int sync) wake_affine_weight() argument 7111 wake_affine(struct sched_domain * sd,struct task_struct * p,int this_cpu,int prev_cpu,int sync) wake_affine() argument 7194 find_idlest_cpu(struct sched_domain * sd,struct task_struct * p,int cpu,int prev_cpu,int sd_flag) find_idlest_cpu() argument 7816 eenv_task_busy_time(struct energy_env * eenv,struct task_struct * p,int prev_cpu) eenv_task_busy_time() argument 7956 find_energy_efficient_cpu(struct task_struct * p,int prev_cpu) find_energy_efficient_cpu() argument 8155 select_task_rq_fair(struct task_struct * p,int prev_cpu,int wake_flags) select_task_rq_fair() argument [all...] |