Lines Matching defs:rt_rq

134 void init_rt_rq(struct rt_rq *rt_rq)  in init_rt_rq()
179 static inline struct rq *rq_of_rt_rq(struct rt_rq *rt_rq) in rq_of_rt_rq()
191 struct rt_rq *rt_rq = rt_se->rt_rq; in rq_of_rt_se() local
218 void init_tg_rt_entry(struct task_group *tg, struct rt_rq *rt_rq, in init_tg_rt_entry()
247 struct rt_rq *rt_rq; in alloc_rt_sched_group() local
294 static inline struct rq *rq_of_rt_rq(struct rt_rq *rt_rq) in rq_of_rt_rq()
365 static void update_rt_migration(struct rt_rq *rt_rq) in update_rt_migration()
378 static void inc_rt_migration(struct sched_rt_entity *rt_se, struct rt_rq *rt_rq) in inc_rt_migration()
395 static void dec_rt_migration(struct sched_rt_entity *rt_se, struct rt_rq *rt_rq) in dec_rt_migration()
472 void inc_rt_migration(struct sched_rt_entity *rt_se, struct rt_rq *rt_rq) in inc_rt_migration()
477 void dec_rt_migration(struct sched_rt_entity *rt_se, struct rt_rq *rt_rq) in dec_rt_migration()
535 static inline u64 sched_rt_runtime(struct rt_rq *rt_rq) in sched_rt_runtime()
543 static inline u64 sched_rt_period(struct rt_rq *rt_rq) in sched_rt_period()
563 #define for_each_rt_rq(rt_rq, iter, rq) \ argument
579 static void sched_rt_rq_enqueue(struct rt_rq *rt_rq) in sched_rt_rq_enqueue()
600 static void sched_rt_rq_dequeue(struct rt_rq *rt_rq) in sched_rt_rq_dequeue()
616 static inline int rt_rq_throttled(struct rt_rq *rt_rq) in rt_rq_throttled()
623 struct rt_rq *rt_rq = group_rt_rq(rt_se); in rt_se_boosted() local
651 static inline struct rt_bandwidth *sched_rt_bandwidth(struct rt_rq *rt_rq) in sched_rt_bandwidth()
658 static inline u64 sched_rt_runtime(struct rt_rq *rt_rq) in sched_rt_runtime()
663 static inline u64 sched_rt_period(struct rt_rq *rt_rq) in sched_rt_period()
670 #define for_each_rt_rq(rt_rq, iter, rq) \ argument
681 static inline void sched_rt_rq_enqueue(struct rt_rq *rt_rq) in sched_rt_rq_enqueue()
692 static inline void sched_rt_rq_dequeue(struct rt_rq *rt_rq) in sched_rt_rq_dequeue()
697 static inline int rt_rq_throttled(struct rt_rq *rt_rq) in rt_rq_throttled()
713 static inline struct rt_bandwidth *sched_rt_bandwidth(struct rt_rq *rt_rq) in sched_rt_bandwidth()
720 bool sched_rt_bandwidth_account(struct rt_rq *rt_rq) in sched_rt_bandwidth_account()
732 static void do_balance_runtime(struct rt_rq *rt_rq) in do_balance_runtime()
788 struct rt_rq *rt_rq; in __disable_runtime() local
869 struct rt_rq *rt_rq; in __enable_runtime() local
890 static void balance_runtime(struct rt_rq *rt_rq) in balance_runtime()
902 static inline void balance_runtime(struct rt_rq *rt_rq) {} in balance_runtime()
926 struct rt_rq *rt_rq = sched_rt_period_rt_rq(rt_b, i); in do_sched_rt_period_timer() local
993 struct rt_rq *rt_rq = group_rt_rq(rt_se); in rt_se_prio() local
1002 static int sched_rt_runtime_exceeded(struct rt_rq *rt_rq) in sched_rt_runtime_exceeded()
1075 struct rt_rq *rt_rq = rt_rq_of_se(rt_se); in update_curr_rt() local
1092 dequeue_top_rt_rq(struct rt_rq *rt_rq, unsigned int count) in dequeue_top_rt_rq()
1109 enqueue_top_rt_rq(struct rt_rq *rt_rq) in enqueue_top_rt_rq()
1133 inc_rt_prio_smp(struct rt_rq *rt_rq, int prio, int prev_prio) in inc_rt_prio_smp()
1149 dec_rt_prio_smp(struct rt_rq *rt_rq, int prio, int prev_prio) in dec_rt_prio_smp()
1167 void inc_rt_prio_smp(struct rt_rq *rt_rq, int prio, int prev_prio) {} in inc_rt_prio_smp()
1169 void dec_rt_prio_smp(struct rt_rq *rt_rq, int prio, int prev_prio) {} in dec_rt_prio_smp()
1175 inc_rt_prio(struct rt_rq *rt_rq, int prio) in inc_rt_prio()
1186 dec_rt_prio(struct rt_rq *rt_rq, int prio) in dec_rt_prio()
1214 static inline void inc_rt_prio(struct rt_rq *rt_rq, int prio) {} in inc_rt_prio()
1215 static inline void dec_rt_prio(struct rt_rq *rt_rq, int prio) {} in dec_rt_prio()
1222 inc_rt_group(struct sched_rt_entity *rt_se, struct rt_rq *rt_rq) in inc_rt_group()
1232 dec_rt_group(struct sched_rt_entity *rt_se, struct rt_rq *rt_rq) in dec_rt_group()
1243 inc_rt_group(struct sched_rt_entity *rt_se, struct rt_rq *rt_rq) in inc_rt_group()
1249 void dec_rt_group(struct sched_rt_entity *rt_se, struct rt_rq *rt_rq) {} in dec_rt_group()
1279 void inc_rt_tasks(struct sched_rt_entity *rt_se, struct rt_rq *rt_rq) in inc_rt_tasks()
1293 void dec_rt_tasks(struct sched_rt_entity *rt_se, struct rt_rq *rt_rq) in dec_rt_tasks()
1341 update_stats_wait_start_rt(struct rt_rq *rt_rq, struct sched_rt_entity *rt_se) in update_stats_wait_start_rt()
1360 update_stats_enqueue_sleeper_rt(struct rt_rq *rt_rq, struct sched_rt_entity *rt_se) in update_stats_enqueue_sleeper_rt()
1379 update_stats_enqueue_rt(struct rt_rq *rt_rq, struct sched_rt_entity *rt_se, in update_stats_enqueue_rt()
1390 update_stats_wait_end_rt(struct rt_rq *rt_rq, struct sched_rt_entity *rt_se) in update_stats_wait_end_rt()
1409 update_stats_dequeue_rt(struct rt_rq *rt_rq, struct sched_rt_entity *rt_se, in update_stats_dequeue_rt()
1436 struct rt_rq *rt_rq = rt_rq_of_se(rt_se); in __enqueue_rt_entity() local
1470 struct rt_rq *rt_rq = rt_rq_of_se(rt_se); in __dequeue_rt_entity() local
1527 struct rt_rq *rt_rq = group_rt_rq(rt_se); in dequeue_rt_entity() local
1570 requeue_rt_entity(struct rt_rq *rt_rq, struct sched_rt_entity *rt_se, int head) in requeue_rt_entity()
1586 struct rt_rq *rt_rq; in requeue_task_rt() local
1750 struct rt_rq *rt_rq = &rq->rt; in set_next_task_rt() local
1773 static struct sched_rt_entity *pick_next_rt_entity(struct rt_rq *rt_rq) in pick_next_rt_entity()
1794 struct rt_rq *rt_rq = &rq->rt; in _pick_next_task_rt() local
1831 struct rt_rq *rt_rq = &rq->rt; in put_prev_task_rt() local
2695 struct rt_rq *rt_rq; in task_is_throttled_rt() local
2887 struct rt_rq *rt_rq = tg->rt_rq[i]; in tg_set_rt_bandwidth() local
2980 struct rt_rq *rt_rq = &cpu_rq(i)->rt; in sched_rt_global_constraints() local
3084 struct rt_rq *rt_rq; in print_rt_stats() local