sched.c (ee2708aedad00544d38dba6df88efeb4a330bd66) | sched.c (ed1ad86b8527f8f864df3c182adbfcd12a445de6) |
---|---|
1// SPDX-License-Identifier: GPL-2.0 2/* Multipath TCP 3 * 4 * Copyright (c) 2022, SUSE. 5 */ 6 7#define pr_fmt(fmt) "MPTCP: " fmt 8 9#include <linux/kernel.h> 10#include <linux/module.h> 11#include <linux/list.h> 12#include <linux/rculist.h> 13#include <linux/spinlock.h> 14#include "protocol.h" 15 16static DEFINE_SPINLOCK(mptcp_sched_list_lock); 17static LIST_HEAD(mptcp_sched_list); 18 | 1// SPDX-License-Identifier: GPL-2.0 2/* Multipath TCP 3 * 4 * Copyright (c) 2022, SUSE. 5 */ 6 7#define pr_fmt(fmt) "MPTCP: " fmt 8 9#include <linux/kernel.h> 10#include <linux/module.h> 11#include <linux/list.h> 12#include <linux/rculist.h> 13#include <linux/spinlock.h> 14#include "protocol.h" 15 16static DEFINE_SPINLOCK(mptcp_sched_list_lock); 17static LIST_HEAD(mptcp_sched_list); 18 |
19static int mptcp_sched_default_get_subflow(struct mptcp_sock *msk, 20 struct mptcp_sched_data *data) 21{ 22 struct sock *ssk; 23 24 ssk = data->reinject ? mptcp_subflow_get_retrans(msk) : 25 mptcp_subflow_get_send(msk); 26 if (!ssk) 27 return -EINVAL; 28 29 mptcp_subflow_set_scheduled(mptcp_subflow_ctx(ssk), true); 30 return 0; 31} 32 33static struct mptcp_sched_ops mptcp_sched_default = { 34 .get_subflow = mptcp_sched_default_get_subflow, 35 .name = "default", 36 .owner = THIS_MODULE, 37}; 38 |
|
19/* Must be called with rcu read lock held */ 20struct mptcp_sched_ops *mptcp_sched_find(const char *name) 21{ 22 struct mptcp_sched_ops *sched, *ret = NULL; 23 24 list_for_each_entry_rcu(sched, &mptcp_sched_list, list) { 25 if (!strcmp(sched->name, name)) { 26 ret = sched; --- 18 unchanged lines hidden (view full) --- 45 spin_unlock(&mptcp_sched_list_lock); 46 47 pr_debug("%s registered", sched->name); 48 return 0; 49} 50 51void mptcp_unregister_scheduler(struct mptcp_sched_ops *sched) 52{ | 39/* Must be called with rcu read lock held */ 40struct mptcp_sched_ops *mptcp_sched_find(const char *name) 41{ 42 struct mptcp_sched_ops *sched, *ret = NULL; 43 44 list_for_each_entry_rcu(sched, &mptcp_sched_list, list) { 45 if (!strcmp(sched->name, name)) { 46 ret = sched; --- 18 unchanged lines hidden (view full) --- 65 spin_unlock(&mptcp_sched_list_lock); 66 67 pr_debug("%s registered", sched->name); 68 return 0; 69} 70 71void mptcp_unregister_scheduler(struct mptcp_sched_ops *sched) 72{ |
73 if (sched == &mptcp_sched_default) 74 return; 75 |
|
53 spin_lock(&mptcp_sched_list_lock); 54 list_del_rcu(&sched->list); 55 spin_unlock(&mptcp_sched_list_lock); 56} 57 | 76 spin_lock(&mptcp_sched_list_lock); 77 list_del_rcu(&sched->list); 78 spin_unlock(&mptcp_sched_list_lock); 79} 80 |
81void mptcp_sched_init(void) 82{ 83 mptcp_register_scheduler(&mptcp_sched_default); 84} 85 |
|
58int mptcp_init_sched(struct mptcp_sock *msk, 59 struct mptcp_sched_ops *sched) 60{ 61 if (!sched) | 86int mptcp_init_sched(struct mptcp_sock *msk, 87 struct mptcp_sched_ops *sched) 88{ 89 if (!sched) |
62 goto out; | 90 sched = &mptcp_sched_default; |
63 64 if (!bpf_try_module_get(sched, sched->owner)) 65 return -EBUSY; 66 67 msk->sched = sched; 68 if (msk->sched->init) 69 msk->sched->init(msk); 70 71 pr_debug("sched=%s", msk->sched->name); 72 | 91 92 if (!bpf_try_module_get(sched, sched->owner)) 93 return -EBUSY; 94 95 msk->sched = sched; 96 if (msk->sched->init) 97 msk->sched->init(msk); 98 99 pr_debug("sched=%s", msk->sched->name); 100 |
73out: | |
74 return 0; 75} 76 77void mptcp_release_sched(struct mptcp_sock *msk) 78{ 79 struct mptcp_sched_ops *sched = msk->sched; 80 81 if (!sched) --- 30 unchanged lines hidden (view full) --- 112 return -EINVAL; 113 } 114 115 mptcp_for_each_subflow(msk, subflow) { 116 if (READ_ONCE(subflow->scheduled)) 117 return 0; 118 } 119 | 101 return 0; 102} 103 104void mptcp_release_sched(struct mptcp_sock *msk) 105{ 106 struct mptcp_sched_ops *sched = msk->sched; 107 108 if (!sched) --- 30 unchanged lines hidden (view full) --- 139 return -EINVAL; 140 } 141 142 mptcp_for_each_subflow(msk, subflow) { 143 if (READ_ONCE(subflow->scheduled)) 144 return 0; 145 } 146 |
120 if (!msk->sched) { 121 struct sock *ssk; 122 123 ssk = mptcp_subflow_get_send(msk); 124 if (!ssk) 125 return -EINVAL; 126 mptcp_subflow_set_scheduled(mptcp_subflow_ctx(ssk), true); 127 return 0; 128 } 129 | |
130 data.reinject = false; | 147 data.reinject = false; |
148 if (msk->sched == &mptcp_sched_default || !msk->sched) 149 return mptcp_sched_default_get_subflow(msk, &data); |
|
131 return msk->sched->get_subflow(msk, &data); 132} 133 134int mptcp_sched_get_retrans(struct mptcp_sock *msk) 135{ 136 struct mptcp_subflow_context *subflow; 137 struct mptcp_sched_data data; 138 139 msk_owned_by_me(msk); 140 141 /* the following check is moved out of mptcp_subflow_get_retrans */ 142 if (__mptcp_check_fallback(msk)) 143 return -EINVAL; 144 145 mptcp_for_each_subflow(msk, subflow) { 146 if (READ_ONCE(subflow->scheduled)) 147 return 0; 148 } 149 | 150 return msk->sched->get_subflow(msk, &data); 151} 152 153int mptcp_sched_get_retrans(struct mptcp_sock *msk) 154{ 155 struct mptcp_subflow_context *subflow; 156 struct mptcp_sched_data data; 157 158 msk_owned_by_me(msk); 159 160 /* the following check is moved out of mptcp_subflow_get_retrans */ 161 if (__mptcp_check_fallback(msk)) 162 return -EINVAL; 163 164 mptcp_for_each_subflow(msk, subflow) { 165 if (READ_ONCE(subflow->scheduled)) 166 return 0; 167 } 168 |
150 if (!msk->sched) { 151 struct sock *ssk; 152 153 ssk = mptcp_subflow_get_retrans(msk); 154 if (!ssk) 155 return -EINVAL; 156 mptcp_subflow_set_scheduled(mptcp_subflow_ctx(ssk), true); 157 return 0; 158 } 159 | |
160 data.reinject = true; | 169 data.reinject = true; |
170 if (msk->sched == &mptcp_sched_default || !msk->sched) 171 return mptcp_sched_default_get_subflow(msk, &data); |
|
161 return msk->sched->get_subflow(msk, &data); 162} | 172 return msk->sched->get_subflow(msk, &data); 173} |