1 /* 2 * QEMU block throttling group infrastructure 3 * 4 * Copyright (C) Nodalink, EURL. 2014 5 * Copyright (C) Igalia, S.L. 2015 6 * 7 * Authors: 8 * Benoît Canet <benoit.canet@nodalink.com> 9 * Alberto Garcia <berto@igalia.com> 10 * 11 * This program is free software; you can redistribute it and/or 12 * modify it under the terms of the GNU General Public License as 13 * published by the Free Software Foundation; either version 2 or 14 * (at your option) version 3 of the License. 15 * 16 * This program is distributed in the hope that it will be useful, 17 * but WITHOUT ANY WARRANTY; without even the implied warranty of 18 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the 19 * GNU General Public License for more details. 20 * 21 * You should have received a copy of the GNU General Public License 22 * along with this program; if not, see <http://www.gnu.org/licenses/>. 23 */ 24 25 #include "qemu/osdep.h" 26 #include "sysemu/block-backend.h" 27 #include "block/throttle-groups.h" 28 #include "qemu/throttle-options.h" 29 #include "qemu/queue.h" 30 #include "qemu/thread.h" 31 #include "sysemu/qtest.h" 32 #include "qapi/error.h" 33 #include "qapi/qapi-visit-block-core.h" 34 #include "qom/object.h" 35 #include "qom/object_interfaces.h" 36 37 static void throttle_group_obj_init(Object *obj); 38 static void throttle_group_obj_complete(UserCreatable *obj, Error **errp); 39 static void timer_cb(ThrottleGroupMember *tgm, bool is_write); 40 41 /* The ThrottleGroup structure (with its ThrottleState) is shared 42 * among different ThrottleGroupMembers and it's independent from 43 * AioContext, so in order to use it from different threads it needs 44 * its own locking. 45 * 46 * This locking is however handled internally in this file, so it's 47 * transparent to outside users. 48 * 49 * The whole ThrottleGroup structure is private and invisible to 50 * outside users, that only use it through its ThrottleState. 51 * 52 * In addition to the ThrottleGroup structure, ThrottleGroupMember has 53 * fields that need to be accessed by other members of the group and 54 * therefore also need to be protected by this lock. Once a 55 * ThrottleGroupMember is registered in a group those fields can be accessed 56 * by other threads any time. 57 * 58 * Again, all this is handled internally and is mostly transparent to 59 * the outside. The 'throttle_timers' field however has an additional 60 * constraint because it may be temporarily invalid (see for example 61 * blk_set_aio_context()). Therefore in this file a thread will 62 * access some other ThrottleGroupMember's timers only after verifying that 63 * that ThrottleGroupMember has throttled requests in the queue. 64 */ 65 typedef struct ThrottleGroup { 66 Object parent_obj; 67 68 /* refuse individual property change if initialization is complete */ 69 bool is_initialized; 70 char *name; /* This is constant during the lifetime of the group */ 71 72 QemuMutex lock; /* This lock protects the following four fields */ 73 ThrottleState ts; 74 QLIST_HEAD(, ThrottleGroupMember) head; 75 ThrottleGroupMember *tokens[2]; 76 bool any_timer_armed[2]; 77 QEMUClockType clock_type; 78 79 /* This field is protected by the global QEMU mutex */ 80 QTAILQ_ENTRY(ThrottleGroup) list; 81 } ThrottleGroup; 82 83 /* This is protected by the global QEMU mutex */ 84 static QTAILQ_HEAD(, ThrottleGroup) throttle_groups = 85 QTAILQ_HEAD_INITIALIZER(throttle_groups); 86 87 88 /* This function reads throttle_groups and must be called under the global 89 * mutex. 90 */ 91 static ThrottleGroup *throttle_group_by_name(const char *name) 92 { 93 ThrottleGroup *iter; 94 95 /* Look for an existing group with that name */ 96 QTAILQ_FOREACH(iter, &throttle_groups, list) { 97 if (!g_strcmp0(name, iter->name)) { 98 return iter; 99 } 100 } 101 102 return NULL; 103 } 104 105 /* This function reads throttle_groups and must be called under the global 106 * mutex. 107 */ 108 bool throttle_group_exists(const char *name) 109 { 110 return throttle_group_by_name(name) != NULL; 111 } 112 113 /* Increments the reference count of a ThrottleGroup given its name. 114 * 115 * If no ThrottleGroup is found with the given name a new one is 116 * created. 117 * 118 * This function edits throttle_groups and must be called under the global 119 * mutex. 120 * 121 * @name: the name of the ThrottleGroup 122 * @ret: the ThrottleState member of the ThrottleGroup 123 */ 124 ThrottleState *throttle_group_incref(const char *name) 125 { 126 ThrottleGroup *tg = NULL; 127 128 /* Look for an existing group with that name */ 129 tg = throttle_group_by_name(name); 130 131 if (tg) { 132 object_ref(OBJECT(tg)); 133 } else { 134 /* Create a new one if not found */ 135 /* new ThrottleGroup obj will have a refcnt = 1 */ 136 tg = THROTTLE_GROUP(object_new(TYPE_THROTTLE_GROUP)); 137 tg->name = g_strdup(name); 138 throttle_group_obj_complete(USER_CREATABLE(tg), &error_abort); 139 } 140 141 return &tg->ts; 142 } 143 144 /* Decrease the reference count of a ThrottleGroup. 145 * 146 * When the reference count reaches zero the ThrottleGroup is 147 * destroyed. 148 * 149 * This function edits throttle_groups and must be called under the global 150 * mutex. 151 * 152 * @ts: The ThrottleGroup to unref, given by its ThrottleState member 153 */ 154 void throttle_group_unref(ThrottleState *ts) 155 { 156 ThrottleGroup *tg = container_of(ts, ThrottleGroup, ts); 157 object_unref(OBJECT(tg)); 158 } 159 160 /* Get the name from a ThrottleGroupMember's group. The name (and the pointer) 161 * is guaranteed to remain constant during the lifetime of the group. 162 * 163 * @tgm: a ThrottleGroupMember 164 * @ret: the name of the group. 165 */ 166 const char *throttle_group_get_name(ThrottleGroupMember *tgm) 167 { 168 ThrottleGroup *tg = container_of(tgm->throttle_state, ThrottleGroup, ts); 169 return tg->name; 170 } 171 172 /* Return the next ThrottleGroupMember in the round-robin sequence, simulating 173 * a circular list. 174 * 175 * This assumes that tg->lock is held. 176 * 177 * @tgm: the current ThrottleGroupMember 178 * @ret: the next ThrottleGroupMember in the sequence 179 */ 180 static ThrottleGroupMember *throttle_group_next_tgm(ThrottleGroupMember *tgm) 181 { 182 ThrottleState *ts = tgm->throttle_state; 183 ThrottleGroup *tg = container_of(ts, ThrottleGroup, ts); 184 ThrottleGroupMember *next = QLIST_NEXT(tgm, round_robin); 185 186 if (!next) { 187 next = QLIST_FIRST(&tg->head); 188 } 189 190 return next; 191 } 192 193 /* 194 * Return whether a ThrottleGroupMember has pending requests. 195 * 196 * This assumes that tg->lock is held. 197 * 198 * @tgm: the ThrottleGroupMember 199 * @is_write: the type of operation (read/write) 200 * @ret: whether the ThrottleGroupMember has pending requests. 201 */ 202 static inline bool tgm_has_pending_reqs(ThrottleGroupMember *tgm, 203 bool is_write) 204 { 205 return tgm->pending_reqs[is_write]; 206 } 207 208 /* Return the next ThrottleGroupMember in the round-robin sequence with pending 209 * I/O requests. 210 * 211 * This assumes that tg->lock is held. 212 * 213 * @tgm: the current ThrottleGroupMember 214 * @is_write: the type of operation (read/write) 215 * @ret: the next ThrottleGroupMember with pending requests, or tgm if 216 * there is none. 217 */ 218 static ThrottleGroupMember *next_throttle_token(ThrottleGroupMember *tgm, 219 bool is_write) 220 { 221 ThrottleState *ts = tgm->throttle_state; 222 ThrottleGroup *tg = container_of(ts, ThrottleGroup, ts); 223 ThrottleGroupMember *token, *start; 224 225 /* If this member has its I/O limits disabled then it means that 226 * it's being drained. Skip the round-robin search and return tgm 227 * immediately if it has pending requests. Otherwise we could be 228 * forcing it to wait for other member's throttled requests. */ 229 if (tgm_has_pending_reqs(tgm, is_write) && 230 atomic_read(&tgm->io_limits_disabled)) { 231 return tgm; 232 } 233 234 start = token = tg->tokens[is_write]; 235 236 /* get next bs round in round robin style */ 237 token = throttle_group_next_tgm(token); 238 while (token != start && !tgm_has_pending_reqs(token, is_write)) { 239 token = throttle_group_next_tgm(token); 240 } 241 242 /* If no IO are queued for scheduling on the next round robin token 243 * then decide the token is the current tgm because chances are 244 * the current tgm got the current request queued. 245 */ 246 if (token == start && !tgm_has_pending_reqs(token, is_write)) { 247 token = tgm; 248 } 249 250 /* Either we return the original TGM, or one with pending requests */ 251 assert(token == tgm || tgm_has_pending_reqs(token, is_write)); 252 253 return token; 254 } 255 256 /* Check if the next I/O request for a ThrottleGroupMember needs to be 257 * throttled or not. If there's no timer set in this group, set one and update 258 * the token accordingly. 259 * 260 * This assumes that tg->lock is held. 261 * 262 * @tgm: the current ThrottleGroupMember 263 * @is_write: the type of operation (read/write) 264 * @ret: whether the I/O request needs to be throttled or not 265 */ 266 static bool throttle_group_schedule_timer(ThrottleGroupMember *tgm, 267 bool is_write) 268 { 269 ThrottleState *ts = tgm->throttle_state; 270 ThrottleGroup *tg = container_of(ts, ThrottleGroup, ts); 271 ThrottleTimers *tt = &tgm->throttle_timers; 272 bool must_wait; 273 274 if (atomic_read(&tgm->io_limits_disabled)) { 275 return false; 276 } 277 278 /* Check if any of the timers in this group is already armed */ 279 if (tg->any_timer_armed[is_write]) { 280 return true; 281 } 282 283 must_wait = throttle_schedule_timer(ts, tt, is_write); 284 285 /* If a timer just got armed, set tgm as the current token */ 286 if (must_wait) { 287 tg->tokens[is_write] = tgm; 288 tg->any_timer_armed[is_write] = true; 289 } 290 291 return must_wait; 292 } 293 294 /* Start the next pending I/O request for a ThrottleGroupMember. Return whether 295 * any request was actually pending. 296 * 297 * @tgm: the current ThrottleGroupMember 298 * @is_write: the type of operation (read/write) 299 */ 300 static bool coroutine_fn throttle_group_co_restart_queue(ThrottleGroupMember *tgm, 301 bool is_write) 302 { 303 bool ret; 304 305 qemu_co_mutex_lock(&tgm->throttled_reqs_lock); 306 ret = qemu_co_queue_next(&tgm->throttled_reqs[is_write]); 307 qemu_co_mutex_unlock(&tgm->throttled_reqs_lock); 308 309 return ret; 310 } 311 312 /* Look for the next pending I/O request and schedule it. 313 * 314 * This assumes that tg->lock is held. 315 * 316 * @tgm: the current ThrottleGroupMember 317 * @is_write: the type of operation (read/write) 318 */ 319 static void schedule_next_request(ThrottleGroupMember *tgm, bool is_write) 320 { 321 ThrottleState *ts = tgm->throttle_state; 322 ThrottleGroup *tg = container_of(ts, ThrottleGroup, ts); 323 bool must_wait; 324 ThrottleGroupMember *token; 325 326 /* Check if there's any pending request to schedule next */ 327 token = next_throttle_token(tgm, is_write); 328 if (!tgm_has_pending_reqs(token, is_write)) { 329 return; 330 } 331 332 /* Set a timer for the request if it needs to be throttled */ 333 must_wait = throttle_group_schedule_timer(token, is_write); 334 335 /* If it doesn't have to wait, queue it for immediate execution */ 336 if (!must_wait) { 337 /* Give preference to requests from the current tgm */ 338 if (qemu_in_coroutine() && 339 throttle_group_co_restart_queue(tgm, is_write)) { 340 token = tgm; 341 } else { 342 ThrottleTimers *tt = &token->throttle_timers; 343 int64_t now = qemu_clock_get_ns(tg->clock_type); 344 timer_mod(tt->timers[is_write], now); 345 tg->any_timer_armed[is_write] = true; 346 } 347 tg->tokens[is_write] = token; 348 } 349 } 350 351 /* Check if an I/O request needs to be throttled, wait and set a timer 352 * if necessary, and schedule the next request using a round robin 353 * algorithm. 354 * 355 * @tgm: the current ThrottleGroupMember 356 * @bytes: the number of bytes for this I/O 357 * @is_write: the type of operation (read/write) 358 */ 359 void coroutine_fn throttle_group_co_io_limits_intercept(ThrottleGroupMember *tgm, 360 unsigned int bytes, 361 bool is_write) 362 { 363 bool must_wait; 364 ThrottleGroupMember *token; 365 ThrottleGroup *tg = container_of(tgm->throttle_state, ThrottleGroup, ts); 366 qemu_mutex_lock(&tg->lock); 367 368 /* First we check if this I/O has to be throttled. */ 369 token = next_throttle_token(tgm, is_write); 370 must_wait = throttle_group_schedule_timer(token, is_write); 371 372 /* Wait if there's a timer set or queued requests of this type */ 373 if (must_wait || tgm->pending_reqs[is_write]) { 374 tgm->pending_reqs[is_write]++; 375 qemu_mutex_unlock(&tg->lock); 376 qemu_co_mutex_lock(&tgm->throttled_reqs_lock); 377 qemu_co_queue_wait(&tgm->throttled_reqs[is_write], 378 &tgm->throttled_reqs_lock); 379 qemu_co_mutex_unlock(&tgm->throttled_reqs_lock); 380 qemu_mutex_lock(&tg->lock); 381 tgm->pending_reqs[is_write]--; 382 } 383 384 /* The I/O will be executed, so do the accounting */ 385 throttle_account(tgm->throttle_state, is_write, bytes); 386 387 /* Schedule the next request */ 388 schedule_next_request(tgm, is_write); 389 390 qemu_mutex_unlock(&tg->lock); 391 } 392 393 typedef struct { 394 ThrottleGroupMember *tgm; 395 bool is_write; 396 } RestartData; 397 398 static void coroutine_fn throttle_group_restart_queue_entry(void *opaque) 399 { 400 RestartData *data = opaque; 401 ThrottleGroupMember *tgm = data->tgm; 402 ThrottleState *ts = tgm->throttle_state; 403 ThrottleGroup *tg = container_of(ts, ThrottleGroup, ts); 404 bool is_write = data->is_write; 405 bool empty_queue; 406 407 empty_queue = !throttle_group_co_restart_queue(tgm, is_write); 408 409 /* If the request queue was empty then we have to take care of 410 * scheduling the next one */ 411 if (empty_queue) { 412 qemu_mutex_lock(&tg->lock); 413 schedule_next_request(tgm, is_write); 414 qemu_mutex_unlock(&tg->lock); 415 } 416 417 g_free(data); 418 419 atomic_dec(&tgm->restart_pending); 420 aio_wait_kick(); 421 } 422 423 static void throttle_group_restart_queue(ThrottleGroupMember *tgm, bool is_write) 424 { 425 Coroutine *co; 426 RestartData *rd = g_new0(RestartData, 1); 427 428 rd->tgm = tgm; 429 rd->is_write = is_write; 430 431 /* This function is called when a timer is fired or when 432 * throttle_group_restart_tgm() is called. Either way, there can 433 * be no timer pending on this tgm at this point */ 434 assert(!timer_pending(tgm->throttle_timers.timers[is_write])); 435 436 atomic_inc(&tgm->restart_pending); 437 438 co = qemu_coroutine_create(throttle_group_restart_queue_entry, rd); 439 aio_co_enter(tgm->aio_context, co); 440 } 441 442 void throttle_group_restart_tgm(ThrottleGroupMember *tgm) 443 { 444 int i; 445 446 if (tgm->throttle_state) { 447 for (i = 0; i < 2; i++) { 448 QEMUTimer *t = tgm->throttle_timers.timers[i]; 449 if (timer_pending(t)) { 450 /* If there's a pending timer on this tgm, fire it now */ 451 timer_del(t); 452 timer_cb(tgm, i); 453 } else { 454 /* Else run the next request from the queue manually */ 455 throttle_group_restart_queue(tgm, i); 456 } 457 } 458 } 459 } 460 461 /* Update the throttle configuration for a particular group. Similar 462 * to throttle_config(), but guarantees atomicity within the 463 * throttling group. 464 * 465 * @tgm: a ThrottleGroupMember that is a member of the group 466 * @cfg: the configuration to set 467 */ 468 void throttle_group_config(ThrottleGroupMember *tgm, ThrottleConfig *cfg) 469 { 470 ThrottleState *ts = tgm->throttle_state; 471 ThrottleGroup *tg = container_of(ts, ThrottleGroup, ts); 472 qemu_mutex_lock(&tg->lock); 473 throttle_config(ts, tg->clock_type, cfg); 474 qemu_mutex_unlock(&tg->lock); 475 476 throttle_group_restart_tgm(tgm); 477 } 478 479 /* Get the throttle configuration from a particular group. Similar to 480 * throttle_get_config(), but guarantees atomicity within the 481 * throttling group. 482 * 483 * @tgm: a ThrottleGroupMember that is a member of the group 484 * @cfg: the configuration will be written here 485 */ 486 void throttle_group_get_config(ThrottleGroupMember *tgm, ThrottleConfig *cfg) 487 { 488 ThrottleState *ts = tgm->throttle_state; 489 ThrottleGroup *tg = container_of(ts, ThrottleGroup, ts); 490 qemu_mutex_lock(&tg->lock); 491 throttle_get_config(ts, cfg); 492 qemu_mutex_unlock(&tg->lock); 493 } 494 495 /* ThrottleTimers callback. This wakes up a request that was waiting 496 * because it had been throttled. 497 * 498 * @tgm: the ThrottleGroupMember whose request had been throttled 499 * @is_write: the type of operation (read/write) 500 */ 501 static void timer_cb(ThrottleGroupMember *tgm, bool is_write) 502 { 503 ThrottleState *ts = tgm->throttle_state; 504 ThrottleGroup *tg = container_of(ts, ThrottleGroup, ts); 505 506 /* The timer has just been fired, so we can update the flag */ 507 qemu_mutex_lock(&tg->lock); 508 tg->any_timer_armed[is_write] = false; 509 qemu_mutex_unlock(&tg->lock); 510 511 /* Run the request that was waiting for this timer */ 512 throttle_group_restart_queue(tgm, is_write); 513 } 514 515 static void read_timer_cb(void *opaque) 516 { 517 timer_cb(opaque, false); 518 } 519 520 static void write_timer_cb(void *opaque) 521 { 522 timer_cb(opaque, true); 523 } 524 525 /* Register a ThrottleGroupMember from the throttling group, also initializing 526 * its timers and updating its throttle_state pointer to point to it. If a 527 * throttling group with that name does not exist yet, it will be created. 528 * 529 * This function edits throttle_groups and must be called under the global 530 * mutex. 531 * 532 * @tgm: the ThrottleGroupMember to insert 533 * @groupname: the name of the group 534 * @ctx: the AioContext to use 535 */ 536 void throttle_group_register_tgm(ThrottleGroupMember *tgm, 537 const char *groupname, 538 AioContext *ctx) 539 { 540 int i; 541 ThrottleState *ts = throttle_group_incref(groupname); 542 ThrottleGroup *tg = container_of(ts, ThrottleGroup, ts); 543 544 tgm->throttle_state = ts; 545 tgm->aio_context = ctx; 546 atomic_set(&tgm->restart_pending, 0); 547 548 qemu_mutex_lock(&tg->lock); 549 /* If the ThrottleGroup is new set this ThrottleGroupMember as the token */ 550 for (i = 0; i < 2; i++) { 551 if (!tg->tokens[i]) { 552 tg->tokens[i] = tgm; 553 } 554 } 555 556 QLIST_INSERT_HEAD(&tg->head, tgm, round_robin); 557 558 throttle_timers_init(&tgm->throttle_timers, 559 tgm->aio_context, 560 tg->clock_type, 561 read_timer_cb, 562 write_timer_cb, 563 tgm); 564 qemu_co_mutex_init(&tgm->throttled_reqs_lock); 565 qemu_co_queue_init(&tgm->throttled_reqs[0]); 566 qemu_co_queue_init(&tgm->throttled_reqs[1]); 567 568 qemu_mutex_unlock(&tg->lock); 569 } 570 571 /* Unregister a ThrottleGroupMember from its group, removing it from the list, 572 * destroying the timers and setting the throttle_state pointer to NULL. 573 * 574 * The ThrottleGroupMember must not have pending throttled requests, so the 575 * caller has to drain them first. 576 * 577 * The group will be destroyed if it's empty after this operation. 578 * 579 * @tgm the ThrottleGroupMember to remove 580 */ 581 void throttle_group_unregister_tgm(ThrottleGroupMember *tgm) 582 { 583 ThrottleState *ts = tgm->throttle_state; 584 ThrottleGroup *tg = container_of(ts, ThrottleGroup, ts); 585 ThrottleGroupMember *token; 586 int i; 587 588 if (!ts) { 589 /* Discard already unregistered tgm */ 590 return; 591 } 592 593 /* Wait for throttle_group_restart_queue_entry() coroutines to finish */ 594 AIO_WAIT_WHILE(tgm->aio_context, atomic_read(&tgm->restart_pending) > 0); 595 596 qemu_mutex_lock(&tg->lock); 597 for (i = 0; i < 2; i++) { 598 assert(tgm->pending_reqs[i] == 0); 599 assert(qemu_co_queue_empty(&tgm->throttled_reqs[i])); 600 assert(!timer_pending(tgm->throttle_timers.timers[i])); 601 if (tg->tokens[i] == tgm) { 602 token = throttle_group_next_tgm(tgm); 603 /* Take care of the case where this is the last tgm in the group */ 604 if (token == tgm) { 605 token = NULL; 606 } 607 tg->tokens[i] = token; 608 } 609 } 610 611 /* remove the current tgm from the list */ 612 QLIST_REMOVE(tgm, round_robin); 613 throttle_timers_destroy(&tgm->throttle_timers); 614 qemu_mutex_unlock(&tg->lock); 615 616 throttle_group_unref(&tg->ts); 617 tgm->throttle_state = NULL; 618 } 619 620 void throttle_group_attach_aio_context(ThrottleGroupMember *tgm, 621 AioContext *new_context) 622 { 623 ThrottleTimers *tt = &tgm->throttle_timers; 624 throttle_timers_attach_aio_context(tt, new_context); 625 tgm->aio_context = new_context; 626 } 627 628 void throttle_group_detach_aio_context(ThrottleGroupMember *tgm) 629 { 630 ThrottleGroup *tg = container_of(tgm->throttle_state, ThrottleGroup, ts); 631 ThrottleTimers *tt = &tgm->throttle_timers; 632 int i; 633 634 /* Requests must have been drained */ 635 assert(tgm->pending_reqs[0] == 0 && tgm->pending_reqs[1] == 0); 636 assert(qemu_co_queue_empty(&tgm->throttled_reqs[0])); 637 assert(qemu_co_queue_empty(&tgm->throttled_reqs[1])); 638 639 /* Kick off next ThrottleGroupMember, if necessary */ 640 qemu_mutex_lock(&tg->lock); 641 for (i = 0; i < 2; i++) { 642 if (timer_pending(tt->timers[i])) { 643 tg->any_timer_armed[i] = false; 644 schedule_next_request(tgm, i); 645 } 646 } 647 qemu_mutex_unlock(&tg->lock); 648 649 throttle_timers_detach_aio_context(tt); 650 tgm->aio_context = NULL; 651 } 652 653 #undef THROTTLE_OPT_PREFIX 654 #define THROTTLE_OPT_PREFIX "x-" 655 656 /* Helper struct and array for QOM property setter/getter */ 657 typedef struct { 658 const char *name; 659 BucketType type; 660 enum { 661 AVG, 662 MAX, 663 BURST_LENGTH, 664 IOPS_SIZE, 665 } category; 666 } ThrottleParamInfo; 667 668 static ThrottleParamInfo properties[] = { 669 { 670 THROTTLE_OPT_PREFIX QEMU_OPT_IOPS_TOTAL, 671 THROTTLE_OPS_TOTAL, AVG, 672 }, 673 { 674 THROTTLE_OPT_PREFIX QEMU_OPT_IOPS_TOTAL_MAX, 675 THROTTLE_OPS_TOTAL, MAX, 676 }, 677 { 678 THROTTLE_OPT_PREFIX QEMU_OPT_IOPS_TOTAL_MAX_LENGTH, 679 THROTTLE_OPS_TOTAL, BURST_LENGTH, 680 }, 681 { 682 THROTTLE_OPT_PREFIX QEMU_OPT_IOPS_READ, 683 THROTTLE_OPS_READ, AVG, 684 }, 685 { 686 THROTTLE_OPT_PREFIX QEMU_OPT_IOPS_READ_MAX, 687 THROTTLE_OPS_READ, MAX, 688 }, 689 { 690 THROTTLE_OPT_PREFIX QEMU_OPT_IOPS_READ_MAX_LENGTH, 691 THROTTLE_OPS_READ, BURST_LENGTH, 692 }, 693 { 694 THROTTLE_OPT_PREFIX QEMU_OPT_IOPS_WRITE, 695 THROTTLE_OPS_WRITE, AVG, 696 }, 697 { 698 THROTTLE_OPT_PREFIX QEMU_OPT_IOPS_WRITE_MAX, 699 THROTTLE_OPS_WRITE, MAX, 700 }, 701 { 702 THROTTLE_OPT_PREFIX QEMU_OPT_IOPS_WRITE_MAX_LENGTH, 703 THROTTLE_OPS_WRITE, BURST_LENGTH, 704 }, 705 { 706 THROTTLE_OPT_PREFIX QEMU_OPT_BPS_TOTAL, 707 THROTTLE_BPS_TOTAL, AVG, 708 }, 709 { 710 THROTTLE_OPT_PREFIX QEMU_OPT_BPS_TOTAL_MAX, 711 THROTTLE_BPS_TOTAL, MAX, 712 }, 713 { 714 THROTTLE_OPT_PREFIX QEMU_OPT_BPS_TOTAL_MAX_LENGTH, 715 THROTTLE_BPS_TOTAL, BURST_LENGTH, 716 }, 717 { 718 THROTTLE_OPT_PREFIX QEMU_OPT_BPS_READ, 719 THROTTLE_BPS_READ, AVG, 720 }, 721 { 722 THROTTLE_OPT_PREFIX QEMU_OPT_BPS_READ_MAX, 723 THROTTLE_BPS_READ, MAX, 724 }, 725 { 726 THROTTLE_OPT_PREFIX QEMU_OPT_BPS_READ_MAX_LENGTH, 727 THROTTLE_BPS_READ, BURST_LENGTH, 728 }, 729 { 730 THROTTLE_OPT_PREFIX QEMU_OPT_BPS_WRITE, 731 THROTTLE_BPS_WRITE, AVG, 732 }, 733 { 734 THROTTLE_OPT_PREFIX QEMU_OPT_BPS_WRITE_MAX, 735 THROTTLE_BPS_WRITE, MAX, 736 }, 737 { 738 THROTTLE_OPT_PREFIX QEMU_OPT_BPS_WRITE_MAX_LENGTH, 739 THROTTLE_BPS_WRITE, BURST_LENGTH, 740 }, 741 { 742 THROTTLE_OPT_PREFIX QEMU_OPT_IOPS_SIZE, 743 0, IOPS_SIZE, 744 } 745 }; 746 747 /* This function edits throttle_groups and must be called under the global 748 * mutex */ 749 static void throttle_group_obj_init(Object *obj) 750 { 751 ThrottleGroup *tg = THROTTLE_GROUP(obj); 752 753 tg->clock_type = QEMU_CLOCK_REALTIME; 754 if (qtest_enabled()) { 755 /* For testing block IO throttling only */ 756 tg->clock_type = QEMU_CLOCK_VIRTUAL; 757 } 758 tg->is_initialized = false; 759 qemu_mutex_init(&tg->lock); 760 throttle_init(&tg->ts); 761 QLIST_INIT(&tg->head); 762 } 763 764 /* This function edits throttle_groups and must be called under the global 765 * mutex */ 766 static void throttle_group_obj_complete(UserCreatable *obj, Error **errp) 767 { 768 ThrottleGroup *tg = THROTTLE_GROUP(obj); 769 ThrottleConfig cfg; 770 771 /* set group name to object id if it exists */ 772 if (!tg->name && tg->parent_obj.parent) { 773 tg->name = object_get_canonical_path_component(OBJECT(obj)); 774 } 775 /* We must have a group name at this point */ 776 assert(tg->name); 777 778 /* error if name is duplicate */ 779 if (throttle_group_exists(tg->name)) { 780 error_setg(errp, "A group with this name already exists"); 781 return; 782 } 783 784 /* check validity */ 785 throttle_get_config(&tg->ts, &cfg); 786 if (!throttle_is_valid(&cfg, errp)) { 787 return; 788 } 789 throttle_config(&tg->ts, tg->clock_type, &cfg); 790 QTAILQ_INSERT_TAIL(&throttle_groups, tg, list); 791 tg->is_initialized = true; 792 } 793 794 /* This function edits throttle_groups and must be called under the global 795 * mutex */ 796 static void throttle_group_obj_finalize(Object *obj) 797 { 798 ThrottleGroup *tg = THROTTLE_GROUP(obj); 799 if (tg->is_initialized) { 800 QTAILQ_REMOVE(&throttle_groups, tg, list); 801 } 802 qemu_mutex_destroy(&tg->lock); 803 g_free(tg->name); 804 } 805 806 static void throttle_group_set(Object *obj, Visitor *v, const char * name, 807 void *opaque, Error **errp) 808 809 { 810 ThrottleGroup *tg = THROTTLE_GROUP(obj); 811 ThrottleConfig *cfg; 812 ThrottleParamInfo *info = opaque; 813 Error *local_err = NULL; 814 int64_t value; 815 816 /* If we have finished initialization, don't accept individual property 817 * changes through QOM. Throttle configuration limits must be set in one 818 * transaction, as certain combinations are invalid. 819 */ 820 if (tg->is_initialized) { 821 error_setg(&local_err, "Property cannot be set after initialization"); 822 goto ret; 823 } 824 825 visit_type_int64(v, name, &value, &local_err); 826 if (local_err) { 827 goto ret; 828 } 829 if (value < 0) { 830 error_setg(&local_err, "Property values cannot be negative"); 831 goto ret; 832 } 833 834 cfg = &tg->ts.cfg; 835 switch (info->category) { 836 case AVG: 837 cfg->buckets[info->type].avg = value; 838 break; 839 case MAX: 840 cfg->buckets[info->type].max = value; 841 break; 842 case BURST_LENGTH: 843 if (value > UINT_MAX) { 844 error_setg(&local_err, "%s value must be in the" 845 "range [0, %u]", info->name, UINT_MAX); 846 goto ret; 847 } 848 cfg->buckets[info->type].burst_length = value; 849 break; 850 case IOPS_SIZE: 851 cfg->op_size = value; 852 break; 853 } 854 855 ret: 856 error_propagate(errp, local_err); 857 return; 858 859 } 860 861 static void throttle_group_get(Object *obj, Visitor *v, const char *name, 862 void *opaque, Error **errp) 863 { 864 ThrottleGroup *tg = THROTTLE_GROUP(obj); 865 ThrottleConfig cfg; 866 ThrottleParamInfo *info = opaque; 867 int64_t value; 868 869 throttle_get_config(&tg->ts, &cfg); 870 switch (info->category) { 871 case AVG: 872 value = cfg.buckets[info->type].avg; 873 break; 874 case MAX: 875 value = cfg.buckets[info->type].max; 876 break; 877 case BURST_LENGTH: 878 value = cfg.buckets[info->type].burst_length; 879 break; 880 case IOPS_SIZE: 881 value = cfg.op_size; 882 break; 883 } 884 885 visit_type_int64(v, name, &value, errp); 886 } 887 888 static void throttle_group_set_limits(Object *obj, Visitor *v, 889 const char *name, void *opaque, 890 Error **errp) 891 892 { 893 ThrottleGroup *tg = THROTTLE_GROUP(obj); 894 ThrottleConfig cfg; 895 ThrottleLimits arg = { 0 }; 896 ThrottleLimits *argp = &arg; 897 Error *local_err = NULL; 898 899 visit_type_ThrottleLimits(v, name, &argp, &local_err); 900 if (local_err) { 901 goto ret; 902 } 903 qemu_mutex_lock(&tg->lock); 904 throttle_get_config(&tg->ts, &cfg); 905 throttle_limits_to_config(argp, &cfg, &local_err); 906 if (local_err) { 907 goto unlock; 908 } 909 throttle_config(&tg->ts, tg->clock_type, &cfg); 910 911 unlock: 912 qemu_mutex_unlock(&tg->lock); 913 ret: 914 error_propagate(errp, local_err); 915 return; 916 } 917 918 static void throttle_group_get_limits(Object *obj, Visitor *v, 919 const char *name, void *opaque, 920 Error **errp) 921 { 922 ThrottleGroup *tg = THROTTLE_GROUP(obj); 923 ThrottleConfig cfg; 924 ThrottleLimits arg = { 0 }; 925 ThrottleLimits *argp = &arg; 926 927 qemu_mutex_lock(&tg->lock); 928 throttle_get_config(&tg->ts, &cfg); 929 qemu_mutex_unlock(&tg->lock); 930 931 throttle_config_to_limits(&cfg, argp); 932 933 visit_type_ThrottleLimits(v, name, &argp, errp); 934 } 935 936 static bool throttle_group_can_be_deleted(UserCreatable *uc) 937 { 938 return OBJECT(uc)->ref == 1; 939 } 940 941 static void throttle_group_obj_class_init(ObjectClass *klass, void *class_data) 942 { 943 size_t i = 0; 944 UserCreatableClass *ucc = USER_CREATABLE_CLASS(klass); 945 946 ucc->complete = throttle_group_obj_complete; 947 ucc->can_be_deleted = throttle_group_can_be_deleted; 948 949 /* individual properties */ 950 for (i = 0; i < sizeof(properties) / sizeof(ThrottleParamInfo); i++) { 951 object_class_property_add(klass, 952 properties[i].name, 953 "int", 954 throttle_group_get, 955 throttle_group_set, 956 NULL, &properties[i], 957 &error_abort); 958 } 959 960 /* ThrottleLimits */ 961 object_class_property_add(klass, 962 "limits", "ThrottleLimits", 963 throttle_group_get_limits, 964 throttle_group_set_limits, 965 NULL, NULL, 966 &error_abort); 967 } 968 969 static const TypeInfo throttle_group_info = { 970 .name = TYPE_THROTTLE_GROUP, 971 .parent = TYPE_OBJECT, 972 .class_init = throttle_group_obj_class_init, 973 .instance_size = sizeof(ThrottleGroup), 974 .instance_init = throttle_group_obj_init, 975 .instance_finalize = throttle_group_obj_finalize, 976 .interfaces = (InterfaceInfo[]) { 977 { TYPE_USER_CREATABLE }, 978 { } 979 }, 980 }; 981 982 static void throttle_groups_init(void) 983 { 984 type_register_static(&throttle_group_info); 985 } 986 987 type_init(throttle_groups_init); 988