qemu/block/throttle-groups.c
<<
>>
Prefs
   1/*
   2 * QEMU block throttling group infrastructure
   3 *
   4 * Copyright (C) Nodalink, EURL. 2014
   5 * Copyright (C) Igalia, S.L. 2015
   6 *
   7 * Authors:
   8 *   BenoƮt Canet <benoit.canet@nodalink.com>
   9 *   Alberto Garcia <berto@igalia.com>
  10 *
  11 * This program is free software; you can redistribute it and/or
  12 * modify it under the terms of the GNU General Public License as
  13 * published by the Free Software Foundation; either version 2 or
  14 * (at your option) version 3 of the License.
  15 *
  16 * This program is distributed in the hope that it will be useful,
  17 * but WITHOUT ANY WARRANTY; without even the implied warranty of
  18 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
  19 * GNU General Public License for more details.
  20 *
  21 * You should have received a copy of the GNU General Public License
  22 * along with this program; if not, see <http://www.gnu.org/licenses/>.
  23 */
  24
  25#include "qemu/osdep.h"
  26#include "sysemu/block-backend.h"
  27#include "block/throttle-groups.h"
  28#include "qemu/throttle-options.h"
  29#include "qemu/queue.h"
  30#include "qemu/thread.h"
  31#include "sysemu/qtest.h"
  32#include "qapi/error.h"
  33#include "qapi-visit.h"
  34#include "qom/object.h"
  35#include "qom/object_interfaces.h"
  36
  37static void throttle_group_obj_init(Object *obj);
  38static void throttle_group_obj_complete(UserCreatable *obj, Error **errp);
  39
  40/* The ThrottleGroup structure (with its ThrottleState) is shared
  41 * among different ThrottleGroupMembers and it's independent from
  42 * AioContext, so in order to use it from different threads it needs
  43 * its own locking.
  44 *
  45 * This locking is however handled internally in this file, so it's
  46 * transparent to outside users.
  47 *
  48 * The whole ThrottleGroup structure is private and invisible to
  49 * outside users, that only use it through its ThrottleState.
  50 *
  51 * In addition to the ThrottleGroup structure, ThrottleGroupMember has
  52 * fields that need to be accessed by other members of the group and
  53 * therefore also need to be protected by this lock. Once a
  54 * ThrottleGroupMember is registered in a group those fields can be accessed
  55 * by other threads any time.
  56 *
  57 * Again, all this is handled internally and is mostly transparent to
  58 * the outside. The 'throttle_timers' field however has an additional
  59 * constraint because it may be temporarily invalid (see for example
  60 * blk_set_aio_context()). Therefore in this file a thread will
  61 * access some other ThrottleGroupMember's timers only after verifying that
  62 * that ThrottleGroupMember has throttled requests in the queue.
  63 */
  64typedef struct ThrottleGroup {
  65    Object parent_obj;
  66
  67    /* refuse individual property change if initialization is complete */
  68    bool is_initialized;
  69    char *name; /* This is constant during the lifetime of the group */
  70
  71    QemuMutex lock; /* This lock protects the following four fields */
  72    ThrottleState ts;
  73    QLIST_HEAD(, ThrottleGroupMember) head;
  74    ThrottleGroupMember *tokens[2];
  75    bool any_timer_armed[2];
  76    QEMUClockType clock_type;
  77
  78    /* This field is protected by the global QEMU mutex */
  79    QTAILQ_ENTRY(ThrottleGroup) list;
  80} ThrottleGroup;
  81
  82/* This is protected by the global QEMU mutex */
  83static QTAILQ_HEAD(, ThrottleGroup) throttle_groups =
  84    QTAILQ_HEAD_INITIALIZER(throttle_groups);
  85
  86
  87/* This function reads throttle_groups and must be called under the global
  88 * mutex.
  89 */
  90static ThrottleGroup *throttle_group_by_name(const char *name)
  91{
  92    ThrottleGroup *iter;
  93
  94    /* Look for an existing group with that name */
  95    QTAILQ_FOREACH(iter, &throttle_groups, list) {
  96        if (!g_strcmp0(name, iter->name)) {
  97            return iter;
  98        }
  99    }
 100
 101    return NULL;
 102}
 103
 104/* This function reads throttle_groups and must be called under the global
 105 * mutex.
 106 */
 107bool throttle_group_exists(const char *name)
 108{
 109    return throttle_group_by_name(name) != NULL;
 110}
 111
 112/* Increments the reference count of a ThrottleGroup given its name.
 113 *
 114 * If no ThrottleGroup is found with the given name a new one is
 115 * created.
 116 *
 117 * This function edits throttle_groups and must be called under the global
 118 * mutex.
 119 *
 120 * @name: the name of the ThrottleGroup
 121 * @ret:  the ThrottleState member of the ThrottleGroup
 122 */
 123ThrottleState *throttle_group_incref(const char *name)
 124{
 125    ThrottleGroup *tg = NULL;
 126
 127    /* Look for an existing group with that name */
 128    tg = throttle_group_by_name(name);
 129
 130    if (tg) {
 131        object_ref(OBJECT(tg));
 132    } else {
 133        /* Create a new one if not found */
 134        /* new ThrottleGroup obj will have a refcnt = 1 */
 135        tg = THROTTLE_GROUP(object_new(TYPE_THROTTLE_GROUP));
 136        tg->name = g_strdup(name);
 137        throttle_group_obj_complete(USER_CREATABLE(tg), &error_abort);
 138    }
 139
 140    return &tg->ts;
 141}
 142
 143/* Decrease the reference count of a ThrottleGroup.
 144 *
 145 * When the reference count reaches zero the ThrottleGroup is
 146 * destroyed.
 147 *
 148 * This function edits throttle_groups and must be called under the global
 149 * mutex.
 150 *
 151 * @ts:  The ThrottleGroup to unref, given by its ThrottleState member
 152 */
 153void throttle_group_unref(ThrottleState *ts)
 154{
 155    ThrottleGroup *tg = container_of(ts, ThrottleGroup, ts);
 156    object_unref(OBJECT(tg));
 157}
 158
 159/* Get the name from a ThrottleGroupMember's group. The name (and the pointer)
 160 * is guaranteed to remain constant during the lifetime of the group.
 161 *
 162 * @tgm:  a ThrottleGroupMember
 163 * @ret:  the name of the group.
 164 */
 165const char *throttle_group_get_name(ThrottleGroupMember *tgm)
 166{
 167    ThrottleGroup *tg = container_of(tgm->throttle_state, ThrottleGroup, ts);
 168    return tg->name;
 169}
 170
 171/* Return the next ThrottleGroupMember in the round-robin sequence, simulating
 172 * a circular list.
 173 *
 174 * This assumes that tg->lock is held.
 175 *
 176 * @tgm: the current ThrottleGroupMember
 177 * @ret: the next ThrottleGroupMember in the sequence
 178 */
 179static ThrottleGroupMember *throttle_group_next_tgm(ThrottleGroupMember *tgm)
 180{
 181    ThrottleState *ts = tgm->throttle_state;
 182    ThrottleGroup *tg = container_of(ts, ThrottleGroup, ts);
 183    ThrottleGroupMember *next = QLIST_NEXT(tgm, round_robin);
 184
 185    if (!next) {
 186        next = QLIST_FIRST(&tg->head);
 187    }
 188
 189    return next;
 190}
 191
 192/*
 193 * Return whether a ThrottleGroupMember has pending requests.
 194 *
 195 * This assumes that tg->lock is held.
 196 *
 197 * @tgm:        the ThrottleGroupMember
 198 * @is_write:   the type of operation (read/write)
 199 * @ret:        whether the ThrottleGroupMember has pending requests.
 200 */
 201static inline bool tgm_has_pending_reqs(ThrottleGroupMember *tgm,
 202                                        bool is_write)
 203{
 204    return tgm->pending_reqs[is_write];
 205}
 206
 207/* Return the next ThrottleGroupMember in the round-robin sequence with pending
 208 * I/O requests.
 209 *
 210 * This assumes that tg->lock is held.
 211 *
 212 * @tgm:       the current ThrottleGroupMember
 213 * @is_write:  the type of operation (read/write)
 214 * @ret:       the next ThrottleGroupMember with pending requests, or tgm if
 215 *             there is none.
 216 */
 217static ThrottleGroupMember *next_throttle_token(ThrottleGroupMember *tgm,
 218                                                bool is_write)
 219{
 220    ThrottleState *ts = tgm->throttle_state;
 221    ThrottleGroup *tg = container_of(ts, ThrottleGroup, ts);
 222    ThrottleGroupMember *token, *start;
 223
 224    start = token = tg->tokens[is_write];
 225
 226    /* get next bs round in round robin style */
 227    token = throttle_group_next_tgm(token);
 228    while (token != start && !tgm_has_pending_reqs(token, is_write)) {
 229        token = throttle_group_next_tgm(token);
 230    }
 231
 232    /* If no IO are queued for scheduling on the next round robin token
 233     * then decide the token is the current tgm because chances are
 234     * the current tgm got the current request queued.
 235     */
 236    if (token == start && !tgm_has_pending_reqs(token, is_write)) {
 237        token = tgm;
 238    }
 239
 240    /* Either we return the original TGM, or one with pending requests */
 241    assert(token == tgm || tgm_has_pending_reqs(token, is_write));
 242
 243    return token;
 244}
 245
 246/* Check if the next I/O request for a ThrottleGroupMember needs to be
 247 * throttled or not. If there's no timer set in this group, set one and update
 248 * the token accordingly.
 249 *
 250 * This assumes that tg->lock is held.
 251 *
 252 * @tgm:        the current ThrottleGroupMember
 253 * @is_write:   the type of operation (read/write)
 254 * @ret:        whether the I/O request needs to be throttled or not
 255 */
 256static bool throttle_group_schedule_timer(ThrottleGroupMember *tgm,
 257                                          bool is_write)
 258{
 259    ThrottleState *ts = tgm->throttle_state;
 260    ThrottleGroup *tg = container_of(ts, ThrottleGroup, ts);
 261    ThrottleTimers *tt = &tgm->throttle_timers;
 262    bool must_wait;
 263
 264    if (atomic_read(&tgm->io_limits_disabled)) {
 265        return false;
 266    }
 267
 268    /* Check if any of the timers in this group is already armed */
 269    if (tg->any_timer_armed[is_write]) {
 270        return true;
 271    }
 272
 273    must_wait = throttle_schedule_timer(ts, tt, is_write);
 274
 275    /* If a timer just got armed, set tgm as the current token */
 276    if (must_wait) {
 277        tg->tokens[is_write] = tgm;
 278        tg->any_timer_armed[is_write] = true;
 279    }
 280
 281    return must_wait;
 282}
 283
 284/* Start the next pending I/O request for a ThrottleGroupMember. Return whether
 285 * any request was actually pending.
 286 *
 287 * @tgm:       the current ThrottleGroupMember
 288 * @is_write:  the type of operation (read/write)
 289 */
 290static bool coroutine_fn throttle_group_co_restart_queue(ThrottleGroupMember *tgm,
 291                                                         bool is_write)
 292{
 293    bool ret;
 294
 295    qemu_co_mutex_lock(&tgm->throttled_reqs_lock);
 296    ret = qemu_co_queue_next(&tgm->throttled_reqs[is_write]);
 297    qemu_co_mutex_unlock(&tgm->throttled_reqs_lock);
 298
 299    return ret;
 300}
 301
 302/* Look for the next pending I/O request and schedule it.
 303 *
 304 * This assumes that tg->lock is held.
 305 *
 306 * @tgm:       the current ThrottleGroupMember
 307 * @is_write:  the type of operation (read/write)
 308 */
 309static void schedule_next_request(ThrottleGroupMember *tgm, bool is_write)
 310{
 311    ThrottleState *ts = tgm->throttle_state;
 312    ThrottleGroup *tg = container_of(ts, ThrottleGroup, ts);
 313    bool must_wait;
 314    ThrottleGroupMember *token;
 315
 316    /* Check if there's any pending request to schedule next */
 317    token = next_throttle_token(tgm, is_write);
 318    if (!tgm_has_pending_reqs(token, is_write)) {
 319        return;
 320    }
 321
 322    /* Set a timer for the request if it needs to be throttled */
 323    must_wait = throttle_group_schedule_timer(token, is_write);
 324
 325    /* If it doesn't have to wait, queue it for immediate execution */
 326    if (!must_wait) {
 327        /* Give preference to requests from the current tgm */
 328        if (qemu_in_coroutine() &&
 329            throttle_group_co_restart_queue(tgm, is_write)) {
 330            token = tgm;
 331        } else {
 332            ThrottleTimers *tt = &token->throttle_timers;
 333            int64_t now = qemu_clock_get_ns(tg->clock_type);
 334            timer_mod(tt->timers[is_write], now);
 335            tg->any_timer_armed[is_write] = true;
 336        }
 337        tg->tokens[is_write] = token;
 338    }
 339}
 340
 341/* Check if an I/O request needs to be throttled, wait and set a timer
 342 * if necessary, and schedule the next request using a round robin
 343 * algorithm.
 344 *
 345 * @tgm:       the current ThrottleGroupMember
 346 * @bytes:     the number of bytes for this I/O
 347 * @is_write:  the type of operation (read/write)
 348 */
 349void coroutine_fn throttle_group_co_io_limits_intercept(ThrottleGroupMember *tgm,
 350                                                        unsigned int bytes,
 351                                                        bool is_write)
 352{
 353    bool must_wait;
 354    ThrottleGroupMember *token;
 355    ThrottleGroup *tg = container_of(tgm->throttle_state, ThrottleGroup, ts);
 356    qemu_mutex_lock(&tg->lock);
 357
 358    /* First we check if this I/O has to be throttled. */
 359    token = next_throttle_token(tgm, is_write);
 360    must_wait = throttle_group_schedule_timer(token, is_write);
 361
 362    /* Wait if there's a timer set or queued requests of this type */
 363    if (must_wait || tgm->pending_reqs[is_write]) {
 364        tgm->pending_reqs[is_write]++;
 365        qemu_mutex_unlock(&tg->lock);
 366        qemu_co_mutex_lock(&tgm->throttled_reqs_lock);
 367        qemu_co_queue_wait(&tgm->throttled_reqs[is_write],
 368                           &tgm->throttled_reqs_lock);
 369        qemu_co_mutex_unlock(&tgm->throttled_reqs_lock);
 370        qemu_mutex_lock(&tg->lock);
 371        tgm->pending_reqs[is_write]--;
 372    }
 373
 374    /* The I/O will be executed, so do the accounting */
 375    throttle_account(tgm->throttle_state, is_write, bytes);
 376
 377    /* Schedule the next request */
 378    schedule_next_request(tgm, is_write);
 379
 380    qemu_mutex_unlock(&tg->lock);
 381}
 382
 383typedef struct {
 384    ThrottleGroupMember *tgm;
 385    bool is_write;
 386} RestartData;
 387
 388static void coroutine_fn throttle_group_restart_queue_entry(void *opaque)
 389{
 390    RestartData *data = opaque;
 391    ThrottleGroupMember *tgm = data->tgm;
 392    ThrottleState *ts = tgm->throttle_state;
 393    ThrottleGroup *tg = container_of(ts, ThrottleGroup, ts);
 394    bool is_write = data->is_write;
 395    bool empty_queue;
 396
 397    empty_queue = !throttle_group_co_restart_queue(tgm, is_write);
 398
 399    /* If the request queue was empty then we have to take care of
 400     * scheduling the next one */
 401    if (empty_queue) {
 402        qemu_mutex_lock(&tg->lock);
 403        schedule_next_request(tgm, is_write);
 404        qemu_mutex_unlock(&tg->lock);
 405    }
 406
 407    g_free(data);
 408}
 409
 410static void throttle_group_restart_queue(ThrottleGroupMember *tgm, bool is_write)
 411{
 412    Coroutine *co;
 413    RestartData *rd = g_new0(RestartData, 1);
 414
 415    rd->tgm = tgm;
 416    rd->is_write = is_write;
 417
 418    co = qemu_coroutine_create(throttle_group_restart_queue_entry, rd);
 419    aio_co_enter(tgm->aio_context, co);
 420}
 421
 422void throttle_group_restart_tgm(ThrottleGroupMember *tgm)
 423{
 424    if (tgm->throttle_state) {
 425        throttle_group_restart_queue(tgm, 0);
 426        throttle_group_restart_queue(tgm, 1);
 427    }
 428}
 429
 430/* Update the throttle configuration for a particular group. Similar
 431 * to throttle_config(), but guarantees atomicity within the
 432 * throttling group.
 433 *
 434 * @tgm:    a ThrottleGroupMember that is a member of the group
 435 * @cfg: the configuration to set
 436 */
 437void throttle_group_config(ThrottleGroupMember *tgm, ThrottleConfig *cfg)
 438{
 439    ThrottleState *ts = tgm->throttle_state;
 440    ThrottleGroup *tg = container_of(ts, ThrottleGroup, ts);
 441    qemu_mutex_lock(&tg->lock);
 442    throttle_config(ts, tg->clock_type, cfg);
 443    qemu_mutex_unlock(&tg->lock);
 444
 445    throttle_group_restart_tgm(tgm);
 446}
 447
 448/* Get the throttle configuration from a particular group. Similar to
 449 * throttle_get_config(), but guarantees atomicity within the
 450 * throttling group.
 451 *
 452 * @tgm:    a ThrottleGroupMember that is a member of the group
 453 * @cfg: the configuration will be written here
 454 */
 455void throttle_group_get_config(ThrottleGroupMember *tgm, ThrottleConfig *cfg)
 456{
 457    ThrottleState *ts = tgm->throttle_state;
 458    ThrottleGroup *tg = container_of(ts, ThrottleGroup, ts);
 459    qemu_mutex_lock(&tg->lock);
 460    throttle_get_config(ts, cfg);
 461    qemu_mutex_unlock(&tg->lock);
 462}
 463
 464/* ThrottleTimers callback. This wakes up a request that was waiting
 465 * because it had been throttled.
 466 *
 467 * @tgm:       the ThrottleGroupMember whose request had been throttled
 468 * @is_write:  the type of operation (read/write)
 469 */
 470static void timer_cb(ThrottleGroupMember *tgm, bool is_write)
 471{
 472    ThrottleState *ts = tgm->throttle_state;
 473    ThrottleGroup *tg = container_of(ts, ThrottleGroup, ts);
 474
 475    /* The timer has just been fired, so we can update the flag */
 476    qemu_mutex_lock(&tg->lock);
 477    tg->any_timer_armed[is_write] = false;
 478    qemu_mutex_unlock(&tg->lock);
 479
 480    /* Run the request that was waiting for this timer */
 481    throttle_group_restart_queue(tgm, is_write);
 482}
 483
 484static void read_timer_cb(void *opaque)
 485{
 486    timer_cb(opaque, false);
 487}
 488
 489static void write_timer_cb(void *opaque)
 490{
 491    timer_cb(opaque, true);
 492}
 493
 494/* Register a ThrottleGroupMember from the throttling group, also initializing
 495 * its timers and updating its throttle_state pointer to point to it. If a
 496 * throttling group with that name does not exist yet, it will be created.
 497 *
 498 * This function edits throttle_groups and must be called under the global
 499 * mutex.
 500 *
 501 * @tgm:       the ThrottleGroupMember to insert
 502 * @groupname: the name of the group
 503 * @ctx:       the AioContext to use
 504 */
 505void throttle_group_register_tgm(ThrottleGroupMember *tgm,
 506                                 const char *groupname,
 507                                 AioContext *ctx)
 508{
 509    int i;
 510    ThrottleState *ts = throttle_group_incref(groupname);
 511    ThrottleGroup *tg = container_of(ts, ThrottleGroup, ts);
 512
 513    tgm->throttle_state = ts;
 514    tgm->aio_context = ctx;
 515
 516    qemu_mutex_lock(&tg->lock);
 517    /* If the ThrottleGroup is new set this ThrottleGroupMember as the token */
 518    for (i = 0; i < 2; i++) {
 519        if (!tg->tokens[i]) {
 520            tg->tokens[i] = tgm;
 521        }
 522    }
 523
 524    QLIST_INSERT_HEAD(&tg->head, tgm, round_robin);
 525
 526    throttle_timers_init(&tgm->throttle_timers,
 527                         tgm->aio_context,
 528                         tg->clock_type,
 529                         read_timer_cb,
 530                         write_timer_cb,
 531                         tgm);
 532    qemu_co_mutex_init(&tgm->throttled_reqs_lock);
 533    qemu_co_queue_init(&tgm->throttled_reqs[0]);
 534    qemu_co_queue_init(&tgm->throttled_reqs[1]);
 535
 536    qemu_mutex_unlock(&tg->lock);
 537}
 538
 539/* Unregister a ThrottleGroupMember from its group, removing it from the list,
 540 * destroying the timers and setting the throttle_state pointer to NULL.
 541 *
 542 * The ThrottleGroupMember must not have pending throttled requests, so the
 543 * caller has to drain them first.
 544 *
 545 * The group will be destroyed if it's empty after this operation.
 546 *
 547 * @tgm the ThrottleGroupMember to remove
 548 */
 549void throttle_group_unregister_tgm(ThrottleGroupMember *tgm)
 550{
 551    ThrottleState *ts = tgm->throttle_state;
 552    ThrottleGroup *tg = container_of(ts, ThrottleGroup, ts);
 553    ThrottleGroupMember *token;
 554    int i;
 555
 556    if (!ts) {
 557        /* Discard already unregistered tgm */
 558        return;
 559    }
 560
 561    assert(tgm->pending_reqs[0] == 0 && tgm->pending_reqs[1] == 0);
 562    assert(qemu_co_queue_empty(&tgm->throttled_reqs[0]));
 563    assert(qemu_co_queue_empty(&tgm->throttled_reqs[1]));
 564
 565    qemu_mutex_lock(&tg->lock);
 566    for (i = 0; i < 2; i++) {
 567        if (tg->tokens[i] == tgm) {
 568            token = throttle_group_next_tgm(tgm);
 569            /* Take care of the case where this is the last tgm in the group */
 570            if (token == tgm) {
 571                token = NULL;
 572            }
 573            tg->tokens[i] = token;
 574        }
 575    }
 576
 577    /* remove the current tgm from the list */
 578    QLIST_REMOVE(tgm, round_robin);
 579    throttle_timers_destroy(&tgm->throttle_timers);
 580    qemu_mutex_unlock(&tg->lock);
 581
 582    throttle_group_unref(&tg->ts);
 583    tgm->throttle_state = NULL;
 584}
 585
 586void throttle_group_attach_aio_context(ThrottleGroupMember *tgm,
 587                                       AioContext *new_context)
 588{
 589    ThrottleTimers *tt = &tgm->throttle_timers;
 590    throttle_timers_attach_aio_context(tt, new_context);
 591    tgm->aio_context = new_context;
 592}
 593
 594void throttle_group_detach_aio_context(ThrottleGroupMember *tgm)
 595{
 596    ThrottleGroup *tg = container_of(tgm->throttle_state, ThrottleGroup, ts);
 597    ThrottleTimers *tt = &tgm->throttle_timers;
 598    int i;
 599
 600    /* Requests must have been drained */
 601    assert(tgm->pending_reqs[0] == 0 && tgm->pending_reqs[1] == 0);
 602    assert(qemu_co_queue_empty(&tgm->throttled_reqs[0]));
 603    assert(qemu_co_queue_empty(&tgm->throttled_reqs[1]));
 604
 605    /* Kick off next ThrottleGroupMember, if necessary */
 606    qemu_mutex_lock(&tg->lock);
 607    for (i = 0; i < 2; i++) {
 608        if (timer_pending(tt->timers[i])) {
 609            tg->any_timer_armed[i] = false;
 610            schedule_next_request(tgm, i);
 611        }
 612    }
 613    qemu_mutex_unlock(&tg->lock);
 614
 615    throttle_timers_detach_aio_context(tt);
 616    tgm->aio_context = NULL;
 617}
 618
 619#undef THROTTLE_OPT_PREFIX
 620#define THROTTLE_OPT_PREFIX "x-"
 621
 622/* Helper struct and array for QOM property setter/getter */
 623typedef struct {
 624    const char *name;
 625    BucketType type;
 626    enum {
 627        AVG,
 628        MAX,
 629        BURST_LENGTH,
 630        IOPS_SIZE,
 631    } category;
 632} ThrottleParamInfo;
 633
 634static ThrottleParamInfo properties[] = {
 635    {
 636        THROTTLE_OPT_PREFIX QEMU_OPT_IOPS_TOTAL,
 637        THROTTLE_OPS_TOTAL, AVG,
 638    },
 639    {
 640        THROTTLE_OPT_PREFIX QEMU_OPT_IOPS_TOTAL_MAX,
 641        THROTTLE_OPS_TOTAL, MAX,
 642    },
 643    {
 644        THROTTLE_OPT_PREFIX QEMU_OPT_IOPS_TOTAL_MAX_LENGTH,
 645        THROTTLE_OPS_TOTAL, BURST_LENGTH,
 646    },
 647    {
 648        THROTTLE_OPT_PREFIX QEMU_OPT_IOPS_READ,
 649        THROTTLE_OPS_READ, AVG,
 650    },
 651    {
 652        THROTTLE_OPT_PREFIX QEMU_OPT_IOPS_READ_MAX,
 653        THROTTLE_OPS_READ, MAX,
 654    },
 655    {
 656        THROTTLE_OPT_PREFIX QEMU_OPT_IOPS_READ_MAX_LENGTH,
 657        THROTTLE_OPS_READ, BURST_LENGTH,
 658    },
 659    {
 660        THROTTLE_OPT_PREFIX QEMU_OPT_IOPS_WRITE,
 661        THROTTLE_OPS_WRITE, AVG,
 662    },
 663    {
 664        THROTTLE_OPT_PREFIX QEMU_OPT_IOPS_WRITE_MAX,
 665        THROTTLE_OPS_WRITE, MAX,
 666    },
 667    {
 668        THROTTLE_OPT_PREFIX QEMU_OPT_IOPS_WRITE_MAX_LENGTH,
 669        THROTTLE_OPS_WRITE, BURST_LENGTH,
 670    },
 671    {
 672        THROTTLE_OPT_PREFIX QEMU_OPT_BPS_TOTAL,
 673        THROTTLE_BPS_TOTAL, AVG,
 674    },
 675    {
 676        THROTTLE_OPT_PREFIX QEMU_OPT_BPS_TOTAL_MAX,
 677        THROTTLE_BPS_TOTAL, MAX,
 678    },
 679    {
 680        THROTTLE_OPT_PREFIX QEMU_OPT_BPS_TOTAL_MAX_LENGTH,
 681        THROTTLE_BPS_TOTAL, BURST_LENGTH,
 682    },
 683    {
 684        THROTTLE_OPT_PREFIX QEMU_OPT_BPS_READ,
 685        THROTTLE_BPS_READ, AVG,
 686    },
 687    {
 688        THROTTLE_OPT_PREFIX QEMU_OPT_BPS_READ_MAX,
 689        THROTTLE_BPS_READ, MAX,
 690    },
 691    {
 692        THROTTLE_OPT_PREFIX QEMU_OPT_BPS_READ_MAX_LENGTH,
 693        THROTTLE_BPS_READ, BURST_LENGTH,
 694    },
 695    {
 696        THROTTLE_OPT_PREFIX QEMU_OPT_BPS_WRITE,
 697        THROTTLE_BPS_WRITE, AVG,
 698    },
 699    {
 700        THROTTLE_OPT_PREFIX QEMU_OPT_BPS_WRITE_MAX,
 701        THROTTLE_BPS_WRITE, MAX,
 702    },
 703    {
 704        THROTTLE_OPT_PREFIX QEMU_OPT_BPS_WRITE_MAX_LENGTH,
 705        THROTTLE_BPS_WRITE, BURST_LENGTH,
 706    },
 707    {
 708        THROTTLE_OPT_PREFIX QEMU_OPT_IOPS_SIZE,
 709        0, IOPS_SIZE,
 710    }
 711};
 712
 713/* This function edits throttle_groups and must be called under the global
 714 * mutex */
 715static void throttle_group_obj_init(Object *obj)
 716{
 717    ThrottleGroup *tg = THROTTLE_GROUP(obj);
 718
 719    tg->clock_type = QEMU_CLOCK_REALTIME;
 720    if (qtest_enabled()) {
 721        /* For testing block IO throttling only */
 722        tg->clock_type = QEMU_CLOCK_VIRTUAL;
 723    }
 724    tg->is_initialized = false;
 725    qemu_mutex_init(&tg->lock);
 726    throttle_init(&tg->ts);
 727    QLIST_INIT(&tg->head);
 728}
 729
 730/* This function edits throttle_groups and must be called under the global
 731 * mutex */
 732static void throttle_group_obj_complete(UserCreatable *obj, Error **errp)
 733{
 734    ThrottleGroup *tg = THROTTLE_GROUP(obj);
 735    ThrottleConfig cfg;
 736
 737    /* set group name to object id if it exists */
 738    if (!tg->name && tg->parent_obj.parent) {
 739        tg->name = object_get_canonical_path_component(OBJECT(obj));
 740    }
 741    /* We must have a group name at this point */
 742    assert(tg->name);
 743
 744    /* error if name is duplicate */
 745    if (throttle_group_exists(tg->name)) {
 746        error_setg(errp, "A group with this name already exists");
 747        return;
 748    }
 749
 750    /* check validity */
 751    throttle_get_config(&tg->ts, &cfg);
 752    if (!throttle_is_valid(&cfg, errp)) {
 753        return;
 754    }
 755    throttle_config(&tg->ts, tg->clock_type, &cfg);
 756    QTAILQ_INSERT_TAIL(&throttle_groups, tg, list);
 757    tg->is_initialized = true;
 758}
 759
 760/* This function edits throttle_groups and must be called under the global
 761 * mutex */
 762static void throttle_group_obj_finalize(Object *obj)
 763{
 764    ThrottleGroup *tg = THROTTLE_GROUP(obj);
 765    if (tg->is_initialized) {
 766        QTAILQ_REMOVE(&throttle_groups, tg, list);
 767    }
 768    qemu_mutex_destroy(&tg->lock);
 769    g_free(tg->name);
 770}
 771
 772static void throttle_group_set(Object *obj, Visitor *v, const char * name,
 773                               void *opaque, Error **errp)
 774
 775{
 776    ThrottleGroup *tg = THROTTLE_GROUP(obj);
 777    ThrottleConfig *cfg;
 778    ThrottleParamInfo *info = opaque;
 779    Error *local_err = NULL;
 780    int64_t value;
 781
 782    /* If we have finished initialization, don't accept individual property
 783     * changes through QOM. Throttle configuration limits must be set in one
 784     * transaction, as certain combinations are invalid.
 785     */
 786    if (tg->is_initialized) {
 787        error_setg(&local_err, "Property cannot be set after initialization");
 788        goto ret;
 789    }
 790
 791    visit_type_int64(v, name, &value, &local_err);
 792    if (local_err) {
 793        goto ret;
 794    }
 795    if (value < 0) {
 796        error_setg(&local_err, "Property values cannot be negative");
 797        goto ret;
 798    }
 799
 800    cfg = &tg->ts.cfg;
 801    switch (info->category) {
 802    case AVG:
 803        cfg->buckets[info->type].avg = value;
 804        break;
 805    case MAX:
 806        cfg->buckets[info->type].max = value;
 807        break;
 808    case BURST_LENGTH:
 809        if (value > UINT_MAX) {
 810            error_setg(&local_err, "%s value must be in the"
 811                       "range [0, %u]", info->name, UINT_MAX);
 812            goto ret;
 813        }
 814        cfg->buckets[info->type].burst_length = value;
 815        break;
 816    case IOPS_SIZE:
 817        cfg->op_size = value;
 818        break;
 819    }
 820
 821ret:
 822    error_propagate(errp, local_err);
 823    return;
 824
 825}
 826
 827static void throttle_group_get(Object *obj, Visitor *v, const char *name,
 828                               void *opaque, Error **errp)
 829{
 830    ThrottleGroup *tg = THROTTLE_GROUP(obj);
 831    ThrottleConfig cfg;
 832    ThrottleParamInfo *info = opaque;
 833    int64_t value;
 834
 835    throttle_get_config(&tg->ts, &cfg);
 836    switch (info->category) {
 837    case AVG:
 838        value = cfg.buckets[info->type].avg;
 839        break;
 840    case MAX:
 841        value = cfg.buckets[info->type].max;
 842        break;
 843    case BURST_LENGTH:
 844        value = cfg.buckets[info->type].burst_length;
 845        break;
 846    case IOPS_SIZE:
 847        value = cfg.op_size;
 848        break;
 849    }
 850
 851    visit_type_int64(v, name, &value, errp);
 852}
 853
 854static void throttle_group_set_limits(Object *obj, Visitor *v,
 855                                      const char *name, void *opaque,
 856                                      Error **errp)
 857
 858{
 859    ThrottleGroup *tg = THROTTLE_GROUP(obj);
 860    ThrottleConfig cfg;
 861    ThrottleLimits arg = { 0 };
 862    ThrottleLimits *argp = &arg;
 863    Error *local_err = NULL;
 864
 865    visit_type_ThrottleLimits(v, name, &argp, &local_err);
 866    if (local_err) {
 867        goto ret;
 868    }
 869    qemu_mutex_lock(&tg->lock);
 870    throttle_get_config(&tg->ts, &cfg);
 871    throttle_limits_to_config(argp, &cfg, &local_err);
 872    if (local_err) {
 873        goto unlock;
 874    }
 875    throttle_config(&tg->ts, tg->clock_type, &cfg);
 876
 877unlock:
 878    qemu_mutex_unlock(&tg->lock);
 879ret:
 880    error_propagate(errp, local_err);
 881    return;
 882}
 883
 884static void throttle_group_get_limits(Object *obj, Visitor *v,
 885                                      const char *name, void *opaque,
 886                                      Error **errp)
 887{
 888    ThrottleGroup *tg = THROTTLE_GROUP(obj);
 889    ThrottleConfig cfg;
 890    ThrottleLimits arg = { 0 };
 891    ThrottleLimits *argp = &arg;
 892
 893    qemu_mutex_lock(&tg->lock);
 894    throttle_get_config(&tg->ts, &cfg);
 895    qemu_mutex_unlock(&tg->lock);
 896
 897    throttle_config_to_limits(&cfg, argp);
 898
 899    visit_type_ThrottleLimits(v, name, &argp, errp);
 900}
 901
 902static bool throttle_group_can_be_deleted(UserCreatable *uc)
 903{
 904    return OBJECT(uc)->ref == 1;
 905}
 906
 907static void throttle_group_obj_class_init(ObjectClass *klass, void *class_data)
 908{
 909    size_t i = 0;
 910    UserCreatableClass *ucc = USER_CREATABLE_CLASS(klass);
 911
 912    ucc->complete = throttle_group_obj_complete;
 913    ucc->can_be_deleted = throttle_group_can_be_deleted;
 914
 915    /* individual properties */
 916    for (i = 0; i < sizeof(properties) / sizeof(ThrottleParamInfo); i++) {
 917        object_class_property_add(klass,
 918                                  properties[i].name,
 919                                  "int",
 920                                  throttle_group_get,
 921                                  throttle_group_set,
 922                                  NULL, &properties[i],
 923                                  &error_abort);
 924    }
 925
 926    /* ThrottleLimits */
 927    object_class_property_add(klass,
 928                              "limits", "ThrottleLimits",
 929                              throttle_group_get_limits,
 930                              throttle_group_set_limits,
 931                              NULL, NULL,
 932                              &error_abort);
 933}
 934
 935static const TypeInfo throttle_group_info = {
 936    .name = TYPE_THROTTLE_GROUP,
 937    .parent = TYPE_OBJECT,
 938    .class_init = throttle_group_obj_class_init,
 939    .instance_size = sizeof(ThrottleGroup),
 940    .instance_init = throttle_group_obj_init,
 941    .instance_finalize = throttle_group_obj_finalize,
 942    .interfaces = (InterfaceInfo[]) {
 943        { TYPE_USER_CREATABLE },
 944        { }
 945    },
 946};
 947
 948static void throttle_groups_init(void)
 949{
 950    type_register_static(&throttle_group_info);
 951}
 952
 953type_init(throttle_groups_init);
 954