| /openbmc/qemu/block/ |
| H A D | io_uring.c | 53 AioContext *aio_context; member 223 assert(luringcb->co->ctx == s->aio_context); in luring_process_completions() 433 s->aio_context = NULL; in luring_detach_aio_context() 438 s->aio_context = new_context; in luring_attach_aio_context() 440 aio_set_fd_handler(s->aio_context, s->ring.ring_fd, in luring_attach_aio_context()
|
| H A D | linux-aio.c | 58 AioContext *aio_context; member 107 assert(laiocb->co->ctx == laiocb->ctx->aio_context); in qemu_laio_process_completion() 346 uint64_t max_batch = s->aio_context->aio_max_batch ?: DEFAULT_MAX_BATCH; in laio_max_batch() 450 s->aio_context = NULL; in laio_detach_aio_context() 455 s->aio_context = new_context; in laio_attach_aio_context()
|
| H A D | throttle-groups.c | 446 aio_co_enter(tgm->aio_context, co); in throttle_group_restart_queue() 552 tgm->aio_context = ctx; in throttle_group_register_tgm() 567 tgm->aio_context, in throttle_group_register_tgm() 598 AIO_WAIT_WHILE(tgm->aio_context, qatomic_read(&tgm->restart_pending) > 0); in throttle_group_unregister_tgm() 629 tgm->aio_context = new_context; in throttle_group_attach_aio_context() 655 tgm->aio_context = NULL; in throttle_group_detach_aio_context()
|
| H A D | nfs.c | 60 AioContext *aio_context; member 194 aio_set_fd_handler(client->aio_context, nfs_get_fd(client->context), in nfs_set_events() 365 aio_set_fd_handler(client->aio_context, nfs_get_fd(client->context), in nfs_detach_aio_context() 375 client->aio_context = new_context; in nfs_attach_aio_context() 383 aio_set_fd_handler(client->aio_context, nfs_get_fd(client->context), in nfs_client_close() 606 client->aio_context = bdrv_get_aio_context(bs); in nfs_file_open() 643 client->aio_context = qemu_get_aio_context(); in nfs_file_co_create()
|
| H A D | curl.c | 118 AioContext *aio_context; member 135 aio_set_fd_handler(s->aio_context, socket->fd, in curl_drop_socket() 179 aio_set_fd_handler(s->aio_context, fd, in curl_sock_cb() 183 aio_set_fd_handler(s->aio_context, fd, in curl_sock_cb() 187 aio_set_fd_handler(s->aio_context, fd, in curl_sock_cb() 192 aio_set_fd_handler(s->aio_context, fd, in curl_sock_cb() 617 s->aio_context = new_context; in curl_attach_aio_context() 798 s->aio_context = bdrv_get_aio_context(bs); in curl_open()
|
| H A D | nvme.c | 100 AioContext *aio_context; member 222 AioContext *aio_context, in nvme_create_queue_pair() argument 237 trace_nvme_create_queue_pair(idx, q, size, aio_context, in nvme_create_queue_pair() 251 q->completion_bh = aio_bh_new(aio_context, nvme_process_completion_bh, q); in nvme_create_queue_pair() 348 replay_bh_schedule_oneshot_event(q->s->aio_context, in nvme_wake_free_req_locked() 497 if (qemu_get_current_aio_context() == q->s->aio_context) { in nvme_deferred_fn() 500 aio_bh_schedule_oneshot(q->s->aio_context, in nvme_deferred_fn() 537 AioContext *aio_context = bdrv_get_aio_context(bs); in nvme_admin_cmd_sync() local 546 AIO_WAIT_WHILE(aio_context, ret == -EINPROGRESS); in nvme_admin_cmd_sync() 761 AioContext *aio_context = bdrv_get_aio_context(bs); in nvme_init() local [all …]
|
| H A D | iscsi.c | 68 AioContext *aio_context; member 177 acb->bh = aio_bh_new(acb->iscsilun->aio_context, iscsi_bh_cb, acb); in iscsi_schedule_bh() 266 aio_timer_init(iTask->iscsilun->aio_context, in iscsi_co_generic_cb() 288 replay_bh_schedule_oneshot_event(iTask->iscsilun->aio_context, in iscsi_co_generic_cb() 365 aio_set_fd_handler(iscsilun->aio_context, iscsi_get_fd(iscsi), in iscsi_set_events() 1542 aio_set_fd_handler(iscsilun->aio_context, iscsi_get_fd(iscsilun->iscsi), in iscsi_detach_aio_context() 1561 iscsilun->aio_context = new_context; in iscsi_attach_aio_context() 1565 iscsilun->nop_timer = aio_timer_new(iscsilun->aio_context, in iscsi_attach_aio_context() 1573 iscsilun->event_timer = aio_timer_new(iscsilun->aio_context, in iscsi_attach_aio_context() 1906 iscsilun->aio_context = bdrv_get_aio_context(bs); in iscsi_open() [all …]
|
| H A D | gluster.c | 59 AioContext *aio_context; member 746 aio_co_schedule(acb->aio_context, acb->coroutine); in gluster_finish_aiocb() 1009 acb.aio_context = bdrv_get_aio_context(bs); in qemu_gluster_co_pwrite_zeroes() 1187 acb.aio_context = bdrv_get_aio_context(bs); in qemu_gluster_co_rw() 1254 acb.aio_context = bdrv_get_aio_context(bs); in qemu_gluster_co_flush_to_disk() 1302 acb.aio_context = bdrv_get_aio_context(bs); in qemu_gluster_co_pdiscard()
|
| H A D | commit.c | 444 s->base = blk_new(s->common.job.aio_context, in commit_start() 456 s->top = blk_new(s->common.job.aio_context, 0, BLK_PERM_ALL); in commit_start()
|
| H A D | replication.c | 150 assert(commit_job->aio_context == qemu_get_current_aio_context()); in replication_close()
|
| H A D | trace-events | 157 nvme_create_queue_pair(unsigned q_index, void *q, size_t size, void *aio_context, int fd) "index %u…
|
| H A D | mirror.c | 1975 s->target = blk_new(s->common.job.aio_context, in mirror_start_job()
|
| /openbmc/qemu/ |
| H A D | blockdev.c | 1585 AioContext *aio_context; in external_snapshot_abort() local 1591 aio_context = bdrv_get_aio_context(state->old_bs); in external_snapshot_abort() 1609 if (aio_context != tmp_context) { in external_snapshot_abort() 1611 aio_context, NULL, in external_snapshot_abort() 1644 AioContext *aio_context, 1664 AioContext *aio_context; in drive_backup_action() local 1691 aio_context = bdrv_get_aio_context(bs); in drive_backup_action() 1778 ret = bdrv_try_change_aio_context(target_bs, aio_context, NULL, errp); in drive_backup_action() 1794 bs, target_bs, aio_context, in drive_backup_action() 1850 AioContext *aio_context; in blockdev_backup_action() local [all …]
|
| H A D | job.c | 380 job->aio_context = ctx; in job_set_aio_context() 421 job->aio_context = ctx; in job_create() 591 next_aio_context = job->aio_context; in job_do_yield_locked() 601 next_aio_context = job->aio_context; in job_do_yield_locked() 1109 assert(job->aio_context == qemu_get_current_aio_context()); in job_co_entry() 1134 aio_co_enter(job->aio_context, job->co); in job_start() 1262 AIO_WAIT_WHILE_UNLOCKED(job->aio_context, in job_finish_sync_locked()
|
| H A D | blockjob.c | 179 return job->job.aio_context; in child_job_get_parent_aio_context() 627 return job->job.aio_context; in block_job_get_aio_context()
|
| H A D | block.c | 426 bs->aio_context = qemu_get_aio_context(); in bdrv_new() 7486 return bs ? bs->aio_context : qemu_get_aio_context(); in bdrv_get_aio_context() 7544 bs->aio_context = NULL; in bdrv_detach_aio_context() 7553 bs->aio_context = new_context; in bdrv_attach_aio_context()
|
| H A D | qemu-img.c | 868 AioContext *aio_context = block_job_get_aio_context(job); in run_block_job() local 876 aio_poll(aio_context, true); in run_block_job()
|
| /openbmc/qemu/include/block/ |
| H A D | throttle-groups.h | 37 AioContext *aio_context; member
|
| H A D | block_int-common.h | 1107 AioContext *aio_context; /* event loop used for fd handlers, timers, etc */ member
|
| /openbmc/qemu/include/qemu/ |
| H A D | throttle.h | 126 AioContext *aio_context,
|
| H A D | job.h | 86 AioContext *aio_context; member
|
| /openbmc/qemu/util/ |
| H A D | qemu-timer.c | 674 AioContext *aio_context; in qemu_clock_advance_virtual_time() local 675 aio_context = qemu_get_aio_context(); in qemu_clock_advance_virtual_time() 684 timerlist_run_timers(aio_context->tlg.tl[QEMU_CLOCK_VIRTUAL]); in qemu_clock_advance_virtual_time()
|
| H A D | throttle.c | 236 AioContext *aio_context, in throttle_timers_init() argument 249 throttle_timers_attach_aio_context(tt, aio_context); in throttle_timers_init()
|
| /openbmc/qemu/tests/unit/ |
| H A D | test-block-iothread.c | 517 g_assert(qemu_get_current_aio_context() == job->aio_context); in test_job_run() 528 g_assert(qemu_get_current_aio_context() == job->aio_context); in test_job_run() 773 g_assert(job->aio_context == ctx); in test_propagate_mirror()
|