1 /* 2 * QEMU I/O channels 3 * 4 * Copyright (c) 2015 Red Hat, Inc. 5 * 6 * This library is free software; you can redistribute it and/or 7 * modify it under the terms of the GNU Lesser General Public 8 * License as published by the Free Software Foundation; either 9 * version 2.1 of the License, or (at your option) any later version. 10 * 11 * This library is distributed in the hope that it will be useful, 12 * but WITHOUT ANY WARRANTY; without even the implied warranty of 13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU 14 * Lesser General Public License for more details. 15 * 16 * You should have received a copy of the GNU Lesser General Public 17 * License along with this library; if not, see <http://www.gnu.org/licenses/>. 18 * 19 */ 20 21 #include "qemu/osdep.h" 22 #include "io/channel.h" 23 #include "qapi/error.h" 24 #include "qemu/main-loop.h" 25 #include "qemu/module.h" 26 #include "qemu/iov.h" 27 28 bool qio_channel_has_feature(QIOChannel *ioc, 29 QIOChannelFeature feature) 30 { 31 return ioc->features & (1 << feature); 32 } 33 34 35 void qio_channel_set_feature(QIOChannel *ioc, 36 QIOChannelFeature feature) 37 { 38 ioc->features |= (1 << feature); 39 } 40 41 42 void qio_channel_set_name(QIOChannel *ioc, 43 const char *name) 44 { 45 g_free(ioc->name); 46 ioc->name = g_strdup(name); 47 } 48 49 50 ssize_t qio_channel_readv_full(QIOChannel *ioc, 51 const struct iovec *iov, 52 size_t niov, 53 int **fds, 54 size_t *nfds, 55 Error **errp) 56 { 57 QIOChannelClass *klass = QIO_CHANNEL_GET_CLASS(ioc); 58 59 if ((fds || nfds) && 60 !qio_channel_has_feature(ioc, QIO_CHANNEL_FEATURE_FD_PASS)) { 61 error_setg_errno(errp, EINVAL, 62 "Channel does not support file descriptor passing"); 63 return -1; 64 } 65 66 return klass->io_readv(ioc, iov, niov, fds, nfds, errp); 67 } 68 69 70 ssize_t qio_channel_writev_full(QIOChannel *ioc, 71 const struct iovec *iov, 72 size_t niov, 73 int *fds, 74 size_t nfds, 75 Error **errp) 76 { 77 QIOChannelClass *klass = QIO_CHANNEL_GET_CLASS(ioc); 78 79 if ((fds || nfds) && 80 !qio_channel_has_feature(ioc, QIO_CHANNEL_FEATURE_FD_PASS)) { 81 error_setg_errno(errp, EINVAL, 82 "Channel does not support file descriptor passing"); 83 return -1; 84 } 85 86 return klass->io_writev(ioc, iov, niov, fds, nfds, errp); 87 } 88 89 90 int qio_channel_readv_all_eof(QIOChannel *ioc, 91 const struct iovec *iov, 92 size_t niov, 93 Error **errp) 94 { 95 return qio_channel_readv_full_all_eof(ioc, iov, niov, NULL, NULL, errp); 96 } 97 98 int qio_channel_readv_all(QIOChannel *ioc, 99 const struct iovec *iov, 100 size_t niov, 101 Error **errp) 102 { 103 return qio_channel_readv_full_all(ioc, iov, niov, NULL, NULL, errp); 104 } 105 106 int qio_channel_readv_full_all_eof(QIOChannel *ioc, 107 const struct iovec *iov, 108 size_t niov, 109 int **fds, size_t *nfds, 110 Error **errp) 111 { 112 int ret = -1; 113 struct iovec *local_iov = g_new(struct iovec, niov); 114 struct iovec *local_iov_head = local_iov; 115 unsigned int nlocal_iov = niov; 116 int **local_fds = fds; 117 size_t *local_nfds = nfds; 118 bool partial = false; 119 120 if (nfds) { 121 *nfds = 0; 122 } 123 124 if (fds) { 125 *fds = NULL; 126 } 127 128 nlocal_iov = iov_copy(local_iov, nlocal_iov, 129 iov, niov, 130 0, iov_size(iov, niov)); 131 132 while ((nlocal_iov > 0) || local_fds) { 133 ssize_t len; 134 len = qio_channel_readv_full(ioc, local_iov, nlocal_iov, local_fds, 135 local_nfds, errp); 136 if (len == QIO_CHANNEL_ERR_BLOCK) { 137 if (qemu_in_coroutine()) { 138 qio_channel_yield(ioc, G_IO_IN); 139 } else { 140 qio_channel_wait(ioc, G_IO_IN); 141 } 142 continue; 143 } 144 145 if (len == 0) { 146 if (local_nfds && *local_nfds) { 147 /* 148 * Got some FDs, but no data yet. This isn't an EOF 149 * scenario (yet), so carry on to try to read data 150 * on next loop iteration 151 */ 152 goto next_iter; 153 } else if (!partial) { 154 /* No fds and no data - EOF before any data read */ 155 ret = 0; 156 goto cleanup; 157 } else { 158 len = -1; 159 error_setg(errp, 160 "Unexpected end-of-file before all data were read"); 161 /* Fallthrough into len < 0 handling */ 162 } 163 } 164 165 if (len < 0) { 166 /* Close any FDs we previously received */ 167 if (nfds && fds) { 168 size_t i; 169 for (i = 0; i < (*nfds); i++) { 170 close((*fds)[i]); 171 } 172 g_free(*fds); 173 *fds = NULL; 174 *nfds = 0; 175 } 176 goto cleanup; 177 } 178 179 if (nlocal_iov) { 180 iov_discard_front(&local_iov, &nlocal_iov, len); 181 } 182 183 next_iter: 184 partial = true; 185 local_fds = NULL; 186 local_nfds = NULL; 187 } 188 189 ret = 1; 190 191 cleanup: 192 g_free(local_iov_head); 193 return ret; 194 } 195 196 int qio_channel_readv_full_all(QIOChannel *ioc, 197 const struct iovec *iov, 198 size_t niov, 199 int **fds, size_t *nfds, 200 Error **errp) 201 { 202 int ret = qio_channel_readv_full_all_eof(ioc, iov, niov, fds, nfds, errp); 203 204 if (ret == 0) { 205 error_setg(errp, "Unexpected end-of-file before all data were read"); 206 return -1; 207 } 208 if (ret == 1) { 209 return 0; 210 } 211 212 return ret; 213 } 214 215 int qio_channel_writev_all(QIOChannel *ioc, 216 const struct iovec *iov, 217 size_t niov, 218 Error **errp) 219 { 220 return qio_channel_writev_full_all(ioc, iov, niov, NULL, 0, errp); 221 } 222 223 int qio_channel_writev_full_all(QIOChannel *ioc, 224 const struct iovec *iov, 225 size_t niov, 226 int *fds, size_t nfds, 227 Error **errp) 228 { 229 int ret = -1; 230 struct iovec *local_iov = g_new(struct iovec, niov); 231 struct iovec *local_iov_head = local_iov; 232 unsigned int nlocal_iov = niov; 233 234 nlocal_iov = iov_copy(local_iov, nlocal_iov, 235 iov, niov, 236 0, iov_size(iov, niov)); 237 238 while (nlocal_iov > 0) { 239 ssize_t len; 240 len = qio_channel_writev_full(ioc, local_iov, nlocal_iov, fds, nfds, 241 errp); 242 if (len == QIO_CHANNEL_ERR_BLOCK) { 243 if (qemu_in_coroutine()) { 244 qio_channel_yield(ioc, G_IO_OUT); 245 } else { 246 qio_channel_wait(ioc, G_IO_OUT); 247 } 248 continue; 249 } 250 if (len < 0) { 251 goto cleanup; 252 } 253 254 iov_discard_front(&local_iov, &nlocal_iov, len); 255 256 fds = NULL; 257 nfds = 0; 258 } 259 260 ret = 0; 261 cleanup: 262 g_free(local_iov_head); 263 return ret; 264 } 265 266 ssize_t qio_channel_readv(QIOChannel *ioc, 267 const struct iovec *iov, 268 size_t niov, 269 Error **errp) 270 { 271 return qio_channel_readv_full(ioc, iov, niov, NULL, NULL, errp); 272 } 273 274 275 ssize_t qio_channel_writev(QIOChannel *ioc, 276 const struct iovec *iov, 277 size_t niov, 278 Error **errp) 279 { 280 return qio_channel_writev_full(ioc, iov, niov, NULL, 0, errp); 281 } 282 283 284 ssize_t qio_channel_read(QIOChannel *ioc, 285 char *buf, 286 size_t buflen, 287 Error **errp) 288 { 289 struct iovec iov = { .iov_base = buf, .iov_len = buflen }; 290 return qio_channel_readv_full(ioc, &iov, 1, NULL, NULL, errp); 291 } 292 293 294 ssize_t qio_channel_write(QIOChannel *ioc, 295 const char *buf, 296 size_t buflen, 297 Error **errp) 298 { 299 struct iovec iov = { .iov_base = (char *)buf, .iov_len = buflen }; 300 return qio_channel_writev_full(ioc, &iov, 1, NULL, 0, errp); 301 } 302 303 304 int qio_channel_read_all_eof(QIOChannel *ioc, 305 char *buf, 306 size_t buflen, 307 Error **errp) 308 { 309 struct iovec iov = { .iov_base = buf, .iov_len = buflen }; 310 return qio_channel_readv_all_eof(ioc, &iov, 1, errp); 311 } 312 313 314 int qio_channel_read_all(QIOChannel *ioc, 315 char *buf, 316 size_t buflen, 317 Error **errp) 318 { 319 struct iovec iov = { .iov_base = buf, .iov_len = buflen }; 320 return qio_channel_readv_all(ioc, &iov, 1, errp); 321 } 322 323 324 int qio_channel_write_all(QIOChannel *ioc, 325 const char *buf, 326 size_t buflen, 327 Error **errp) 328 { 329 struct iovec iov = { .iov_base = (char *)buf, .iov_len = buflen }; 330 return qio_channel_writev_all(ioc, &iov, 1, errp); 331 } 332 333 334 int qio_channel_set_blocking(QIOChannel *ioc, 335 bool enabled, 336 Error **errp) 337 { 338 QIOChannelClass *klass = QIO_CHANNEL_GET_CLASS(ioc); 339 return klass->io_set_blocking(ioc, enabled, errp); 340 } 341 342 343 int qio_channel_close(QIOChannel *ioc, 344 Error **errp) 345 { 346 QIOChannelClass *klass = QIO_CHANNEL_GET_CLASS(ioc); 347 return klass->io_close(ioc, errp); 348 } 349 350 351 GSource *qio_channel_create_watch(QIOChannel *ioc, 352 GIOCondition condition) 353 { 354 QIOChannelClass *klass = QIO_CHANNEL_GET_CLASS(ioc); 355 GSource *ret = klass->io_create_watch(ioc, condition); 356 357 if (ioc->name) { 358 g_source_set_name(ret, ioc->name); 359 } 360 361 return ret; 362 } 363 364 365 void qio_channel_set_aio_fd_handler(QIOChannel *ioc, 366 AioContext *ctx, 367 IOHandler *io_read, 368 IOHandler *io_write, 369 void *opaque) 370 { 371 QIOChannelClass *klass = QIO_CHANNEL_GET_CLASS(ioc); 372 373 klass->io_set_aio_fd_handler(ioc, ctx, io_read, io_write, opaque); 374 } 375 376 guint qio_channel_add_watch_full(QIOChannel *ioc, 377 GIOCondition condition, 378 QIOChannelFunc func, 379 gpointer user_data, 380 GDestroyNotify notify, 381 GMainContext *context) 382 { 383 GSource *source; 384 guint id; 385 386 source = qio_channel_create_watch(ioc, condition); 387 388 g_source_set_callback(source, (GSourceFunc)func, user_data, notify); 389 390 id = g_source_attach(source, context); 391 g_source_unref(source); 392 393 return id; 394 } 395 396 guint qio_channel_add_watch(QIOChannel *ioc, 397 GIOCondition condition, 398 QIOChannelFunc func, 399 gpointer user_data, 400 GDestroyNotify notify) 401 { 402 return qio_channel_add_watch_full(ioc, condition, func, 403 user_data, notify, NULL); 404 } 405 406 GSource *qio_channel_add_watch_source(QIOChannel *ioc, 407 GIOCondition condition, 408 QIOChannelFunc func, 409 gpointer user_data, 410 GDestroyNotify notify, 411 GMainContext *context) 412 { 413 GSource *source; 414 guint id; 415 416 id = qio_channel_add_watch_full(ioc, condition, func, 417 user_data, notify, context); 418 source = g_main_context_find_source_by_id(context, id); 419 g_source_ref(source); 420 return source; 421 } 422 423 424 int qio_channel_shutdown(QIOChannel *ioc, 425 QIOChannelShutdown how, 426 Error **errp) 427 { 428 QIOChannelClass *klass = QIO_CHANNEL_GET_CLASS(ioc); 429 430 if (!klass->io_shutdown) { 431 error_setg(errp, "Data path shutdown not supported"); 432 return -1; 433 } 434 435 return klass->io_shutdown(ioc, how, errp); 436 } 437 438 439 void qio_channel_set_delay(QIOChannel *ioc, 440 bool enabled) 441 { 442 QIOChannelClass *klass = QIO_CHANNEL_GET_CLASS(ioc); 443 444 if (klass->io_set_delay) { 445 klass->io_set_delay(ioc, enabled); 446 } 447 } 448 449 450 void qio_channel_set_cork(QIOChannel *ioc, 451 bool enabled) 452 { 453 QIOChannelClass *klass = QIO_CHANNEL_GET_CLASS(ioc); 454 455 if (klass->io_set_cork) { 456 klass->io_set_cork(ioc, enabled); 457 } 458 } 459 460 461 off_t qio_channel_io_seek(QIOChannel *ioc, 462 off_t offset, 463 int whence, 464 Error **errp) 465 { 466 QIOChannelClass *klass = QIO_CHANNEL_GET_CLASS(ioc); 467 468 if (!klass->io_seek) { 469 error_setg(errp, "Channel does not support random access"); 470 return -1; 471 } 472 473 return klass->io_seek(ioc, offset, whence, errp); 474 } 475 476 477 static void qio_channel_restart_read(void *opaque) 478 { 479 QIOChannel *ioc = opaque; 480 Coroutine *co = ioc->read_coroutine; 481 482 /* Assert that aio_co_wake() reenters the coroutine directly */ 483 assert(qemu_get_current_aio_context() == 484 qemu_coroutine_get_aio_context(co)); 485 aio_co_wake(co); 486 } 487 488 static void qio_channel_restart_write(void *opaque) 489 { 490 QIOChannel *ioc = opaque; 491 Coroutine *co = ioc->write_coroutine; 492 493 /* Assert that aio_co_wake() reenters the coroutine directly */ 494 assert(qemu_get_current_aio_context() == 495 qemu_coroutine_get_aio_context(co)); 496 aio_co_wake(co); 497 } 498 499 static void qio_channel_set_aio_fd_handlers(QIOChannel *ioc) 500 { 501 IOHandler *rd_handler = NULL, *wr_handler = NULL; 502 AioContext *ctx; 503 504 if (ioc->read_coroutine) { 505 rd_handler = qio_channel_restart_read; 506 } 507 if (ioc->write_coroutine) { 508 wr_handler = qio_channel_restart_write; 509 } 510 511 ctx = ioc->ctx ? ioc->ctx : iohandler_get_aio_context(); 512 qio_channel_set_aio_fd_handler(ioc, ctx, rd_handler, wr_handler, ioc); 513 } 514 515 void qio_channel_attach_aio_context(QIOChannel *ioc, 516 AioContext *ctx) 517 { 518 assert(!ioc->read_coroutine); 519 assert(!ioc->write_coroutine); 520 ioc->ctx = ctx; 521 } 522 523 void qio_channel_detach_aio_context(QIOChannel *ioc) 524 { 525 ioc->read_coroutine = NULL; 526 ioc->write_coroutine = NULL; 527 qio_channel_set_aio_fd_handlers(ioc); 528 ioc->ctx = NULL; 529 } 530 531 void coroutine_fn qio_channel_yield(QIOChannel *ioc, 532 GIOCondition condition) 533 { 534 assert(qemu_in_coroutine()); 535 if (condition == G_IO_IN) { 536 assert(!ioc->read_coroutine); 537 ioc->read_coroutine = qemu_coroutine_self(); 538 } else if (condition == G_IO_OUT) { 539 assert(!ioc->write_coroutine); 540 ioc->write_coroutine = qemu_coroutine_self(); 541 } else { 542 abort(); 543 } 544 qio_channel_set_aio_fd_handlers(ioc); 545 qemu_coroutine_yield(); 546 547 /* Allow interrupting the operation by reentering the coroutine other than 548 * through the aio_fd_handlers. */ 549 if (condition == G_IO_IN && ioc->read_coroutine) { 550 ioc->read_coroutine = NULL; 551 qio_channel_set_aio_fd_handlers(ioc); 552 } else if (condition == G_IO_OUT && ioc->write_coroutine) { 553 ioc->write_coroutine = NULL; 554 qio_channel_set_aio_fd_handlers(ioc); 555 } 556 } 557 558 559 static gboolean qio_channel_wait_complete(QIOChannel *ioc, 560 GIOCondition condition, 561 gpointer opaque) 562 { 563 GMainLoop *loop = opaque; 564 565 g_main_loop_quit(loop); 566 return FALSE; 567 } 568 569 570 void qio_channel_wait(QIOChannel *ioc, 571 GIOCondition condition) 572 { 573 GMainContext *ctxt = g_main_context_new(); 574 GMainLoop *loop = g_main_loop_new(ctxt, TRUE); 575 GSource *source; 576 577 source = qio_channel_create_watch(ioc, condition); 578 579 g_source_set_callback(source, 580 (GSourceFunc)qio_channel_wait_complete, 581 loop, 582 NULL); 583 584 g_source_attach(source, ctxt); 585 586 g_main_loop_run(loop); 587 588 g_source_unref(source); 589 g_main_loop_unref(loop); 590 g_main_context_unref(ctxt); 591 } 592 593 594 static void qio_channel_finalize(Object *obj) 595 { 596 QIOChannel *ioc = QIO_CHANNEL(obj); 597 598 g_free(ioc->name); 599 600 #ifdef _WIN32 601 if (ioc->event) { 602 CloseHandle(ioc->event); 603 } 604 #endif 605 } 606 607 static const TypeInfo qio_channel_info = { 608 .parent = TYPE_OBJECT, 609 .name = TYPE_QIO_CHANNEL, 610 .instance_size = sizeof(QIOChannel), 611 .instance_finalize = qio_channel_finalize, 612 .abstract = true, 613 .class_size = sizeof(QIOChannelClass), 614 }; 615 616 617 static void qio_channel_register_types(void) 618 { 619 type_register_static(&qio_channel_info); 620 } 621 622 623 type_init(qio_channel_register_types); 624