xref: /openbmc/qemu/monitor/monitor.c (revision ebe15582)
1 /*
2  * QEMU monitor
3  *
4  * Copyright (c) 2003-2004 Fabrice Bellard
5  *
6  * Permission is hereby granted, free of charge, to any person obtaining a copy
7  * of this software and associated documentation files (the "Software"), to deal
8  * in the Software without restriction, including without limitation the rights
9  * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
10  * copies of the Software, and to permit persons to whom the Software is
11  * furnished to do so, subject to the following conditions:
12  *
13  * The above copyright notice and this permission notice shall be included in
14  * all copies or substantial portions of the Software.
15  *
16  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
17  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
18  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
19  * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
20  * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
21  * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
22  * THE SOFTWARE.
23  */
24 
25 #include "qemu/osdep.h"
26 #include "monitor-internal.h"
27 #include "qapi/error.h"
28 #include "qapi/qapi-emit-events.h"
29 #include "qapi/qmp/qdict.h"
30 #include "qapi/qmp/qstring.h"
31 #include "qemu/error-report.h"
32 #include "qemu/option.h"
33 #include "sysemu/qtest.h"
34 #include "sysemu/sysemu.h"
35 #include "trace.h"
36 
37 /*
38  * To prevent flooding clients, events can be throttled. The
39  * throttling is calculated globally, rather than per-Monitor
40  * instance.
41  */
42 typedef struct MonitorQAPIEventState {
43     QAPIEvent event;    /* Throttling state for this event type and... */
44     QDict *data;        /* ... data, see qapi_event_throttle_equal() */
45     QEMUTimer *timer;   /* Timer for handling delayed events */
46     QDict *qdict;       /* Delayed event (if any) */
47 } MonitorQAPIEventState;
48 
49 typedef struct {
50     int64_t rate;       /* Minimum time (in ns) between two events */
51 } MonitorQAPIEventConf;
52 
53 /* Shared monitor I/O thread */
54 IOThread *mon_iothread;
55 
56 /* Bottom half to dispatch the requests received from I/O thread */
57 QEMUBH *qmp_dispatcher_bh;
58 
59 /* Protects mon_list, monitor_qapi_event_state, monitor_destroyed.  */
60 QemuMutex monitor_lock;
61 static GHashTable *monitor_qapi_event_state;
62 
63 MonitorList mon_list;
64 int mon_refcount;
65 static bool monitor_destroyed;
66 
67 __thread Monitor *cur_mon;
68 
69 /**
70  * Is the current monitor, if any, a QMP monitor?
71  */
72 bool monitor_cur_is_qmp(void)
73 {
74     return cur_mon && monitor_is_qmp(cur_mon);
75 }
76 
77 /**
78  * Is @mon is using readline?
79  * Note: not all HMP monitors use readline, e.g., gdbserver has a
80  * non-interactive HMP monitor, so readline is not used there.
81  */
82 static inline bool monitor_uses_readline(const MonitorHMP *mon)
83 {
84     return mon->use_readline;
85 }
86 
87 static inline bool monitor_is_hmp_non_interactive(const Monitor *mon)
88 {
89     if (monitor_is_qmp(mon)) {
90         return false;
91     }
92 
93     return !monitor_uses_readline(container_of(mon, MonitorHMP, common));
94 }
95 
96 static void monitor_flush_locked(Monitor *mon);
97 
98 static gboolean monitor_unblocked(GIOChannel *chan, GIOCondition cond,
99                                   void *opaque)
100 {
101     Monitor *mon = opaque;
102 
103     qemu_mutex_lock(&mon->mon_lock);
104     mon->out_watch = 0;
105     monitor_flush_locked(mon);
106     qemu_mutex_unlock(&mon->mon_lock);
107     return FALSE;
108 }
109 
110 /* Caller must hold mon->mon_lock */
111 static void monitor_flush_locked(Monitor *mon)
112 {
113     int rc;
114     size_t len;
115     const char *buf;
116 
117     if (mon->skip_flush) {
118         return;
119     }
120 
121     buf = qstring_get_str(mon->outbuf);
122     len = qstring_get_length(mon->outbuf);
123 
124     if (len && !mon->mux_out) {
125         rc = qemu_chr_fe_write(&mon->chr, (const uint8_t *) buf, len);
126         if ((rc < 0 && errno != EAGAIN) || (rc == len)) {
127             /* all flushed or error */
128             qobject_unref(mon->outbuf);
129             mon->outbuf = qstring_new();
130             return;
131         }
132         if (rc > 0) {
133             /* partial write */
134             QString *tmp = qstring_from_str(buf + rc);
135             qobject_unref(mon->outbuf);
136             mon->outbuf = tmp;
137         }
138         if (mon->out_watch == 0) {
139             mon->out_watch =
140                 qemu_chr_fe_add_watch(&mon->chr, G_IO_OUT | G_IO_HUP,
141                                       monitor_unblocked, mon);
142         }
143     }
144 }
145 
146 void monitor_flush(Monitor *mon)
147 {
148     qemu_mutex_lock(&mon->mon_lock);
149     monitor_flush_locked(mon);
150     qemu_mutex_unlock(&mon->mon_lock);
151 }
152 
153 /* flush at every end of line */
154 int monitor_puts(Monitor *mon, const char *str)
155 {
156     int i;
157     char c;
158 
159     qemu_mutex_lock(&mon->mon_lock);
160     for (i = 0; str[i]; i++) {
161         c = str[i];
162         if (c == '\n') {
163             qstring_append_chr(mon->outbuf, '\r');
164         }
165         qstring_append_chr(mon->outbuf, c);
166         if (c == '\n') {
167             monitor_flush_locked(mon);
168         }
169     }
170     qemu_mutex_unlock(&mon->mon_lock);
171 
172     return i;
173 }
174 
175 int monitor_vprintf(Monitor *mon, const char *fmt, va_list ap)
176 {
177     char *buf;
178     int n;
179 
180     if (!mon) {
181         return -1;
182     }
183 
184     if (monitor_is_qmp(mon)) {
185         return -1;
186     }
187 
188     buf = g_strdup_vprintf(fmt, ap);
189     n = monitor_puts(mon, buf);
190     g_free(buf);
191     return n;
192 }
193 
194 int monitor_printf(Monitor *mon, const char *fmt, ...)
195 {
196     int ret;
197 
198     va_list ap;
199     va_start(ap, fmt);
200     ret = monitor_vprintf(mon, fmt, ap);
201     va_end(ap);
202     return ret;
203 }
204 
205 /*
206  * Print to current monitor if we have one, else to stderr.
207  */
208 int error_vprintf(const char *fmt, va_list ap)
209 {
210     if (cur_mon && !monitor_cur_is_qmp()) {
211         return monitor_vprintf(cur_mon, fmt, ap);
212     }
213     return vfprintf(stderr, fmt, ap);
214 }
215 
216 int error_vprintf_unless_qmp(const char *fmt, va_list ap)
217 {
218     if (!cur_mon) {
219         return vfprintf(stderr, fmt, ap);
220     }
221     if (!monitor_cur_is_qmp()) {
222         return monitor_vprintf(cur_mon, fmt, ap);
223     }
224     return -1;
225 }
226 
227 
228 static MonitorQAPIEventConf monitor_qapi_event_conf[QAPI_EVENT__MAX] = {
229     /* Limit guest-triggerable events to 1 per second */
230     [QAPI_EVENT_RTC_CHANGE]        = { 1000 * SCALE_MS },
231     [QAPI_EVENT_WATCHDOG]          = { 1000 * SCALE_MS },
232     [QAPI_EVENT_BALLOON_CHANGE]    = { 1000 * SCALE_MS },
233     [QAPI_EVENT_QUORUM_REPORT_BAD] = { 1000 * SCALE_MS },
234     [QAPI_EVENT_QUORUM_FAILURE]    = { 1000 * SCALE_MS },
235     [QAPI_EVENT_VSERPORT_CHANGE]   = { 1000 * SCALE_MS },
236 };
237 
238 /*
239  * Return the clock to use for recording an event's time.
240  * It's QEMU_CLOCK_REALTIME, except for qtests it's
241  * QEMU_CLOCK_VIRTUAL, to support testing rate limits.
242  * Beware: result is invalid before configure_accelerator().
243  */
244 static inline QEMUClockType monitor_get_event_clock(void)
245 {
246     return qtest_enabled() ? QEMU_CLOCK_VIRTUAL : QEMU_CLOCK_REALTIME;
247 }
248 
249 /*
250  * Broadcast an event to all monitors.
251  * @qdict is the event object.  Its member "event" must match @event.
252  * Caller must hold monitor_lock.
253  */
254 static void monitor_qapi_event_emit(QAPIEvent event, QDict *qdict)
255 {
256     Monitor *mon;
257     MonitorQMP *qmp_mon;
258 
259     trace_monitor_protocol_event_emit(event, qdict);
260     QTAILQ_FOREACH(mon, &mon_list, entry) {
261         if (!monitor_is_qmp(mon)) {
262             continue;
263         }
264 
265         qmp_mon = container_of(mon, MonitorQMP, common);
266         if (qmp_mon->commands != &qmp_cap_negotiation_commands) {
267             qmp_send_response(qmp_mon, qdict);
268         }
269     }
270 }
271 
272 static void monitor_qapi_event_handler(void *opaque);
273 
274 /*
275  * Queue a new event for emission to Monitor instances,
276  * applying any rate limiting if required.
277  */
278 static void
279 monitor_qapi_event_queue_no_reenter(QAPIEvent event, QDict *qdict)
280 {
281     MonitorQAPIEventConf *evconf;
282     MonitorQAPIEventState *evstate;
283 
284     assert(event < QAPI_EVENT__MAX);
285     evconf = &monitor_qapi_event_conf[event];
286     trace_monitor_protocol_event_queue(event, qdict, evconf->rate);
287 
288     qemu_mutex_lock(&monitor_lock);
289 
290     if (!evconf->rate) {
291         /* Unthrottled event */
292         monitor_qapi_event_emit(event, qdict);
293     } else {
294         QDict *data = qobject_to(QDict, qdict_get(qdict, "data"));
295         MonitorQAPIEventState key = { .event = event, .data = data };
296 
297         evstate = g_hash_table_lookup(monitor_qapi_event_state, &key);
298         assert(!evstate || timer_pending(evstate->timer));
299 
300         if (evstate) {
301             /*
302              * Timer is pending for (at least) evconf->rate ns after
303              * last send.  Store event for sending when timer fires,
304              * replacing a prior stored event if any.
305              */
306             qobject_unref(evstate->qdict);
307             evstate->qdict = qobject_ref(qdict);
308         } else {
309             /*
310              * Last send was (at least) evconf->rate ns ago.
311              * Send immediately, and arm the timer to call
312              * monitor_qapi_event_handler() in evconf->rate ns.  Any
313              * events arriving before then will be delayed until then.
314              */
315             int64_t now = qemu_clock_get_ns(monitor_get_event_clock());
316 
317             monitor_qapi_event_emit(event, qdict);
318 
319             evstate = g_new(MonitorQAPIEventState, 1);
320             evstate->event = event;
321             evstate->data = qobject_ref(data);
322             evstate->qdict = NULL;
323             evstate->timer = timer_new_ns(monitor_get_event_clock(),
324                                           monitor_qapi_event_handler,
325                                           evstate);
326             g_hash_table_add(monitor_qapi_event_state, evstate);
327             timer_mod_ns(evstate->timer, now + evconf->rate);
328         }
329     }
330 
331     qemu_mutex_unlock(&monitor_lock);
332 }
333 
334 void qapi_event_emit(QAPIEvent event, QDict *qdict)
335 {
336     /*
337      * monitor_qapi_event_queue_no_reenter() is not reentrant: it
338      * would deadlock on monitor_lock.  Work around by queueing
339      * events in thread-local storage.
340      * TODO: remove this, make it re-enter safe.
341      */
342     typedef struct MonitorQapiEvent {
343         QAPIEvent event;
344         QDict *qdict;
345         QSIMPLEQ_ENTRY(MonitorQapiEvent) entry;
346     } MonitorQapiEvent;
347     static __thread QSIMPLEQ_HEAD(, MonitorQapiEvent) event_queue;
348     static __thread bool reentered;
349     MonitorQapiEvent *ev;
350 
351     if (!reentered) {
352         QSIMPLEQ_INIT(&event_queue);
353     }
354 
355     ev = g_new(MonitorQapiEvent, 1);
356     ev->qdict = qobject_ref(qdict);
357     ev->event = event;
358     QSIMPLEQ_INSERT_TAIL(&event_queue, ev, entry);
359     if (reentered) {
360         return;
361     }
362 
363     reentered = true;
364 
365     while ((ev = QSIMPLEQ_FIRST(&event_queue)) != NULL) {
366         QSIMPLEQ_REMOVE_HEAD(&event_queue, entry);
367         monitor_qapi_event_queue_no_reenter(ev->event, ev->qdict);
368         qobject_unref(ev->qdict);
369         g_free(ev);
370     }
371 
372     reentered = false;
373 }
374 
375 /*
376  * This function runs evconf->rate ns after sending a throttled
377  * event.
378  * If another event has since been stored, send it.
379  */
380 static void monitor_qapi_event_handler(void *opaque)
381 {
382     MonitorQAPIEventState *evstate = opaque;
383     MonitorQAPIEventConf *evconf = &monitor_qapi_event_conf[evstate->event];
384 
385     trace_monitor_protocol_event_handler(evstate->event, evstate->qdict);
386     qemu_mutex_lock(&monitor_lock);
387 
388     if (evstate->qdict) {
389         int64_t now = qemu_clock_get_ns(monitor_get_event_clock());
390 
391         monitor_qapi_event_emit(evstate->event, evstate->qdict);
392         qobject_unref(evstate->qdict);
393         evstate->qdict = NULL;
394         timer_mod_ns(evstate->timer, now + evconf->rate);
395     } else {
396         g_hash_table_remove(monitor_qapi_event_state, evstate);
397         qobject_unref(evstate->data);
398         timer_free(evstate->timer);
399         g_free(evstate);
400     }
401 
402     qemu_mutex_unlock(&monitor_lock);
403 }
404 
405 static unsigned int qapi_event_throttle_hash(const void *key)
406 {
407     const MonitorQAPIEventState *evstate = key;
408     unsigned int hash = evstate->event * 255;
409 
410     if (evstate->event == QAPI_EVENT_VSERPORT_CHANGE) {
411         hash += g_str_hash(qdict_get_str(evstate->data, "id"));
412     }
413 
414     if (evstate->event == QAPI_EVENT_QUORUM_REPORT_BAD) {
415         hash += g_str_hash(qdict_get_str(evstate->data, "node-name"));
416     }
417 
418     return hash;
419 }
420 
421 static gboolean qapi_event_throttle_equal(const void *a, const void *b)
422 {
423     const MonitorQAPIEventState *eva = a;
424     const MonitorQAPIEventState *evb = b;
425 
426     if (eva->event != evb->event) {
427         return FALSE;
428     }
429 
430     if (eva->event == QAPI_EVENT_VSERPORT_CHANGE) {
431         return !strcmp(qdict_get_str(eva->data, "id"),
432                        qdict_get_str(evb->data, "id"));
433     }
434 
435     if (eva->event == QAPI_EVENT_QUORUM_REPORT_BAD) {
436         return !strcmp(qdict_get_str(eva->data, "node-name"),
437                        qdict_get_str(evb->data, "node-name"));
438     }
439 
440     return TRUE;
441 }
442 
443 int monitor_suspend(Monitor *mon)
444 {
445     if (monitor_is_hmp_non_interactive(mon)) {
446         return -ENOTTY;
447     }
448 
449     atomic_inc(&mon->suspend_cnt);
450 
451     if (mon->use_io_thread) {
452         /*
453          * Kick I/O thread to make sure this takes effect.  It'll be
454          * evaluated again in prepare() of the watch object.
455          */
456         aio_notify(iothread_get_aio_context(mon_iothread));
457     }
458 
459     trace_monitor_suspend(mon, 1);
460     return 0;
461 }
462 
463 static void monitor_accept_input(void *opaque)
464 {
465     Monitor *mon = opaque;
466 
467     qemu_chr_fe_accept_input(&mon->chr);
468 }
469 
470 void monitor_resume(Monitor *mon)
471 {
472     if (monitor_is_hmp_non_interactive(mon)) {
473         return;
474     }
475 
476     if (atomic_dec_fetch(&mon->suspend_cnt) == 0) {
477         AioContext *ctx;
478 
479         if (mon->use_io_thread) {
480             ctx = iothread_get_aio_context(mon_iothread);
481         } else {
482             ctx = qemu_get_aio_context();
483         }
484 
485         if (!monitor_is_qmp(mon)) {
486             MonitorHMP *hmp_mon = container_of(mon, MonitorHMP, common);
487             assert(hmp_mon->rs);
488             readline_show_prompt(hmp_mon->rs);
489         }
490 
491         aio_bh_schedule_oneshot(ctx, monitor_accept_input, mon);
492     }
493 
494     trace_monitor_suspend(mon, -1);
495 }
496 
497 int monitor_can_read(void *opaque)
498 {
499     Monitor *mon = opaque;
500 
501     return !atomic_mb_read(&mon->suspend_cnt);
502 }
503 
504 void monitor_list_append(Monitor *mon)
505 {
506     qemu_mutex_lock(&monitor_lock);
507     /*
508      * This prevents inserting new monitors during monitor_cleanup().
509      * A cleaner solution would involve the main thread telling other
510      * threads to terminate, waiting for their termination.
511      */
512     if (!monitor_destroyed) {
513         QTAILQ_INSERT_HEAD(&mon_list, mon, entry);
514         mon = NULL;
515     }
516     qemu_mutex_unlock(&monitor_lock);
517 
518     if (mon) {
519         monitor_data_destroy(mon);
520         g_free(mon);
521     }
522 }
523 
524 static void monitor_iothread_init(void)
525 {
526     mon_iothread = iothread_create("mon_iothread", &error_abort);
527 }
528 
529 void monitor_data_init(Monitor *mon, bool is_qmp, bool skip_flush,
530                        bool use_io_thread)
531 {
532     if (use_io_thread && !mon_iothread) {
533         monitor_iothread_init();
534     }
535     qemu_mutex_init(&mon->mon_lock);
536     mon->is_qmp = is_qmp;
537     mon->outbuf = qstring_new();
538     mon->skip_flush = skip_flush;
539     mon->use_io_thread = use_io_thread;
540 }
541 
542 void monitor_data_destroy(Monitor *mon)
543 {
544     g_free(mon->mon_cpu_path);
545     qemu_chr_fe_deinit(&mon->chr, false);
546     if (monitor_is_qmp(mon)) {
547         monitor_data_destroy_qmp(container_of(mon, MonitorQMP, common));
548     } else {
549         readline_free(container_of(mon, MonitorHMP, common)->rs);
550     }
551     qobject_unref(mon->outbuf);
552     qemu_mutex_destroy(&mon->mon_lock);
553 }
554 
555 void monitor_cleanup(void)
556 {
557     /*
558      * We need to explicitly stop the I/O thread (but not destroy it),
559      * clean up the monitor resources, then destroy the I/O thread since
560      * we need to unregister from chardev below in
561      * monitor_data_destroy(), and chardev is not thread-safe yet
562      */
563     if (mon_iothread) {
564         iothread_stop(mon_iothread);
565     }
566 
567     /* Flush output buffers and destroy monitors */
568     qemu_mutex_lock(&monitor_lock);
569     monitor_destroyed = true;
570     while (!QTAILQ_EMPTY(&mon_list)) {
571         Monitor *mon = QTAILQ_FIRST(&mon_list);
572         QTAILQ_REMOVE(&mon_list, mon, entry);
573         /* Permit QAPI event emission from character frontend release */
574         qemu_mutex_unlock(&monitor_lock);
575         monitor_flush(mon);
576         monitor_data_destroy(mon);
577         qemu_mutex_lock(&monitor_lock);
578         g_free(mon);
579     }
580     qemu_mutex_unlock(&monitor_lock);
581 
582     /* QEMUBHs needs to be deleted before destroying the I/O thread */
583     qemu_bh_delete(qmp_dispatcher_bh);
584     qmp_dispatcher_bh = NULL;
585     if (mon_iothread) {
586         iothread_destroy(mon_iothread);
587         mon_iothread = NULL;
588     }
589 }
590 
591 static void monitor_qapi_event_init(void)
592 {
593     monitor_qapi_event_state = g_hash_table_new(qapi_event_throttle_hash,
594                                                 qapi_event_throttle_equal);
595 }
596 
597 void monitor_init_globals_core(void)
598 {
599     monitor_qapi_event_init();
600     qemu_mutex_init(&monitor_lock);
601 
602     /*
603      * The dispatcher BH must run in the main loop thread, since we
604      * have commands assuming that context.  It would be nice to get
605      * rid of those assumptions.
606      */
607     qmp_dispatcher_bh = aio_bh_new(iohandler_get_aio_context(),
608                                    monitor_qmp_bh_dispatcher,
609                                    NULL);
610 }
611 
612 QemuOptsList qemu_mon_opts = {
613     .name = "mon",
614     .implied_opt_name = "chardev",
615     .head = QTAILQ_HEAD_INITIALIZER(qemu_mon_opts.head),
616     .desc = {
617         {
618             .name = "mode",
619             .type = QEMU_OPT_STRING,
620         },{
621             .name = "chardev",
622             .type = QEMU_OPT_STRING,
623         },{
624             .name = "pretty",
625             .type = QEMU_OPT_BOOL,
626         },
627         { /* end of list */ }
628     },
629 };
630