mirror of
				https://github.com/qemu/qemu.git
				synced 2025-10-30 19:15:42 +00:00 
			
		
		
		
	 06831001ac
			
		
	
	
		06831001ac
		
	
	
	
	
		
			
			qatomic_mb_read and qatomic_mb_set were the very first atomic primitives introduced for QEMU; their semantics are unclear and they provide a false sense of safety. The last use of qatomic_mb_read() has been removed, so delete it. qatomic_mb_set() instead can survive as an optimized qatomic_set()+smp_mb(), similar to Linux's smp_store_mb(), but rename it to qatomic_set_mb() to match the order of the two operations. Reviewed-by: Richard Henderson <richard.henderson@linaro.org> Signed-off-by: Paolo Bonzini <pbonzini@redhat.com>
		
			
				
	
	
		
			573 lines
		
	
	
		
			18 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			573 lines
		
	
	
		
			18 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| /*
 | |
|  * QEMU monitor
 | |
|  *
 | |
|  * Copyright (c) 2003-2004 Fabrice Bellard
 | |
|  *
 | |
|  * Permission is hereby granted, free of charge, to any person obtaining a copy
 | |
|  * of this software and associated documentation files (the "Software"), to deal
 | |
|  * in the Software without restriction, including without limitation the rights
 | |
|  * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
 | |
|  * copies of the Software, and to permit persons to whom the Software is
 | |
|  * furnished to do so, subject to the following conditions:
 | |
|  *
 | |
|  * The above copyright notice and this permission notice shall be included in
 | |
|  * all copies or substantial portions of the Software.
 | |
|  *
 | |
|  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
 | |
|  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
 | |
|  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
 | |
|  * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
 | |
|  * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
 | |
|  * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
 | |
|  * THE SOFTWARE.
 | |
|  */
 | |
| 
 | |
| #include "qemu/osdep.h"
 | |
| 
 | |
| #include "chardev/char-io.h"
 | |
| #include "monitor-internal.h"
 | |
| #include "qapi/error.h"
 | |
| #include "qapi/qapi-commands-control.h"
 | |
| #include "qapi/qmp/qdict.h"
 | |
| #include "qapi/qmp/qjson.h"
 | |
| #include "qapi/qmp/qlist.h"
 | |
| #include "trace.h"
 | |
| 
 | |
| /*
 | |
|  * qmp_dispatcher_co_busy is used for synchronisation between the
 | |
|  * monitor thread and the main thread to ensure that the dispatcher
 | |
|  * coroutine never gets scheduled a second time when it's already
 | |
|  * scheduled (scheduling the same coroutine twice is forbidden).
 | |
|  *
 | |
|  * It is true if the coroutine will process at least one more request
 | |
|  * before going to sleep.  Either it has been kicked already, or it
 | |
|  * is active and processing requests.  Additional requests may therefore
 | |
|  * be pushed onto mon->qmp_requests, and @qmp_dispatcher_co_shutdown may
 | |
|  * be set without further ado.  @qmp_dispatcher_co must not be woken up
 | |
|  * in this case.
 | |
|  *
 | |
|  * If false, you have to wake up @qmp_dispatcher_co after pushing new
 | |
|  * requests. You also have to set @qmp_dispatcher_co_busy to true
 | |
|  * before waking up the coroutine.
 | |
|  *
 | |
|  * The coroutine will automatically change this variable back to false
 | |
|  * before it yields.  Nobody else may set the variable to false.
 | |
|  *
 | |
|  * Access must be atomic for thread safety.
 | |
|  */
 | |
| static bool qmp_dispatcher_co_busy = true;
 | |
| 
 | |
| struct QMPRequest {
 | |
|     /* Owner of the request */
 | |
|     MonitorQMP *mon;
 | |
|     /*
 | |
|      * Request object to be handled or Error to be reported
 | |
|      * (exactly one of them is non-null)
 | |
|      */
 | |
|     QObject *req;
 | |
|     Error *err;
 | |
| };
 | |
| typedef struct QMPRequest QMPRequest;
 | |
| 
 | |
| QmpCommandList qmp_commands, qmp_cap_negotiation_commands;
 | |
| 
 | |
| static bool qmp_oob_enabled(MonitorQMP *mon)
 | |
| {
 | |
|     return mon->capab[QMP_CAPABILITY_OOB];
 | |
| }
 | |
| 
 | |
| static void monitor_qmp_caps_reset(MonitorQMP *mon)
 | |
| {
 | |
|     memset(mon->capab_offered, 0, sizeof(mon->capab_offered));
 | |
|     memset(mon->capab, 0, sizeof(mon->capab));
 | |
|     mon->capab_offered[QMP_CAPABILITY_OOB] = mon->common.use_io_thread;
 | |
| }
 | |
| 
 | |
| static void qmp_request_free(QMPRequest *req)
 | |
| {
 | |
|     qobject_unref(req->req);
 | |
|     error_free(req->err);
 | |
|     g_free(req);
 | |
| }
 | |
| 
 | |
| /* Caller must hold mon->qmp.qmp_queue_lock */
 | |
| static void monitor_qmp_cleanup_req_queue_locked(MonitorQMP *mon)
 | |
| {
 | |
|     while (!g_queue_is_empty(mon->qmp_requests)) {
 | |
|         qmp_request_free(g_queue_pop_head(mon->qmp_requests));
 | |
|     }
 | |
| }
 | |
| 
 | |
| static void monitor_qmp_cleanup_queue_and_resume(MonitorQMP *mon)
 | |
| {
 | |
|     QEMU_LOCK_GUARD(&mon->qmp_queue_lock);
 | |
| 
 | |
|     /*
 | |
|      * Same condition as in monitor_qmp_dispatcher_co(), but before
 | |
|      * removing an element from the queue (hence no `- 1`).
 | |
|      * Also, the queue should not be empty either, otherwise the
 | |
|      * monitor hasn't been suspended yet (or was already resumed).
 | |
|      */
 | |
|     bool need_resume = (!qmp_oob_enabled(mon) ||
 | |
|         mon->qmp_requests->length == QMP_REQ_QUEUE_LEN_MAX)
 | |
|         && !g_queue_is_empty(mon->qmp_requests);
 | |
| 
 | |
|     monitor_qmp_cleanup_req_queue_locked(mon);
 | |
| 
 | |
|     if (need_resume) {
 | |
|         /*
 | |
|          * handle_qmp_command() suspended the monitor because the
 | |
|          * request queue filled up, to be resumed when the queue has
 | |
|          * space again.  We just emptied it; resume the monitor.
 | |
|          *
 | |
|          * Without this, the monitor would remain suspended forever
 | |
|          * when we get here while the monitor is suspended.  An
 | |
|          * unfortunately timed CHR_EVENT_CLOSED can do the trick.
 | |
|          */
 | |
|         monitor_resume(&mon->common);
 | |
|     }
 | |
| 
 | |
| }
 | |
| 
 | |
| void qmp_send_response(MonitorQMP *mon, const QDict *rsp)
 | |
| {
 | |
|     const QObject *data = QOBJECT(rsp);
 | |
|     GString *json;
 | |
| 
 | |
|     json = qobject_to_json_pretty(data, mon->pretty);
 | |
|     assert(json != NULL);
 | |
|     trace_monitor_qmp_respond(mon, json->str);
 | |
| 
 | |
|     g_string_append_c(json, '\n');
 | |
|     monitor_puts(&mon->common, json->str);
 | |
| 
 | |
|     g_string_free(json, true);
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Emit QMP response @rsp to @mon.
 | |
|  * Null @rsp can only happen for commands with QCO_NO_SUCCESS_RESP.
 | |
|  * Nothing is emitted then.
 | |
|  */
 | |
| static void monitor_qmp_respond(MonitorQMP *mon, QDict *rsp)
 | |
| {
 | |
|     if (rsp) {
 | |
|         qmp_send_response(mon, rsp);
 | |
|     }
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Runs outside of coroutine context for OOB commands, but in
 | |
|  * coroutine context for everything else.
 | |
|  */
 | |
| static void monitor_qmp_dispatch(MonitorQMP *mon, QObject *req)
 | |
| {
 | |
|     QDict *rsp;
 | |
|     QDict *error;
 | |
| 
 | |
|     rsp = qmp_dispatch(mon->commands, req, qmp_oob_enabled(mon),
 | |
|                        &mon->common);
 | |
| 
 | |
|     if (mon->commands == &qmp_cap_negotiation_commands) {
 | |
|         error = qdict_get_qdict(rsp, "error");
 | |
|         if (error
 | |
|             && !g_strcmp0(qdict_get_try_str(error, "class"),
 | |
|                     QapiErrorClass_str(ERROR_CLASS_COMMAND_NOT_FOUND))) {
 | |
|             /* Provide a more useful error message */
 | |
|             qdict_del(error, "desc");
 | |
|             qdict_put_str(error, "desc", "Expecting capabilities negotiation"
 | |
|                           " with 'qmp_capabilities'");
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     monitor_qmp_respond(mon, rsp);
 | |
|     qobject_unref(rsp);
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Pop a QMP request from a monitor request queue.
 | |
|  * Return the request, or NULL all request queues are empty.
 | |
|  * We are using round-robin fashion to pop the request, to avoid
 | |
|  * processing commands only on a very busy monitor.  To achieve that,
 | |
|  * when we process one request on a specific monitor, we put that
 | |
|  * monitor to the end of mon_list queue.
 | |
|  *
 | |
|  * Note: if the function returned with non-NULL, then the caller will
 | |
|  * be with qmp_mon->qmp_queue_lock held, and the caller is responsible
 | |
|  * to release it.
 | |
|  */
 | |
| static QMPRequest *monitor_qmp_requests_pop_any_with_lock(void)
 | |
| {
 | |
|     QMPRequest *req_obj = NULL;
 | |
|     Monitor *mon;
 | |
|     MonitorQMP *qmp_mon;
 | |
| 
 | |
|     QTAILQ_FOREACH(mon, &mon_list, entry) {
 | |
|         if (!monitor_is_qmp(mon)) {
 | |
|             continue;
 | |
|         }
 | |
| 
 | |
|         qmp_mon = container_of(mon, MonitorQMP, common);
 | |
|         qemu_mutex_lock(&qmp_mon->qmp_queue_lock);
 | |
|         req_obj = g_queue_pop_head(qmp_mon->qmp_requests);
 | |
|         if (req_obj) {
 | |
|             /* With the lock of corresponding queue held */
 | |
|             break;
 | |
|         }
 | |
|         qemu_mutex_unlock(&qmp_mon->qmp_queue_lock);
 | |
|     }
 | |
| 
 | |
|     if (req_obj) {
 | |
|         /*
 | |
|          * We found one request on the monitor. Degrade this monitor's
 | |
|          * priority to lowest by re-inserting it to end of queue.
 | |
|          */
 | |
|         QTAILQ_REMOVE(&mon_list, mon, entry);
 | |
|         QTAILQ_INSERT_TAIL(&mon_list, mon, entry);
 | |
|     }
 | |
| 
 | |
|     return req_obj;
 | |
| }
 | |
| 
 | |
| static QMPRequest *monitor_qmp_dispatcher_pop_any(void)
 | |
| {
 | |
|     while (true) {
 | |
|         /*
 | |
|          * To avoid double scheduling, busy is true on entry to
 | |
|          * monitor_qmp_dispatcher_co(), and must be set again before
 | |
|          * aio_co_wake()-ing it.
 | |
|          */
 | |
|         assert(qatomic_read(&qmp_dispatcher_co_busy) == true);
 | |
| 
 | |
|         /*
 | |
|          * Mark the dispatcher as not busy already here so that we
 | |
|          * don't miss any new requests coming in the middle of our
 | |
|          * processing.
 | |
|          *
 | |
|          * Clear qmp_dispatcher_co_busy before reading request.
 | |
|          */
 | |
|         qatomic_set_mb(&qmp_dispatcher_co_busy, false);
 | |
| 
 | |
|         WITH_QEMU_LOCK_GUARD(&monitor_lock) {
 | |
|             QMPRequest *req_obj;
 | |
| 
 | |
|             /* On shutdown, don't take any more requests from the queue */
 | |
|             if (qmp_dispatcher_co_shutdown) {
 | |
|                 return NULL;
 | |
|             }
 | |
| 
 | |
|             req_obj = monitor_qmp_requests_pop_any_with_lock();
 | |
|             if (req_obj) {
 | |
|                 return req_obj;
 | |
|             }
 | |
|         }
 | |
| 
 | |
|         /*
 | |
|          * No more requests to process.  Wait to be reentered from
 | |
|          * handle_qmp_command() when it pushes more requests, or
 | |
|          * from monitor_cleanup() when it requests shutdown.
 | |
|          */
 | |
|         qemu_coroutine_yield();
 | |
|     }
 | |
| }
 | |
| 
 | |
| void coroutine_fn monitor_qmp_dispatcher_co(void *data)
 | |
| {
 | |
|     QMPRequest *req_obj;
 | |
|     QDict *rsp;
 | |
|     bool oob_enabled;
 | |
|     MonitorQMP *mon;
 | |
| 
 | |
|     while ((req_obj = monitor_qmp_dispatcher_pop_any()) != NULL) {
 | |
|         trace_monitor_qmp_in_band_dequeue(req_obj,
 | |
|                                           req_obj->mon->qmp_requests->length);
 | |
| 
 | |
|         /*
 | |
|          * @req_obj has a request, we hold req_obj->mon->qmp_queue_lock
 | |
|          */
 | |
| 
 | |
|         mon = req_obj->mon;
 | |
| 
 | |
|         /*
 | |
|          * We need to resume the monitor if handle_qmp_command()
 | |
|          * suspended it.  Two cases:
 | |
|          * 1. OOB enabled: mon->qmp_requests has no more space
 | |
|          *    Resume right away, so that OOB commands can get executed while
 | |
|          *    this request is being processed.
 | |
|          * 2. OOB disabled: always
 | |
|          *    Resume only after we're done processing the request,
 | |
|          * We need to save qmp_oob_enabled() for later, because
 | |
|          * qmp_qmp_capabilities() can change it.
 | |
|          */
 | |
|         oob_enabled = qmp_oob_enabled(mon);
 | |
|         if (oob_enabled
 | |
|             && mon->qmp_requests->length == QMP_REQ_QUEUE_LEN_MAX - 1) {
 | |
|             monitor_resume(&mon->common);
 | |
|         }
 | |
| 
 | |
|         /*
 | |
|          * Drop the queue mutex now, before yielding, otherwise we might
 | |
|          * deadlock if the main thread tries to lock it.
 | |
|          */
 | |
|         qemu_mutex_unlock(&mon->qmp_queue_lock);
 | |
| 
 | |
|         if (qatomic_xchg(&qmp_dispatcher_co_busy, true) == true) {
 | |
|             /*
 | |
|              * Someone rescheduled us (probably because a new requests
 | |
|              * came in), but we didn't actually yield. Do that now,
 | |
|              * only to be immediately reentered and removed from the
 | |
|              * list of scheduled coroutines.
 | |
|              */
 | |
|             qemu_coroutine_yield();
 | |
|         }
 | |
| 
 | |
|         /*
 | |
|          * Move the coroutine from iohandler_ctx to qemu_aio_context for
 | |
|          * executing the command handler so that it can make progress if it
 | |
|          * involves an AIO_WAIT_WHILE().
 | |
|          */
 | |
|         aio_co_schedule(qemu_get_aio_context(), qmp_dispatcher_co);
 | |
|         qemu_coroutine_yield();
 | |
| 
 | |
|         /* Process request */
 | |
|         if (req_obj->req) {
 | |
|             if (trace_event_get_state(TRACE_MONITOR_QMP_CMD_IN_BAND)) {
 | |
|                 QDict *qdict = qobject_to(QDict, req_obj->req);
 | |
|                 QObject *id = qdict ? qdict_get(qdict, "id") : NULL;
 | |
|                 GString *id_json;
 | |
| 
 | |
|                 id_json = id ? qobject_to_json(id) : g_string_new(NULL);
 | |
|                 trace_monitor_qmp_cmd_in_band(id_json->str);
 | |
|                 g_string_free(id_json, true);
 | |
|             }
 | |
|             monitor_qmp_dispatch(mon, req_obj->req);
 | |
|         } else {
 | |
|             assert(req_obj->err);
 | |
|             trace_monitor_qmp_err_in_band(error_get_pretty(req_obj->err));
 | |
|             rsp = qmp_error_response(req_obj->err);
 | |
|             req_obj->err = NULL;
 | |
|             monitor_qmp_respond(mon, rsp);
 | |
|             qobject_unref(rsp);
 | |
|         }
 | |
| 
 | |
|         if (!oob_enabled) {
 | |
|             monitor_resume(&mon->common);
 | |
|         }
 | |
| 
 | |
|         qmp_request_free(req_obj);
 | |
| 
 | |
|         /*
 | |
|          * Yield and reschedule so the main loop stays responsive.
 | |
|          *
 | |
|          * Move back to iohandler_ctx so that nested event loops for
 | |
|          * qemu_aio_context don't start new monitor commands.
 | |
|          */
 | |
|         aio_co_schedule(iohandler_get_aio_context(), qmp_dispatcher_co);
 | |
|         qemu_coroutine_yield();
 | |
|     }
 | |
|     qatomic_set(&qmp_dispatcher_co, NULL);
 | |
| }
 | |
| 
 | |
| void qmp_dispatcher_co_wake(void)
 | |
| {
 | |
|     /* Write request before reading qmp_dispatcher_co_busy.  */
 | |
|     smp_mb__before_rmw();
 | |
| 
 | |
|     if (!qatomic_xchg(&qmp_dispatcher_co_busy, true)) {
 | |
|         aio_co_wake(qmp_dispatcher_co);
 | |
|     }
 | |
| }
 | |
| 
 | |
| static void handle_qmp_command(void *opaque, QObject *req, Error *err)
 | |
| {
 | |
|     MonitorQMP *mon = opaque;
 | |
|     QDict *qdict = qobject_to(QDict, req);
 | |
|     QMPRequest *req_obj;
 | |
| 
 | |
|     assert(!req != !err);
 | |
| 
 | |
|     if (req && trace_event_get_state_backends(TRACE_HANDLE_QMP_COMMAND)) {
 | |
|         GString *req_json = qobject_to_json(req);
 | |
|         trace_handle_qmp_command(mon, req_json->str);
 | |
|         g_string_free(req_json, true);
 | |
|     }
 | |
| 
 | |
|     if (qdict && qmp_is_oob(qdict)) {
 | |
|         /* OOB commands are executed immediately */
 | |
|         if (trace_event_get_state(TRACE_MONITOR_QMP_CMD_OUT_OF_BAND)) {
 | |
|             QObject *id = qdict_get(qdict, "id");
 | |
|             GString *id_json;
 | |
| 
 | |
|             id_json = id ? qobject_to_json(id) : g_string_new(NULL);
 | |
|             trace_monitor_qmp_cmd_out_of_band(id_json->str);
 | |
|             g_string_free(id_json, true);
 | |
|         }
 | |
|         monitor_qmp_dispatch(mon, req);
 | |
|         qobject_unref(req);
 | |
|         return;
 | |
|     }
 | |
| 
 | |
|     req_obj = g_new0(QMPRequest, 1);
 | |
|     req_obj->mon = mon;
 | |
|     req_obj->req = req;
 | |
|     req_obj->err = err;
 | |
| 
 | |
|     /* Protect qmp_requests and fetching its length. */
 | |
|     WITH_QEMU_LOCK_GUARD(&mon->qmp_queue_lock) {
 | |
| 
 | |
|         /*
 | |
|          * Suspend the monitor when we can't queue more requests after
 | |
|          * this one.  Dequeuing in monitor_qmp_dispatcher_co() or
 | |
|          * monitor_qmp_cleanup_queue_and_resume() will resume it.
 | |
|          * Note that when OOB is disabled, we queue at most one command,
 | |
|          * for backward compatibility.
 | |
|          */
 | |
|         if (!qmp_oob_enabled(mon) ||
 | |
|             mon->qmp_requests->length == QMP_REQ_QUEUE_LEN_MAX - 1) {
 | |
|             monitor_suspend(&mon->common);
 | |
|         }
 | |
| 
 | |
|         /*
 | |
|          * Put the request to the end of queue so that requests will be
 | |
|          * handled in time order.  Ownership for req_obj, req,
 | |
|          * etc. will be delivered to the handler side.
 | |
|          */
 | |
|         trace_monitor_qmp_in_band_enqueue(req_obj, mon,
 | |
|                                           mon->qmp_requests->length);
 | |
|         assert(mon->qmp_requests->length < QMP_REQ_QUEUE_LEN_MAX);
 | |
|         g_queue_push_tail(mon->qmp_requests, req_obj);
 | |
|     }
 | |
| 
 | |
|     /* Kick the dispatcher routine */
 | |
|     qmp_dispatcher_co_wake();
 | |
| }
 | |
| 
 | |
| static void monitor_qmp_read(void *opaque, const uint8_t *buf, int size)
 | |
| {
 | |
|     MonitorQMP *mon = opaque;
 | |
| 
 | |
|     json_message_parser_feed(&mon->parser, (const char *) buf, size);
 | |
| }
 | |
| 
 | |
| static QDict *qmp_greeting(MonitorQMP *mon)
 | |
| {
 | |
|     QList *cap_list = qlist_new();
 | |
|     QObject *ver = NULL;
 | |
|     QDict *args;
 | |
|     QMPCapability cap;
 | |
| 
 | |
|     args = qdict_new();
 | |
|     qmp_marshal_query_version(args, &ver, NULL);
 | |
|     qobject_unref(args);
 | |
| 
 | |
|     for (cap = 0; cap < QMP_CAPABILITY__MAX; cap++) {
 | |
|         if (mon->capab_offered[cap]) {
 | |
|             qlist_append_str(cap_list, QMPCapability_str(cap));
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     return qdict_from_jsonf_nofail(
 | |
|         "{'QMP': {'version': %p, 'capabilities': %p}}",
 | |
|         ver, cap_list);
 | |
| }
 | |
| 
 | |
| static void monitor_qmp_event(void *opaque, QEMUChrEvent event)
 | |
| {
 | |
|     QDict *data;
 | |
|     MonitorQMP *mon = opaque;
 | |
| 
 | |
|     switch (event) {
 | |
|     case CHR_EVENT_OPENED:
 | |
|         mon->commands = &qmp_cap_negotiation_commands;
 | |
|         monitor_qmp_caps_reset(mon);
 | |
|         data = qmp_greeting(mon);
 | |
|         qmp_send_response(mon, data);
 | |
|         qobject_unref(data);
 | |
|         mon_refcount++;
 | |
|         break;
 | |
|     case CHR_EVENT_CLOSED:
 | |
|         /*
 | |
|          * Note: this is only useful when the output of the chardev
 | |
|          * backend is still open.  For example, when the backend is
 | |
|          * stdio, it's possible that stdout is still open when stdin
 | |
|          * is closed.
 | |
|          */
 | |
|         monitor_qmp_cleanup_queue_and_resume(mon);
 | |
|         json_message_parser_destroy(&mon->parser);
 | |
|         json_message_parser_init(&mon->parser, handle_qmp_command,
 | |
|                                  mon, NULL);
 | |
|         mon_refcount--;
 | |
|         monitor_fdsets_cleanup();
 | |
|         break;
 | |
|     case CHR_EVENT_BREAK:
 | |
|     case CHR_EVENT_MUX_IN:
 | |
|     case CHR_EVENT_MUX_OUT:
 | |
|         /* Ignore */
 | |
|         break;
 | |
|     }
 | |
| }
 | |
| 
 | |
| void monitor_data_destroy_qmp(MonitorQMP *mon)
 | |
| {
 | |
|     json_message_parser_destroy(&mon->parser);
 | |
|     qemu_mutex_destroy(&mon->qmp_queue_lock);
 | |
|     monitor_qmp_cleanup_req_queue_locked(mon);
 | |
|     g_queue_free(mon->qmp_requests);
 | |
| }
 | |
| 
 | |
| static void monitor_qmp_setup_handlers_bh(void *opaque)
 | |
| {
 | |
|     MonitorQMP *mon = opaque;
 | |
|     GMainContext *context;
 | |
| 
 | |
|     assert(mon->common.use_io_thread);
 | |
|     context = iothread_get_g_main_context(mon_iothread);
 | |
|     assert(context);
 | |
|     qemu_chr_fe_set_handlers(&mon->common.chr, monitor_can_read,
 | |
|                              monitor_qmp_read, monitor_qmp_event,
 | |
|                              NULL, &mon->common, context, true);
 | |
|     monitor_list_append(&mon->common);
 | |
| }
 | |
| 
 | |
| void monitor_init_qmp(Chardev *chr, bool pretty, Error **errp)
 | |
| {
 | |
|     MonitorQMP *mon = g_new0(MonitorQMP, 1);
 | |
| 
 | |
|     if (!qemu_chr_fe_init(&mon->common.chr, chr, errp)) {
 | |
|         g_free(mon);
 | |
|         return;
 | |
|     }
 | |
|     qemu_chr_fe_set_echo(&mon->common.chr, true);
 | |
| 
 | |
|     /* Note: we run QMP monitor in I/O thread when @chr supports that */
 | |
|     monitor_data_init(&mon->common, true, false,
 | |
|                       qemu_chr_has_feature(chr, QEMU_CHAR_FEATURE_GCONTEXT));
 | |
| 
 | |
|     mon->pretty = pretty;
 | |
| 
 | |
|     qemu_mutex_init(&mon->qmp_queue_lock);
 | |
|     mon->qmp_requests = g_queue_new();
 | |
| 
 | |
|     json_message_parser_init(&mon->parser, handle_qmp_command, mon, NULL);
 | |
|     if (mon->common.use_io_thread) {
 | |
|         /*
 | |
|          * Make sure the old iowatch is gone.  It's possible when
 | |
|          * e.g. the chardev is in client mode, with wait=on.
 | |
|          */
 | |
|         remove_fd_in_watch(chr);
 | |
|         /*
 | |
|          * We can't call qemu_chr_fe_set_handlers() directly here
 | |
|          * since chardev might be running in the monitor I/O
 | |
|          * thread.  Schedule a bottom half.
 | |
|          */
 | |
|         aio_bh_schedule_oneshot(iothread_get_aio_context(mon_iothread),
 | |
|                                 monitor_qmp_setup_handlers_bh, mon);
 | |
|         /* The bottom half will add @mon to @mon_list */
 | |
|     } else {
 | |
|         qemu_chr_fe_set_handlers(&mon->common.chr, monitor_can_read,
 | |
|                                  monitor_qmp_read, monitor_qmp_event,
 | |
|                                  NULL, &mon->common, NULL, true);
 | |
|         monitor_list_append(&mon->common);
 | |
|     }
 | |
| }
 |