mirror of
https://gitlab.com/libvirt/libvirt.git
synced 2025-01-07 05:25:25 +00:00
0db5912617
With -blockdev it will be possible that a block job loses the disk that was used to start it to a guest-initiated hot-unplug. Don't emit the block job events in that case as we can't report the top level source or disk target for an unplugged (and potentially replugged with different source) disk. Eventually when we add machinery for tracking jobs globally for a VM the event will be reinstated via the domain job event. Signed-off-by: Peter Krempa <pkrempa@redhat.com> Reviewed-by: Ján Tomko <jtomko@redhat.com>
474 lines
13 KiB
C
474 lines
13 KiB
C
/*
|
|
* qemu_blockjob.c: helper functions for QEMU block jobs
|
|
*
|
|
* Copyright (C) 2006-2015 Red Hat, Inc.
|
|
* Copyright (C) 2006 Daniel P. Berrange
|
|
*
|
|
* This library is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2.1 of the License, or (at your option) any later version.
|
|
*
|
|
* This library is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
* License along with this library. If not, see
|
|
* <http://www.gnu.org/licenses/>.
|
|
*/
|
|
|
|
#include <config.h>
|
|
|
|
#include "internal.h"
|
|
|
|
#include "qemu_blockjob.h"
|
|
#include "qemu_block.h"
|
|
#include "qemu_domain.h"
|
|
|
|
#include "conf/domain_conf.h"
|
|
#include "conf/domain_event.h"
|
|
|
|
#include "virlog.h"
|
|
#include "virstoragefile.h"
|
|
#include "virthread.h"
|
|
#include "virtime.h"
|
|
#include "locking/domain_lock.h"
|
|
#include "viralloc.h"
|
|
#include "virstring.h"
|
|
#include "qemu_security.h"
|
|
|
|
#define VIR_FROM_THIS VIR_FROM_QEMU
|
|
|
|
VIR_LOG_INIT("qemu.qemu_blockjob");
|
|
|
|
|
|
static virClassPtr qemuBlockJobDataClass;
|
|
|
|
|
|
static void
|
|
qemuBlockJobDataDispose(void *obj)
|
|
{
|
|
qemuBlockJobDataPtr job = obj;
|
|
|
|
VIR_FREE(job->name);
|
|
VIR_FREE(job->errmsg);
|
|
}
|
|
|
|
|
|
static int
|
|
qemuBlockJobDataOnceInit(void)
|
|
{
|
|
if (!VIR_CLASS_NEW(qemuBlockJobData, virClassForObject()))
|
|
return -1;
|
|
|
|
return 0;
|
|
}
|
|
|
|
|
|
VIR_ONCE_GLOBAL_INIT(qemuBlockJobData);
|
|
|
|
static qemuBlockJobDataPtr
|
|
qemuBlockJobDataNew(qemuBlockJobType type,
|
|
const char *name)
|
|
{
|
|
qemuBlockJobDataPtr job = NULL;
|
|
qemuBlockJobDataPtr ret = NULL;
|
|
|
|
if (qemuBlockJobDataInitialize() < 0)
|
|
return NULL;
|
|
|
|
if (!(job = virObjectNew(qemuBlockJobDataClass)))
|
|
return NULL;
|
|
|
|
if (VIR_STRDUP(job->name, name) < 0)
|
|
goto cleanup;
|
|
|
|
job->state = QEMU_BLOCKJOB_STATE_NEW;
|
|
job->newstate = -1;
|
|
job->type = type;
|
|
|
|
VIR_STEAL_PTR(ret, job);
|
|
|
|
cleanup:
|
|
virObjectUnref(job);
|
|
return ret;
|
|
}
|
|
|
|
|
|
/**
|
|
* qemuBlockJobDiskNew:
|
|
* @disk: disk definition
|
|
*
|
|
* Start/associate a new blockjob with @disk.
|
|
*
|
|
* Returns 0 on success and -1 on failure.
|
|
*/
|
|
qemuBlockJobDataPtr
|
|
qemuBlockJobDiskNew(virDomainDiskDefPtr disk,
|
|
qemuBlockJobType type,
|
|
const char *jobname)
|
|
{
|
|
qemuBlockJobDataPtr job = NULL;
|
|
|
|
if (!(job = qemuBlockJobDataNew(type, jobname)))
|
|
return NULL;
|
|
|
|
job->disk = disk;
|
|
if (disk)
|
|
QEMU_DOMAIN_DISK_PRIVATE(disk)->blockjob = virObjectRef(job);
|
|
|
|
return job;
|
|
}
|
|
|
|
|
|
/**
|
|
* qemuBlockJobDiskGetJob:
|
|
* @disk: disk definition
|
|
*
|
|
* Get a reference to the block job data object associated with @disk.
|
|
*/
|
|
qemuBlockJobDataPtr
|
|
qemuBlockJobDiskGetJob(virDomainDiskDefPtr disk)
|
|
{
|
|
qemuBlockJobDataPtr job = QEMU_DOMAIN_DISK_PRIVATE(disk)->blockjob;
|
|
|
|
if (!job)
|
|
return NULL;
|
|
|
|
return virObjectRef(job);
|
|
}
|
|
|
|
|
|
/**
|
|
* qemuBlockJobStarted:
|
|
* @job: job data
|
|
*
|
|
* Mark @job as started in qemu.
|
|
*/
|
|
void
|
|
qemuBlockJobStarted(qemuBlockJobDataPtr job)
|
|
{
|
|
if (job->state == QEMU_BLOCKJOB_STATE_NEW)
|
|
job->state = QEMU_BLOCKJOB_STATE_RUNNING;
|
|
}
|
|
|
|
|
|
static void
|
|
qemuBlockJobTerminate(qemuBlockJobDataPtr job)
|
|
{
|
|
qemuDomainDiskPrivatePtr diskPriv;
|
|
|
|
if (job->disk) {
|
|
diskPriv = QEMU_DOMAIN_DISK_PRIVATE(job->disk);
|
|
|
|
if (job == diskPriv->blockjob) {
|
|
virObjectUnref(diskPriv->blockjob);
|
|
diskPriv->blockjob = NULL;
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
/**
|
|
* qemuBlockJobStartupFinalize:
|
|
* @job: job being started
|
|
*
|
|
* Cancels and clears the job private data if the job was not started with
|
|
* qemu (see qemuBlockJobStarted) or just clears up the local reference
|
|
* to @job if it was started.
|
|
*/
|
|
void
|
|
qemuBlockJobStartupFinalize(qemuBlockJobDataPtr job)
|
|
{
|
|
if (!job)
|
|
return;
|
|
|
|
if (job->state == QEMU_BLOCKJOB_STATE_NEW)
|
|
qemuBlockJobTerminate(job);
|
|
|
|
virObjectUnref(job);
|
|
}
|
|
|
|
|
|
bool
|
|
qemuBlockJobIsRunning(qemuBlockJobDataPtr job)
|
|
{
|
|
return job->state == QEMU_BLOCKJOB_STATE_RUNNING ||
|
|
job->state == QEMU_BLOCKJOB_STATE_READY;
|
|
}
|
|
|
|
|
|
/**
|
|
* qemuBlockJobEmitEvents:
|
|
*
|
|
* Emits the VIR_DOMAIN_EVENT_ID_BLOCK_JOB and VIR_DOMAIN_EVENT_ID_BLOCK_JOB_2
|
|
* for a block job. The former event is emitted only for local disks.
|
|
*/
|
|
static void
|
|
qemuBlockJobEmitEvents(virQEMUDriverPtr driver,
|
|
virDomainObjPtr vm,
|
|
virDomainDiskDefPtr disk,
|
|
virDomainBlockJobType type,
|
|
virConnectDomainEventBlockJobStatus status)
|
|
{
|
|
virObjectEventPtr event = NULL;
|
|
virObjectEventPtr event2 = NULL;
|
|
|
|
/* don't emit events for jobs without disk */
|
|
if (!disk)
|
|
return;
|
|
|
|
/* don't emit events for internal jobs and states */
|
|
if (type >= VIR_DOMAIN_BLOCK_JOB_TYPE_LAST ||
|
|
status >= VIR_DOMAIN_BLOCK_JOB_LAST)
|
|
return;
|
|
|
|
if (virStorageSourceIsLocalStorage(disk->src) &&
|
|
!virStorageSourceIsEmpty(disk->src)) {
|
|
event = virDomainEventBlockJobNewFromObj(vm, virDomainDiskGetSource(disk),
|
|
type, status);
|
|
virObjectEventStateQueue(driver->domainEventState, event);
|
|
}
|
|
|
|
event2 = virDomainEventBlockJob2NewFromObj(vm, disk->dst, type, status);
|
|
virObjectEventStateQueue(driver->domainEventState, event2);
|
|
}
|
|
|
|
|
|
static void
|
|
qemuBlockJobEventProcessLegacyCompleted(virQEMUDriverPtr driver,
|
|
virDomainObjPtr vm,
|
|
qemuBlockJobDataPtr job,
|
|
int asyncJob)
|
|
{
|
|
virDomainDiskDefPtr disk = job->disk;
|
|
virDomainDiskDefPtr persistDisk = NULL;
|
|
|
|
if (!disk)
|
|
return;
|
|
|
|
if (disk->mirrorState == VIR_DOMAIN_DISK_MIRROR_STATE_PIVOT) {
|
|
if (vm->newDef) {
|
|
virStorageSourcePtr copy = NULL;
|
|
|
|
if ((persistDisk = virDomainDiskByName(vm->newDef,
|
|
disk->dst, false))) {
|
|
copy = virStorageSourceCopy(disk->mirror, false);
|
|
if (!copy ||
|
|
virStorageSourceInitChainElement(copy,
|
|
persistDisk->src,
|
|
true) < 0) {
|
|
VIR_WARN("Unable to update persistent definition "
|
|
"on vm %s after block job",
|
|
vm->def->name);
|
|
virObjectUnref(copy);
|
|
copy = NULL;
|
|
persistDisk = NULL;
|
|
}
|
|
}
|
|
if (copy) {
|
|
virObjectUnref(persistDisk->src);
|
|
persistDisk->src = copy;
|
|
}
|
|
}
|
|
|
|
/* XXX We want to revoke security labels as well as audit that
|
|
* revocation, before dropping the original source. But it gets
|
|
* tricky if both source and mirror share common backing files (we
|
|
* want to only revoke the non-shared portion of the chain); so for
|
|
* now, we leak the access to the original. */
|
|
virDomainLockImageDetach(driver->lockManager, vm, disk->src);
|
|
|
|
/* Move secret driver metadata */
|
|
if (qemuSecurityMoveImageMetadata(driver, vm, disk->src, disk->mirror) < 0)
|
|
VIR_WARN("Unable to move disk metadata on vm %s", vm->def->name);
|
|
|
|
virObjectUnref(disk->src);
|
|
disk->src = disk->mirror;
|
|
} else {
|
|
if (disk->mirror) {
|
|
virDomainLockImageDetach(driver->lockManager, vm, disk->mirror);
|
|
virObjectUnref(disk->mirror);
|
|
}
|
|
}
|
|
|
|
/* Recompute the cached backing chain to match our
|
|
* updates. Better would be storing the chain ourselves
|
|
* rather than reprobing, but we haven't quite completed
|
|
* that conversion to use our XML tracking. */
|
|
disk->mirror = NULL;
|
|
disk->mirrorState = VIR_DOMAIN_DISK_MIRROR_STATE_NONE;
|
|
disk->mirrorJob = VIR_DOMAIN_BLOCK_JOB_TYPE_UNKNOWN;
|
|
disk->src->id = 0;
|
|
virStorageSourceBackingStoreClear(disk->src);
|
|
ignore_value(qemuDomainDetermineDiskChain(driver, vm, disk, NULL, true));
|
|
ignore_value(qemuBlockNodeNamesDetect(driver, vm, asyncJob));
|
|
qemuBlockJobTerminate(job);
|
|
}
|
|
|
|
|
|
/**
|
|
* qemuBlockJobEventProcessLegacy:
|
|
* @driver: qemu driver
|
|
* @vm: domain
|
|
* @job: job to process events for
|
|
*
|
|
* Update disk's mirror state in response to a block job event
|
|
* from QEMU. For mirror state's that must survive libvirt
|
|
* restart, also update the domain's status XML.
|
|
*/
|
|
static void
|
|
qemuBlockJobEventProcessLegacy(virQEMUDriverPtr driver,
|
|
virDomainObjPtr vm,
|
|
qemuBlockJobDataPtr job,
|
|
int asyncJob)
|
|
{
|
|
VIR_AUTOUNREF(virQEMUDriverConfigPtr) cfg = virQEMUDriverGetConfig(driver);
|
|
virDomainDiskDefPtr disk = job->disk;
|
|
|
|
VIR_DEBUG("disk=%s, mirrorState=%s, type=%d, state=%d, newstate=%d",
|
|
disk->dst,
|
|
NULLSTR(virDomainDiskMirrorStateTypeToString(disk->mirrorState)),
|
|
job->type,
|
|
job->state,
|
|
job->newstate);
|
|
|
|
if (job->newstate == -1)
|
|
return;
|
|
|
|
qemuBlockJobEmitEvents(driver, vm, disk, job->type, job->newstate);
|
|
|
|
/* If we completed a block pull or commit, then update the XML
|
|
* to match. */
|
|
switch ((virConnectDomainEventBlockJobStatus) job->newstate) {
|
|
case VIR_DOMAIN_BLOCK_JOB_COMPLETED:
|
|
qemuBlockJobEventProcessLegacyCompleted(driver, vm, job, asyncJob);
|
|
break;
|
|
|
|
case VIR_DOMAIN_BLOCK_JOB_READY:
|
|
disk->mirrorState = VIR_DOMAIN_DISK_MIRROR_STATE_READY;
|
|
break;
|
|
|
|
case VIR_DOMAIN_BLOCK_JOB_FAILED:
|
|
case VIR_DOMAIN_BLOCK_JOB_CANCELED:
|
|
if (disk->mirror) {
|
|
virDomainLockImageDetach(driver->lockManager, vm, disk->mirror);
|
|
virObjectUnref(disk->mirror);
|
|
disk->mirror = NULL;
|
|
}
|
|
disk->mirrorState = VIR_DOMAIN_DISK_MIRROR_STATE_NONE;
|
|
disk->mirrorJob = VIR_DOMAIN_BLOCK_JOB_TYPE_UNKNOWN;
|
|
qemuBlockJobTerminate(job);
|
|
break;
|
|
|
|
case VIR_DOMAIN_BLOCK_JOB_LAST:
|
|
break;
|
|
}
|
|
|
|
job->state = job->newstate;
|
|
job->newstate = -1;
|
|
|
|
if (virDomainSaveStatus(driver->xmlopt, cfg->stateDir, vm, driver->caps) < 0)
|
|
VIR_WARN("Unable to save status on vm %s after block job", vm->def->name);
|
|
|
|
if (job->state == VIR_DOMAIN_BLOCK_JOB_COMPLETED && vm->newDef) {
|
|
if (virDomainSaveConfig(cfg->configDir, driver->caps, vm->newDef) < 0)
|
|
VIR_WARN("Unable to update persistent definition on vm %s "
|
|
"after block job", vm->def->name);
|
|
}
|
|
}
|
|
|
|
|
|
/**
|
|
* qemuBlockJobUpdate:
|
|
* @vm: domain
|
|
* @job: job data
|
|
* @asyncJob: current qemu asynchronous job type
|
|
*
|
|
* Update disk's mirror state in response to a block job event stored in
|
|
* blockJobStatus by qemuProcessHandleBlockJob event handler.
|
|
*
|
|
* Returns the block job event processed or -1 if there was no pending event.
|
|
*/
|
|
int
|
|
qemuBlockJobUpdate(virDomainObjPtr vm,
|
|
qemuBlockJobDataPtr job,
|
|
int asyncJob)
|
|
{
|
|
qemuDomainObjPrivatePtr priv = vm->privateData;
|
|
|
|
if (job->newstate == -1)
|
|
return -1;
|
|
|
|
qemuBlockJobEventProcessLegacy(priv->driver, vm, job, asyncJob);
|
|
|
|
return job->state;
|
|
}
|
|
|
|
|
|
/**
|
|
* qemuBlockJobSyncBegin:
|
|
* @job: block job data
|
|
* @disk: domain disk
|
|
*
|
|
* Begin a new synchronous block job for @disk. The synchronous
|
|
* block job is ended by a call to qemuBlockJobSyncEnd, or by
|
|
* the guest quitting.
|
|
*
|
|
* During a synchronous block job, a block job event for @disk
|
|
* will not be processed asynchronously. Instead, it will be
|
|
* processed only when qemuBlockJobUpdate or qemuBlockJobSyncEnd
|
|
* is called.
|
|
*/
|
|
void
|
|
qemuBlockJobSyncBegin(qemuBlockJobDataPtr job)
|
|
{
|
|
const char *diskdst = NULL;
|
|
|
|
if (job->disk)
|
|
diskdst = job->disk->dst;
|
|
|
|
VIR_DEBUG("disk=%s", NULLSTR(diskdst));
|
|
job->synchronous = true;
|
|
}
|
|
|
|
|
|
/**
|
|
* qemuBlockJobSyncEnd:
|
|
* @vm: domain
|
|
* @disk: domain disk
|
|
*
|
|
* End a synchronous block job for @disk. Any pending block job event
|
|
* for the disk is processed. Note that it's not necessary to call this function
|
|
* in case the block job was not started successfully if
|
|
* qemuBlockJobStartupFinalize will be called.
|
|
*/
|
|
void
|
|
qemuBlockJobSyncEnd(virDomainObjPtr vm,
|
|
qemuBlockJobDataPtr job,
|
|
int asyncJob)
|
|
{
|
|
const char *diskdst = NULL;
|
|
|
|
if (job->disk)
|
|
diskdst = job->disk->dst;
|
|
|
|
VIR_DEBUG("disk=%s", NULLSTR(diskdst));
|
|
qemuBlockJobUpdate(vm, job, asyncJob);
|
|
job->synchronous = false;
|
|
}
|
|
|
|
|
|
qemuBlockJobDataPtr
|
|
qemuBlockJobGetByDisk(virDomainDiskDefPtr disk)
|
|
{
|
|
qemuBlockJobDataPtr job = QEMU_DOMAIN_DISK_PRIVATE(disk)->blockjob;
|
|
|
|
if (!job)
|
|
return NULL;
|
|
|
|
return virObjectRef(job);
|
|
}
|