2021-05-12 17:01:31 +00:00
|
|
|
/*
|
|
|
|
* Copyright Intel Corp. 2020-2021
|
|
|
|
*
|
|
|
|
* ch_process.c: Process controller for Cloud-Hypervisor driver
|
|
|
|
*
|
|
|
|
* This library is free software; you can redistribute it and/or
|
|
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
|
|
* License as published by the Free Software Foundation; either
|
|
|
|
* version 2.1 of the License, or (at your option) any later version.
|
|
|
|
*
|
|
|
|
* This library is distributed in the hope that it will be useful,
|
|
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
|
|
* Lesser General Public License for more details.
|
|
|
|
*
|
|
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
|
|
* License along with this library. If not, see
|
|
|
|
* <http://www.gnu.org/licenses/>.
|
|
|
|
*/
|
|
|
|
|
|
|
|
#include <config.h>
|
|
|
|
|
|
|
|
#include <unistd.h>
|
|
|
|
#include <fcntl.h>
|
|
|
|
|
|
|
|
#include "ch_domain.h"
|
|
|
|
#include "ch_monitor.h"
|
|
|
|
#include "ch_process.h"
|
2022-01-25 16:19:54 +00:00
|
|
|
#include "domain_cgroup.h"
|
2021-05-12 17:01:31 +00:00
|
|
|
#include "virerror.h"
|
2021-09-08 18:01:20 +00:00
|
|
|
#include "virjson.h"
|
2021-05-12 17:01:31 +00:00
|
|
|
#include "virlog.h"
|
|
|
|
|
|
|
|
#define VIR_FROM_THIS VIR_FROM_CH
|
|
|
|
|
|
|
|
VIR_LOG_INIT("ch.ch_process");
|
|
|
|
|
|
|
|
#define START_SOCKET_POSTFIX ": starting up socket\n"
|
|
|
|
#define START_VM_POSTFIX ": starting up vm\n"
|
|
|
|
|
|
|
|
|
|
|
|
static virCHMonitor *
|
|
|
|
virCHProcessConnectMonitor(virCHDriver *driver,
|
|
|
|
virDomainObj *vm)
|
|
|
|
{
|
|
|
|
virCHMonitor *monitor = NULL;
|
|
|
|
virCHDriverConfig *cfg = virCHDriverGetConfig(driver);
|
|
|
|
|
|
|
|
monitor = virCHMonitorNew(vm, cfg->stateDir);
|
|
|
|
|
|
|
|
virObjectUnref(cfg);
|
|
|
|
return monitor;
|
|
|
|
}
|
|
|
|
|
2021-09-08 18:01:20 +00:00
|
|
|
static void
|
|
|
|
virCHProcessUpdateConsoleDevice(virDomainObj *vm,
|
|
|
|
virJSONValue *config,
|
|
|
|
const char *device)
|
|
|
|
{
|
|
|
|
const char *path;
|
|
|
|
virDomainChrDef *chr = NULL;
|
|
|
|
virJSONValue *dev, *file;
|
|
|
|
|
|
|
|
if (!config)
|
|
|
|
return;
|
|
|
|
|
|
|
|
dev = virJSONValueObjectGet(config, device);
|
|
|
|
if (!dev) {
|
|
|
|
virReportError(VIR_ERR_INTERNAL_ERROR,
|
|
|
|
_("missing '%s' in 'config' from cloud-hypervisor"),
|
|
|
|
device);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
file = virJSONValueObjectGet(dev, "file");
|
|
|
|
if (!file) {
|
|
|
|
virReportError(VIR_ERR_INTERNAL_ERROR,
|
|
|
|
_("missing 'file' in '%s' from cloud-hypervisor"),
|
|
|
|
device);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
path = virJSONValueGetString(file);
|
|
|
|
if (!path) {
|
|
|
|
virReportError(VIR_ERR_INTERNAL_ERROR,
|
|
|
|
_("unable to parse contents of 'file' field in '%s' from cloud-hypervisor"),
|
|
|
|
device);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (STREQ(device, "console")) {
|
|
|
|
chr = vm->def->consoles[0];
|
|
|
|
} else if (STREQ(device, "serial")) {
|
|
|
|
chr = vm->def->serials[0];
|
|
|
|
}
|
|
|
|
|
|
|
|
if (chr && chr->source)
|
|
|
|
chr->source->data.file.path = g_strdup(path);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
virCHProcessUpdateConsole(virDomainObj *vm,
|
|
|
|
virJSONValue *info)
|
|
|
|
{
|
|
|
|
virJSONValue *config;
|
|
|
|
|
|
|
|
config = virJSONValueObjectGet(info, "config");
|
|
|
|
if (!config) {
|
|
|
|
virReportError(VIR_ERR_INTERNAL_ERROR, "%s",
|
|
|
|
_("missing 'config' in info query result from cloud-hypervisor"));
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (vm->def->nconsoles > 0)
|
|
|
|
virCHProcessUpdateConsoleDevice(vm, config, "console");
|
|
|
|
if (vm->def->nserials > 0)
|
|
|
|
virCHProcessUpdateConsoleDevice(vm, config, "serial");
|
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
|
|
|
virCHProcessUpdateInfo(virDomainObj *vm)
|
|
|
|
{
|
2021-12-01 08:26:27 +00:00
|
|
|
g_autoptr(virJSONValue) info = NULL;
|
2021-09-08 18:01:20 +00:00
|
|
|
virCHDomainObjPrivate *priv = vm->privateData;
|
|
|
|
if (virCHMonitorGetInfo(priv->monitor, &info) < 0)
|
|
|
|
return -1;
|
|
|
|
|
|
|
|
virCHProcessUpdateConsole(vm, info);
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2022-01-25 16:19:54 +00:00
|
|
|
static int
|
|
|
|
virCHProcessGetAllCpuAffinity(virBitmap **cpumapRet)
|
|
|
|
{
|
|
|
|
*cpumapRet = NULL;
|
|
|
|
|
|
|
|
if (!virHostCPUHasBitmap())
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
if (!(*cpumapRet = virHostCPUGetOnlineBitmap()))
|
|
|
|
return -1;
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
#if defined(WITH_SCHED_GETAFFINITY) || defined(WITH_BSD_CPU_AFFINITY)
|
|
|
|
static int
|
|
|
|
virCHProcessInitCpuAffinity(virDomainObj *vm)
|
|
|
|
{
|
|
|
|
g_autoptr(virBitmap) cpumapToSet = NULL;
|
|
|
|
virDomainNumatuneMemMode mem_mode;
|
|
|
|
virCHDomainObjPrivate *priv = vm->privateData;
|
|
|
|
|
|
|
|
if (!vm->pid) {
|
|
|
|
virReportError(VIR_ERR_INTERNAL_ERROR, "%s",
|
|
|
|
_("Cannot setup CPU affinity until process is started"));
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (virDomainNumaGetNodeCount(vm->def->numa) <= 1 &&
|
|
|
|
virDomainNumatuneGetMode(vm->def->numa, -1, &mem_mode) == 0 &&
|
|
|
|
mem_mode == VIR_DOMAIN_NUMATUNE_MEM_STRICT) {
|
|
|
|
virBitmap *nodeset = NULL;
|
|
|
|
|
|
|
|
if (virDomainNumatuneMaybeGetNodeset(vm->def->numa,
|
|
|
|
priv->autoNodeset,
|
|
|
|
&nodeset, -1) < 0)
|
|
|
|
return -1;
|
|
|
|
|
|
|
|
if (virNumaNodesetToCPUset(nodeset, &cpumapToSet) < 0)
|
|
|
|
return -1;
|
|
|
|
} else if (vm->def->cputune.emulatorpin) {
|
|
|
|
if (!(cpumapToSet = virBitmapNewCopy(vm->def->cputune.emulatorpin)))
|
|
|
|
return -1;
|
|
|
|
} else {
|
|
|
|
if (virCHProcessGetAllCpuAffinity(&cpumapToSet) < 0)
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (cpumapToSet && virProcessSetAffinity(vm->pid, cpumapToSet, false) < 0) {
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
#else /* !defined(WITH_SCHED_GETAFFINITY) && !defined(WITH_BSD_CPU_AFFINITY) */
|
|
|
|
static int
|
|
|
|
virCHProcessInitCpuAffinity(virDomainObj *vm G_GNUC_UNUSED)
|
|
|
|
{
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
#endif /* !defined(WITH_SCHED_GETAFFINITY) && !defined(WITH_BSD_CPU_AFFINITY) */
|
|
|
|
|
|
|
|
/**
|
|
|
|
* virCHProcessSetupPid:
|
|
|
|
*
|
|
|
|
* This function sets resource properties (affinity, cgroups,
|
|
|
|
* scheduler) for any PID associated with a domain. It should be used
|
|
|
|
* to set up emulator PIDs as well as vCPU and I/O thread pids to
|
|
|
|
* ensure they are all handled the same way.
|
|
|
|
*
|
|
|
|
* Returns 0 on success, -1 on error.
|
|
|
|
*/
|
|
|
|
static int
|
|
|
|
virCHProcessSetupPid(virDomainObj *vm,
|
|
|
|
pid_t pid,
|
|
|
|
virCgroupThreadName nameval,
|
|
|
|
int id,
|
|
|
|
virBitmap *cpumask,
|
|
|
|
unsigned long long period,
|
|
|
|
long long quota,
|
|
|
|
virDomainThreadSchedParam *sched)
|
|
|
|
{
|
|
|
|
virCHDomainObjPrivate *priv = vm->privateData;
|
|
|
|
virDomainNumatuneMemMode mem_mode;
|
|
|
|
g_autoptr(virCgroup) cgroup = NULL;
|
|
|
|
virBitmap *use_cpumask = NULL;
|
|
|
|
virBitmap *affinity_cpumask = NULL;
|
|
|
|
g_autoptr(virBitmap) hostcpumap = NULL;
|
|
|
|
g_autofree char *mem_mask = NULL;
|
|
|
|
int ret = -1;
|
|
|
|
|
|
|
|
if ((period || quota) &&
|
|
|
|
!virCgroupHasController(priv->cgroup, VIR_CGROUP_CONTROLLER_CPU)) {
|
|
|
|
virReportError(VIR_ERR_CONFIG_UNSUPPORTED, "%s",
|
|
|
|
_("cgroup cpu is required for scheduler tuning"));
|
|
|
|
goto cleanup;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Infer which cpumask shall be used. */
|
|
|
|
if (cpumask) {
|
|
|
|
use_cpumask = cpumask;
|
|
|
|
} else if (vm->def->placement_mode == VIR_DOMAIN_CPU_PLACEMENT_MODE_AUTO) {
|
|
|
|
use_cpumask = priv->autoCpuset;
|
|
|
|
} else if (vm->def->cpumask) {
|
|
|
|
use_cpumask = vm->def->cpumask;
|
|
|
|
} else {
|
|
|
|
/* we can't assume cloud-hypervisor itself is running on all pCPUs,
|
|
|
|
* so we need to explicitly set the spawned instance to all pCPUs. */
|
|
|
|
if (virCHProcessGetAllCpuAffinity(&hostcpumap) < 0)
|
|
|
|
goto cleanup;
|
|
|
|
affinity_cpumask = hostcpumap;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* If CPU cgroup controller is not initialized here, then we need
|
|
|
|
* neither period nor quota settings. And if CPUSET controller is
|
|
|
|
* not initialized either, then there's nothing to do anyway.
|
|
|
|
*/
|
|
|
|
if (virCgroupHasController(priv->cgroup, VIR_CGROUP_CONTROLLER_CPU) ||
|
|
|
|
virCgroupHasController(priv->cgroup, VIR_CGROUP_CONTROLLER_CPUSET)) {
|
|
|
|
|
|
|
|
if (virDomainNumatuneGetMode(vm->def->numa, -1, &mem_mode) == 0 &&
|
2022-04-01 12:30:05 +00:00
|
|
|
(mem_mode == VIR_DOMAIN_NUMATUNE_MEM_STRICT ||
|
|
|
|
mem_mode == VIR_DOMAIN_NUMATUNE_MEM_RESTRICTIVE) &&
|
2022-01-25 16:19:54 +00:00
|
|
|
virDomainNumatuneMaybeFormatNodeset(vm->def->numa,
|
|
|
|
priv->autoNodeset,
|
|
|
|
&mem_mask, -1) < 0)
|
|
|
|
goto cleanup;
|
|
|
|
|
|
|
|
if (virCgroupNewThread(priv->cgroup, nameval, id, true, &cgroup) < 0)
|
|
|
|
goto cleanup;
|
|
|
|
|
|
|
|
if (virCgroupHasController(priv->cgroup, VIR_CGROUP_CONTROLLER_CPUSET)) {
|
|
|
|
if (use_cpumask &&
|
|
|
|
virDomainCgroupSetupCpusetCpus(cgroup, use_cpumask) < 0)
|
|
|
|
goto cleanup;
|
|
|
|
|
|
|
|
if (mem_mask && virCgroupSetCpusetMems(cgroup, mem_mask) < 0)
|
|
|
|
goto cleanup;
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
if (virDomainCgroupSetupVcpuBW(cgroup, period, quota) < 0)
|
|
|
|
goto cleanup;
|
|
|
|
|
|
|
|
/* Move the thread to the sub dir */
|
|
|
|
VIR_INFO("Adding pid %d to cgroup", pid);
|
|
|
|
if (virCgroupAddThread(cgroup, pid) < 0)
|
|
|
|
goto cleanup;
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!affinity_cpumask)
|
|
|
|
affinity_cpumask = use_cpumask;
|
|
|
|
|
|
|
|
/* Setup legacy affinity. */
|
|
|
|
if (affinity_cpumask
|
|
|
|
&& virProcessSetAffinity(pid, affinity_cpumask, false) < 0)
|
|
|
|
goto cleanup;
|
|
|
|
|
|
|
|
/* Set scheduler type and priority, but not for the main thread. */
|
|
|
|
if (sched &&
|
|
|
|
nameval != VIR_CGROUP_THREAD_EMULATOR &&
|
|
|
|
virProcessSetScheduler(pid, sched->policy, sched->priority) < 0)
|
|
|
|
goto cleanup;
|
|
|
|
|
|
|
|
ret = 0;
|
|
|
|
cleanup:
|
|
|
|
if (ret < 0 && cgroup)
|
|
|
|
virCgroupRemove(cgroup);
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2022-01-25 16:19:58 +00:00
|
|
|
static int
|
|
|
|
virCHProcessSetupIOThread(virDomainObj *vm,
|
|
|
|
virDomainIOThreadInfo *iothread)
|
|
|
|
{
|
|
|
|
virCHDomainObjPrivate *priv = vm->privateData;
|
|
|
|
|
|
|
|
return virCHProcessSetupPid(vm, iothread->iothread_id,
|
|
|
|
VIR_CGROUP_THREAD_IOTHREAD,
|
|
|
|
iothread->iothread_id,
|
|
|
|
priv->autoCpuset, /* This should be updated when CLH supports accepting
|
|
|
|
iothread settings from input domain definition */
|
|
|
|
vm->def->cputune.iothread_period,
|
|
|
|
vm->def->cputune.iothread_quota,
|
|
|
|
NULL); /* CLH doesn't allow choosing a scheduler for iothreads.*/
|
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
|
|
|
virCHProcessSetupIOThreads(virDomainObj *vm)
|
|
|
|
{
|
|
|
|
virCHDomainObjPrivate *priv = vm->privateData;
|
|
|
|
virDomainIOThreadInfo **iothreads = NULL;
|
|
|
|
size_t i;
|
2022-02-02 16:26:48 +00:00
|
|
|
int niothreads;
|
|
|
|
|
|
|
|
if ((niothreads = virCHMonitorGetIOThreads(priv->monitor, &iothreads)) < 0)
|
|
|
|
return -1;
|
2022-01-25 16:19:58 +00:00
|
|
|
|
|
|
|
for (i = 0; i < niothreads; i++) {
|
|
|
|
VIR_DEBUG("IOThread index = %ld , tid = %d", i, iothreads[i]->iothread_id);
|
|
|
|
if (virCHProcessSetupIOThread(vm, iothreads[i]) < 0)
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
|
|
|
virCHProcessSetupEmulatorThread(virDomainObj *vm,
|
|
|
|
virCHMonitorEmuThreadInfo emuthread)
|
|
|
|
{
|
|
|
|
return virCHProcessSetupPid(vm, emuthread.tid,
|
|
|
|
VIR_CGROUP_THREAD_EMULATOR, 0,
|
|
|
|
vm->def->cputune.emulatorpin,
|
|
|
|
vm->def->cputune.emulator_period,
|
|
|
|
vm->def->cputune.emulator_quota,
|
|
|
|
vm->def->cputune.emulatorsched);
|
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
|
|
|
virCHProcessSetupEmulatorThreads(virDomainObj *vm)
|
|
|
|
{
|
|
|
|
int thd_index = 0;
|
|
|
|
virCHDomainObjPrivate *priv = vm->privateData;
|
|
|
|
|
|
|
|
/* Cloud-hypervisor start 4 Emulator threads by default:
|
|
|
|
* vmm
|
|
|
|
* cloud-hypervisor
|
|
|
|
* http-server
|
|
|
|
* signal_handler */
|
|
|
|
for (thd_index = 0; thd_index < priv->monitor->nthreads; thd_index++) {
|
|
|
|
if (priv->monitor->threads[thd_index].type == virCHThreadTypeEmulator) {
|
|
|
|
VIR_DEBUG("Setup tid = %d (%s) Emulator thread",
|
|
|
|
priv->monitor->threads[thd_index].emuInfo.tid,
|
|
|
|
priv->monitor->threads[thd_index].emuInfo.thrName);
|
|
|
|
|
|
|
|
if (virCHProcessSetupEmulatorThread(vm,
|
|
|
|
priv->monitor->threads[thd_index].emuInfo) < 0)
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2022-01-25 16:19:54 +00:00
|
|
|
/**
|
|
|
|
* virCHProcessSetupVcpu:
|
|
|
|
* @vm: domain object
|
|
|
|
* @vcpuid: id of VCPU to set defaults
|
|
|
|
*
|
|
|
|
* This function sets resource properties (cgroups, affinity, scheduler) for a
|
|
|
|
* vCPU. This function expects that the vCPU is online and the vCPU pids were
|
|
|
|
* correctly detected at the point when it's called.
|
|
|
|
*
|
|
|
|
* Returns 0 on success, -1 on error.
|
|
|
|
*/
|
|
|
|
int
|
|
|
|
virCHProcessSetupVcpu(virDomainObj *vm,
|
|
|
|
unsigned int vcpuid)
|
|
|
|
{
|
|
|
|
pid_t vcpupid = virCHDomainGetVcpuPid(vm, vcpuid);
|
|
|
|
virDomainVcpuDef *vcpu = virDomainDefGetVcpu(vm->def, vcpuid);
|
|
|
|
|
|
|
|
return virCHProcessSetupPid(vm, vcpupid, VIR_CGROUP_THREAD_VCPU,
|
|
|
|
vcpuid, vcpu->cpumask,
|
|
|
|
vm->def->cputune.period,
|
|
|
|
vm->def->cputune.quota, &vcpu->sched);
|
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
|
|
|
virCHProcessSetupVcpus(virDomainObj *vm)
|
|
|
|
{
|
|
|
|
virDomainVcpuDef *vcpu;
|
|
|
|
unsigned int maxvcpus = virDomainDefGetVcpusMax(vm->def);
|
|
|
|
size_t i;
|
|
|
|
|
|
|
|
if ((vm->def->cputune.period || vm->def->cputune.quota) &&
|
|
|
|
!virCgroupHasController(((virCHDomainObjPrivate *) vm->privateData)->
|
|
|
|
cgroup, VIR_CGROUP_CONTROLLER_CPU)) {
|
|
|
|
virReportError(VIR_ERR_CONFIG_UNSUPPORTED, "%s",
|
|
|
|
_("cgroup cpu is required for scheduler tuning"));
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!virCHDomainHasVcpuPids(vm)) {
|
|
|
|
/* If any CPU has custom affinity that differs from the
|
|
|
|
* VM default affinity, we must reject it */
|
|
|
|
for (i = 0; i < maxvcpus; i++) {
|
|
|
|
vcpu = virDomainDefGetVcpu(vm->def, i);
|
|
|
|
|
|
|
|
if (!vcpu->online)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
if (vcpu->cpumask &&
|
|
|
|
!virBitmapEqual(vm->def->cpumask, vcpu->cpumask)) {
|
|
|
|
virReportError(VIR_ERR_OPERATION_INVALID, "%s",
|
|
|
|
_("cpu affinity is not supported"));
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
for (i = 0; i < maxvcpus; i++) {
|
|
|
|
vcpu = virDomainDefGetVcpu(vm->def, i);
|
|
|
|
|
|
|
|
if (!vcpu->online)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
if (virCHProcessSetupVcpu(vm, i) < 0)
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2021-05-12 17:01:31 +00:00
|
|
|
/**
|
|
|
|
* virCHProcessStart:
|
|
|
|
* @driver: pointer to driver structure
|
|
|
|
* @vm: pointer to virtual machine structure
|
|
|
|
* @reason: reason for switching vm to running state
|
|
|
|
*
|
|
|
|
* Starts Cloud-Hypervisor listen on a local socket
|
|
|
|
*
|
|
|
|
* Returns 0 on success or -1 in case of error
|
|
|
|
*/
|
2022-01-25 16:19:54 +00:00
|
|
|
int
|
|
|
|
virCHProcessStart(virCHDriver *driver,
|
|
|
|
virDomainObj *vm,
|
|
|
|
virDomainRunningReason reason)
|
2021-05-12 17:01:31 +00:00
|
|
|
{
|
|
|
|
int ret = -1;
|
|
|
|
virCHDomainObjPrivate *priv = vm->privateData;
|
2022-01-25 16:19:54 +00:00
|
|
|
g_autoptr(virCHDriverConfig) cfg = virCHDriverGetConfig(priv->driver);
|
2021-12-10 20:34:41 +00:00
|
|
|
g_autofree int *nicindexes = NULL;
|
|
|
|
size_t nnicindexes = 0;
|
2021-05-12 17:01:31 +00:00
|
|
|
|
2022-02-10 15:17:20 +00:00
|
|
|
if (virDomainObjIsActive(vm)) {
|
|
|
|
virReportError(VIR_ERR_OPERATION_INVALID, "%s",
|
|
|
|
_("VM is already active"));
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
|
2021-05-12 17:01:31 +00:00
|
|
|
if (!priv->monitor) {
|
|
|
|
/* And we can get the first monitor connection now too */
|
|
|
|
if (!(priv->monitor = virCHProcessConnectMonitor(driver, vm))) {
|
|
|
|
virReportError(VIR_ERR_INTERNAL_ERROR, "%s",
|
2022-01-25 16:19:54 +00:00
|
|
|
_("failed to create connection to CH socket"));
|
2021-05-12 17:01:31 +00:00
|
|
|
goto cleanup;
|
|
|
|
}
|
|
|
|
|
2021-12-10 20:34:41 +00:00
|
|
|
if (virCHMonitorCreateVM(priv->monitor,
|
|
|
|
&nnicindexes, &nicindexes) < 0) {
|
2021-05-12 17:01:31 +00:00
|
|
|
virReportError(VIR_ERR_INTERNAL_ERROR, "%s",
|
2022-01-25 16:19:54 +00:00
|
|
|
_("failed to create guest VM"));
|
2021-05-12 17:01:31 +00:00
|
|
|
goto cleanup;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2022-01-25 16:19:54 +00:00
|
|
|
vm->pid = priv->monitor->pid;
|
|
|
|
vm->def->id = vm->pid;
|
|
|
|
priv->machineName = virCHDomainGetMachineName(vm);
|
|
|
|
|
|
|
|
if (virDomainCgroupSetupCgroup("ch", vm,
|
|
|
|
nnicindexes, nicindexes,
|
|
|
|
&priv->cgroup,
|
|
|
|
cfg->cgroupControllers,
|
|
|
|
0, /*maxThreadsPerProc*/
|
|
|
|
priv->driver->privileged,
|
|
|
|
priv->machineName) < 0)
|
|
|
|
goto cleanup;
|
|
|
|
|
|
|
|
if (virCHProcessInitCpuAffinity(vm) < 0)
|
|
|
|
goto cleanup;
|
|
|
|
|
2021-05-12 17:01:31 +00:00
|
|
|
if (virCHMonitorBootVM(priv->monitor) < 0) {
|
|
|
|
virReportError(VIR_ERR_INTERNAL_ERROR, "%s",
|
|
|
|
_("failed to boot guest VM"));
|
|
|
|
goto cleanup;
|
|
|
|
}
|
|
|
|
|
2022-01-25 16:19:54 +00:00
|
|
|
virCHDomainRefreshThreadInfo(vm);
|
2021-09-08 18:01:20 +00:00
|
|
|
|
2022-01-25 16:19:58 +00:00
|
|
|
VIR_DEBUG("Setting emulator tuning/settings");
|
|
|
|
if (virCHProcessSetupEmulatorThreads(vm) < 0)
|
|
|
|
goto cleanup;
|
|
|
|
|
|
|
|
VIR_DEBUG("Setting iothread tuning/settings");
|
|
|
|
if (virCHProcessSetupIOThreads(vm) < 0)
|
|
|
|
goto cleanup;
|
|
|
|
|
2022-01-25 16:19:54 +00:00
|
|
|
VIR_DEBUG("Setting global CPU cgroup (if required)");
|
|
|
|
if (virDomainCgroupSetupGlobalCpuCgroup(vm,
|
2022-04-01 11:14:59 +00:00
|
|
|
priv->cgroup) < 0)
|
2022-01-25 16:19:54 +00:00
|
|
|
goto cleanup;
|
2021-09-08 18:01:20 +00:00
|
|
|
|
2022-01-25 16:19:54 +00:00
|
|
|
VIR_DEBUG("Setting vCPU tuning/settings");
|
|
|
|
if (virCHProcessSetupVcpus(vm) < 0)
|
|
|
|
goto cleanup;
|
|
|
|
|
|
|
|
virCHProcessUpdateInfo(vm);
|
2021-05-12 17:01:31 +00:00
|
|
|
virDomainObjSetState(vm, VIR_DOMAIN_RUNNING, reason);
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
cleanup:
|
|
|
|
if (ret)
|
|
|
|
virCHProcessStop(driver, vm, VIR_DOMAIN_SHUTOFF_FAILED);
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2022-01-25 16:19:54 +00:00
|
|
|
int
|
|
|
|
virCHProcessStop(virCHDriver *driver G_GNUC_UNUSED,
|
|
|
|
virDomainObj *vm,
|
|
|
|
virDomainShutoffReason reason)
|
2021-05-12 17:01:31 +00:00
|
|
|
{
|
2022-01-25 16:19:54 +00:00
|
|
|
int ret;
|
|
|
|
int retries = 0;
|
2021-05-12 17:01:31 +00:00
|
|
|
virCHDomainObjPrivate *priv = vm->privateData;
|
|
|
|
|
|
|
|
VIR_DEBUG("Stopping VM name=%s pid=%d reason=%d",
|
|
|
|
vm->def->name, (int)vm->pid, (int)reason);
|
|
|
|
|
|
|
|
if (priv->monitor) {
|
2022-01-28 17:42:45 +00:00
|
|
|
g_clear_pointer(&priv->monitor, virCHMonitorClose);
|
2021-05-12 17:01:31 +00:00
|
|
|
}
|
|
|
|
|
2022-01-25 16:19:54 +00:00
|
|
|
retry:
|
|
|
|
if ((ret = virDomainCgroupRemoveCgroup(vm,
|
|
|
|
priv->cgroup,
|
|
|
|
priv->machineName)) < 0) {
|
|
|
|
if (ret == -EBUSY && (retries++ < 5)) {
|
|
|
|
g_usleep(200*1000);
|
|
|
|
goto retry;
|
|
|
|
}
|
|
|
|
VIR_WARN("Failed to remove cgroup for %s",
|
|
|
|
vm->def->name);
|
|
|
|
}
|
|
|
|
|
lib: Be consistent about vm->pid
The virDomainObj struct has @pid member where the domain's
hypervisor PID is stored (e.g. QEMU/bhyve/libvirt_lxc/... PID).
However, we are not consistent when it comes to shutoff state.
Initially, because virDomainObjNew() uses g_new0() the @pid is
initialized to 0. But when domain is shut off, some functions set
it to -1 (virBhyveProcessStop, virCHProcessStop, qemuProcessStop,
..).
In other places, the @pid is tested to be 0, on some other places
it's tested for being negative and in the rest for being
positive.
To solve this inconsistency we can stick with either value, -1 or
0. I've chosen the latter as it's safer IMO. For instance if by
mistake we'd kill(vm->pid, SIGTERM) we would kill ourselves
instead of init's process group.
Signed-off-by: Michal Privoznik <mprivozn@redhat.com>
Reviewed-by: Jonathon Jongsma <jjongsma@redhat.com>
2022-05-26 07:07:56 +00:00
|
|
|
vm->pid = 0;
|
2021-05-12 17:01:31 +00:00
|
|
|
vm->def->id = -1;
|
2021-12-10 20:34:41 +00:00
|
|
|
g_clear_pointer(&priv->machineName, g_free);
|
2021-05-12 17:01:31 +00:00
|
|
|
|
|
|
|
virDomainObjSetState(vm, VIR_DOMAIN_SHUTOFF, reason);
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|