mirror of
https://github.com/cloud-hypervisor/cloud-hypervisor.git
synced 2025-02-22 03:12:27 +00:00
vm-virtio: Implement reset() for virtio-vsock
The virtio specification defines a device can be reset, which was not supported by this virtio-vsock implementation. The reason it is needed is to support unbinding this device from the guest driver, and rebind it to vfio-pci driver. Signed-off-by: Sebastien Boeuf <sebastien.boeuf@intel.com>
This commit is contained in:
parent
3e750de43f
commit
dac7737919
@ -87,7 +87,7 @@ pub struct VsockEpollHandler<B: VsockBackend> {
|
|||||||
pub queue_evts: Vec<EventFd>,
|
pub queue_evts: Vec<EventFd>,
|
||||||
pub kill_evt: EventFd,
|
pub kill_evt: EventFd,
|
||||||
pub interrupt_cb: Arc<VirtioInterrupt>,
|
pub interrupt_cb: Arc<VirtioInterrupt>,
|
||||||
pub backend: B,
|
pub backend: Arc<RwLock<B>>,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl<B> VsockEpollHandler<B>
|
impl<B> VsockEpollHandler<B>
|
||||||
@ -118,7 +118,7 @@ where
|
|||||||
for avail_desc in self.queues[0].iter(&mem) {
|
for avail_desc in self.queues[0].iter(&mem) {
|
||||||
let used_len = match VsockPacket::from_rx_virtq_head(&avail_desc) {
|
let used_len = match VsockPacket::from_rx_virtq_head(&avail_desc) {
|
||||||
Ok(mut pkt) => {
|
Ok(mut pkt) => {
|
||||||
if self.backend.recv_pkt(&mut pkt).is_ok() {
|
if self.backend.write().unwrap().recv_pkt(&mut pkt).is_ok() {
|
||||||
pkt.hdr().len() as u32 + pkt.len()
|
pkt.hdr().len() as u32 + pkt.len()
|
||||||
} else {
|
} else {
|
||||||
// We are using a consuming iterator over the virtio buffers, so, if we can't
|
// We are using a consuming iterator over the virtio buffers, so, if we can't
|
||||||
@ -168,7 +168,7 @@ where
|
|||||||
}
|
}
|
||||||
};
|
};
|
||||||
|
|
||||||
if self.backend.send_pkt(&pkt).is_err() {
|
if self.backend.write().unwrap().send_pkt(&pkt).is_err() {
|
||||||
self.queues[1].go_to_previous_position();
|
self.queues[1].go_to_previous_position();
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
@ -217,8 +217,11 @@ where
|
|||||||
epoll::ctl(
|
epoll::ctl(
|
||||||
epoll_fd,
|
epoll_fd,
|
||||||
epoll::ControlOptions::EPOLL_CTL_ADD,
|
epoll::ControlOptions::EPOLL_CTL_ADD,
|
||||||
self.backend.get_polled_fd(),
|
self.backend.read().unwrap().get_polled_fd(),
|
||||||
epoll::Event::new(self.backend.get_polled_evset(), u64::from(BACKEND_EVENT)),
|
epoll::Event::new(
|
||||||
|
self.backend.read().unwrap().get_polled_evset(),
|
||||||
|
u64::from(BACKEND_EVENT),
|
||||||
|
),
|
||||||
)
|
)
|
||||||
.map_err(DeviceError::EpollCtl)?;
|
.map_err(DeviceError::EpollCtl)?;
|
||||||
epoll::ctl(
|
epoll::ctl(
|
||||||
@ -284,7 +287,7 @@ where
|
|||||||
event_type: "rx queue event",
|
event_type: "rx queue event",
|
||||||
underlying: e,
|
underlying: e,
|
||||||
});
|
});
|
||||||
} else if self.backend.has_pending_rx() {
|
} else if self.backend.read().unwrap().has_pending_rx() {
|
||||||
self.process_rx()?;
|
self.process_rx()?;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -301,7 +304,7 @@ where
|
|||||||
// The backend may have queued up responses to the packets we sent during TX queue
|
// The backend may have queued up responses to the packets we sent during TX queue
|
||||||
// processing. If that happened, we need to fetch those responses and place them
|
// processing. If that happened, we need to fetch those responses and place them
|
||||||
// into RX buffers.
|
// into RX buffers.
|
||||||
if self.backend.has_pending_rx() {
|
if self.backend.read().unwrap().has_pending_rx() {
|
||||||
self.process_rx()?;
|
self.process_rx()?;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -318,14 +321,14 @@ where
|
|||||||
}
|
}
|
||||||
BACKEND_EVENT => {
|
BACKEND_EVENT => {
|
||||||
debug!("vsock: backend event");
|
debug!("vsock: backend event");
|
||||||
self.backend.notify(evset);
|
self.backend.write().unwrap().notify(evset);
|
||||||
// After the backend has been kicked, it might've freed up some resources, so we
|
// After the backend has been kicked, it might've freed up some resources, so we
|
||||||
// can attempt to send it more data to process.
|
// can attempt to send it more data to process.
|
||||||
// In particular, if `self.backend.send_pkt()` halted the TX queue processing (by
|
// In particular, if `self.backend.send_pkt()` halted the TX queue processing (by
|
||||||
// reurning an error) at some point in the past, now is the time to try walking the
|
// reurning an error) at some point in the past, now is the time to try walking the
|
||||||
// TX queue again.
|
// TX queue again.
|
||||||
self.process_tx()?;
|
self.process_tx()?;
|
||||||
if self.backend.has_pending_rx() {
|
if self.backend.read().unwrap().has_pending_rx() {
|
||||||
self.process_rx()?;
|
self.process_rx()?;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -349,10 +352,12 @@ where
|
|||||||
/// Virtio device exposing virtual socket to the guest.
|
/// Virtio device exposing virtual socket to the guest.
|
||||||
pub struct Vsock<B: VsockBackend> {
|
pub struct Vsock<B: VsockBackend> {
|
||||||
cid: u64,
|
cid: u64,
|
||||||
backend: Option<B>,
|
backend: Arc<RwLock<B>>,
|
||||||
kill_evt: Option<EventFd>,
|
kill_evt: Option<EventFd>,
|
||||||
avail_features: u64,
|
avail_features: u64,
|
||||||
acked_features: u64,
|
acked_features: u64,
|
||||||
|
queue_evts: Option<Vec<EventFd>>,
|
||||||
|
interrupt_cb: Option<Arc<VirtioInterrupt>>,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl<B> Vsock<B>
|
impl<B> Vsock<B>
|
||||||
@ -366,10 +371,12 @@ where
|
|||||||
|
|
||||||
Ok(Vsock {
|
Ok(Vsock {
|
||||||
cid,
|
cid,
|
||||||
backend: Some(backend),
|
backend: Arc::new(RwLock::new(backend)),
|
||||||
kill_evt: None,
|
kill_evt: None,
|
||||||
avail_features,
|
avail_features,
|
||||||
acked_features: 0u64,
|
acked_features: 0u64,
|
||||||
|
queue_evts: None,
|
||||||
|
interrupt_cb: None,
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -388,7 +395,7 @@ where
|
|||||||
|
|
||||||
impl<B> VirtioDevice for Vsock<B>
|
impl<B> VirtioDevice for Vsock<B>
|
||||||
where
|
where
|
||||||
B: VsockBackend + 'static,
|
B: VsockBackend + Sync + 'static,
|
||||||
{
|
{
|
||||||
fn device_type(&self) -> u32 {
|
fn device_type(&self) -> u32 {
|
||||||
VirtioDeviceType::TYPE_VSOCK as u32
|
VirtioDeviceType::TYPE_VSOCK as u32
|
||||||
@ -481,13 +488,28 @@ where
|
|||||||
};
|
};
|
||||||
self.kill_evt = Some(self_kill_evt);
|
self.kill_evt = Some(self_kill_evt);
|
||||||
|
|
||||||
|
// Save the interrupt EventFD as we need to return it on reset
|
||||||
|
// but clone it to pass into the thread.
|
||||||
|
self.interrupt_cb = Some(interrupt_cb.clone());
|
||||||
|
|
||||||
|
let mut tmp_queue_evts: Vec<EventFd> = Vec::new();
|
||||||
|
for queue_evt in queue_evts.iter() {
|
||||||
|
// Save the queue EventFD as we need to return it on reset
|
||||||
|
// but clone it to pass into the thread.
|
||||||
|
tmp_queue_evts.push(queue_evt.try_clone().map_err(|e| {
|
||||||
|
error!("failed to clone queue EventFd: {}", e);
|
||||||
|
ActivateError::BadActivate
|
||||||
|
})?);
|
||||||
|
}
|
||||||
|
self.queue_evts = Some(tmp_queue_evts);
|
||||||
|
|
||||||
let mut handler = VsockEpollHandler {
|
let mut handler = VsockEpollHandler {
|
||||||
mem,
|
mem,
|
||||||
queues,
|
queues,
|
||||||
queue_evts,
|
queue_evts,
|
||||||
kill_evt,
|
kill_evt,
|
||||||
interrupt_cb,
|
interrupt_cb,
|
||||||
backend: self.backend.take().unwrap(),
|
backend: self.backend.clone(),
|
||||||
};
|
};
|
||||||
|
|
||||||
let worker_result = thread::Builder::new()
|
let worker_result = thread::Builder::new()
|
||||||
@ -501,6 +523,19 @@ where
|
|||||||
|
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
|
fn reset(&mut self) -> Option<(Arc<VirtioInterrupt>, Vec<EventFd>)> {
|
||||||
|
if let Some(kill_evt) = self.kill_evt.take() {
|
||||||
|
// Ignore the result because there is nothing we can do about it.
|
||||||
|
let _ = kill_evt.write(1);
|
||||||
|
}
|
||||||
|
|
||||||
|
// Return the interrupt and queue EventFDs
|
||||||
|
Some((
|
||||||
|
self.interrupt_cb.take().unwrap(),
|
||||||
|
self.queue_evts.take().unwrap(),
|
||||||
|
))
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
#[cfg(test)]
|
#[cfg(test)]
|
||||||
@ -626,7 +661,7 @@ mod tests {
|
|||||||
let test_ctx = TestContext::new();
|
let test_ctx = TestContext::new();
|
||||||
let mut ctx = test_ctx.create_epoll_handler_context();
|
let mut ctx = test_ctx.create_epoll_handler_context();
|
||||||
|
|
||||||
ctx.handler.backend.set_pending_rx(false);
|
ctx.handler.backend.write().unwrap().set_pending_rx(false);
|
||||||
ctx.signal_txq_event();
|
ctx.signal_txq_event();
|
||||||
|
|
||||||
// The available TX descriptor should have been used.
|
// The available TX descriptor should have been used.
|
||||||
@ -642,7 +677,7 @@ mod tests {
|
|||||||
let test_ctx = TestContext::new();
|
let test_ctx = TestContext::new();
|
||||||
let mut ctx = test_ctx.create_epoll_handler_context();
|
let mut ctx = test_ctx.create_epoll_handler_context();
|
||||||
|
|
||||||
ctx.handler.backend.set_pending_rx(true);
|
ctx.handler.backend.write().unwrap().set_pending_rx(true);
|
||||||
ctx.signal_txq_event();
|
ctx.signal_txq_event();
|
||||||
|
|
||||||
// Both available RX and TX descriptors should have been used.
|
// Both available RX and TX descriptors should have been used.
|
||||||
@ -657,8 +692,12 @@ mod tests {
|
|||||||
let test_ctx = TestContext::new();
|
let test_ctx = TestContext::new();
|
||||||
let mut ctx = test_ctx.create_epoll_handler_context();
|
let mut ctx = test_ctx.create_epoll_handler_context();
|
||||||
|
|
||||||
ctx.handler.backend.set_pending_rx(false);
|
ctx.handler.backend.write().unwrap().set_pending_rx(false);
|
||||||
ctx.handler.backend.set_tx_err(Some(VsockError::NoData));
|
ctx.handler
|
||||||
|
.backend
|
||||||
|
.write()
|
||||||
|
.unwrap()
|
||||||
|
.set_tx_err(Some(VsockError::NoData));
|
||||||
ctx.signal_txq_event();
|
ctx.signal_txq_event();
|
||||||
|
|
||||||
// Both RX and TX queues should be untouched.
|
// Both RX and TX queues should be untouched.
|
||||||
@ -679,7 +718,7 @@ mod tests {
|
|||||||
// The available descriptor should have been consumed, but no packet should have
|
// The available descriptor should have been consumed, but no packet should have
|
||||||
// reached the backend.
|
// reached the backend.
|
||||||
assert_eq!(ctx.guest_txvq.used.idx.get(), 1);
|
assert_eq!(ctx.guest_txvq.used.idx.get(), 1);
|
||||||
assert_eq!(ctx.handler.backend.tx_ok_cnt, 0);
|
assert_eq!(ctx.handler.backend.read().unwrap().tx_ok_cnt, 0);
|
||||||
}
|
}
|
||||||
|
|
||||||
// Test case: spurious TXQ_EVENT.
|
// Test case: spurious TXQ_EVENT.
|
||||||
@ -707,8 +746,12 @@ mod tests {
|
|||||||
let test_ctx = TestContext::new();
|
let test_ctx = TestContext::new();
|
||||||
let mut ctx = test_ctx.create_epoll_handler_context();
|
let mut ctx = test_ctx.create_epoll_handler_context();
|
||||||
|
|
||||||
ctx.handler.backend.set_pending_rx(true);
|
ctx.handler.backend.write().unwrap().set_pending_rx(true);
|
||||||
ctx.handler.backend.set_rx_err(Some(VsockError::NoData));
|
ctx.handler
|
||||||
|
.backend
|
||||||
|
.write()
|
||||||
|
.unwrap()
|
||||||
|
.set_rx_err(Some(VsockError::NoData));
|
||||||
ctx.signal_rxq_event();
|
ctx.signal_rxq_event();
|
||||||
|
|
||||||
// The available RX buffer should've been left untouched.
|
// The available RX buffer should've been left untouched.
|
||||||
@ -723,7 +766,7 @@ mod tests {
|
|||||||
let test_ctx = TestContext::new();
|
let test_ctx = TestContext::new();
|
||||||
let mut ctx = test_ctx.create_epoll_handler_context();
|
let mut ctx = test_ctx.create_epoll_handler_context();
|
||||||
|
|
||||||
ctx.handler.backend.set_pending_rx(true);
|
ctx.handler.backend.write().unwrap().set_pending_rx(true);
|
||||||
ctx.signal_rxq_event();
|
ctx.signal_rxq_event();
|
||||||
|
|
||||||
// The available RX buffer should have been used.
|
// The available RX buffer should have been used.
|
||||||
@ -741,14 +784,14 @@ mod tests {
|
|||||||
// The chain should've been processed, without employing the backend.
|
// The chain should've been processed, without employing the backend.
|
||||||
assert!(ctx.handler.process_rx().is_ok());
|
assert!(ctx.handler.process_rx().is_ok());
|
||||||
assert_eq!(ctx.guest_rxvq.used.idx.get(), 1);
|
assert_eq!(ctx.guest_rxvq.used.idx.get(), 1);
|
||||||
assert_eq!(ctx.handler.backend.rx_ok_cnt, 0);
|
assert_eq!(ctx.handler.backend.read().unwrap().rx_ok_cnt, 0);
|
||||||
}
|
}
|
||||||
|
|
||||||
// Test case: spurious RXQ_EVENT.
|
// Test case: spurious RXQ_EVENT.
|
||||||
{
|
{
|
||||||
let test_ctx = TestContext::new();
|
let test_ctx = TestContext::new();
|
||||||
let mut ctx = test_ctx.create_epoll_handler_context();
|
let mut ctx = test_ctx.create_epoll_handler_context();
|
||||||
ctx.handler.backend.set_pending_rx(false);
|
ctx.handler.backend.write().unwrap().set_pending_rx(false);
|
||||||
match ctx
|
match ctx
|
||||||
.handler
|
.handler
|
||||||
.handle_event(RX_QUEUE_EVENT, epoll::Events::EPOLLIN)
|
.handle_event(RX_QUEUE_EVENT, epoll::Events::EPOLLIN)
|
||||||
@ -765,7 +808,7 @@ mod tests {
|
|||||||
{
|
{
|
||||||
let test_ctx = TestContext::new();
|
let test_ctx = TestContext::new();
|
||||||
let mut ctx = test_ctx.create_epoll_handler_context();
|
let mut ctx = test_ctx.create_epoll_handler_context();
|
||||||
ctx.handler.backend.set_pending_rx(false);
|
ctx.handler.backend.write().unwrap().set_pending_rx(false);
|
||||||
match ctx
|
match ctx
|
||||||
.handler
|
.handler
|
||||||
.handle_event(EVT_QUEUE_EVENT, epoll::Events::EPOLLIN)
|
.handle_event(EVT_QUEUE_EVENT, epoll::Events::EPOLLIN)
|
||||||
@ -785,13 +828,16 @@ mod tests {
|
|||||||
let test_ctx = TestContext::new();
|
let test_ctx = TestContext::new();
|
||||||
let mut ctx = test_ctx.create_epoll_handler_context();
|
let mut ctx = test_ctx.create_epoll_handler_context();
|
||||||
|
|
||||||
ctx.handler.backend.set_pending_rx(true);
|
ctx.handler.backend.write().unwrap().set_pending_rx(true);
|
||||||
ctx.handler
|
ctx.handler
|
||||||
.handle_event(BACKEND_EVENT, epoll::Events::EPOLLIN)
|
.handle_event(BACKEND_EVENT, epoll::Events::EPOLLIN)
|
||||||
.unwrap();
|
.unwrap();
|
||||||
|
|
||||||
// The backend should've received this event.
|
// The backend should've received this event.
|
||||||
assert_eq!(ctx.handler.backend.evset, Some(epoll::Events::EPOLLIN));
|
assert_eq!(
|
||||||
|
ctx.handler.backend.read().unwrap().evset,
|
||||||
|
Some(epoll::Events::EPOLLIN)
|
||||||
|
);
|
||||||
// TX queue processing should've been triggered.
|
// TX queue processing should've been triggered.
|
||||||
assert_eq!(ctx.guest_txvq.used.idx.get(), 1);
|
assert_eq!(ctx.guest_txvq.used.idx.get(), 1);
|
||||||
// RX queue processing should've been triggered.
|
// RX queue processing should've been triggered.
|
||||||
@ -805,13 +851,16 @@ mod tests {
|
|||||||
let test_ctx = TestContext::new();
|
let test_ctx = TestContext::new();
|
||||||
let mut ctx = test_ctx.create_epoll_handler_context();
|
let mut ctx = test_ctx.create_epoll_handler_context();
|
||||||
|
|
||||||
ctx.handler.backend.set_pending_rx(false);
|
ctx.handler.backend.write().unwrap().set_pending_rx(false);
|
||||||
ctx.handler
|
ctx.handler
|
||||||
.handle_event(BACKEND_EVENT, epoll::Events::EPOLLIN)
|
.handle_event(BACKEND_EVENT, epoll::Events::EPOLLIN)
|
||||||
.unwrap();
|
.unwrap();
|
||||||
|
|
||||||
// The backend should've received this event.
|
// The backend should've received this event.
|
||||||
assert_eq!(ctx.handler.backend.evset, Some(epoll::Events::EPOLLIN));
|
assert_eq!(
|
||||||
|
ctx.handler.backend.read().unwrap().evset,
|
||||||
|
Some(epoll::Events::EPOLLIN)
|
||||||
|
);
|
||||||
// TX queue processing should've been triggered.
|
// TX queue processing should've been triggered.
|
||||||
assert_eq!(ctx.guest_txvq.used.idx.get(), 1);
|
assert_eq!(ctx.guest_txvq.used.idx.get(), 1);
|
||||||
// The RX queue should've been left untouched.
|
// The RX queue should've been left untouched.
|
||||||
|
@ -307,7 +307,7 @@ mod tests {
|
|||||||
queue_evts,
|
queue_evts,
|
||||||
kill_evt: EventFd::new(EFD_NONBLOCK).unwrap(),
|
kill_evt: EventFd::new(EFD_NONBLOCK).unwrap(),
|
||||||
interrupt_cb,
|
interrupt_cb,
|
||||||
backend: TestBackend::new(),
|
backend: Arc::new(RwLock::new(TestBackend::new())),
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
Loading…
x
Reference in New Issue
Block a user