mirror of
https://github.com/cloud-hypervisor/cloud-hypervisor.git
synced 2024-12-22 05:35:20 +00:00
arch: create memory mapping by the actual memory info
The original codes did not consider that the previous memory region might not be full and always set it to the maximum size. This commit fixes this problem by creating memory mappings based on the actual memory details in both E820 on x86_64 and fdt on aarch64. Fixes: #5463 Signed-off-by: Yu Li <liyu.yukiteru@bytedance.com>
This commit is contained in:
parent
499989fb17
commit
1017157bb6
@ -260,32 +260,51 @@ fn create_memory_node(
|
||||
fdt.end_node(memory_node)?;
|
||||
}
|
||||
} else {
|
||||
let last_addr = guest_mem.last_addr().raw_value();
|
||||
if last_addr < super::layout::MEM_32BIT_RESERVED_START.raw_value() {
|
||||
// Case 1: all RAM is under the hole
|
||||
let mem_size = last_addr - super::layout::RAM_START.raw_value() + 1;
|
||||
let mem_reg_prop = [super::layout::RAM_START.raw_value(), mem_size];
|
||||
let memory_node = fdt.begin_node("memory")?;
|
||||
fdt.property_string("device_type", "memory")?;
|
||||
fdt.property_array_u64("reg", &mem_reg_prop)?;
|
||||
fdt.end_node(memory_node)?;
|
||||
} else {
|
||||
// Case 2: RAM is split by the hole
|
||||
// Region 1: RAM before the hole
|
||||
let mem_size = super::layout::MEM_32BIT_RESERVED_START.raw_value()
|
||||
- super::layout::RAM_START.raw_value();
|
||||
let mem_reg_prop = [super::layout::RAM_START.raw_value(), mem_size];
|
||||
let memory_node_name = format!("memory@{:x}", super::layout::RAM_START.raw_value());
|
||||
let memory_node = fdt.begin_node(&memory_node_name)?;
|
||||
fdt.property_string("device_type", "memory")?;
|
||||
fdt.property_array_u64("reg", &mem_reg_prop)?;
|
||||
fdt.end_node(memory_node)?;
|
||||
let ram_regions = {
|
||||
let mut ram_regions = Vec::new();
|
||||
let mut current_start = 0;
|
||||
let mut current_end = 0;
|
||||
|
||||
// Region 2: RAM after the hole
|
||||
let mem_size = last_addr - super::layout::RAM_64BIT_START.raw_value() + 1;
|
||||
let mem_reg_prop = [super::layout::RAM_64BIT_START.raw_value(), mem_size];
|
||||
let memory_node_name =
|
||||
format!("memory@{:x}", super::layout::RAM_64BIT_START.raw_value());
|
||||
for (start, size) in guest_mem
|
||||
.iter()
|
||||
.map(|m| (m.start_addr().raw_value(), m.len()))
|
||||
{
|
||||
if current_end == start {
|
||||
// This zone is continuous with the previous one.
|
||||
current_end += size;
|
||||
} else {
|
||||
if current_start < current_end {
|
||||
ram_regions.push((current_start, current_end));
|
||||
}
|
||||
|
||||
current_start = start;
|
||||
current_end = start + size;
|
||||
}
|
||||
}
|
||||
|
||||
if current_start < current_end {
|
||||
ram_regions.push((current_start, current_end));
|
||||
}
|
||||
|
||||
ram_regions
|
||||
};
|
||||
|
||||
for ((region_start, region_end), layout_start) in ram_regions
|
||||
.iter()
|
||||
.zip([super::layout::RAM_START, super::layout::RAM_64BIT_START])
|
||||
{
|
||||
let layout_start = layout_start.raw_value();
|
||||
if &layout_start < region_start || region_end <= &layout_start {
|
||||
warn!(
|
||||
"Layout start addr {:#x} cannot fit ram region {:#x} ~ {:#x}",
|
||||
layout_start, region_start, region_end
|
||||
);
|
||||
continue;
|
||||
}
|
||||
|
||||
let mem_size = region_end - layout_start;
|
||||
let mem_reg_prop = [layout_start, mem_size];
|
||||
let memory_node_name = format!("memory@{:x}", layout_start);
|
||||
let memory_node = fdt.begin_node(&memory_node_name)?;
|
||||
fdt.property_string("device_type", "memory")?;
|
||||
fdt.property_array_u64("reg", &mem_reg_prop)?;
|
||||
|
@ -972,30 +972,55 @@ fn configure_pvh(
|
||||
// Create the memory map entries.
|
||||
add_memmap_entry(&mut memmap, 0, layout::EBDA_START.raw_value(), E820_RAM);
|
||||
|
||||
let mem_end = guest_mem.last_addr();
|
||||
// Merge continuous zones to one region
|
||||
let ram_regions = {
|
||||
let mut ram_regions = Vec::new();
|
||||
let mut current_start = 0;
|
||||
let mut current_end = 0;
|
||||
|
||||
if mem_end < layout::MEM_32BIT_RESERVED_START {
|
||||
add_memmap_entry(
|
||||
&mut memmap,
|
||||
layout::HIGH_RAM_START.raw_value(),
|
||||
mem_end.unchecked_offset_from(layout::HIGH_RAM_START) + 1,
|
||||
E820_RAM,
|
||||
);
|
||||
} else {
|
||||
add_memmap_entry(
|
||||
&mut memmap,
|
||||
layout::HIGH_RAM_START.raw_value(),
|
||||
layout::MEM_32BIT_RESERVED_START.unchecked_offset_from(layout::HIGH_RAM_START),
|
||||
E820_RAM,
|
||||
);
|
||||
if mem_end > layout::RAM_64BIT_START {
|
||||
add_memmap_entry(
|
||||
&mut memmap,
|
||||
layout::RAM_64BIT_START.raw_value(),
|
||||
mem_end.unchecked_offset_from(layout::RAM_64BIT_START) + 1,
|
||||
E820_RAM,
|
||||
);
|
||||
for (start, size) in guest_mem
|
||||
.iter()
|
||||
.map(|m| (m.start_addr().raw_value(), m.len()))
|
||||
{
|
||||
if current_end == start {
|
||||
// This zone is continuous with the previous one.
|
||||
current_end += size;
|
||||
} else {
|
||||
if current_start < current_end {
|
||||
ram_regions.push((current_start, current_end));
|
||||
}
|
||||
|
||||
current_start = start;
|
||||
current_end = start + size;
|
||||
}
|
||||
}
|
||||
|
||||
if current_start < current_end {
|
||||
ram_regions.push((current_start, current_end));
|
||||
}
|
||||
|
||||
ram_regions
|
||||
};
|
||||
|
||||
for ((region_start, region_end), layout_start) in ram_regions
|
||||
.iter()
|
||||
.zip([layout::HIGH_RAM_START, layout::RAM_64BIT_START])
|
||||
{
|
||||
let layout_start = layout_start.raw_value();
|
||||
if &layout_start < region_start || region_end <= &layout_start {
|
||||
return Err(super::Error::MemmapTableSetup);
|
||||
}
|
||||
|
||||
info!(
|
||||
"create_memmap_entry, start: 0x{:08x}, end: 0x{:08x})",
|
||||
layout_start, region_end
|
||||
);
|
||||
add_memmap_entry(
|
||||
&mut memmap,
|
||||
layout_start,
|
||||
region_end - layout_start,
|
||||
E820_RAM,
|
||||
);
|
||||
}
|
||||
|
||||
add_memmap_entry(
|
||||
|
Loading…
Reference in New Issue
Block a user