#![cfg_attr(minimal_rt, no_std, no_main)]
#![expect(unsafe_code)]
mod arch;
mod boot_logger;
mod cmdline;
mod dt;
mod host_params;
mod hypercall;
mod rt;
mod sidecar;
mod single_threaded;
use crate::arch::setup_vtl2_memory;
use crate::arch::setup_vtl2_vp;
#[cfg(target_arch = "x86_64")]
use crate::arch::tdx::get_tdx_tsc_reftime;
use crate::arch::verify_imported_regions_hash;
use crate::boot_logger::boot_logger_init;
use crate::boot_logger::log;
use crate::hypercall::hvcall;
use crate::single_threaded::off_stack;
use arrayvec::ArrayString;
use arrayvec::ArrayVec;
use boot_logger::LoggerType;
use core::fmt::Write;
use dt::BootTimes;
use dt::write_dt;
use host_params::COMMAND_LINE_SIZE;
use host_params::PartitionInfo;
use host_params::shim_params::IsolationType;
use host_params::shim_params::ShimParams;
use hvdef::Vtl;
use loader_defs::linux::SETUP_DTB;
use loader_defs::linux::setup_data;
use loader_defs::shim::ShimParamsRaw;
use memory_range::MemoryRange;
use memory_range::RangeWalkResult;
use memory_range::merge_adjacent_ranges;
use memory_range::walk_ranges;
use minimal_rt::enlightened_panic::enable_enlightened_panic;
use sidecar::SidecarConfig;
use sidecar_defs::SidecarOutput;
use sidecar_defs::SidecarParams;
use single_threaded::OffStackRef;
use zerocopy::FromBytes;
use zerocopy::FromZeros;
use zerocopy::Immutable;
use zerocopy::IntoBytes;
use zerocopy::KnownLayout;
#[derive(Debug)]
struct CommandLineTooLong;
impl From<core::fmt::Error> for CommandLineTooLong {
fn from(_: core::fmt::Error) -> Self {
Self
}
}
fn build_kernel_command_line(
params: &ShimParams,
cmdline: &mut ArrayString<COMMAND_LINE_SIZE>,
partition_info: &PartitionInfo,
can_trust_host: bool,
sidecar: Option<&SidecarConfig<'_>>,
) -> Result<(), CommandLineTooLong> {
const KERNEL_PARAMETERS: &[&str] = &[
"loglevel=8",
"log_buf_len=128K",
"printk.time=1",
"console_msg_format=syslog",
"uio_hv_generic.no_mask=1",
"coredump_filter=0x33",
"cpufreq.off=1",
"cpuidle.off=1",
"cryptomgr.notests",
"idle=halt",
"initcall_blacklist=init_real_mode,sbf_init",
"lpj=3000000",
"no_timer_check",
"noxsave",
"oops=panic",
"panic_on_warn=0",
"panic_print=0",
"panic=-1",
"printk.devkmsg=on",
"reboot=t",
"rootfstype=tmpfs",
"sysctl.vm.compaction_proactiveness=0",
"tsc=reliable",
"unknown_nmi_panic=1",
"vfio_pci.ids=1414:00ba",
"vfio.enable_unsafe_noiommu_mode=1",
"rdinit=/underhill-init",
"OPENHCL_NVME_VFIO=1",
"hv_storvsc.storvsc_vcpus_per_sub_channel=2048",
"hv_storvsc.storvsc_max_hw_queues=2",
"hv_storvsc.storvsc_ringbuffer_size=0x8000",
"MIMALLOC_ARENA_EAGER_COMMIT=0",
];
const X86_KERNEL_PARAMETERS: &[&str] = &[
"clearcpuid=pcid",
"iommu=off",
"pci=off",
];
const AARCH64_KERNEL_PARAMETERS: &[&str] = &[];
for p in KERNEL_PARAMETERS {
write!(cmdline, "{p} ")?;
}
let arch_parameters = if cfg!(target_arch = "x86_64") {
X86_KERNEL_PARAMETERS
} else {
AARCH64_KERNEL_PARAMETERS
};
for p in arch_parameters {
write!(cmdline, "{p} ")?;
}
const HARDWARE_ISOLATED_KERNEL_PARAMETERS: &[&str] = &[
"swiotlb=4096,1",
];
const NON_HARDWARE_ISOLATED_KERNEL_PARAMETERS: &[&str] = &[
"swiotlb=1,1",
];
if params.isolation_type.is_hardware_isolated() {
for p in HARDWARE_ISOLATED_KERNEL_PARAMETERS {
write!(cmdline, "{p} ")?;
}
} else {
for p in NON_HARDWARE_ISOLATED_KERNEL_PARAMETERS {
write!(cmdline, "{p} ")?;
}
}
let console = if partition_info.com3_serial_available && can_trust_host {
"ttyS2,115200"
} else {
"ttynull"
};
write!(cmdline, "console={console} ")?;
if params.isolation_type != IsolationType::None {
write!(
cmdline,
"{}=1 ",
underhill_confidentiality::OPENHCL_CONFIDENTIAL_ENV_VAR_NAME
)?;
}
if partition_info.nvme_keepalive && !partition_info.vtl2_pool_memory.is_empty() {
write!(cmdline, "OPENHCL_NVME_KEEP_ALIVE=1 ")?;
}
if let Some(sidecar) = sidecar {
write!(cmdline, "{} ", sidecar.kernel_command_line())?;
}
if can_trust_host {
let old_cmdline = &partition_info.cmdline;
if !old_cmdline.contains("hv_vmbus.message_connection_id=") {
write!(
cmdline,
"hv_vmbus.message_connection_id=0x{:x} ",
partition_info.vmbus_vtl2.connection_id
)?;
}
cmdline.write_str(old_cmdline)?;
}
Ok(())
}
const FDT_SIZE: usize = 256 * 1024;
#[repr(C, align(4096))]
#[derive(FromBytes, IntoBytes, Immutable, KnownLayout)]
struct Fdt {
header: setup_data,
data: [u8; FDT_SIZE - size_of::<setup_data>()],
}
fn shim_parameters(shim_params_raw_offset: isize) -> ShimParams {
unsafe extern "C" {
static __ehdr_start: u8;
}
let shim_base = core::ptr::addr_of!(__ehdr_start) as usize;
let raw_shim_params = unsafe {
&*(shim_base.wrapping_add_signed(shim_params_raw_offset) as *const ShimParamsRaw)
};
ShimParams::new(shim_base as u64, raw_shim_params)
}
pub const MAX_RESERVED_MEM_RANGES: usize = 5 + sidecar_defs::MAX_NODES;
#[derive(Clone, Copy, Debug, PartialEq, Eq)]
enum ReservedMemoryType {
Vtl2Config,
Vtl2Reserved,
SidecarImage,
SidecarNode,
Vtl2GpaPool,
}
fn reserved_memory_regions(
partition_info: &PartitionInfo,
sidecar: Option<&SidecarConfig<'_>>,
) -> OffStackRef<'static, impl AsRef<[(MemoryRange, ReservedMemoryType)]> + use<>> {
let mut reserved = off_stack!(ArrayVec<(MemoryRange, ReservedMemoryType), MAX_RESERVED_MEM_RANGES>, ArrayVec::new_const());
reserved.clear();
reserved.extend(
partition_info
.vtl2_config_regions()
.map(|r| (r, ReservedMemoryType::Vtl2Config)),
);
if let Some(sidecar) = sidecar {
reserved.push((sidecar.image, ReservedMemoryType::SidecarImage));
reserved.extend(sidecar.node_params.iter().map(|x| {
(
MemoryRange::new(x.memory_base..x.memory_base + x.memory_size),
ReservedMemoryType::SidecarNode,
)
}));
}
if !partition_info.vtl2_reserved_region.is_empty() {
reserved.push((
partition_info.vtl2_reserved_region,
ReservedMemoryType::Vtl2Reserved,
));
}
if partition_info.vtl2_pool_memory != MemoryRange::EMPTY {
reserved.push((
partition_info.vtl2_pool_memory,
ReservedMemoryType::Vtl2GpaPool,
));
}
reserved
.as_mut()
.sort_unstable_by_key(|(r, _typ)| r.start());
let mut flattened = off_stack!(ArrayVec<(MemoryRange, ReservedMemoryType), MAX_RESERVED_MEM_RANGES>, ArrayVec::new_const());
flattened.clear();
flattened.extend(merge_adjacent_ranges(reserved.iter().copied()));
flattened
}
#[cfg_attr(not(target_arch = "x86_64"), expect(dead_code))]
mod x86_boot {
use crate::PageAlign;
use crate::ReservedMemoryType;
use crate::host_params::PartitionInfo;
use crate::single_threaded::OffStackRef;
use crate::single_threaded::off_stack;
use crate::zeroed;
use core::mem::size_of;
use core::ops::Range;
use core::ptr;
use loader_defs::linux::E820_RAM;
use loader_defs::linux::E820_RESERVED;
use loader_defs::linux::SETUP_E820_EXT;
use loader_defs::linux::boot_params;
use loader_defs::linux::e820entry;
use loader_defs::linux::setup_data;
use memory_range::MemoryRange;
use memory_range::RangeWalkResult;
use memory_range::walk_ranges;
use zerocopy::FromZeros;
use zerocopy::Immutable;
use zerocopy::KnownLayout;
#[repr(C)]
#[derive(FromZeros, Immutable, KnownLayout)]
pub struct E820Ext {
pub header: setup_data,
pub entries: [e820entry; 512],
}
fn add_e820_entry(
entry: Option<&mut e820entry>,
range: MemoryRange,
typ: u32,
) -> Result<(), BuildE820MapError> {
*entry.ok_or(BuildE820MapError::OutOfE820Entries)? = e820entry {
addr: range.start().into(),
size: range.len().into(),
typ: typ.into(),
};
Ok(())
}
#[derive(Debug)]
pub enum BuildE820MapError {
ReservedRegionNotCovered,
OutOfE820Entries,
}
pub fn build_e820_map(
boot_params: &mut boot_params,
ext: &mut E820Ext,
partition_info: &PartitionInfo,
reserved: &[(MemoryRange, ReservedMemoryType)],
) -> Result<bool, BuildE820MapError> {
boot_params.e820_entries = 0;
let mut entries = boot_params
.e820_map
.iter_mut()
.chain(ext.entries.iter_mut());
let mut n = 0;
for (range, r) in walk_ranges(
partition_info.vtl2_ram.iter().map(|e| (e.range, ())),
reserved.iter().map(|&(r, _)| (r, ())),
) {
match r {
RangeWalkResult::Neither => {}
RangeWalkResult::Left(_) => {
add_e820_entry(entries.next(), range, E820_RAM)?;
n += 1;
}
RangeWalkResult::Right(_) => {
return Err(BuildE820MapError::ReservedRegionNotCovered);
}
RangeWalkResult::Both(_, _) => {
add_e820_entry(entries.next(), range, E820_RESERVED)?;
n += 1;
}
}
}
let base = n.min(boot_params.e820_map.len());
boot_params.e820_entries = base as u8;
if base < n {
ext.header.len = ((n - base) * size_of::<e820entry>()) as u32;
Ok(true)
} else {
Ok(false)
}
}
pub fn build_boot_params(
partition_info: &PartitionInfo,
reserved_memory: &[(MemoryRange, ReservedMemoryType)],
initrd: Range<u64>,
cmdline: &str,
setup_data_head: *const setup_data,
setup_data_tail: &mut &mut setup_data,
) -> OffStackRef<'static, PageAlign<boot_params>> {
let mut boot_params_storage = off_stack!(PageAlign<boot_params>, zeroed());
let boot_params = &mut boot_params_storage.0;
boot_params.hdr.type_of_loader = 0xff; boot_params.hdr.hardware_subarch = 1.into();
boot_params.hdr.ramdisk_image = (initrd.start as u32).into();
boot_params.ext_ramdisk_image = (initrd.start >> 32) as u32;
let initrd_len = initrd.end - initrd.start;
boot_params.hdr.ramdisk_size = (initrd_len as u32).into();
boot_params.ext_ramdisk_size = (initrd_len >> 32) as u32;
let e820_ext = OffStackRef::leak(off_stack!(E820Ext, zeroed()));
let used_ext = build_e820_map(boot_params, e820_ext, partition_info, reserved_memory)
.expect("building e820 map must succeed");
if used_ext {
e820_ext.header.ty = SETUP_E820_EXT;
setup_data_tail.next = ptr::from_ref(&e820_ext.header) as u64;
*setup_data_tail = &mut e820_ext.header;
}
let cmd_line_addr = cmdline.as_ptr() as u64;
boot_params.hdr.cmd_line_ptr = (cmd_line_addr as u32).into();
boot_params.ext_cmd_line_ptr = (cmd_line_addr >> 32) as u32;
boot_params.hdr.setup_data = (setup_data_head as u64).into();
boot_params_storage
}
}
#[cfg(target_arch = "x86_64")]
fn build_cc_blob_sev_info(
cc_blob: &mut loader_defs::linux::cc_blob_sev_info,
shim_params: &ShimParams,
) {
cc_blob.magic = loader_defs::linux::CC_BLOB_SEV_INFO_MAGIC;
cc_blob.version = 0;
cc_blob._reserved = 0;
cc_blob.secrets_phys = shim_params.secrets_start();
cc_blob.secrets_len = hvdef::HV_PAGE_SIZE as u32;
cc_blob._rsvd1 = 0;
cc_blob.cpuid_phys = shim_params.cpuid_start();
cc_blob.cpuid_len = hvdef::HV_PAGE_SIZE as u32;
cc_blob._rsvd2 = 0;
}
#[repr(C, align(4096))]
#[derive(FromZeros, Immutable, KnownLayout)]
struct PageAlign<T>(T);
const fn zeroed<T: FromZeros>() -> T {
unsafe { core::mem::MaybeUninit::<T>::zeroed().assume_init() }
}
fn get_ref_time(isolation: IsolationType) -> Option<u64> {
match isolation {
#[cfg(target_arch = "x86_64")]
IsolationType::Tdx => get_tdx_tsc_reftime(),
#[cfg(target_arch = "x86_64")]
IsolationType::Snp => None,
_ => Some(minimal_rt::reftime::reference_time()),
}
}
fn shim_main(shim_params_raw_offset: isize) -> ! {
let p = shim_parameters(shim_params_raw_offset);
if !p.isolation_type.is_hardware_isolated() {
hvcall().initialize();
if p.isolation_type == IsolationType::None {
enable_enlightened_panic();
}
}
let static_options =
cmdline::parse_boot_command_line(p.command_line().command_line().unwrap_or(""));
if let Some(typ) = static_options.logger {
boot_logger_init(p.isolation_type, typ);
log!("openhcl_boot: early debugging enabled");
}
let can_trust_host =
p.isolation_type == IsolationType::None || static_options.confidential_debug;
let boot_reftime = get_ref_time(p.isolation_type);
let mut dt_storage = off_stack!(PartitionInfo, PartitionInfo::new());
let partition_info = match PartitionInfo::read_from_dt(&p, &mut dt_storage, can_trust_host) {
Ok(Some(val)) => val,
Ok(None) => panic!("host did not provide a device tree"),
Err(e) => panic!("unable to read device tree params {}", e),
};
if !p.isolation_type.is_hardware_isolated()
&& hvcall().vtl() == Vtl::Vtl2
&& hvdef::HvRegisterVsmCapabilities::from(
hvcall()
.get_register(hvdef::HvAllArchRegisterName::VsmCapabilities.into())
.expect("failed to query vsm capabilities")
.as_u64(),
)
.vtl0_alias_map_available()
{
if partition_info.vtl0_alias_map.is_none() {
partition_info.vtl0_alias_map =
Some(1 << (arch::physical_address_bits(p.isolation_type) - 1));
}
} else {
partition_info.vtl0_alias_map = None;
}
if can_trust_host {
let dynamic_options = cmdline::parse_boot_command_line(&partition_info.cmdline);
if let Some(typ) = dynamic_options.logger {
boot_logger_init(p.isolation_type, typ);
} else if partition_info.com3_serial_available && cfg!(target_arch = "x86_64") {
boot_logger_init(p.isolation_type, LoggerType::Serial);
}
}
log!("openhcl_boot: entered shim_main");
if partition_info.cpus.is_empty() {
panic!("no cpus");
}
validate_vp_hw_ids(partition_info);
setup_vtl2_vp(partition_info);
setup_vtl2_memory(&p, partition_info);
verify_imported_regions_hash(&p);
let mut sidecar_params = off_stack!(PageAlign<SidecarParams>, zeroed());
let mut sidecar_output = off_stack!(PageAlign<SidecarOutput>, zeroed());
let sidecar = sidecar::start_sidecar(
&p,
partition_info,
&mut sidecar_params.0,
&mut sidecar_output.0,
);
let mut cmdline = off_stack!(ArrayString<COMMAND_LINE_SIZE>, ArrayString::new_const());
build_kernel_command_line(
&p,
&mut cmdline,
partition_info,
can_trust_host,
sidecar.as_ref(),
)
.unwrap();
let mut fdt = off_stack!(Fdt, zeroed());
fdt.header.len = fdt.data.len() as u32;
fdt.header.ty = SETUP_DTB;
#[cfg(target_arch = "x86_64")]
let mut setup_data_tail = &mut fdt.header;
#[cfg(target_arch = "x86_64")]
let setup_data_head = core::ptr::from_ref(setup_data_tail);
#[cfg(target_arch = "x86_64")]
if p.isolation_type == IsolationType::Snp {
let cc_blob = OffStackRef::leak(off_stack!(loader_defs::linux::cc_blob_sev_info, zeroed()));
build_cc_blob_sev_info(cc_blob, &p);
let cc_data = OffStackRef::leak(off_stack!(loader_defs::linux::cc_setup_data, zeroed()));
cc_data.header.len = size_of::<loader_defs::linux::cc_setup_data>() as u32;
cc_data.header.ty = loader_defs::linux::SETUP_CC_BLOB;
cc_data.cc_blob_address = core::ptr::from_ref(&*cc_blob) as u32;
setup_data_tail.next = core::ptr::from_ref(&*cc_data) as u64;
setup_data_tail = &mut cc_data.header;
}
let reserved_memory = reserved_memory_regions(partition_info, sidecar.as_ref());
let initrd = p.initrd_base..p.initrd_base + p.initrd_size;
let computed_crc = crc32fast::hash(p.initrd());
assert_eq!(
computed_crc, p.initrd_crc,
"computed initrd crc does not match build time calculated crc"
);
#[cfg(target_arch = "x86_64")]
let boot_params = x86_boot::build_boot_params(
partition_info,
reserved_memory.as_ref(),
initrd.clone(),
&cmdline,
setup_data_head,
&mut setup_data_tail,
);
let boot_times = boot_reftime.map(|start| BootTimes {
start,
end: get_ref_time(p.isolation_type).unwrap_or(0),
});
for (range, result) in walk_ranges(
partition_info.vtl2_ram.iter().map(|r| (r.range, ())),
p.imported_regions(),
) {
match result {
RangeWalkResult::Neither | RangeWalkResult::Left(_) | RangeWalkResult::Both(_, _) => {}
RangeWalkResult::Right(accepted) => {
assert!(
accepted,
"range {:#x?} not in vtl2 ram was not preaccepted at launch",
range
);
}
}
}
write_dt(
&mut fdt.data,
partition_info,
reserved_memory.as_ref(),
p.imported_regions().map(|r| {
r.0
}),
initrd,
&cmdline,
sidecar.as_ref(),
boot_times,
)
.unwrap();
rt::verify_stack_cookie();
log!("uninitializing hypercalls, about to jump to kernel");
hvcall().uninitialize();
cfg_if::cfg_if! {
if #[cfg(target_arch = "x86_64")] {
let kernel_entry: extern "C" fn(u64, &loader_defs::linux::boot_params) -> ! =
unsafe { core::mem::transmute(p.kernel_entry_address) };
kernel_entry(0, &boot_params.0)
} else if #[cfg(target_arch = "aarch64")] {
let kernel_entry: extern "C" fn(fdt_data: *const u8, mbz0: u64, mbz1: u64, mbz2: u64) -> ! =
unsafe { core::mem::transmute(p.kernel_entry_address) };
unsafe {
core::arch::asm!(
"
mrs {0}, sctlr_el1
bic {0}, {0}, #0x1
msr sctlr_el1, {0}
tlbi vmalle1
dsb sy
isb sy",
lateout(reg) _,
);
}
kernel_entry(fdt.data.as_ptr(), 0, 0, 0)
} else {
panic!("unsupported arch")
}
}
}
fn validate_vp_hw_ids(partition_info: &PartitionInfo) {
use host_params::MAX_CPU_COUNT;
use hypercall::HwId;
if partition_info.isolation.is_hardware_isolated() {
return;
}
if hvcall().vtl() != Vtl::Vtl2 {
return;
}
let mut hw_ids = off_stack!(ArrayVec<HwId, MAX_CPU_COUNT>, ArrayVec::new_const());
hw_ids.clear();
hw_ids.extend(partition_info.cpus.iter().map(|c| c.reg as _));
let mut vp_indexes = off_stack!(ArrayVec<u32, MAX_CPU_COUNT>, ArrayVec::new_const());
vp_indexes.clear();
if let Err(err) = hvcall().get_vp_index_from_hw_id(&hw_ids, &mut vp_indexes) {
panic!(
"failed to get VP index for hardware ID {:#x}: {}",
hw_ids[vp_indexes.len().min(hw_ids.len() - 1)],
err
);
}
if let Some((i, &vp_index)) = vp_indexes
.iter()
.enumerate()
.find(|&(i, vp_index)| i as u32 != *vp_index)
{
panic!(
"CPU hardware ID {:#x} does not correspond to VP index {}",
hw_ids[i], vp_index
);
}
}
#[cfg(not(minimal_rt))]
fn main() {
unimplemented!("build with MINIMAL_RT_BUILD to produce a working boot loader");
}
#[cfg(test)]
mod test {
use super::x86_boot::E820Ext;
use super::x86_boot::build_e820_map;
use crate::ReservedMemoryType;
use crate::dt::write_dt;
use crate::host_params::MAX_CPU_COUNT;
use crate::host_params::PartitionInfo;
use crate::host_params::shim_params::IsolationType;
use crate::reserved_memory_regions;
use arrayvec::ArrayString;
use arrayvec::ArrayVec;
use core::ops::Range;
use host_fdt_parser::CpuEntry;
use host_fdt_parser::MemoryEntry;
use host_fdt_parser::VmbusInfo;
use igvm_defs::MemoryMapEntryType;
use loader_defs::linux::E820_RAM;
use loader_defs::linux::E820_RESERVED;
use loader_defs::linux::boot_params;
use loader_defs::linux::e820entry;
use memory_range::MemoryRange;
use memory_range::RangeWalkResult;
use memory_range::walk_ranges;
use zerocopy::FromZeros;
const HIGH_MMIO_GAP_END: u64 = 0x1000000000; const VMBUS_MMIO_GAP_SIZE: u64 = 0x10000000; const HIGH_MMIO_GAP_START: u64 = HIGH_MMIO_GAP_END - VMBUS_MMIO_GAP_SIZE;
fn new_partition_info(cpu_count: usize) -> PartitionInfo {
let mut cpus: ArrayVec<CpuEntry, MAX_CPU_COUNT> = ArrayVec::new();
for id in 0..(cpu_count as u64) {
cpus.push(CpuEntry { reg: id, vnode: 0 });
}
let mut mmio = ArrayVec::new();
mmio.push(
MemoryRange::try_new(HIGH_MMIO_GAP_START..HIGH_MMIO_GAP_END).expect("valid range"),
);
PartitionInfo {
vtl2_ram: ArrayVec::new(),
vtl2_full_config_region: MemoryRange::EMPTY,
vtl2_config_region_reclaim: MemoryRange::EMPTY,
vtl2_reserved_region: MemoryRange::EMPTY,
vtl2_pool_memory: MemoryRange::EMPTY,
vtl2_used_ranges: ArrayVec::new(),
partition_ram: ArrayVec::new(),
isolation: IsolationType::None,
bsp_reg: cpus[0].reg as u32,
cpus,
cmdline: ArrayString::new(),
vmbus_vtl2: VmbusInfo {
mmio,
connection_id: 0,
},
vmbus_vtl0: VmbusInfo {
mmio: ArrayVec::new(),
connection_id: 0,
},
com3_serial_available: false,
gic: None,
memory_allocation_mode: host_fdt_parser::MemoryAllocationMode::Host,
entropy: None,
vtl0_alias_map: None,
nvme_keepalive: false,
}
}
#[test]
#[cfg_attr(
target_arch = "aarch64",
ignore = "TODO: investigate why this doesn't always work on ARM"
)]
fn fdt_cpu_scaling() {
const MAX_CPUS: usize = 2048;
let mut buf = [0; 0x40000];
write_dt(
&mut buf,
&new_partition_info(MAX_CPUS),
&[],
[],
0..0,
&ArrayString::from("test").unwrap_or_default(),
None,
None,
)
.unwrap();
}
#[test]
#[ignore = "TODO: temporarily broken"]
fn fdt_dtc_check_content() {
const MAX_CPUS: usize = 2;
const BUF_SIZE: usize = 0x1000;
let dtb_data_spans: [(usize, &[u8]); 2] = [
(
0,
b"\xd0\x0d\xfe\xed\x00\x00\x10\x00\x00\x00\x04\x38\x00\x00\x00\x38\
\x00\x00\x00\x28\x00\x00\x00\x11\x00\x00\x00\x10\x00\x00\x00\x00\
\x00\x00\x00\x4a\x00\x00\x01\x6c\x00\x00\x00\x00\x00\x00\x00\x00\
\x00\x00\x00\x00\x00\x00\x00\x00\x23\x61\x64\x64\x72\x65\x73\x73\
\x2d\x63\x65\x6c\x6c\x73\x00\x23\x73\x69\x7a\x65\x2d\x63\x65\x6c\
\x6c\x73\x00\x6d\x6f\x64\x65\x6c\x00\x72\x65\x67\x00\x64\x65\x76\
\x69\x63\x65\x5f\x74\x79\x70\x65\x00\x73\x74\x61\x74\x75\x73\x00\
\x63\x6f\x6d\x70\x61\x74\x69\x62\x6c\x65\x00\x72\x61\x6e\x67\x65\
\x73",
),
(
0x430,
b"\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x01\x00\x00\x00\x00\
\x00\x00\x00\x03\x00\x00\x00\x04\x00\x00\x00\x00\x00\x00\x00\x02\
\x00\x00\x00\x03\x00\x00\x00\x04\x00\x00\x00\x0f\x00\x00\x00\x00\
\x00\x00\x00\x03\x00\x00\x00\x0f\x00\x00\x00\x1b\x6d\x73\x66\x74\
\x2c\x75\x6e\x64\x65\x72\x68\x69\x6c\x6c\x00\x00\x00\x00\x00\x01\
\x63\x70\x75\x73\x00\x00\x00\x00\x00\x00\x00\x03\x00\x00\x00\x04\
\x00\x00\x00\x00\x00\x00\x00\x01\x00\x00\x00\x03\x00\x00\x00\x04\
\x00\x00\x00\x0f\x00\x00\x00\x00\x00\x00\x00\x01\x63\x70\x75\x40\
\x30\x00\x00\x00\x00\x00\x00\x03\x00\x00\x00\x04\x00\x00\x00\x25\
\x63\x70\x75\x00\x00\x00\x00\x03\x00\x00\x00\x04\x00\x00\x00\x21\
\x00\x00\x00\x00\x00\x00\x00\x03\x00\x00\x00\x05\x00\x00\x00\x31\
\x6f\x6b\x61\x79\x00\x00\x00\x00\x00\x00\x00\x02\x00\x00\x00\x01\
\x63\x70\x75\x40\x31\x00\x00\x00\x00\x00\x00\x03\x00\x00\x00\x04\
\x00\x00\x00\x25\x63\x70\x75\x00\x00\x00\x00\x03\x00\x00\x00\x04\
\x00\x00\x00\x21\x00\x00\x00\x01\x00\x00\x00\x03\x00\x00\x00\x05\
\x00\x00\x00\x31\x6f\x6b\x61\x79\x00\x00\x00\x00\x00\x00\x00\x02\
\x00\x00\x00\x02\x00\x00\x00\x01\x76\x6d\x62\x75\x73\x00\x00\x00\
\x00\x00\x00\x03\x00\x00\x00\x04\x00\x00\x00\x00\x00\x00\x00\x02\
\x00\x00\x00\x03\x00\x00\x00\x04\x00\x00\x00\x0f\x00\x00\x00\x01\
\x00\x00\x00\x03\x00\x00\x00\x0b\x00\x00\x00\x38\x6d\x73\x66\x74\
\x2c\x76\x6d\x62\x75\x73\x00\x00\x00\x00\x00\x03\x00\x00\x00\x14\
\x00\x00\x00\x43\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x0f\
\xf0\x00\x00\x00\x10\x00\x00\x00\x00\x00\x00\x02\x00\x00\x00\x02\
\x00\x00\x00\x09",
),
];
let mut sample_buf = [0u8; BUF_SIZE];
for (span_start, bytes) in dtb_data_spans {
sample_buf[span_start..span_start + bytes.len()].copy_from_slice(bytes);
}
let mut buf = [0u8; BUF_SIZE];
write_dt(
&mut buf,
&new_partition_info(MAX_CPUS),
&[],
[],
0..0,
&ArrayString::from("test").unwrap_or_default(),
None,
None,
)
.unwrap();
assert!(sample_buf == buf);
}
#[test]
#[ignore = "enabling the test requires installing additional software, \
and developers will experience a break."]
fn fdt_dtc_decompile() {
const MAX_CPUS: usize = 2048;
let mut buf = [0; 0x40000];
write_dt(
&mut buf,
&new_partition_info(MAX_CPUS),
&[],
[],
0..0,
&ArrayString::from("test").unwrap_or_default(),
None,
None,
)
.unwrap();
let input_dtb_file_name = "openhcl_boot.dtb";
let output_dts_file_name = "openhcl_boot.dts";
std::fs::write(input_dtb_file_name, buf).unwrap();
let success = std::process::Command::new("dtc")
.args([input_dtb_file_name, "-I", "dtb", "-o", output_dts_file_name])
.status()
.unwrap()
.success();
assert!(success);
}
fn partition_info_ram_ranges(
ram: &[Range<u64>],
parameter_range: MemoryRange,
reclaim: Option<Range<u64>>,
) -> PartitionInfo {
let mut info = PartitionInfo::new();
info.vtl2_ram = ram
.iter()
.map(|r| MemoryEntry {
range: MemoryRange::try_new(r.clone()).unwrap(),
mem_type: MemoryMapEntryType::VTL2_PROTECTABLE,
vnode: 0,
})
.collect();
info.vtl2_full_config_region = parameter_range;
info.vtl2_config_region_reclaim = reclaim
.map(|r| MemoryRange::try_new(r).unwrap())
.unwrap_or(MemoryRange::EMPTY);
info
}
fn check_e820(boot_params: &boot_params, ext: &E820Ext, expected: &[(Range<u64>, u32)]) {
let actual = boot_params.e820_map[..boot_params.e820_entries as usize]
.iter()
.chain(
ext.entries
.iter()
.take((ext.header.len as usize) / size_of::<e820entry>()),
);
assert_eq!(actual.clone().count(), expected.len());
for (actual, (expected_range, expected_type)) in actual.zip(expected.iter()) {
let addr: u64 = actual.addr.into();
let size: u64 = actual.size.into();
let typ: u32 = actual.typ.into();
assert_eq!(addr, expected_range.start);
assert_eq!(size, expected_range.end - expected_range.start);
assert_eq!(typ, *expected_type);
}
}
const ONE_MB: u64 = 0x10_0000;
#[test]
fn test_e820_basic() {
let mut boot_params: boot_params = FromZeros::new_zeroed();
let mut ext = FromZeros::new_zeroed();
let parameter_range = MemoryRange::try_new(2 * ONE_MB..3 * ONE_MB).unwrap();
let partition_info =
partition_info_ram_ranges(&[ONE_MB..4 * ONE_MB], parameter_range, None);
assert!(
build_e820_map(
&mut boot_params,
&mut ext,
&partition_info,
reserved_memory_regions(&partition_info, None).as_ref(),
)
.is_ok()
);
check_e820(
&boot_params,
&ext,
&[
(ONE_MB..2 * ONE_MB, E820_RAM),
(2 * ONE_MB..3 * ONE_MB, E820_RESERVED),
(3 * ONE_MB..4 * ONE_MB, E820_RAM),
],
);
let mut boot_params: boot_params = FromZeros::new_zeroed();
let mut ext = FromZeros::new_zeroed();
let parameter_range = MemoryRange::try_new(2 * ONE_MB..5 * ONE_MB).unwrap();
let partition_info = partition_info_ram_ranges(
&[ONE_MB..6 * ONE_MB],
parameter_range,
Some(3 * ONE_MB..4 * ONE_MB),
);
assert!(
build_e820_map(
&mut boot_params,
&mut ext,
&partition_info,
reserved_memory_regions(&partition_info, None).as_ref(),
)
.is_ok()
);
check_e820(
&boot_params,
&ext,
&[
(ONE_MB..2 * ONE_MB, E820_RAM),
(2 * ONE_MB..3 * ONE_MB, E820_RESERVED),
(3 * ONE_MB..4 * ONE_MB, E820_RAM),
(4 * ONE_MB..5 * ONE_MB, E820_RESERVED),
(5 * ONE_MB..6 * ONE_MB, E820_RAM),
],
);
let mut boot_params: boot_params = FromZeros::new_zeroed();
let mut ext = FromZeros::new_zeroed();
let parameter_range = MemoryRange::try_new(2 * ONE_MB..5 * ONE_MB).unwrap();
let partition_info = partition_info_ram_ranges(
&[ONE_MB..4 * ONE_MB, 4 * ONE_MB..10 * ONE_MB],
parameter_range,
Some(3 * ONE_MB..4 * ONE_MB),
);
assert!(
build_e820_map(
&mut boot_params,
&mut ext,
&partition_info,
reserved_memory_regions(&partition_info, None).as_ref(),
)
.is_ok()
);
check_e820(
&boot_params,
&ext,
&[
(ONE_MB..2 * ONE_MB, E820_RAM),
(2 * ONE_MB..3 * ONE_MB, E820_RESERVED),
(3 * ONE_MB..4 * ONE_MB, E820_RAM),
(4 * ONE_MB..5 * ONE_MB, E820_RESERVED),
(5 * ONE_MB..10 * ONE_MB, E820_RAM),
],
);
let mut boot_params: boot_params = FromZeros::new_zeroed();
let mut ext = FromZeros::new_zeroed();
let parameter_range = MemoryRange::try_new(2 * ONE_MB..5 * ONE_MB).unwrap();
let partition_info = partition_info_ram_ranges(
&[
ONE_MB..2 * ONE_MB,
2 * ONE_MB..3 * ONE_MB,
3 * ONE_MB..4 * ONE_MB,
4 * ONE_MB..5 * ONE_MB,
5 * ONE_MB..6 * ONE_MB,
6 * ONE_MB..7 * ONE_MB,
7 * ONE_MB..8 * ONE_MB,
],
parameter_range,
Some(3 * ONE_MB..4 * ONE_MB),
);
assert!(
build_e820_map(
&mut boot_params,
&mut ext,
&partition_info,
reserved_memory_regions(&partition_info, None).as_ref(),
)
.is_ok()
);
check_e820(
&boot_params,
&ext,
&[
(ONE_MB..2 * ONE_MB, E820_RAM),
(2 * ONE_MB..3 * ONE_MB, E820_RESERVED),
(3 * ONE_MB..4 * ONE_MB, E820_RAM),
(4 * ONE_MB..5 * ONE_MB, E820_RESERVED),
(5 * ONE_MB..6 * ONE_MB, E820_RAM),
(6 * ONE_MB..7 * ONE_MB, E820_RAM),
(7 * ONE_MB..8 * ONE_MB, E820_RAM),
],
);
}
#[test]
fn test_e820_param_not_covered() {
let mut boot_params: boot_params = FromZeros::new_zeroed();
let mut ext = FromZeros::new_zeroed();
let parameter_range = MemoryRange::try_new(5 * ONE_MB..6 * ONE_MB).unwrap();
let partition_info =
partition_info_ram_ranges(&[ONE_MB..4 * ONE_MB], parameter_range, None);
assert!(
build_e820_map(
&mut boot_params,
&mut ext,
&partition_info,
reserved_memory_regions(&partition_info, None).as_ref(),
)
.is_err()
);
let mut boot_params: boot_params = FromZeros::new_zeroed();
let mut ext = FromZeros::new_zeroed();
let parameter_range = MemoryRange::try_new(3 * ONE_MB..6 * ONE_MB).unwrap();
let partition_info =
partition_info_ram_ranges(&[ONE_MB..4 * ONE_MB], parameter_range, None);
assert!(
build_e820_map(
&mut boot_params,
&mut ext,
&partition_info,
reserved_memory_regions(&partition_info, None).as_ref(),
)
.is_err()
);
let mut boot_params: boot_params = FromZeros::new_zeroed();
let mut ext = FromZeros::new_zeroed();
let parameter_range = MemoryRange::try_new(2 * ONE_MB..5 * ONE_MB).unwrap();
let partition_info =
partition_info_ram_ranges(&[4 * ONE_MB..6 * ONE_MB], parameter_range, None);
assert!(
build_e820_map(
&mut boot_params,
&mut ext,
&partition_info,
reserved_memory_regions(&partition_info, None).as_ref(),
)
.is_err()
);
let mut boot_params: boot_params = FromZeros::new_zeroed();
let mut ext = FromZeros::new_zeroed();
let parameter_range = MemoryRange::try_new(2 * ONE_MB..8 * ONE_MB).unwrap();
let partition_info =
partition_info_ram_ranges(&[4 * ONE_MB..6 * ONE_MB], parameter_range, None);
assert!(
build_e820_map(
&mut boot_params,
&mut ext,
&partition_info,
reserved_memory_regions(&partition_info, None).as_ref(),
)
.is_err()
);
let mut boot_params: boot_params = FromZeros::new_zeroed();
let mut ext = FromZeros::new_zeroed();
let parameter_range = MemoryRange::try_new(2 * ONE_MB..8 * ONE_MB).unwrap();
let partition_info = partition_info_ram_ranges(
&[ONE_MB..6 * ONE_MB, 7 * ONE_MB..10 * ONE_MB],
parameter_range,
None,
);
assert!(
build_e820_map(
&mut boot_params,
&mut ext,
&partition_info,
reserved_memory_regions(&partition_info, None).as_ref(),
)
.is_err()
);
}
#[test]
fn test_e820_huge() {
let mut boot_params: boot_params = FromZeros::new_zeroed();
let mut ext = FromZeros::new_zeroed();
let ram = MemoryRange::new(0..32 * ONE_MB);
let partition_info = partition_info_ram_ranges(&[ram.into()], MemoryRange::EMPTY, None);
let reserved = (0..256)
.map(|i| {
(
MemoryRange::from_4k_gpn_range(i * 8 + 1..i * 8 + 3),
ReservedMemoryType::Vtl2Config,
)
})
.collect::<Vec<_>>();
build_e820_map(&mut boot_params, &mut ext, &partition_info, &reserved).unwrap();
assert!(ext.header.len > 0);
let expected = walk_ranges([(ram, ())], reserved.iter().map(|&(r, _)| (r, ())))
.flat_map(|(range, r)| match r {
RangeWalkResult::Neither => None,
RangeWalkResult::Left(_) => Some((range.into(), E820_RAM)),
RangeWalkResult::Right(_) => unreachable!(),
RangeWalkResult::Both(_, _) => Some((range.into(), E820_RESERVED)),
})
.collect::<Vec<_>>();
check_e820(&boot_params, &ext, &expected);
}
}