use super::VpContextBuilder;
use super::VpContextState;
use crate::vp_context_builder::VpContextPageState;
use igvm_defs::PAGE_SIZE_4K;
use loader::importer::SegmentRegister;
use loader::importer::X86Register;
use std::mem::offset_of;
use x86defs::X64_EFER_LME;
use x86defs::X86X_MSR_DEFAULT_PAT;
use zerocopy::Immutable;
use zerocopy::IntoBytes;
use zerocopy::KnownLayout;
#[repr(C)]
#[derive(Debug, Default, Clone, Copy, IntoBytes, Immutable, KnownLayout)]
pub struct TdxTrampolineContext {
start_gate: u32,
data_selector: u16,
static_gdt_limit: u16,
static_gdt_base: u32,
task_selector: u16,
idtr_limit: u16,
idtr_base: u64,
initial_rip: u64,
code_selector: u16,
padding_2: [u16; 2],
gdtr_limit: u16,
gdtr_base: u64,
rsp: u64,
rbp: u64,
rsi: u64,
r8: u64,
r9: u64,
r10: u64,
r11: u64,
cr0: u64,
cr3: u64,
cr4: u64,
transition_cr3: u32,
padding_3: u32,
static_gdt: [u8; 16],
}
#[derive(Debug)]
pub struct TdxHardwareContext {
trampoline_context: TdxTrampolineContext,
accept_lower_1mb: bool,
}
impl TdxHardwareContext {
pub fn new(accept_lower_1mb: bool) -> Self {
Self {
trampoline_context: TdxTrampolineContext::default(),
accept_lower_1mb,
}
}
}
impl VpContextBuilder for TdxHardwareContext {
type Register = X86Register;
fn import_vp_register(&mut self, register: X86Register) {
let mut set_data_selector = |reg: SegmentRegister| {
if self.trampoline_context.data_selector == 0 {
self.trampoline_context.data_selector = reg.selector;
} else if self.trampoline_context.data_selector != reg.selector {
panic!("data selectors must be the same");
}
};
match register {
X86Register::Gdtr(reg) => {
self.trampoline_context.gdtr_base = reg.base;
self.trampoline_context.gdtr_limit = reg.limit;
}
X86Register::Idtr(reg) => {
self.trampoline_context.idtr_base = reg.base;
self.trampoline_context.idtr_limit = reg.limit;
}
X86Register::Ds(reg)
| X86Register::Es(reg)
| X86Register::Fs(reg)
| X86Register::Gs(reg)
| X86Register::Ss(reg) => set_data_selector(reg),
X86Register::Cs(reg) => self.trampoline_context.code_selector = reg.selector,
X86Register::Tr(reg) => {
self.trampoline_context.task_selector = reg.selector;
}
X86Register::Cr0(cr0) => self.trampoline_context.cr0 = cr0,
X86Register::Cr3(cr3) => {
let cr3_u32: u32 = cr3.try_into().expect("cr3 must fit in u32");
self.trampoline_context.transition_cr3 = cr3_u32;
self.trampoline_context.cr3 = cr3;
}
X86Register::Cr4(cr4) => self.trampoline_context.cr4 = cr4,
X86Register::Efer(efer) => {
if efer & X64_EFER_LME == 0 {
panic!("EFER LME must be set for tdx")
}
}
X86Register::Pat(pat) => {
if pat != X86X_MSR_DEFAULT_PAT {
panic!("PAT must be default for tdx")
}
}
X86Register::Rbp(rbp) => self.trampoline_context.rbp = rbp,
X86Register::Rip(rip) => self.trampoline_context.initial_rip = rip,
X86Register::Rsi(rsi) => self.trampoline_context.rsi = rsi,
X86Register::Rsp(rsp) => self.trampoline_context.rsp = rsp,
X86Register::R8(r8) => self.trampoline_context.r8 = r8,
X86Register::R9(r9) => self.trampoline_context.r9 = r9,
X86Register::R10(r10) => self.trampoline_context.r10 = r10,
X86Register::R11(r11) => self.trampoline_context.r11 = r11,
X86Register::R12(_) => panic!("r12 not allowed for tdx"),
X86Register::Rflags(_) => panic!("rflags not allowed for tdx"),
X86Register::MtrrDefType(_)
| X86Register::MtrrPhysBase0(_)
| X86Register::MtrrPhysMask0(_)
| X86Register::MtrrPhysBase1(_)
| X86Register::MtrrPhysMask1(_)
| X86Register::MtrrPhysBase2(_)
| X86Register::MtrrPhysMask2(_)
| X86Register::MtrrPhysBase3(_)
| X86Register::MtrrPhysMask3(_)
| X86Register::MtrrPhysBase4(_)
| X86Register::MtrrPhysMask4(_)
| X86Register::MtrrFix64k00000(_)
| X86Register::MtrrFix16k80000(_)
| X86Register::MtrrFix4kE0000(_)
| X86Register::MtrrFix4kE8000(_)
| X86Register::MtrrFix4kF0000(_)
| X86Register::MtrrFix4kF8000(_) => {
tracing::warn!(?register, "Ignoring MTRR register for TDX.")
}
}
}
fn set_vp_context_memory(&mut self, _page_base: u64) {
unimplemented!("not supported for TDX");
}
fn finalize(&mut self, state: &mut Vec<VpContextState>) {
self.trampoline_context.static_gdt[0x08] = 0xFF;
self.trampoline_context.static_gdt[0x09] = 0xFF;
self.trampoline_context.static_gdt[0x0D] = 0x9B;
self.trampoline_context.static_gdt[0x0E] = 0xA0;
self.trampoline_context.static_gdt_limit = 0xF;
self.trampoline_context.static_gdt_base =
0xFFFFF000 + offset_of!(TdxTrampolineContext, static_gdt) as u32;
let mut byte_offset = 0xFF0;
let mut reset_page = vec![0xCCu8; PAGE_SIZE_4K as usize];
let trampoline_context = self.trampoline_context.as_bytes();
reset_page[0..trampoline_context.len()].copy_from_slice(trampoline_context);
let copy_instr =
|trampoline_page: &mut Vec<u8>, byte_offset, instruction: &[u8]| -> usize {
trampoline_page[byte_offset..byte_offset + instruction.len()]
.copy_from_slice(instruction);
byte_offset + instruction.len()
};
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0xE9]);
let mut relative_offset =
(trampoline_context.len() as u32).wrapping_sub((byte_offset + 4) as u32);
copy_instr(&mut reset_page, byte_offset, relative_offset.as_bytes());
byte_offset = trampoline_context.len();
let l0_offset = byte_offset;
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0x3B, 0x35]);
relative_offset = 0xFFFFF000 + offset_of!(TdxTrampolineContext, start_gate) as u32;
byte_offset = copy_instr(&mut reset_page, byte_offset, relative_offset.as_bytes());
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0x75]);
let jne_l0_offset = (l0_offset.wrapping_sub(byte_offset + 1)) as u8;
byte_offset = copy_instr(&mut reset_page, byte_offset, &[jne_l0_offset]);
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0x0F, 0x01, 0x15]);
relative_offset = 0xFFFFF000 + offset_of!(TdxTrampolineContext, static_gdt_limit) as u32;
byte_offset = copy_instr(&mut reset_page, byte_offset, relative_offset.as_bytes());
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0x8B, 0x05]);
relative_offset = 0xFFFFF000 + offset_of!(TdxTrampolineContext, cr4) as u32;
byte_offset = copy_instr(&mut reset_page, byte_offset, relative_offset.as_bytes());
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0x0F, 0x22, 0xE0]);
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0x8B, 0x05]);
relative_offset = 0xFFFFF000 + offset_of!(TdxTrampolineContext, transition_cr3) as u32;
byte_offset = copy_instr(&mut reset_page, byte_offset, relative_offset.as_bytes());
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0x0F, 0x22, 0xD8]);
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0x8B, 0x05]);
relative_offset = 0xFFFFF000 + offset_of!(TdxTrampolineContext, cr0) as u32;
byte_offset = copy_instr(&mut reset_page, byte_offset, relative_offset.as_bytes());
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0x0F, 0x22, 0xC0]);
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0xEA]);
relative_offset = 0xFFFFF000 + byte_offset as u32 + 6;
byte_offset = copy_instr(&mut reset_page, byte_offset, relative_offset.as_bytes());
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0x08, 0x00]);
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0x48, 0x8B, 0x05]);
relative_offset =
(offset_of!(TdxTrampolineContext, cr3) as u32).wrapping_sub((byte_offset + 4) as u32);
byte_offset = copy_instr(&mut reset_page, byte_offset, relative_offset.as_bytes());
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0x0F, 0x22, 0xD8]);
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0x66, 0x8B, 0x05]);
relative_offset = (offset_of!(TdxTrampolineContext, gdtr_limit) as u32)
.wrapping_sub((byte_offset + 4) as u32);
byte_offset = copy_instr(&mut reset_page, byte_offset, relative_offset.as_bytes());
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0x66, 0x85, 0xC0]);
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0x74]);
byte_offset += 1;
let l4_offset = byte_offset as u32;
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0x0F, 0x01, 0x15]);
relative_offset = (offset_of!(TdxTrampolineContext, gdtr_limit) as u32)
.wrapping_sub((byte_offset + 4) as u32);
byte_offset = copy_instr(&mut reset_page, byte_offset, relative_offset.as_bytes());
reset_page[l0_offset.wrapping_sub(1)] = (byte_offset.wrapping_sub(l0_offset)) as u8;
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0x66, 0x8B, 0x05]);
relative_offset = (offset_of!(TdxTrampolineContext, idtr_limit) as u32)
.wrapping_sub((byte_offset + 4) as u32);
byte_offset = copy_instr(&mut reset_page, byte_offset, relative_offset.as_bytes());
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0x66, 0x85, 0xC0]);
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0x74]);
byte_offset += 1;
let jump_offset = byte_offset;
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0x0F, 0x01, 0x1D]);
relative_offset = (offset_of!(TdxTrampolineContext, idtr_limit) as u32)
.wrapping_sub((byte_offset + 4) as u32);
byte_offset = copy_instr(&mut reset_page, byte_offset, relative_offset.as_bytes());
reset_page[jump_offset.wrapping_sub(1)] = (byte_offset.wrapping_sub(jump_offset)) as u8;
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0x66, 0x8B, 0x05]);
relative_offset = (offset_of!(TdxTrampolineContext, data_selector) as u32)
.wrapping_sub((byte_offset + 4) as u32);
byte_offset = copy_instr(&mut reset_page, byte_offset, relative_offset.as_bytes());
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0x8E, 0xD0]);
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0x8E, 0xD8]);
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0x8E, 0xC0]);
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0x8E, 0xE0]);
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0x8E, 0xE8]);
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0x66, 0x8B, 0x05]);
relative_offset = (offset_of!(TdxTrampolineContext, task_selector) as u32)
.wrapping_sub((byte_offset + 4) as u32);
byte_offset = copy_instr(&mut reset_page, byte_offset, relative_offset.as_bytes());
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0x66, 0x85, 0xC0]);
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0x74]);
byte_offset += 1;
let jump_offset = byte_offset;
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0x0F, 0x00, 0xD8]);
reset_page[jump_offset.wrapping_sub(1)] = (byte_offset.wrapping_sub(jump_offset)) as u8;
reset_page[(l4_offset as usize).wrapping_sub(1)] =
(byte_offset.wrapping_sub(l4_offset as usize)) as u8;
if self.accept_lower_1mb {
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0x85, 0xF6]);
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0x75]);
byte_offset += 1;
let l3_offset = byte_offset;
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0x33, 0xC9]);
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0x33, 0xD2]);
byte_offset = copy_instr(
&mut reset_page,
byte_offset,
&[0xBF, 0x00, 0x00, 0x10, 0x00],
);
let jump_offset = byte_offset;
byte_offset = copy_instr(
&mut reset_page,
byte_offset,
&[0xB8, 0x06, 0x00, 0x00, 0x00],
);
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0x66, 0x0F, 0x01, 0xCC]);
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0x48, 0x85, 0xC0]);
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0x0F, 0x85]);
byte_offset += 4;
let relative_offset = 0xFEF - byte_offset;
copy_instr(
&mut reset_page,
byte_offset.wrapping_sub(4),
relative_offset.as_bytes(),
);
byte_offset = copy_instr(
&mut reset_page,
byte_offset,
&[0x81, 0xC1, 0x00, 0x10, 0x00, 0x00],
);
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0x3B, 0xCF]);
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0x72]);
byte_offset += 1;
reset_page[byte_offset.wrapping_sub(1)] = (jump_offset.wrapping_sub(byte_offset)) as u8;
reset_page[l3_offset.wrapping_sub(1)] = (byte_offset.wrapping_sub(l3_offset)) as u8;
}
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0x48, 0x8B, 0x25]);
relative_offset =
(offset_of!(TdxTrampolineContext, rsp) as u32).wrapping_sub((byte_offset + 4) as u32);
byte_offset = copy_instr(&mut reset_page, byte_offset, relative_offset.as_bytes());
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0x48, 0x8B, 0x2D]);
relative_offset =
(offset_of!(TdxTrampolineContext, rbp) as u32).wrapping_sub((byte_offset + 4) as u32);
byte_offset = copy_instr(&mut reset_page, byte_offset, relative_offset.as_bytes());
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0x8B, 0xCE]);
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0x48, 0x8B, 0x35]);
relative_offset =
(offset_of!(TdxTrampolineContext, rsi) as u32).wrapping_sub((byte_offset + 4) as u32);
byte_offset = copy_instr(&mut reset_page, byte_offset, relative_offset.as_bytes());
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0x4C, 0x8B, 0x05]);
relative_offset =
(offset_of!(TdxTrampolineContext, r8) as u32).wrapping_sub((byte_offset + 4) as u32);
byte_offset = copy_instr(&mut reset_page, byte_offset, relative_offset.as_bytes());
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0x4C, 0x8B, 0x0D]);
relative_offset =
(offset_of!(TdxTrampolineContext, r9) as u32).wrapping_sub((byte_offset + 4) as u32);
byte_offset = copy_instr(&mut reset_page, byte_offset, relative_offset.as_bytes());
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0x4C, 0x8B, 0x15]);
relative_offset =
(offset_of!(TdxTrampolineContext, r10) as u32).wrapping_sub((byte_offset + 4) as u32);
byte_offset = copy_instr(&mut reset_page, byte_offset, relative_offset.as_bytes());
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0x4C, 0x8B, 0x1D]);
relative_offset =
(offset_of!(TdxTrampolineContext, r11) as u32).wrapping_sub((byte_offset + 4) as u32);
byte_offset = copy_instr(&mut reset_page, byte_offset, relative_offset.as_bytes());
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0x66, 0x8B, 0x05]);
relative_offset = (offset_of!(TdxTrampolineContext, code_selector) as u32)
.wrapping_sub((byte_offset + 4) as u32);
byte_offset = copy_instr(&mut reset_page, byte_offset, relative_offset.as_bytes());
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0x66, 0x85, 0xC0]);
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0x74]);
byte_offset += 1;
let jump_offset = byte_offset;
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0x48, 0xFF, 0x2D]);
relative_offset = (offset_of!(TdxTrampolineContext, initial_rip) as u32)
.wrapping_sub((byte_offset + 4) as u32);
byte_offset = copy_instr(&mut reset_page, byte_offset, relative_offset.as_bytes());
reset_page[jump_offset.wrapping_sub(1)] = (byte_offset.wrapping_sub(jump_offset)) as u8;
byte_offset = copy_instr(&mut reset_page, byte_offset, &[0x48, 0xFF, 0x25]);
relative_offset = (offset_of!(TdxTrampolineContext, initial_rip) as u32)
.wrapping_sub((byte_offset + 4) as u32);
copy_instr(&mut reset_page, byte_offset, relative_offset.as_bytes());
state.push(VpContextState::Page(VpContextPageState {
page_base: 0xFFFFF,
page_count: 1,
acceptance: loader::importer::BootPageAcceptance::Exclusive,
data: reset_page,
}));
}
}