use core::{convert::TryFrom, fmt};
use ax_page_table_entry::{GenericPTE, MappingFlags};
use ax_page_table_multiarch::{PageTable64, PagingMetaData};
use bit_field::BitField;
use crate::{GuestPhysAddr, HostPhysAddr};
bitflags::bitflags! {
struct EPTFlags: u64 {
const READ = 1 << 0;
const WRITE = 1 << 1;
const EXECUTE = 1 << 2;
const MEM_TYPE_MASK = 0b111 << 3;
const IGNORE_PAT = 1 << 6;
const HUGE_PAGE = 1 << 7;
const ACCESSED = 1 << 8;
const DIRTY = 1 << 9;
const EXECUTE_FOR_USER = 1 << 10;
}
}
numeric_enum_macro::numeric_enum! {
#[repr(u8)]
#[derive(Debug, PartialEq, Clone, Copy)]
enum EPTMemType {
Uncached = 0,
WriteCombining = 1,
WriteThrough = 4,
WriteProtected = 5,
WriteBack = 6,
}
}
impl EPTFlags {
fn set_mem_type(&mut self, mem_type: EPTMemType) {
let mut bits = self.bits();
bits.set_bits(3..6, mem_type as u64);
*self = Self::from_bits_truncate(bits)
}
fn mem_type(&self) -> Result<EPTMemType, u8> {
EPTMemType::try_from(self.bits().get_bits(3..6) as u8)
}
}
impl From<MappingFlags> for EPTFlags {
fn from(f: MappingFlags) -> Self {
if f.is_empty() {
return Self::empty();
}
let mut ret = Self::empty();
if f.contains(MappingFlags::READ) {
ret |= Self::READ;
}
if f.contains(MappingFlags::WRITE) {
ret |= Self::WRITE;
}
if f.contains(MappingFlags::EXECUTE) {
ret |= Self::EXECUTE;
}
if !f.contains(MappingFlags::DEVICE) {
ret.set_mem_type(EPTMemType::WriteBack);
}
ret
}
}
impl From<EPTFlags> for MappingFlags {
fn from(f: EPTFlags) -> Self {
let mut ret = MappingFlags::empty();
if f.contains(EPTFlags::READ) {
ret |= Self::READ;
}
if f.contains(EPTFlags::WRITE) {
ret |= Self::WRITE;
}
if f.contains(EPTFlags::EXECUTE) {
ret |= Self::EXECUTE;
}
if let Ok(EPTMemType::Uncached) = f.mem_type() {
ret |= Self::DEVICE;
}
ret
}
}
#[derive(Clone, Copy)]
#[repr(transparent)]
pub struct EPTEntry(u64);
impl EPTEntry {
const PHYS_ADDR_MASK: u64 = 0x000f_ffff_ffff_f000; }
impl GenericPTE for EPTEntry {
fn new_page(paddr: HostPhysAddr, flags: MappingFlags, is_huge: bool) -> Self {
let mut flags = EPTFlags::from(flags);
if is_huge {
flags |= EPTFlags::HUGE_PAGE;
}
Self(flags.bits() | (paddr.as_usize() as u64 & Self::PHYS_ADDR_MASK))
}
fn new_table(paddr: HostPhysAddr) -> Self {
let flags = EPTFlags::READ | EPTFlags::WRITE | EPTFlags::EXECUTE;
Self(flags.bits() | (paddr.as_usize() as u64 & Self::PHYS_ADDR_MASK))
}
fn paddr(&self) -> HostPhysAddr {
HostPhysAddr::from((self.0 & Self::PHYS_ADDR_MASK) as usize)
}
fn flags(&self) -> MappingFlags {
EPTFlags::from_bits_truncate(self.0).into()
}
fn set_paddr(&mut self, paddr: HostPhysAddr) {
self.0 = (self.0 & !Self::PHYS_ADDR_MASK) | (paddr.as_usize() as u64 & Self::PHYS_ADDR_MASK)
}
fn set_flags(&mut self, flags: MappingFlags, is_huge: bool) {
let mut flags = EPTFlags::from(flags);
if is_huge {
flags |= EPTFlags::HUGE_PAGE;
}
self.0 = (self.0 & Self::PHYS_ADDR_MASK) | flags.bits()
}
fn is_unused(&self) -> bool {
self.0 == 0
}
fn is_present(&self) -> bool {
self.0 & 0x7 != 0 }
fn is_huge(&self) -> bool {
EPTFlags::from_bits_truncate(self.0).contains(EPTFlags::HUGE_PAGE)
}
fn clear(&mut self) {
self.0 = 0
}
fn bits(self) -> usize {
self.0 as usize
}
}
impl fmt::Debug for EPTEntry {
fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result {
f.debug_struct("EPTEntry")
.field("raw", &self.0)
.field("hpaddr", &self.paddr())
.field("flags", &self.flags())
.field("mem_type", &EPTFlags::from_bits_truncate(self.0).mem_type())
.finish()
}
}
pub struct ExtendedPageTableMetadata;
impl PagingMetaData for ExtendedPageTableMetadata {
const LEVELS: usize = 4;
const PA_MAX_BITS: usize = 52;
const VA_MAX_BITS: usize = 48;
type VirtAddr = GuestPhysAddr;
#[allow(unused_variables)]
fn flush_tlb(vaddr: Option<GuestPhysAddr>) {
#[cfg(target_os = "none")]
{
if let Some(vaddr) = vaddr {
unsafe { x86::tlb::flush(vaddr.into()) }
} else {
unsafe { x86::tlb::flush_all() }
}
}
}
}
pub type ExtendedPageTable<H> = PageTable64<ExtendedPageTableMetadata, EPTEntry, H>;