1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152
use crate::{process::Pfn, FileWrapper, ProcResult};
use bitflags::bitflags;
use std::{
io::{BufReader, Read, Seek, SeekFrom},
mem::size_of,
path::Path,
};
#[cfg(feature = "serde1")]
use serde::{Deserialize, Serialize};
//const fn genmask(high: usize, low: usize) -> u64 {
// let mask_bits = size_of::<u64>() * 8;
// (!0 - (1 << low) + 1) & (!0 >> (mask_bits - 1 - high))
//}
bitflags! {
/// Represents the fields and flags in a page table entry for a memory page.
#[cfg_attr(feature = "serde1", derive(Serialize, Deserialize))]
pub struct PhysicalPageFlags: u64 {
/// The page is being locked for exclusive access, e.g. by undergoing read/write IO
const LOCKED = 1 << 0;
/// IO error occurred
const ERROR = 1 << 1;
/// The page has been referenced since last LRU list enqueue/requeue
const REFERENCED = 1 << 2;
/// The page has up-to-date data. ie. for file backed page: (in-memory data revision >= on-disk one)
const UPTODATE = 1 << 3;
/// The page has been written to, hence contains new data. i.e. for file backed page: (in-memory data revision > on-disk one)
const DIRTY = 1 << 4;
/// The page is in one of the LRU lists
const LRU = 1 << 5;
/// The page is in the active LRU list
const ACTIVE = 1 << 6;
/// The page is managed by the SLAB/SLOB/SLUB/SLQB kernel memory allocator. When compound page is used, SLUB/SLQB will only set this flag on the head page; SLOB will not flag it at all
const SLAB = 1 << 7;
/// The page is being synced to disk
const WRITEBACK = 1 << 8;
/// The page will be reclaimed soon after its pageout IO completed
const RECLAIM = 1 << 9;
/// A free memory block managed by the buddy system allocator. The buddy system organizes free memory in blocks of various orders. An order N block has 2^N physically contiguous pages, with the BUDDY flag set for and _only_ for the first page
const BUDDY = 1 << 10;
/// A memory mapped page
const MMAP = 1 << 11;
/// A memory mapped page that is not part of a file
const ANON = 1 << 12;
/// The page is mapped to swap space, i.e. has an associated swap entry
const SWAPCACHE = 1 << 13;
/// The page is backed by swap/RAM
const SWAPBACKED = 1 << 14;
/// A compound page with order N consists of 2^N physically contiguous pages. A compound page with order 2 takes the form of “HTTT”, where H donates its head page and T donates its tail page(s). The major consumers of compound pages are hugeTLB pages (<https://www.kernel.org/doc/html/latest/admin-guide/mm/hugetlbpage.html#hugetlbpage>), the SLUB etc. memory allocators and various device drivers. However in this interface, only huge/giga pages are made visible to end users
const COMPOUND_HEAD = 1 << 15;
/// A compound page tail (see description above)
const COMPOUND_TAIL = 1 << 16;
/// This is an integral part of a HugeTLB page
const HUGE = 1 << 17;
/// The page is in the unevictable (non-)LRU list It is somehow pinned and not a candidate for LRU page reclaims, e.g. ramfs pages, shmctl(SHM_LOCK) and mlock() memory segments
const UNEVICTABLE = 1 << 18;
/// Hardware detected memory corruption on this page: don’t touch the data!
const HWPOISON = 1 << 19;
/// No page frame exists at the requested address
const NOPAGE = 1 << 20;
/// Identical memory pages dynamically shared between one or more processes
const KSM = 1 << 21;
/// Contiguous pages which construct transparent hugepages
const THP = 1 << 22;
/// The page is logically offline
const OFFLINE = 1 << 23;
/// Zero page for pfn_zero or huge_zero page
const ZERO_PAGE = 1 << 24;
/// The page has not been accessed since it was marked idle (see <https://www.kernel.org/doc/html/latest/admin-guide/mm/idle_page_tracking.html#idle-page-tracking>). Note that this flag may be stale in case the page was accessed via a PTE. To make sure the flag is up-to-date one has to read /sys/kernel/mm/page_idle/bitmap first
const IDLE = 1 << 25;
/// The page is in use as a page table
const PGTABLE = 1 << 26;
}
}
impl PhysicalPageFlags {
pub(crate) fn parse_info(info: u64) -> Self {
PhysicalPageFlags::from_bits_truncate(info)
}
}
/// Parse physical memory flags accessing `/proc/kpageflags`.
///
/// Require root or CAP_SYS_ADMIN
pub struct KPageFlags {
reader: BufReader<FileWrapper>,
}
impl KPageFlags {
/// Get a parser from default `/proc/kpageflags`
///
/// Return `Err` if process is not running as root or don't have CAP_SYS_ADMIN
pub fn new() -> ProcResult<Self> {
Self::from_custom_root("/proc")
}
/// Get a parser from custom `/proc`
///
/// Return `Err` if process is not running as root or don't have CAP_SYS_ADMIN
pub fn from_custom_root<P: AsRef<Path>>(root: P) -> ProcResult<Self> {
let mut path = root.as_ref().to_path_buf();
path.push("kpageflags");
let reader = BufReader::new(FileWrapper::open(path)?);
Ok(Self { reader })
}
/// Retrieve information in the page table entry for the PFN (page frame number) at index `page_index`.
/// If you need to retrieve multiple PFNs, opt for [Self::get_range_info()] instead.
///
/// Return Err if the PFN is not in RAM (see [crate::iomem()]):
/// Io(Error { kind: UnexpectedEof, message: "failed to fill whole buffer" }, None)
pub fn get_info(&mut self, pfn: Pfn) -> ProcResult<PhysicalPageFlags> {
self.get_range_info(pfn, Pfn(pfn.0 + 1))
.map(|mut vec| vec.pop().unwrap())
}
/// Retrieve information in the page table entry for the PFNs within range `start` (included) and `end` (excluded) PFNs.
///
/// Return Err if any PFN is not in RAM (see [crate::iomem()]):
/// Io(Error { kind: UnexpectedEof, message: "failed to fill whole buffer" }, None)
pub fn get_range_info(&mut self, start: Pfn, end: Pfn) -> ProcResult<Vec<PhysicalPageFlags>> {
let start_position = start.0 * size_of::<PhysicalPageFlags>() as u64;
self.reader.seek(SeekFrom::Start(start_position))?;
let mut page_infos = Vec::with_capacity((end.0 - start.0) as usize);
for _ in start.0..end.0 {
let mut info_bytes = [0; size_of::<PhysicalPageFlags>()];
self.reader.read_exact(&mut info_bytes)?;
page_infos.push(PhysicalPageFlags::parse_info(u64::from_ne_bytes(info_bytes)));
}
Ok(page_infos)
}
}
#[cfg(test)]
mod tests {
use super::*;
#[test]
fn test_kpageflags_parsing() {
let pagemap_entry: u64 = 0b0000000000000000000000000000000000000000000000000000000000000001;
let info = PhysicalPageFlags::parse_info(pagemap_entry);
assert!(info == PhysicalPageFlags::LOCKED);
}
}