Permalink
Cannot retrieve contributors at this time
Join GitHub today
GitHub is home to over 28 million developers working together to host and review code, manage projects, and build software together.
Sign up
Fetching contributors…
| //! Memory Management Unit - Translates virtual memory addresses to | |
| //! physical addresses. Memory is grouped into tabulated Pages. This module | |
| //! defines the Page(uint) and Table<U> implementations. | |
| use core::mem::size_of; | |
| use core::ptr::copy_nonoverlapping_memory; | |
| use core::fmt; | |
| use core::prelude::*; | |
| use core; | |
| use kernel::mm::physical; | |
| use kernel::mm::physical::Phys; | |
| use util::rt; | |
| use kernel; | |
| pub type Frame = [u8, ..PAGE_SIZE]; | |
| bitflags!(flags Flags: uint { | |
| const PRESENT = 1 << 0, | |
| const RW = 1 << 1, | |
| const USER = 1 << 2, | |
| const ACCESSED = 1 << 5, | |
| const HUGE = 1 << 7 | |
| }) | |
| #[repr(packed)] | |
| pub struct Page(uint); | |
| const PAGE_SIZE: uint = 0x1000; | |
| const PAGE_SIZE_LOG2: uint = 12; | |
| const ENTRIES: uint = 1024; | |
| const DIR_VADDR: uint = 0xFFFFF000; | |
| struct VMemLayout { | |
| temp1: PageDirectory, // @ 0xFF7FF000 | |
| temp_tables: [PageTable, ..ENTRIES - 1], // @ 0xFF800000 | |
| temp: PageDirectory, // @ 0xFFBFF000 | |
| tables: [PageTable, ..ENTRIES - 1], // @ 0xFFC00000 | |
| dir: PageDirectory // @ 0xFFFFF000 | |
| } | |
| static VMEM: *mut VMemLayout = 0xFF7FF000u as *mut VMemLayout; | |
| // U: underlying element type | |
| #[repr(packed)] | |
| struct Table<U> { | |
| entries: [Page, ..ENTRIES] | |
| } | |
| #[repr(packed)] | |
| struct Directory<U = PageTable> { | |
| entries: [U, ..ENTRIES] | |
| } | |
| pub type PageTable = Table<Page>; | |
| pub type PageDirectory = Table<Table<Page>>; | |
| pub unsafe fn init() { | |
| let dir: Phys<PageDirectory> = physical::zero_alloc_frames(1); | |
| let table: Phys<PageTable> = physical::alloc_frames(1); | |
| (*table.as_ptr()).identity_map(0, PRESENT | RW); | |
| (*dir.as_ptr()).set_addr(0 as *mut u8, table, PRESENT | RW); | |
| // Map the directory as its own last table. | |
| // When accessing its virtual address(...) | |
| (*dir.as_ptr()).map_self(dir); | |
| kernel::int_table.map(|mut t| { | |
| use super::exception::{Fault, exception_handler}; | |
| t.set_isr(Fault::PageFault, true, exception_handler()); | |
| }); | |
| switch_directory(dir); | |
| enable_paging(); | |
| } | |
| pub fn switch_directory(dir: Phys<PageDirectory>) { | |
| use super::CR3; | |
| CR3::write(dir.as_ptr()); | |
| } | |
| fn enable_paging() { | |
| use super::{CR0, CR0_PG}; | |
| CR0::write(CR0 | CR0_PG); | |
| } | |
| pub unsafe fn map(page_ptr: *mut u8, len: uint, flags: Flags) { | |
| (*VMEM).dir.map(page_ptr, len, flags); | |
| } | |
| #[inline] | |
| fn flush_tlb<T>(addr: T) { | |
| unsafe { | |
| asm!("invlpg [$0]" :: "r"(addr) : "memory" : "volatile", "intel") | |
| } | |
| } | |
| impl Page { | |
| fn new<T>(addr: Phys<T>, flags: Flags) -> Page { | |
| Page(addr.to_uint()) | flags | |
| } | |
| fn at_frame(i: uint, flags: Flags) -> Page { | |
| Page(i * PAGE_SIZE) | flags | |
| } | |
| fn physical<P>(&self) -> Phys<P> { | |
| let &Page(p) = self; | |
| Phys::at(p & 0xFFFFF000) | |
| } | |
| fn is_present(self) -> bool { | |
| self.contains(PRESENT) | |
| } | |
| fn contains(&self, flags: Flags) -> bool { | |
| let &Page(bits) = self; | |
| (bits & flags.bits) == flags.bits | |
| } | |
| } | |
| impl core::ops::BitOr<Flags, Page> for Page { | |
| #[inline(always)] | |
| fn bitor(&self, other: &Flags) -> Page { | |
| let &Page(bits) = self; | |
| Page(bits | other.bits) | |
| } | |
| } | |
| impl fmt::Show for Page { | |
| fn fmt(&self, fmt: &mut fmt::Formatter) -> fmt::Result { | |
| let &Page(p) = self; | |
| let page = p & 0xFFFFF000; | |
| let (p, r, u, a) = ( | |
| if self.contains(PRESENT) { 'P' } else { ' ' }, | |
| if self.contains(RW) { 'R' } else { ' ' }, | |
| if self.contains(USER) { 'U' } else { ' ' }, | |
| if self.contains(ACCESSED) { 'A' } else { ' ' } | |
| ); | |
| write!(fmt, "0x{:x}({}{}{}{})", page, p, r, u, a) | |
| } | |
| } | |
| impl<U> Table<U> { | |
| fn set_addr<S, T>(&mut self, vaddr: *mut S, phys: Phys<T>, flags: Flags) { | |
| // FIXME error: internal compiler error: missing default for a not explicitely provided type param | |
| self.set(vaddr as uint, Page::new(phys, flags)); | |
| flush_tlb(vaddr); | |
| } | |
| fn set(&mut self, addr: uint, page: Page) { // TODO addr: Phys<T> | |
| // update entry, based on the underlying type (page, table) | |
| let size = size_of::<U>() / size_of::<Page>() * PAGE_SIZE; | |
| let index = (addr / size) % ENTRIES; | |
| self.entries[index] = page; | |
| } | |
| fn get(&self, addr: uint) -> Page { | |
| let size = size_of::<U>() / size_of::<Page>() * PAGE_SIZE; | |
| let index = (addr / size) % ENTRIES; | |
| self.entries[index] | |
| } | |
| } | |
| impl Table<Page> { | |
| fn identity_map(&mut self, start: uint, flags: Flags) { | |
| for i in range(0, ENTRIES) { | |
| self.entries[i] = Page::at_frame(start + i, flags); | |
| } | |
| } | |
| } | |
| // Can't impl on typedefs. Rust #9767 | |
| impl Table<Table<Page>> { | |
| fn fetch_table<T>(&mut self, vptr: *mut T, flags: Flags) -> *mut PageTable { | |
| match self.get(vptr as uint) { | |
| table @ Page(_) if table.is_present() => { | |
| table.physical().as_ptr() | |
| } | |
| _ => unsafe { // allocate table | |
| let table: Phys<PageTable> = physical::zero_alloc_frames(1); | |
| self.set_addr(vptr, table, flags); // page fault | |
| // flush_tlb(table); | |
| table.as_ptr() | |
| } | |
| } | |
| } | |
| pub unsafe fn set_page<T>(&mut self, vptr: *mut T, phys: Phys<T>, flags: Flags) -> *mut T { | |
| let table = self.fetch_table(vptr, flags); | |
| (*table).set_addr(vptr, phys, flags); | |
| vptr | |
| } | |
| pub unsafe fn map_frame(&mut self, vptr: *mut u8, flags: Flags) { | |
| self.set_page(vptr, physical::alloc_frames(1), flags | PRESENT); | |
| } | |
| pub fn map(&mut self, mut page_ptr: *mut u8, len: uint, flags: Flags) { | |
| // TODO: optimize with uints? | |
| unsafe { | |
| let end = page_ptr.offset(len as int); | |
| while page_ptr < end { | |
| let frame = physical::alloc_frames(1); | |
| self.set_page(page_ptr, frame, flags | PRESENT); | |
| (*VMEM).dir.set_page(page_ptr, frame, flags | PRESENT); | |
| page_ptr = page_ptr.offset(PAGE_SIZE as int); | |
| } | |
| } | |
| } | |
| fn map_self(&mut self, this: Phys<PageDirectory>) { | |
| self.set(DIR_VADDR as uint, Page::new(this, PRESENT | RW)); | |
| } | |
| pub fn clone(&self) -> Phys<PageDirectory> { | |
| unsafe { | |
| // new directory | |
| let dir_phys: Phys<PageDirectory> = physical::zero_alloc_frames(1); | |
| let &VMemLayout { ref mut temp1, ref mut dir, .. } = &mut *VMEM; | |
| dir.set_page(temp1, dir_phys, PRESENT | RW); | |
| temp1.map_self(dir_phys); | |
| let cnt = 0xC0000000 / (ENTRIES * PAGE_SIZE); | |
| copy_nonoverlapping_memory(&mut temp1.entries[0] as *mut Page, &self.entries as *const Page, cnt); | |
| dir_phys | |
| } | |
| } | |
| } | |
| pub fn clone_directory() -> Phys<PageDirectory> { | |
| unsafe { | |
| (*VMEM).dir.clone() | |
| } | |
| } |