diff --git a/src/arch/aarch64/memory/mod.rs b/src/arch/aarch64/memory/mod.rs
index 62ce461..ac10ab7 100644
--- a/src/arch/aarch64/memory/mod.rs
+++ b/src/arch/aarch64/memory/mod.rs
@@ -35,3 +35,293 @@ pub trait FrameAllocator {
fn allocate_frame(&mut self) -> Option;
fn deallocate_frame(&mut self, frame: Frame);
}
+
+// --------------------------------------------
+
+/*
+ * MIT License
+ *
+ * Copyright (c) 2019 Andre Richter
+ *
+ * Permission is hereby granted, free of charge, to any person obtaining a copy
+ * of this software and associated documentation files (the "Software"), to deal
+ * in the Software without restriction, including without limitation the rights
+ * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
+ * copies of the Software, and to permit persons to whom the Software is
+ * furnished to do so, subject to the following conditions:
+ *
+ * The above copyright notice and this permission notice shall be included in all
+ * copies or substantial portions of the Software.
+ *
+ * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
+ * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
+ * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
+ * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
+ * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
+ * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
+ * SOFTWARE.
+ */
+
+use crate::println;
+use core::fmt;
+use core::ops::RangeInclusive;
+
+/// System memory map.
+#[rustfmt::skip]
+pub mod map {
+ pub const START: usize = 0x0000_0000;
+ pub const END: usize = 0x3FFF_FFFF;
+
+ pub mod physical {
+ pub const MMIO_BASE: usize = 0x3F00_0000;
+ pub const VIDEOCORE_MBOX_BASE: usize = MMIO_BASE + 0x0000_B880;
+ pub const GPIO_BASE: usize = MMIO_BASE + 0x0020_0000;
+ pub const PL011_UART_BASE: usize = MMIO_BASE + 0x0020_1000;
+ pub const MINI_UART_BASE: usize = MMIO_BASE + 0x0021_5000;
+ pub const MMIO_END: usize = super::END;
+ }
+
+ pub mod virt {
+ pub const KERN_STACK_START: usize = super::START;
+ pub const KERN_STACK_END: usize = 0x0007_FFFF;
+
+ // The second 2 MiB block.
+ pub const DMA_HEAP_START: usize = 0x0020_0000;
+ pub const DMA_HEAP_END: usize = 0x005F_FFFF;
+ }
+}
+
+/// Types used for compiling the virtual memory layout of the kernel using
+/// address ranges.
+pub mod kernel_mem_range {
+ use core::ops::RangeInclusive;
+
+ #[derive(Copy, Clone)]
+ pub enum MemAttributes {
+ CacheableDRAM,
+ NonCacheableDRAM,
+ Device,
+ }
+
+ #[derive(Copy, Clone)]
+ pub enum AccessPermissions {
+ ReadOnly,
+ ReadWrite,
+ }
+
+ #[allow(dead_code)]
+ #[derive(Copy, Clone)]
+ pub enum Translation {
+ Identity,
+ Offset(usize),
+ }
+
+ #[derive(Copy, Clone)]
+ pub struct AttributeFields {
+ pub mem_attributes: MemAttributes,
+ pub acc_perms: AccessPermissions,
+ pub execute_never: bool,
+ }
+
+ impl Default for AttributeFields {
+ fn default() -> AttributeFields {
+ AttributeFields {
+ mem_attributes: MemAttributes::CacheableDRAM,
+ acc_perms: AccessPermissions::ReadWrite,
+ execute_never: true,
+ }
+ }
+ }
+
+ pub struct Descriptor {
+ pub name: &'static str,
+ pub virtual_range: fn() -> RangeInclusive,
+ pub translation: Translation,
+ pub attribute_fields: AttributeFields,
+ }
+}
+
+use kernel_mem_range::*;
+
+/// A virtual memory layout that is agnostic of the paging granularity that the
+/// hardware MMU will use.
+///
+/// Contains only special ranges, aka anything that is _not_ normal cacheable
+/// DRAM.
+static KERNEL_VIRTUAL_LAYOUT: [Descriptor; 5] = [
+ Descriptor {
+ name: "Kernel stack",
+ virtual_range: || {
+ RangeInclusive::new(map::virt::KERN_STACK_START, map::virt::KERN_STACK_END)
+ },
+ translation: Translation::Identity,
+ attribute_fields: AttributeFields {
+ mem_attributes: MemAttributes::CacheableDRAM,
+ acc_perms: AccessPermissions::ReadWrite,
+ execute_never: true,
+ },
+ },
+ Descriptor {
+ name: "Kernel code and RO data",
+ virtual_range: || {
+ // Using the linker script, we ensure that the RO area is consecutive and 4
+ // KiB aligned, and we export the boundaries via symbols:
+ //
+ // [__ro_start, __ro_end)
+ extern "C" {
+ // The inclusive start of the read-only area, aka the address of the
+ // first byte of the area.
+ static __ro_start: u64;
+
+ // The exclusive end of the read-only area, aka the address of
+ // the first byte _after_ the RO area.
+ static __ro_end: u64;
+ }
+
+ unsafe {
+ // Notice the subtraction to turn the exclusive end into an
+ // inclusive end
+ RangeInclusive::new(
+ &__ro_start as *const _ as usize,
+ &__ro_end as *const _ as usize - 1,
+ )
+ }
+ },
+ translation: Translation::Identity,
+ attribute_fields: AttributeFields {
+ mem_attributes: MemAttributes::CacheableDRAM,
+ acc_perms: AccessPermissions::ReadOnly,
+ execute_never: false,
+ },
+ },
+ Descriptor {
+ name: "Kernel data and BSS",
+ virtual_range: || {
+ extern "C" {
+ static __ro_end: u64;
+ static __bss_end: u64;
+ }
+
+ unsafe {
+ RangeInclusive::new(
+ &__ro_end as *const _ as usize,
+ &__bss_end as *const _ as usize - 1,
+ )
+ }
+ },
+ translation: Translation::Identity,
+ attribute_fields: AttributeFields {
+ mem_attributes: MemAttributes::CacheableDRAM,
+ acc_perms: AccessPermissions::ReadWrite,
+ execute_never: true,
+ },
+ },
+ Descriptor {
+ name: "DMA heap pool",
+ virtual_range: || RangeInclusive::new(map::virt::DMA_HEAP_START, map::virt::DMA_HEAP_END),
+ translation: Translation::Identity,
+ attribute_fields: AttributeFields {
+ mem_attributes: MemAttributes::NonCacheableDRAM,
+ acc_perms: AccessPermissions::ReadWrite,
+ execute_never: true,
+ },
+ },
+ Descriptor {
+ name: "Device MMIO",
+ virtual_range: || RangeInclusive::new(map::physical::MMIO_BASE, map::physical::MMIO_END),
+ translation: Translation::Identity,
+ attribute_fields: AttributeFields {
+ mem_attributes: MemAttributes::Device,
+ acc_perms: AccessPermissions::ReadWrite,
+ execute_never: true,
+ },
+ },
+];
+
+/// For a given virtual address, find and return the output address and
+/// according attributes.
+///
+/// If the address is not covered in VIRTUAL_LAYOUT, return a default for normal
+/// cacheable DRAM.
+fn get_virt_addr_properties(virt_addr: usize) -> Result<(usize, AttributeFields), &'static str> {
+ if virt_addr > map::END {
+ return Err("Address out of range.");
+ }
+
+ for i in KERNEL_VIRTUAL_LAYOUT.iter() {
+ if (i.virtual_range)().contains(&virt_addr) {
+ let output_addr = match i.translation {
+ Translation::Identity => virt_addr,
+ Translation::Offset(a) => a + (virt_addr - (i.virtual_range)().start()),
+ };
+
+ return Ok((output_addr, i.attribute_fields));
+ }
+ }
+
+ Ok((virt_addr, AttributeFields::default()))
+}
+
+/// Human-readable output of a Descriptor.
+impl fmt::Display for Descriptor {
+ fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result {
+ // Call the function to which self.range points, and dereference the
+ // result, which causes Rust to copy the value.
+ let start = *(self.virtual_range)().start();
+ let end = *(self.virtual_range)().end();
+ let size = end - start + 1;
+
+ // log2(1024)
+ const KIB_RSHIFT: u32 = 10;
+
+ // log2(1024 * 1024)
+ const MIB_RSHIFT: u32 = 20;
+
+ let (size, unit) = if (size >> MIB_RSHIFT) > 0 {
+ (size >> MIB_RSHIFT, "MiB")
+ } else if (size >> KIB_RSHIFT) > 0 {
+ (size >> KIB_RSHIFT, "KiB")
+ } else {
+ (size, "Byte")
+ };
+
+ let attr = match self.attribute_fields.mem_attributes {
+ MemAttributes::CacheableDRAM => "C",
+ MemAttributes::NonCacheableDRAM => "NC",
+ MemAttributes::Device => "Dev",
+ };
+
+ let acc_p = match self.attribute_fields.acc_perms {
+ AccessPermissions::ReadOnly => "RO",
+ AccessPermissions::ReadWrite => "RW",
+ };
+
+ let xn = if self.attribute_fields.execute_never {
+ "PXN"
+ } else {
+ "PX"
+ };
+
+ write!(
+ f,
+ " {:#010X} - {:#010X} | {: >3} {} | {: <3} {} {: <3} | {}",
+ start, end, size, unit, attr, acc_p, xn, self.name
+ )
+ }
+}
+
+/// Print the kernel memory layout.
+pub fn print_layout() {
+ println!("[i] Kernel memory layout:");
+
+ for i in KERNEL_VIRTUAL_LAYOUT.iter() {
+ println!("{}", i);
+ }
+}
+
+/// Calculate the next possible aligned address without sanity checking the
+/// input parameters.
+#[inline]
+fn aligned_addr_unchecked(addr: usize, alignment: usize) -> usize {
+ (addr + (alignment - 1)) & !(alignment - 1)
+}
diff --git a/src/arch/aarch64/mod.rs b/src/arch/aarch64/mod.rs
index 586ace7..90e638c 100644
--- a/src/arch/aarch64/mod.rs
+++ b/src/arch/aarch64/mod.rs
@@ -1,7 +1,7 @@
// mod arch::aarch64
mod boot;
-mod memory;
+pub mod memory;
pub mod mmu;
pub mod traps;
pub use self::memory::{PhysicalAddress, VirtualAddress};