blob: 2ccea8c1a48f42afb77f2adcbea1972e9cb0d909 [file] [log] [blame]
Imre Kisd5b96fd2024-09-11 17:04:32 +02001// SPDX-FileCopyrightText: Copyright 2024 Arm Limited and/or its affiliates <open-source-office@arm.com>
2// SPDX-License-Identifier: MIT OR Apache-2.0
3
Imre Kis64d112f2025-01-20 12:59:01 +01004//! Objects for representing physical and virtual addresses
5
Imre Kis5f960442024-11-29 16:49:43 +01006use core::{fmt, ops::Range};
Imre Kisd5b96fd2024-09-11 17:04:32 +02007
Imre Kisc9a55ff2025-01-17 15:06:50 +01008use crate::TranslationRegime;
9
Imre Kis631127d2024-11-21 13:09:01 +010010use super::TranslationGranule;
Imre Kisd5b96fd2024-09-11 17:04:32 +020011
Imre Kis64d112f2025-01-20 12:59:01 +010012/// Physical address object
Imre Kis5f960442024-11-29 16:49:43 +010013#[derive(PartialEq, Eq, PartialOrd, Ord, Clone, Copy)]
Imre Kisd5b96fd2024-09-11 17:04:32 +020014pub struct PhysicalAddress(pub(super) usize);
15
16impl PhysicalAddress {
Imre Kis1278c9f2025-01-15 19:48:36 +010017 /// Create a new PhysicalAddress from the raw address value
18 ///
19 /// # Safety
20 /// The address has to be a valid physical address
Imre Kisd5b96fd2024-09-11 17:04:32 +020021 pub const unsafe fn new(address: usize) -> Self {
22 Self(address)
23 }
24
Imre Kis64d112f2025-01-20 12:59:01 +010025 /// Add offset to the physical address and check for overflow
Imre Kisd5b96fd2024-09-11 17:04:32 +020026 pub const fn add_offset(self, offset: usize) -> Option<Self> {
27 if let Some(address) = self.0.checked_add(offset) {
28 Some(Self(address))
29 } else {
30 None
31 }
32 }
33
Imre Kis64d112f2025-01-20 12:59:01 +010034 /// Identity map physical address to virtual address
Imre Kisd5b96fd2024-09-11 17:04:32 +020035 pub const fn identity_va(self) -> VirtualAddress {
36 VirtualAddress(self.0)
37 }
38
Imre Kis64d112f2025-01-20 12:59:01 +010039 /// Calculate difference of physical addresses
Imre Kisd5b96fd2024-09-11 17:04:32 +020040 pub const fn diff(self, rhs: Self) -> Option<usize> {
41 self.0.checked_sub(rhs.0)
42 }
43}
44
45impl From<PhysicalAddress> for usize {
46 fn from(value: PhysicalAddress) -> Self {
47 value.0
48 }
49}
50
51impl From<PhysicalAddress> for u64 {
52 fn from(value: PhysicalAddress) -> Self {
53 value.0 as u64
54 }
55}
56
Imre Kis5f960442024-11-29 16:49:43 +010057impl fmt::Debug for PhysicalAddress {
58 fn fmt(&self, f: &mut fmt::Formatter<'_>) -> core::fmt::Result {
59 f.debug_tuple("PA")
60 .field(&format_args!("{:#x}", self.0))
61 .finish()
62 }
63}
64
Imre Kis64d112f2025-01-20 12:59:01 +010065/// Virtual address object
Imre Kis5f960442024-11-29 16:49:43 +010066#[derive(PartialEq, Eq, PartialOrd, Ord, Clone, Copy)]
Imre Kisd5b96fd2024-09-11 17:04:32 +020067pub struct VirtualAddress(pub(super) usize);
68
69impl VirtualAddress {
Imre Kis1278c9f2025-01-15 19:48:36 +010070 /// Create a new VirtualAddress from the raw address value
71 ///
72 /// # Safety
73 /// The address has to be a valid virtual address
Imre Kisd5b96fd2024-09-11 17:04:32 +020074 pub const unsafe fn new(address: usize) -> Self {
75 Self(address)
76 }
77
Imre Kis64d112f2025-01-20 12:59:01 +010078 /// Add offset to the virtual address and check for overflow
Imre Kisd5b96fd2024-09-11 17:04:32 +020079 pub const fn add_offset(self, offset: usize) -> Option<Self> {
80 if let Some(address) = self.0.checked_add(offset) {
81 Some(Self(address))
82 } else {
83 None
84 }
85 }
86
Imre Kis64d112f2025-01-20 12:59:01 +010087 /// Identity map virtual address to physical address
Imre Kisd5b96fd2024-09-11 17:04:32 +020088 pub const fn identity_pa(self) -> PhysicalAddress {
89 PhysicalAddress(self.0)
90 }
91
Imre Kis64d112f2025-01-20 12:59:01 +010092 /// Mask the lower bits of the virtual address for the given granule and level
Imre Kis631127d2024-11-21 13:09:01 +010093 pub const fn mask_for_level<const VA_BITS: usize>(
94 self,
95 translation_granule: TranslationGranule<VA_BITS>,
96 level: isize,
97 ) -> Self {
98 Self(self.0 & (translation_granule.block_size_at_level(level) - 1))
Imre Kisd5b96fd2024-09-11 17:04:32 +020099 }
100
Imre Kis64d112f2025-01-20 12:59:01 +0100101 /// Calculate the index of the virtual address in a translation table at the
102 /// given granule and level.
Imre Kis631127d2024-11-21 13:09:01 +0100103 pub const fn get_level_index<const VA_BITS: usize>(
104 self,
105 translation_granule: TranslationGranule<VA_BITS>,
106 level: isize,
107 ) -> usize {
108 self.0 >> translation_granule.total_bits_at_level(level)
Imre Kisd5b96fd2024-09-11 17:04:32 +0200109 }
110
Imre Kis64d112f2025-01-20 12:59:01 +0100111 /// Check if the address is valid in the translation regime, i.e. if the top bits match the
112 /// VA range.
Imre Kisc9a55ff2025-01-17 15:06:50 +0100113 pub fn is_valid_in_regime<const VA_BITS: usize>(&self, regime: TranslationRegime) -> bool {
114 let mask = Self::get_upper_bit_mask::<VA_BITS>();
115 let required_upper_bits = if regime.is_upper_va_range() { mask } else { 0 };
116
117 (self.0 & mask) == required_upper_bits
118 }
119
Imre Kis64d112f2025-01-20 12:59:01 +0100120 /// Sets the upper bits of the virtual address according to the translation regime.
121 /// Fill with '1' bits for upper VA range, fill with '0' bits for lower VA range.
Imre Kisc9a55ff2025-01-17 15:06:50 +0100122 pub fn set_upper_bits<const VA_BITS: usize>(self, regime: TranslationRegime) -> Self {
123 let mask = Self::get_upper_bit_mask::<VA_BITS>();
124
125 Self(if regime.is_upper_va_range() {
126 self.0 | mask
127 } else {
128 self.0 & !mask
129 })
130 }
131
Imre Kis64d112f2025-01-20 12:59:01 +0100132 /// Remove top bits, i.e fill top bits with zeroes.
Imre Kisc9a55ff2025-01-17 15:06:50 +0100133 pub fn remove_upper_bits<const VA_BITS: usize>(self) -> Self {
134 Self(self.0 & !Self::get_upper_bit_mask::<VA_BITS>())
135 }
136
Imre Kis64d112f2025-01-20 12:59:01 +0100137 /// Mask bits in the address
Imre Kisd5b96fd2024-09-11 17:04:32 +0200138 pub const fn mask_bits(self, mask: usize) -> Self {
139 Self(self.0 & mask)
140 }
141
Imre Kis64d112f2025-01-20 12:59:01 +0100142 /// Calculate difference of virtual addresses
Imre Kisd5b96fd2024-09-11 17:04:32 +0200143 pub const fn diff(self, rhs: Self) -> Option<usize> {
144 self.0.checked_sub(rhs.0)
145 }
Imre Kisf0370e82024-11-18 16:24:55 +0100146
Imre Kis64d112f2025-01-20 12:59:01 +0100147 /// Align address to the next aligned address
Imre Kisf0370e82024-11-18 16:24:55 +0100148 pub const fn align_up(self, alignment: usize) -> Self {
149 Self(self.0.next_multiple_of(alignment))
150 }
Imre Kisc9a55ff2025-01-17 15:06:50 +0100151
152 const fn get_upper_bit_mask<const VA_BITS: usize>() -> usize {
153 !((1 << VA_BITS) - 1)
154 }
Imre Kisd5b96fd2024-09-11 17:04:32 +0200155}
156
157impl From<VirtualAddress> for usize {
158 fn from(value: VirtualAddress) -> Self {
159 value.0
160 }
161}
162
163impl From<VirtualAddress> for u64 {
164 fn from(value: VirtualAddress) -> Self {
165 value.0 as u64
166 }
167}
168
Imre Kis5f960442024-11-29 16:49:43 +0100169impl fmt::Debug for VirtualAddress {
170 fn fmt(&self, f: &mut fmt::Formatter<'_>) -> core::fmt::Result {
171 f.debug_tuple("VA")
172 .field(&format_args!("{:#x}", self.0))
173 .finish()
174 }
175}
176
Imre Kis64d112f2025-01-20 12:59:01 +0100177/// Represents a virtual address range
Imre Kisc9a55ff2025-01-17 15:06:50 +0100178#[derive(Debug)]
Imre Kisd5b96fd2024-09-11 17:04:32 +0200179pub struct VirtualAddressRange {
180 pub(super) start: VirtualAddress,
181 pub(super) end: VirtualAddress,
182}
183
184impl VirtualAddressRange {
185 pub fn new(start: VirtualAddress, end: VirtualAddress) -> Self {
186 Self { start, end }
187 }
188
Imre Kis1278c9f2025-01-15 19:48:36 +0100189 /// Create a new VirtualAddressRange from the raw address values
190 ///
191 /// # Safety
192 /// The addresses have to be valid virtual addresses
Imre Kisd5b96fd2024-09-11 17:04:32 +0200193 pub unsafe fn from_range(value: Range<usize>) -> Self {
194 Self::new(
195 VirtualAddress::new(value.start),
196 VirtualAddress::new(value.end),
197 )
198 }
199
Imre Kis64d112f2025-01-20 12:59:01 +0100200 /// The length of the range in bytes
Imre Kisd5b96fd2024-09-11 17:04:32 +0200201 pub fn len(&self) -> Option<usize> {
202 self.end.diff(self.start)
203 }
204
Imre Kis64d112f2025-01-20 12:59:01 +0100205 /// Create an iterator which returns virtual addresses in the range of the given step in bytes.
Imre Kisd5b96fd2024-09-11 17:04:32 +0200206 pub fn step_by(self, step: usize) -> VirtualAddressIterator {
207 VirtualAddressIterator {
208 next: self.start,
209 end: self.end,
210 step,
211 }
212 }
213}
214
Imre Kis64d112f2025-01-20 12:59:01 +0100215/// Iterator for walking the virtual address range using the given step.
Imre Kisd5b96fd2024-09-11 17:04:32 +0200216pub struct VirtualAddressIterator {
217 next: VirtualAddress,
218 end: VirtualAddress,
219 step: usize,
220}
221
222impl Iterator for VirtualAddressIterator {
223 type Item = VirtualAddress;
224
225 fn next(&mut self) -> Option<Self::Item> {
226 if self.next < self.end {
227 let current = self.next;
228
229 self.next = if let Some(next) = self.next.add_offset(self.step) {
230 next
231 } else {
232 self.end
233 };
234
235 Some(current)
236 } else {
237 None
238 }
239 }
240}