blob: eeea904d1c53129c737c81fb5b82cd8f0cfb54ff [file] [log] [blame]
Andrew Scull18834872018-10-12 11:48:09 +01001/*
Andrew Walbran692b3252019-03-07 15:51:31 +00002 * Copyright 2018 The Hafnium Authors.
Andrew Scull18834872018-10-12 11:48:09 +01003 *
4 * Licensed under the Apache License, Version 2.0 (the "License");
5 * you may not use this file except in compliance with the License.
6 * You may obtain a copy of the License at
7 *
8 * https://www.apache.org/licenses/LICENSE-2.0
9 *
10 * Unless required by applicable law or agreed to in writing, software
11 * distributed under the License is distributed on an "AS IS" BASIS,
12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13 * See the License for the specific language governing permissions and
14 * limitations under the License.
15 */
16
Andrew Scull18c78fc2018-08-20 12:57:41 +010017#include "hf/cpu.h"
Wedson Almeida Filho987c0ff2018-06-20 16:34:38 +010018
Andrew Scull04502e42018-09-03 14:54:52 +010019#include <stdalign.h>
20
Andrew Scull18c78fc2018-08-20 12:57:41 +010021#include "hf/arch/cpu.h"
22
23#include "hf/api.h"
24#include "hf/dlog.h"
Andrew Walbranb037d5b2019-06-25 17:19:41 +010025#include "hf/spci.h"
Andrew Scull8d9e1212019-04-05 13:52:55 +010026#include "hf/std.h"
Andrew Scull18c78fc2018-08-20 12:57:41 +010027#include "hf/vm.h"
Wedson Almeida Filho987c0ff2018-06-20 16:34:38 +010028
Andrew Scull19503262018-09-20 14:48:39 +010029#include "vmapi/hf/call.h"
30
Andrew Scull23e93a82018-10-26 14:56:04 +010031#define STACK_SIZE PAGE_SIZE
32
Wedson Almeida Filho3fcbcff2018-07-10 23:53:39 +010033/* The stack to be used by the CPUs. */
Andrew Scull37402872018-10-24 14:23:06 +010034alignas(2 * sizeof(uintreg_t)) static char callstacks[MAX_CPUS][STACK_SIZE];
Wedson Almeida Filho3fcbcff2018-07-10 23:53:39 +010035
36/* State of all supported CPUs. The stack of the first one is initialized. */
37struct cpu cpus[MAX_CPUS] = {
38 {
39 .is_on = 1,
Andrew Scullf3d45592018-09-20 14:30:22 +010040 .stack_bottom = &callstacks[0][STACK_SIZE],
Wedson Almeida Filho3fcbcff2018-07-10 23:53:39 +010041 },
42};
43
Andrew Scullbb3ab6c2018-11-26 20:38:49 +000044static uint32_t cpu_count = 1;
Wedson Almeida Filho3fcbcff2018-07-10 23:53:39 +010045
Andrew Scullbb3ab6c2018-11-26 20:38:49 +000046static void cpu_init(struct cpu *c)
47{
48 /* TODO: Assumes that c is zeroed out already. */
49 sl_init(&c->lock);
50 c->irq_disable_count = 1;
51}
52
53void cpu_module_init(const uint64_t *cpu_ids, size_t count)
54{
55 uint32_t i;
56 uint32_t j;
57 uint64_t boot_cpu_id = cpus[0].id;
58 bool found_boot_cpu = false;
59
60 cpu_count = count;
61
62 /*
63 * Initialize CPUs with the IDs from the configuration passed in. The
64 * CPUs after the boot CPU are initialized in reverse order. The boot
65 * CPU is initialized when it is found or in place of the last CPU if it
66 * is not found.
67 */
68 j = cpu_count;
69 for (i = 0; i < cpu_count; ++i) {
70 struct cpu *c;
71 uint64_t id = cpu_ids[i];
72
73 if (found_boot_cpu || id != boot_cpu_id) {
74 c = &cpus[--j];
75 } else {
76 found_boot_cpu = true;
77 c = &cpus[0];
78 }
Wedson Almeida Filho81568c42019-01-04 13:33:02 +000079
Wedson Almeida Filho3fcbcff2018-07-10 23:53:39 +010080 cpu_init(c);
Andrew Scullbb3ab6c2018-11-26 20:38:49 +000081 c->id = id;
Andrew Scullf3d45592018-09-20 14:30:22 +010082 c->stack_bottom = &callstacks[i][STACK_SIZE];
Wedson Almeida Filho3fcbcff2018-07-10 23:53:39 +010083 }
Andrew Scullbb3ab6c2018-11-26 20:38:49 +000084
85 if (!found_boot_cpu) {
86 /* Boot CPU was initialized but with wrong ID. */
87 dlog("Boot CPU's ID not found in config.");
88 cpus[0].id = boot_cpu_id;
89 }
Wedson Almeida Filho3fcbcff2018-07-10 23:53:39 +010090}
91
92size_t cpu_index(struct cpu *c)
93{
Wedson Almeida Filho03e767a2018-07-30 15:32:03 +010094 return c - cpus;
Wedson Almeida Filho3fcbcff2018-07-10 23:53:39 +010095}
96
Wedson Almeida Filho987c0ff2018-06-20 16:34:38 +010097void cpu_irq_enable(struct cpu *c)
98{
99 c->irq_disable_count--;
Andrew Scull7364a8e2018-07-19 15:39:29 +0100100 if (!c->irq_disable_count) {
Wedson Almeida Filho987c0ff2018-06-20 16:34:38 +0100101 arch_irq_enable();
Andrew Scull7364a8e2018-07-19 15:39:29 +0100102 }
Wedson Almeida Filho987c0ff2018-06-20 16:34:38 +0100103}
104
105void cpu_irq_disable(struct cpu *c)
106{
Andrew Scull7364a8e2018-07-19 15:39:29 +0100107 if (!c->irq_disable_count) {
Wedson Almeida Filho987c0ff2018-06-20 16:34:38 +0100108 arch_irq_disable();
Andrew Scull7364a8e2018-07-19 15:39:29 +0100109 }
Wedson Almeida Filho987c0ff2018-06-20 16:34:38 +0100110 c->irq_disable_count++;
111}
112
Wedson Almeida Filho87009642018-07-02 10:20:07 +0100113/**
114 * Turns CPU on and returns the previous state.
115 */
Andrew Scull37402872018-10-24 14:23:06 +0100116bool cpu_on(struct cpu *c, ipaddr_t entry, uintreg_t arg)
Wedson Almeida Filho987c0ff2018-06-20 16:34:38 +0100117{
Wedson Almeida Filho87009642018-07-02 10:20:07 +0100118 bool prev;
119
Wedson Almeida Filho987c0ff2018-06-20 16:34:38 +0100120 sl_lock(&c->lock);
Wedson Almeida Filho87009642018-07-02 10:20:07 +0100121 prev = c->is_on;
122 c->is_on = true;
123 sl_unlock(&c->lock);
124
125 if (!prev) {
Andrew Walbran42347a92019-05-09 13:59:03 +0100126 struct vm *vm = vm_find(HF_PRIMARY_VM_ID);
Andrew Walbrane1310df2019-04-29 17:28:28 +0100127 struct vcpu *vcpu = vm_get_vcpu(vm, cpu_index(c));
Andrew Walbranb58f8992019-04-15 12:29:31 +0100128 struct vcpu_locked vcpu_locked;
Wedson Almeida Filho81568c42019-01-04 13:33:02 +0000129
Andrew Walbranb58f8992019-04-15 12:29:31 +0100130 vcpu_locked = vcpu_lock(vcpu);
131 vcpu_on(vcpu_locked, entry, arg);
132 vcpu_unlock(&vcpu_locked);
Wedson Almeida Filho987c0ff2018-06-20 16:34:38 +0100133 }
Wedson Almeida Filho87009642018-07-02 10:20:07 +0100134
135 return prev;
Wedson Almeida Filho987c0ff2018-06-20 16:34:38 +0100136}
137
Wedson Almeida Filho03e767a2018-07-30 15:32:03 +0100138/**
139 * Prepares the CPU for turning itself off.
Wedson Almeida Filho987c0ff2018-06-20 16:34:38 +0100140 */
141void cpu_off(struct cpu *c)
142{
Wedson Almeida Filho987c0ff2018-06-20 16:34:38 +0100143 sl_lock(&c->lock);
Wedson Almeida Filho87009642018-07-02 10:20:07 +0100144 c->is_on = false;
Wedson Almeida Filho987c0ff2018-06-20 16:34:38 +0100145 sl_unlock(&c->lock);
Wedson Almeida Filho03e767a2018-07-30 15:32:03 +0100146}
Wedson Almeida Filho987c0ff2018-06-20 16:34:38 +0100147
Wedson Almeida Filho03e767a2018-07-30 15:32:03 +0100148/**
149 * Searches for a CPU based on its id.
150 */
Andrew Scullbb3ab6c2018-11-26 20:38:49 +0000151struct cpu *cpu_find(uint64_t id)
Wedson Almeida Filho03e767a2018-07-30 15:32:03 +0100152{
153 size_t i;
154
Andrew Scullbb3ab6c2018-11-26 20:38:49 +0000155 for (i = 0; i < cpu_count; i++) {
Wedson Almeida Filho03e767a2018-07-30 15:32:03 +0100156 if (cpus[i].id == id) {
Andrew Scullf3d45592018-09-20 14:30:22 +0100157 return &cpus[i];
Wedson Almeida Filho03e767a2018-07-30 15:32:03 +0100158 }
159 }
160
161 return NULL;
Wedson Almeida Filho987c0ff2018-06-20 16:34:38 +0100162}
163
Andrew Walbranb58f8992019-04-15 12:29:31 +0100164/**
165 * Locks the given vCPU and updates `locked` to hold the newly locked vCPU.
166 */
167struct vcpu_locked vcpu_lock(struct vcpu *vcpu)
168{
169 struct vcpu_locked locked = {
170 .vcpu = vcpu,
171 };
172
173 sl_lock(&vcpu->lock);
174
175 return locked;
176}
177
178/**
179 * Unlocks a vCPU previously locked with vpu_lock, and updates `locked` to
180 * reflect the fact that the vCPU is no longer locked.
181 */
182void vcpu_unlock(struct vcpu_locked *locked)
183{
184 sl_unlock(&locked->vcpu->lock);
185 locked->vcpu = NULL;
186}
187
Wedson Almeida Filho87009642018-07-02 10:20:07 +0100188void vcpu_init(struct vcpu *vcpu, struct vm *vm)
Wedson Almeida Filho987c0ff2018-06-20 16:34:38 +0100189{
Andrew Scull2b5fbad2019-04-05 13:55:56 +0100190 memset_s(vcpu, sizeof(*vcpu), 0, sizeof(*vcpu));
Wedson Almeida Filho87009642018-07-02 10:20:07 +0100191 sl_init(&vcpu->lock);
Wedson Almeida Filho03306112018-11-26 00:08:03 +0000192 vcpu->regs_available = true;
Wedson Almeida Filho987c0ff2018-06-20 16:34:38 +0100193 vcpu->vm = vm;
Andrew Sculld6ee1102019-04-05 22:12:42 +0100194 vcpu->state = VCPU_STATE_OFF;
Wedson Almeida Filho987c0ff2018-06-20 16:34:38 +0100195}
Wedson Almeida Filho87009642018-07-02 10:20:07 +0100196
Andrew Walbranb58f8992019-04-15 12:29:31 +0100197/**
198 * Initialise the registers for the given vCPU and set the state to
199 * VCPU_STATE_READY. The caller must hold the vCPU lock while calling this.
200 */
201void vcpu_on(struct vcpu_locked vcpu, ipaddr_t entry, uintreg_t arg)
Wedson Almeida Filho87009642018-07-02 10:20:07 +0100202{
Andrew Walbranb58f8992019-04-15 12:29:31 +0100203 arch_regs_set_pc_arg(&vcpu.vcpu->regs, entry, arg);
204 vcpu.vcpu->state = VCPU_STATE_READY;
Wedson Almeida Filho87009642018-07-02 10:20:07 +0100205}
206
Andrew Walbranb037d5b2019-06-25 17:19:41 +0100207spci_vcpu_index_t vcpu_index(const struct vcpu *vcpu)
Wedson Almeida Filho9d5040f2018-10-29 08:41:27 +0000208{
Andrew Walbranb037d5b2019-06-25 17:19:41 +0100209 size_t index = vcpu - vcpu->vm->vcpus;
210
211 assert(index < UINT16_MAX);
212 return index;
Wedson Almeida Filho9d5040f2018-10-29 08:41:27 +0000213}
Wedson Almeida Filho99d2d4c2019-02-14 12:53:46 +0000214
215/**
Andrew Walbran33645652019-04-15 12:29:31 +0100216 * Check whether the given vcpu_state is an off state, for the purpose of
217 * turning vCPUs on and off. Note that aborted still counts as on in this
218 * context.
Andrew Walbran9a43fee2019-04-18 17:42:32 +0100219 */
Andrew Walbran33645652019-04-15 12:29:31 +0100220bool vcpu_is_off(struct vcpu_locked vcpu)
221{
222 switch (vcpu.vcpu->state) {
223 case VCPU_STATE_OFF:
224 return true;
225 case VCPU_STATE_READY:
226 case VCPU_STATE_RUNNING:
227 case VCPU_STATE_BLOCKED_MAILBOX:
228 case VCPU_STATE_BLOCKED_INTERRUPT:
229 case VCPU_STATE_ABORTED:
230 /*
231 * Aborted still counts as ON for the purposes of PSCI,
232 * because according to the PSCI specification (section
233 * 5.7.1) a core is only considered to be off if it has
234 * been turned off with a CPU_OFF call or hasn't yet
235 * been turned on with a CPU_ON call.
236 */
237 return false;
238 }
239}
240
241/**
242 * Starts a vCPU of a secondary VM.
243 *
244 * Returns true if the secondary was reset and started, or false if it was
245 * already on and so nothing was done.
246 */
247bool vcpu_secondary_reset_and_start(struct vcpu *vcpu, ipaddr_t entry,
Andrew Walbran9a43fee2019-04-18 17:42:32 +0100248 uintreg_t arg)
249{
Andrew Walbranb58f8992019-04-15 12:29:31 +0100250 struct vcpu_locked vcpu_locked;
Andrew Walbran9a43fee2019-04-18 17:42:32 +0100251 struct vm *vm = vcpu->vm;
Andrew Walbran33645652019-04-15 12:29:31 +0100252 bool vcpu_was_off;
Andrew Walbran9a43fee2019-04-18 17:42:32 +0100253
254 assert(vm->id != HF_PRIMARY_VM_ID);
255
Andrew Walbranb58f8992019-04-15 12:29:31 +0100256 vcpu_locked = vcpu_lock(vcpu);
Andrew Walbran33645652019-04-15 12:29:31 +0100257 vcpu_was_off = vcpu_is_off(vcpu_locked);
258 if (vcpu_was_off) {
259 /*
260 * Set vCPU registers to a clean state ready for boot. As this
261 * is a secondary which can migrate between pCPUs, the ID of the
262 * vCPU is defined as the index and does not match the ID of the
263 * pCPU it is running on.
264 */
265 arch_regs_reset(&vcpu->regs, false, vm->id, vcpu_index(vcpu),
266 vm->ptable.root);
267 vcpu_on(vcpu_locked, entry, arg);
268 }
Andrew Walbranb58f8992019-04-15 12:29:31 +0100269 vcpu_unlock(&vcpu_locked);
Andrew Walbran33645652019-04-15 12:29:31 +0100270
271 return vcpu_was_off;
Andrew Walbran9a43fee2019-04-18 17:42:32 +0100272}
273
274/**
Wedson Almeida Filho99d2d4c2019-02-14 12:53:46 +0000275 * Handles a page fault. It does so by determining if it's a legitimate or
276 * spurious fault, and recovering from the latter.
277 *
278 * Returns true if the caller should resume the current vcpu, or false if its VM
279 * should be aborted.
280 */
281bool vcpu_handle_page_fault(const struct vcpu *current,
282 struct vcpu_fault_info *f)
283{
284 struct vm *vm = current->vm;
Wedson Almeida Filho99d2d4c2019-02-14 12:53:46 +0000285 int mode;
286 int mask = f->mode | MM_MODE_INVALID;
Andrew Sculld3cfaad2019-04-04 11:34:10 +0100287 bool resume;
Wedson Almeida Filho99d2d4c2019-02-14 12:53:46 +0000288
289 sl_lock(&vm->lock);
290
291 /*
292 * Check if this is a legitimate fault, i.e., if the page table doesn't
293 * allow the access attemped by the VM.
Andrew Sculld3cfaad2019-04-04 11:34:10 +0100294 *
295 * Otherwise, this is a spurious fault, likely because another CPU is
296 * updating the page table. It is responsible for issuing global TLB
297 * invalidations while holding the VM lock, so we don't need to do
298 * anything else to recover from it. (Acquiring/releasing the lock
299 * ensured that the invalidations have completed.)
Wedson Almeida Filho99d2d4c2019-02-14 12:53:46 +0000300 */
Andrew Sculld3cfaad2019-04-04 11:34:10 +0100301 resume = mm_vm_get_mode(&vm->ptable, f->ipaddr, ipa_add(f->ipaddr, 1),
302 &mode) &&
303 (mode & mask) == f->mode;
Wedson Almeida Filho99d2d4c2019-02-14 12:53:46 +0000304
Wedson Almeida Filho99d2d4c2019-02-14 12:53:46 +0000305 sl_unlock(&vm->lock);
Andrew Sculld3cfaad2019-04-04 11:34:10 +0100306
307 if (!resume) {
Wedson Almeida Filho99d2d4c2019-02-14 12:53:46 +0000308 dlog("Stage-2 page fault: pc=0x%x, vmid=%u, vcpu=%u, "
Andrew Sculld3cfaad2019-04-04 11:34:10 +0100309 "vaddr=0x%x, ipaddr=0x%x, mode=0x%x\n",
Wedson Almeida Filho99d2d4c2019-02-14 12:53:46 +0000310 f->pc, vm->id, vcpu_index(current), f->vaddr, f->ipaddr,
Andrew Sculld3cfaad2019-04-04 11:34:10 +0100311 f->mode);
Wedson Almeida Filho99d2d4c2019-02-14 12:53:46 +0000312 }
Andrew Sculld3cfaad2019-04-04 11:34:10 +0100313
314 return resume;
Wedson Almeida Filho99d2d4c2019-02-14 12:53:46 +0000315}