blob: 5b6ac0db3aea195b10856569808863eaeb0ee4d2 [file] [log] [blame]
Andrew Scull18834872018-10-12 11:48:09 +01001/*
Andrew Walbran692b3252019-03-07 15:51:31 +00002 * Copyright 2018 The Hafnium Authors.
Andrew Scull18834872018-10-12 11:48:09 +01003 *
Andrew Walbrane959ec12020-06-17 15:01:09 +01004 * Use of this source code is governed by a BSD-style
5 * license that can be found in the LICENSE file or at
6 * https://opensource.org/licenses/BSD-3-Clause.
Andrew Scull18834872018-10-12 11:48:09 +01007 */
8
Andrew Scullfbc938a2018-08-20 14:09:28 +01009#pragma once
Andrew Scullf35a5c92018-08-07 18:09:46 +010010
Andrew Scull6d2db332018-10-10 15:28:17 +010011#include "hf/abi.h"
Andrew Walbranb5ab43c2020-04-30 11:32:54 +010012#include "hf/ffa.h"
Andrew Scull6d2db332018-10-10 15:28:17 +010013#include "hf/types.h"
Andrew Scullf35a5c92018-08-07 18:09:46 +010014
Andrew Scull5ac05f02018-08-10 17:23:22 +010015/**
Fuad Tabba77a4b012019-11-15 12:13:08 +000016 * This function must be implemented to trigger the architecture-specific
Andrew Scull5ac05f02018-08-10 17:23:22 +010017 * mechanism to call to the hypervisor.
Andrew Scullf35a5c92018-08-07 18:09:46 +010018 */
Andrew Walbran4e6bcc72019-09-11 13:57:22 +010019int64_t hf_call(uint64_t arg0, uint64_t arg1, uint64_t arg2, uint64_t arg3);
Andrew Walbranb5ab43c2020-04-30 11:32:54 +010020struct ffa_value ffa_call(struct ffa_value args);
Andrew Scullf35a5c92018-08-07 18:09:46 +010021
Andrew Scull5ac05f02018-08-10 17:23:22 +010022/**
Andrew Scull55c4d8b2018-12-18 18:50:18 +000023 * Returns the VM's own ID.
Andrew Scull5ac05f02018-08-10 17:23:22 +010024 */
Andrew Walbranb5ab43c2020-04-30 11:32:54 +010025static inline struct ffa_value ffa_id_get(void)
Andrew Walbrand230f662019-10-07 18:03:36 +010026{
Andrew Walbranb5ab43c2020-04-30 11:32:54 +010027 return ffa_call((struct ffa_value){.func = FFA_ID_GET_32});
Andrew Walbrand230f662019-10-07 18:03:36 +010028}
29
30/**
Olivier Deprez421677d2021-06-18 12:18:53 +020031 * Returns the SPMC FF-A ID at NS virtual/physical and secure virtual
32 * FF-A instances.
33 * DEN0077A FF-A v1.1 Beta0 section 13.9 FFA_SPM_ID_GET.
34 */
35static inline struct ffa_value ffa_spm_id_get(void)
36{
37 return ffa_call((struct ffa_value){.func = FFA_SPM_ID_GET_32});
38}
39
40/**
Fuad Tabbae4efcc32020-07-16 15:37:27 +010041 * Requests information for partitions instantiated in the system. The
42 * information is returned in the RX buffer of the caller as an array of
43 * partition information descriptors (struct ffa_partition_info).
44 *
45 * A Null UUID (UUID that is all zeros) returns information for all partitions,
46 * whereas a non-Null UUID returns information only for partitions that match.
47 *
48 * Returns:
49 * - FFA_SUCCESS on success. The count of partition information descriptors
50 * populated in the RX buffer is returned in arg2 (register w2).
51 * - FFA_BUSY if the caller's RX buffer is not free.
52 * - FFA_NO_MEMORY if the results do not fit in the callers RX buffer.
53 * - FFA_INVALID_PARAMETERS for an unrecognized UUID.
54 */
55static inline struct ffa_value ffa_partition_info_get(
56 const struct ffa_uuid *uuid)
57{
58 return ffa_call((struct ffa_value){.func = FFA_PARTITION_INFO_GET_32,
59 .arg1 = uuid->uuid[0],
60 .arg2 = uuid->uuid[1],
61 .arg3 = uuid->uuid[2],
62 .arg4 = uuid->uuid[3]});
63}
64
65/**
Olivier Deprez175161a2021-11-23 18:41:09 +010066 * DEN0077A FF-A v1.1 Beta0 section 18.3.2.1
67 * Registers vCPU secondary entry point for the caller VM.
68 * Called from secure virtual FF-A instance.
69 */
70static inline struct ffa_value ffa_secondary_ep_register(uintptr_t address)
71{
72 return ffa_call((struct ffa_value){.func = FFA_SECONDARY_EP_REGISTER_64,
73 .arg1 = address});
74}
75
76/**
Andrew Walbrand230f662019-10-07 18:03:36 +010077 * Returns the VM's own ID.
78 */
Andrew Walbranb5ab43c2020-04-30 11:32:54 +010079static inline ffa_vm_id_t hf_vm_get_id(void)
Andrew Scull5ac05f02018-08-10 17:23:22 +010080{
Andrew Walbranb5ab43c2020-04-30 11:32:54 +010081 return ffa_id_get().arg2;
Andrew Scull5ac05f02018-08-10 17:23:22 +010082}
83
84/**
Andrew Walbran27faff32019-10-02 18:20:57 +010085 * Runs the given vCPU of the given VM.
86 */
Andrew Walbranb5ab43c2020-04-30 11:32:54 +010087static inline struct ffa_value ffa_run(ffa_vm_id_t vm_id,
88 ffa_vcpu_index_t vcpu_idx)
Andrew Walbran27faff32019-10-02 18:20:57 +010089{
Andrew Walbranb5ab43c2020-04-30 11:32:54 +010090 return ffa_call((struct ffa_value){.func = FFA_RUN_32,
91 ffa_vm_vcpu(vm_id, vcpu_idx)});
Andrew Walbran27faff32019-10-02 18:20:57 +010092}
93
94/**
Fuad Tabbab0ef2a42019-12-19 11:19:25 +000095 * Hints that the vCPU is willing to yield its current use of the physical CPU.
Andrew Walbranb5ab43c2020-04-30 11:32:54 +010096 * This call always returns FFA_SUCCESS.
Andrew Scull55c4d8b2018-12-18 18:50:18 +000097 */
Andrew Walbranb5ab43c2020-04-30 11:32:54 +010098static inline struct ffa_value ffa_yield(void)
Andrew Scull55c4d8b2018-12-18 18:50:18 +000099{
Andrew Walbranb5ab43c2020-04-30 11:32:54 +0100100 return ffa_call((struct ffa_value){.func = FFA_YIELD_32});
Andrew Scull55c4d8b2018-12-18 18:50:18 +0000101}
102
103/**
Andrew Scull5ac05f02018-08-10 17:23:22 +0100104 * Configures the pages to send/receive data through. The pages must not be
105 * shared.
Andrew Walbran54afb502018-11-26 16:01:11 +0000106 *
Wedson Almeida Filhoea62e2e2019-01-09 19:14:59 +0000107 * Returns:
Andrew Walbranb5ab43c2020-04-30 11:32:54 +0100108 * - FFA_ERROR FFA_INVALID_PARAMETERS if the given addresses are not properly
Andrew Walbranbfffb0f2019-11-05 14:02:34 +0000109 * aligned or are the same.
Andrew Walbranb5ab43c2020-04-30 11:32:54 +0100110 * - FFA_ERROR FFA_NO_MEMORY if the hypervisor was unable to map the buffers
Fuad Tabba9dc276f2020-07-16 09:29:32 +0100111 * due to insufficient page table memory.
Andrew Walbranb5ab43c2020-04-30 11:32:54 +0100112 * - FFA_ERROR FFA_DENIED if the pages are already mapped or are not owned by
Andrew Walbranbfffb0f2019-11-05 14:02:34 +0000113 * the caller.
Andrew Walbranb5ab43c2020-04-30 11:32:54 +0100114 * - FFA_SUCCESS on success if no further action is needed.
115 * - FFA_RX_RELEASE if it was called by the primary VM and the primary VM now
Andrew Walbranbfffb0f2019-11-05 14:02:34 +0000116 * needs to wake up or kick waiters.
Andrew Scull5ac05f02018-08-10 17:23:22 +0100117 */
Andrew Walbranb5ab43c2020-04-30 11:32:54 +0100118static inline struct ffa_value ffa_rxtx_map(hf_ipaddr_t send, hf_ipaddr_t recv)
Andrew Scull5ac05f02018-08-10 17:23:22 +0100119{
Andrew Walbranb5ab43c2020-04-30 11:32:54 +0100120 return ffa_call(
121 (struct ffa_value){.func = FFA_RXTX_MAP_64,
122 .arg1 = send,
123 .arg2 = recv,
124 .arg3 = HF_MAILBOX_SIZE / FFA_PAGE_SIZE});
Andrew Scull5ac05f02018-08-10 17:23:22 +0100125}
126
127/**
Daniel Boulby9e420ca2021-07-07 15:03:49 +0100128 * Unmaps the RX/TX buffer pair of an endpoint or Hypervisor from the
129 * translation regime of the callee.
130 *
131 * Returns:
132 * - FFA_ERROR FFA_INVALID_PARAMETERS if there is no buffer pair registered on
133 * behalf of the caller.
134 * - FFA_SUCCESS on success if no further action is needed.
135 */
136static inline struct ffa_value ffa_rxtx_unmap(void)
137{
138 /* Note that allocator ID MBZ at virtual instance. */
139 return ffa_call((struct ffa_value){.func = FFA_RXTX_UNMAP_32});
140}
141
142/**
Andrew Scullaa039b32018-10-04 15:02:26 +0100143 * Copies data from the sender's send buffer to the recipient's receive buffer.
Andrew Walbran54afb502018-11-26 16:01:11 +0000144 *
Wedson Almeida Filho17c997f2019-01-09 18:50:09 +0000145 * If the recipient's receive buffer is busy, it can optionally register the
146 * caller to be notified when the recipient's receive buffer becomes available.
147 *
Andrew Walbran70bc8622019-10-07 14:15:58 +0100148 * Attributes may include:
Andrew Walbranb5ab43c2020-04-30 11:32:54 +0100149 * - FFA_MSG_SEND_NOTIFY, to notify the caller when it should try again.
150 * - FFA_MSG_SEND_LEGACY_MEMORY_*, to send a legacy architected memory sharing
Andrew Walbran70bc8622019-10-07 14:15:58 +0100151 * message.
152 *
Andrew Walbranb5ab43c2020-04-30 11:32:54 +0100153 * Returns FFA_SUCCESS if the message is sent, or an error code otherwise:
Andrew Walbran70bc8622019-10-07 14:15:58 +0100154 * - INVALID_PARAMETERS: one or more of the parameters do not conform.
Jose Marinhoa1dfeda2019-02-27 16:46:03 +0000155 * - BUSY: the message could not be delivered either because the mailbox
Andrew Walbran70bc8622019-10-07 14:15:58 +0100156 * was full or the target VM is not yet set up.
Andrew Scull5ac05f02018-08-10 17:23:22 +0100157 */
Andrew Walbranb5ab43c2020-04-30 11:32:54 +0100158static inline struct ffa_value ffa_msg_send(ffa_vm_id_t sender_vm_id,
159 ffa_vm_id_t target_vm_id,
160 uint32_t size, uint32_t attributes)
Andrew Scull5ac05f02018-08-10 17:23:22 +0100161{
Andrew Walbranb5ab43c2020-04-30 11:32:54 +0100162 return ffa_call((struct ffa_value){
163 .func = FFA_MSG_SEND_32,
Andrew Walbran70bc8622019-10-07 14:15:58 +0100164 .arg1 = ((uint64_t)sender_vm_id << 16) | target_vm_id,
165 .arg3 = size,
166 .arg4 = attributes});
Andrew Scull5ac05f02018-08-10 17:23:22 +0100167}
168
Andrew Walbranb5ab43c2020-04-30 11:32:54 +0100169static inline struct ffa_value ffa_mem_donate(uint32_t length,
Andrew Walbrana65a1322020-04-06 19:32:32 +0100170 uint32_t fragment_length)
Andrew Walbrane908c4a2019-12-02 17:13:47 +0000171{
Andrew Walbranb5ab43c2020-04-30 11:32:54 +0100172 return ffa_call((struct ffa_value){.func = FFA_MEM_DONATE_32,
173 .arg1 = length,
174 .arg2 = fragment_length});
Andrew Walbrane908c4a2019-12-02 17:13:47 +0000175}
176
Andrew Walbranb5ab43c2020-04-30 11:32:54 +0100177static inline struct ffa_value ffa_mem_lend(uint32_t length,
178 uint32_t fragment_length)
Andrew Walbrane908c4a2019-12-02 17:13:47 +0000179{
Andrew Walbranb5ab43c2020-04-30 11:32:54 +0100180 return ffa_call((struct ffa_value){.func = FFA_MEM_LEND_32,
181 .arg1 = length,
182 .arg2 = fragment_length});
Andrew Walbrane908c4a2019-12-02 17:13:47 +0000183}
184
Andrew Walbranb5ab43c2020-04-30 11:32:54 +0100185static inline struct ffa_value ffa_mem_share(uint32_t length,
186 uint32_t fragment_length)
Andrew Walbran82d6d152019-12-24 15:02:06 +0000187{
Andrew Walbranb5ab43c2020-04-30 11:32:54 +0100188 return ffa_call((struct ffa_value){.func = FFA_MEM_SHARE_32,
189 .arg1 = length,
190 .arg2 = fragment_length});
Andrew Walbran82d6d152019-12-24 15:02:06 +0000191}
192
Andrew Walbranb5ab43c2020-04-30 11:32:54 +0100193static inline struct ffa_value ffa_mem_retrieve_req(uint32_t length,
194 uint32_t fragment_length)
Andrew Walbran5de9c3d2020-02-10 13:35:29 +0000195{
Andrew Walbranb5ab43c2020-04-30 11:32:54 +0100196 return ffa_call((struct ffa_value){.func = FFA_MEM_RETRIEVE_REQ_32,
197 .arg1 = length,
198 .arg2 = fragment_length});
Andrew Walbran5de9c3d2020-02-10 13:35:29 +0000199}
200
Andrew Walbranb5ab43c2020-04-30 11:32:54 +0100201static inline struct ffa_value ffa_mem_relinquish(void)
Andrew Walbran5de9c3d2020-02-10 13:35:29 +0000202{
Andrew Walbranb5ab43c2020-04-30 11:32:54 +0100203 return ffa_call((struct ffa_value){.func = FFA_MEM_RELINQUISH_32});
204}
205
206static inline struct ffa_value ffa_mem_reclaim(ffa_memory_handle_t handle,
207 ffa_memory_region_flags_t flags)
208{
209 return ffa_call((struct ffa_value){.func = FFA_MEM_RECLAIM_32,
210 .arg1 = (uint32_t)handle,
211 .arg2 = (uint32_t)(handle >> 32),
212 .arg3 = flags});
Andrew Walbran5de9c3d2020-02-10 13:35:29 +0000213}
214
Andrew Walbranca808b12020-05-15 17:22:28 +0100215static inline struct ffa_value ffa_mem_frag_rx(ffa_memory_handle_t handle,
216 uint32_t fragment_offset)
217{
218 /* Note that sender MBZ at virtual instance. */
219 return ffa_call((struct ffa_value){.func = FFA_MEM_FRAG_RX_32,
220 .arg1 = (uint32_t)handle,
221 .arg2 = (uint32_t)(handle >> 32),
222 .arg3 = fragment_offset});
223}
224
225static inline struct ffa_value ffa_mem_frag_tx(ffa_memory_handle_t handle,
226 uint32_t fragment_length)
227{
228 /* Note that sender MBZ at virtual instance. */
229 return ffa_call((struct ffa_value){.func = FFA_MEM_FRAG_TX_32,
230 .arg1 = (uint32_t)handle,
231 .arg2 = (uint32_t)(handle >> 32),
232 .arg3 = fragment_length});
233}
234
Andrew Scull5ac05f02018-08-10 17:23:22 +0100235/**
Andrew Walbran0de4f162019-09-03 16:44:20 +0100236 * Called by secondary VMs to receive a message. This will block until a message
237 * is received.
Andrew Scull5ac05f02018-08-10 17:23:22 +0100238 *
Andrew Scullaa039b32018-10-04 15:02:26 +0100239 * The mailbox must be cleared before a new message can be received.
Andrew Walbran9311c9a2019-03-12 16:59:04 +0000240 *
Andrew Walbran0de4f162019-09-03 16:44:20 +0100241 * If no message is immediately available and there are no enabled and pending
242 * interrupts (irrespective of whether interrupts are enabled globally), then
243 * this will block until a message is available or an enabled interrupt becomes
Fuad Tabbab0ef2a42019-12-19 11:19:25 +0000244 * pending. This matches the behaviour of the WFI instruction on AArch64, except
Andrew Walbran0de4f162019-09-03 16:44:20 +0100245 * that a message becoming available is also treated like a wake-up event.
Andrew Walbranc8500812019-06-26 10:36:48 +0100246 *
247 * Returns:
Andrew Walbranb5ab43c2020-04-30 11:32:54 +0100248 * - FFA_MSG_SEND if a message is successfully received.
249 * - FFA_ERROR FFA_NOT_SUPPORTED if called from the primary VM.
250 * - FFA_ERROR FFA_INTERRUPTED if an interrupt happened during the call.
Andrew Scull5ac05f02018-08-10 17:23:22 +0100251 */
Andrew Walbranb5ab43c2020-04-30 11:32:54 +0100252static inline struct ffa_value ffa_msg_wait(void)
Andrew Scull5ac05f02018-08-10 17:23:22 +0100253{
Andrew Walbranb5ab43c2020-04-30 11:32:54 +0100254 return ffa_call((struct ffa_value){.func = FFA_MSG_WAIT_32});
Andrew Walbran0de4f162019-09-03 16:44:20 +0100255}
256
257/**
258 * Called by secondary VMs to receive a message. The call will return whether or
259 * not a message is available.
260 *
261 * The mailbox must be cleared before a new message can be received.
262 *
263 * Returns:
Andrew Walbranb5ab43c2020-04-30 11:32:54 +0100264 * - FFA_MSG_SEND if a message is successfully received.
265 * - FFA_ERROR FFA_NOT_SUPPORTED if called from the primary VM.
266 * - FFA_ERROR FFA_INTERRUPTED if an interrupt happened during the call.
267 * - FFA_ERROR FFA_RETRY if there was no pending message.
Andrew Walbran0de4f162019-09-03 16:44:20 +0100268 */
Andrew Walbranb5ab43c2020-04-30 11:32:54 +0100269static inline struct ffa_value ffa_msg_poll(void)
Andrew Walbran0de4f162019-09-03 16:44:20 +0100270{
Andrew Walbranb5ab43c2020-04-30 11:32:54 +0100271 return ffa_call((struct ffa_value){.func = FFA_MSG_POLL_32});
Andrew Scull5ac05f02018-08-10 17:23:22 +0100272}
273
274/**
Andrew Walbran8a0f5ca2019-11-05 13:12:23 +0000275 * Releases the caller's mailbox so that a new message can be received. The
276 * caller must have copied out all data they wish to preserve as new messages
277 * will overwrite the old and will arrive asynchronously.
Andrew Walbran54afb502018-11-26 16:01:11 +0000278 *
Wedson Almeida Filhoea62e2e2019-01-09 19:14:59 +0000279 * Returns:
Andrew Walbranb5ab43c2020-04-30 11:32:54 +0100280 * - FFA_ERROR FFA_DENIED on failure, if the mailbox hasn't been read.
281 * - FFA_SUCCESS on success if no further action is needed.
282 * - FFA_RX_RELEASE if it was called by the primary VM and the primary VM now
Andrew Walbran8a0f5ca2019-11-05 13:12:23 +0000283 * needs to wake up or kick waiters. Waiters should be retrieved by calling
Wedson Almeida Filhoea62e2e2019-01-09 19:14:59 +0000284 * hf_mailbox_waiter_get.
Andrew Scull5ac05f02018-08-10 17:23:22 +0100285 */
Andrew Walbranb5ab43c2020-04-30 11:32:54 +0100286static inline struct ffa_value ffa_rx_release(void)
Andrew Scull5ac05f02018-08-10 17:23:22 +0100287{
Andrew Walbranb5ab43c2020-04-30 11:32:54 +0100288 return ffa_call((struct ffa_value){.func = FFA_RX_RELEASE_32});
Andrew Scull5ac05f02018-08-10 17:23:22 +0100289}
Andrew Walbran318f5732018-11-20 16:23:42 +0000290
291/**
Wedson Almeida Filhoea62e2e2019-01-09 19:14:59 +0000292 * Retrieves the next VM whose mailbox became writable. For a VM to be notified
293 * by this function, the caller must have called api_mailbox_send before with
294 * the notify argument set to true, and this call must have failed because the
295 * mailbox was not available.
296 *
Wedson Almeida Filhob790f652019-01-22 23:41:56 +0000297 * It should be called repeatedly to retrieve a list of VMs.
Wedson Almeida Filhoea62e2e2019-01-09 19:14:59 +0000298 *
299 * Returns -1 if no VM became writable, or the id of the VM whose mailbox
300 * became writable.
301 */
302static inline int64_t hf_mailbox_writable_get(void)
303{
304 return hf_call(HF_MAILBOX_WRITABLE_GET, 0, 0, 0);
305}
306
307/**
308 * Retrieves the next VM waiting to be notified that the mailbox of the
309 * specified VM became writable. Only primary VMs are allowed to call this.
310 *
Wedson Almeida Filhob790f652019-01-22 23:41:56 +0000311 * Returns -1 on failure or if there are no waiters; the VM id of the next
312 * waiter otherwise.
Wedson Almeida Filhoea62e2e2019-01-09 19:14:59 +0000313 */
Andrew Walbranb5ab43c2020-04-30 11:32:54 +0100314static inline int64_t hf_mailbox_waiter_get(ffa_vm_id_t vm_id)
Wedson Almeida Filhoea62e2e2019-01-09 19:14:59 +0000315{
316 return hf_call(HF_MAILBOX_WAITER_GET, vm_id, 0, 0);
317}
318
319/**
Andrew Walbran318f5732018-11-20 16:23:42 +0000320 * Enables or disables a given interrupt ID.
321 *
322 * Returns 0 on success, or -1 if the intid is invalid.
323 */
Manish Pandey35e452f2021-02-18 21:36:34 +0000324static inline int64_t hf_interrupt_enable(uint32_t intid, bool enable,
325 enum interrupt_type type)
Andrew Walbran318f5732018-11-20 16:23:42 +0000326{
Manish Pandey35e452f2021-02-18 21:36:34 +0000327 return hf_call(HF_INTERRUPT_ENABLE, intid, enable, type);
Andrew Walbran318f5732018-11-20 16:23:42 +0000328}
329
330/**
331 * Gets the ID of the pending interrupt (if any) and acknowledge it.
332 *
333 * Returns HF_INVALID_INTID if there are no pending interrupts.
334 */
Wedson Almeida Filhoc559d132019-01-09 19:33:40 +0000335static inline uint32_t hf_interrupt_get(void)
Andrew Walbran318f5732018-11-20 16:23:42 +0000336{
Wedson Almeida Filhoc559d132019-01-09 19:33:40 +0000337 return hf_call(HF_INTERRUPT_GET, 0, 0, 0);
Andrew Walbran318f5732018-11-20 16:23:42 +0000338}
339
340/**
341 * Injects a virtual interrupt of the given ID into the given target vCPU.
342 * This doesn't cause the vCPU to actually be run immediately; it will be taken
343 * when the vCPU is next run, which is up to the scheduler.
344 *
Andrew Walbran3d84a262018-12-13 14:41:19 +0000345 * Returns:
346 * - -1 on failure because the target VM or vCPU doesn't exist, the interrupt
347 * ID is invalid, or the current VM is not allowed to inject interrupts to
348 * the target VM.
349 * - 0 on success if no further action is needed.
350 * - 1 if it was called by the primary VM and the primary VM now needs to wake
351 * up or kick the target vCPU.
Andrew Walbran318f5732018-11-20 16:23:42 +0000352 */
Andrew Walbranb5ab43c2020-04-30 11:32:54 +0100353static inline int64_t hf_interrupt_inject(ffa_vm_id_t target_vm_id,
354 ffa_vcpu_index_t target_vcpu_idx,
Andrew Walbran318f5732018-11-20 16:23:42 +0000355 uint32_t intid)
356{
Wedson Almeida Filhoc559d132019-01-09 19:33:40 +0000357 return hf_call(HF_INTERRUPT_INJECT, target_vm_id, target_vcpu_idx,
Andrew Walbran318f5732018-11-20 16:23:42 +0000358 intid);
359}
Andrew Scull6386f252018-12-06 13:29:10 +0000360
361/**
Andrew Walbranc1ad4ce2019-05-09 11:41:39 +0100362 * Sends a character to the debug log for the VM.
363 *
364 * Returns 0 on success, or -1 if it failed for some reason.
365 */
366static inline int64_t hf_debug_log(char c)
367{
368 return hf_call(HF_DEBUG_LOG, c, 0, 0);
369}
370
Andrew Walbranb5ab43c2020-04-30 11:32:54 +0100371/** Obtains the Hafnium's version of the implemented FF-A specification. */
372static inline int32_t ffa_version(uint32_t requested_version)
Jose Marinhofc0b2b62019-06-06 11:18:45 +0100373{
Andrew Walbranb5ab43c2020-04-30 11:32:54 +0100374 return ffa_call((struct ffa_value){.func = FFA_VERSION_32,
375 .arg1 = requested_version})
Andrew Walbran9fd29072020-04-22 12:12:14 +0100376 .func;
Jose Marinhofc0b2b62019-06-06 11:18:45 +0100377}
Jose Marinhoc0f4ff22019-10-09 10:37:42 +0100378
379/**
380 * Discovery function returning information about the implementation of optional
Andrew Walbranb5ab43c2020-04-30 11:32:54 +0100381 * FF-A interfaces.
Jose Marinhoc0f4ff22019-10-09 10:37:42 +0100382 *
383 * Returns:
Andrew Walbranb5ab43c2020-04-30 11:32:54 +0100384 * - FFA_SUCCESS in .func if the optional interface with function_id is
Jose Marinhoc0f4ff22019-10-09 10:37:42 +0100385 * implemented.
Andrew Walbranb5ab43c2020-04-30 11:32:54 +0100386 * - FFA_ERROR in .func if the optional interface with function_id is not
Jose Marinhoc0f4ff22019-10-09 10:37:42 +0100387 * implemented.
388 */
Andrew Walbranb5ab43c2020-04-30 11:32:54 +0100389static inline struct ffa_value ffa_features(uint32_t function_id)
Jose Marinhoc0f4ff22019-10-09 10:37:42 +0100390{
Andrew Walbranb5ab43c2020-04-30 11:32:54 +0100391 return ffa_call((struct ffa_value){.func = FFA_FEATURES_32,
392 .arg1 = function_id});
Jose Marinhoc0f4ff22019-10-09 10:37:42 +0100393}
Olivier Deprezee9d6a92019-11-26 09:14:11 +0000394
395static inline struct ffa_value ffa_msg_send_direct_req(
396 ffa_vm_id_t sender_vm_id, ffa_vm_id_t target_vm_id, uint32_t arg3,
397 uint32_t arg4, uint32_t arg5, uint32_t arg6, uint32_t arg7)
398{
399 return ffa_call((struct ffa_value){
400 .func = FFA_MSG_SEND_DIRECT_REQ_32,
401 .arg1 = ((uint64_t)sender_vm_id << 16) | target_vm_id,
402 .arg3 = arg3,
403 .arg4 = arg4,
404 .arg5 = arg5,
405 .arg6 = arg6,
406 .arg7 = arg7,
407 });
408}
409
410static inline struct ffa_value ffa_msg_send_direct_resp(
411 ffa_vm_id_t sender_vm_id, ffa_vm_id_t target_vm_id, uint32_t arg3,
412 uint32_t arg4, uint32_t arg5, uint32_t arg6, uint32_t arg7)
413{
414 return ffa_call((struct ffa_value){
415 .func = FFA_MSG_SEND_DIRECT_RESP_32,
416 .arg1 = ((uint64_t)sender_vm_id << 16) | target_vm_id,
417 .arg3 = arg3,
418 .arg4 = arg4,
419 .arg5 = arg5,
420 .arg6 = arg6,
421 .arg7 = arg7,
422 });
423}
J-Alvesef69ac92021-08-26 09:21:27 +0100424
425static inline struct ffa_value ffa_notification_bind(
426 ffa_vm_id_t sender_vm_id, ffa_vm_id_t receiver_vm_id, uint32_t flags,
427 ffa_notifications_bitmap_t bitmap)
428{
429 return ffa_call((struct ffa_value){
430 .func = FFA_NOTIFICATION_BIND_32,
431 .arg1 = (sender_vm_id << 16) | (receiver_vm_id),
432 .arg2 = flags,
433 .arg3 = (uint32_t)(bitmap),
434 .arg4 = (uint32_t)(bitmap >> 32),
435 });
436}
437
438static inline struct ffa_value ffa_notification_unbind(
439 ffa_vm_id_t sender_vm_id, ffa_vm_id_t receiver_vm_id,
440 ffa_notifications_bitmap_t bitmap)
441{
442 return ffa_call((struct ffa_value){
443 .func = FFA_NOTIFICATION_UNBIND_32,
444 .arg1 = (sender_vm_id << 16) | (receiver_vm_id),
445 .arg3 = (uint32_t)(bitmap),
446 .arg4 = (uint32_t)(bitmap >> 32),
447 });
448}
449
450static inline struct ffa_value ffa_notification_set(
451 ffa_vm_id_t sender_vm_id, ffa_vm_id_t receiver_vm_id, uint32_t flags,
452 ffa_notifications_bitmap_t bitmap)
453{
454 return ffa_call((struct ffa_value){
455 .func = FFA_NOTIFICATION_SET_32,
456 .arg1 = (sender_vm_id << 16) | (receiver_vm_id),
457 .arg2 = flags,
458 .arg3 = (uint32_t)(bitmap),
459 .arg4 = (uint32_t)(bitmap >> 32),
460 });
461}
462
463static inline struct ffa_value ffa_notification_get(ffa_vm_id_t receiver_vm_id,
464 ffa_vcpu_index_t vcpu_id,
465 uint32_t flags)
466{
467 return ffa_call((struct ffa_value){
468 .func = FFA_NOTIFICATION_GET_32,
J-Alvesbe6e3032021-11-30 14:54:12 +0000469 .arg1 = (vcpu_id << 16) | (receiver_vm_id),
J-Alvesef69ac92021-08-26 09:21:27 +0100470 .arg2 = flags,
471 });
472}
473
474static inline struct ffa_value ffa_notification_info_get(void)
475{
476 return ffa_call((struct ffa_value){
477 .func = FFA_NOTIFICATION_INFO_GET_64,
478 });
479}
Raghu Krishnamurthyea6d25f2021-09-14 15:27:06 -0700480
481static inline struct ffa_value ffa_mem_perm_get(uint64_t base_va)
482{
483 return ffa_call((struct ffa_value){.func = FFA_MEM_PERM_GET_32,
484 .arg1 = base_va});
485}
486
487static inline struct ffa_value ffa_mem_perm_set(uint64_t base_va,
488 uint32_t page_count,
489 uint32_t mem_perm)
490{
491 return ffa_call((struct ffa_value){.func = FFA_MEM_PERM_SET_32,
492 .arg1 = base_va,
493 .arg2 = page_count,
494 .arg3 = mem_perm});
495}