VirtualBox

source: vbox/trunk/include/iprt/memobj.h

Last change on this file was 106061, checked in by vboxsync, 2 months ago

Copyright year updates by scm.

  • Property eol-style set to native
  • Property svn:eol-style set to native
  • Property svn:keywords set to Id Revision
File size: 39.0 KB
Line 
1/** @file
2 * IPRT - Memory Objects (Ring-0).
3 */
4
5/*
6 * Copyright (C) 2006-2024 Oracle and/or its affiliates.
7 *
8 * This file is part of VirtualBox base platform packages, as
9 * available from https://www.virtualbox.org.
10 *
11 * This program is free software; you can redistribute it and/or
12 * modify it under the terms of the GNU General Public License
13 * as published by the Free Software Foundation, in version 3 of the
14 * License.
15 *
16 * This program is distributed in the hope that it will be useful, but
17 * WITHOUT ANY WARRANTY; without even the implied warranty of
18 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
19 * General Public License for more details.
20 *
21 * You should have received a copy of the GNU General Public License
22 * along with this program; if not, see <https://www.gnu.org/licenses>.
23 *
24 * The contents of this file may alternatively be used under the terms
25 * of the Common Development and Distribution License Version 1.0
26 * (CDDL), a copy of it is provided in the "COPYING.CDDL" file included
27 * in the VirtualBox distribution, in which case the provisions of the
28 * CDDL are applicable instead of those of the GPL.
29 *
30 * You may elect to license modified versions of this file under the
31 * terms and conditions of either the GPL or the CDDL or both.
32 *
33 * SPDX-License-Identifier: GPL-3.0-only OR CDDL-1.0
34 */
35
36#ifndef IPRT_INCLUDED_memobj_h
37#define IPRT_INCLUDED_memobj_h
38#ifndef RT_WITHOUT_PRAGMA_ONCE
39# pragma once
40#endif
41
42#include <iprt/cdefs.h>
43#include <iprt/types.h>
44
45RT_C_DECLS_BEGIN
46
47/** @defgroup grp_rt_memobj RTMemObj - Memory Object Manipulation (Ring-0)
48 * @ingroup grp_rt
49 * @{
50 */
51
52/** @def RTMEM_TAG
53 * The default allocation tag used by the RTMem allocation APIs.
54 *
55 * When not defined before the inclusion of iprt/memobj.h or iprt/mem.h, this
56 * will default to the pointer to the current file name. The memory API will
57 * make of use of this as pointer to a volatile but read-only string.
58 */
59#ifndef RTMEM_TAG
60# define RTMEM_TAG (__FILE__)
61#endif
62
63#ifdef IN_RING0
64
65/**
66 * Checks if this is mapping or not.
67 *
68 * @returns true if it's a mapping, otherwise false.
69 * @param MemObj The ring-0 memory object handle.
70 */
71RTR0DECL(bool) RTR0MemObjIsMapping(RTR0MEMOBJ MemObj);
72
73/**
74 * Gets the address of a ring-0 memory object.
75 *
76 * @returns The address of the memory object.
77 * @returns NULL if the handle is invalid (asserts in strict builds) or if there isn't any mapping.
78 * @param MemObj The ring-0 memory object handle.
79 */
80RTR0DECL(void *) RTR0MemObjAddress(RTR0MEMOBJ MemObj);
81
82/**
83 * Gets the ring-3 address of a ring-0 memory object.
84 *
85 * This only applies to ring-0 memory object with ring-3 mappings of some kind, i.e.
86 * locked user memory, reserved user address space and user mappings. This API should
87 * not be used on any other objects.
88 *
89 * @returns The address of the memory object.
90 * @returns NIL_RTR3PTR if the handle is invalid or if it's not an object with a ring-3 mapping.
91 * Strict builds will assert in both cases.
92 * @param MemObj The ring-0 memory object handle.
93 */
94RTR0DECL(RTR3PTR) RTR0MemObjAddressR3(RTR0MEMOBJ MemObj);
95
96/**
97 * Gets the size of a ring-0 memory object.
98 *
99 * The returned value may differ from the one specified to the API creating the
100 * object because of alignment adjustments. The minimal alignment currently
101 * employed by any API is PAGE_SIZE, so the result can safely be shifted by
102 * PAGE_SHIFT to calculate a page count.
103 *
104 * @returns The object size.
105 * @returns 0 if the handle is invalid (asserts in strict builds) or if there isn't any mapping.
106 * @param MemObj The ring-0 memory object handle.
107 */
108RTR0DECL(size_t) RTR0MemObjSize(RTR0MEMOBJ MemObj);
109
110/**
111 * Get the physical address of an page in the memory object.
112 *
113 * @returns The physical address.
114 * @returns NIL_RTHCPHYS if the object doesn't contain fixed physical pages.
115 * @returns NIL_RTHCPHYS if the iPage is out of range.
116 * @returns NIL_RTHCPHYS if the object handle isn't valid.
117 * @param MemObj The ring-0 memory object handle.
118 * @param iPage The page number within the object.
119 */
120RTR0DECL(RTHCPHYS) RTR0MemObjGetPagePhysAddr(RTR0MEMOBJ MemObj, size_t iPage);
121
122/**
123 * Checks whether the allocation was zero initialized or not.
124 *
125 * This only works on allocations. It is not meaningful for mappings, reserved
126 * memory and entered physical address, and will return false for these.
127 *
128 * @returns true if the allocation was initialized to zero at allocation time,
129 * false if not or query not meaningful to the object type.
130 * @param hMemObj The ring-0 memory object.
131 *
132 * @remarks It can be expected that memory allocated in the same fashion will
133 * have the same initialization state. So, if this returns true for
134 * one allocation it will return true for all other similarly made
135 * allocations.
136 */
137RTR0DECL(bool) RTR0MemObjWasZeroInitialized(RTR0MEMOBJ hMemObj);
138
139/**
140 * Initializes the allocation to zero.
141 *
142 * This only works on allocations, locked ring-0 memory and ring-0 mappings. It
143 * will return VERR_WRONG_TYPE if applied to any memory reservation,
144 * ring-3 mapping or ring-3 locking object.
145 *
146 * @returns IPRT status code.
147 * @param hMemObj The ring-0 memory object.
148 * @param fForce If @c true, always zero the allocation, if @c false
149 * it is only done when RTR0MemObjWasZeroInitialized()
150 * would return false.
151 */
152RTR0DECL(int) RTR0MemObjZeroInitialize(RTR0MEMOBJ hMemObj, bool fForce);
153
154/**
155 * Frees a ring-0 memory object.
156 *
157 * @returns IPRT status code.
158 * @retval VERR_INVALID_HANDLE if
159 * @param MemObj The ring-0 memory object to be freed. NULL is accepted.
160 * @param fFreeMappings Whether or not to free mappings of the object.
161 */
162RTR0DECL(int) RTR0MemObjFree(RTR0MEMOBJ MemObj, bool fFreeMappings);
163
164/**
165 * Allocates page aligned virtual kernel memory (default tag).
166 *
167 * The memory is taken from a non paged (= fixed physical memory backing) pool.
168 *
169 * @returns IPRT status code.
170 * @param pMemObj Where to store the ring-0 memory object handle.
171 * @param cb Number of bytes to allocate. This is rounded up to nearest page.
172 * @param fExecutable Flag indicating whether it should be permitted to
173 * executed code in the memory object. The user must
174 * use RTR0MemObjProtect after initialization the
175 * allocation to actually make it executable.
176 */
177#define RTR0MemObjAllocPage(pMemObj, cb, fExecutable) \
178 RTR0MemObjAllocPageTag((pMemObj), (cb), (fExecutable), RTMEM_TAG)
179
180/**
181 * Allocates page aligned virtual kernel memory (custom tag).
182 *
183 * The memory is taken from a non paged (= fixed physical memory backing) pool.
184 *
185 * @returns IPRT status code.
186 * @param pMemObj Where to store the ring-0 memory object handle.
187 * @param cb Number of bytes to allocate. This is rounded up to nearest page.
188 * @param fExecutable Flag indicating whether it should be permitted to
189 * executed code in the memory object. The user must
190 * use RTR0MemObjProtect after initialization the
191 * allocation to actually make it executable.
192 * @param pszTag Allocation tag used for statistics and such.
193 */
194RTR0DECL(int) RTR0MemObjAllocPageTag(PRTR0MEMOBJ pMemObj, size_t cb, bool fExecutable, const char *pszTag);
195
196/**
197 * Allocates large page aligned virtual kernel memory (default tag).
198 *
199 * Each large page in the allocation is backed by a contiguous chunk of physical
200 * memory aligned to the page size. The memory is taken from a non paged (=
201 * fixed physical memory backing) pool.
202 *
203 * On some hosts we only support allocating a single large page at a time, they
204 * will return VERR_NOT_SUPPORTED if @a cb is larger than @a cbLargePage.
205 *
206 * @returns IPRT status code.
207 * @retval VERR_TRY_AGAIN instead of VERR_NO_MEMORY when
208 * RTMEMOBJ_ALLOC_LARGE_F_FAST is set and supported.
209 * @param pMemObj Where to store the ring-0 memory object handle.
210 * @param cb Number of bytes to allocate. This is rounded up to
211 * nearest large page.
212 * @param cbLargePage The large page size. The allowed values varies from
213 * architecture to architecture and the paging mode
214 * used by the OS.
215 * @param fFlags Flags, RTMEMOBJ_ALLOC_LARGE_F_XXX.
216 *
217 * @note The implicit kernel mapping of this allocation does not necessarily
218 * have to be aligned on a @a cbLargePage boundrary.
219 */
220#define RTR0MemObjAllocLarge(pMemObj, cb, cbLargePage, fFlags) \
221 RTR0MemObjAllocLargeTag((pMemObj), (cb), (cbLargePage), (fFlags), RTMEM_TAG)
222
223/**
224 * Allocates large page aligned virtual kernel memory (custom tag).
225 *
226 * Each large page in the allocation is backed by a contiguous chunk of physical
227 * memory aligned to the page size. The memory is taken from a non paged (=
228 * fixed physical memory backing) pool.
229 *
230 * On some hosts we only support allocating a single large page at a time, they
231 * will return VERR_NOT_SUPPORTED if @a cb is larger than @a cbLargePage.
232 *
233 * @returns IPRT status code.
234 * @retval VERR_TRY_AGAIN instead of VERR_NO_MEMORY when
235 * RTMEMOBJ_ALLOC_LARGE_F_FAST is set and supported.
236 * @param pMemObj Where to store the ring-0 memory object handle.
237 * @param cb Number of bytes to allocate. This is rounded up to
238 * nearest large page.
239 * @param cbLargePage The large page size. The allowed values varies from
240 * architecture to architecture and the paging mode
241 * used by the OS.
242 * @param fFlags Flags, RTMEMOBJ_ALLOC_LARGE_F_XXX.
243 * @param pszTag Allocation tag used for statistics and such.
244 *
245 * @note The implicit kernel mapping of this allocation does not necessarily
246 * have to be aligned on a @a cbLargePage boundrary.
247 */
248RTR0DECL(int) RTR0MemObjAllocLargeTag(PRTR0MEMOBJ pMemObj, size_t cb, size_t cbLargePage, uint32_t fFlags, const char *pszTag);
249
250/** @name RTMEMOBJ_ALLOC_LARGE_F_XXX
251 * @{ */
252/** Indicates that it is okay to fail if there aren't enough large pages handy,
253 * cancelling any expensive search and reshuffling of memory (when supported).
254 * @note This flag can't be realized on all OSes. (Those who do support it
255 * will return VERR_TRY_AGAIN instead of VERR_NO_MEMORY if they
256 * cannot satisfy the request.) */
257#define RTMEMOBJ_ALLOC_LARGE_F_FAST RT_BIT_32(0)
258/** Mask with valid bits. */
259#define RTMEMOBJ_ALLOC_LARGE_F_VALID_MASK UINT32_C(0x00000001)
260/** @} */
261
262/**
263 * Allocates page aligned virtual kernel memory with physical backing below 4GB
264 * (default tag).
265 *
266 * The physical memory backing the allocation is fixed.
267 *
268 * @returns IPRT status code.
269 * @param pMemObj Where to store the ring-0 memory object handle.
270 * @param cb Number of bytes to allocate. This is rounded up to nearest page.
271 * @param fExecutable Flag indicating whether it should be permitted to
272 * executed code in the memory object. The user must
273 * use RTR0MemObjProtect after initialization the
274 * allocation to actually make it executable.
275 */
276#define RTR0MemObjAllocLow(pMemObj, cb, fExecutable) \
277 RTR0MemObjAllocLowTag((pMemObj), (cb), (fExecutable), RTMEM_TAG)
278
279/**
280 * Allocates page aligned virtual kernel memory with physical backing below 4GB
281 * (custom tag).
282 *
283 * The physical memory backing the allocation is fixed.
284 *
285 * @returns IPRT status code.
286 * @param pMemObj Where to store the ring-0 memory object handle.
287 * @param cb Number of bytes to allocate. This is rounded up to nearest page.
288 * @param fExecutable Flag indicating whether it should be permitted to
289 * executed code in the memory object. The user must
290 * use RTR0MemObjProtect after initialization the
291 * allocation to actually make it executable.
292 * @param pszTag Allocation tag used for statistics and such.
293 */
294RTR0DECL(int) RTR0MemObjAllocLowTag(PRTR0MEMOBJ pMemObj, size_t cb, bool fExecutable, const char *pszTag);
295
296/**
297 * Allocates page aligned virtual kernel memory with contiguous physical backing
298 * (default tag).
299 *
300 * The physical memory backing the allocation is fixed.
301 *
302 * @returns IPRT status code.
303 * @param pMemObj Where to store the ring-0 memory object handle.
304 * @param cb Number of bytes to allocate. This is rounded up to nearest page.
305 * @param PhysHighest The highest permitable address (inclusive).
306 * Pass NIL_RTHCPHYS if any address is acceptable.
307 * @param fExecutable Flag indicating whether it should be permitted to
308 * executed code in the memory object. The user must
309 * use RTR0MemObjProtect after initialization the
310 * allocation to actually make it executable.
311 */
312#define RTR0MemObjAllocCont(pMemObj, cb, PhysHighest, fExecutable) \
313 RTR0MemObjAllocContTag((pMemObj), (cb), (PhysHighest), (fExecutable), RTMEM_TAG)
314
315/**
316 * Allocates page aligned virtual kernel memory with contiguous physical
317 * backing (custom tag).
318 *
319 * The physical memory backing the allocation is fixed.
320 *
321 * @returns IPRT status code.
322 * @param pMemObj Where to store the ring-0 memory object handle.
323 * @param cb Number of bytes to allocate. This is rounded up to nearest page.
324 * @param PhysHighest The highest permitable address (inclusive).
325 * Pass NIL_RTHCPHYS if any address is acceptable.
326 * @param fExecutable Flag indicating whether it should be permitted to
327 * executed code in the memory object. The user must
328 * use RTR0MemObjProtect after initialization the
329 * allocation to actually make it executable.
330 * @param pszTag Allocation tag used for statistics and such.
331 */
332RTR0DECL(int) RTR0MemObjAllocContTag(PRTR0MEMOBJ pMemObj, size_t cb, RTHCPHYS PhysHighest, bool fExecutable, const char *pszTag);
333
334/**
335 * Locks a range of user virtual memory (default tag).
336 *
337 * @returns IPRT status code.
338 * @param pMemObj Where to store the ring-0 memory object handle.
339 * @param R3Ptr User virtual address. This is rounded down to a page
340 * boundary.
341 * @param cb Number of bytes to lock. This is rounded up to
342 * nearest page boundary.
343 * @param fAccess The desired access, a combination of RTMEM_PROT_READ
344 * and RTMEM_PROT_WRITE.
345 * @param R0Process The process to lock pages in. NIL_RTR0PROCESS is an
346 * alias for the current one.
347 *
348 * @remarks RTR0MemGetAddressR3() and RTR0MemGetAddress() will return therounded
349 * down address.
350 *
351 * @remarks Linux: This API requires that the memory begin locked is in a memory
352 * mapping that is not required in any forked off child process. This
353 * is not intented as permanent restriction, feel free to help out
354 * lifting it.
355 */
356#define RTR0MemObjLockUser(pMemObj, R3Ptr, cb, fAccess, R0Process) \
357 RTR0MemObjLockUserTag((pMemObj), (R3Ptr), (cb), (fAccess), (R0Process), RTMEM_TAG)
358
359/**
360 * Locks a range of user virtual memory (custom tag).
361 *
362 * @returns IPRT status code.
363 * @param pMemObj Where to store the ring-0 memory object handle.
364 * @param R3Ptr User virtual address. This is rounded down to a page
365 * boundary.
366 * @param cb Number of bytes to lock. This is rounded up to
367 * nearest page boundary.
368 * @param fAccess The desired access, a combination of RTMEM_PROT_READ
369 * and RTMEM_PROT_WRITE.
370 * @param R0Process The process to lock pages in. NIL_RTR0PROCESS is an
371 * alias for the current one.
372 * @param pszTag Allocation tag used for statistics and such.
373 *
374 * @remarks RTR0MemGetAddressR3() and RTR0MemGetAddress() will return therounded
375 * down address.
376 *
377 * @remarks Linux: This API requires that the memory begin locked is in a memory
378 * mapping that is not required in any forked off child process. This
379 * is not intented as permanent restriction, feel free to help out
380 * lifting it.
381 */
382RTR0DECL(int) RTR0MemObjLockUserTag(PRTR0MEMOBJ pMemObj, RTR3PTR R3Ptr, size_t cb, uint32_t fAccess,
383 RTR0PROCESS R0Process, const char *pszTag);
384
385/**
386 * Locks a range of kernel virtual memory (default tag).
387 *
388 * @returns IPRT status code.
389 * @param pMemObj Where to store the ring-0 memory object handle.
390 * @param pv Kernel virtual address. This is rounded down to a page boundary.
391 * @param cb Number of bytes to lock. This is rounded up to nearest page boundary.
392 * @param fAccess The desired access, a combination of RTMEM_PROT_READ
393 * and RTMEM_PROT_WRITE.
394 *
395 * @remark RTR0MemGetAddress() will return the rounded down address.
396 */
397#define RTR0MemObjLockKernel(pMemObj, pv, cb, fAccess) \
398 RTR0MemObjLockKernelTag((pMemObj), (pv), (cb), (fAccess), RTMEM_TAG)
399
400/**
401 * Locks a range of kernel virtual memory (custom tag).
402 *
403 * @returns IPRT status code.
404 * @param pMemObj Where to store the ring-0 memory object handle.
405 * @param pv Kernel virtual address. This is rounded down to a page boundary.
406 * @param cb Number of bytes to lock. This is rounded up to nearest page boundary.
407 * @param fAccess The desired access, a combination of RTMEM_PROT_READ
408 * and RTMEM_PROT_WRITE.
409 * @param pszTag Allocation tag used for statistics and such.
410 *
411 * @remark RTR0MemGetAddress() will return the rounded down address.
412 */
413RTR0DECL(int) RTR0MemObjLockKernelTag(PRTR0MEMOBJ pMemObj, void *pv, size_t cb, uint32_t fAccess, const char *pszTag);
414
415/**
416 * Allocates contiguous page aligned physical memory without (necessarily) any
417 * kernel mapping (default tag).
418 *
419 * @returns IPRT status code.
420 * @param pMemObj Where to store the ring-0 memory object handle.
421 * @param cb Number of bytes to allocate. This is rounded up to nearest page.
422 * @param PhysHighest The highest permitable address (inclusive).
423 * Pass NIL_RTHCPHYS if any address is acceptable.
424 */
425#define RTR0MemObjAllocPhys(pMemObj, cb, PhysHighest) \
426 RTR0MemObjAllocPhysTag((pMemObj), (cb), (PhysHighest), RTMEM_TAG)
427
428/**
429 * Allocates contiguous page aligned physical memory without (necessarily) any
430 * kernel mapping (custom tag).
431 *
432 * @returns IPRT status code.
433 * @param pMemObj Where to store the ring-0 memory object handle.
434 * @param cb Number of bytes to allocate. This is rounded up to nearest page.
435 * @param PhysHighest The highest permitable address (inclusive).
436 * Pass NIL_RTHCPHYS if any address is acceptable.
437 * @param pszTag Allocation tag used for statistics and such.
438 */
439RTR0DECL(int) RTR0MemObjAllocPhysTag(PRTR0MEMOBJ pMemObj, size_t cb, RTHCPHYS PhysHighest, const char *pszTag);
440
441/**
442 * Allocates contiguous physical memory without (necessarily) any kernel mapping
443 * (default tag).
444 *
445 * @returns IPRT status code.
446 * @param pMemObj Where to store the ring-0 memory object handle.
447 * @param cb Number of bytes to allocate. This is rounded up to nearest page.
448 * @param PhysHighest The highest permitable address (inclusive).
449 * Pass NIL_RTHCPHYS if any address is acceptable.
450 * @param uAlignment The alignment of the reserved memory.
451 * Supported values are 0 (alias for PAGE_SIZE), PAGE_SIZE, _2M, _4M and _1G.
452 */
453#define RTR0MemObjAllocPhysEx(pMemObj, cb, PhysHighest, uAlignment) \
454 RTR0MemObjAllocPhysExTag((pMemObj), (cb), (PhysHighest), (uAlignment), RTMEM_TAG)
455
456/**
457 * Allocates contiguous physical memory without (necessarily) any kernel mapping
458 * (custom tag).
459 *
460 * @returns IPRT status code.
461 * @param pMemObj Where to store the ring-0 memory object handle.
462 * @param cb Number of bytes to allocate. This is rounded up to nearest page.
463 * @param PhysHighest The highest permitable address (inclusive).
464 * Pass NIL_RTHCPHYS if any address is acceptable.
465 * @param uAlignment The alignment of the reserved memory.
466 * Supported values are 0 (alias for PAGE_SIZE), PAGE_SIZE, _2M, _4M and _1G.
467 * @param pszTag Allocation tag used for statistics and such.
468 */
469RTR0DECL(int) RTR0MemObjAllocPhysExTag(PRTR0MEMOBJ pMemObj, size_t cb, RTHCPHYS PhysHighest, size_t uAlignment, const char *pszTag);
470
471/**
472 * Allocates non-contiguous page aligned physical memory without (necessarily)
473 * any kernel mapping (default tag).
474 *
475 * This API is for allocating huge amounts of pages and will return
476 * VERR_NOT_SUPPORTED if this cannot be implemented in a satisfactory
477 * manner.
478 *
479 * @returns IPRT status code.
480 * @retval VERR_NOT_SUPPORTED if it's not possible to allocated unmapped
481 * physical memory on this platform. The caller should expect
482 * this error and have a fallback strategy for it.
483 *
484 * @param pMemObj Where to store the ring-0 memory object handle.
485 * @param cb Number of bytes to allocate. This is rounded up to nearest page.
486 * @param PhysHighest The highest permitable address (inclusive).
487 * Pass NIL_RTHCPHYS if any address is acceptable.
488 */
489#define RTR0MemObjAllocPhysNC(pMemObj, cb, PhysHighest) \
490 RTR0MemObjAllocPhysNCTag((pMemObj), (cb), (PhysHighest), RTMEM_TAG)
491
492/**
493 * Allocates non-contiguous page aligned physical memory without (necessarily)
494 * any kernel mapping (custom tag).
495 *
496 * This API is for allocating huge amounts of pages and will return
497 * VERR_NOT_SUPPORTED if this cannot be implemented in a satisfactory
498 * manner.
499 *
500 * @returns IPRT status code.
501 * @retval VERR_NOT_SUPPORTED if it's not possible to allocated unmapped
502 * physical memory on this platform. The caller should expect
503 * this error and have a fallback strategy for it.
504 *
505 * @param pMemObj Where to store the ring-0 memory object handle.
506 * @param cb Number of bytes to allocate. This is rounded up to nearest page.
507 * @param PhysHighest The highest permitable address (inclusive).
508 * Pass NIL_RTHCPHYS if any address is acceptable.
509 * @param pszTag Allocation tag used for statistics and such.
510 */
511RTR0DECL(int) RTR0MemObjAllocPhysNCTag(PRTR0MEMOBJ pMemObj, size_t cb, RTHCPHYS PhysHighest, const char *pszTag);
512
513/** Memory cache policy for RTR0MemObjEnterPhys.
514 * @{
515 */
516/** Default caching policy -- don't care. */
517#define RTMEM_CACHE_POLICY_DONT_CARE UINT32_C(0)
518/** MMIO caching policy -- uncachable. */
519#define RTMEM_CACHE_POLICY_MMIO UINT32_C(1)
520/** @} */
521
522/**
523 * Creates a page aligned, contiguous, physical memory object (default tag).
524 *
525 * No physical memory is allocated, we trust you do know what you're doing.
526 *
527 * @returns IPRT status code.
528 * @param pMemObj Where to store the ring-0 memory object handle.
529 * @param Phys The physical address to start at. This is rounded down to the
530 * nearest page boundary.
531 * @param cb The size of the object in bytes. This is rounded up to nearest page boundary.
532 * @param uCachePolicy One of the RTMEM_CACHE_XXX modes.
533 */
534#define RTR0MemObjEnterPhys(pMemObj, Phys, cb, uCachePolicy) \
535 RTR0MemObjEnterPhysTag((pMemObj), (Phys), (cb), (uCachePolicy), RTMEM_TAG)
536
537/**
538 * Creates a page aligned, contiguous, physical memory object (custom tag).
539 *
540 * No physical memory is allocated, we trust you do know what you're doing.
541 *
542 * @returns IPRT status code.
543 * @param pMemObj Where to store the ring-0 memory object handle.
544 * @param Phys The physical address to start at. This is rounded down to the
545 * nearest page boundary.
546 * @param cb The size of the object in bytes. This is rounded up to nearest page boundary.
547 * @param uCachePolicy One of the RTMEM_CACHE_XXX modes.
548 * @param pszTag Allocation tag used for statistics and such.
549 */
550RTR0DECL(int) RTR0MemObjEnterPhysTag(PRTR0MEMOBJ pMemObj, RTHCPHYS Phys, size_t cb, uint32_t uCachePolicy, const char *pszTag);
551
552/**
553 * Reserves kernel virtual address space (default tag).
554 *
555 * If this function fails with VERR_NOT_SUPPORTED, the idea is that you
556 * can use RTR0MemObjEnterPhys() + RTR0MemObjMapKernel() as a fallback if
557 * you have a safe physical address range to make use of...
558 *
559 * @returns IPRT status code.
560 * @param pMemObj Where to store the ring-0 memory object handle.
561 * @param pvFixed Requested address. (void *)-1 means any address. This must match the alignment.
562 * @param cb The number of bytes to reserve. This is rounded up to nearest page.
563 * @param uAlignment The alignment of the reserved memory.
564 * Supported values are 0 (alias for PAGE_SIZE), PAGE_SIZE, _2M and _4M.
565 */
566#define RTR0MemObjReserveKernel(pMemObj, pvFixed, cb, uAlignment) \
567 RTR0MemObjReserveKernelTag((pMemObj), (pvFixed), (cb), (uAlignment), RTMEM_TAG)
568
569/**
570 * Reserves kernel virtual address space (custom tag).
571 *
572 * If this function fails with VERR_NOT_SUPPORTED, the idea is that you
573 * can use RTR0MemObjEnterPhys() + RTR0MemObjMapKernel() as a fallback if
574 * you have a safe physical address range to make use of...
575 *
576 * @returns IPRT status code.
577 * @param pMemObj Where to store the ring-0 memory object handle.
578 * @param pvFixed Requested address. (void *)-1 means any address. This must match the alignment.
579 * @param cb The number of bytes to reserve. This is rounded up to nearest page.
580 * @param uAlignment The alignment of the reserved memory.
581 * Supported values are 0 (alias for PAGE_SIZE), PAGE_SIZE, _2M and _4M.
582 * @param pszTag Allocation tag used for statistics and such.
583 */
584RTR0DECL(int) RTR0MemObjReserveKernelTag(PRTR0MEMOBJ pMemObj, void *pvFixed, size_t cb, size_t uAlignment, const char *pszTag);
585
586/**
587 * Reserves user virtual address space in the current process (default tag).
588 *
589 * @returns IPRT status code.
590 * @param pMemObj Where to store the ring-0 memory object handle.
591 * @param R3PtrFixed Requested address. (RTR3PTR)-1 means any address. This must match the alignment.
592 * @param cb The number of bytes to reserve. This is rounded up to nearest PAGE_SIZE.
593 * @param uAlignment The alignment of the reserved memory.
594 * Supported values are 0 (alias for PAGE_SIZE), PAGE_SIZE, _2M and _4M.
595 * @param R0Process The process to reserve the memory in.
596 * NIL_RTR0PROCESS is an alias for the current one.
597 */
598#define RTR0MemObjReserveUser(pMemObj, R3PtrFixed, cb, uAlignment, R0Process) \
599 RTR0MemObjReserveUserTag((pMemObj), (R3PtrFixed), (cb), (uAlignment), (R0Process), RTMEM_TAG)
600
601/**
602 * Reserves user virtual address space in the current process (custom tag).
603 *
604 * @returns IPRT status code.
605 * @param pMemObj Where to store the ring-0 memory object handle.
606 * @param R3PtrFixed Requested address. (RTR3PTR)-1 means any address. This must match the alignment.
607 * @param cb The number of bytes to reserve. This is rounded up to nearest PAGE_SIZE.
608 * @param uAlignment The alignment of the reserved memory.
609 * Supported values are 0 (alias for PAGE_SIZE), PAGE_SIZE, _2M and _4M.
610 * @param R0Process The process to reserve the memory in.
611 * NIL_RTR0PROCESS is an alias for the current one.
612 * @param pszTag Allocation tag used for statistics and such.
613 */
614RTR0DECL(int) RTR0MemObjReserveUserTag(PRTR0MEMOBJ pMemObj, RTR3PTR R3PtrFixed, size_t cb, size_t uAlignment,
615 RTR0PROCESS R0Process, const char *pszTag);
616
617/**
618 * Maps a memory object into kernel virtual address space (default tag).
619 *
620 * This is the same as calling RTR0MemObjMapKernelEx with cbSub and offSub set
621 * to zero.
622 *
623 * @returns IPRT status code.
624 * @param pMemObj Where to store the ring-0 memory object handle of the mapping object.
625 * @param MemObjToMap The object to be map.
626 * @param pvFixed Requested address. (void *)-1 means any address. This must match the alignment.
627 * @param uAlignment The alignment of the reserved memory.
628 * Supported values are 0 (alias for PAGE_SIZE), PAGE_SIZE, _2M and _4M.
629 * @param fProt Combination of RTMEM_PROT_* flags (except RTMEM_PROT_NONE).
630 */
631#define RTR0MemObjMapKernel(pMemObj, MemObjToMap, pvFixed, uAlignment, fProt) \
632 RTR0MemObjMapKernelTag((pMemObj), (MemObjToMap), (pvFixed), (uAlignment), (fProt), RTMEM_TAG)
633
634/**
635 * Maps a memory object into kernel virtual address space (custom tag).
636 *
637 * This is the same as calling RTR0MemObjMapKernelEx with cbSub and offSub set
638 * to zero.
639 *
640 * @returns IPRT status code.
641 * @param pMemObj Where to store the ring-0 memory object handle of the mapping object.
642 * @param MemObjToMap The object to be map.
643 * @param pvFixed Requested address. (void *)-1 means any address. This must match the alignment.
644 * @param uAlignment The alignment of the reserved memory.
645 * Supported values are 0 (alias for PAGE_SIZE), PAGE_SIZE, _2M and _4M.
646 * @param fProt Combination of RTMEM_PROT_* flags (except RTMEM_PROT_NONE).
647 * @param pszTag Allocation tag used for statistics and such.
648 */
649RTR0DECL(int) RTR0MemObjMapKernelTag(PRTR0MEMOBJ pMemObj, RTR0MEMOBJ MemObjToMap, void *pvFixed,
650 size_t uAlignment, unsigned fProt, const char *pszTag);
651
652/**
653 * Maps a memory object into kernel virtual address space (default tag).
654 *
655 * The ability to map subsections of the object into kernel space is currently
656 * not implemented on all platforms. All/Most of platforms supports mapping the
657 * whole object into kernel space.
658 *
659 * @returns IPRT status code.
660 * @retval VERR_NOT_SUPPORTED if it's not possible to map a subsection of a
661 * memory object on this platform. When you hit this, try implement it.
662 *
663 * @param pMemObj Where to store the ring-0 memory object handle of the mapping object.
664 * @param MemObjToMap The object to be map.
665 * @param pvFixed Requested address. (void *)-1 means any address. This must match the alignment.
666 * @param uAlignment The alignment of the reserved memory.
667 * Supported values are 0 (alias for PAGE_SIZE), PAGE_SIZE, _2M and _4M.
668 * @param fProt Combination of RTMEM_PROT_* flags (except RTMEM_PROT_NONE).
669 * @param offSub Where in the object to start mapping. If non-zero
670 * the value must be page aligned and cbSub must be
671 * non-zero as well.
672 * @param cbSub The size of the part of the object to be mapped. If
673 * zero the entire object is mapped. The value must be
674 * page aligned.
675 */
676#define RTR0MemObjMapKernelEx(pMemObj, MemObjToMap, pvFixed, uAlignment, fProt, offSub, cbSub) \
677 RTR0MemObjMapKernelExTag((pMemObj), (MemObjToMap), (pvFixed), (uAlignment), (fProt), (offSub), (cbSub), RTMEM_TAG)
678
679/**
680 * Maps a memory object into kernel virtual address space (custom tag).
681 *
682 * The ability to map subsections of the object into kernel space is currently
683 * not implemented on all platforms. All/Most of platforms supports mapping the
684 * whole object into kernel space.
685 *
686 * @returns IPRT status code.
687 * @retval VERR_NOT_SUPPORTED if it's not possible to map a subsection of a
688 * memory object on this platform. When you hit this, try implement it.
689 *
690 * @param pMemObj Where to store the ring-0 memory object handle of the mapping object.
691 * @param MemObjToMap The object to be map.
692 * @param pvFixed Requested address. (void *)-1 means any address. This must match the alignment.
693 * @param uAlignment The alignment of the reserved memory.
694 * Supported values are 0 (alias for PAGE_SIZE), PAGE_SIZE, _2M and _4M.
695 * @param fProt Combination of RTMEM_PROT_* flags (except RTMEM_PROT_NONE).
696 * @param offSub Where in the object to start mapping. If non-zero
697 * the value must be page aligned and cbSub must be
698 * non-zero as well.
699 * @param cbSub The size of the part of the object to be mapped. If
700 * zero the entire object is mapped. The value must be
701 * page aligned.
702 * @param pszTag Allocation tag used for statistics and such.
703 */
704RTR0DECL(int) RTR0MemObjMapKernelExTag(PRTR0MEMOBJ pMemObj, RTR0MEMOBJ MemObjToMap, void *pvFixed, size_t uAlignment,
705 unsigned fProt, size_t offSub, size_t cbSub, const char *pszTag);
706
707/**
708 * Maps a memory object into user virtual address space in the current process
709 * (default tag).
710 *
711 * @returns IPRT status code.
712 * @param pMemObj Where to store the ring-0 memory object handle of the mapping object.
713 * @param MemObjToMap The object to be map.
714 * @param R3PtrFixed Requested address. (RTR3PTR)-1 means any address. This must match the alignment.
715 * @param uAlignment The alignment of the reserved memory.
716 * Supported values are 0 (alias for PAGE_SIZE), PAGE_SIZE, _2M and _4M.
717 * @param fProt Combination of RTMEM_PROT_* flags (except RTMEM_PROT_NONE).
718 * @param R0Process The process to map the memory into. NIL_RTR0PROCESS
719 * is an alias for the current one.
720 */
721#define RTR0MemObjMapUser(pMemObj, MemObjToMap, R3PtrFixed, uAlignment, fProt, R0Process) \
722 RTR0MemObjMapUserTag((pMemObj), (MemObjToMap), (R3PtrFixed), (uAlignment), (fProt), (R0Process), RTMEM_TAG)
723
724/**
725 * Maps a memory object into user virtual address space in the current process
726 * (custom tag).
727 *
728 * @returns IPRT status code.
729 * @param pMemObj Where to store the ring-0 memory object handle of the mapping object.
730 * @param MemObjToMap The object to be map.
731 * @param R3PtrFixed Requested address. (RTR3PTR)-1 means any address. This must match the alignment.
732 * @param uAlignment The alignment of the reserved memory.
733 * Supported values are 0 (alias for PAGE_SIZE), PAGE_SIZE, _2M and _4M.
734 * @param fProt Combination of RTMEM_PROT_* flags (except RTMEM_PROT_NONE).
735 * @param R0Process The process to map the memory into. NIL_RTR0PROCESS
736 * is an alias for the current one.
737 * @param pszTag Allocation tag used for statistics and such.
738 */
739RTR0DECL(int) RTR0MemObjMapUserTag(PRTR0MEMOBJ pMemObj, RTR0MEMOBJ MemObjToMap, RTR3PTR R3PtrFixed,
740 size_t uAlignment, unsigned fProt, RTR0PROCESS R0Process, const char *pszTag);
741
742/**
743 * Maps a memory object into user virtual address space in the current process
744 * (default tag).
745 *
746 * @returns IPRT status code.
747 * @param pMemObj Where to store the ring-0 memory object handle of the mapping object.
748 * @param MemObjToMap The object to be map.
749 * @param R3PtrFixed Requested address. (RTR3PTR)-1 means any address. This must match the alignment.
750 * @param uAlignment The alignment of the reserved memory.
751 * Supported values are 0 (alias for PAGE_SIZE), PAGE_SIZE, _2M and _4M.
752 * @param fProt Combination of RTMEM_PROT_* flags (except RTMEM_PROT_NONE).
753 * @param R0Process The process to map the memory into. NIL_RTR0PROCESS
754 * is an alias for the current one.
755 * @param offSub Where in the object to start mapping. If non-zero
756 * the value must be page aligned and cbSub must be
757 * non-zero as well.
758 * @param cbSub The size of the part of the object to be mapped. If
759 * zero the entire object is mapped. The value must be
760 * page aligned.
761 */
762#define RTR0MemObjMapUserEx(pMemObj, MemObjToMap, R3PtrFixed, uAlignment, fProt, R0Process, offSub, cbSub) \
763 RTR0MemObjMapUserExTag((pMemObj), (MemObjToMap), (R3PtrFixed), (uAlignment), (fProt), (R0Process), \
764 (offSub), (cbSub), RTMEM_TAG)
765
766/**
767 * Maps a memory object into user virtual address space in the current process
768 * (custom tag).
769 *
770 * @returns IPRT status code.
771 * @param pMemObj Where to store the ring-0 memory object handle of the mapping object.
772 * @param MemObjToMap The object to be map.
773 * @param R3PtrFixed Requested address. (RTR3PTR)-1 means any address. This must match the alignment.
774 * @param uAlignment The alignment of the reserved memory.
775 * Supported values are 0 (alias for PAGE_SIZE), PAGE_SIZE, _2M and _4M.
776 * @param fProt Combination of RTMEM_PROT_* flags (except RTMEM_PROT_NONE).
777 * @param R0Process The process to map the memory into. NIL_RTR0PROCESS
778 * is an alias for the current one.
779 * @param offSub Where in the object to start mapping. If non-zero
780 * the value must be page aligned and cbSub must be
781 * non-zero as well.
782 * @param cbSub The size of the part of the object to be mapped. If
783 * zero the entire object is mapped. The value must be
784 * page aligned.
785 * @param pszTag Allocation tag used for statistics and such.
786 */
787RTR0DECL(int) RTR0MemObjMapUserExTag(PRTR0MEMOBJ pMemObj, RTR0MEMOBJ MemObjToMap, RTR3PTR R3PtrFixed, size_t uAlignment,
788 unsigned fProt, RTR0PROCESS R0Process, size_t offSub, size_t cbSub, const char *pszTag);
789
790/**
791 * Change the page level protection of one or more pages in a memory object.
792 *
793 * @returns IPRT status code.
794 * @retval VERR_NOT_SUPPORTED if the OS doesn't provide any way to manipulate
795 * page level protection. The caller must handle this status code
796 * gracefully. (Note that it may also occur if the implementation is
797 * missing, in which case just go ahead and implement it.)
798 *
799 * @param hMemObj Memory object handle.
800 * @param offSub Offset into the memory object. Must be page aligned.
801 * @param cbSub Number of bytes to change the protection of. Must be
802 * page aligned.
803 * @param fProt Combination of RTMEM_PROT_* flags.
804 */
805RTR0DECL(int) RTR0MemObjProtect(RTR0MEMOBJ hMemObj, size_t offSub, size_t cbSub, uint32_t fProt);
806
807#endif /* IN_RING0 */
808
809/** @} */
810
811RT_C_DECLS_END
812
813#endif /* !IPRT_INCLUDED_memobj_h */
814
Note: See TracBrowser for help on using the repository browser.

© 2024 Oracle Support Privacy / Do Not Sell My Info Terms of Use Trademark Policy Automated Access Etiquette