All Downloads are FREE. Search and download functionalities are using the official Maven repository.

org.lwjgl.system.rpmalloc.RPMallocConfig Maven / Gradle / Ivy

Go to download

A public domain cross platform lock free thread caching 16-byte aligned memory allocator implemented in C.

There is a newer version: 3.3.4
Show newest version
/*
 * Copyright LWJGL. All rights reserved.
 * License terms: https://www.lwjgl.org/license
 * MACHINE GENERATED FILE, DO NOT EDIT
 */
package org.lwjgl.system.rpmalloc;

import org.jspecify.annotations.*;

import java.nio.*;

import org.lwjgl.*;
import org.lwjgl.system.*;

import static org.lwjgl.system.Checks.*;
import static org.lwjgl.system.MemoryUtil.*;
import static org.lwjgl.system.MemoryStack.*;

/**
 * This struct enables configuration of a memory mapper providing map/unmap of memory pages. Defaults to {@code VirtualAlloc}/{@code mmap} if none
 * provided. This allows rpmalloc to be used in contexts where memory is provided by internal means.
 * 
 * 

Page size may be set explicitly in initialization. This allows the allocator to be used as a sub-allocator where the page granularity should be lower * to reduce risk of wasting unused memory ranges.

* *

If rpmalloc is built with {@code ENABLE_GUARDS}, {@code memory_overwrite} may be set to detect writes before or after allocated memory blocks. This is * not enabled in the default LWJGL build.

* *

Layout

* *

 * struct rpmalloc_config_t {
 *     void * (*{@link RPMemoryMapCallbackI memory_map}) (size_t size, size_t *offset);
 *     void (*{@link RPMemoryUnmapCallbackI memory_unmap}) (void *address, size_t size, size_t offset, int release);
 *     void (*{@link RPErrorCallbackI error_callback}) (char const *message);
 *     int (*{@link RPMapFailCallbackI map_fail_callback}) (size_t size);
 *     size_t {@link #page_size};
 *     size_t {@link #span_size};
 *     size_t {@link #span_map_count};
 *     int {@link #enable_huge_pages};
 *     char const * {@link #page_name};
 *     char const * {@link #huge_page_name};
 * }
*/ @NativeType("struct rpmalloc_config_t") public class RPMallocConfig extends Struct implements NativeResource { /** The struct size in bytes. */ public static final int SIZEOF; /** The struct alignment in bytes. */ public static final int ALIGNOF; /** The struct member offsets. */ public static final int MEMORY_MAP, MEMORY_UNMAP, ERROR_CALLBACK, MAP_FAIL_CALLBACK, PAGE_SIZE, SPAN_SIZE, SPAN_MAP_COUNT, ENABLE_HUGE_PAGES, PAGE_NAME, HUGE_PAGE_NAME; static { Layout layout = __struct( __member(POINTER_SIZE), __member(POINTER_SIZE), __member(POINTER_SIZE), __member(POINTER_SIZE), __member(POINTER_SIZE), __member(POINTER_SIZE), __member(POINTER_SIZE), __member(4), __member(POINTER_SIZE), __member(POINTER_SIZE) ); SIZEOF = layout.getSize(); ALIGNOF = layout.getAlignment(); MEMORY_MAP = layout.offsetof(0); MEMORY_UNMAP = layout.offsetof(1); ERROR_CALLBACK = layout.offsetof(2); MAP_FAIL_CALLBACK = layout.offsetof(3); PAGE_SIZE = layout.offsetof(4); SPAN_SIZE = layout.offsetof(5); SPAN_MAP_COUNT = layout.offsetof(6); ENABLE_HUGE_PAGES = layout.offsetof(7); PAGE_NAME = layout.offsetof(8); HUGE_PAGE_NAME = layout.offsetof(9); } protected RPMallocConfig(long address, @Nullable ByteBuffer container) { super(address, container); } @Override protected RPMallocConfig create(long address, @Nullable ByteBuffer container) { return new RPMallocConfig(address, container); } /** * Creates a {@code RPMallocConfig} instance at the current position of the specified {@link ByteBuffer} container. Changes to the buffer's content will be * visible to the struct instance and vice versa. * *

The created instance holds a strong reference to the container object.

*/ public RPMallocConfig(ByteBuffer container) { super(memAddress(container), __checkContainer(container, SIZEOF)); } @Override public int sizeof() { return SIZEOF; } /** the memory map callback function */ @NativeType("void * (*) (size_t, size_t *)") public @Nullable RPMemoryMapCallback memory_map() { return nmemory_map(address()); } /** the memory unmap callback function */ @NativeType("void (*) (void *, size_t, size_t, int)") public @Nullable RPMemoryUnmapCallback memory_unmap() { return nmemory_unmap(address()); } /** the error callback function */ @NativeType("void (*) (char const *)") public @Nullable RPErrorCallback error_callback() { return nerror_callback(address()); } /** the map fail callback function */ @NativeType("int (*) (size_t)") public @Nullable RPMapFailCallback map_fail_callback() { return nmap_fail_callback(address()); } /** * the size of memory pages. * *

The page size MUST be a power of two in {@code [512,16384]} range (29 to 214) unless 0 - set to 0 to use system page size. All * memory mapping requests to {@code memory_map} will be made with size set to a multiple of the page size.

* *

Used if {@code RPMALLOC_CONFIGURABLE} is defined to 1, otherwise system page size is used.

*/ @NativeType("size_t") public long page_size() { return npage_size(address()); } /** * size of a span of memory blocks. * *

MUST be a power of two, and in {@code [4096,262144]} range (unless 0 - set to 0 to use the default span size).

* *

Used if {@code RPMALLOC_CONFIGURABLE} is defined to 1.

*/ @NativeType("size_t") public long span_size() { return nspan_size(address()); } /** * number of spans to map at each request to map new virtual memory blocks. * *

This can be used to minimize the system call overhead at the cost of virtual memory address space. The extra mapped pages will not be written until * actually used, so physical committed memory should not be affected in the default implementation.

* *

Will be aligned to a multiple of spans that match memory page size in case of huge pages.

*/ @NativeType("size_t") public long span_map_count() { return nspan_map_count(address()); } /** * enable use of large/huge pages. * *

If this flag is set to non-zero and page size is zero, the allocator will try to enable huge pages and auto detect the configuration. If this is set to * non-zero and page_size is also non-zero, the allocator will assume huge pages have been configured and enabled prior to initializing the allocator.

* *

For Windows, see large-page-support. For Linux, see * hugetlbpage.txt.

*/ @NativeType("int") public boolean enable_huge_pages() { return nenable_huge_pages(address()) != 0; } /** allocated pages name for systems supporting it to be able to distinguish among anonymous regions */ @NativeType("char const *") public @Nullable ByteBuffer page_name() { return npage_name(address()); } /** allocated pages name for systems supporting it to be able to distinguish among anonymous regions */ @NativeType("char const *") public @Nullable String page_nameString() { return npage_nameString(address()); } /** huge allocated pages name for systems supporting it to be able to distinguish among anonymous regions */ @NativeType("char const *") public @Nullable ByteBuffer huge_page_name() { return nhuge_page_name(address()); } /** huge allocated pages name for systems supporting it to be able to distinguish among anonymous regions */ @NativeType("char const *") public @Nullable String huge_page_nameString() { return nhuge_page_nameString(address()); } /** Sets the specified value to the {@link #memory_map} field. */ public RPMallocConfig memory_map(@Nullable @NativeType("void * (*) (size_t, size_t *)") RPMemoryMapCallbackI value) { nmemory_map(address(), value); return this; } /** Sets the specified value to the {@link #memory_unmap} field. */ public RPMallocConfig memory_unmap(@Nullable @NativeType("void (*) (void *, size_t, size_t, int)") RPMemoryUnmapCallbackI value) { nmemory_unmap(address(), value); return this; } /** Sets the specified value to the {@link #error_callback} field. */ public RPMallocConfig error_callback(@Nullable @NativeType("void (*) (char const *)") RPErrorCallbackI value) { nerror_callback(address(), value); return this; } /** Sets the specified value to the {@link #map_fail_callback} field. */ public RPMallocConfig map_fail_callback(@Nullable @NativeType("int (*) (size_t)") RPMapFailCallbackI value) { nmap_fail_callback(address(), value); return this; } /** Sets the specified value to the {@link #page_size} field. */ public RPMallocConfig page_size(@NativeType("size_t") long value) { npage_size(address(), value); return this; } /** Sets the specified value to the {@link #span_size} field. */ public RPMallocConfig span_size(@NativeType("size_t") long value) { nspan_size(address(), value); return this; } /** Sets the specified value to the {@link #span_map_count} field. */ public RPMallocConfig span_map_count(@NativeType("size_t") long value) { nspan_map_count(address(), value); return this; } /** Sets the specified value to the {@link #enable_huge_pages} field. */ public RPMallocConfig enable_huge_pages(@NativeType("int") boolean value) { nenable_huge_pages(address(), value ? 1 : 0); return this; } /** Sets the address of the specified encoded string to the {@link #page_name} field. */ public RPMallocConfig page_name(@Nullable @NativeType("char const *") ByteBuffer value) { npage_name(address(), value); return this; } /** Sets the address of the specified encoded string to the {@link #huge_page_name} field. */ public RPMallocConfig huge_page_name(@Nullable @NativeType("char const *") ByteBuffer value) { nhuge_page_name(address(), value); return this; } /** Initializes this struct with the specified values. */ public RPMallocConfig set( @Nullable RPMemoryMapCallbackI memory_map, @Nullable RPMemoryUnmapCallbackI memory_unmap, @Nullable RPErrorCallbackI error_callback, @Nullable RPMapFailCallbackI map_fail_callback, long page_size, long span_size, long span_map_count, boolean enable_huge_pages, @Nullable ByteBuffer page_name, @Nullable ByteBuffer huge_page_name ) { memory_map(memory_map); memory_unmap(memory_unmap); error_callback(error_callback); map_fail_callback(map_fail_callback); page_size(page_size); span_size(span_size); span_map_count(span_map_count); enable_huge_pages(enable_huge_pages); page_name(page_name); huge_page_name(huge_page_name); return this; } /** * Copies the specified struct data to this struct. * * @param src the source struct * * @return this struct */ public RPMallocConfig set(RPMallocConfig src) { memCopy(src.address(), address(), SIZEOF); return this; } // ----------------------------------- /** Returns a new {@code RPMallocConfig} instance allocated with {@link MemoryUtil#memAlloc memAlloc}. The instance must be explicitly freed. */ public static RPMallocConfig malloc() { return new RPMallocConfig(nmemAllocChecked(SIZEOF), null); } /** Returns a new {@code RPMallocConfig} instance allocated with {@link MemoryUtil#memCalloc memCalloc}. The instance must be explicitly freed. */ public static RPMallocConfig calloc() { return new RPMallocConfig(nmemCallocChecked(1, SIZEOF), null); } /** Returns a new {@code RPMallocConfig} instance allocated with {@link BufferUtils}. */ public static RPMallocConfig create() { ByteBuffer container = BufferUtils.createByteBuffer(SIZEOF); return new RPMallocConfig(memAddress(container), container); } /** Returns a new {@code RPMallocConfig} instance for the specified memory address. */ public static RPMallocConfig create(long address) { return new RPMallocConfig(address, null); } /** Like {@link #create(long) create}, but returns {@code null} if {@code address} is {@code NULL}. */ public static @Nullable RPMallocConfig createSafe(long address) { return address == NULL ? null : new RPMallocConfig(address, null); } // ----------------------------------- /** Deprecated for removal in 3.4.0. Use {@link #malloc(MemoryStack)} instead. */ @Deprecated public static RPMallocConfig mallocStack() { return malloc(stackGet()); } /** Deprecated for removal in 3.4.0. Use {@link #calloc(MemoryStack)} instead. */ @Deprecated public static RPMallocConfig callocStack() { return calloc(stackGet()); } /** Deprecated for removal in 3.4.0. Use {@link #malloc(MemoryStack)} instead. */ @Deprecated public static RPMallocConfig mallocStack(MemoryStack stack) { return malloc(stack); } /** Deprecated for removal in 3.4.0. Use {@link #calloc(MemoryStack)} instead. */ @Deprecated public static RPMallocConfig callocStack(MemoryStack stack) { return calloc(stack); } /** Deprecated for removal in 3.4.0. Use {@link #malloc(int, MemoryStack)} instead. */ /** * Returns a new {@code RPMallocConfig} instance allocated on the specified {@link MemoryStack}. * * @param stack the stack from which to allocate */ public static RPMallocConfig malloc(MemoryStack stack) { return new RPMallocConfig(stack.nmalloc(ALIGNOF, SIZEOF), null); } /** * Returns a new {@code RPMallocConfig} instance allocated on the specified {@link MemoryStack} and initializes all its bits to zero. * * @param stack the stack from which to allocate */ public static RPMallocConfig calloc(MemoryStack stack) { return new RPMallocConfig(stack.ncalloc(ALIGNOF, 1, SIZEOF), null); } // ----------------------------------- /** Unsafe version of {@link #memory_map}. */ public static @Nullable RPMemoryMapCallback nmemory_map(long struct) { return RPMemoryMapCallback.createSafe(memGetAddress(struct + RPMallocConfig.MEMORY_MAP)); } /** Unsafe version of {@link #memory_unmap}. */ public static @Nullable RPMemoryUnmapCallback nmemory_unmap(long struct) { return RPMemoryUnmapCallback.createSafe(memGetAddress(struct + RPMallocConfig.MEMORY_UNMAP)); } /** Unsafe version of {@link #error_callback}. */ public static @Nullable RPErrorCallback nerror_callback(long struct) { return RPErrorCallback.createSafe(memGetAddress(struct + RPMallocConfig.ERROR_CALLBACK)); } /** Unsafe version of {@link #map_fail_callback}. */ public static @Nullable RPMapFailCallback nmap_fail_callback(long struct) { return RPMapFailCallback.createSafe(memGetAddress(struct + RPMallocConfig.MAP_FAIL_CALLBACK)); } /** Unsafe version of {@link #page_size}. */ public static long npage_size(long struct) { return memGetAddress(struct + RPMallocConfig.PAGE_SIZE); } /** Unsafe version of {@link #span_size}. */ public static long nspan_size(long struct) { return memGetAddress(struct + RPMallocConfig.SPAN_SIZE); } /** Unsafe version of {@link #span_map_count}. */ public static long nspan_map_count(long struct) { return memGetAddress(struct + RPMallocConfig.SPAN_MAP_COUNT); } /** Unsafe version of {@link #enable_huge_pages}. */ public static int nenable_huge_pages(long struct) { return memGetInt(struct + RPMallocConfig.ENABLE_HUGE_PAGES); } /** Unsafe version of {@link #page_name}. */ public static @Nullable ByteBuffer npage_name(long struct) { return memByteBufferNT1Safe(memGetAddress(struct + RPMallocConfig.PAGE_NAME)); } /** Unsafe version of {@link #page_nameString}. */ public static @Nullable String npage_nameString(long struct) { return memASCIISafe(memGetAddress(struct + RPMallocConfig.PAGE_NAME)); } /** Unsafe version of {@link #huge_page_name}. */ public static @Nullable ByteBuffer nhuge_page_name(long struct) { return memByteBufferNT1Safe(memGetAddress(struct + RPMallocConfig.HUGE_PAGE_NAME)); } /** Unsafe version of {@link #huge_page_nameString}. */ public static @Nullable String nhuge_page_nameString(long struct) { return memASCIISafe(memGetAddress(struct + RPMallocConfig.HUGE_PAGE_NAME)); } /** Unsafe version of {@link #memory_map(RPMemoryMapCallbackI) memory_map}. */ public static void nmemory_map(long struct, @Nullable RPMemoryMapCallbackI value) { memPutAddress(struct + RPMallocConfig.MEMORY_MAP, memAddressSafe(value)); } /** Unsafe version of {@link #memory_unmap(RPMemoryUnmapCallbackI) memory_unmap}. */ public static void nmemory_unmap(long struct, @Nullable RPMemoryUnmapCallbackI value) { memPutAddress(struct + RPMallocConfig.MEMORY_UNMAP, memAddressSafe(value)); } /** Unsafe version of {@link #error_callback(RPErrorCallbackI) error_callback}. */ public static void nerror_callback(long struct, @Nullable RPErrorCallbackI value) { memPutAddress(struct + RPMallocConfig.ERROR_CALLBACK, memAddressSafe(value)); } /** Unsafe version of {@link #map_fail_callback(RPMapFailCallbackI) map_fail_callback}. */ public static void nmap_fail_callback(long struct, @Nullable RPMapFailCallbackI value) { memPutAddress(struct + RPMallocConfig.MAP_FAIL_CALLBACK, memAddressSafe(value)); } /** Unsafe version of {@link #page_size(long) page_size}. */ public static void npage_size(long struct, long value) { memPutAddress(struct + RPMallocConfig.PAGE_SIZE, value); } /** Unsafe version of {@link #span_size(long) span_size}. */ public static void nspan_size(long struct, long value) { memPutAddress(struct + RPMallocConfig.SPAN_SIZE, value); } /** Unsafe version of {@link #span_map_count(long) span_map_count}. */ public static void nspan_map_count(long struct, long value) { memPutAddress(struct + RPMallocConfig.SPAN_MAP_COUNT, value); } /** Unsafe version of {@link #enable_huge_pages(boolean) enable_huge_pages}. */ public static void nenable_huge_pages(long struct, int value) { memPutInt(struct + RPMallocConfig.ENABLE_HUGE_PAGES, value); } /** Unsafe version of {@link #page_name(ByteBuffer) page_name}. */ public static void npage_name(long struct, @Nullable ByteBuffer value) { if (CHECKS) { checkNT1Safe(value); } memPutAddress(struct + RPMallocConfig.PAGE_NAME, memAddressSafe(value)); } /** Unsafe version of {@link #huge_page_name(ByteBuffer) huge_page_name}. */ public static void nhuge_page_name(long struct, @Nullable ByteBuffer value) { if (CHECKS) { checkNT1Safe(value); } memPutAddress(struct + RPMallocConfig.HUGE_PAGE_NAME, memAddressSafe(value)); } }




© 2015 - 2025 Weber Informatics LLC | Privacy Policy