mirror of
https://github.com/odin-lang/Odin.git
synced 2026-01-06 21:17:40 +00:00
708 lines
17 KiB
Odin
708 lines
17 KiB
Odin
#shared_global_scope;
|
|
|
|
#import "os.odin";
|
|
#import "fmt.odin";
|
|
#import "mem.odin";
|
|
#import "utf8.odin";
|
|
#import "hash.odin";
|
|
|
|
// IMPORTANT NOTE(bill): `type_info` & `type_info_val` cannot be used within a
|
|
// #shared_global_scope due to the internals of the compiler.
|
|
// This could change at a later date if the all these data structures are
|
|
// implemented within the compiler rather than in this "preload" file
|
|
|
|
|
|
// IMPORTANT NOTE(bill): Do not change the order of any of this data
|
|
// The compiler relies upon this _exact_ order
|
|
Type_Info_Enum_Value :: raw_union {
|
|
f: f64,
|
|
i: i64,
|
|
}
|
|
// NOTE(bill): This must match the compiler's
|
|
Calling_Convention :: enum {
|
|
ODIN = 0,
|
|
C = 1,
|
|
STD = 2,
|
|
FAST = 3,
|
|
}
|
|
|
|
Type_Info_Record :: struct #ordered {
|
|
types: []^Type_Info,
|
|
names: []string,
|
|
offsets: []int, // offsets may not be used in tuples
|
|
size: int, // in bytes
|
|
align: int, // in bytes
|
|
packed: bool,
|
|
ordered: bool,
|
|
custom_align: bool,
|
|
}
|
|
|
|
Type_Info :: union {
|
|
Named{name: string, base: ^Type_Info},
|
|
Integer{size: int, signed: bool},
|
|
Float{size: int},
|
|
String{},
|
|
Boolean{},
|
|
Any{},
|
|
Pointer{
|
|
elem: ^Type_Info, // nil -> rawptr
|
|
},
|
|
Procedure{
|
|
params: ^Type_Info, // Type_Info.Tuple
|
|
results: ^Type_Info, // Type_Info.Tuple
|
|
variadic: bool,
|
|
convention: Calling_Convention,
|
|
},
|
|
Array{
|
|
elem: ^Type_Info,
|
|
elem_size: int,
|
|
count: int,
|
|
},
|
|
Dynamic_Array{elem: ^Type_Info, elem_size: int},
|
|
Slice {elem: ^Type_Info, elem_size: int},
|
|
Vector {elem: ^Type_Info, elem_size, count, align: int},
|
|
Tuple {using record: Type_Info_Record}, // Only really used for procedures
|
|
Struct {using record: Type_Info_Record},
|
|
Raw_Union {using record: Type_Info_Record},
|
|
Union{
|
|
common_fields: struct {
|
|
types: []^Type_Info,
|
|
names: []string,
|
|
offsets: []int, // offsets may not be used in tuples
|
|
},
|
|
variant_names: []string,
|
|
variant_types: []^Type_Info,
|
|
size: int,
|
|
align: int,
|
|
},
|
|
Enum{
|
|
base: ^Type_Info,
|
|
names: []string,
|
|
values: []Type_Info_Enum_Value,
|
|
},
|
|
Map{
|
|
key: ^Type_Info,
|
|
value: ^Type_Info,
|
|
generated_struct: ^Type_Info,
|
|
count: int, // == 0 if dynamic
|
|
},
|
|
}
|
|
|
|
|
|
// NOTE(bill): only the ones that are needed (not all types)
|
|
// This will be set by the compiler
|
|
__type_table: []Type_Info;
|
|
|
|
type_info_base :: proc(info: ^Type_Info) -> ^Type_Info {
|
|
if info == nil {
|
|
return nil;
|
|
}
|
|
base := info;
|
|
match i in base {
|
|
case Type_Info.Named:
|
|
base = i.base;
|
|
}
|
|
return base;
|
|
}
|
|
|
|
|
|
type_info_base_without_enum :: proc(info: ^Type_Info) -> ^Type_Info {
|
|
if info == nil {
|
|
return nil;
|
|
}
|
|
base := info;
|
|
match i in base {
|
|
case Type_Info.Named:
|
|
base = i.base;
|
|
case Type_Info.Enum:
|
|
base = i.base;
|
|
}
|
|
return base;
|
|
}
|
|
|
|
|
|
|
|
assume :: proc(cond: bool) #foreign __llvm_core "llvm.assume";
|
|
|
|
__debug_trap :: proc() #foreign __llvm_core "llvm.debugtrap";
|
|
__trap :: proc() #foreign __llvm_core "llvm.trap";
|
|
read_cycle_counter :: proc() -> u64 #foreign __llvm_core "llvm.readcyclecounter";
|
|
|
|
|
|
|
|
|
|
// IMPORTANT NOTE(bill): Must be in this order (as the compiler relies upon it)
|
|
Allocator_Mode :: enum u8 {
|
|
ALLOC,
|
|
FREE,
|
|
FREE_ALL,
|
|
RESIZE,
|
|
}
|
|
Allocator_Proc :: #type proc(allocator_data: rawptr, mode: Allocator_Mode,
|
|
size, alignment: int,
|
|
old_memory: rawptr, old_size: int, flags: u64) -> rawptr;
|
|
Allocator :: struct #ordered {
|
|
procedure: Allocator_Proc,
|
|
data: rawptr,
|
|
}
|
|
|
|
Context :: struct #ordered {
|
|
thread_id: int,
|
|
|
|
allocator: Allocator,
|
|
|
|
user_data: rawptr,
|
|
user_index: int,
|
|
}
|
|
|
|
#thread_local __context: Context;
|
|
|
|
|
|
DEFAULT_ALIGNMENT :: align_of([vector 4]f32);
|
|
|
|
|
|
__check_context :: proc() {
|
|
c := ^__context;
|
|
|
|
if c.allocator.procedure == nil {
|
|
c.allocator = default_allocator();
|
|
}
|
|
if c.thread_id == 0 {
|
|
c.thread_id = os.current_thread_id();
|
|
}
|
|
}
|
|
|
|
alloc :: proc(size: int) -> rawptr #inline { return alloc_align(size, DEFAULT_ALIGNMENT); }
|
|
|
|
alloc_align :: proc(size, alignment: int) -> rawptr #inline {
|
|
__check_context();
|
|
a := context.allocator;
|
|
return a.procedure(a.data, Allocator_Mode.ALLOC, size, alignment, nil, 0, 0);
|
|
}
|
|
|
|
free_ptr_with_allocator :: proc(a: Allocator, ptr: rawptr) #inline {
|
|
if ptr == nil {
|
|
return;
|
|
}
|
|
if a.procedure == nil {
|
|
return;
|
|
}
|
|
a.procedure(a.data, Allocator_Mode.FREE, 0, 0, ptr, 0, 0);
|
|
}
|
|
|
|
free_ptr :: proc(ptr: rawptr) #inline {
|
|
__check_context();
|
|
free_ptr_with_allocator(context.allocator, ptr);
|
|
}
|
|
|
|
free_all :: proc() #inline {
|
|
__check_context();
|
|
a := context.allocator;
|
|
a.procedure(a.data, Allocator_Mode.FREE_ALL, 0, 0, nil, 0, 0);
|
|
}
|
|
|
|
|
|
resize :: proc(ptr: rawptr, old_size, new_size: int) -> rawptr #inline { return resize_align(ptr, old_size, new_size, DEFAULT_ALIGNMENT); }
|
|
resize_align :: proc(ptr: rawptr, old_size, new_size, alignment: int) -> rawptr #inline {
|
|
__check_context();
|
|
a := context.allocator;
|
|
return a.procedure(a.data, Allocator_Mode.RESIZE, new_size, alignment, ptr, old_size, 0);
|
|
}
|
|
|
|
|
|
|
|
default_resize_align :: proc(old_memory: rawptr, old_size, new_size, alignment: int) -> rawptr {
|
|
if old_memory == nil {
|
|
return alloc_align(new_size, alignment);
|
|
}
|
|
|
|
if new_size == 0 {
|
|
free(old_memory);
|
|
return nil;
|
|
}
|
|
|
|
if new_size == old_size {
|
|
return old_memory;
|
|
}
|
|
|
|
new_memory := alloc_align(new_size, alignment);
|
|
if new_memory == nil {
|
|
return nil;
|
|
}
|
|
|
|
mem.copy(new_memory, old_memory, min(old_size, new_size));;
|
|
free(old_memory);
|
|
return new_memory;
|
|
}
|
|
|
|
|
|
default_allocator_proc :: proc(allocator_data: rawptr, mode: Allocator_Mode,
|
|
size, alignment: int,
|
|
old_memory: rawptr, old_size: int, flags: u64) -> rawptr {
|
|
using Allocator_Mode;
|
|
|
|
match mode {
|
|
case ALLOC:
|
|
return os.heap_alloc(size);
|
|
|
|
case FREE:
|
|
os.heap_free(old_memory);
|
|
return nil;
|
|
|
|
case FREE_ALL:
|
|
// NOTE(bill): Does nothing
|
|
|
|
case RESIZE:
|
|
ptr := os.heap_resize(old_memory, size);
|
|
assert(ptr != nil);
|
|
return ptr;
|
|
}
|
|
|
|
return nil;
|
|
}
|
|
|
|
default_allocator :: proc() -> Allocator {
|
|
return Allocator{
|
|
procedure = default_allocator_proc,
|
|
data = nil,
|
|
};
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
__string_eq :: proc(a, b: string) -> bool {
|
|
if a.count != b.count {
|
|
return false;
|
|
}
|
|
if a.data == b.data {
|
|
return true;
|
|
}
|
|
return __string_cmp(a, b) == 0;
|
|
}
|
|
|
|
__string_cmp :: proc(a, b: string) -> int {
|
|
return mem.compare(cast([]byte)a, cast([]byte)b);
|
|
}
|
|
|
|
__string_ne :: proc(a, b: string) -> bool #inline { return !__string_eq(a, b); }
|
|
__string_lt :: proc(a, b: string) -> bool #inline { return __string_cmp(a, b) < 0; }
|
|
__string_gt :: proc(a, b: string) -> bool #inline { return __string_cmp(a, b) > 0; }
|
|
__string_le :: proc(a, b: string) -> bool #inline { return __string_cmp(a, b) <= 0; }
|
|
__string_ge :: proc(a, b: string) -> bool #inline { return __string_cmp(a, b) >= 0; }
|
|
|
|
|
|
__assert :: proc(file: string, line, column: int, msg: string) #inline {
|
|
fmt.fprintf(os.stderr, "%s(%d:%d) Runtime assertion: %s\n",
|
|
file, line, column, msg);
|
|
__debug_trap();
|
|
}
|
|
__panic :: proc(file: string, line, column: int, msg: string) #inline {
|
|
fmt.fprintf(os.stderr, "%s(%d:%d) Panic: %s\n",
|
|
file, line, column, msg);
|
|
__debug_trap();
|
|
}
|
|
__bounds_check_error :: proc(file: string, line, column: int, index, count: int) {
|
|
if 0 <= index && index < count {
|
|
return;
|
|
}
|
|
fmt.fprintf(os.stderr, "%s(%d:%d) Index %d is out of bounds range 0..%d\n",
|
|
file, line, column, index, count);
|
|
__debug_trap();
|
|
}
|
|
|
|
__slice_expr_error :: proc(file: string, line, column: int, low, high, max: int) {
|
|
if 0 <= low && low <= high && high <= max {
|
|
return;
|
|
}
|
|
fmt.fprintf(os.stderr, "%s(%d:%d) Invalid slice indices: [%d..%d..%d]\n",
|
|
file, line, column, low, high, max);
|
|
__debug_trap();
|
|
}
|
|
__substring_expr_error :: proc(file: string, line, column: int, low, high: int) {
|
|
if 0 <= low && low <= high {
|
|
return;
|
|
}
|
|
fmt.fprintf(os.stderr, "%s(%d:%d) Invalid substring indices: [%d..%d]\n",
|
|
file, line, column, low, high);
|
|
__debug_trap();
|
|
}
|
|
__union_cast_check :: proc(ok: bool, file: string, line, column: int, from, to: ^Type_Info) {
|
|
if !ok {
|
|
fmt.fprintf(os.stderr, "%s(%d:%d) Invalid `union_cast` from %T to %T\n",
|
|
file, line, column, from, to);
|
|
__debug_trap();
|
|
}
|
|
}
|
|
|
|
__string_decode_rune :: proc(s: string) -> (rune, int) #inline {
|
|
return utf8.decode_rune(s);
|
|
}
|
|
|
|
|
|
__mem_set :: proc(data: rawptr, value: i32, len: int) -> rawptr {
|
|
llvm_memset_64bit :: proc(dst: rawptr, val: byte, len: int, align: i32, is_volatile: bool) #foreign __llvm_core "llvm.memset.p0i8.i64";
|
|
llvm_memset_64bit(data, cast(byte)value, len, 1, false);
|
|
return data;
|
|
}
|
|
__mem_zero :: proc(data: rawptr, len: int) -> rawptr {
|
|
return __mem_set(data, 0, len);
|
|
}
|
|
__mem_copy :: proc(dst, src: rawptr, len: int) -> rawptr {
|
|
// NOTE(bill): This _must_ be implemented like C's memmove
|
|
llvm_memmove_64bit :: proc(dst, src: rawptr, len: int, align: i32, is_volatile: bool) #foreign __llvm_core "llvm.memmove.p0i8.p0i8.i64";
|
|
llvm_memmove_64bit(dst, src, len, 1, false);
|
|
return dst;
|
|
}
|
|
__mem_copy_non_overlapping :: proc(dst, src: rawptr, len: int) -> rawptr {
|
|
// NOTE(bill): This _must_ be implemented like C's memcpy
|
|
llvm_memcpy_64bit :: proc(dst, src: rawptr, len: int, align: i32, is_volatile: bool) #foreign __llvm_core "llvm.memcpy.p0i8.p0i8.i64";
|
|
llvm_memcpy_64bit(dst, src, len, 1, false);
|
|
return dst;
|
|
}
|
|
|
|
__mem_compare :: proc(a, b: ^byte, n: int) -> int {
|
|
for i in 0..n {
|
|
match {
|
|
case (a+i)^ < (b+i)^:
|
|
return -1;
|
|
case (a+i)^ > (b+i)^:
|
|
return +1;
|
|
}
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
|
|
Raw_Any :: struct #ordered {
|
|
type_info: ^Type_Info,
|
|
data: rawptr,
|
|
}
|
|
|
|
Raw_String :: struct #ordered {
|
|
data: ^byte,
|
|
count: int,
|
|
};
|
|
|
|
Raw_Slice :: struct #ordered {
|
|
data: rawptr,
|
|
count: int,
|
|
capacity: int,
|
|
};
|
|
|
|
Raw_Dynamic_Array :: struct #ordered {
|
|
data: rawptr,
|
|
count: int,
|
|
capacity: int,
|
|
allocator: Allocator,
|
|
};
|
|
|
|
Raw_Dynamic_Map :: struct #ordered {
|
|
hashes: [dynamic]int,
|
|
entries: Raw_Dynamic_Array,
|
|
};
|
|
|
|
|
|
|
|
__dynamic_array_reserve :: proc(array_: rawptr, elem_size, elem_align: int, capacity: int) -> bool {
|
|
array := cast(^Raw_Dynamic_Array)array_;
|
|
|
|
if capacity <= array.capacity {
|
|
return true;
|
|
}
|
|
|
|
__check_context();
|
|
if array.allocator.procedure == nil {
|
|
array.allocator = context.allocator;
|
|
}
|
|
assert(array.allocator.procedure != nil);
|
|
|
|
old_size := array.capacity * elem_size;
|
|
new_size := capacity * elem_size;
|
|
allocator := array.allocator;
|
|
|
|
new_data := allocator.procedure(allocator.data, Allocator_Mode.RESIZE, new_size, elem_align, array.data, old_size, 0);
|
|
if new_data == nil {
|
|
return false;
|
|
}
|
|
|
|
array.data = new_data;
|
|
array.capacity = capacity;
|
|
return true;
|
|
}
|
|
|
|
|
|
__dynamic_array_append :: proc(array_: rawptr, elem_size, elem_align: int,
|
|
items: rawptr, item_count: int) -> int {
|
|
array := cast(^Raw_Dynamic_Array)array_;
|
|
|
|
if item_count <= 0 || items == nil {
|
|
return array.count;
|
|
}
|
|
|
|
|
|
ok := true;
|
|
if array.capacity <= array.count+item_count {
|
|
capacity := 2 * array.capacity + max(8, item_count);
|
|
ok = __dynamic_array_reserve(array, elem_size, elem_align, capacity);
|
|
}
|
|
if !ok {
|
|
// TODO(bill): Better error handling for failed reservation
|
|
return array.count;
|
|
}
|
|
data := cast(^byte)array.data;
|
|
assert(data != nil);
|
|
mem.copy(data + (elem_size*array.count), items, elem_size * item_count);
|
|
array.count += item_count;
|
|
return array.count;
|
|
}
|
|
|
|
__dynamic_array_append_nothing :: proc(array_: rawptr, elem_size, elem_align: int) -> int {
|
|
array := cast(^Raw_Dynamic_Array)array_;
|
|
|
|
ok := true;
|
|
if array.capacity <= array.count+1 {
|
|
capacity := 2 * array.capacity + max(8, 1);
|
|
ok = __dynamic_array_reserve(array, elem_size, elem_align, capacity);
|
|
}
|
|
if !ok {
|
|
// TODO(bill): Better error handling for failed reservation
|
|
return array.count;
|
|
}
|
|
data := cast(^byte)array.data;
|
|
assert(data != nil);
|
|
mem.zero(data + (elem_size*array.count), elem_size);
|
|
array.count++;
|
|
return array.count;
|
|
}
|
|
|
|
__slice_append :: proc(slice_: rawptr, elem_size, elem_align: int,
|
|
items: rawptr, item_count: int) -> int {
|
|
slice := cast(^Raw_Slice)slice_;
|
|
|
|
if item_count <= 0 || items == nil {
|
|
return slice.count;
|
|
}
|
|
|
|
item_count = min(slice.capacity-slice.count, item_count);
|
|
if item_count > 0 {
|
|
data := cast(^byte)slice.data;
|
|
assert(data != nil);
|
|
mem.copy(data + (elem_size*slice.count), items, elem_size * item_count);
|
|
slice.count += item_count;
|
|
}
|
|
return slice.count;
|
|
}
|
|
|
|
|
|
// Map stuff
|
|
|
|
__default_hash :: proc(data: []byte) -> u64 {
|
|
return hash.fnv64a(data);
|
|
}
|
|
__default_hash_string :: proc(s: string) -> u64 {
|
|
return __default_hash(cast([]byte)s);
|
|
}
|
|
|
|
__Map_Key :: struct #ordered {
|
|
hash: u64,
|
|
str: string,
|
|
}
|
|
|
|
__Map_Find_Result :: struct #ordered {
|
|
hash_index: int,
|
|
entry_prev: int,
|
|
entry_index: int,
|
|
}
|
|
|
|
__Map_Entry_Header :: struct #ordered {
|
|
key: __Map_Key,
|
|
next: int,
|
|
/*
|
|
value: Value_Type,
|
|
*/
|
|
}
|
|
|
|
__Map_Header :: struct #ordered {
|
|
m: ^Raw_Dynamic_Map,
|
|
is_key_string: bool,
|
|
entry_size: int,
|
|
entry_align: int,
|
|
value_offset: int,
|
|
}
|
|
|
|
__dynamic_map_reserve :: proc(using header: __Map_Header, capacity: int) -> bool {
|
|
h := __dynamic_array_reserve(^m.hashes, size_of(int), align_of(int), capacity);
|
|
e := __dynamic_array_reserve(^m.entries, entry_size, entry_align, capacity);
|
|
return h && e;
|
|
}
|
|
|
|
__dynamic_map_rehash :: proc(using header: __Map_Header, new_count: int) {
|
|
new_header := header;
|
|
nm: Raw_Dynamic_Map;
|
|
new_header.m = ^nm;
|
|
|
|
reserve(nm.hashes, new_count);
|
|
nm.hashes.count = nm.hashes.capacity;
|
|
__dynamic_array_reserve(^nm.entries, entry_size, entry_align, m.entries.count);
|
|
for _, i in nm.hashes {
|
|
nm.hashes[i] = -1;
|
|
}
|
|
|
|
for i := 0; i < nm.entries.count; i++ {
|
|
entry_header := __dynamic_map_get_entry(new_header, i);
|
|
data := cast(^byte)entry_header;
|
|
|
|
if nm.hashes.count == 0 {
|
|
__dynamic_map_grow(new_header);
|
|
}
|
|
|
|
fr := __dynamic_map_find(new_header, entry_header.key);
|
|
j := __dynamic_map_add_entry(new_header, entry_header.key);
|
|
if fr.entry_prev < 0 {
|
|
nm.hashes[fr.hash_index] = j;
|
|
} else {
|
|
e := __dynamic_map_get_entry(new_header, fr.entry_prev);
|
|
e.next = j;
|
|
}
|
|
|
|
e := __dynamic_map_get_entry(new_header, j);
|
|
e.next = fr.entry_index;
|
|
ndata := cast(^byte)e;
|
|
mem.copy(ndata+value_offset, data+value_offset, entry_size-value_offset);
|
|
if __dynamic_map_full(new_header) {
|
|
__dynamic_map_grow(new_header);
|
|
}
|
|
}
|
|
free_ptr_with_allocator(header.m.hashes.allocator, header.m.hashes.data);
|
|
free_ptr_with_allocator(header.m.entries.allocator, header.m.entries.data);
|
|
header.m^ = nm;
|
|
}
|
|
|
|
__dynamic_map_get :: proc(h: __Map_Header, key: __Map_Key) -> rawptr {
|
|
index := __dynamic_map_find(h, key).entry_index;
|
|
if index >= 0 {
|
|
data := cast(^byte)__dynamic_map_get_entry(h, index);
|
|
val := data + h.value_offset;
|
|
return val;
|
|
}
|
|
return nil;
|
|
}
|
|
|
|
__dynamic_map_set :: proc(using h: __Map_Header, key: __Map_Key, value: rawptr) {
|
|
index: int;
|
|
|
|
if m.hashes.count == 0 {
|
|
__dynamic_map_grow(h);
|
|
}
|
|
fr := __dynamic_map_find(h, key);
|
|
if fr.entry_index >= 0 {
|
|
index = fr.entry_index;
|
|
} else {
|
|
index = __dynamic_map_add_entry(h, key);
|
|
if fr.entry_prev >= 0 {
|
|
entry := __dynamic_map_get_entry(h, fr.entry_prev);
|
|
entry.next = index;
|
|
} else {
|
|
m.hashes[fr.hash_index] = index;
|
|
}
|
|
}
|
|
{
|
|
data := cast(^byte)__dynamic_map_get_entry(h, index);
|
|
val := data+value_offset;
|
|
mem.copy(val, value, entry_size-value_offset);
|
|
}
|
|
|
|
if __dynamic_map_full(h) {
|
|
__dynamic_map_grow(h);
|
|
}
|
|
}
|
|
|
|
|
|
__dynamic_map_grow :: proc(using h: __Map_Header) {
|
|
new_count := 2*m.entries.count + 8;
|
|
__dynamic_map_rehash(h, new_count);
|
|
}
|
|
|
|
__dynamic_map_full :: proc(using h: __Map_Header) -> bool {
|
|
return cast(int)(0.75 * cast(f64)m.hashes.count) <= m.entries.count;
|
|
}
|
|
|
|
|
|
__dynamic_map_hash_equal :: proc(h: __Map_Header, a, b: __Map_Key) -> bool {
|
|
if a.hash == b.hash {
|
|
if h.is_key_string {
|
|
return a.str == b.str;
|
|
}
|
|
return true;
|
|
}
|
|
return false;
|
|
}
|
|
|
|
__dynamic_map_find :: proc(using h: __Map_Header, key: __Map_Key) -> __Map_Find_Result {
|
|
fr := __Map_Find_Result{-1, -1, -1};
|
|
if m.hashes.count > 0 {
|
|
fr.hash_index = cast(int)(key.hash % cast(u64)m.hashes.count);
|
|
fr.entry_index = m.hashes[fr.hash_index];
|
|
for fr.entry_index >= 0 {
|
|
entry := __dynamic_map_get_entry(h, fr.entry_index);
|
|
if __dynamic_map_hash_equal(h, entry.key, key) {
|
|
return fr;
|
|
}
|
|
fr.entry_prev = fr.entry_index;
|
|
fr.entry_index = entry.next;
|
|
}
|
|
}
|
|
return fr;
|
|
}
|
|
|
|
__dynamic_map_add_entry :: proc(using h: __Map_Header, key: __Map_Key) -> int {
|
|
prev := m.entries.count;
|
|
c := __dynamic_array_append_nothing(^m.entries, entry_size, entry_align);
|
|
if c != prev {
|
|
end := __dynamic_map_get_entry(h, c-1);
|
|
end.key = key;
|
|
end.next = -1;
|
|
}
|
|
return prev;
|
|
}
|
|
|
|
|
|
__dynamic_map_delete :: proc(using h: __Map_Header, key: __Map_Key) {
|
|
fr := __dynamic_map_find(h, key);
|
|
if fr.entry_index >= 0 {
|
|
__dynamic_map_erase(h, fr);
|
|
}
|
|
}
|
|
|
|
__dynamic_map_get_entry :: proc(using h: __Map_Header, index: int) -> ^__Map_Entry_Header {
|
|
data := cast(^byte)m.entries.data + index*entry_size;
|
|
return cast(^__Map_Entry_Header)data;
|
|
}
|
|
|
|
__dynamic_map_erase :: proc(using h: __Map_Header, fr: __Map_Find_Result) {
|
|
if fr.entry_prev < 0 {
|
|
m.hashes[fr.hash_index] = __dynamic_map_get_entry(h, fr.entry_index).next;
|
|
} else {
|
|
__dynamic_map_get_entry(h, fr.entry_prev).next = __dynamic_map_get_entry(h, fr.entry_index).next;
|
|
}
|
|
|
|
if fr.entry_index == m.entries.count-1 {
|
|
m.entries.count--;
|
|
}
|
|
mem.copy(__dynamic_map_get_entry(h, fr.entry_index), __dynamic_map_get_entry(h, m.entries.count-1), entry_size);
|
|
last := __dynamic_map_find(h, __dynamic_map_get_entry(h, fr.entry_index).key);
|
|
if last.entry_prev >= 0 {
|
|
__dynamic_map_get_entry(h, last.entry_prev).next = fr.entry_index;
|
|
} else {
|
|
m.hashes[last.hash_index] = fr.entry_index;
|
|
}
|
|
}
|