This commit is contained in:
2025-05-31 16:29:42 -04:00
parent 9b19fb69c9
commit e678b15d11

View File

@@ -40,7 +40,6 @@ enum {
#define def_struct(symbol) struct symbol symbol; struct symbol #define def_struct(symbol) struct symbol symbol; struct symbol
#define def_union(symbol) union symbol symbol; union symbol #define def_union(symbol) union symbol symbol; union symbol
#define fn(symbol) symbol #define fn(symbol) symbol
#define fn_type(symbol, return_type, ...) return_type (symbol) (__VA_ARGS__)
#define opt_args(symbol, ...) &(symbol){__VA_ARGS__} #define opt_args(symbol, ...) &(symbol){__VA_ARGS__}
#define ret_type(type) type #define ret_type(type) type
#define local_persist static #define local_persist static
@@ -59,8 +58,6 @@ enum {
#define giga(n) (cast(SSIZE, n) << 30) #define giga(n) (cast(SSIZE, n) << 30)
#define tera(n) (cast(SSIZE, n) << 40) #define tera(n) (cast(SSIZE, n) << 40)
#define do_for(init, cond, post, body) do { init for (; (cond); (post)) body } while(0)
#define range_iter(type, iter, m_begin, op, m_end) \ #define range_iter(type, iter, m_begin, op, m_end) \
tmpl(Iter_Range,type) iter = { \ tmpl(Iter_Range,type) iter = { \
.r = {(m_begin), (m_end)}, \ .r = {(m_begin), (m_end)}, \
@@ -92,9 +89,9 @@ typedef def_range(SSIZE);
if (! (cond)) \ if (! (cond)) \
{ \ { \
assert_handler( \ assert_handler( \
stringify(cond), \ stringify(cond), \
__FILE__, \ __FILE__, \
__func__, \ __func__, \
cast(S64, __LINE__), \ cast(S64, __LINE__), \
msg, \ msg, \
## __VA_ARGS__); \ ## __VA_ARGS__); \
@@ -530,9 +527,9 @@ Str8Cache str8cache__make( Opts_str8cache_init* opts);
#define str8cache_init(cache, ...) str8cache__init(cache, opt_args(Opts_str8cache_init)) #define str8cache_init(cache, ...) str8cache__init(cache, opt_args(Opts_str8cache_init))
#define str8cache_make(...) str8cache__make( opt_args(Opts_str8cache_init)) #define str8cache_make(...) str8cache__make( opt_args(Opts_str8cache_init))
void str8cache_clear(KT1CX_Str8 kt); void str8cache_clear(KT1CX_Str8 kt);
Str8* str8cache_get (KT1CX_Str8 kt, U64 key); Str8 str8cache_get (KT1CX_Str8 kt, U64 key);
Str8* str8cache_set (KT1CX_Str8* kt, U64 key, Str8 value, AllocatorInfo str_reserve, AllocatorInfo backing_cells); Str8 str8cache_set (KT1CX_Str8* kt, U64 key, Str8 value, AllocatorInfo str_reserve, AllocatorInfo backing_cells);
Str8 cache_str8(Str8Cache* cache, Str8 str); Str8 cache_str8(Str8Cache* cache, Str8 str);
@@ -655,6 +652,11 @@ Str8 watl_dump_listing(AllocatorInfo buffer, Slice_WATL_Line lines);
#pragma region Implementation #pragma region Implementation
#pragma region Memory Operations #pragma region Memory Operations
// #include <memory.h>
void* __cdecl memcpy (void* _Dst, void const* _Src, USIZE _Size);
void* __cdecl memmove(void* _Dst, void const* _Src, USIZE _Size);
void* __cdecl memset (void* _Dst, int _Val, USIZE _Size);
inline inline
void* memory_copy(void* restrict dest, void const* restrict src, USIZE length) { void* memory_copy(void* restrict dest, void const* restrict src, USIZE length) {
if (dest == nullptr || src == nullptr || length == 0) { return nullptr; } if (dest == nullptr || src == nullptr || length == 0) { return nullptr; }
@@ -829,10 +831,8 @@ void farena_allocator_proc(AllocatorProc_In in, AllocatorProc_Out* out)
{ {
case AllocatorOp_Alloc: case AllocatorOp_Alloc:
case AllocatorOp_Alloc_NoZero: case AllocatorOp_Alloc_NoZero:
out->continuity_break = false; out->allocation = farena__push(arena, in.requested_size, 1, &(Opts_farena){.type_name = lit("Byte"), .alignment = in.alignment});
out->allocation = farena__push(arena, in.requested_size, 1, &(Opts_farena){.type_name = lit("Byte"), .alignment = in.alignment}); memory_zero(out->allocation.ptr, out->allocation.len * cast(SSIZE, in.op));
out->left = arena->used;
memory_zero(out->allocation.ptr, out->allocation.len * in.op);
break; break;
case AllocatorOp_Free: case AllocatorOp_Free:
@@ -997,14 +997,16 @@ Slice_Byte varena__push(VArena* vm, SSIZE amount, SSIZE type_width, Opts_varena*
SSIZE aligned_size = align_pow2(requested_size, alignment); SSIZE aligned_size = align_pow2(requested_size, alignment);
SSIZE current_offset = vm->reserve_start + vm->commit_used; SSIZE current_offset = vm->reserve_start + vm->commit_used;
SSIZE to_be_used = vm->commit_used + aligned_size; SSIZE to_be_used = vm->commit_used + aligned_size;
SSIZE reserve_left = vm->reserve - vm->committed; SSIZE reserve_left = vm->reserve - current_offset;
assert(to_be_used < reserve_left);
SSIZE header_offset = vm->reserve_start - cast(SSIZE, vm); SSIZE header_offset = vm->reserve_start - cast(SSIZE, vm);
SSIZE commit_left = vm->committed - vm->commit_used - header_offset; SSIZE commit_left = vm->committed - vm->commit_used - header_offset;
B32 exhausted = commit_left < to_be_used; B32 exhausted = commit_left < to_be_used;
assert(to_be_used < reserve_left);
if (exhausted) if (exhausted)
{ {
SSIZE next_commit_size = reserve_left > 0 ? max(vm->commit_size, to_be_used) : cast(SSIZE, align_pow2( abs(reserve_left), os_system_info()->target_page_size)); SSIZE next_commit_size = reserve_left > 0 ?
max(vm->commit_size, to_be_used)
: cast(SSIZE, align_pow2( reserve_left, os_system_info()->target_page_size));
if (next_commit_size) { if (next_commit_size) {
Byte* next_commit_start = cast(Byte*, cast(SSIZE, vm) + vm->committed); Byte* next_commit_start = cast(Byte*, cast(SSIZE, vm) + vm->committed);
B32 no_large_pages = (vm->flags & VArenaFlag_NoLargePages) != 0; B32 no_large_pages = (vm->flags & VArenaFlag_NoLargePages) != 0;
@@ -1015,6 +1017,7 @@ Slice_Byte varena__push(VArena* vm, SSIZE amount, SSIZE type_width, Opts_varena*
vm->committed += next_commit_size; vm->committed += next_commit_size;
} }
} }
vm->commit_used += to_be_used;
return (Slice_Byte){.ptr = cast(Byte*, current_offset), .len = requested_size}; return (Slice_Byte){.ptr = cast(Byte*, current_offset), .len = requested_size};
} }
inline void varena_release(VArena* arena) { os_vmem_release(arena, arena->reserve); } inline void varena_release(VArena* arena) { os_vmem_release(arena, arena->reserve); }
@@ -1032,8 +1035,8 @@ void varena_allocator_proc(AllocatorProc_In in, AllocatorProc_Out* out)
{ {
case AllocatorOp_Alloc: case AllocatorOp_Alloc:
case AllocatorOp_Alloc_NoZero: case AllocatorOp_Alloc_NoZero:
out->allocation = varena_push_array(vm, Byte, in.requested_size); out->allocation = varena_push_array(vm, Byte, in.requested_size, .alignment = in.alignment);
memory_zero(out->allocation.ptr, out->allocation.len * in.op); memory_zero(out->allocation.ptr, out->allocation.len * cast(SSIZE, in.op));
break; break;
case AllocatorOp_Free: case AllocatorOp_Free:
@@ -1106,7 +1109,8 @@ Arena* arena__make(Opts_arena_make* opts) {
return arena; return arena;
} }
Slice_Byte arena__push(Arena* arena, SSIZE amount, SSIZE type_width, Opts_arena* opts) { Slice_Byte arena__push(Arena* arena, SSIZE amount, SSIZE type_width, Opts_arena* opts) {
assert(opts != nullptr); assert(arena != nullptr);
assert(opts != nullptr);
Arena* active = arena->current; Arena* active = arena->current;
SSIZE size_requested = amount * type_width; SSIZE size_requested = amount * type_width;
SSIZE alignment = opts->alignment ? opts->alignment : MEMORY_ALIGNMENT_DEFAULT; SSIZE alignment = opts->alignment ? opts->alignment : MEMORY_ALIGNMENT_DEFAULT;
@@ -1128,15 +1132,15 @@ Slice_Byte arena__push(Arena* arena, SSIZE amount, SSIZE type_width, Opts_arena*
sll_stack_push_n(arena->current, new_arena, prev); sll_stack_push_n(arena->current, new_arena, prev);
active = arena->current; active = arena->current;
} }
Byte* result = cast(Byte*, active->pos + pos_pre);
Slice_Byte vresult = varena_push_array(active->backing, Byte, size_aligned, .alignment = alignment); Slice_Byte vresult = varena_push_array(active->backing, Byte, size_aligned, .alignment = alignment);
active->pos = pos_pst;
assert(result == vresult.ptr);
slice_assert(vresult); slice_assert(vresult);
assert(cast(Byte*, pos_pst) == vresult.ptr);
active->pos = pos_pst;
return vresult; return vresult;
} }
inline inline
void arena_release(Arena* arena) { void arena_release(Arena* arena) {
assert(arena != nullptr);
Arena* curr = arena->current; Arena* curr = arena->current;
Arena* prev = nullptr; Arena* prev = nullptr;
for (; curr != nullptr; curr = prev) { for (; curr != nullptr; curr = prev) {
@@ -1146,7 +1150,7 @@ void arena_release(Arena* arena) {
} }
inline void arena_reset(Arena* arena) { arena_rewind(arena, (AllocatorSP){.type_sig = arena_allocator_proc, .slot = 0}); } inline void arena_reset(Arena* arena) { arena_rewind(arena, (AllocatorSP){.type_sig = arena_allocator_proc, .slot = 0}); }
void arena_rewind(Arena* arena, AllocatorSP save_point) { void arena_rewind(Arena* arena, AllocatorSP save_point) {
assert(arena); assert(arena != nullptr);
assert(save_point.type_sig == arena_allocator_proc); assert(save_point.type_sig == arena_allocator_proc);
SSIZE header_size = align_pow2(size_of(Arena), MEMORY_ALIGNMENT_DEFAULT); SSIZE header_size = align_pow2(size_of(Arena), MEMORY_ALIGNMENT_DEFAULT);
Arena* curr = arena->current; Arena* curr = arena->current;
@@ -1165,13 +1169,15 @@ void arena_rewind(Arena* arena, AllocatorSP save_point) {
inline AllocatorSP arena_save(Arena* arena) { return (AllocatorSP){arena_allocator_proc, arena->current->pos}; }; inline AllocatorSP arena_save(Arena* arena) { return (AllocatorSP){arena_allocator_proc, arena->current->pos}; };
void arena_allocator_proc(AllocatorProc_In in, AllocatorProc_Out* out) void arena_allocator_proc(AllocatorProc_In in, AllocatorProc_Out* out)
{ {
assert(out != nullptr);
Arena* arena = cast(Arena*, in.data); Arena* arena = cast(Arena*, in.data);
assert(arena != nullptr);
switch (in.op) switch (in.op)
{ {
case AllocatorOp_Alloc: case AllocatorOp_Alloc:
case AllocatorOp_Alloc_NoZero: case AllocatorOp_Alloc_NoZero:
out->allocation = arena_push_array(arena, Byte, in.requested_size); out->allocation = arena_push_array(arena, Byte, in.requested_size, .alignment = in.alignment);
memory_zero(out->allocation.ptr, out->allocation.len * in.op); memory_zero(out->allocation.ptr, out->allocation.len * cast(SSIZE, in.op));
break; break;
case AllocatorOp_Free: case AllocatorOp_Free:
break; break;
@@ -1284,7 +1290,8 @@ Slice_Byte kt1cx__get(KT1CX_Byte kt, U64 key, KT1CX_ByteMeta m) {
process_slots: process_slots:
KT1CX_Byte_Slot* slot = cast(KT1CX_Byte_Slot*, slot_cursor + m.slot_key_offset); KT1CX_Byte_Slot* slot = cast(KT1CX_Byte_Slot*, slot_cursor + m.slot_key_offset);
if (slot->occupied && slot->key == key) { if (slot->occupied && slot->key == key) {
return (Slice_Byte){slot_cursor, m.type_width}; Slice_Byte slot_value = {slot_cursor, m.type_width};
return slot_value;
} }
} }
Byte* next = slot_cursor + m.cell_next_offset; Byte* next = slot_cursor + m.cell_next_offset;
@@ -1294,7 +1301,7 @@ Slice_Byte kt1cx__get(KT1CX_Byte kt, U64 key, KT1CX_ByteMeta m) {
goto process_slots; goto process_slots;
} }
else { else {
return (Slice_Byte){}; return (Slice_Byte){0};
} }
} }
} }
@@ -1329,12 +1336,12 @@ Slice_Byte kt1cx__set(KT1CX_Byte* kt, U64 key, Slice_Byte value, AllocatorInfo b
KT1CX_Byte_Slot* slot = cast(KT1CX_Byte_Slot*, cell.ptr + m.slot_key_offset); KT1CX_Byte_Slot* slot = cast(KT1CX_Byte_Slot*, cell.ptr + m.slot_key_offset);
slot->occupied = true; slot->occupied = true;
slot->key = key; slot->key = key;
Slice_Byte slot_value = {slot_cursor, m.type_width}; Slice_Byte slot_value = {cell.ptr, m.type_width};
return slot_value; return slot_value;
} }
} }
assert_msg(false, "impossible path"); assert_msg(false, "impossible path");
return (Slice_Byte){}; return (Slice_Byte){0};
} }
#pragma endregion Key Table #pragma endregion Key Table
@@ -1350,15 +1357,13 @@ Str8 str8_from_u32(AllocatorInfo ainfo, U32 num, U32 radix, U8 min_digits, U8 di
{ {
Str8 result = {0}; Str8 result = {0};
Str8 prefix = {0}; Str8 prefix = {0};
switch (radix) switch (radix) {
{
case 16: { prefix = lit("0x"); } break; case 16: { prefix = lit("0x"); } break;
case 8: { prefix = lit("0o"); } break; case 8: { prefix = lit("0o"); } break;
case 2: { prefix = lit("0b"); } break; case 2: { prefix = lit("0b"); } break;
} }
U8 digit_group_size = 3; U8 digit_group_size = 3;
switch (radix) switch (radix) {
{
default: break; default: break;
case 2: case 2:
case 8: case 8:
@@ -1381,17 +1386,15 @@ Str8 str8_from_u32(AllocatorInfo ainfo, U32 num, U32 radix, U8 min_digits, U8 di
needed_digits += 1; needed_digits += 1;
} }
} }
needed_leading_zeros = (min_digits > needed_digits) ? min_digits - needed_digits : 0; needed_leading_zeros = (min_digits > needed_digits) ? min_digits - needed_digits : 0;
U32 needed_separators = 0; U32 needed_separators = 0;
if (digit_group_separator != 0) if (digit_group_separator != 0)
{ {
needed_separators = (needed_digits + needed_leading_zeros) / digit_group_size; needed_separators = (needed_digits + needed_leading_zeros) / digit_group_size;
if (needed_separators > 0 && (needed_digits + needed_leading_zeros) % digit_group_size == 0) { if (needed_separators > 0 && (needed_digits + needed_leading_zeros) % digit_group_size == 0) {
needed_separators -= 1; needed_separators -= 1;
} }
} }
result = alloc_slice(ainfo, UTF8, prefix.len + needed_leading_zeros + needed_separators + needed_digits); result = alloc_slice(ainfo, UTF8, prefix.len + needed_leading_zeros + needed_separators + needed_digits);
slice_assert(result); slice_assert(result);
} }
@@ -1431,12 +1434,10 @@ Str8 str8__fmt_kt1l(AllocatorInfo ainfo, Slice_Byte buffer, KT1L_Str8 table, Str
slice_assert(table); slice_assert(table);
slice_assert(fmt_template); slice_assert(fmt_template);
assert(ainfo.proc != nullptr ? (allocator_query(ainfo) & AllocatorQuery_Grow) != 0 : true); assert(ainfo.proc != nullptr ? (allocator_query(ainfo) & AllocatorQuery_Grow) != 0 : true);
UTF8* cursor_buffer = buffer.ptr; UTF8* cursor_buffer = buffer.ptr;
SSIZE buffer_remaining = buffer.len; SSIZE buffer_remaining = buffer.len;
char curr_code = * fmt_template.ptr; char curr_code = * fmt_template.ptr;
UTF8* cursor_fmt = fmt_template.ptr; UTF8* cursor_fmt = fmt_template.ptr;
SSIZE left_fmt = fmt_template.len; SSIZE left_fmt = fmt_template.len;
while (left_fmt && buffer_remaining) while (left_fmt && buffer_remaining)
@@ -1557,7 +1558,7 @@ void str8cache_clear(KT1CX_Str8 kt) {
}); });
} }
inline inline
Str8* str8cache_get(KT1CX_Str8 kt, U64 key) { Str8 str8cache_get(KT1CX_Str8 kt, U64 key) {
Slice_Byte result = kt1cx__get(pcast(KT1CX_Byte, kt), key Slice_Byte result = kt1cx__get(pcast(KT1CX_Byte, kt), key
, (KT1CX_ByteMeta){ , (KT1CX_ByteMeta){
.slot_size = size_of(KT1CX_Slot_Str8), .slot_size = size_of(KT1CX_Slot_Str8),
@@ -1568,10 +1569,14 @@ Str8* str8cache_get(KT1CX_Str8 kt, U64 key) {
.type_width = size_of(Str8), .type_width = size_of(Str8),
.type_name = lit(stringify(Str8)) .type_name = lit(stringify(Str8))
}); });
return cast(Str8*, result.ptr); return pcast(Str8, result);
} }
inline inline
Str8* str8cache_set(KT1CX_Str8* kt, U64 key, Str8 value, AllocatorInfo str_reserve, AllocatorInfo backing_cells) { Str8 str8cache_set(KT1CX_Str8* kt, U64 key, Str8 value, AllocatorInfo str_reserve, AllocatorInfo backing_cells) {
assert(kt != nullptr);
assert(str_reserve.proc != nullptr);
assert(backing_cells.proc != nullptr);
slice_assert(value);
Slice_Byte entry = kt1cx__set(pcast(KT1CX_Byte*, kt), key, slice_byte(value), backing_cells, (KT1CX_ByteMeta){ Slice_Byte entry = kt1cx__set(pcast(KT1CX_Byte*, kt), key, slice_byte(value), backing_cells, (KT1CX_ByteMeta){
.slot_size = size_of(KT1CX_Slot_Str8), .slot_size = size_of(KT1CX_Slot_Str8),
.slot_key_offset = offset_of(KT1CX_Slot_Str8, key), .slot_key_offset = offset_of(KT1CX_Slot_Str8, key),
@@ -1581,19 +1586,20 @@ Str8* str8cache_set(KT1CX_Str8* kt, U64 key, Str8 value, AllocatorInfo str_reser
.type_width = size_of(Str8), .type_width = size_of(Str8),
.type_name = lit(stringify(Str8)) .type_name = lit(stringify(Str8))
}); });
Str8* result = pcast(Str8*, entry.ptr); slice_assert(entry);
if (result->ptr == nullptr && result->len == 0) { Str8 result = { entry.ptr, 0 };
* result = alloc_slice(str_reserve, UTF8, value.len); B32 is_empty = (result.ptr == nullptr) && (result.len == 0);
if (is_empty) {
result = alloc_slice(str_reserve, UTF8, value.len);
slice_copy(result, value);
} }
slice_assert(* result);
return result; return result;
} }
inline inline
Str8 cache_str8(Str8Cache* cache, Str8 str) { Str8 cache_str8(Str8Cache* cache, Str8 str) {
U64 key = 0; hash64_djb8(& key, slice_byte(str)); U64 key = 0; hash64_djb8(& key, slice_byte(str));
Str8* result = str8cache_set(& cache->kt, key, str, cache->str_reserve, cache->cell_reserve); Str8 result = str8cache_set(& cache->kt, key, str, cache->str_reserve, cache->cell_reserve);
assert(result != nullptr); return result;
return * result;
} }
inline inline