mirror of
https://github.com/odin-lang/Odin.git
synced 2026-05-04 20:14:44 +00:00
ZLIB: Remove superfluous code.
This commit is contained in:
@@ -10,6 +10,7 @@ package compress
|
||||
|
||||
import "core:io"
|
||||
import "core:image"
|
||||
import "core:bytes"
|
||||
|
||||
/*
|
||||
These settings bound how much compression algorithms will allocate for their output buffer.
|
||||
@@ -44,8 +45,6 @@ when size_of(uintptr) == 8 {
|
||||
}
|
||||
|
||||
|
||||
// when #config(TRACY_ENABLE, false) { import tracy "shared:odin-tracy" }
|
||||
|
||||
Error :: union {
|
||||
General_Error,
|
||||
Deflate_Error,
|
||||
@@ -132,8 +131,7 @@ Context :: struct #packed {
|
||||
input: io.Stream,
|
||||
input_data: []u8,
|
||||
|
||||
output: io.Stream,
|
||||
output_buf: ^[dynamic]u8,
|
||||
output: ^bytes.Buffer,
|
||||
bytes_written: i64,
|
||||
|
||||
/*
|
||||
@@ -159,6 +157,10 @@ Context :: struct #packed {
|
||||
input_refills_from_stream: b8,
|
||||
output_to_stream: b8,
|
||||
reserved_flag: b8,
|
||||
|
||||
bit_buffer_stuff: [3]u64,
|
||||
|
||||
|
||||
}
|
||||
// #assert(size_of(Context) == 128);
|
||||
|
||||
@@ -188,8 +190,6 @@ Code_Buffer :: struct #packed {
|
||||
|
||||
@(optimization_mode="speed")
|
||||
read_slice :: #force_inline proc(z: ^Context, size: int) -> (res: []u8, err: io.Error) {
|
||||
when #config(TRACY_ENABLE, false) { tracy.ZoneN("Read Slice"); }
|
||||
|
||||
#no_bounds_check {
|
||||
if len(z.input_data) >= size {
|
||||
res = z.input_data[:size];
|
||||
@@ -220,8 +220,6 @@ read_slice :: #force_inline proc(z: ^Context, size: int) -> (res: []u8, err: io.
|
||||
|
||||
@(optimization_mode="speed")
|
||||
read_data :: #force_inline proc(z: ^Context, $T: typeid) -> (res: T, err: io.Error) {
|
||||
when #config(TRACY_ENABLE, false) { tracy.ZoneN("Read Data"); }
|
||||
|
||||
b, e := read_slice(z, size_of(T));
|
||||
if e == .None {
|
||||
return (^T)(&b[0])^, .None;
|
||||
@@ -232,8 +230,6 @@ read_data :: #force_inline proc(z: ^Context, $T: typeid) -> (res: T, err: io.Err
|
||||
|
||||
@(optimization_mode="speed")
|
||||
read_u8 :: #force_inline proc(z: ^Context) -> (res: u8, err: io.Error) {
|
||||
when #config(TRACY_ENABLE, false) { tracy.ZoneN("Read u8"); }
|
||||
|
||||
#no_bounds_check {
|
||||
if len(z.input_data) >= 1 {
|
||||
res = z.input_data[0];
|
||||
@@ -252,8 +248,6 @@ read_u8 :: #force_inline proc(z: ^Context) -> (res: u8, err: io.Error) {
|
||||
|
||||
@(optimization_mode="speed")
|
||||
peek_data :: #force_inline proc(z: ^Context, $T: typeid) -> (res: T, err: io.Error) {
|
||||
when #config(TRACY_ENABLE, false) { tracy.ZoneN("Peek Data"); }
|
||||
|
||||
size :: size_of(T);
|
||||
|
||||
#no_bounds_check {
|
||||
@@ -304,8 +298,6 @@ peek_back_byte :: #force_inline proc(cb: ^Code_Buffer, offset: i64) -> (res: u8,
|
||||
// Generalized bit reader LSB
|
||||
@(optimization_mode="speed")
|
||||
refill_lsb :: proc(z: ^Context, cb: ^Code_Buffer, width := i8(24)) {
|
||||
when #config(TRACY_ENABLE, false) { tracy.ZoneN("Refill LSB"); }
|
||||
|
||||
refill := u64(width);
|
||||
|
||||
for {
|
||||
|
||||
@@ -110,10 +110,11 @@ load_from_slice :: proc(slice: []u8, buf: ^bytes.Buffer, known_gzip_size := -1,
|
||||
stream := bytes.reader_to_stream(&r);
|
||||
|
||||
ctx := &compress.Context{
|
||||
input = stream,
|
||||
input = stream,
|
||||
input_data = slice,
|
||||
input_fully_in_memory = true,
|
||||
input_refills_from_stream = true,
|
||||
output = buf,
|
||||
};
|
||||
|
||||
err = load_from_stream(ctx, buf, known_gzip_size, expected_output_size, allocator);
|
||||
@@ -138,8 +139,7 @@ load_from_stream :: proc(ctx: ^compress.Context, buf: ^bytes.Buffer, known_gzip_
|
||||
|
||||
input_data_consumed := 0;
|
||||
|
||||
ws := bytes.buffer_to_stream(buf);
|
||||
ctx.output = ws;
|
||||
ctx.output = buf;
|
||||
|
||||
if expected_output_size > GZIP_MAX_PAYLOAD_SIZE {
|
||||
return E_GZIP.Payload_Size_Exceeds_Max_Payload;
|
||||
@@ -365,6 +365,9 @@ load_from_stream :: proc(ctx: ^compress.Context, buf: ^bytes.Buffer, known_gzip_
|
||||
payload_u32le, footer_error = compress.read_data(ctx, u32le);
|
||||
|
||||
payload := bytes.buffer_to_bytes(buf);
|
||||
|
||||
// fmt.printf("GZIP payload: %v\n", string(payload));
|
||||
|
||||
crc32 := u32le(hash.crc32(payload));
|
||||
|
||||
if crc32 != payload_crc {
|
||||
|
||||
@@ -13,6 +13,7 @@ package zlib
|
||||
|
||||
import "core:bytes"
|
||||
import "core:fmt"
|
||||
import "core:compress"
|
||||
|
||||
main :: proc() {
|
||||
|
||||
@@ -37,6 +38,7 @@ main :: proc() {
|
||||
};
|
||||
OUTPUT_SIZE :: 438;
|
||||
|
||||
fmt.printf("size_of(Context): %v\n", size_of(compress.Context));
|
||||
|
||||
buf: bytes.Buffer;
|
||||
|
||||
|
||||
@@ -13,13 +13,10 @@ import "core:compress"
|
||||
|
||||
import "core:mem"
|
||||
import "core:io"
|
||||
import "core:bytes"
|
||||
import "core:hash"
|
||||
|
||||
import "core:bytes"
|
||||
// import "core:fmt"
|
||||
|
||||
// when #config(TRACY_ENABLE, false) { import tracy "shared:odin-tracy" }
|
||||
|
||||
/*
|
||||
zlib.inflate decompresses a ZLIB stream passed in as a []u8 or io.Stream.
|
||||
Returns: Error.
|
||||
@@ -33,8 +30,6 @@ import "core:hash"
|
||||
`Context.rolling_hash` if not inlining it is still faster.
|
||||
|
||||
*/
|
||||
INLINE_ADLER :: false;
|
||||
|
||||
Context :: compress.Context;
|
||||
Code_Buffer :: compress.Code_Buffer;
|
||||
|
||||
@@ -172,89 +167,48 @@ grow_buffer :: proc(buf: ^[dynamic]u8) -> (err: compress.Error) {
|
||||
|
||||
@(optimization_mode="speed")
|
||||
write_byte :: #force_inline proc(z: ^Context, cb: ^Code_Buffer, c: u8) -> (err: io.Error) #no_bounds_check {
|
||||
when #config(TRACY_ENABLE, false) { tracy.ZoneN("Write Byte"); }
|
||||
c := c;
|
||||
|
||||
if !z.output_to_stream {
|
||||
/*
|
||||
Resize if needed.
|
||||
*/
|
||||
if int(z.bytes_written) + 1 >= len(z.output_buf) {
|
||||
grow_buffer(z.output_buf);
|
||||
/*
|
||||
Resize if needed.
|
||||
*/
|
||||
if int(z.bytes_written) + 1 >= len(z.output.buf) {
|
||||
e := grow_buffer(&z.output.buf);
|
||||
if e != nil {
|
||||
return .Short_Write;
|
||||
}
|
||||
|
||||
#no_bounds_check {
|
||||
z.output_buf[z.bytes_written] = c;
|
||||
cb.last[z.bytes_written & cb.window_mask] = c;
|
||||
}
|
||||
z.bytes_written += 1;
|
||||
|
||||
when INLINE_ADLER {
|
||||
buf := transmute([]u8)mem.Raw_Slice{data=&c, len=1};
|
||||
z.rolling_hash = hash.adler32(buf, z.rolling_hash);
|
||||
}
|
||||
} else {
|
||||
buf := transmute([]u8)mem.Raw_Slice{data=&c, len=1};
|
||||
when INLINE_ADLER {
|
||||
z.rolling_hash = hash.adler32(buf, z.rolling_hash);
|
||||
}
|
||||
|
||||
_, e := z.output->impl_write(buf);
|
||||
if e != .None {
|
||||
return e;
|
||||
}
|
||||
|
||||
#no_bounds_check cb.last[z.bytes_written & cb.window_mask] = c;
|
||||
z.bytes_written += 1;
|
||||
}
|
||||
|
||||
#no_bounds_check {
|
||||
z.output.buf[z.bytes_written] = c;
|
||||
cb.last[z.bytes_written & cb.window_mask] = c;
|
||||
}
|
||||
z.bytes_written += 1;
|
||||
return .None;
|
||||
}
|
||||
|
||||
@(optimization_mode="speed")
|
||||
repl_byte :: proc(z: ^Context, cb: ^Code_Buffer, count: u16, c: u8) -> (err: io.Error) #no_bounds_check {
|
||||
when #config(TRACY_ENABLE, false) { tracy.ZoneN("Repl Byte"); }
|
||||
/*
|
||||
TODO(Jeroen): Once we have a magic ring buffer, we can just peek/write into it
|
||||
without having to worry about wrapping, so no need for a temp allocation to give to
|
||||
the output stream, just give it _that_ slice.
|
||||
*/
|
||||
|
||||
if !z.output_to_stream {
|
||||
/*
|
||||
Resize if needed.
|
||||
*/
|
||||
if int(z.bytes_written) + int(count) >= len(z.output_buf) {
|
||||
grow_buffer(z.output_buf);
|
||||
/*
|
||||
Resize if needed.
|
||||
*/
|
||||
if int(z.bytes_written) + int(count) >= len(z.output.buf) {
|
||||
e := grow_buffer(&z.output.buf);
|
||||
if e != nil {
|
||||
return .Short_Write;
|
||||
}
|
||||
}
|
||||
|
||||
#no_bounds_check {
|
||||
for _ in 0..<count {
|
||||
z.output_buf[z.bytes_written] = c;
|
||||
cb.last[z.bytes_written & cb.window_mask] = c;
|
||||
z.bytes_written += 1;
|
||||
}
|
||||
}
|
||||
|
||||
when INLINE_ADLER {
|
||||
/* TODO(Jeroen): Test */
|
||||
buf := z.output_buf[bytes_written:][:count];
|
||||
z.rolling_hash = hash.adler32(buf, z.rolling_hash);
|
||||
}
|
||||
|
||||
} else {
|
||||
buf := make([]u8, count, context.temp_allocator);
|
||||
#no_bounds_check for i in 0..<count {
|
||||
buf[i] = c;
|
||||
#no_bounds_check {
|
||||
for _ in 0..<count {
|
||||
z.output.buf[z.bytes_written] = c;
|
||||
cb.last[z.bytes_written & cb.window_mask] = c;
|
||||
z.bytes_written += 1;
|
||||
}
|
||||
when INLINE_ADLER { z.rolling_hash = hash.adler32(buf, z.rolling_hash); }
|
||||
|
||||
_, e := z.output->impl_write(buf);
|
||||
if e != .None {
|
||||
return e;
|
||||
}
|
||||
}
|
||||
|
||||
return .None;
|
||||
@@ -262,7 +216,6 @@ repl_byte :: proc(z: ^Context, cb: ^Code_Buffer, count: u16, c: u8) -> (err: io.
|
||||
|
||||
@(optimization_mode="speed")
|
||||
repl_bytes :: proc(z: ^Context, cb: ^Code_Buffer, count: u16, distance: u16) -> (err: io.Error) {
|
||||
when #config(TRACY_ENABLE, false) { tracy.ZoneN("Repl Bytes"); }
|
||||
/*
|
||||
TODO(Jeroen): Once we have a magic ring buffer, we can just peek/write into it
|
||||
without having to worry about wrapping, so no need for a temp allocation to give to
|
||||
@@ -271,40 +224,20 @@ repl_bytes :: proc(z: ^Context, cb: ^Code_Buffer, count: u16, distance: u16) ->
|
||||
|
||||
offset := z.bytes_written - i64(distance);
|
||||
|
||||
if !z.output_to_stream {
|
||||
if int(z.bytes_written) + int(count) >= len(z.output_buf) {
|
||||
grow_buffer(z.output_buf);
|
||||
if int(z.bytes_written) + int(count) >= len(z.output.buf) {
|
||||
e := grow_buffer(&z.output.buf);
|
||||
if e != nil {
|
||||
return .Short_Write;
|
||||
}
|
||||
}
|
||||
|
||||
#no_bounds_check {
|
||||
for _ in 0..<count {
|
||||
c := cb.last[offset & cb.window_mask];
|
||||
z.output_buf[z.bytes_written] = c;
|
||||
cb.last[z.bytes_written & cb.window_mask] = c;
|
||||
z.bytes_written += 1; offset += 1;
|
||||
}
|
||||
}
|
||||
|
||||
when INLINE_ADLER {
|
||||
/* TODO(Jeroen): Test */
|
||||
buf := z.output_buf[bytes_written:][:count];
|
||||
z.rolling_hash = hash.adler32(buf, z.rolling_hash);
|
||||
}
|
||||
} else {
|
||||
buf := make([]u8, count, context.temp_allocator);
|
||||
#no_bounds_check for i in 0..<count {
|
||||
#no_bounds_check {
|
||||
for _ in 0..<count {
|
||||
c := cb.last[offset & cb.window_mask];
|
||||
|
||||
z.output.buf[z.bytes_written] = c;
|
||||
cb.last[z.bytes_written & cb.window_mask] = c;
|
||||
buf[i] = c;
|
||||
z.bytes_written += 1; offset += 1;
|
||||
}
|
||||
when INLINE_ADLER { z.rolling_hash = hash.adler32(buf, z.rolling_hash); }
|
||||
|
||||
_, e := z.output->impl_write(buf);
|
||||
if e != .None {
|
||||
return e;
|
||||
}
|
||||
}
|
||||
|
||||
return .None;
|
||||
@@ -317,7 +250,6 @@ allocate_huffman_table :: proc(allocator := context.allocator) -> (z: ^Huffman_T
|
||||
|
||||
@(optimization_mode="speed")
|
||||
build_huffman :: proc(z: ^Huffman_Table, code_lengths: []u8) -> (err: Error) {
|
||||
when #config(TRACY_ENABLE, false) { tracy.ZoneN("Build Huffman Table"); }
|
||||
sizes: [HUFFMAN_MAX_BITS+1]int;
|
||||
next_code: [HUFFMAN_MAX_BITS]int;
|
||||
|
||||
@@ -377,7 +309,6 @@ build_huffman :: proc(z: ^Huffman_Table, code_lengths: []u8) -> (err: Error) {
|
||||
|
||||
@(optimization_mode="speed")
|
||||
decode_huffman_slowpath :: proc(z: ^Context, cb: ^Code_Buffer, t: ^Huffman_Table) -> (r: u16, err: Error) #no_bounds_check {
|
||||
when #config(TRACY_ENABLE, false) { tracy.ZoneN("Decode Huffman Slow"); }
|
||||
code := u16(compress.peek_bits_lsb(z, cb, 16));
|
||||
|
||||
k := int(z_bit_reverse(code, 16));
|
||||
@@ -409,7 +340,6 @@ decode_huffman_slowpath :: proc(z: ^Context, cb: ^Code_Buffer, t: ^Huffman_Table
|
||||
|
||||
@(optimization_mode="speed")
|
||||
decode_huffman :: proc(z: ^Context, cb: ^Code_Buffer, t: ^Huffman_Table) -> (r: u16, err: Error) #no_bounds_check {
|
||||
when #config(TRACY_ENABLE, false) { tracy.ZoneN("Decode Huffman"); }
|
||||
if cb.num_bits < 16 {
|
||||
if cb.num_bits > 63 {
|
||||
return 0, E_ZLIB.Code_Buffer_Malformed;
|
||||
@@ -430,7 +360,6 @@ decode_huffman :: proc(z: ^Context, cb: ^Code_Buffer, t: ^Huffman_Table) -> (r:
|
||||
|
||||
@(optimization_mode="speed")
|
||||
parse_huffman_block :: proc(z: ^Context, cb: ^Code_Buffer, z_repeat, z_offset: ^Huffman_Table) -> (err: Error) #no_bounds_check {
|
||||
when #config(TRACY_ENABLE, false) { tracy.ZoneN("Parse Huffman Block"); }
|
||||
#no_bounds_check for {
|
||||
value, e := decode_huffman(z, cb, z_repeat);
|
||||
if e != nil {
|
||||
@@ -573,10 +502,7 @@ inflate_from_stream :: proc(using ctx: ^Context, raw := false, expected_output_s
|
||||
compress.discard_to_next_byte_lsb(cb);
|
||||
adler32 := compress.read_bits_lsb(ctx, cb, 8) << 24 | compress.read_bits_lsb(ctx, cb, 8) << 16 | compress.read_bits_lsb(ctx, cb, 8) << 8 | compress.read_bits_lsb(ctx, cb, 8);
|
||||
|
||||
when !INLINE_ADLER {
|
||||
buf := (^bytes.Buffer)(ctx.output.stream_data).buf[:];
|
||||
ctx.rolling_hash = hash.adler32(buf);
|
||||
}
|
||||
ctx.rolling_hash = hash.adler32(ctx.output.buf[:]);
|
||||
|
||||
if ctx.rolling_hash != u32(adler32) {
|
||||
return E_General.Checksum_Failed;
|
||||
@@ -587,26 +513,28 @@ inflate_from_stream :: proc(using ctx: ^Context, raw := false, expected_output_s
|
||||
|
||||
@(optimization_mode="speed")
|
||||
inflate_from_stream_raw :: proc(z: ^Context, cb: ^Code_Buffer, expected_output_size := -1, allocator := context.allocator) -> (err: Error) #no_bounds_check {
|
||||
when #config(TRACY_ENABLE, false) { tracy.ZoneN("Inflate Raw"); }
|
||||
expected_output_size := expected_output_size;
|
||||
|
||||
buf := (^bytes.Buffer)(z.output.stream_data);
|
||||
z.output_buf = &buf.buf;
|
||||
|
||||
// fmt.printf("ZLIB: Expected Payload Size: %v\n", expected_output_size);
|
||||
|
||||
if expected_output_size > -1 && expected_output_size <= compress.COMPRESS_OUTPUT_ALLOCATE_MAX {
|
||||
reserve(z.output_buf, expected_output_size);
|
||||
resize (z.output_buf, expected_output_size);
|
||||
} else {
|
||||
reserve(z.output_buf, compress.COMPRESS_OUTPUT_ALLOCATE_MIN);
|
||||
if expected_output_size <= 0 {
|
||||
/*
|
||||
Always set up a minimum allocation size.
|
||||
*/
|
||||
expected_output_size = compress.COMPRESS_OUTPUT_ALLOCATE_MIN;
|
||||
}
|
||||
|
||||
// reserve(&z.output_buf, compress.COMPRESS_OUTPUT_ALLOCATE_MIN);
|
||||
// resize (&z.output_buf, compress.COMPRESS_OUTPUT_ALLOCATE_MIN);
|
||||
// fmt.printf("ZLIB: buf: %v\n", buf);
|
||||
// fmt.printf("ZLIB: output_buf: %v\n", z.output_buf);
|
||||
// fmt.printf("ZLIB: z.output: %v\n", z.output);
|
||||
// fmt.printf("\nZLIB: Expected Payload Size: %v\n\n", expected_output_size);
|
||||
|
||||
if expected_output_size > 0 && expected_output_size <= compress.COMPRESS_OUTPUT_ALLOCATE_MAX {
|
||||
/*
|
||||
Try to pre-allocate the output buffer.
|
||||
*/
|
||||
reserve(&z.output.buf, expected_output_size);
|
||||
resize (&z.output.buf, expected_output_size);
|
||||
};
|
||||
|
||||
if len(z.output.buf) != expected_output_size {
|
||||
return .Resize_Failed;
|
||||
}
|
||||
|
||||
cb.num_bits = 0;
|
||||
cb.code_buffer = 0;
|
||||
@@ -651,7 +579,6 @@ inflate_from_stream_raw :: proc(z: ^Context, cb: ^Code_Buffer, expected_output_s
|
||||
|
||||
switch type {
|
||||
case 0:
|
||||
when #config(TRACY_ENABLE, false) { tracy.ZoneN("Literal Block"); }
|
||||
// Uncompressed block
|
||||
|
||||
// Discard bits until next byte boundary
|
||||
@@ -680,7 +607,6 @@ inflate_from_stream_raw :: proc(z: ^Context, cb: ^Code_Buffer, expected_output_s
|
||||
case 3:
|
||||
return E_Deflate.BType_3;
|
||||
case:
|
||||
when #config(TRACY_ENABLE, false) { tracy.ZoneN("Huffman Block"); }
|
||||
// log.debugf("Err: %v | Final: %v | Type: %v\n", err, final, type);
|
||||
if type == 1 {
|
||||
// Use fixed code lengths.
|
||||
@@ -784,8 +710,8 @@ inflate_from_stream_raw :: proc(z: ^Context, cb: ^Code_Buffer, expected_output_s
|
||||
}
|
||||
|
||||
// fmt.printf("ZLIB: Bytes written: %v\n", z.bytes_written);
|
||||
if int(z.bytes_written) != len(buf.buf) {
|
||||
resize(&buf.buf, int(z.bytes_written));
|
||||
if int(z.bytes_written) != len(z.output.buf) {
|
||||
resize(&z.output.buf, int(z.bytes_written));
|
||||
}
|
||||
|
||||
return nil;
|
||||
@@ -801,9 +727,7 @@ inflate_from_byte_array :: proc(input: []u8, buf: ^bytes.Buffer, raw := false, e
|
||||
ctx.input_data = input;
|
||||
ctx.input_fully_in_memory = true;
|
||||
|
||||
buf := buf;
|
||||
ws := bytes.buffer_to_stream(buf);
|
||||
ctx.output = ws;
|
||||
ctx.output = buf;
|
||||
|
||||
err = inflate_from_stream(ctx=&ctx, raw=raw, expected_output_size=expected_output_size);
|
||||
|
||||
@@ -820,9 +744,7 @@ inflate_from_byte_array_raw :: proc(input: []u8, buf: ^bytes.Buffer, cb: ^Code_B
|
||||
ctx.input_data = input;
|
||||
ctx.input_fully_in_memory = true;
|
||||
|
||||
buf := buf;
|
||||
ws := bytes.buffer_to_stream(buf);
|
||||
ctx.output = ws;
|
||||
ctx.output = buf;
|
||||
|
||||
return inflate_from_stream_raw(z=&ctx, cb=cb, expected_output_size=expected_output_size);
|
||||
}
|
||||
|
||||
Reference in New Issue
Block a user