mirror of
https://codeberg.org/ziglang/zig.git
synced 2025-12-06 05:44:20 +00:00
lzma2 Decoder already checks if decoding is finished or not inside the process function, `range_decoder`finish does not mean the decoder has finished, also need to check `ld.rep[0] == 0xFFFF_FFFF`, which was already done inside the proccess function. This fix delete the redundant `isFinish()` check for `range_decoder`.
282 lines
8 KiB
Zig
282 lines
8 KiB
Zig
const std = @import("../std.zig");
|
|
const Allocator = std.mem.Allocator;
|
|
const ArrayList = std.ArrayList;
|
|
const lzma = std.compress.lzma;
|
|
const Writer = std.Io.Writer;
|
|
const Reader = std.Io.Reader;
|
|
|
|
/// An accumulating buffer for LZ sequences
|
|
pub const AccumBuffer = struct {
|
|
/// Buffer
|
|
buf: ArrayList(u8),
|
|
/// Buffer memory limit
|
|
memlimit: usize,
|
|
/// Total number of bytes sent through the buffer
|
|
len: usize,
|
|
|
|
pub fn init(memlimit: usize) AccumBuffer {
|
|
return .{
|
|
.buf = .{},
|
|
.memlimit = memlimit,
|
|
.len = 0,
|
|
};
|
|
}
|
|
|
|
pub fn appendByte(self: *AccumBuffer, allocator: Allocator, byte: u8) !void {
|
|
try self.buf.append(allocator, byte);
|
|
self.len += 1;
|
|
}
|
|
|
|
/// Reset the internal dictionary
|
|
pub fn reset(self: *AccumBuffer, writer: *Writer) !void {
|
|
try writer.writeAll(self.buf.items);
|
|
self.buf.clearRetainingCapacity();
|
|
self.len = 0;
|
|
}
|
|
|
|
/// Retrieve the last byte or return a default
|
|
pub fn lastOr(self: AccumBuffer, lit: u8) u8 {
|
|
const buf_len = self.buf.items.len;
|
|
return if (buf_len == 0)
|
|
lit
|
|
else
|
|
self.buf.items[buf_len - 1];
|
|
}
|
|
|
|
/// Retrieve the n-th last byte
|
|
pub fn lastN(self: AccumBuffer, dist: usize) !u8 {
|
|
const buf_len = self.buf.items.len;
|
|
if (dist > buf_len) {
|
|
return error.CorruptInput;
|
|
}
|
|
|
|
return self.buf.items[buf_len - dist];
|
|
}
|
|
|
|
/// Append a literal
|
|
pub fn appendLiteral(
|
|
self: *AccumBuffer,
|
|
allocator: Allocator,
|
|
lit: u8,
|
|
writer: *Writer,
|
|
) !void {
|
|
_ = writer;
|
|
if (self.len >= self.memlimit) {
|
|
return error.CorruptInput;
|
|
}
|
|
try self.buf.append(allocator, lit);
|
|
self.len += 1;
|
|
}
|
|
|
|
/// Fetch an LZ sequence (length, distance) from inside the buffer
|
|
pub fn appendLz(
|
|
self: *AccumBuffer,
|
|
allocator: Allocator,
|
|
len: usize,
|
|
dist: usize,
|
|
writer: *Writer,
|
|
) !void {
|
|
_ = writer;
|
|
|
|
const buf_len = self.buf.items.len;
|
|
if (dist > buf_len) return error.CorruptInput;
|
|
|
|
try self.buf.ensureUnusedCapacity(allocator, len);
|
|
const buffer = self.buf.allocatedSlice();
|
|
const src = buffer[buf_len - dist ..][0..len];
|
|
const dst = buffer[buf_len..][0..len];
|
|
|
|
// This is not a @memmove; it intentionally repeats patterns caused by
|
|
// iterating one byte at a time.
|
|
for (dst, src) |*d, s| d.* = s;
|
|
|
|
self.buf.items.len = buf_len + len;
|
|
self.len += len;
|
|
}
|
|
|
|
pub fn finish(self: *AccumBuffer, writer: *Writer) !void {
|
|
try writer.writeAll(self.buf.items);
|
|
self.buf.clearRetainingCapacity();
|
|
}
|
|
|
|
pub fn deinit(self: *AccumBuffer, allocator: Allocator) void {
|
|
self.buf.deinit(allocator);
|
|
self.* = undefined;
|
|
}
|
|
};
|
|
|
|
pub const Decode = struct {
|
|
lzma_decode: lzma.Decode,
|
|
|
|
pub fn init(gpa: Allocator) !Decode {
|
|
return .{ .lzma_decode = try lzma.Decode.init(gpa, .{ .lc = 0, .lp = 0, .pb = 0 }) };
|
|
}
|
|
|
|
pub fn deinit(self: *Decode, gpa: Allocator) void {
|
|
self.lzma_decode.deinit(gpa);
|
|
self.* = undefined;
|
|
}
|
|
|
|
/// Returns how many compressed bytes were consumed.
|
|
pub fn decompress(d: *Decode, reader: *Reader, allocating: *Writer.Allocating) !u64 {
|
|
const gpa = allocating.allocator;
|
|
|
|
var accum = AccumBuffer.init(std.math.maxInt(usize));
|
|
defer accum.deinit(gpa);
|
|
|
|
var n_read: u64 = 0;
|
|
|
|
while (true) {
|
|
const status = try reader.takeByte();
|
|
n_read += 1;
|
|
|
|
switch (status) {
|
|
0 => break,
|
|
1 => n_read += try parseUncompressed(reader, allocating, &accum, true),
|
|
2 => n_read += try parseUncompressed(reader, allocating, &accum, false),
|
|
else => n_read += try d.parseLzma(reader, allocating, &accum, status),
|
|
}
|
|
}
|
|
|
|
try accum.finish(&allocating.writer);
|
|
return n_read;
|
|
}
|
|
|
|
fn parseLzma(
|
|
d: *Decode,
|
|
reader: *Reader,
|
|
allocating: *Writer.Allocating,
|
|
accum: *AccumBuffer,
|
|
status: u8,
|
|
) !u64 {
|
|
if (status & 0x80 == 0) return error.CorruptInput;
|
|
|
|
const Reset = struct {
|
|
dict: bool,
|
|
state: bool,
|
|
props: bool,
|
|
};
|
|
|
|
const reset: Reset = switch ((status >> 5) & 0x3) {
|
|
0 => .{
|
|
.dict = false,
|
|
.state = false,
|
|
.props = false,
|
|
},
|
|
1 => .{
|
|
.dict = false,
|
|
.state = true,
|
|
.props = false,
|
|
},
|
|
2 => .{
|
|
.dict = false,
|
|
.state = true,
|
|
.props = true,
|
|
},
|
|
3 => .{
|
|
.dict = true,
|
|
.state = true,
|
|
.props = true,
|
|
},
|
|
else => unreachable,
|
|
};
|
|
|
|
var n_read: u64 = 0;
|
|
|
|
const unpacked_size = blk: {
|
|
var tmp: u64 = status & 0x1F;
|
|
tmp <<= 16;
|
|
tmp |= try reader.takeInt(u16, .big);
|
|
n_read += 2;
|
|
break :blk tmp + 1;
|
|
};
|
|
|
|
const packed_size = blk: {
|
|
const tmp: u17 = try reader.takeInt(u16, .big);
|
|
n_read += 2;
|
|
break :blk tmp + 1;
|
|
};
|
|
|
|
if (reset.dict) try accum.reset(&allocating.writer);
|
|
|
|
const ld = &d.lzma_decode;
|
|
|
|
if (reset.state) {
|
|
var new_props = ld.properties;
|
|
|
|
if (reset.props) {
|
|
var props = try reader.takeByte();
|
|
n_read += 1;
|
|
if (props >= 225) {
|
|
return error.CorruptInput;
|
|
}
|
|
|
|
const lc = @as(u4, @intCast(props % 9));
|
|
props /= 9;
|
|
const lp = @as(u3, @intCast(props % 5));
|
|
props /= 5;
|
|
const pb = @as(u3, @intCast(props));
|
|
|
|
if (lc + lp > 4) {
|
|
return error.CorruptInput;
|
|
}
|
|
|
|
new_props = .{ .lc = lc, .lp = lp, .pb = pb };
|
|
}
|
|
|
|
try ld.resetState(allocating.allocator, new_props);
|
|
}
|
|
|
|
const expected_unpacked_size = accum.len + unpacked_size;
|
|
const start_count = n_read;
|
|
var range_decoder = try lzma.RangeDecoder.initCounting(reader, &n_read);
|
|
|
|
while (true) {
|
|
if (accum.len >= expected_unpacked_size) break;
|
|
switch (try ld.process(reader, allocating, accum, &range_decoder, &n_read)) {
|
|
.more => continue,
|
|
.finished => break,
|
|
}
|
|
}
|
|
if (accum.len != expected_unpacked_size) return error.DecompressedSizeMismatch;
|
|
if (n_read - start_count != packed_size) return error.CompressedSizeMismatch;
|
|
|
|
return n_read;
|
|
}
|
|
|
|
fn parseUncompressed(
|
|
reader: *Reader,
|
|
allocating: *Writer.Allocating,
|
|
accum: *AccumBuffer,
|
|
reset_dict: bool,
|
|
) !usize {
|
|
const unpacked_size = @as(u17, try reader.takeInt(u16, .big)) + 1;
|
|
|
|
if (reset_dict) try accum.reset(&allocating.writer);
|
|
|
|
const gpa = allocating.allocator;
|
|
|
|
for (0..unpacked_size) |_| {
|
|
try accum.appendByte(gpa, try reader.takeByte());
|
|
}
|
|
return 2 + unpacked_size;
|
|
}
|
|
};
|
|
|
|
test "decompress hello world stream" {
|
|
const expected = "Hello\nWorld!\n";
|
|
const compressed = &[_]u8{ 0x01, 0x00, 0x05, 0x48, 0x65, 0x6C, 0x6C, 0x6F, 0x0A, 0x02, 0x00, 0x06, 0x57, 0x6F, 0x72, 0x6C, 0x64, 0x21, 0x0A, 0x00 };
|
|
|
|
const gpa = std.testing.allocator;
|
|
|
|
var decode = try Decode.init(gpa);
|
|
defer decode.deinit(gpa);
|
|
|
|
var stream: std.Io.Reader = .fixed(compressed);
|
|
var result: std.Io.Writer.Allocating = .init(gpa);
|
|
defer result.deinit();
|
|
|
|
const n_read = try decode.decompress(&stream, &result);
|
|
try std.testing.expectEqual(compressed.len, n_read);
|
|
try std.testing.expectEqualStrings(expected, result.written());
|
|
}
|