zig/lib/compiler/resinator/cvtres.zig
2025-10-02 17:44:52 -04:00

1113 lines
43 KiB
Zig

const std = @import("std");
const Allocator = std.mem.Allocator;
const res = @import("res.zig");
const NameOrOrdinal = res.NameOrOrdinal;
const MemoryFlags = res.MemoryFlags;
const Language = res.Language;
const numPaddingBytesNeeded = @import("compile.zig").Compiler.numPaddingBytesNeeded;
pub const Resource = struct {
type_value: NameOrOrdinal,
name_value: NameOrOrdinal,
data_version: u32,
memory_flags: MemoryFlags,
language: Language,
version: u32,
characteristics: u32,
data: []const u8,
pub fn deinit(self: Resource, allocator: Allocator) void {
self.name_value.deinit(allocator);
self.type_value.deinit(allocator);
allocator.free(self.data);
}
/// Returns true if all fields match the expected value of the resource at the
/// start of all .res files that distinguishes the .res file as 32-bit (as
/// opposed to 16-bit).
pub fn is32BitPreface(self: Resource) bool {
if (self.type_value != .ordinal or self.type_value.ordinal != 0) return false;
if (self.name_value != .ordinal or self.name_value.ordinal != 0) return false;
if (self.data_version != 0) return false;
if (@as(u16, @bitCast(self.memory_flags)) != 0) return false;
if (@as(u16, @bitCast(self.language)) != 0) return false;
if (self.version != 0) return false;
if (self.characteristics != 0) return false;
if (self.data.len != 0) return false;
return true;
}
pub fn isDlgInclude(resource: Resource) bool {
return resource.type_value == .ordinal and resource.type_value.ordinal == @intFromEnum(res.RT.DLGINCLUDE);
}
};
pub const ParsedResources = struct {
list: std.ArrayList(Resource) = .empty,
allocator: Allocator,
pub fn init(allocator: Allocator) ParsedResources {
return .{ .allocator = allocator };
}
pub fn deinit(self: *ParsedResources) void {
for (self.list.items) |*resource| {
resource.deinit(self.allocator);
}
self.list.deinit(self.allocator);
}
};
pub const ParseResOptions = struct {
skip_zero_data_resources: bool = true,
skip_dlginclude_resources: bool = true,
max_size: u64,
};
/// The returned ParsedResources should be freed by calling its `deinit` function.
pub fn parseRes(allocator: Allocator, reader: *std.Io.Reader, options: ParseResOptions) !ParsedResources {
var resources = ParsedResources.init(allocator);
errdefer resources.deinit();
try parseResInto(&resources, reader, options);
return resources;
}
pub fn parseResInto(resources: *ParsedResources, reader: *std.Io.Reader, options: ParseResOptions) !void {
const allocator = resources.allocator;
var bytes_remaining: u64 = options.max_size;
{
const first_resource_and_size = try parseResource(allocator, reader, bytes_remaining);
defer first_resource_and_size.resource.deinit(allocator);
if (!first_resource_and_size.resource.is32BitPreface()) return error.InvalidPreface;
bytes_remaining -= first_resource_and_size.total_size;
}
while (bytes_remaining != 0) {
const resource_and_size = try parseResource(allocator, reader, bytes_remaining);
if (options.skip_zero_data_resources and resource_and_size.resource.data.len == 0) {
resource_and_size.resource.deinit(allocator);
} else if (options.skip_dlginclude_resources and resource_and_size.resource.isDlgInclude()) {
resource_and_size.resource.deinit(allocator);
} else {
errdefer resource_and_size.resource.deinit(allocator);
try resources.list.append(allocator, resource_and_size.resource);
}
bytes_remaining -= resource_and_size.total_size;
}
}
pub const ResourceAndSize = struct {
resource: Resource,
total_size: u64,
};
pub fn parseResource(allocator: Allocator, reader: *std.Io.Reader, max_size: u64) !ResourceAndSize {
const data_size = try reader.takeInt(u32, .little);
const header_size = try reader.takeInt(u32, .little);
const total_size: u64 = @as(u64, header_size) + data_size;
if (total_size > max_size) return error.ImpossibleSize;
const remaining_header_bytes = try reader.take(header_size -| 8);
var remaining_header_reader: std.Io.Reader = .fixed(remaining_header_bytes);
const type_value = try parseNameOrOrdinal(allocator, &remaining_header_reader);
errdefer type_value.deinit(allocator);
const name_value = try parseNameOrOrdinal(allocator, &remaining_header_reader);
errdefer name_value.deinit(allocator);
const padding_after_name = numPaddingBytesNeeded(@intCast(remaining_header_reader.seek));
try remaining_header_reader.discardAll(padding_after_name);
std.debug.assert(remaining_header_reader.seek % 4 == 0);
const data_version = try remaining_header_reader.takeInt(u32, .little);
const memory_flags: MemoryFlags = @bitCast(try remaining_header_reader.takeInt(u16, .little));
const language: Language = @bitCast(try remaining_header_reader.takeInt(u16, .little));
const version = try remaining_header_reader.takeInt(u32, .little);
const characteristics = try remaining_header_reader.takeInt(u32, .little);
if (remaining_header_reader.seek != remaining_header_reader.end) return error.HeaderSizeMismatch;
const data = try allocator.alloc(u8, data_size);
errdefer allocator.free(data);
try reader.readSliceAll(data);
const padding_after_data = numPaddingBytesNeeded(@intCast(data_size));
try reader.discardAll(padding_after_data);
return .{
.resource = .{
.name_value = name_value,
.type_value = type_value,
.language = language,
.memory_flags = memory_flags,
.version = version,
.characteristics = characteristics,
.data_version = data_version,
.data = data,
},
.total_size = header_size + data.len + padding_after_data,
};
}
pub fn parseNameOrOrdinal(allocator: Allocator, reader: *std.Io.Reader) !NameOrOrdinal {
const first_code_unit = try reader.takeInt(u16, .little);
if (first_code_unit == 0xFFFF) {
const ordinal_value = try reader.takeInt(u16, .little);
return .{ .ordinal = ordinal_value };
}
var name_buf = try std.ArrayList(u16).initCapacity(allocator, 16);
errdefer name_buf.deinit(allocator);
var code_unit = first_code_unit;
while (code_unit != 0) {
try name_buf.append(allocator, std.mem.nativeToLittle(u16, code_unit));
code_unit = try reader.takeInt(u16, .little);
}
return .{ .name = try name_buf.toOwnedSliceSentinel(allocator, 0) };
}
pub const CoffOptions = struct {
target: std.coff.IMAGE.FILE.MACHINE = .AMD64,
/// If true, zeroes will be written to all timestamp fields
reproducible: bool = true,
/// If true, the MEM_WRITE flag will not be set in the .rsrc section header
read_only: bool = false,
/// If non-null, a symbol with this name and storage class EXTERNAL will be added to the symbol table.
define_external_symbol: ?[]const u8 = null,
/// Re-use data offsets for resources with data that is identical.
fold_duplicate_data: bool = false,
};
pub const Diagnostics = union {
none: void,
/// Contains the index of the second resource in a duplicate resource pair.
duplicate_resource: usize,
/// Contains the index of the resource that either has data that's too long or
/// caused the total data to overflow.
overflow_resource: usize,
};
pub fn writeCoff(allocator: Allocator, writer: *std.Io.Writer, resources: []const Resource, options: CoffOptions, diagnostics: ?*Diagnostics) !void {
var resource_tree = ResourceTree.init(allocator, options);
defer resource_tree.deinit();
for (resources, 0..) |*resource, i| {
resource_tree.put(resource, i) catch |err| {
switch (err) {
error.DuplicateResource => {
if (diagnostics) |d_ptr| d_ptr.* = .{ .duplicate_resource = i };
},
error.ResourceDataTooLong, error.TotalResourceDataTooLong => {
if (diagnostics) |d_ptr| d_ptr.* = .{ .overflow_resource = i };
},
else => {},
}
return err;
};
}
const lengths = resource_tree.dataLengths();
const byte_size_of_relocation = 10;
const relocations_len: u32 = @intCast(byte_size_of_relocation * resources.len);
const pointer_to_rsrc01_data = @sizeOf(std.coff.Header) + (@sizeOf(std.coff.SectionHeader) * 2);
const pointer_to_relocations = pointer_to_rsrc01_data + lengths.rsrc01;
const pointer_to_rsrc02_data = pointer_to_relocations + relocations_len;
const pointer_to_symbol_table = pointer_to_rsrc02_data + lengths.rsrc02;
const timestamp: i64 = if (options.reproducible) 0 else std.time.timestamp();
const size_of_optional_header = 0;
const machine_type: std.coff.IMAGE.FILE.MACHINE = options.target;
const flags = std.coff.Header.Flags{
.@"32BIT_MACHINE" = true,
};
const number_of_symbols = 5 + @as(u32, @intCast(resources.len)) + @intFromBool(options.define_external_symbol != null);
const coff_header = std.coff.Header{
.machine = machine_type,
.number_of_sections = 2,
.time_date_stamp = @as(u32, @truncate(@as(u64, @bitCast(timestamp)))),
.pointer_to_symbol_table = pointer_to_symbol_table,
.number_of_symbols = number_of_symbols,
.size_of_optional_header = size_of_optional_header,
.flags = flags,
};
try writer.writeStruct(coff_header, .little);
const rsrc01_header = std.coff.SectionHeader{
.name = ".rsrc$01".*,
.virtual_size = 0,
.virtual_address = 0,
.size_of_raw_data = lengths.rsrc01,
.pointer_to_raw_data = pointer_to_rsrc01_data,
.pointer_to_relocations = if (relocations_len != 0) pointer_to_relocations else 0,
.pointer_to_linenumbers = 0,
.number_of_relocations = @intCast(resources.len),
.number_of_linenumbers = 0,
.flags = .{
.CNT_INITIALIZED_DATA = true,
.MEM_WRITE = !options.read_only,
.MEM_READ = true,
},
};
try writer.writeStruct(rsrc01_header, .little);
const rsrc02_header = std.coff.SectionHeader{
.name = ".rsrc$02".*,
.virtual_size = 0,
.virtual_address = 0,
.size_of_raw_data = lengths.rsrc02,
.pointer_to_raw_data = pointer_to_rsrc02_data,
.pointer_to_relocations = 0,
.pointer_to_linenumbers = 0,
.number_of_relocations = 0,
.number_of_linenumbers = 0,
.flags = .{
.CNT_INITIALIZED_DATA = true,
.MEM_WRITE = !options.read_only,
.MEM_READ = true,
},
};
try writer.writeStruct(rsrc02_header, .little);
// TODO: test surrogate pairs
try resource_tree.sort();
var string_table = StringTable{};
defer string_table.deinit(allocator);
const resource_symbols = try resource_tree.writeCoff(
allocator,
writer,
resources,
lengths,
&string_table,
);
defer allocator.free(resource_symbols);
try writeSymbol(writer, .{
.name = "@feat.00".*,
.value = 0x11,
.section_number = .ABSOLUTE,
.type = .{
.base_type = .NULL,
.complex_type = .NULL,
},
.storage_class = .STATIC,
.number_of_aux_symbols = 0,
});
try writeSymbol(writer, .{
.name = ".rsrc$01".*,
.value = 0,
.section_number = @enumFromInt(1),
.type = .{
.base_type = .NULL,
.complex_type = .NULL,
},
.storage_class = .STATIC,
.number_of_aux_symbols = 1,
});
try writeSectionDefinition(writer, .{
.length = lengths.rsrc01,
.number_of_relocations = @intCast(resources.len),
.number_of_linenumbers = 0,
.checksum = 0,
.number = 0,
.selection = .NONE,
.unused = .{0} ** 3,
});
try writeSymbol(writer, .{
.name = ".rsrc$02".*,
.value = 0,
.section_number = @enumFromInt(2),
.type = .{
.base_type = .NULL,
.complex_type = .NULL,
},
.storage_class = .STATIC,
.number_of_aux_symbols = 1,
});
try writeSectionDefinition(writer, .{
.length = lengths.rsrc02,
.number_of_relocations = 0,
.number_of_linenumbers = 0,
.checksum = 0,
.number = 0,
.selection = .NONE,
.unused = .{0} ** 3,
});
for (resource_symbols) |resource_symbol| {
try writeSymbol(writer, resource_symbol);
}
if (options.define_external_symbol) |external_symbol_name| {
const name_bytes: [8]u8 = name_bytes: {
if (external_symbol_name.len > 8) {
const string_table_offset: u32 = try string_table.put(allocator, external_symbol_name);
var bytes = [_]u8{0} ** 8;
std.mem.writeInt(u32, bytes[4..8], string_table_offset, .little);
break :name_bytes bytes;
} else {
var symbol_shortname = [_]u8{0} ** 8;
@memcpy(symbol_shortname[0..external_symbol_name.len], external_symbol_name);
break :name_bytes symbol_shortname;
}
};
try writeSymbol(writer, .{
.name = name_bytes,
.value = 0,
.section_number = .ABSOLUTE,
.type = .{
.base_type = .NULL,
.complex_type = .NULL,
},
.storage_class = .EXTERNAL,
.number_of_aux_symbols = 0,
});
}
try writer.writeInt(u32, string_table.totalByteLength(), .little);
try writer.writeAll(string_table.bytes.items);
}
fn writeSymbol(writer: *std.Io.Writer, symbol: std.coff.Symbol) !void {
try writer.writeAll(&symbol.name);
try writer.writeInt(u32, symbol.value, .little);
try writer.writeInt(u16, @intFromEnum(symbol.section_number), .little);
try writer.writeInt(u8, @intFromEnum(symbol.type.base_type), .little);
try writer.writeInt(u8, @intFromEnum(symbol.type.complex_type), .little);
try writer.writeInt(u8, @intFromEnum(symbol.storage_class), .little);
try writer.writeInt(u8, symbol.number_of_aux_symbols, .little);
}
fn writeSectionDefinition(writer: *std.Io.Writer, def: std.coff.SectionDefinition) !void {
try writer.writeInt(u32, def.length, .little);
try writer.writeInt(u16, def.number_of_relocations, .little);
try writer.writeInt(u16, def.number_of_linenumbers, .little);
try writer.writeInt(u32, def.checksum, .little);
try writer.writeInt(u16, def.number, .little);
try writer.writeInt(u8, @intFromEnum(def.selection), .little);
try writer.writeAll(&def.unused);
}
pub const ResourceDirectoryTable = extern struct {
characteristics: u32,
timestamp: u32,
major_version: u16,
minor_version: u16,
number_of_name_entries: u16,
number_of_id_entries: u16,
};
pub const ResourceDirectoryEntry = extern struct {
entry: packed union {
name_offset: packed struct(u32) {
address: u31,
/// This is undocumented in the PE/COFF spec, but the high bit
/// is set by cvtres.exe for string addresses
to_string: bool = true,
},
integer_id: u32,
},
offset: packed struct(u32) {
address: u31,
to_subdirectory: bool,
},
pub fn writeCoff(self: ResourceDirectoryEntry, writer: *std.Io.Writer) !void {
try writer.writeInt(u32, @bitCast(self.entry), .little);
try writer.writeInt(u32, @bitCast(self.offset), .little);
}
};
pub const ResourceDataEntry = extern struct {
data_rva: u32,
size: u32,
codepage: u32,
reserved: u32 = 0,
};
/// type -> name -> language
const ResourceTree = struct {
type_to_name_map: std.ArrayHashMapUnmanaged(NameOrOrdinal, NameToLanguageMap, NameOrOrdinalHashContext, true),
rsrc_string_table: std.ArrayHashMapUnmanaged(NameOrOrdinal, void, NameOrOrdinalHashContext, true),
deduplicated_data: std.StringArrayHashMapUnmanaged(u32),
data_offsets: std.ArrayList(u32),
rsrc02_len: u32,
coff_options: CoffOptions,
allocator: Allocator,
const RelocatableResource = struct {
resource: *const Resource,
original_index: usize,
};
const LanguageToResourceMap = std.AutoArrayHashMapUnmanaged(Language, RelocatableResource);
const NameToLanguageMap = std.ArrayHashMapUnmanaged(NameOrOrdinal, LanguageToResourceMap, NameOrOrdinalHashContext, true);
const NameOrOrdinalHashContext = struct {
pub fn hash(self: @This(), v: NameOrOrdinal) u32 {
_ = self;
var hasher = std.hash.Wyhash.init(0);
const tag = std.meta.activeTag(v);
hasher.update(std.mem.asBytes(&tag));
switch (v) {
.name => |name| {
hasher.update(std.mem.sliceAsBytes(name));
},
.ordinal => |*ordinal| {
hasher.update(std.mem.asBytes(ordinal));
},
}
return @truncate(hasher.final());
}
pub fn eql(self: @This(), a: NameOrOrdinal, b: NameOrOrdinal, b_index: usize) bool {
_ = self;
_ = b_index;
const tag_a = std.meta.activeTag(a);
const tag_b = std.meta.activeTag(b);
if (tag_a != tag_b) return false;
return switch (a) {
.name => std.mem.eql(u16, a.name, b.name),
.ordinal => a.ordinal == b.ordinal,
};
}
};
pub fn init(allocator: Allocator, coff_options: CoffOptions) ResourceTree {
return .{
.type_to_name_map = .empty,
.rsrc_string_table = .empty,
.deduplicated_data = .empty,
.data_offsets = .empty,
.rsrc02_len = 0,
.coff_options = coff_options,
.allocator = allocator,
};
}
pub fn deinit(self: *ResourceTree) void {
for (self.type_to_name_map.values()) |*name_to_lang_map| {
for (name_to_lang_map.values()) |*lang_to_resources_map| {
lang_to_resources_map.deinit(self.allocator);
}
name_to_lang_map.deinit(self.allocator);
}
self.type_to_name_map.deinit(self.allocator);
self.rsrc_string_table.deinit(self.allocator);
self.deduplicated_data.deinit(self.allocator);
self.data_offsets.deinit(self.allocator);
}
pub fn put(self: *ResourceTree, resource: *const Resource, original_index: usize) !void {
const name_to_lang_map = blk: {
const gop_result = try self.type_to_name_map.getOrPut(self.allocator, resource.type_value);
if (!gop_result.found_existing) {
gop_result.value_ptr.* = .empty;
}
break :blk gop_result.value_ptr;
};
const lang_to_resources_map = blk: {
const gop_result = try name_to_lang_map.getOrPut(self.allocator, resource.name_value);
if (!gop_result.found_existing) {
gop_result.value_ptr.* = .empty;
}
break :blk gop_result.value_ptr;
};
{
const gop_result = try lang_to_resources_map.getOrPut(self.allocator, resource.language);
if (gop_result.found_existing) return error.DuplicateResource;
gop_result.value_ptr.* = .{
.original_index = original_index,
.resource = resource,
};
}
// Resize the data_offsets list to accommodate the index, but only if necessary
try self.data_offsets.resize(self.allocator, @max(self.data_offsets.items.len, original_index + 1));
if (self.coff_options.fold_duplicate_data) {
const gop_result = try self.deduplicated_data.getOrPut(self.allocator, resource.data);
if (!gop_result.found_existing) {
gop_result.value_ptr.* = self.rsrc02_len;
try self.incrementRsrc02Len(resource);
}
self.data_offsets.items[original_index] = gop_result.value_ptr.*;
} else {
self.data_offsets.items[original_index] = self.rsrc02_len;
try self.incrementRsrc02Len(resource);
}
if (resource.type_value == .name and !self.rsrc_string_table.contains(resource.type_value)) {
try self.rsrc_string_table.putNoClobber(self.allocator, resource.type_value, {});
}
if (resource.name_value == .name and !self.rsrc_string_table.contains(resource.name_value)) {
try self.rsrc_string_table.putNoClobber(self.allocator, resource.name_value, {});
}
}
fn incrementRsrc02Len(self: *ResourceTree, resource: *const Resource) !void {
// Note: This @intCast is only safe if we assume that the resource was parsed from a .res file,
// since the maximum data length for a resource in the .res file format is maxInt(u32).
// TODO: Either codify this properly or use std.math.cast and return an error.
const data_len: u32 = @intCast(resource.data.len);
const data_len_including_padding: u32 = std.math.cast(u32, std.mem.alignForward(u33, data_len, 8)) orelse {
return error.ResourceDataTooLong;
};
// TODO: Verify that this corresponds to an actual PE/COFF limitation for resource data
// in the final linked binary. The limit may turn out to be shorter than u32 max if both
// the tree data and the resource data lengths together need to fit within a u32,
// or it may be longer in which case we would want to add more .rsrc$NN sections
// to the object file for the data that overflows .rsrc$02.
self.rsrc02_len = std.math.add(u32, self.rsrc02_len, data_len_including_padding) catch {
return error.TotalResourceDataTooLong;
};
}
const Lengths = struct {
level1: u32,
level2: u32,
level3: u32,
data_entries: u32,
strings: u32,
padding: u32,
rsrc01: u32,
rsrc02: u32,
fn stringsStart(self: Lengths) u32 {
return self.rsrc01 - self.strings - self.padding;
}
};
pub fn dataLengths(self: *const ResourceTree) Lengths {
var lengths: Lengths = .{
.level1 = 0,
.level2 = 0,
.level3 = 0,
.data_entries = 0,
.strings = 0,
.padding = 0,
.rsrc01 = undefined,
.rsrc02 = self.rsrc02_len,
};
lengths.level1 += @sizeOf(ResourceDirectoryTable);
for (self.type_to_name_map.values()) |name_to_lang_map| {
lengths.level1 += @sizeOf(ResourceDirectoryEntry);
lengths.level2 += @sizeOf(ResourceDirectoryTable);
for (name_to_lang_map.values()) |lang_to_resources_map| {
lengths.level2 += @sizeOf(ResourceDirectoryEntry);
lengths.level3 += @sizeOf(ResourceDirectoryTable);
for (lang_to_resources_map.values()) |_| {
lengths.level3 += @sizeOf(ResourceDirectoryEntry);
lengths.data_entries += @sizeOf(ResourceDataEntry);
}
}
}
for (self.rsrc_string_table.keys()) |v| {
lengths.strings += @sizeOf(u16); // string length
lengths.strings += @intCast(v.name.len * @sizeOf(u16));
}
lengths.rsrc01 = lengths.level1 + lengths.level2 + lengths.level3 + lengths.data_entries + lengths.strings;
lengths.padding = @intCast((4 -% lengths.rsrc01) % 4);
lengths.rsrc01 += lengths.padding;
return lengths;
}
pub fn sort(self: *ResourceTree) !void {
const NameOrOrdinalSortContext = struct {
keys: []NameOrOrdinal,
pub fn lessThan(ctx: @This(), a_index: usize, b_index: usize) bool {
const a = ctx.keys[a_index];
const b = ctx.keys[b_index];
if (std.meta.activeTag(a) != std.meta.activeTag(b)) {
return if (a == .name) true else false;
}
switch (a) {
.name => {
const n = @min(a.name.len, b.name.len);
for (a.name[0..n], b.name[0..n]) |a_c, b_c| {
switch (std.math.order(std.mem.littleToNative(u16, a_c), std.mem.littleToNative(u16, b_c))) {
.eq => continue,
.lt => return true,
.gt => return false,
}
}
return a.name.len < b.name.len;
},
.ordinal => {
return a.ordinal < b.ordinal;
},
}
}
};
self.type_to_name_map.sortUnstable(NameOrOrdinalSortContext{ .keys = self.type_to_name_map.keys() });
for (self.type_to_name_map.values()) |*name_to_lang_map| {
name_to_lang_map.sortUnstable(NameOrOrdinalSortContext{ .keys = name_to_lang_map.keys() });
}
const LangSortContext = struct {
keys: []Language,
pub fn lessThan(ctx: @This(), a_index: usize, b_index: usize) bool {
return @as(u16, @bitCast(ctx.keys[a_index])) < @as(u16, @bitCast(ctx.keys[b_index]));
}
};
for (self.type_to_name_map.values()) |*name_to_lang_map| {
for (name_to_lang_map.values()) |*lang_to_resource_map| {
lang_to_resource_map.sortUnstable(LangSortContext{ .keys = lang_to_resource_map.keys() });
}
}
}
pub fn writeCoff(
self: *const ResourceTree,
allocator: Allocator,
w: *std.Io.Writer,
resources_in_data_order: []const Resource,
lengths: Lengths,
coff_string_table: *StringTable,
) ![]const std.coff.Symbol {
if (self.type_to_name_map.count() == 0) {
try w.splatByteAll(0, 16);
return &.{};
}
var level2_list: std.ArrayList(*const NameToLanguageMap) = .empty;
defer level2_list.deinit(allocator);
var level3_list: std.ArrayList(*const LanguageToResourceMap) = .empty;
defer level3_list.deinit(allocator);
var resources_list: std.ArrayList(*const RelocatableResource) = .empty;
defer resources_list.deinit(allocator);
var relocations = Relocations.init(allocator);
defer relocations.deinit();
var string_offsets = try allocator.alloc(u31, self.rsrc_string_table.count());
const strings_start = lengths.stringsStart();
defer allocator.free(string_offsets);
{
var string_address: u31 = @intCast(strings_start);
for (self.rsrc_string_table.keys(), 0..) |v, i| {
string_offsets[i] = string_address;
string_address += @sizeOf(u16) + @as(u31, @intCast(v.name.len * @sizeOf(u16)));
}
}
const level2_start = lengths.level1;
var level2_address = level2_start;
{
const counts = entryTypeCounts(self.type_to_name_map.keys());
const table = ResourceDirectoryTable{
.characteristics = 0,
.timestamp = 0,
.major_version = 0,
.minor_version = 0,
.number_of_id_entries = counts.ids,
.number_of_name_entries = counts.names,
};
try w.writeStruct(table, .little);
var it = self.type_to_name_map.iterator();
while (it.next()) |entry| {
const type_value = entry.key_ptr;
const dir_entry = ResourceDirectoryEntry{
.entry = switch (type_value.*) {
.name => .{ .name_offset = .{ .address = string_offsets[self.rsrc_string_table.getIndex(type_value.*).?] } },
.ordinal => .{ .integer_id = type_value.ordinal },
},
.offset = .{
.address = @intCast(level2_address),
.to_subdirectory = true,
},
};
try dir_entry.writeCoff(w);
level2_address += @sizeOf(ResourceDirectoryTable) + @as(u32, @intCast(entry.value_ptr.count() * @sizeOf(ResourceDirectoryEntry)));
const name_to_lang_map = entry.value_ptr;
try level2_list.append(allocator, name_to_lang_map);
}
}
const level3_start = level2_start + lengths.level2;
var level3_address = level3_start;
for (level2_list.items) |name_to_lang_map| {
const counts = entryTypeCounts(name_to_lang_map.keys());
const table = ResourceDirectoryTable{
.characteristics = 0,
.timestamp = 0,
.major_version = 0,
.minor_version = 0,
.number_of_id_entries = counts.ids,
.number_of_name_entries = counts.names,
};
try w.writeStruct(table, .little);
var it = name_to_lang_map.iterator();
while (it.next()) |entry| {
const name_value = entry.key_ptr;
const dir_entry = ResourceDirectoryEntry{
.entry = switch (name_value.*) {
.name => .{ .name_offset = .{ .address = string_offsets[self.rsrc_string_table.getIndex(name_value.*).?] } },
.ordinal => .{ .integer_id = name_value.ordinal },
},
.offset = .{
.address = @intCast(level3_address),
.to_subdirectory = true,
},
};
try dir_entry.writeCoff(w);
level3_address += @sizeOf(ResourceDirectoryTable) + @as(u32, @intCast(entry.value_ptr.count() * @sizeOf(ResourceDirectoryEntry)));
const lang_to_resources_map = entry.value_ptr;
try level3_list.append(allocator, lang_to_resources_map);
}
}
var reloc_addresses = try allocator.alloc(u32, resources_in_data_order.len);
defer allocator.free(reloc_addresses);
const data_entries_start = level3_start + lengths.level3;
var data_entry_address = data_entries_start;
for (level3_list.items) |lang_to_resources_map| {
const counts = EntryTypeCounts{
.names = 0,
.ids = @intCast(lang_to_resources_map.count()),
};
const table = ResourceDirectoryTable{
.characteristics = 0,
.timestamp = 0,
.major_version = 0,
.minor_version = 0,
.number_of_id_entries = counts.ids,
.number_of_name_entries = counts.names,
};
try w.writeStruct(table, .little);
var it = lang_to_resources_map.iterator();
while (it.next()) |entry| {
const lang = entry.key_ptr.*;
const dir_entry = ResourceDirectoryEntry{
.entry = .{ .integer_id = lang.asInt() },
.offset = .{
.address = @intCast(data_entry_address),
.to_subdirectory = false,
},
};
const reloc_resource = entry.value_ptr;
reloc_addresses[reloc_resource.original_index] = @intCast(data_entry_address);
try dir_entry.writeCoff(w);
data_entry_address += @sizeOf(ResourceDataEntry);
try resources_list.append(allocator, reloc_resource);
}
}
for (resources_list.items, 0..) |reloc_resource, i| {
// TODO: This logic works but is convoluted, would be good to clean this up
const orig_resource = &resources_in_data_order[reloc_resource.original_index];
const address: u32 = reloc_addresses[i];
try relocations.add(address, self.data_offsets.items[i]);
const data_entry = ResourceDataEntry{
.data_rva = 0, // relocation
.size = @intCast(orig_resource.data.len),
.codepage = 0,
};
try w.writeStruct(data_entry, .little);
}
for (self.rsrc_string_table.keys()) |v| {
const str = v.name;
try w.writeInt(u16, @intCast(str.len), .little);
try w.writeAll(std.mem.sliceAsBytes(str));
}
try w.splatByteAll(0, lengths.padding);
for (relocations.list.items) |relocation| {
try writeRelocation(w, std.coff.Relocation{
.virtual_address = relocation.relocation_address,
.symbol_table_index = relocation.symbol_index,
.type = supported_targets.rvaRelocationTypeIndicator(self.coff_options.target).?,
});
}
if (self.coff_options.fold_duplicate_data) {
for (self.deduplicated_data.keys()) |data| {
const padding_bytes: u4 = @intCast((8 -% data.len) % 8);
try w.writeAll(data);
try w.splatByteAll(0, padding_bytes);
}
} else {
for (resources_in_data_order) |resource| {
const padding_bytes: u4 = @intCast((8 -% resource.data.len) % 8);
try w.writeAll(resource.data);
try w.splatByteAll(0, padding_bytes);
}
}
var symbols = try allocator.alloc(std.coff.Symbol, resources_list.items.len);
errdefer allocator.free(symbols);
for (relocations.list.items, 0..) |relocation, i| {
// cvtres.exe writes the symbol names as $R<data offset as hexadecimal>.
//
// When the data offset would exceed 6 hex digits in cvtres.exe, it
// truncates the value down to 6 hex digits. This is bad behavior, since
// e.g. an initial resource with exactly 16 MiB of data and the
// resource following it would both have the symbol name $R000000.
//
// Instead, if the offset would exceed 6 hexadecimal digits,
// we put the longer name in the string table.
//
// Another option would be to adopt llvm-cvtres' behavior
// of $R000001, $R000002, etc. rather than using data offset values.
var name_buf: [8]u8 = undefined;
if (relocation.data_offset > std.math.maxInt(u24)) {
const name_slice = try std.fmt.allocPrint(allocator, "$R{X}", .{relocation.data_offset});
defer allocator.free(name_slice);
const string_table_offset: u32 = try coff_string_table.put(allocator, name_slice);
std.mem.writeInt(u32, name_buf[0..4], 0, .little);
std.mem.writeInt(u32, name_buf[4..8], string_table_offset, .little);
} else {
const name_slice = std.fmt.bufPrint(&name_buf, "$R{X:0>6}", .{relocation.data_offset}) catch unreachable;
std.debug.assert(name_slice.len == 8);
}
symbols[i] = .{
.name = name_buf,
.value = relocation.data_offset,
.section_number = @enumFromInt(2),
.type = .{
.base_type = .NULL,
.complex_type = .NULL,
},
.storage_class = .STATIC,
.number_of_aux_symbols = 0,
};
}
return symbols;
}
fn writeRelocation(writer: *std.Io.Writer, relocation: std.coff.Relocation) !void {
try writer.writeInt(u32, relocation.virtual_address, .little);
try writer.writeInt(u32, relocation.symbol_table_index, .little);
try writer.writeInt(u16, relocation.type, .little);
}
const EntryTypeCounts = struct {
names: u16,
ids: u16,
};
fn entryTypeCounts(s: []const NameOrOrdinal) EntryTypeCounts {
var names: u16 = 0;
var ordinals: u16 = 0;
for (s) |v| {
switch (v) {
.name => names += 1,
.ordinal => ordinals += 1,
}
}
return .{ .names = names, .ids = ordinals };
}
};
const Relocation = struct {
symbol_index: u32,
data_offset: u32,
relocation_address: u32,
};
const Relocations = struct {
allocator: Allocator,
list: std.ArrayList(Relocation) = .empty,
cur_symbol_index: u32 = 5,
pub fn init(allocator: Allocator) Relocations {
return .{ .allocator = allocator };
}
pub fn deinit(self: *Relocations) void {
self.list.deinit(self.allocator);
}
pub fn add(self: *Relocations, relocation_address: u32, data_offset: u32) !void {
try self.list.append(self.allocator, .{
.symbol_index = self.cur_symbol_index,
.data_offset = data_offset,
.relocation_address = relocation_address,
});
self.cur_symbol_index += 1;
}
};
/// Does not do deduplication (only because there's no chance of duplicate strings in this
/// instance).
const StringTable = struct {
bytes: std.ArrayList(u8) = .empty,
pub fn deinit(self: *StringTable, allocator: Allocator) void {
self.bytes.deinit(allocator);
}
/// Returns the byte offset of the string in the string table
pub fn put(self: *StringTable, allocator: Allocator, string: []const u8) !u32 {
const null_terminated_len = string.len + 1;
const start_offset = self.totalByteLength();
if (start_offset + null_terminated_len > std.math.maxInt(u32)) {
return error.StringTableOverflow;
}
try self.bytes.ensureUnusedCapacity(allocator, null_terminated_len);
self.bytes.appendSliceAssumeCapacity(string);
self.bytes.appendAssumeCapacity(0);
return start_offset;
}
/// Returns the total byte count of the string table, including the byte count of the size field
pub fn totalByteLength(self: StringTable) u32 {
return @intCast(4 + self.bytes.items.len);
}
};
pub const supported_targets = struct {
/// Enum containing a mixture of names that come from:
/// - Machine Types constants in the PE format spec:
/// https://learn.microsoft.com/en-us/windows/win32/debug/pe-format#machine-types
/// - cvtres.exe /machine options
/// - Zig/LLVM arch names
/// All field names are lowercase regardless of their casing used in the above origins.
pub const Arch = enum {
// cvtres.exe /machine names
x64,
x86,
/// Note: Following cvtres.exe's lead, this corresponds to ARMNT, not ARM
arm,
arm64,
arm64ec,
arm64x,
ia64,
ebc,
// PE/COFF MACHINE constant names not covered above
amd64,
i386,
armnt,
// Zig/LLVM names not already covered above
x86_64,
aarch64,
pub fn toCoffMachineType(arch: Arch) std.coff.IMAGE.FILE.MACHINE {
return switch (arch) {
.x64, .amd64, .x86_64 => .AMD64,
.x86, .i386 => .I386,
.arm, .armnt => .ARMNT,
.arm64, .aarch64 => .ARM64,
.arm64ec => .ARM64EC,
.arm64x => .ARM64X,
.ia64 => .IA64,
.ebc => .EBC,
};
}
pub fn description(arch: Arch) []const u8 {
return switch (arch) {
.x64, .amd64, .x86_64 => "64-bit X86",
.x86, .i386 => "32-bit X86",
.arm, .armnt => "ARM Thumb-2 little endian",
.arm64, .aarch64 => "ARM64/AArch64 little endian",
.arm64ec => "ARM64 \"Emulation Compatible\"",
.arm64x => "ARM64 and ARM64EC together",
.ia64 => "64-bit Intel Itanium",
.ebc => "EFI Byte Code",
};
}
pub const ordered_for_display: []const Arch = &.{
.x64,
.x86_64,
.amd64,
.x86,
.i386,
.arm64,
.aarch64,
.arm,
.armnt,
.arm64ec,
.arm64x,
.ia64,
.ebc,
};
comptime {
for (@typeInfo(Arch).@"enum".fields) |enum_field| {
_ = std.mem.indexOfScalar(Arch, ordered_for_display, @enumFromInt(enum_field.value)) orelse {
@compileError(std.fmt.comptimePrint("'{s}' missing from ordered_for_display", .{enum_field.name}));
};
}
}
pub const longest_name = blk: {
var len = 0;
for (@typeInfo(Arch).@"enum".fields) |field| {
if (field.name.len > len) len = field.name.len;
}
break :blk len;
};
pub fn fromStringIgnoreCase(str: []const u8) ?Arch {
if (str.len > longest_name) return null;
var lower_buf: [longest_name]u8 = undefined;
const lower = std.ascii.lowerString(&lower_buf, str);
return std.meta.stringToEnum(Arch, lower);
}
test fromStringIgnoreCase {
try std.testing.expectEqual(.x64, Arch.fromStringIgnoreCase("x64").?);
try std.testing.expectEqual(.x64, Arch.fromStringIgnoreCase("X64").?);
try std.testing.expectEqual(.aarch64, Arch.fromStringIgnoreCase("Aarch64").?);
try std.testing.expectEqual(null, Arch.fromStringIgnoreCase("armzzz"));
try std.testing.expectEqual(null, Arch.fromStringIgnoreCase("long string that is longer than any field"));
}
};
// https://learn.microsoft.com/en-us/windows/win32/debug/pe-format#type-indicators
pub fn rvaRelocationTypeIndicator(target: std.coff.IMAGE.FILE.MACHINE) ?u16 {
return switch (target) {
.AMD64 => @intFromEnum(std.coff.IMAGE.REL.AMD64.ADDR32NB),
.I386 => @intFromEnum(std.coff.IMAGE.REL.I386.DIR32NB),
.ARMNT => @intFromEnum(std.coff.IMAGE.REL.ARM.ADDR32NB),
.ARM64, .ARM64EC, .ARM64X => @intFromEnum(std.coff.IMAGE.REL.ARM64.ADDR32NB),
.IA64 => @intFromEnum(std.coff.IMAGE.REL.IA64.DIR32NB),
.EBC => 0x1, // This is what cvtres.exe writes for this target, unsure where it comes from
else => null,
};
}
pub fn isSupported(target: std.coff.IMAGE.FILE.MACHINE) bool {
return rvaRelocationTypeIndicator(target) != null;
}
comptime {
// Enforce two things:
// 1. Arch enum field names are all lowercase (necessary for how fromStringIgnoreCase is implemented)
// 2. All enum fields in Arch have an associated RVA relocation type when converted to a coff.IMAGE.FILE.MACHINE
for (@typeInfo(Arch).@"enum".fields) |enum_field| {
const all_lower = all_lower: for (enum_field.name) |c| {
if (std.ascii.isUpper(c)) break :all_lower false;
} else break :all_lower true;
if (!all_lower) @compileError(std.fmt.comptimePrint("Arch field is not all lowercase: {s}", .{enum_field.name}));
const coff_machine = @field(Arch, enum_field.name).toCoffMachineType();
_ = rvaRelocationTypeIndicator(coff_machine) orelse {
@compileError(std.fmt.comptimePrint("No RVA relocation for Arch: {s}", .{enum_field.name}));
};
}
}
};