zig/lib/compiler/resinator/compile.zig

3356 lines
166 KiB
Zig
Raw Permalink Blame History

const std = @import("std");
const builtin = @import("builtin");
const Allocator = std.mem.Allocator;
const Node = @import("ast.zig").Node;
const lex = @import("lex.zig");
const Parser = @import("parse.zig").Parser;
const ResourceType = @import("rc.zig").ResourceType;
const Token = @import("lex.zig").Token;
const literals = @import("literals.zig");
const Number = literals.Number;
const SourceBytes = literals.SourceBytes;
const Diagnostics = @import("errors.zig").Diagnostics;
const ErrorDetails = @import("errors.zig").ErrorDetails;
const MemoryFlags = @import("res.zig").MemoryFlags;
const rc = @import("rc.zig");
const res = @import("res.zig");
const ico = @import("ico.zig");
const ani = @import("ani.zig");
const bmp = @import("bmp.zig");
const WORD = std.os.windows.WORD;
const DWORD = std.os.windows.DWORD;
const utils = @import("utils.zig");
const NameOrOrdinal = res.NameOrOrdinal;
const SupportedCodePage = @import("code_pages.zig").SupportedCodePage;
const CodePageLookup = @import("ast.zig").CodePageLookup;
const SourceMappings = @import("source_mapping.zig").SourceMappings;
const windows1252 = @import("windows1252.zig");
const lang = @import("lang.zig");
const code_pages = @import("code_pages.zig");
const errors = @import("errors.zig");
const native_endian = builtin.cpu.arch.endian();
pub const CompileOptions = struct {
cwd: std.fs.Dir,
diagnostics: *Diagnostics,
source_mappings: ?*SourceMappings = null,
/// List of paths (absolute or relative to `cwd`) for every file that the resources within the .rc file depend on.
dependencies: ?*Dependencies = null,
default_code_page: SupportedCodePage = .windows1252,
/// If true, the first #pragma code_page directive only sets the input code page, but not the output code page.
/// This check must be done before comments are removed from the file.
disjoint_code_page: bool = false,
ignore_include_env_var: bool = false,
extra_include_paths: []const []const u8 = &.{},
/// This is just an API convenience to allow separately passing 'system' (i.e. those
/// that would normally be gotten from the INCLUDE env var) include paths. This is mostly
/// intended for use when setting `ignore_include_env_var = true`. When `ignore_include_env_var`
/// is false, `system_include_paths` will be searched before the paths in the INCLUDE env var.
system_include_paths: []const []const u8 = &.{},
default_language_id: ?u16 = null,
// TODO: Implement verbose output
verbose: bool = false,
null_terminate_string_table_strings: bool = false,
/// Note: This is a u15 to ensure that the maximum number of UTF-16 code units
/// plus a null-terminator can always fit into a u16.
max_string_literal_codepoints: u15 = lex.default_max_string_literal_codepoints,
silent_duplicate_control_ids: bool = false,
warn_instead_of_error_on_invalid_code_page: bool = false,
};
pub const Dependencies = struct {
list: std.ArrayList([]const u8),
allocator: Allocator,
pub fn init(allocator: Allocator) Dependencies {
return .{
.list = .empty,
.allocator = allocator,
};
}
pub fn deinit(self: *Dependencies) void {
for (self.list.items) |item| {
self.allocator.free(item);
}
self.list.deinit(self.allocator);
}
};
pub fn compile(allocator: Allocator, source: []const u8, writer: *std.Io.Writer, options: CompileOptions) !void {
var lexer = lex.Lexer.init(source, .{
.default_code_page = options.default_code_page,
.source_mappings = options.source_mappings,
.max_string_literal_codepoints = options.max_string_literal_codepoints,
});
var parser = Parser.init(&lexer, .{
.warn_instead_of_error_on_invalid_code_page = options.warn_instead_of_error_on_invalid_code_page,
.disjoint_code_page = options.disjoint_code_page,
});
var tree = try parser.parse(allocator, options.diagnostics);
defer tree.deinit();
var search_dirs: std.ArrayList(SearchDir) = .empty;
defer {
for (search_dirs.items) |*search_dir| {
search_dir.deinit(allocator);
}
search_dirs.deinit(allocator);
}
if (options.source_mappings) |source_mappings| {
const root_path = source_mappings.files.get(source_mappings.root_filename_offset);
// If dirname returns null, then the root path will be the same as
// the cwd so we don't need to add it as a distinct search path.
if (std.fs.path.dirname(root_path)) |root_dir_path| {
var root_dir = try options.cwd.openDir(root_dir_path, .{});
errdefer root_dir.close();
try search_dirs.append(allocator, .{ .dir = root_dir, .path = try allocator.dupe(u8, root_dir_path) });
}
}
// Re-open the passed in cwd since we want to be able to close it (std.fs.cwd() shouldn't be closed)
const cwd_dir = options.cwd.openDir(".", .{}) catch |err| {
try options.diagnostics.append(.{
.err = .failed_to_open_cwd,
.token = .{
.id = .invalid,
.start = 0,
.end = 0,
.line_number = 1,
},
.code_page = .utf8,
.print_source_line = false,
.extra = .{ .file_open_error = .{
.err = ErrorDetails.FileOpenError.enumFromError(err),
.filename_string_index = undefined,
} },
});
return error.CompileError;
};
try search_dirs.append(allocator, .{ .dir = cwd_dir, .path = null });
for (options.extra_include_paths) |extra_include_path| {
var dir = openSearchPathDir(options.cwd, extra_include_path) catch {
// TODO: maybe a warning that the search path is skipped?
continue;
};
errdefer dir.close();
try search_dirs.append(allocator, .{ .dir = dir, .path = try allocator.dupe(u8, extra_include_path) });
}
for (options.system_include_paths) |system_include_path| {
var dir = openSearchPathDir(options.cwd, system_include_path) catch {
// TODO: maybe a warning that the search path is skipped?
continue;
};
errdefer dir.close();
try search_dirs.append(allocator, .{ .dir = dir, .path = try allocator.dupe(u8, system_include_path) });
}
if (!options.ignore_include_env_var) {
const INCLUDE = std.process.getEnvVarOwned(allocator, "INCLUDE") catch "";
defer allocator.free(INCLUDE);
// The only precedence here is llvm-rc which also uses the platform-specific
// delimiter. There's no precedence set by `rc.exe` since it's Windows-only.
const delimiter = switch (builtin.os.tag) {
.windows => ';',
else => ':',
};
var it = std.mem.tokenizeScalar(u8, INCLUDE, delimiter);
while (it.next()) |search_path| {
var dir = openSearchPathDir(options.cwd, search_path) catch continue;
errdefer dir.close();
try search_dirs.append(allocator, .{ .dir = dir, .path = try allocator.dupe(u8, search_path) });
}
}
var arena_allocator = std.heap.ArenaAllocator.init(allocator);
defer arena_allocator.deinit();
const arena = arena_allocator.allocator();
var compiler = Compiler{
.source = source,
.arena = arena,
.allocator = allocator,
.cwd = options.cwd,
.diagnostics = options.diagnostics,
.dependencies = options.dependencies,
.input_code_pages = &tree.input_code_pages,
.output_code_pages = &tree.output_code_pages,
// This is only safe because we know search_dirs won't be modified past this point
.search_dirs = search_dirs.items,
.null_terminate_string_table_strings = options.null_terminate_string_table_strings,
.silent_duplicate_control_ids = options.silent_duplicate_control_ids,
};
if (options.default_language_id) |default_language_id| {
compiler.state.language = res.Language.fromInt(default_language_id);
}
try compiler.writeRoot(tree.root(), writer);
}
pub const Compiler = struct {
source: []const u8,
arena: Allocator,
allocator: Allocator,
cwd: std.fs.Dir,
state: State = .{},
diagnostics: *Diagnostics,
dependencies: ?*Dependencies,
input_code_pages: *const CodePageLookup,
output_code_pages: *const CodePageLookup,
search_dirs: []SearchDir,
null_terminate_string_table_strings: bool,
silent_duplicate_control_ids: bool,
pub const State = struct {
icon_id: u16 = 1,
string_tables: StringTablesByLanguage = .{},
language: res.Language = .{},
font_dir: FontDir = .{},
version: u32 = 0,
characteristics: u32 = 0,
};
pub fn writeRoot(self: *Compiler, root: *Node.Root, writer: *std.Io.Writer) !void {
try writeEmptyResource(writer);
for (root.body) |node| {
try self.writeNode(node, writer);
}
// now write the FONTDIR (if it has anything in it)
try self.state.font_dir.writeResData(self, writer);
if (self.state.font_dir.fonts.items.len != 0) {
// The Win32 RC compiler may write a different FONTDIR resource than us,
// due to it sometimes writing a non-zero-length device name/face name
// whereas we *always* write them both as zero-length.
//
// In practical terms, this doesn't matter, since for various reasons the format
// of the FONTDIR cannot be relied on and is seemingly not actually used by anything
// anymore. We still want to emit some sort of diagnostic for the purposes of being able
// to know that our .RES is intentionally not meant to be byte-for-byte identical with
// the rc.exe output.
//
// By using the hint type here, we allow this diagnostic to be detected in code,
// but it will not be printed since the end-user doesn't need to care.
try self.addErrorDetails(.{
.err = .result_contains_fontdir,
.type = .hint,
.token = .{
.id = .invalid,
.start = 0,
.end = 0,
.line_number = 1,
},
});
}
// once we've written every else out, we can write out the finalized STRINGTABLE resources
var string_tables_it = self.state.string_tables.tables.iterator();
while (string_tables_it.next()) |string_table_entry| {
var string_table_it = string_table_entry.value_ptr.blocks.iterator();
while (string_table_it.next()) |entry| {
try entry.value_ptr.writeResData(self, string_table_entry.key_ptr.*, entry.key_ptr.*, writer);
}
}
}
pub fn writeNode(self: *Compiler, node: *Node, writer: *std.Io.Writer) !void {
switch (node.id) {
.root => unreachable, // writeRoot should be called directly instead
.resource_external => try self.writeResourceExternal(@alignCast(@fieldParentPtr("base", node)), writer),
.resource_raw_data => try self.writeResourceRawData(@alignCast(@fieldParentPtr("base", node)), writer),
.literal => unreachable, // this is context dependent and should be handled by its parent
.binary_expression => unreachable,
.grouped_expression => unreachable,
.not_expression => unreachable,
.invalid => {}, // no-op, currently only used for dangling literals at EOF
.accelerators => try self.writeAccelerators(@alignCast(@fieldParentPtr("base", node)), writer),
.accelerator => unreachable, // handled by writeAccelerators
.dialog => try self.writeDialog(@alignCast(@fieldParentPtr("base", node)), writer),
.control_statement => unreachable,
.toolbar => try self.writeToolbar(@alignCast(@fieldParentPtr("base", node)), writer),
.menu => try self.writeMenu(@alignCast(@fieldParentPtr("base", node)), writer),
.menu_item => unreachable,
.menu_item_separator => unreachable,
.menu_item_ex => unreachable,
.popup => unreachable,
.popup_ex => unreachable,
.version_info => try self.writeVersionInfo(@alignCast(@fieldParentPtr("base", node)), writer),
.version_statement => unreachable,
.block => unreachable,
.block_value => unreachable,
.block_value_value => unreachable,
.string_table => try self.writeStringTable(@alignCast(@fieldParentPtr("base", node))),
.string_table_string => unreachable, // handled by writeStringTable
.language_statement => self.writeLanguageStatement(@alignCast(@fieldParentPtr("base", node))),
.font_statement => unreachable,
.simple_statement => self.writeTopLevelSimpleStatement(@alignCast(@fieldParentPtr("base", node))),
}
}
/// Returns the filename encoded as UTF-8 (allocated by self.allocator)
pub fn evaluateFilenameExpression(self: *Compiler, expression_node: *Node) ![]u8 {
switch (expression_node.id) {
.literal => {
const literal_node = expression_node.cast(.literal).?;
switch (literal_node.token.id) {
.literal, .number => {
const slice = literal_node.token.slice(self.source);
const code_page = self.input_code_pages.getForToken(literal_node.token);
var buf = try std.ArrayList(u8).initCapacity(self.allocator, slice.len);
errdefer buf.deinit(self.allocator);
var index: usize = 0;
while (code_page.codepointAt(index, slice)) |codepoint| : (index += codepoint.byte_len) {
const c = codepoint.value;
if (c == code_pages.Codepoint.invalid) {
try buf.appendSlice(self.allocator, "<EFBFBD>");
} else {
// Anything that is not returned as an invalid codepoint must be encodable as UTF-8.
const utf8_len = std.unicode.utf8CodepointSequenceLength(c) catch unreachable;
try buf.ensureUnusedCapacity(self.allocator, utf8_len);
_ = std.unicode.utf8Encode(c, buf.unusedCapacitySlice()) catch unreachable;
buf.items.len += utf8_len;
}
}
return buf.toOwnedSlice(self.allocator);
},
.quoted_ascii_string, .quoted_wide_string => {
const slice = literal_node.token.slice(self.source);
const column = literal_node.token.calculateColumn(self.source, 8, null);
const bytes = SourceBytes{ .slice = slice, .code_page = self.input_code_pages.getForToken(literal_node.token) };
var buf: std.ArrayList(u8) = .empty;
errdefer buf.deinit(self.allocator);
// Filenames are sort-of parsed as if they were wide strings, but the max escape width of
// hex/octal escapes is still determined by the L prefix. Since we want to end up with
// UTF-8, we can parse either string type directly to UTF-8.
var parser = literals.IterativeStringParser.init(bytes, .{
.start_column = column,
.diagnostics = self.errContext(literal_node.token),
// TODO: Re-evaluate this. It's not been tested whether or not using the actual
// output code page would make more sense.
.output_code_page = .windows1252,
});
while (try parser.nextUnchecked()) |parsed| {
const c = parsed.codepoint;
if (c == code_pages.Codepoint.invalid) {
try buf.appendSlice(self.allocator, "<EFBFBD>");
} else {
var codepoint_buf: [4]u8 = undefined;
// If the codepoint cannot be encoded, we fall back to <20>
if (std.unicode.utf8Encode(c, &codepoint_buf)) |len| {
try buf.appendSlice(self.allocator, codepoint_buf[0..len]);
} else |_| {
try buf.appendSlice(self.allocator, "<EFBFBD>");
}
}
}
return buf.toOwnedSlice(self.allocator);
},
else => unreachable, // no other token types should be in a filename literal node
}
},
.binary_expression => {
const binary_expression_node = expression_node.cast(.binary_expression).?;
return self.evaluateFilenameExpression(binary_expression_node.right);
},
.grouped_expression => {
const grouped_expression_node = expression_node.cast(.grouped_expression).?;
return self.evaluateFilenameExpression(grouped_expression_node.expression);
},
else => unreachable,
}
}
/// https://learn.microsoft.com/en-us/windows/win32/menurc/searching-for-files
///
/// Searches, in this order:
/// Directory of the 'root' .rc file (if different from CWD)
/// CWD
/// extra_include_paths (resolved relative to CWD)
/// system_include_paths (resolve relative to CWD)
/// INCLUDE environment var paths (only if ignore_include_env_var is false; resolved relative to CWD)
///
/// Note: The CWD being searched *in addition to* the directory of the 'root' .rc file
/// is also how the Win32 RC compiler preprocessor searches for includes, but that
/// differs from how the clang preprocessor searches for includes.
///
/// Note: This will always return the first matching file that can be opened.
/// This matches the Win32 RC compiler, which will fail with an error if the first
/// matching file is invalid. That is, it does not do the `cmd` PATH searching
/// thing of continuing to look for matching files until it finds a valid
/// one if a matching file is invalid.
fn searchForFile(self: *Compiler, path: []const u8) !std.fs.File {
// If the path is absolute, then it is not resolved relative to any search
// paths, so there's no point in checking them.
//
// This behavior was determined/confirmed with the following test:
// - A `test.rc` file with the contents `1 RCDATA "/test.bin"`
// - A `test.bin` file at `C:\test.bin`
// - A `test.bin` file at `inc\test.bin` relative to the .rc file
// - Invoking `rc` with `rc /i inc test.rc`
//
// This results in a .res file with the contents of `C:\test.bin`, not
// the contents of `inc\test.bin`. Further, if `C:\test.bin` is deleted,
// then it start failing to find `/test.bin`, meaning that it does not resolve
// `/test.bin` relative to include paths and instead only treats it as
// an absolute path.
if (std.fs.path.isAbsolute(path)) {
const file = try utils.openFileNotDir(std.fs.cwd(), path, .{});
errdefer file.close();
if (self.dependencies) |dependencies| {
const duped_path = try dependencies.allocator.dupe(u8, path);
errdefer dependencies.allocator.free(duped_path);
try dependencies.list.append(dependencies.allocator, duped_path);
}
}
var first_error: ?std.fs.File.OpenError = null;
for (self.search_dirs) |search_dir| {
if (utils.openFileNotDir(search_dir.dir, path, .{})) |file| {
errdefer file.close();
if (self.dependencies) |dependencies| {
const searched_file_path = try std.fs.path.join(dependencies.allocator, &.{
search_dir.path orelse "", path,
});
errdefer dependencies.allocator.free(searched_file_path);
try dependencies.list.append(dependencies.allocator, searched_file_path);
}
return file;
} else |err| if (first_error == null) {
first_error = err;
}
}
return first_error orelse error.FileNotFound;
}
/// Returns a Windows-1252 encoded string regardless of the current output code page.
/// All codepoints are encoded as a maximum of 2 bytes, where unescaped codepoints
/// >= 0x10000 are encoded as `??` and everything else is encoded as 1 byte.
pub fn parseDlgIncludeString(self: *Compiler, token: Token) ![]u8 {
const bytes = self.sourceBytesForToken(token);
const output_code_page = self.output_code_pages.getForToken(token);
var buf = try std.ArrayList(u8).initCapacity(self.allocator, bytes.slice.len);
errdefer buf.deinit(self.allocator);
var iterative_parser = literals.IterativeStringParser.init(bytes, .{
.start_column = token.calculateColumn(self.source, 8, null),
.diagnostics = self.errContext(token),
// TODO: Potentially re-evaluate this, it's not been tested whether or not
// using the actual output code page would make more sense.
.output_code_page = .windows1252,
});
// This is similar to the logic in parseQuotedString, but ends up with everything
// encoded as Windows-1252. This effectively consolidates the two-step process
// of rc.exe into one step, since rc.exe's preprocessor converts to UTF-16 (this
// is when invalid sequences are replaced by the replacement character (U+FFFD)),
// and then that's run through the parser. Our preprocessor keeps things in their
// original encoding, meaning we emulate the <encoding> -> UTF-16 -> Windows-1252
// results all at once.
while (try iterative_parser.next()) |parsed| {
const c = parsed.codepoint;
switch (iterative_parser.declared_string_type) {
.wide => {
if (windows1252.bestFitFromCodepoint(c)) |best_fit| {
try buf.append(self.allocator, best_fit);
} else if (c < 0x10000 or c == code_pages.Codepoint.invalid or parsed.escaped_surrogate_pair) {
try buf.append(self.allocator, '?');
} else {
try buf.appendSlice(self.allocator, "??");
}
},
.ascii => {
if (parsed.from_escaped_integer) {
const truncated: u8 = @truncate(c);
switch (output_code_page) {
.utf8 => switch (truncated) {
0...0x7F => try buf.append(self.allocator, truncated),
else => try buf.append(self.allocator, '?'),
},
.windows1252 => {
try buf.append(self.allocator, truncated);
},
}
} else {
if (windows1252.bestFitFromCodepoint(c)) |best_fit| {
try buf.append(self.allocator, best_fit);
} else if (c < 0x10000 or c == code_pages.Codepoint.invalid) {
try buf.append(self.allocator, '?');
} else {
try buf.appendSlice(self.allocator, "??");
}
}
},
}
}
return buf.toOwnedSlice(self.allocator);
}
pub fn writeResourceExternal(self: *Compiler, node: *Node.ResourceExternal, writer: *std.Io.Writer) !void {
// Init header with data size zero for now, will need to fill it in later
var header = try self.resourceHeader(node.id, node.type, .{});
defer header.deinit(self.allocator);
const maybe_predefined_type = header.predefinedResourceType();
// DLGINCLUDE has special handling that doesn't actually need the file to exist
if (maybe_predefined_type != null and maybe_predefined_type.? == .DLGINCLUDE) {
const filename_token = node.filename.cast(.literal).?.token;
const parsed_filename = try self.parseDlgIncludeString(filename_token);
defer self.allocator.free(parsed_filename);
// NUL within the parsed string acts as a terminator
const parsed_filename_terminated = std.mem.sliceTo(parsed_filename, 0);
header.applyMemoryFlags(node.common_resource_attributes, self.source);
// This is effectively limited by `max_string_literal_codepoints` which is a u15.
// Each codepoint within a DLGINCLUDE string is encoded as a maximum of
// 2 bytes, which means that the maximum byte length of a DLGINCLUDE string is
// (including the NUL terminator): 32,767 * 2 + 1 = 65,535 or exactly the u16 max.
header.data_size = @intCast(parsed_filename_terminated.len + 1);
try header.write(writer, self.errContext(node.id));
try writer.writeAll(parsed_filename_terminated);
try writer.writeByte(0);
try writeDataPadding(writer, header.data_size);
return;
}
const filename_utf8 = try self.evaluateFilenameExpression(node.filename);
defer self.allocator.free(filename_utf8);
// TODO: More robust checking of the validity of the filename.
// This currently only checks for NUL bytes, but it should probably also check for
// platform-specific invalid characters like '*', '?', '"', '<', '>', '|' (Windows)
// Related: https://github.com/ziglang/zig/pull/14533#issuecomment-1416888193
if (std.mem.indexOfScalar(u8, filename_utf8, 0) != null) {
return self.addErrorDetailsAndFail(.{
.err = .invalid_filename,
.token = node.filename.getFirstToken(),
.token_span_end = node.filename.getLastToken(),
.extra = .{ .number = 0 },
});
}
// Allow plain number literals, but complex number expressions are evaluated strangely
// and almost certainly lead to things not intended by the user (e.g. '(1+-1)' evaluates
// to the filename '-1'), so error if the filename node is a grouped/binary expression.
// Note: This is done here instead of during parsing so that we can easily include
// the evaluated filename as part of the error messages.
if (node.filename.id != .literal) {
const filename_string_index = try self.diagnostics.putString(filename_utf8);
try self.addErrorDetails(.{
.err = .number_expression_as_filename,
.token = node.filename.getFirstToken(),
.token_span_end = node.filename.getLastToken(),
.extra = .{ .number = filename_string_index },
});
return self.addErrorDetailsAndFail(.{
.err = .number_expression_as_filename,
.type = .note,
.token = node.filename.getFirstToken(),
.token_span_end = node.filename.getLastToken(),
.print_source_line = false,
.extra = .{ .number = filename_string_index },
});
}
// From here on out, we know that the filename must be comprised of a single token,
// so get it here to simplify future usage.
const filename_token = node.filename.getFirstToken();
const file_handle = self.searchForFile(filename_utf8) catch |err| switch (err) {
error.OutOfMemory => |e| return e,
else => |e| {
const filename_string_index = try self.diagnostics.putString(filename_utf8);
return self.addErrorDetailsAndFail(.{
.err = .file_open_error,
.token = filename_token,
.extra = .{ .file_open_error = .{
.err = ErrorDetails.FileOpenError.enumFromError(e),
.filename_string_index = filename_string_index,
} },
});
},
};
defer file_handle.close();
var file_buffer: [2048]u8 = undefined;
var file_reader = file_handle.reader(&file_buffer);
if (maybe_predefined_type) |predefined_type| {
switch (predefined_type) {
.GROUP_ICON, .GROUP_CURSOR => {
// Check for animated icon first
if (ani.isAnimatedIcon(&file_reader.interface)) {
// Animated icons are just put into the resource unmodified,
// and the resource type changes to ANIICON/ANICURSOR
const new_predefined_type: res.RT = switch (predefined_type) {
.GROUP_ICON => .ANIICON,
.GROUP_CURSOR => .ANICURSOR,
else => unreachable,
};
header.type_value.ordinal = @intFromEnum(new_predefined_type);
header.memory_flags = MemoryFlags.defaults(new_predefined_type);
header.applyMemoryFlags(node.common_resource_attributes, self.source);
header.data_size = std.math.cast(u32, try file_reader.getSize()) orelse {
return self.addErrorDetailsAndFail(.{
.err = .resource_data_size_exceeds_max,
.token = node.id,
});
};
try header.write(writer, self.errContext(node.id));
try file_reader.seekTo(0);
try writeResourceData(writer, &file_reader.interface, header.data_size);
return;
}
// isAnimatedIcon moved the file cursor so reset to the start
try file_reader.seekTo(0);
const icon_dir = ico.read(self.allocator, &file_reader.interface, try file_reader.getSize()) catch |err| switch (err) {
error.OutOfMemory => |e| return e,
else => |e| {
return self.iconReadError(
e,
filename_utf8,
filename_token,
predefined_type,
);
},
};
defer icon_dir.deinit();
// This limit is inherent to the ico format since number of entries is a u16 field.
std.debug.assert(icon_dir.entries.len <= std.math.maxInt(u16));
// Note: The Win32 RC compiler will compile the resource as whatever type is
// in the icon_dir regardless of the type of resource specified in the .rc.
// This leads to unusable .res files when the types mismatch, so
// we error instead.
const res_types_match = switch (predefined_type) {
.GROUP_ICON => icon_dir.image_type == .icon,
.GROUP_CURSOR => icon_dir.image_type == .cursor,
else => unreachable,
};
if (!res_types_match) {
return self.addErrorDetailsAndFail(.{
.err = .icon_dir_and_resource_type_mismatch,
.token = filename_token,
.extra = .{ .resource = switch (predefined_type) {
.GROUP_ICON => .icon,
.GROUP_CURSOR => .cursor,
else => unreachable,
} },
});
}
// Memory flags affect the RT_ICON and the RT_GROUP_ICON differently
var icon_memory_flags = MemoryFlags.defaults(res.RT.ICON);
applyToMemoryFlags(&icon_memory_flags, node.common_resource_attributes, self.source);
applyToGroupMemoryFlags(&header.memory_flags, node.common_resource_attributes, self.source);
const first_icon_id = self.state.icon_id;
const entry_type = if (predefined_type == .GROUP_ICON) @intFromEnum(res.RT.ICON) else @intFromEnum(res.RT.CURSOR);
for (icon_dir.entries, 0..) |*entry, entry_i_usize| {
// We know that the entry index must fit within a u16, so
// cast it here to simplify usage sites.
const entry_i: u16 = @intCast(entry_i_usize);
var full_data_size = entry.data_size_in_bytes;
if (icon_dir.image_type == .cursor) {
full_data_size = std.math.add(u32, full_data_size, 4) catch {
return self.addErrorDetailsAndFail(.{
.err = .resource_data_size_exceeds_max,
.token = node.id,
});
};
}
const image_header = ResourceHeader{
.type_value = .{ .ordinal = entry_type },
.name_value = .{ .ordinal = self.state.icon_id },
.data_size = full_data_size,
.memory_flags = icon_memory_flags,
.language = self.state.language,
.version = self.state.version,
.characteristics = self.state.characteristics,
};
try image_header.write(writer, self.errContext(node.id));
// From https://learn.microsoft.com/en-us/windows/win32/menurc/localheader:
// > The LOCALHEADER structure is the first data written to the RT_CURSOR
// > resource if a RESDIR structure contains information about a cursor.
// where LOCALHEADER is `struct { WORD xHotSpot; WORD yHotSpot; }`
if (icon_dir.image_type == .cursor) {
try writer.writeInt(u16, entry.type_specific_data.cursor.hotspot_x, .little);
try writer.writeInt(u16, entry.type_specific_data.cursor.hotspot_y, .little);
}
try file_reader.seekTo(entry.data_offset_from_start_of_file);
var header_bytes: [16]u8 align(@alignOf(ico.BitmapHeader)) = (file_reader.interface.takeArray(16) catch {
return self.iconReadError(
error.UnexpectedEOF,
filename_utf8,
filename_token,
predefined_type,
);
}).*;
const image_format = ico.ImageFormat.detect(&header_bytes);
if (!image_format.validate(&header_bytes)) {
return self.iconReadError(
error.InvalidHeader,
filename_utf8,
filename_token,
predefined_type,
);
}
switch (image_format) {
.riff => switch (icon_dir.image_type) {
.icon => {
// The Win32 RC compiler treats this as an error, but icon dirs
// with RIFF encoded icons within them work ~okay (they work
// in some places but not others, they may not animate, etc) if they are
// allowed to be compiled.
try self.addErrorDetails(.{
.err = .rc_would_error_on_icon_dir,
.type = .warning,
.token = filename_token,
.extra = .{ .icon_dir = .{ .icon_type = .icon, .icon_format = .riff, .index = entry_i } },
});
try self.addErrorDetails(.{
.err = .rc_would_error_on_icon_dir,
.type = .note,
.print_source_line = false,
.token = filename_token,
.extra = .{ .icon_dir = .{ .icon_type = .icon, .icon_format = .riff, .index = entry_i } },
});
},
.cursor => {
// The Win32 RC compiler errors in this case too, but we only error
// here because the cursor would fail to be loaded at runtime if we
// compiled it.
return self.addErrorDetailsAndFail(.{
.err = .format_not_supported_in_icon_dir,
.token = filename_token,
.extra = .{ .icon_dir = .{ .icon_type = .cursor, .icon_format = .riff, .index = entry_i } },
});
},
},
.png => switch (icon_dir.image_type) {
.icon => {
// PNG always seems to have 1 for color planes no matter what
entry.type_specific_data.icon.color_planes = 1;
// These seem to be the only values of num_colors that
// get treated specially
entry.type_specific_data.icon.bits_per_pixel = switch (entry.num_colors) {
2 => 1,
8 => 3,
16 => 4,
else => entry.type_specific_data.icon.bits_per_pixel,
};
},
.cursor => {
// The Win32 RC compiler treats this as an error, but cursor dirs
// with PNG encoded icons within them work fine if they are
// allowed to be compiled.
try self.addErrorDetails(.{
.err = .rc_would_error_on_icon_dir,
.type = .warning,
.token = filename_token,
.extra = .{ .icon_dir = .{ .icon_type = .cursor, .icon_format = .png, .index = entry_i } },
});
},
},
.dib => {
const bitmap_header: *ico.BitmapHeader = @ptrCast(@alignCast(&header_bytes));
if (native_endian == .big) {
std.mem.byteSwapAllFields(ico.BitmapHeader, bitmap_header);
}
const bitmap_version = ico.BitmapHeader.Version.get(bitmap_header.bcSize);
// The Win32 RC compiler only allows headers with
// `bcSize == sizeof(BITMAPINFOHEADER)`, but it seems unlikely
// that there's a good reason for that outside of too-old
// bitmap headers.
// TODO: Need to test V4 and V5 bitmaps to check they actually work
if (bitmap_version == .@"win2.0") {
return self.addErrorDetailsAndFail(.{
.err = .rc_would_error_on_bitmap_version,
.token = filename_token,
.extra = .{ .icon_dir = .{
.icon_type = if (icon_dir.image_type == .icon) .icon else .cursor,
.icon_format = image_format,
.index = entry_i,
.bitmap_version = bitmap_version,
} },
});
} else if (bitmap_version != .@"nt3.1") {
try self.addErrorDetails(.{
.err = .rc_would_error_on_bitmap_version,
.type = .warning,
.token = filename_token,
.extra = .{ .icon_dir = .{
.icon_type = if (icon_dir.image_type == .icon) .icon else .cursor,
.icon_format = image_format,
.index = entry_i,
.bitmap_version = bitmap_version,
} },
});
}
switch (icon_dir.image_type) {
.icon => {
// The values in the icon's BITMAPINFOHEADER always take precedence over
// the values in the IconDir, but not in the LOCALHEADER (see above).
entry.type_specific_data.icon.color_planes = bitmap_header.bcPlanes;
entry.type_specific_data.icon.bits_per_pixel = bitmap_header.bcBitCount;
},
.cursor => {
// Only cursors get the width/height from BITMAPINFOHEADER (icons don't)
entry.width = @intCast(bitmap_header.bcWidth);
entry.height = @intCast(bitmap_header.bcHeight);
entry.type_specific_data.cursor.hotspot_x = bitmap_header.bcPlanes;
entry.type_specific_data.cursor.hotspot_y = bitmap_header.bcBitCount;
},
}
},
}
try file_reader.seekTo(entry.data_offset_from_start_of_file);
try writeResourceDataNoPadding(writer, &file_reader.interface, entry.data_size_in_bytes);
try writeDataPadding(writer, full_data_size);
if (self.state.icon_id == std.math.maxInt(u16)) {
try self.addErrorDetails(.{
.err = .max_icon_ids_exhausted,
.print_source_line = false,
.token = filename_token,
.extra = .{ .icon_dir = .{
.icon_type = if (icon_dir.image_type == .icon) .icon else .cursor,
.icon_format = image_format,
.index = entry_i,
} },
});
return self.addErrorDetailsAndFail(.{
.err = .max_icon_ids_exhausted,
.type = .note,
.token = filename_token,
.extra = .{ .icon_dir = .{
.icon_type = if (icon_dir.image_type == .icon) .icon else .cursor,
.icon_format = image_format,
.index = entry_i,
} },
});
}
self.state.icon_id += 1;
}
header.data_size = icon_dir.getResDataSize();
try header.write(writer, self.errContext(node.id));
try icon_dir.writeResData(writer, first_icon_id);
try writeDataPadding(writer, header.data_size);
return;
},
.RCDATA,
.HTML,
.MESSAGETABLE,
.DLGINIT,
.PLUGPLAY,
.VXD,
// Note: All of the below can only be specified by using a number
// as the resource type.
.MANIFEST,
.CURSOR,
.ICON,
.ANICURSOR,
.ANIICON,
.FONTDIR,
=> {
header.applyMemoryFlags(node.common_resource_attributes, self.source);
},
.BITMAP => {
header.applyMemoryFlags(node.common_resource_attributes, self.source);
const file_size = try file_reader.getSize();
const bitmap_info = bmp.read(&file_reader.interface, file_size) catch |err| {
const filename_string_index = try self.diagnostics.putString(filename_utf8);
return self.addErrorDetailsAndFail(.{
.err = .bmp_read_error,
.token = filename_token,
.extra = .{ .bmp_read_error = .{
.err = ErrorDetails.BitmapReadError.enumFromError(err),
.filename_string_index = filename_string_index,
} },
});
};
if (bitmap_info.getActualPaletteByteLen() > bitmap_info.getExpectedPaletteByteLen()) {
const num_ignored_bytes = bitmap_info.getActualPaletteByteLen() - bitmap_info.getExpectedPaletteByteLen();
var number_as_bytes: [8]u8 = undefined;
std.mem.writeInt(u64, &number_as_bytes, num_ignored_bytes, native_endian);
const value_string_index = try self.diagnostics.putString(&number_as_bytes);
try self.addErrorDetails(.{
.err = .bmp_ignored_palette_bytes,
.type = .warning,
.token = filename_token,
.extra = .{ .number = value_string_index },
});
} else if (bitmap_info.getActualPaletteByteLen() < bitmap_info.getExpectedPaletteByteLen()) {
const num_padding_bytes = bitmap_info.getExpectedPaletteByteLen() - bitmap_info.getActualPaletteByteLen();
var number_as_bytes: [8]u8 = undefined;
std.mem.writeInt(u64, &number_as_bytes, num_padding_bytes, native_endian);
const value_string_index = try self.diagnostics.putString(&number_as_bytes);
try self.addErrorDetails(.{
.err = .bmp_missing_palette_bytes,
.type = .err,
.token = filename_token,
.extra = .{ .number = value_string_index },
});
const pixel_data_len = bitmap_info.getPixelDataLen(file_size);
// TODO: This is a hack, but we know we have already added
// at least one entry to the diagnostics strings, so we can
// get away with using 0 to mean 'no string' here.
var miscompiled_bytes_string_index: u32 = 0;
if (pixel_data_len > 0) {
const miscompiled_bytes = @min(pixel_data_len, num_padding_bytes);
std.mem.writeInt(u64, &number_as_bytes, miscompiled_bytes, native_endian);
miscompiled_bytes_string_index = try self.diagnostics.putString(&number_as_bytes);
}
return self.addErrorDetailsAndFail(.{
.err = .rc_would_miscompile_bmp_palette_padding,
.type = .note,
.print_source_line = false,
.token = filename_token,
.extra = .{ .number = miscompiled_bytes_string_index },
});
}
// TODO: It might be possible that the calculation done in this function
// could underflow if the underlying file is modified while reading
// it, but need to think about it more to determine if that's a
// real possibility
const bmp_bytes_to_write: u32 = @intCast(bitmap_info.getExpectedByteLen(file_size));
header.data_size = bmp_bytes_to_write;
try header.write(writer, self.errContext(node.id));
try file_reader.seekTo(bmp.file_header_len);
try writeResourceDataNoPadding(writer, &file_reader.interface, bitmap_info.dib_header_size);
if (bitmap_info.getBitmasksByteLen() > 0) {
try writeResourceDataNoPadding(writer, &file_reader.interface, bitmap_info.getBitmasksByteLen());
}
if (bitmap_info.getExpectedPaletteByteLen() > 0) {
try writeResourceDataNoPadding(writer, &file_reader.interface, @intCast(bitmap_info.getActualPaletteByteLen()));
}
try file_reader.seekTo(bitmap_info.pixel_data_offset);
const pixel_bytes: u32 = @intCast(file_size - bitmap_info.pixel_data_offset);
try writeResourceDataNoPadding(writer, &file_reader.interface, pixel_bytes);
try writeDataPadding(writer, bmp_bytes_to_write);
return;
},
.FONT => {
if (self.state.font_dir.ids.get(header.name_value.ordinal) != null) {
// Add warning and skip this resource
// Note: The Win32 compiler prints this as an error but it doesn't fail the compilation
// and the duplicate resource is skipped.
try self.addErrorDetails(.{
.err = .font_id_already_defined,
.token = node.id,
.type = .warning,
.extra = .{ .number = header.name_value.ordinal },
});
try self.addErrorDetails(.{
.err = .font_id_already_defined,
.token = self.state.font_dir.ids.get(header.name_value.ordinal).?,
.type = .note,
.extra = .{ .number = header.name_value.ordinal },
});
return;
}
header.applyMemoryFlags(node.common_resource_attributes, self.source);
const file_size = try file_reader.getSize();
if (file_size > std.math.maxInt(u32)) {
return self.addErrorDetailsAndFail(.{
.err = .resource_data_size_exceeds_max,
.token = node.id,
});
}
// We now know that the data size will fit in a u32
header.data_size = @intCast(file_size);
try header.write(writer, self.errContext(node.id));
// Slurp the first 148 bytes separately so we can store them in the FontDir
var font_dir_header_buf: [148]u8 = @splat(0);
const populated_len: u32 = @intCast(try file_reader.interface.readSliceShort(&font_dir_header_buf));
// Write only the populated bytes slurped from the header
try writer.writeAll(font_dir_header_buf[0..populated_len]);
// Then write the rest of the bytes and the padding
try writeResourceDataNoPadding(writer, &file_reader.interface, header.data_size - populated_len);
try writeDataPadding(writer, header.data_size);
try self.state.font_dir.add(self.arena, FontDir.Font{
.id = header.name_value.ordinal,
.header_bytes = font_dir_header_buf,
}, node.id);
return;
},
.ACCELERATOR, // Cannot use an external file, enforced by the parser
.DIALOG, // Cannot use an external file, enforced by the parser
.DLGINCLUDE, // Handled specially above
.MENU, // Cannot use an external file, enforced by the parser
.STRING, // Parser error if this resource is specified as a number
.TOOLBAR, // Cannot use an external file, enforced by the parser
.VERSION, // Cannot use an external file, enforced by the parser
=> unreachable,
_ => unreachable,
}
} else {
header.applyMemoryFlags(node.common_resource_attributes, self.source);
}
// Fallback to just writing out the entire contents of the file
const data_size = try file_reader.getSize();
if (data_size > std.math.maxInt(u32)) {
return self.addErrorDetailsAndFail(.{
.err = .resource_data_size_exceeds_max,
.token = node.id,
});
}
// We now know that the data size will fit in a u32
header.data_size = @intCast(data_size);
try header.write(writer, self.errContext(node.id));
try writeResourceData(writer, &file_reader.interface, header.data_size);
}
fn iconReadError(
self: *Compiler,
err: ico.ReadError,
filename: []const u8,
token: Token,
predefined_type: res.RT,
) error{ CompileError, OutOfMemory } {
const filename_string_index = try self.diagnostics.putString(filename);
return self.addErrorDetailsAndFail(.{
.err = .icon_read_error,
.token = token,
.extra = .{ .icon_read_error = .{
.err = ErrorDetails.IconReadError.enumFromError(err),
.icon_type = switch (predefined_type) {
.GROUP_ICON => .icon,
.GROUP_CURSOR => .cursor,
else => unreachable,
},
.filename_string_index = filename_string_index,
} },
});
}
pub const DataType = enum {
number,
ascii_string,
wide_string,
};
pub const Data = union(DataType) {
number: Number,
ascii_string: []const u8,
wide_string: [:0]const u16,
pub fn deinit(self: Data, allocator: Allocator) void {
switch (self) {
.wide_string => |wide_string| {
allocator.free(wide_string);
},
.ascii_string => |ascii_string| {
allocator.free(ascii_string);
},
else => {},
}
}
pub fn write(self: Data, writer: *std.Io.Writer) !void {
switch (self) {
.number => |number| switch (number.is_long) {
false => try writer.writeInt(WORD, number.asWord(), .little),
true => try writer.writeInt(DWORD, number.value, .little),
},
.ascii_string => |ascii_string| {
try writer.writeAll(ascii_string);
},
.wide_string => |wide_string| {
try writer.writeAll(std.mem.sliceAsBytes(wide_string));
},
}
}
};
/// Assumes that the node is a number or number expression
pub fn evaluateNumberExpression(expression_node: *Node, source: []const u8, code_page_lookup: *const CodePageLookup) Number {
switch (expression_node.id) {
.literal => {
const literal_node = expression_node.cast(.literal).?;
std.debug.assert(literal_node.token.id == .number);
const bytes = SourceBytes{
.slice = literal_node.token.slice(source),
.code_page = code_page_lookup.getForToken(literal_node.token),
};
return literals.parseNumberLiteral(bytes);
},
.binary_expression => {
const binary_expression_node = expression_node.cast(.binary_expression).?;
const lhs = evaluateNumberExpression(binary_expression_node.left, source, code_page_lookup);
const rhs = evaluateNumberExpression(binary_expression_node.right, source, code_page_lookup);
const operator_char = binary_expression_node.operator.slice(source)[0];
return lhs.evaluateOperator(operator_char, rhs);
},
.grouped_expression => {
const grouped_expression_node = expression_node.cast(.grouped_expression).?;
return evaluateNumberExpression(grouped_expression_node.expression, source, code_page_lookup);
},
else => unreachable,
}
}
const FlagsNumber = struct {
value: u32,
not_mask: u32 = 0xFFFFFFFF,
pub fn evaluateOperator(lhs: FlagsNumber, operator_char: u8, rhs: FlagsNumber) FlagsNumber {
const result = switch (operator_char) {
'-' => lhs.value -% rhs.value,
'+' => lhs.value +% rhs.value,
'|' => lhs.value | rhs.value,
'&' => lhs.value & rhs.value,
else => unreachable, // invalid operator, this would be a lexer/parser bug
};
return .{
.value = result,
.not_mask = lhs.not_mask & rhs.not_mask,
};
}
pub fn applyNotMask(self: FlagsNumber) u32 {
return self.value & self.not_mask;
}
};
pub fn evaluateFlagsExpressionWithDefault(default: u32, expression_node: *Node, source: []const u8, code_page_lookup: *const CodePageLookup) u32 {
var context = FlagsExpressionContext{ .initial_value = default };
const number = evaluateFlagsExpression(expression_node, source, code_page_lookup, &context);
return number.value;
}
pub const FlagsExpressionContext = struct {
initial_value: u32 = 0,
initial_value_used: bool = false,
};
/// Assumes that the node is a number expression (which can contain not_expressions)
pub fn evaluateFlagsExpression(expression_node: *Node, source: []const u8, code_page_lookup: *const CodePageLookup, context: *FlagsExpressionContext) FlagsNumber {
switch (expression_node.id) {
.literal => {
const literal_node = expression_node.cast(.literal).?;
std.debug.assert(literal_node.token.id == .number);
const bytes = SourceBytes{
.slice = literal_node.token.slice(source),
.code_page = code_page_lookup.getForToken(literal_node.token),
};
var value = literals.parseNumberLiteral(bytes).value;
if (!context.initial_value_used) {
context.initial_value_used = true;
value |= context.initial_value;
}
return .{ .value = value };
},
.binary_expression => {
const binary_expression_node = expression_node.cast(.binary_expression).?;
const lhs = evaluateFlagsExpression(binary_expression_node.left, source, code_page_lookup, context);
const rhs = evaluateFlagsExpression(binary_expression_node.right, source, code_page_lookup, context);
const operator_char = binary_expression_node.operator.slice(source)[0];
const result = lhs.evaluateOperator(operator_char, rhs);
return .{ .value = result.applyNotMask() };
},
.grouped_expression => {
const grouped_expression_node = expression_node.cast(.grouped_expression).?;
return evaluateFlagsExpression(grouped_expression_node.expression, source, code_page_lookup, context);
},
.not_expression => {
const not_expression = expression_node.cast(.not_expression).?;
const bytes = SourceBytes{
.slice = not_expression.number_token.slice(source),
.code_page = code_page_lookup.getForToken(not_expression.number_token),
};
const not_number = literals.parseNumberLiteral(bytes);
if (!context.initial_value_used) {
context.initial_value_used = true;
return .{ .value = context.initial_value & ~not_number.value };
}
return .{ .value = 0, .not_mask = ~not_number.value };
},
else => unreachable,
}
}
pub fn evaluateDataExpression(self: *Compiler, expression_node: *Node) !Data {
switch (expression_node.id) {
.literal => {
const literal_node = expression_node.cast(.literal).?;
switch (literal_node.token.id) {
.number => {
const number = evaluateNumberExpression(expression_node, self.source, self.input_code_pages);
return .{ .number = number };
},
.quoted_ascii_string => {
const column = literal_node.token.calculateColumn(self.source, 8, null);
const bytes = SourceBytes{
.slice = literal_node.token.slice(self.source),
.code_page = self.input_code_pages.getForToken(literal_node.token),
};
const parsed = try literals.parseQuotedAsciiString(self.allocator, bytes, .{
.start_column = column,
.diagnostics = self.errContext(literal_node.token),
.output_code_page = self.output_code_pages.getForToken(literal_node.token),
});
errdefer self.allocator.free(parsed);
return .{ .ascii_string = parsed };
},
.quoted_wide_string => {
const column = literal_node.token.calculateColumn(self.source, 8, null);
const bytes = SourceBytes{
.slice = literal_node.token.slice(self.source),
.code_page = self.input_code_pages.getForToken(literal_node.token),
};
const parsed_string = try literals.parseQuotedWideString(self.allocator, bytes, .{
.start_column = column,
.diagnostics = self.errContext(literal_node.token),
.output_code_page = self.output_code_pages.getForToken(literal_node.token),
});
errdefer self.allocator.free(parsed_string);
return .{ .wide_string = parsed_string };
},
else => unreachable, // no other token types should be in a data literal node
}
},
.binary_expression, .grouped_expression => {
const result = evaluateNumberExpression(expression_node, self.source, self.input_code_pages);
return .{ .number = result };
},
.not_expression => unreachable,
else => unreachable,
}
}
pub fn writeResourceRawData(self: *Compiler, node: *Node.ResourceRawData, writer: *std.Io.Writer) !void {
var data_buffer: std.Io.Writer.Allocating = .init(self.allocator);
defer data_buffer.deinit();
for (node.raw_data) |expression| {
const data = try self.evaluateDataExpression(expression);
defer data.deinit(self.allocator);
try data.write(&data_buffer.writer);
}
// TODO: Limit data_buffer in some way to error when writing more than u32 max bytes
const data_len: u32 = std.math.cast(u32, data_buffer.written().len) orelse {
return self.addErrorDetailsAndFail(.{
.err = .resource_data_size_exceeds_max,
.token = node.id,
});
};
try self.writeResourceHeader(writer, node.id, node.type, data_len, node.common_resource_attributes, self.state.language);
var data_fbs: std.Io.Reader = .fixed(data_buffer.written());
try writeResourceData(writer, &data_fbs, data_len);
}
pub fn writeResourceHeader(self: *Compiler, writer: *std.Io.Writer, id_token: Token, type_token: Token, data_size: u32, common_resource_attributes: []Token, language: res.Language) !void {
var header = try self.resourceHeader(id_token, type_token, .{
.language = language,
.data_size = data_size,
});
defer header.deinit(self.allocator);
header.applyMemoryFlags(common_resource_attributes, self.source);
try header.write(writer, self.errContext(id_token));
}
pub fn writeResourceDataNoPadding(writer: *std.Io.Writer, data_reader: *std.Io.Reader, data_size: u32) !void {
try data_reader.streamExact(writer, data_size);
}
pub fn writeResourceData(writer: *std.Io.Writer, data_reader: *std.Io.Reader, data_size: u32) !void {
try writeResourceDataNoPadding(writer, data_reader, data_size);
try writeDataPadding(writer, data_size);
}
pub fn writeDataPadding(writer: *std.Io.Writer, data_size: u32) !void {
try writer.splatByteAll(0, numPaddingBytesNeeded(data_size));
}
pub fn numPaddingBytesNeeded(data_size: u32) u2 {
// Result is guaranteed to be between 0 and 3.
return @intCast((4 -% data_size) % 4);
}
pub fn evaluateAcceleratorKeyExpression(self: *Compiler, node: *Node, is_virt: bool) !u16 {
if (node.isNumberExpression()) {
return evaluateNumberExpression(node, self.source, self.input_code_pages).asWord();
} else {
std.debug.assert(node.isStringLiteral());
const literal: *Node.Literal = @alignCast(@fieldParentPtr("base", node));
const bytes = SourceBytes{
.slice = literal.token.slice(self.source),
.code_page = self.input_code_pages.getForToken(literal.token),
};
const column = literal.token.calculateColumn(self.source, 8, null);
return res.parseAcceleratorKeyString(bytes, is_virt, .{
.start_column = column,
.diagnostics = self.errContext(literal.token),
.output_code_page = self.output_code_pages.getForToken(literal.token),
});
}
}
pub fn writeAccelerators(self: *Compiler, node: *Node.Accelerators, writer: *std.Io.Writer) !void {
var data_buffer: std.Io.Writer.Allocating = .init(self.allocator);
defer data_buffer.deinit();
try self.writeAcceleratorsData(node, &data_buffer.writer);
// TODO: Limit data_buffer in some way to error when writing more than u32 max bytes
const data_size: u32 = std.math.cast(u32, data_buffer.written().len) orelse {
return self.addErrorDetailsAndFail(.{
.err = .resource_data_size_exceeds_max,
.token = node.id,
});
};
var header = try self.resourceHeader(node.id, node.type, .{
.data_size = data_size,
});
defer header.deinit(self.allocator);
header.applyMemoryFlags(node.common_resource_attributes, self.source);
header.applyOptionalStatements(node.optional_statements, self.source, self.input_code_pages);
try header.write(writer, self.errContext(node.id));
var data_fbs: std.Io.Reader = .fixed(data_buffer.written());
try writeResourceData(writer, &data_fbs, data_size);
}
/// Expects `data_writer` to be a LimitedWriter limited to u32, meaning all writes to
/// the writer within this function could return error.NoSpaceLeft
pub fn writeAcceleratorsData(self: *Compiler, node: *Node.Accelerators, data_writer: *std.Io.Writer) !void {
for (node.accelerators, 0..) |accel_node, i| {
const accelerator: *Node.Accelerator = @alignCast(@fieldParentPtr("base", accel_node));
var modifiers = res.AcceleratorModifiers{};
for (accelerator.type_and_options) |type_or_option| {
const modifier = rc.AcceleratorTypeAndOptions.map.get(type_or_option.slice(self.source)).?;
modifiers.apply(modifier);
}
if ((modifiers.isSet(.control) or modifiers.isSet(.shift)) and !modifiers.isSet(.virtkey)) {
try self.addErrorDetails(.{
.err = .accelerator_shift_or_control_without_virtkey,
.type = .warning,
// We know that one of SHIFT or CONTROL was specified, so there's at least one item
// in this list.
.token = accelerator.type_and_options[0],
.token_span_end = accelerator.type_and_options[accelerator.type_and_options.len - 1],
});
}
if (accelerator.event.isNumberExpression() and !modifiers.explicit_ascii_or_virtkey) {
return self.addErrorDetailsAndFail(.{
.err = .accelerator_type_required,
.token = accelerator.event.getFirstToken(),
.token_span_end = accelerator.event.getLastToken(),
});
}
const key = self.evaluateAcceleratorKeyExpression(accelerator.event, modifiers.isSet(.virtkey)) catch |err| switch (err) {
error.OutOfMemory => |e| return e,
else => |e| {
return self.addErrorDetailsAndFail(.{
.err = .invalid_accelerator_key,
.token = accelerator.event.getFirstToken(),
.token_span_end = accelerator.event.getLastToken(),
.extra = .{ .accelerator_error = .{
.err = ErrorDetails.AcceleratorError.enumFromError(e),
} },
});
},
};
const cmd_id = evaluateNumberExpression(accelerator.idvalue, self.source, self.input_code_pages);
if (i == node.accelerators.len - 1) {
modifiers.markLast();
}
try data_writer.writeByte(modifiers.value);
try data_writer.writeByte(0); // padding
try data_writer.writeInt(u16, key, .little);
try data_writer.writeInt(u16, cmd_id.asWord(), .little);
try data_writer.writeInt(u16, 0, .little); // padding
}
}
const DialogOptionalStatementValues = struct {
style: u32 = res.WS.SYSMENU | res.WS.BORDER | res.WS.POPUP,
exstyle: u32 = 0,
class: ?NameOrOrdinal = null,
menu: ?NameOrOrdinal = null,
font: ?FontStatementValues = null,
caption: ?Token = null,
};
pub fn writeDialog(self: *Compiler, node: *Node.Dialog, writer: *std.Io.Writer) !void {
var data_buffer: std.Io.Writer.Allocating = .init(self.allocator);
defer data_buffer.deinit();
const resource = ResourceType.fromString(.{
.slice = node.type.slice(self.source),
.code_page = self.input_code_pages.getForToken(node.type),
});
std.debug.assert(resource == .dialog or resource == .dialogex);
var optional_statement_values: DialogOptionalStatementValues = .{};
defer {
if (optional_statement_values.class) |class| {
class.deinit(self.allocator);
}
if (optional_statement_values.menu) |menu| {
menu.deinit(self.allocator);
}
}
var last_menu: *Node.SimpleStatement = undefined;
var last_class: *Node.SimpleStatement = undefined;
var last_menu_would_be_forced_ordinal = false;
var last_menu_has_digit_as_first_char = false;
var last_menu_did_uppercase = false;
var last_class_would_be_forced_ordinal = false;
for (node.optional_statements) |optional_statement| {
switch (optional_statement.id) {
.simple_statement => {
const simple_statement: *Node.SimpleStatement = @alignCast(@fieldParentPtr("base", optional_statement));
const statement_identifier = simple_statement.identifier;
const statement_type = rc.OptionalStatements.dialog_map.get(statement_identifier.slice(self.source)) orelse continue;
switch (statement_type) {
.style, .exstyle => {
const style = evaluateFlagsExpressionWithDefault(0, simple_statement.value, self.source, self.input_code_pages);
if (statement_type == .style) {
optional_statement_values.style = style;
} else {
optional_statement_values.exstyle = style;
}
},
.caption => {
std.debug.assert(simple_statement.value.id == .literal);
const literal_node: *Node.Literal = @alignCast(@fieldParentPtr("base", simple_statement.value));
optional_statement_values.caption = literal_node.token;
},
.class => {
const is_duplicate = optional_statement_values.class != null;
const forced_ordinal = is_duplicate and optional_statement_values.class.? == .ordinal;
// In the Win32 RC compiler, if any CLASS values that are interpreted as
// an ordinal exist, it affects all future CLASS statements and forces
// them to be treated as an ordinal no matter what.
if (forced_ordinal) {
last_class_would_be_forced_ordinal = true;
}
// clear out the old one if it exists
if (optional_statement_values.class) |prev| {
prev.deinit(self.allocator);
optional_statement_values.class = null;
}
if (simple_statement.value.isNumberExpression()) {
const class_ordinal = evaluateNumberExpression(simple_statement.value, self.source, self.input_code_pages);
optional_statement_values.class = NameOrOrdinal{ .ordinal = class_ordinal.asWord() };
} else {
std.debug.assert(simple_statement.value.isStringLiteral());
const literal_node: *Node.Literal = @alignCast(@fieldParentPtr("base", simple_statement.value));
const parsed = try self.parseQuotedStringAsWideString(literal_node.token);
optional_statement_values.class = NameOrOrdinal{ .name = parsed };
}
last_class = simple_statement;
},
.menu => {
const is_duplicate = optional_statement_values.menu != null;
const forced_ordinal = is_duplicate and optional_statement_values.menu.? == .ordinal;
// In the Win32 RC compiler, if any MENU values that are interpreted as
// an ordinal exist, it affects all future MENU statements and forces
// them to be treated as an ordinal no matter what.
if (forced_ordinal) {
last_menu_would_be_forced_ordinal = true;
}
// clear out the old one if it exists
if (optional_statement_values.menu) |prev| {
prev.deinit(self.allocator);
optional_statement_values.menu = null;
}
std.debug.assert(simple_statement.value.id == .literal);
const literal_node: *Node.Literal = @alignCast(@fieldParentPtr("base", simple_statement.value));
const token_slice = literal_node.token.slice(self.source);
const bytes = SourceBytes{
.slice = token_slice,
.code_page = self.input_code_pages.getForToken(literal_node.token),
};
optional_statement_values.menu = try NameOrOrdinal.fromString(self.allocator, bytes);
if (optional_statement_values.menu.? == .name) {
if (NameOrOrdinal.maybeNonAsciiOrdinalFromString(bytes)) |win32_rc_ordinal| {
try self.addErrorDetails(.{
.err = .invalid_digit_character_in_ordinal,
.type = .err,
.token = literal_node.token,
});
return self.addErrorDetailsAndFail(.{
.err = .win32_non_ascii_ordinal,
.type = .note,
.token = literal_node.token,
.print_source_line = false,
.extra = .{ .number = win32_rc_ordinal.ordinal },
});
}
}
// Need to keep track of some properties of the value
// in order to emit the appropriate warning(s) later on.
// See where the warning are emitted below (outside this loop)
// for the full explanation.
var did_uppercase = false;
var codepoint_i: usize = 0;
while (bytes.code_page.codepointAt(codepoint_i, bytes.slice)) |codepoint| : (codepoint_i += codepoint.byte_len) {
const c = codepoint.value;
switch (c) {
'a'...'z' => {
did_uppercase = true;
break;
},
else => {},
}
}
last_menu_did_uppercase = did_uppercase;
last_menu_has_digit_as_first_char = std.ascii.isDigit(token_slice[0]);
last_menu = simple_statement;
},
else => {},
}
},
.font_statement => {
const font: *Node.FontStatement = @alignCast(@fieldParentPtr("base", optional_statement));
if (optional_statement_values.font != null) {
optional_statement_values.font.?.node = font;
} else {
optional_statement_values.font = FontStatementValues{ .node = font };
}
if (font.weight) |weight| {
const value = evaluateNumberExpression(weight, self.source, self.input_code_pages);
optional_statement_values.font.?.weight = value.asWord();
}
if (font.italic) |italic| {
const value = evaluateNumberExpression(italic, self.source, self.input_code_pages);
optional_statement_values.font.?.italic = value.asWord() != 0;
}
},
else => {},
}
}
// The Win32 RC compiler miscompiles the value in the following scenario:
// Multiple CLASS parameters are specified and any of them are treated as a number, then
// the last CLASS is always treated as a number no matter what
if (last_class_would_be_forced_ordinal and optional_statement_values.class.? == .name) {
const literal_node: *Node.Literal = @alignCast(@fieldParentPtr("base", last_class.value));
const ordinal_value = res.ForcedOrdinal.fromUtf16Le(optional_statement_values.class.?.name);
try self.addErrorDetails(.{
.err = .rc_would_miscompile_dialog_class,
.type = .warning,
.token = literal_node.token,
.extra = .{ .number = ordinal_value },
});
try self.addErrorDetails(.{
.err = .rc_would_miscompile_dialog_class,
.type = .note,
.print_source_line = false,
.token = literal_node.token,
.extra = .{ .number = ordinal_value },
});
try self.addErrorDetails(.{
.err = .rc_would_miscompile_dialog_menu_or_class_id_forced_ordinal,
.type = .note,
.print_source_line = false,
.token = literal_node.token,
.extra = .{ .menu_or_class = .class },
});
}
// The Win32 RC compiler miscompiles the id in two different scenarios:
// 1. The first character of the ID is a digit, in which case it is always treated as a number
// no matter what (and therefore does not match how the MENU/MENUEX id is parsed)
// 2. Multiple MENU parameters are specified and any of them are treated as a number, then
// the last MENU is always treated as a number no matter what
if ((last_menu_would_be_forced_ordinal or last_menu_has_digit_as_first_char) and optional_statement_values.menu.? == .name) {
const literal_node: *Node.Literal = @alignCast(@fieldParentPtr("base", last_menu.value));
const token_slice = literal_node.token.slice(self.source);
const bytes = SourceBytes{
.slice = token_slice,
.code_page = self.input_code_pages.getForToken(literal_node.token),
};
const ordinal_value = res.ForcedOrdinal.fromBytes(bytes);
try self.addErrorDetails(.{
.err = .rc_would_miscompile_dialog_menu_id,
.type = .warning,
.token = literal_node.token,
.extra = .{ .number = ordinal_value },
});
try self.addErrorDetails(.{
.err = .rc_would_miscompile_dialog_menu_id,
.type = .note,
.print_source_line = false,
.token = literal_node.token,
.extra = .{ .number = ordinal_value },
});
if (last_menu_would_be_forced_ordinal) {
try self.addErrorDetails(.{
.err = .rc_would_miscompile_dialog_menu_or_class_id_forced_ordinal,
.type = .note,
.print_source_line = false,
.token = literal_node.token,
.extra = .{ .menu_or_class = .menu },
});
} else {
try self.addErrorDetails(.{
.err = .rc_would_miscompile_dialog_menu_id_starts_with_digit,
.type = .note,
.print_source_line = false,
.token = literal_node.token,
});
}
}
// The MENU id parsing uses the exact same logic as the MENU/MENUEX resource id parsing,
// which means that it will convert ASCII characters to uppercase during the 'name' parsing.
// This turns out not to matter (`LoadMenu` does a case-insensitive lookup anyway),
// but it still makes sense to share the uppercasing logic since the MENU parameter
// here is just a reference to a MENU/MENUEX id within the .exe.
// So, because this is an intentional but inconsequential-to-the-user difference
// between resinator and the Win32 RC compiler, we only emit a hint instead of
// a warning.
if (last_menu_did_uppercase) {
const literal_node: *Node.Literal = @alignCast(@fieldParentPtr("base", last_menu.value));
try self.addErrorDetails(.{
.err = .dialog_menu_id_was_uppercased,
.type = .hint,
.token = literal_node.token,
});
}
const x = evaluateNumberExpression(node.x, self.source, self.input_code_pages);
const y = evaluateNumberExpression(node.y, self.source, self.input_code_pages);
const width = evaluateNumberExpression(node.width, self.source, self.input_code_pages);
const height = evaluateNumberExpression(node.height, self.source, self.input_code_pages);
// FONT statement requires DS_SETFONT, and if it's not present DS_SETFRONT must be unset
if (optional_statement_values.font) |_| {
optional_statement_values.style |= res.DS.SETFONT;
} else {
optional_statement_values.style &= ~res.DS.SETFONT;
}
// CAPTION statement implies WS_CAPTION
if (optional_statement_values.caption) |_| {
optional_statement_values.style |= res.WS.CAPTION;
}
// NOTE: Dialog header and menu/class/title strings can never exceed u32 bytes
// on their own.
try self.writeDialogHeaderAndStrings(
node,
&data_buffer.writer,
resource,
&optional_statement_values,
x,
y,
width,
height,
);
var controls_by_id = std.AutoHashMap(u32, *const Node.ControlStatement).init(self.allocator);
// Number of controls are guaranteed by the parser to be within maxInt(u16).
try controls_by_id.ensureTotalCapacity(@as(u16, @intCast(node.controls.len)));
defer controls_by_id.deinit();
for (node.controls) |control_node| {
const control: *Node.ControlStatement = @alignCast(@fieldParentPtr("base", control_node));
try self.writeDialogControl(
control,
&data_buffer.writer,
resource,
// We know the data_buffer len is limited to u32 max.
@intCast(data_buffer.written().len),
&controls_by_id,
);
if (data_buffer.written().len > std.math.maxInt(u32)) {
try self.addErrorDetails(.{
.err = .resource_data_size_exceeds_max,
.token = node.id,
});
return self.addErrorDetailsAndFail(.{
.err = .resource_data_size_exceeds_max,
.type = .note,
.token = control.type,
});
}
}
// We know the data_buffer len is limited to u32 max.
const data_size: u32 = @intCast(data_buffer.written().len);
var header = try self.resourceHeader(node.id, node.type, .{
.data_size = data_size,
});
defer header.deinit(self.allocator);
header.applyMemoryFlags(node.common_resource_attributes, self.source);
header.applyOptionalStatements(node.optional_statements, self.source, self.input_code_pages);
try header.write(writer, self.errContext(node.id));
var data_fbs: std.Io.Reader = .fixed(data_buffer.written());
try writeResourceData(writer, &data_fbs, data_size);
}
fn writeDialogHeaderAndStrings(
self: *Compiler,
node: *Node.Dialog,
data_writer: *std.Io.Writer,
resource: ResourceType,
optional_statement_values: *const DialogOptionalStatementValues,
x: Number,
y: Number,
width: Number,
height: Number,
) !void {
// Header
if (resource == .dialogex) {
const help_id: u32 = help_id: {
if (node.help_id == null) break :help_id 0;
break :help_id evaluateNumberExpression(node.help_id.?, self.source, self.input_code_pages).value;
};
try data_writer.writeInt(u16, 1, .little); // version number, always 1
try data_writer.writeInt(u16, 0xFFFF, .little); // signature, always 0xFFFF
try data_writer.writeInt(u32, help_id, .little);
try data_writer.writeInt(u32, optional_statement_values.exstyle, .little);
try data_writer.writeInt(u32, optional_statement_values.style, .little);
} else {
try data_writer.writeInt(u32, optional_statement_values.style, .little);
try data_writer.writeInt(u32, optional_statement_values.exstyle, .little);
}
// This limit is enforced by the parser, so we know the number of controls
// is within the range of a u16.
try data_writer.writeInt(u16, @as(u16, @intCast(node.controls.len)), .little);
try data_writer.writeInt(u16, x.asWord(), .little);
try data_writer.writeInt(u16, y.asWord(), .little);
try data_writer.writeInt(u16, width.asWord(), .little);
try data_writer.writeInt(u16, height.asWord(), .little);
// Menu
if (optional_statement_values.menu) |menu| {
try menu.write(data_writer);
} else {
try data_writer.writeInt(u16, 0, .little);
}
// Class
if (optional_statement_values.class) |class| {
try class.write(data_writer);
} else {
try data_writer.writeInt(u16, 0, .little);
}
// Caption
if (optional_statement_values.caption) |caption| {
const parsed = try self.parseQuotedStringAsWideString(caption);
defer self.allocator.free(parsed);
try data_writer.writeAll(std.mem.sliceAsBytes(parsed[0 .. parsed.len + 1]));
} else {
try data_writer.writeInt(u16, 0, .little);
}
// Font
if (optional_statement_values.font) |font| {
try self.writeDialogFont(resource, font, data_writer);
}
}
fn writeDialogControl(
self: *Compiler,
control: *Node.ControlStatement,
data_writer: *std.Io.Writer,
resource: ResourceType,
bytes_written_so_far: u32,
controls_by_id: *std.AutoHashMap(u32, *const Node.ControlStatement),
) !void {
const control_type = rc.Control.map.get(control.type.slice(self.source)).?;
// Each control must be at a 4-byte boundary. However, the Windows RC
// compiler will miscompile controls if their extra data ends on an odd offset.
// We will avoid the miscompilation and emit a warning.
const num_padding = numPaddingBytesNeeded(bytes_written_so_far);
if (num_padding == 1 or num_padding == 3) {
try self.addErrorDetails(.{
.err = .rc_would_miscompile_control_padding,
.type = .warning,
.token = control.type,
});
try self.addErrorDetails(.{
.err = .rc_would_miscompile_control_padding,
.type = .note,
.print_source_line = false,
.token = control.type,
});
}
try data_writer.splatByteAll(0, num_padding);
const style = if (control.style) |style_expression|
// Certain styles are implied by the control type
evaluateFlagsExpressionWithDefault(res.ControlClass.getImpliedStyle(control_type), style_expression, self.source, self.input_code_pages)
else
res.ControlClass.getImpliedStyle(control_type);
const exstyle = if (control.exstyle) |exstyle_expression|
evaluateFlagsExpressionWithDefault(0, exstyle_expression, self.source, self.input_code_pages)
else
0;
switch (resource) {
.dialog => {
// Note: Reverse order from DIALOGEX
try data_writer.writeInt(u32, style, .little);
try data_writer.writeInt(u32, exstyle, .little);
},
.dialogex => {
const help_id: u32 = if (control.help_id) |help_id_expression|
evaluateNumberExpression(help_id_expression, self.source, self.input_code_pages).value
else
0;
try data_writer.writeInt(u32, help_id, .little);
// Note: Reverse order from DIALOG
try data_writer.writeInt(u32, exstyle, .little);
try data_writer.writeInt(u32, style, .little);
},
else => unreachable,
}
const control_x = evaluateNumberExpression(control.x, self.source, self.input_code_pages);
const control_y = evaluateNumberExpression(control.y, self.source, self.input_code_pages);
const control_width = evaluateNumberExpression(control.width, self.source, self.input_code_pages);
const control_height = evaluateNumberExpression(control.height, self.source, self.input_code_pages);
try data_writer.writeInt(u16, control_x.asWord(), .little);
try data_writer.writeInt(u16, control_y.asWord(), .little);
try data_writer.writeInt(u16, control_width.asWord(), .little);
try data_writer.writeInt(u16, control_height.asWord(), .little);
const control_id = evaluateNumberExpression(control.id, self.source, self.input_code_pages);
switch (resource) {
.dialog => try data_writer.writeInt(u16, control_id.asWord(), .little),
.dialogex => try data_writer.writeInt(u32, control_id.value, .little),
else => unreachable,
}
const control_id_for_map: u32 = switch (resource) {
.dialog => control_id.asWord(),
.dialogex => control_id.value,
else => unreachable,
};
const result = controls_by_id.getOrPutAssumeCapacity(control_id_for_map);
if (result.found_existing) {
if (!self.silent_duplicate_control_ids) {
try self.addErrorDetails(.{
.err = .control_id_already_defined,
.type = .warning,
.token = control.id.getFirstToken(),
.token_span_end = control.id.getLastToken(),
.extra = .{ .number = control_id_for_map },
});
try self.addErrorDetails(.{
.err = .control_id_already_defined,
.type = .note,
.token = result.value_ptr.*.id.getFirstToken(),
.token_span_end = result.value_ptr.*.id.getLastToken(),
.extra = .{ .number = control_id_for_map },
});
}
} else {
result.value_ptr.* = control;
}
if (res.ControlClass.fromControl(control_type)) |control_class| {
const ordinal = NameOrOrdinal{ .ordinal = @intFromEnum(control_class) };
try ordinal.write(data_writer);
} else {
const class_node = control.class.?;
if (class_node.isNumberExpression()) {
const number = evaluateNumberExpression(class_node, self.source, self.input_code_pages);
const ordinal = NameOrOrdinal{ .ordinal = number.asWord() };
// This is different from how the Windows RC compiles ordinals here,
// but I think that's a miscompilation/bug of the Windows implementation.
// The Windows behavior is (where LSB = least significant byte):
// - If the LSB is 0x00 => 0xFFFF0000
// - If the LSB is < 0x80 => 0x000000<LSB>
// - If the LSB is >= 0x80 => 0x0000FF<LSB>
//
// Because of this, we emit a warning about the potential miscompilation
try self.addErrorDetails(.{
.err = .rc_would_miscompile_control_class_ordinal,
.type = .warning,
.token = class_node.getFirstToken(),
.token_span_end = class_node.getLastToken(),
});
try self.addErrorDetails(.{
.err = .rc_would_miscompile_control_class_ordinal,
.type = .note,
.print_source_line = false,
.token = class_node.getFirstToken(),
.token_span_end = class_node.getLastToken(),
});
// And then write out the ordinal using a proper a NameOrOrdinal encoding.
try ordinal.write(data_writer);
} else if (class_node.isStringLiteral()) {
const literal_node: *Node.Literal = @alignCast(@fieldParentPtr("base", class_node));
const parsed = try self.parseQuotedStringAsWideString(literal_node.token);
defer self.allocator.free(parsed);
if (rc.ControlClass.fromWideString(parsed)) |control_class| {
const ordinal = NameOrOrdinal{ .ordinal = @intFromEnum(control_class) };
try ordinal.write(data_writer);
} else {
// NUL acts as a terminator
// TODO: Maybe warn when parsed_terminated.len != parsed.len, since
// it seems unlikely that NUL-termination is something intentional
const parsed_terminated = std.mem.sliceTo(parsed, 0);
const name = NameOrOrdinal{ .name = parsed_terminated };
try name.write(data_writer);
}
} else {
const literal_node: *Node.Literal = @alignCast(@fieldParentPtr("base", class_node));
const literal_slice = literal_node.token.slice(self.source);
// This succeeding is guaranteed by the parser
const control_class = rc.ControlClass.map.get(literal_slice) orelse unreachable;
const ordinal = NameOrOrdinal{ .ordinal = @intFromEnum(control_class) };
try ordinal.write(data_writer);
}
}
if (control.text) |text_token| {
const bytes = SourceBytes{
.slice = text_token.slice(self.source),
.code_page = self.input_code_pages.getForToken(text_token),
};
if (text_token.isStringLiteral()) {
const text = try self.parseQuotedStringAsWideString(text_token);
defer self.allocator.free(text);
const name = NameOrOrdinal{ .name = text };
try name.write(data_writer);
} else {
std.debug.assert(text_token.id == .number);
const number = literals.parseNumberLiteral(bytes);
const ordinal = NameOrOrdinal{ .ordinal = number.asWord() };
try ordinal.write(data_writer);
}
} else {
try NameOrOrdinal.writeEmpty(data_writer);
}
// The extra data byte length must be able to fit within a u16.
var extra_data_buf: std.Io.Writer.Allocating = .init(self.allocator);
defer extra_data_buf.deinit();
for (control.extra_data) |data_expression| {
const data = try self.evaluateDataExpression(data_expression);
defer data.deinit(self.allocator);
try data.write(&extra_data_buf.writer);
if (extra_data_buf.written().len > std.math.maxInt(u16)) {
try self.addErrorDetails(.{
.err = .control_extra_data_size_exceeds_max,
.token = control.type,
});
return self.addErrorDetailsAndFail(.{
.err = .control_extra_data_size_exceeds_max,
.type = .note,
.token = data_expression.getFirstToken(),
.token_span_end = data_expression.getLastToken(),
});
}
}
// We know the extra_data_buf size fits within a u16.
const extra_data_size: u16 = @intCast(extra_data_buf.written().len);
try data_writer.writeInt(u16, extra_data_size, .little);
try data_writer.writeAll(extra_data_buf.written());
}
pub fn writeToolbar(self: *Compiler, node: *Node.Toolbar, writer: *std.Io.Writer) !void {
var data_buffer: std.Io.Writer.Allocating = .init(self.allocator);
defer data_buffer.deinit();
const data_writer = &data_buffer.writer;
const button_width = evaluateNumberExpression(node.button_width, self.source, self.input_code_pages);
const button_height = evaluateNumberExpression(node.button_height, self.source, self.input_code_pages);
// I'm assuming this is some sort of version
// TODO: Try to find something mentioning this
try data_writer.writeInt(u16, 1, .little);
try data_writer.writeInt(u16, button_width.asWord(), .little);
try data_writer.writeInt(u16, button_height.asWord(), .little);
// Number of buttons is guaranteed by the parser to be within maxInt(u16).
try data_writer.writeInt(u16, @as(u16, @intCast(node.buttons.len)), .little);
for (node.buttons) |button_or_sep| {
switch (button_or_sep.id) {
.literal => { // This is always SEPARATOR
std.debug.assert(button_or_sep.cast(.literal).?.token.id == .literal);
try data_writer.writeInt(u16, 0, .little);
},
.simple_statement => {
const value_node = button_or_sep.cast(.simple_statement).?.value;
const value = evaluateNumberExpression(value_node, self.source, self.input_code_pages);
try data_writer.writeInt(u16, value.asWord(), .little);
},
else => unreachable, // This is a bug in the parser
}
}
const data_size: u32 = @intCast(data_buffer.written().len);
var header = try self.resourceHeader(node.id, node.type, .{
.data_size = data_size,
});
defer header.deinit(self.allocator);
header.applyMemoryFlags(node.common_resource_attributes, self.source);
try header.write(writer, self.errContext(node.id));
var data_fbs: std.Io.Reader = .fixed(data_buffer.written());
try writeResourceData(writer, &data_fbs, data_size);
}
/// Weight and italic carry over from previous FONT statements within a single resource,
/// so they need to be parsed ahead-of-time and stored
const FontStatementValues = struct {
weight: u16 = 0,
italic: bool = false,
node: *Node.FontStatement,
};
pub fn writeDialogFont(self: *Compiler, resource: ResourceType, values: FontStatementValues, writer: *std.Io.Writer) !void {
const node = values.node;
const point_size = evaluateNumberExpression(node.point_size, self.source, self.input_code_pages);
try writer.writeInt(u16, point_size.asWord(), .little);
if (resource == .dialogex) {
try writer.writeInt(u16, values.weight, .little);
}
if (resource == .dialogex) {
try writer.writeInt(u8, @intFromBool(values.italic), .little);
}
if (node.char_set) |char_set| {
const value = evaluateNumberExpression(char_set, self.source, self.input_code_pages);
try writer.writeInt(u8, @as(u8, @truncate(value.value)), .little);
} else if (resource == .dialogex) {
try writer.writeInt(u8, 1, .little); // DEFAULT_CHARSET
}
const typeface = try self.parseQuotedStringAsWideString(node.typeface);
defer self.allocator.free(typeface);
try writer.writeAll(std.mem.sliceAsBytes(typeface[0 .. typeface.len + 1]));
}
pub fn writeMenu(self: *Compiler, node: *Node.Menu, writer: *std.Io.Writer) !void {
var data_buffer: std.Io.Writer.Allocating = .init(self.allocator);
defer data_buffer.deinit();
const type_bytes = SourceBytes{
.slice = node.type.slice(self.source),
.code_page = self.input_code_pages.getForToken(node.type),
};
const resource = ResourceType.fromString(type_bytes);
std.debug.assert(resource == .menu or resource == .menuex);
try self.writeMenuData(node, &data_buffer.writer, resource);
// TODO: Limit data_buffer in some way to error when writing more than u32 max bytes
const data_size: u32 = std.math.cast(u32, data_buffer.written().len) orelse {
return self.addErrorDetailsAndFail(.{
.err = .resource_data_size_exceeds_max,
.token = node.id,
});
};
var header = try self.resourceHeader(node.id, node.type, .{
.data_size = data_size,
});
defer header.deinit(self.allocator);
header.applyMemoryFlags(node.common_resource_attributes, self.source);
header.applyOptionalStatements(node.optional_statements, self.source, self.input_code_pages);
try header.write(writer, self.errContext(node.id));
var data_fbs: std.Io.Reader = .fixed(data_buffer.written());
try writeResourceData(writer, &data_fbs, data_size);
}
/// Expects `data_writer` to be a LimitedWriter limited to u32, meaning all writes to
/// the writer within this function could return error.NoSpaceLeft
pub fn writeMenuData(self: *Compiler, node: *Node.Menu, data_writer: *std.Io.Writer, resource: ResourceType) !void {
// menu header
const version: u16 = if (resource == .menu) 0 else 1;
try data_writer.writeInt(u16, version, .little);
const header_size: u16 = if (resource == .menu) 0 else 4;
try data_writer.writeInt(u16, header_size, .little); // cbHeaderSize
// Note: There can be extra bytes at the end of this header (`rgbExtra`),
// but they are always zero-length for us, so we don't write anything
// (the length of the rgbExtra field is inferred from the header_size).
// MENU => rgbExtra: [cbHeaderSize]u8
// MENUEX => rgbExtra: [cbHeaderSize-4]u8
if (resource == .menuex) {
if (node.help_id) |help_id_node| {
const help_id = evaluateNumberExpression(help_id_node, self.source, self.input_code_pages);
try data_writer.writeInt(u32, help_id.value, .little);
} else {
try data_writer.writeInt(u32, 0, .little);
}
}
for (node.items, 0..) |item, i| {
const is_last = i == node.items.len - 1;
try self.writeMenuItem(item, data_writer, is_last);
}
}
pub fn writeMenuItem(self: *Compiler, node: *Node, writer: *std.Io.Writer, is_last_of_parent: bool) !void {
switch (node.id) {
.menu_item_separator => {
// This is the 'alternate compability form' of the separator, see
// https://devblogs.microsoft.com/oldnewthing/20080710-00/?p=21673
//
// The 'correct' way is to set the MF_SEPARATOR flag, but the Win32 RC
// compiler still uses this alternate form, so that's what we use too.
var flags = res.MenuItemFlags{};
if (is_last_of_parent) flags.markLast();
try writer.writeInt(u16, flags.value, .little);
try writer.writeInt(u16, 0, .little); // id
try writer.writeInt(u16, 0, .little); // null-terminated UTF-16 text
},
.menu_item => {
const menu_item: *Node.MenuItem = @alignCast(@fieldParentPtr("base", node));
var flags = res.MenuItemFlags{};
for (menu_item.option_list) |option_token| {
// This failing would be a bug in the parser
const option = rc.MenuItem.Option.map.get(option_token.slice(self.source)) orelse unreachable;
flags.apply(option);
}
if (is_last_of_parent) flags.markLast();
try writer.writeInt(u16, flags.value, .little);
var result = evaluateNumberExpression(menu_item.result, self.source, self.input_code_pages);
try writer.writeInt(u16, result.asWord(), .little);
var text = try self.parseQuotedStringAsWideString(menu_item.text);
defer self.allocator.free(text);
try writer.writeAll(std.mem.sliceAsBytes(text[0 .. text.len + 1]));
},
.popup => {
const popup: *Node.Popup = @alignCast(@fieldParentPtr("base", node));
var flags = res.MenuItemFlags{ .value = res.MF.POPUP };
for (popup.option_list) |option_token| {
// This failing would be a bug in the parser
const option = rc.MenuItem.Option.map.get(option_token.slice(self.source)) orelse unreachable;
flags.apply(option);
}
if (is_last_of_parent) flags.markLast();
try writer.writeInt(u16, flags.value, .little);
var text = try self.parseQuotedStringAsWideString(popup.text);
defer self.allocator.free(text);
try writer.writeAll(std.mem.sliceAsBytes(text[0 .. text.len + 1]));
for (popup.items, 0..) |item, i| {
const is_last = i == popup.items.len - 1;
try self.writeMenuItem(item, writer, is_last);
}
},
inline .menu_item_ex, .popup_ex => |node_type| {
const menu_item: *node_type.Type() = @alignCast(@fieldParentPtr("base", node));
if (menu_item.type) |flags| {
const value = evaluateNumberExpression(flags, self.source, self.input_code_pages);
try writer.writeInt(u32, value.value, .little);
} else {
try writer.writeInt(u32, 0, .little);
}
if (menu_item.state) |state| {
const value = evaluateNumberExpression(state, self.source, self.input_code_pages);
try writer.writeInt(u32, value.value, .little);
} else {
try writer.writeInt(u32, 0, .little);
}
if (menu_item.id) |id| {
const value = evaluateNumberExpression(id, self.source, self.input_code_pages);
try writer.writeInt(u32, value.value, .little);
} else {
try writer.writeInt(u32, 0, .little);
}
var flags: u16 = 0;
if (is_last_of_parent) flags |= comptime @as(u16, @intCast(res.MF.END));
// This constant doesn't seem to have a named #define, it's different than MF_POPUP
if (node_type == .popup_ex) flags |= 0x01;
try writer.writeInt(u16, flags, .little);
var text = try self.parseQuotedStringAsWideString(menu_item.text);
defer self.allocator.free(text);
try writer.writeAll(std.mem.sliceAsBytes(text[0 .. text.len + 1]));
// Only the combination of the flags u16 and the text bytes can cause
// non-DWORD alignment, so we can just use the byte length of those
// two values to realign to DWORD alignment.
const relevant_bytes = 2 + (text.len + 1) * 2;
try writeDataPadding(writer, @intCast(relevant_bytes));
if (node_type == .popup_ex) {
if (menu_item.help_id) |help_id_node| {
const help_id = evaluateNumberExpression(help_id_node, self.source, self.input_code_pages);
try writer.writeInt(u32, help_id.value, .little);
} else {
try writer.writeInt(u32, 0, .little);
}
for (menu_item.items, 0..) |item, i| {
const is_last = i == menu_item.items.len - 1;
try self.writeMenuItem(item, writer, is_last);
}
}
},
else => unreachable,
}
}
pub fn writeVersionInfo(self: *Compiler, node: *Node.VersionInfo, writer: *std.Io.Writer) !void {
// NOTE: The node's length field (which is inclusive of the length of all of its children) is a u16
var data_buffer: std.Io.Writer.Allocating = .init(self.allocator);
defer data_buffer.deinit();
const data_writer = &data_buffer.writer;
try data_writer.writeInt(u16, 0, .little); // placeholder size
try data_writer.writeInt(u16, res.FixedFileInfo.byte_len, .little);
try data_writer.writeInt(u16, res.VersionNode.type_binary, .little);
const key_bytes = std.mem.sliceAsBytes(res.FixedFileInfo.key[0 .. res.FixedFileInfo.key.len + 1]);
try data_writer.writeAll(key_bytes);
// The number of bytes written up to this point is always the same, since the name
// of the node is a constant (FixedFileInfo.key). The total number of bytes
// written so far is 38, so we need 2 padding bytes to get back to DWORD alignment
try data_writer.writeInt(u16, 0, .little);
var fixed_file_info = res.FixedFileInfo{};
for (node.fixed_info) |fixed_info| {
switch (fixed_info.id) {
.version_statement => {
const version_statement: *Node.VersionStatement = @alignCast(@fieldParentPtr("base", fixed_info));
const version_type = rc.VersionInfo.map.get(version_statement.type.slice(self.source)).?;
// Ensure that all parts are cleared for each version, to properly account for
// potential duplicate PRODUCTVERSION/FILEVERSION statements
switch (version_type) {
.file_version => @memset(&fixed_file_info.file_version.parts, 0),
.product_version => @memset(&fixed_file_info.product_version.parts, 0),
else => unreachable,
}
for (version_statement.parts, 0..) |part, i| {
const part_value = evaluateNumberExpression(part, self.source, self.input_code_pages);
if (part_value.is_long) {
try self.addErrorDetails(.{
.err = .rc_would_error_u16_with_l_suffix,
.type = .warning,
.token = part.getFirstToken(),
.token_span_end = part.getLastToken(),
.extra = .{ .statement_with_u16_param = switch (version_type) {
.file_version => .fileversion,
.product_version => .productversion,
else => unreachable,
} },
});
try self.addErrorDetails(.{
.err = .rc_would_error_u16_with_l_suffix,
.print_source_line = false,
.type = .note,
.token = part.getFirstToken(),
.token_span_end = part.getLastToken(),
.extra = .{ .statement_with_u16_param = switch (version_type) {
.file_version => .fileversion,
.product_version => .productversion,
else => unreachable,
} },
});
}
switch (version_type) {
.file_version => {
fixed_file_info.file_version.parts[i] = part_value.asWord();
},
.product_version => {
fixed_file_info.product_version.parts[i] = part_value.asWord();
},
else => unreachable,
}
}
},
.simple_statement => {
const statement: *Node.SimpleStatement = @alignCast(@fieldParentPtr("base", fixed_info));
const statement_type = rc.VersionInfo.map.get(statement.identifier.slice(self.source)).?;
const value = evaluateNumberExpression(statement.value, self.source, self.input_code_pages);
switch (statement_type) {
.file_flags_mask => fixed_file_info.file_flags_mask = value.value,
.file_flags => fixed_file_info.file_flags = value.value,
.file_os => fixed_file_info.file_os = value.value,
.file_type => fixed_file_info.file_type = value.value,
.file_subtype => fixed_file_info.file_subtype = value.value,
else => unreachable,
}
},
else => unreachable,
}
}
try fixed_file_info.write(data_writer);
for (node.block_statements) |statement| {
var overflow = false;
self.writeVersionNode(statement, data_writer) catch |err| switch (err) {
error.NoSpaceLeft => {
overflow = true;
},
else => |e| return e,
};
if (overflow or data_buffer.written().len > std.math.maxInt(u16)) {
try self.addErrorDetails(.{
.err = .version_node_size_exceeds_max,
.token = node.id,
});
return self.addErrorDetailsAndFail(.{
.err = .version_node_size_exceeds_max,
.type = .note,
.token = statement.getFirstToken(),
.token_span_end = statement.getLastToken(),
});
}
}
// We know that data_buffer len is within the limits of a u16, since we check in the block
// statements loop above which is the only place it can overflow.
const data_size: u16 = @intCast(data_buffer.written().len);
// And now that we know the full size of this node (including its children), set its size
std.mem.writeInt(u16, data_buffer.written()[0..2], data_size, .little);
var header = try self.resourceHeader(node.id, node.versioninfo, .{
.data_size = data_size,
});
defer header.deinit(self.allocator);
header.applyMemoryFlags(node.common_resource_attributes, self.source);
try header.write(writer, self.errContext(node.id));
var data_fbs: std.Io.Reader = .fixed(data_buffer.written());
try writeResourceData(writer, &data_fbs, data_size);
}
/// Assumes that writer is Writer.Allocating (specifically, that buffered() gets the entire data)
/// TODO: This function could be nicer if writer was guaranteed to fail if it wrote more than u16 max bytes
pub fn writeVersionNode(self: *Compiler, node: *Node, writer: *std.Io.Writer) !void {
// We can assume that buf.items.len will never be able to exceed the limits of a u16
try writeDataPadding(writer, std.math.cast(u16, writer.buffered().len) orelse return error.NoSpaceLeft);
const node_and_children_size_offset = writer.buffered().len;
try writer.writeInt(u16, 0, .little); // placeholder for size
const data_size_offset = writer.buffered().len;
try writer.writeInt(u16, 0, .little); // placeholder for data size
const data_type_offset = writer.buffered().len;
// Data type is string unless the node contains values that are numbers.
try writer.writeInt(u16, res.VersionNode.type_string, .little);
switch (node.id) {
inline .block, .block_value => |node_type| {
const block_or_value: *node_type.Type() = @alignCast(@fieldParentPtr("base", node));
const parsed_key = try self.parseQuotedStringAsWideString(block_or_value.key);
defer self.allocator.free(parsed_key);
const parsed_key_to_first_null = std.mem.sliceTo(parsed_key, 0);
try writer.writeAll(std.mem.sliceAsBytes(parsed_key_to_first_null[0 .. parsed_key_to_first_null.len + 1]));
var has_number_value: bool = false;
for (block_or_value.values) |value_value_node_uncasted| {
const value_value_node = value_value_node_uncasted.cast(.block_value_value).?;
if (value_value_node.expression.isNumberExpression()) {
has_number_value = true;
break;
}
}
// The units used here are dependent on the type. If there are any numbers, then
// this is a byte count. If there are only strings, then this is a count of
// UTF-16 code units.
//
// The Win32 RC compiler miscompiles this count in the case of values that
// have a mix of numbers and strings. This is detected and a warning is emitted
// during parsing, so we can just do the correct thing here.
var values_size: usize = 0;
try writeDataPadding(writer, std.math.cast(u16, writer.buffered().len) orelse return error.NoSpaceLeft);
for (block_or_value.values, 0..) |value_value_node_uncasted, i| {
const value_value_node = value_value_node_uncasted.cast(.block_value_value).?;
const value_node = value_value_node.expression;
if (value_node.isNumberExpression()) {
const number = evaluateNumberExpression(value_node, self.source, self.input_code_pages);
// This is used to write u16 or u32 depending on the number's suffix
const data_wrapper = Data{ .number = number };
try data_wrapper.write(writer);
// Numbers use byte count
values_size += if (number.is_long) 4 else 2;
} else {
std.debug.assert(value_node.isStringLiteral());
const literal_node = value_node.cast(.literal).?;
const parsed_value = try self.parseQuotedStringAsWideString(literal_node.token);
defer self.allocator.free(parsed_value);
const parsed_to_first_null = std.mem.sliceTo(parsed_value, 0);
try writer.writeAll(std.mem.sliceAsBytes(parsed_to_first_null));
// Strings use UTF-16 code-unit count including the null-terminator, but
// only if there are no number values in the list.
var value_size = parsed_to_first_null.len;
if (has_number_value) value_size *= 2; // 2 bytes per UTF-16 code unit
values_size += value_size;
// The null-terminator is only included if there's a trailing comma
// or this is the last value. If the value evaluates to empty, then
// it never gets a null terminator. If there was an explicit null-terminator
// in the string, we still need to potentially add one since we already
// sliced to the terminator.
const is_last = i == block_or_value.values.len - 1;
const is_empty = parsed_to_first_null.len == 0;
const is_only = block_or_value.values.len == 1;
if ((!is_empty or !is_only) and (is_last or value_value_node.trailing_comma)) {
try writer.writeInt(u16, 0, .little);
values_size += if (has_number_value) 2 else 1;
}
}
}
var data_size_slice = writer.buffered()[data_size_offset..];
std.mem.writeInt(u16, data_size_slice[0..@sizeOf(u16)], @as(u16, @intCast(values_size)), .little);
if (has_number_value) {
const data_type_slice = writer.buffered()[data_type_offset..];
std.mem.writeInt(u16, data_type_slice[0..@sizeOf(u16)], res.VersionNode.type_binary, .little);
}
if (node_type == .block) {
const block = block_or_value;
for (block.children) |child| {
try self.writeVersionNode(child, writer);
}
}
},
else => unreachable,
}
const node_and_children_size = writer.buffered().len - node_and_children_size_offset;
const node_and_children_size_slice = writer.buffered()[node_and_children_size_offset..];
std.mem.writeInt(u16, node_and_children_size_slice[0..@sizeOf(u16)], @as(u16, @intCast(node_and_children_size)), .little);
}
pub fn writeStringTable(self: *Compiler, node: *Node.StringTable) !void {
const language = getLanguageFromOptionalStatements(node.optional_statements, self.source, self.input_code_pages) orelse self.state.language;
for (node.strings) |string_node| {
const string: *Node.StringTableString = @alignCast(@fieldParentPtr("base", string_node));
const string_id_data = try self.evaluateDataExpression(string.id);
const string_id = string_id_data.number.asWord();
self.state.string_tables.set(
self.arena,
language,
string_id,
string.string,
&node.base,
self.source,
self.input_code_pages,
self.state.version,
self.state.characteristics,
) catch |err| switch (err) {
error.StringAlreadyDefined => {
// It might be nice to have these errors point to the ids rather than the
// string tokens, but that would mean storing the id token of each string
// which doesn't seem worth it just for slightly better error messages.
try self.addErrorDetails(.{
.err = .string_already_defined,
.token = string.string,
.extra = .{ .string_and_language = .{ .id = string_id, .language = language } },
});
const existing_def_table = self.state.string_tables.tables.getPtr(language).?;
const existing_definition = existing_def_table.get(string_id).?;
return self.addErrorDetailsAndFail(.{
.err = .string_already_defined,
.type = .note,
.token = existing_definition,
.extra = .{ .string_and_language = .{ .id = string_id, .language = language } },
});
},
error.OutOfMemory => |e| return e,
};
}
}
/// Expects this to be a top-level LANGUAGE statement
pub fn writeLanguageStatement(self: *Compiler, node: *Node.LanguageStatement) void {
const primary = Compiler.evaluateNumberExpression(node.primary_language_id, self.source, self.input_code_pages);
const sublanguage = Compiler.evaluateNumberExpression(node.sublanguage_id, self.source, self.input_code_pages);
self.state.language.primary_language_id = @truncate(primary.value);
self.state.language.sublanguage_id = @truncate(sublanguage.value);
}
/// Expects this to be a top-level VERSION or CHARACTERISTICS statement
pub fn writeTopLevelSimpleStatement(self: *Compiler, node: *Node.SimpleStatement) void {
const value = Compiler.evaluateNumberExpression(node.value, self.source, self.input_code_pages);
const statement_type = rc.TopLevelKeywords.map.get(node.identifier.slice(self.source)).?;
switch (statement_type) {
.characteristics => self.state.characteristics = value.value,
.version => self.state.version = value.value,
else => unreachable,
}
}
pub const ResourceHeaderOptions = struct {
language: ?res.Language = null,
data_size: DWORD = 0,
};
pub fn resourceHeader(self: *Compiler, id_token: Token, type_token: Token, options: ResourceHeaderOptions) !ResourceHeader {
const id_bytes = self.sourceBytesForToken(id_token);
const type_bytes = self.sourceBytesForToken(type_token);
return ResourceHeader.init(
self.allocator,
id_bytes,
type_bytes,
options.data_size,
options.language orelse self.state.language,
self.state.version,
self.state.characteristics,
) catch |err| switch (err) {
error.OutOfMemory => |e| return e,
error.TypeNonAsciiOrdinal => {
const win32_rc_ordinal = NameOrOrdinal.maybeNonAsciiOrdinalFromString(type_bytes).?;
try self.addErrorDetails(.{
.err = .invalid_digit_character_in_ordinal,
.type = .err,
.token = type_token,
});
return self.addErrorDetailsAndFail(.{
.err = .win32_non_ascii_ordinal,
.type = .note,
.token = type_token,
.print_source_line = false,
.extra = .{ .number = win32_rc_ordinal.ordinal },
});
},
error.IdNonAsciiOrdinal => {
const win32_rc_ordinal = NameOrOrdinal.maybeNonAsciiOrdinalFromString(id_bytes).?;
try self.addErrorDetails(.{
.err = .invalid_digit_character_in_ordinal,
.type = .err,
.token = id_token,
});
return self.addErrorDetailsAndFail(.{
.err = .win32_non_ascii_ordinal,
.type = .note,
.token = id_token,
.print_source_line = false,
.extra = .{ .number = win32_rc_ordinal.ordinal },
});
},
};
}
pub const ResourceHeader = struct {
name_value: NameOrOrdinal,
type_value: NameOrOrdinal,
language: res.Language,
memory_flags: MemoryFlags,
data_size: DWORD,
version: DWORD,
characteristics: DWORD,
data_version: DWORD = 0,
pub const InitError = error{ OutOfMemory, IdNonAsciiOrdinal, TypeNonAsciiOrdinal };
pub fn init(allocator: Allocator, id_bytes: SourceBytes, type_bytes: SourceBytes, data_size: DWORD, language: res.Language, version: DWORD, characteristics: DWORD) InitError!ResourceHeader {
const type_value = type: {
const resource_type = ResourceType.fromString(type_bytes);
if (res.RT.fromResource(resource_type)) |rt_constant| {
break :type NameOrOrdinal{ .ordinal = @intFromEnum(rt_constant) };
} else {
break :type try NameOrOrdinal.fromString(allocator, type_bytes);
}
};
errdefer type_value.deinit(allocator);
if (type_value == .name) {
if (NameOrOrdinal.maybeNonAsciiOrdinalFromString(type_bytes)) |_| {
return error.TypeNonAsciiOrdinal;
}
}
const name_value = try NameOrOrdinal.fromString(allocator, id_bytes);
errdefer name_value.deinit(allocator);
if (name_value == .name) {
if (NameOrOrdinal.maybeNonAsciiOrdinalFromString(id_bytes)) |_| {
return error.IdNonAsciiOrdinal;
}
}
const predefined_resource_type = type_value.predefinedResourceType();
return ResourceHeader{
.name_value = name_value,
.type_value = type_value,
.data_size = data_size,
.memory_flags = MemoryFlags.defaults(predefined_resource_type),
.language = language,
.version = version,
.characteristics = characteristics,
};
}
pub fn deinit(self: ResourceHeader, allocator: Allocator) void {
self.name_value.deinit(allocator);
self.type_value.deinit(allocator);
}
pub const SizeInfo = struct {
bytes: u32,
padding_after_name: u2,
};
pub fn calcSize(self: ResourceHeader) error{Overflow}!SizeInfo {
var header_size: u32 = 8;
header_size = try std.math.add(
u32,
header_size,
std.math.cast(u32, self.name_value.byteLen()) orelse return error.Overflow,
);
header_size = try std.math.add(
u32,
header_size,
std.math.cast(u32, self.type_value.byteLen()) orelse return error.Overflow,
);
const padding_after_name = numPaddingBytesNeeded(header_size);
header_size = try std.math.add(u32, header_size, padding_after_name);
header_size = try std.math.add(u32, header_size, 16);
return .{ .bytes = header_size, .padding_after_name = padding_after_name };
}
pub fn writeAssertNoOverflow(self: ResourceHeader, writer: *std.Io.Writer) !void {
return self.writeSizeInfo(writer, self.calcSize() catch unreachable);
}
pub fn write(self: ResourceHeader, writer: *std.Io.Writer, err_ctx: errors.DiagnosticsContext) !void {
const size_info = self.calcSize() catch {
try err_ctx.diagnostics.append(.{
.err = .resource_data_size_exceeds_max,
.code_page = err_ctx.code_page,
.token = err_ctx.token,
});
return error.CompileError;
};
return self.writeSizeInfo(writer, size_info);
}
pub fn writeSizeInfo(self: ResourceHeader, writer: *std.Io.Writer, size_info: SizeInfo) !void {
try writer.writeInt(DWORD, self.data_size, .little); // DataSize
try writer.writeInt(DWORD, size_info.bytes, .little); // HeaderSize
try self.type_value.write(writer); // TYPE
try self.name_value.write(writer); // NAME
try writer.splatByteAll(0, size_info.padding_after_name);
try writer.writeInt(DWORD, self.data_version, .little); // DataVersion
try writer.writeInt(WORD, self.memory_flags.value, .little); // MemoryFlags
try writer.writeInt(WORD, self.language.asInt(), .little); // LanguageId
try writer.writeInt(DWORD, self.version, .little); // Version
try writer.writeInt(DWORD, self.characteristics, .little); // Characteristics
}
pub fn predefinedResourceType(self: ResourceHeader) ?res.RT {
return self.type_value.predefinedResourceType();
}
pub fn applyMemoryFlags(self: *ResourceHeader, tokens: []Token, source: []const u8) void {
applyToMemoryFlags(&self.memory_flags, tokens, source);
}
pub fn applyOptionalStatements(self: *ResourceHeader, statements: []*Node, source: []const u8, code_page_lookup: *const CodePageLookup) void {
applyToOptionalStatements(&self.language, &self.version, &self.characteristics, statements, source, code_page_lookup);
}
};
fn applyToMemoryFlags(flags: *MemoryFlags, tokens: []Token, source: []const u8) void {
for (tokens) |token| {
const attribute = rc.CommonResourceAttributes.map.get(token.slice(source)).?;
flags.set(attribute);
}
}
/// RT_GROUP_ICON and RT_GROUP_CURSOR have their own special rules for memory flags
fn applyToGroupMemoryFlags(flags: *MemoryFlags, tokens: []Token, source: []const u8) void {
// There's probably a cleaner implementation of this, but this will result in the same
// flags as the Win32 RC compiler for all 986,410 K-permutations of memory flags
// for an ICON resource.
//
// This was arrived at by iterating over the permutations and creating a
// list where each line looks something like this:
// MOVEABLE PRELOAD -> 0x1050 (MOVEABLE|PRELOAD|DISCARDABLE)
//
// and then noticing a few things:
// 1. Any permutation that does not have PRELOAD in it just uses the
// default flags.
const initial_flags = flags.*;
var flags_set = std.enums.EnumSet(rc.CommonResourceAttributes).initEmpty();
for (tokens) |token| {
const attribute = rc.CommonResourceAttributes.map.get(token.slice(source)).?;
flags_set.insert(attribute);
}
if (!flags_set.contains(.preload)) return;
// 2. Any permutation of flags where applying only the PRELOAD and LOADONCALL flags
// results in no actual change by the end will just use the default flags.
// For example, `PRELOAD LOADONCALL` will result in default flags, but
// `LOADONCALL PRELOAD` will have PRELOAD set after they are both applied in order.
for (tokens) |token| {
const attribute = rc.CommonResourceAttributes.map.get(token.slice(source)).?;
switch (attribute) {
.preload, .loadoncall => flags.set(attribute),
else => {},
}
}
if (flags.value == initial_flags.value) return;
// 3. If none of DISCARDABLE, SHARED, or PURE is specified, then PRELOAD
// implies `flags &= ~SHARED` and LOADONCALL implies `flags |= SHARED`
const shared_set = comptime blk: {
var set = std.enums.EnumSet(rc.CommonResourceAttributes).initEmpty();
set.insert(.discardable);
set.insert(.shared);
set.insert(.pure);
break :blk set;
};
const discardable_shared_or_pure_specified = flags_set.intersectWith(shared_set).count() != 0;
for (tokens) |token| {
const attribute = rc.CommonResourceAttributes.map.get(token.slice(source)).?;
flags.setGroup(attribute, !discardable_shared_or_pure_specified);
}
}
/// Only handles the 'base' optional statements that are shared between resource types.
fn applyToOptionalStatements(language: *res.Language, version: *u32, characteristics: *u32, statements: []*Node, source: []const u8, code_page_lookup: *const CodePageLookup) void {
for (statements) |node| switch (node.id) {
.language_statement => {
const language_statement: *Node.LanguageStatement = @alignCast(@fieldParentPtr("base", node));
language.* = languageFromLanguageStatement(language_statement, source, code_page_lookup);
},
.simple_statement => {
const simple_statement: *Node.SimpleStatement = @alignCast(@fieldParentPtr("base", node));
const statement_type = rc.OptionalStatements.map.get(simple_statement.identifier.slice(source)) orelse continue;
const result = Compiler.evaluateNumberExpression(simple_statement.value, source, code_page_lookup);
switch (statement_type) {
.version => version.* = result.value,
.characteristics => characteristics.* = result.value,
else => unreachable, // only VERSION and CHARACTERISTICS should be in an optional statements list
}
},
else => {},
};
}
pub fn languageFromLanguageStatement(language_statement: *const Node.LanguageStatement, source: []const u8, code_page_lookup: *const CodePageLookup) res.Language {
const primary = Compiler.evaluateNumberExpression(language_statement.primary_language_id, source, code_page_lookup);
const sublanguage = Compiler.evaluateNumberExpression(language_statement.sublanguage_id, source, code_page_lookup);
return .{
.primary_language_id = @truncate(primary.value),
.sublanguage_id = @truncate(sublanguage.value),
};
}
pub fn getLanguageFromOptionalStatements(statements: []*Node, source: []const u8, code_page_lookup: *const CodePageLookup) ?res.Language {
for (statements) |node| switch (node.id) {
.language_statement => {
const language_statement: *Node.LanguageStatement = @alignCast(@fieldParentPtr("base", node));
return languageFromLanguageStatement(language_statement, source, code_page_lookup);
},
else => continue,
};
return null;
}
pub fn writeEmptyResource(writer: *std.Io.Writer) !void {
const header = ResourceHeader{
.name_value = .{ .ordinal = 0 },
.type_value = .{ .ordinal = 0 },
.language = .{
.primary_language_id = 0,
.sublanguage_id = 0,
},
.memory_flags = .{ .value = 0 },
.data_size = 0,
.version = 0,
.characteristics = 0,
};
try header.writeAssertNoOverflow(writer);
}
pub fn sourceBytesForToken(self: *Compiler, token: Token) SourceBytes {
return .{
.slice = token.slice(self.source),
.code_page = self.input_code_pages.getForToken(token),
};
}
/// Helper that calls parseQuotedStringAsWideString with the relevant context
/// Resulting slice is allocated by `self.allocator`.
pub fn parseQuotedStringAsWideString(self: *Compiler, token: Token) ![:0]u16 {
return literals.parseQuotedStringAsWideString(
self.allocator,
self.sourceBytesForToken(token),
.{
.start_column = token.calculateColumn(self.source, 8, null),
.diagnostics = self.errContext(token),
.output_code_page = self.output_code_pages.getForToken(token),
},
);
}
fn addErrorDetailsWithCodePage(self: *Compiler, details: ErrorDetails) Allocator.Error!void {
try self.diagnostics.append(details);
}
/// Code page is looked up in input_code_pages using the token
fn addErrorDetails(self: *Compiler, details_without_code_page: errors.ErrorDetailsWithoutCodePage) Allocator.Error!void {
const details = ErrorDetails{
.err = details_without_code_page.err,
.code_page = self.input_code_pages.getForToken(details_without_code_page.token),
.token = details_without_code_page.token,
.token_span_start = details_without_code_page.token_span_start,
.token_span_end = details_without_code_page.token_span_end,
.type = details_without_code_page.type,
.print_source_line = details_without_code_page.print_source_line,
.extra = details_without_code_page.extra,
};
try self.addErrorDetailsWithCodePage(details);
}
/// Code page is looked up in input_code_pages using the token
fn addErrorDetailsAndFail(self: *Compiler, details_without_code_page: errors.ErrorDetailsWithoutCodePage) error{ CompileError, OutOfMemory } {
try self.addErrorDetails(details_without_code_page);
return error.CompileError;
}
fn errContext(self: *Compiler, token: Token) errors.DiagnosticsContext {
return .{
.diagnostics = self.diagnostics,
.token = token,
.code_page = self.input_code_pages.getForToken(token),
};
}
};
pub const OpenSearchPathError = std.fs.Dir.OpenError;
fn openSearchPathDir(dir: std.fs.Dir, path: []const u8) OpenSearchPathError!std.fs.Dir {
// Validate the search path to avoid possible unreachable on invalid paths,
// see https://github.com/ziglang/zig/issues/15607 for why this is currently necessary.
try validateSearchPath(path);
return dir.openDir(path, .{});
}
/// Very crude attempt at validating a path. This is imperfect
/// and AFAIK it is effectively impossible to implement perfect path
/// validation, since it ultimately depends on the underlying filesystem.
/// Note that this function won't be necessary if/when
/// https://github.com/ziglang/zig/issues/15607
/// is accepted/implemented.
fn validateSearchPath(path: []const u8) error{BadPathName}!void {
switch (builtin.os.tag) {
.windows => {
// This will return error.BadPathName on non-Win32 namespaced paths
// (e.g. the NT \??\ prefix, the device \\.\ prefix, etc).
// Those path types are something of an unavoidable way to
// still hit unreachable during the openDir call.
var component_iterator = try std.fs.path.componentIterator(path);
while (component_iterator.next()) |component| {
// https://learn.microsoft.com/en-us/windows/win32/fileio/naming-a-file
if (std.mem.indexOfAny(u8, component.name, "\x00<>:\"|?*") != null) return error.BadPathName;
}
},
else => {
if (std.mem.indexOfScalar(u8, path, 0) != null) return error.BadPathName;
},
}
}
pub const SearchDir = struct {
dir: std.fs.Dir,
path: ?[]const u8,
pub fn deinit(self: *SearchDir, allocator: Allocator) void {
self.dir.close();
if (self.path) |path| {
allocator.free(path);
}
}
};
pub const FontDir = struct {
fonts: std.ArrayList(Font) = .empty,
/// To keep track of which ids are set and where they were set from
ids: std.AutoHashMapUnmanaged(u16, Token) = .empty,
pub const Font = struct {
id: u16,
header_bytes: [148]u8,
};
pub fn deinit(self: *FontDir, allocator: Allocator) void {
self.fonts.deinit(allocator);
}
pub fn add(self: *FontDir, allocator: Allocator, font: Font, id_token: Token) !void {
try self.ids.putNoClobber(allocator, font.id, id_token);
try self.fonts.append(allocator, font);
}
pub fn writeResData(self: *FontDir, compiler: *Compiler, writer: *std.Io.Writer) !void {
if (self.fonts.items.len == 0) return;
// We know the number of fonts is limited to maxInt(u16) because fonts
// must have a valid and unique u16 ordinal ID (trying to specify a FONT
// with e.g. id 65537 will wrap around to 1 and be ignored if there's already
// a font with that ID in the file).
const num_fonts: u16 = @intCast(self.fonts.items.len);
// u16 count + [(u16 id + 150 bytes) for each font]
// Note: This works out to a maximum data_size of 9,961,322.
const data_size: u32 = 2 + (2 + 150) * num_fonts;
var header = Compiler.ResourceHeader{
.name_value = try NameOrOrdinal.nameFromString(compiler.allocator, .{ .slice = "FONTDIR", .code_page = .windows1252 }),
.type_value = NameOrOrdinal{ .ordinal = @intFromEnum(res.RT.FONTDIR) },
.memory_flags = res.MemoryFlags.defaults(res.RT.FONTDIR),
.language = compiler.state.language,
.version = compiler.state.version,
.characteristics = compiler.state.characteristics,
.data_size = data_size,
};
defer header.deinit(compiler.allocator);
try header.writeAssertNoOverflow(writer);
try writer.writeInt(u16, num_fonts, .little);
for (self.fonts.items) |font| {
// The format of the FONTDIR is a strange beast.
// Technically, each FONT is seemingly meant to be written as a
// FONTDIRENTRY with two trailing NUL-terminated strings corresponding to
// the 'device name' and 'face name' of the .FNT file, but:
//
// 1. When dealing with .FNT files, the Win32 implementation
// gets the device name and face name from the wrong locations,
// so it's basically never going to write the real device/face name
// strings.
// 2. When dealing with files 76-140 bytes long, the Win32 implementation
// can just crash (if there are no NUL bytes in the file).
// 3. The 32-bit Win32 rc.exe uses a 148 byte size for the portion of
// the FONTDIRENTRY before the NUL-terminated strings, which
// does not match the documented FONTDIRENTRY size that (presumably)
// this format is meant to be using, so anything iterating the
// FONTDIR according to the available documentation will get bogus results.
// 4. The FONT resource can be used for non-.FNT types like TTF and OTF,
// in which case emulating the Win32 behavior of unconditionally
// interpreting the bytes as a .FNT and trying to grab device/face names
// from random bytes in the TTF/OTF file can lead to weird behavior
// and errors in the Win32 implementation (for example, the device/face
// name fields are offsets into the file where the NUL-terminated
// string is located, but the Win32 implementation actually treats
// them as signed so if they are negative then the Win32 implementation
// will error; this happening for TTF fonts would just be a bug
// since the TTF could otherwise be valid)
// 5. The FONTDIR resource doesn't actually seem to be used at all by
// anything that I've found, and instead in Windows 3.0 and newer
// it seems like the FONT resources are always just iterated/accessed
// directly without ever looking at the FONTDIR.
//
// All of these combined means that we:
// - Do not need or want to emulate Win32 behavior here
// - For maximum simplicity and compatibility, we just write the first
// 148 bytes of the file without any interpretation (padded with
// zeroes to get up to 148 bytes if necessary), and then
// unconditionally write two NUL bytes, meaning that we always
// write 'device name' and 'face name' as if they were 0-length
// strings.
//
// This gives us byte-for-byte .RES compatibility in the common case while
// allowing us to avoid any erroneous errors caused by trying to read
// the face/device name from a bogus location. Note that the Win32
// implementation never actually writes the real device/face name here
// anyway (except in the bizarre case that a .FNT file has the proper
// device/face name offsets within a reserved section of the .FNT file)
// so there's no feasible way that anything can actually think that the
// device name/face name in the FONTDIR is reliable.
// First, the ID is written, though
try writer.writeInt(u16, font.id, .little);
try writer.writeAll(&font.header_bytes);
try writer.splatByteAll(0, 2);
}
try Compiler.writeDataPadding(writer, data_size);
}
};
pub const StringTablesByLanguage = struct {
/// String tables for each language are written to the .res file in order depending on
/// when the first STRINGTABLE for the language was defined, and all blocks for a given
/// language are written contiguously.
/// Using an ArrayHashMap here gives us this property for free.
tables: std.AutoArrayHashMapUnmanaged(res.Language, StringTable) = .empty,
pub fn deinit(self: *StringTablesByLanguage, allocator: Allocator) void {
self.tables.deinit(allocator);
}
pub fn set(
self: *StringTablesByLanguage,
allocator: Allocator,
language: res.Language,
id: u16,
string_token: Token,
node: *Node,
source: []const u8,
code_page_lookup: *const CodePageLookup,
version: u32,
characteristics: u32,
) StringTable.SetError!void {
var get_or_put_result = try self.tables.getOrPut(allocator, language);
if (!get_or_put_result.found_existing) {
get_or_put_result.value_ptr.* = StringTable{};
}
return get_or_put_result.value_ptr.set(allocator, id, string_token, node, source, code_page_lookup, version, characteristics);
}
};
pub const StringTable = struct {
/// Blocks are written to the .res file in order depending on when the first string
/// was added to the block (i.e. `STRINGTABLE { 16 "b" 0 "a" }` would then get written
/// with block ID 2 (the one with "b") first and block ID 1 (the one with "a") second).
/// Using an ArrayHashMap here gives us this property for free.
blocks: std.AutoArrayHashMapUnmanaged(u16, Block) = .empty,
pub const Block = struct {
strings: std.ArrayList(Token) = .empty,
set_indexes: std.bit_set.IntegerBitSet(16) = .{ .mask = 0 },
memory_flags: MemoryFlags = MemoryFlags.defaults(res.RT.STRING),
characteristics: u32,
version: u32,
/// Returns the index to insert the string into the `strings` list.
/// Returns null if the string should be appended.
fn getInsertionIndex(self: *Block, index: u8) ?u8 {
std.debug.assert(!self.set_indexes.isSet(index));
const first_set = self.set_indexes.findFirstSet() orelse return null;
if (first_set > index) return 0;
const last_set = 15 - @clz(self.set_indexes.mask);
if (index > last_set) return null;
var bit = first_set + 1;
var insertion_index: u8 = 1;
while (bit != index) : (bit += 1) {
if (self.set_indexes.isSet(bit)) insertion_index += 1;
}
return insertion_index;
}
fn getTokenIndex(self: *Block, string_index: u8) ?u8 {
const count = self.strings.items.len;
if (count == 0) return null;
if (count == 1) return 0;
const first_set = self.set_indexes.findFirstSet() orelse unreachable;
if (first_set == string_index) return 0;
const last_set = 15 - @clz(self.set_indexes.mask);
if (last_set == string_index) return @intCast(count - 1);
if (first_set == last_set) return null;
var bit = first_set + 1;
var token_index: u8 = 1;
while (bit < last_set) : (bit += 1) {
if (!self.set_indexes.isSet(bit)) continue;
if (bit == string_index) return token_index;
token_index += 1;
}
return null;
}
fn dump(self: *Block) void {
var bit_it = self.set_indexes.iterator(.{});
var string_index: usize = 0;
while (bit_it.next()) |bit_index| {
const token = self.strings.items[string_index];
std.debug.print("{}: [{}] {any}\n", .{ bit_index, string_index, token });
string_index += 1;
}
}
pub fn applyAttributes(self: *Block, string_table: *Node.StringTable, source: []const u8, code_page_lookup: *const CodePageLookup) void {
Compiler.applyToMemoryFlags(&self.memory_flags, string_table.common_resource_attributes, source);
var dummy_language: res.Language = undefined;
Compiler.applyToOptionalStatements(&dummy_language, &self.version, &self.characteristics, string_table.optional_statements, source, code_page_lookup);
}
fn trimToDoubleNUL(comptime T: type, str: []const T) []const T {
var last_was_null = false;
for (str, 0..) |c, i| {
if (c == 0) {
if (last_was_null) return str[0 .. i - 1];
last_was_null = true;
} else {
last_was_null = false;
}
}
return str;
}
test "trimToDoubleNUL" {
try std.testing.expectEqualStrings("a\x00b", trimToDoubleNUL(u8, "a\x00b"));
try std.testing.expectEqualStrings("a", trimToDoubleNUL(u8, "a\x00\x00b"));
}
pub fn writeResData(self: *Block, compiler: *Compiler, language: res.Language, block_id: u16, writer: *std.Io.Writer) !void {
var data_buffer: std.Io.Writer.Allocating = .init(compiler.allocator);
defer data_buffer.deinit();
const data_writer = &data_buffer.writer;
var i: u8 = 0;
var string_i: u8 = 0;
while (true) : (i += 1) {
if (!self.set_indexes.isSet(i)) {
try data_writer.writeInt(u16, 0, .little);
if (i == 15) break else continue;
}
const string_token = self.strings.items[string_i];
const slice = string_token.slice(compiler.source);
const column = string_token.calculateColumn(compiler.source, 8, null);
const code_page = compiler.input_code_pages.getForToken(string_token);
const bytes = SourceBytes{ .slice = slice, .code_page = code_page };
const utf16_string = try literals.parseQuotedStringAsWideString(compiler.allocator, bytes, .{
.start_column = column,
.diagnostics = compiler.errContext(string_token),
.output_code_page = compiler.output_code_pages.getForToken(string_token),
});
defer compiler.allocator.free(utf16_string);
const trimmed_string = trim: {
// Two NUL characters in a row act as a terminator
// Note: This is only the case for STRINGTABLE strings
const trimmed = trimToDoubleNUL(u16, utf16_string);
// We also want to trim any trailing NUL characters
break :trim std.mem.trimEnd(u16, trimmed, &[_]u16{0});
};
// String literals are limited to maxInt(u15) codepoints, so these UTF-16 encoded
// strings are limited to maxInt(u15) * 2 = 65,534 code units (since 2 is the
// maximum number of UTF-16 code units per codepoint).
// This leaves room for exactly one NUL terminator.
var string_len_in_utf16_code_units: u16 = @intCast(trimmed_string.len);
// If the option is set, then a NUL terminator is added unconditionally.
// We already trimmed any trailing NULs, so we know it will be a new addition to the string.
if (compiler.null_terminate_string_table_strings) string_len_in_utf16_code_units += 1;
try data_writer.writeInt(u16, string_len_in_utf16_code_units, .little);
try data_writer.writeAll(std.mem.sliceAsBytes(trimmed_string));
if (compiler.null_terminate_string_table_strings) {
try data_writer.writeInt(u16, 0, .little);
}
if (i == 15) break;
string_i += 1;
}
// This intCast will never be able to fail due to the length constraints on string literals.
//
// - STRINGTABLE resource definitions can can only provide one string literal per index.
// - STRINGTABLE strings are limited to maxInt(u16) UTF-16 code units (see 'string_len_in_utf16_code_units'
// above), which means that the maximum number of bytes per string literal is
// 2 * maxInt(u16) = 131,070 (since there are 2 bytes per UTF-16 code unit).
// - Each Block/RT_STRING resource includes exactly 16 strings and each have a 2 byte
// length field, so the maximum number of total bytes in a RT_STRING resource's data is
// 16 * (131,070 + 2) = 2,097,152 which is well within the u32 max.
//
// Note: The string literal maximum length is enforced by the lexer.
const data_size: u32 = @intCast(data_buffer.written().len);
const header = Compiler.ResourceHeader{
.name_value = .{ .ordinal = block_id },
.type_value = .{ .ordinal = @intFromEnum(res.RT.STRING) },
.memory_flags = self.memory_flags,
.language = language,
.version = self.version,
.characteristics = self.characteristics,
.data_size = data_size,
};
// The only variable parts of the header are name and type, which in this case
// we fully control and know are numbers, so they have a fixed size.
try header.writeAssertNoOverflow(writer);
var data_fbs: std.Io.Reader = .fixed(data_buffer.written());
try Compiler.writeResourceData(writer, &data_fbs, data_size);
}
};
pub fn deinit(self: *StringTable, allocator: Allocator) void {
var it = self.blocks.iterator();
while (it.next()) |entry| {
entry.value_ptr.strings.deinit(allocator);
}
self.blocks.deinit(allocator);
}
const SetError = error{StringAlreadyDefined} || Allocator.Error;
pub fn set(
self: *StringTable,
allocator: Allocator,
id: u16,
string_token: Token,
node: *Node,
source: []const u8,
code_page_lookup: *const CodePageLookup,
version: u32,
characteristics: u32,
) SetError!void {
const block_id = (id / 16) + 1;
const string_index: u8 = @intCast(id & 0xF);
var get_or_put_result = try self.blocks.getOrPut(allocator, block_id);
if (!get_or_put_result.found_existing) {
get_or_put_result.value_ptr.* = Block{ .version = version, .characteristics = characteristics };
get_or_put_result.value_ptr.applyAttributes(node.cast(.string_table).?, source, code_page_lookup);
} else {
if (get_or_put_result.value_ptr.set_indexes.isSet(string_index)) {
return error.StringAlreadyDefined;
}
}
var block = get_or_put_result.value_ptr;
if (block.getInsertionIndex(string_index)) |insertion_index| {
try block.strings.insert(allocator, insertion_index, string_token);
} else {
try block.strings.append(allocator, string_token);
}
block.set_indexes.set(string_index);
}
pub fn get(self: *StringTable, id: u16) ?Token {
const block_id = (id / 16) + 1;
const string_index: u8 = @intCast(id & 0xF);
const block = self.blocks.getPtr(block_id) orelse return null;
const token_index = block.getTokenIndex(string_index) orelse return null;
return block.strings.items[token_index];
}
pub fn dump(self: *StringTable) !void {
var it = self.iterator();
while (it.next()) |entry| {
std.debug.print("block: {}\n", .{entry.key_ptr.*});
entry.value_ptr.dump();
}
}
};
test "StringTable" {
const S = struct {
fn makeDummyToken(id: usize) Token {
return Token{
.id = .invalid,
.start = id,
.end = id,
.line_number = id,
};
}
};
const allocator = std.testing.allocator;
var string_table = StringTable{};
defer string_table.deinit(allocator);
var code_page_lookup = CodePageLookup.init(allocator, .windows1252);
defer code_page_lookup.deinit();
var dummy_node = Node.StringTable{
.type = S.makeDummyToken(0),
.common_resource_attributes = &.{},
.optional_statements = &.{},
.begin_token = S.makeDummyToken(0),
.strings = &.{},
.end_token = S.makeDummyToken(0),
};
// randomize an array of ids 0-99
var ids = ids: {
var buf: [100]u16 = undefined;
var i: u16 = 0;
while (i < buf.len) : (i += 1) {
buf[i] = i;
}
break :ids buf;
};
var prng = std.Random.DefaultPrng.init(0);
var random = prng.random();
random.shuffle(u16, &ids);
// set each one in the randomized order
for (ids) |id| {
try string_table.set(allocator, id, S.makeDummyToken(id), &dummy_node.base, "", &code_page_lookup, 0, 0);
}
// make sure each one exists and is the right value when gotten
var id: u16 = 0;
while (id < 100) : (id += 1) {
const dummy = S.makeDummyToken(id);
try std.testing.expectError(error.StringAlreadyDefined, string_table.set(allocator, id, dummy, &dummy_node.base, "", &code_page_lookup, 0, 0));
try std.testing.expectEqual(dummy, string_table.get(id).?);
}
// make sure non-existent string ids are not found
try std.testing.expectEqual(@as(?Token, null), string_table.get(100));
}