Fix compilation issues on zig master
This commit is contained in:
parent
645cc1d7ca
commit
3531bd1b5c
15
Package.zig
15
Package.zig
|
@ -33,7 +33,7 @@ pub fn fetchOneAndUnpack(
|
||||||
var http_client: std.http.Client = .{ .allocator = allocator };
|
var http_client: std.http.Client = .{ .allocator = allocator };
|
||||||
defer http_client.deinit();
|
defer http_client.deinit();
|
||||||
|
|
||||||
var global_cache_directory: std.Build.Cache.Directory = .{
|
const global_cache_directory: std.Build.Cache.Directory = .{
|
||||||
.handle = try std.fs.cwd().makeOpenPath(cache_directory, .{}),
|
.handle = try std.fs.cwd().makeOpenPath(cache_directory, .{}),
|
||||||
.path = cache_directory,
|
.path = cache_directory,
|
||||||
};
|
};
|
||||||
|
@ -111,12 +111,12 @@ pub fn fetchAndUnpack(
|
||||||
const path = try global_cache_directory.join(gpa, &.{tmp_dir_sub_path});
|
const path = try global_cache_directory.join(gpa, &.{tmp_dir_sub_path});
|
||||||
errdefer gpa.free(path);
|
errdefer gpa.free(path);
|
||||||
|
|
||||||
const iterable_dir = try global_cache_directory.handle.makeOpenPathIterable(tmp_dir_sub_path, .{});
|
const iterable_dir = try global_cache_directory.handle.makeOpenPath(tmp_dir_sub_path, .{});
|
||||||
errdefer iterable_dir.close();
|
errdefer iterable_dir.close();
|
||||||
|
|
||||||
break :d .{
|
break :d .{
|
||||||
.path = path,
|
.path = path,
|
||||||
.handle = iterable_dir.dir,
|
.handle = iterable_dir,
|
||||||
};
|
};
|
||||||
};
|
};
|
||||||
defer tmp_directory.closeAndFree(gpa);
|
defer tmp_directory.closeAndFree(gpa);
|
||||||
|
@ -124,10 +124,10 @@ pub fn fetchAndUnpack(
|
||||||
var h = std.http.Headers{ .allocator = gpa };
|
var h = std.http.Headers{ .allocator = gpa };
|
||||||
defer h.deinit();
|
defer h.deinit();
|
||||||
|
|
||||||
var req = try http_client.request(.GET, uri, h, .{});
|
var req = try http_client.open(.GET, uri, h, .{});
|
||||||
defer req.deinit();
|
defer req.deinit();
|
||||||
|
|
||||||
try req.start();
|
try req.send(.{});
|
||||||
try req.wait();
|
try req.wait();
|
||||||
|
|
||||||
if (req.response.status != .ok) {
|
if (req.response.status != .ok) {
|
||||||
|
@ -202,7 +202,8 @@ pub fn fetchAndUnpack(
|
||||||
// Of course, if the ignore rules above omit the file from the package, then everything
|
// Of course, if the ignore rules above omit the file from the package, then everything
|
||||||
// is fine and no error should be raised.
|
// is fine and no error should be raised.
|
||||||
|
|
||||||
break :a try Hasher.computeDirectoryHash(thread_pool, .{ .dir = tmp_directory.handle }, &.{});
|
var options = .{};
|
||||||
|
break :a try Hasher.computeDirectoryHash(thread_pool, tmp_directory.handle, &options);
|
||||||
};
|
};
|
||||||
|
|
||||||
const pkg_dir_sub_path = "p" ++ s ++ Hasher.hexDigest(actual_hash);
|
const pkg_dir_sub_path = "p" ++ s ++ Hasher.hexDigest(actual_hash);
|
||||||
|
@ -464,7 +465,7 @@ test "fetch and unpack" {
|
||||||
var http_client: std.http.Client = .{ .allocator = alloc };
|
var http_client: std.http.Client = .{ .allocator = alloc };
|
||||||
defer http_client.deinit();
|
defer http_client.deinit();
|
||||||
|
|
||||||
var global_cache_directory: std.Build.Cache.Directory = .{
|
const global_cache_directory: std.Build.Cache.Directory = .{
|
||||||
.handle = try std.fs.cwd().makeOpenPath("test-pkg", .{}),
|
.handle = try std.fs.cwd().makeOpenPath("test-pkg", .{}),
|
||||||
.path = "test-pkg",
|
.path = "test-pkg",
|
||||||
};
|
};
|
||||||
|
|
26
build.zig
26
build.zig
|
@ -1,6 +1,6 @@
|
||||||
const std = @import("std");
|
const std = @import("std");
|
||||||
const builtin = @import("builtin");
|
const builtin = @import("builtin");
|
||||||
const Builder = @import("std").build.Builder;
|
const Builder = @import("std").Build;
|
||||||
const Package = @import("Package.zig");
|
const Package = @import("Package.zig");
|
||||||
|
|
||||||
const models_url = "https://github.com/aws/aws-sdk-go-v2/archive/7502ff360b1c3b79cbe117437327f6ff5fb89f65.tar.gz";
|
const models_url = "https://github.com/aws/aws-sdk-go-v2/archive/7502ff360b1c3b79cbe117437327f6ff5fb89f65.tar.gz";
|
||||||
|
@ -79,22 +79,18 @@ pub fn build(b: *Builder) !void {
|
||||||
.optimize = optimize,
|
.optimize = optimize,
|
||||||
});
|
});
|
||||||
const smithy_module = smithy_dep.module("smithy");
|
const smithy_module = smithy_dep.module("smithy");
|
||||||
exe.addModule("smithy", smithy_module); // not sure this should be here...
|
exe.root_module.addImport("smithy", smithy_module); // not sure this should be here...
|
||||||
|
|
||||||
// Expose module to others
|
// Expose module to others
|
||||||
_ = b.addModule("aws", .{
|
_ = b.addModule("aws", .{
|
||||||
.source_file = .{ .path = "src/aws.zig" },
|
.root_source_file = .{ .path = "src/aws.zig" },
|
||||||
.dependencies = &[_]std.build.ModuleDependency{
|
.imports = &.{.{ .name = "smithy", .module = smithy_module }},
|
||||||
.{ .name = "smithy", .module = smithy_module },
|
|
||||||
},
|
|
||||||
});
|
});
|
||||||
|
|
||||||
// Expose module to others
|
// Expose module to others
|
||||||
_ = b.addModule("aws-signing", .{
|
_ = b.addModule("aws-signing", .{
|
||||||
.source_file = .{ .path = "src/aws_signing.zig" },
|
.root_source_file = .{ .path = "src/aws_signing.zig" },
|
||||||
.dependencies = &[_]std.build.ModuleDependency{
|
.imports = &.{.{ .name = "smithy", .module = smithy_module }},
|
||||||
.{ .name = "smithy", .module = smithy_module },
|
|
||||||
},
|
|
||||||
});
|
});
|
||||||
// TODO: This does not work correctly due to https://github.com/ziglang/zig/issues/16354
|
// TODO: This does not work correctly due to https://github.com/ziglang/zig/issues/16354
|
||||||
//
|
//
|
||||||
|
@ -128,10 +124,10 @@ pub fn build(b: *Builder) !void {
|
||||||
.name = "codegen",
|
.name = "codegen",
|
||||||
.root_source_file = .{ .path = "codegen/src/main.zig" },
|
.root_source_file = .{ .path = "codegen/src/main.zig" },
|
||||||
// We need this generated for the host, not the real target
|
// We need this generated for the host, not the real target
|
||||||
// .target = target,
|
.target = b.host,
|
||||||
.optimize = if (b.verbose) .Debug else .ReleaseSafe,
|
.optimize = if (b.verbose) .Debug else .ReleaseSafe,
|
||||||
});
|
});
|
||||||
cg_exe.addModule("smithy", smithy_dep.module("smithy"));
|
cg_exe.root_module.addImport("smithy", smithy_dep.module("smithy"));
|
||||||
var cg_cmd = b.addRunArtifact(cg_exe);
|
var cg_cmd = b.addRunArtifact(cg_exe);
|
||||||
cg_cmd.addArg("--models");
|
cg_cmd.addArg("--models");
|
||||||
cg_cmd.addArg(try std.fs.path.join(
|
cg_cmd.addArg(try std.fs.path.join(
|
||||||
|
@ -139,7 +135,7 @@ pub fn build(b: *Builder) !void {
|
||||||
&[_][]const u8{ b.global_cache_root.path.?, models_dir },
|
&[_][]const u8{ b.global_cache_root.path.?, models_dir },
|
||||||
));
|
));
|
||||||
cg_cmd.addArg("--output");
|
cg_cmd.addArg("--output");
|
||||||
cg_cmd.addDirectoryArg(std.Build.FileSource.relative("src/models"));
|
cg_cmd.addDirectoryArg(std.Build.LazyPath.relative("src/models"));
|
||||||
if (b.verbose)
|
if (b.verbose)
|
||||||
cg_cmd.addArg("--verbose");
|
cg_cmd.addArg("--verbose");
|
||||||
cg_cmd.step.dependOn(&fetch_step.step);
|
cg_cmd.step.dependOn(&fetch_step.step);
|
||||||
|
@ -173,10 +169,10 @@ pub fn build(b: *Builder) !void {
|
||||||
// but does not run it.
|
// but does not run it.
|
||||||
const unit_tests = b.addTest(.{
|
const unit_tests = b.addTest(.{
|
||||||
.root_source_file = .{ .path = "src/aws.zig" },
|
.root_source_file = .{ .path = "src/aws.zig" },
|
||||||
.target = t,
|
.target = b.resolveTargetQuery(t),
|
||||||
.optimize = optimize,
|
.optimize = optimize,
|
||||||
});
|
});
|
||||||
unit_tests.addModule("smithy", smithy_dep.module("smithy"));
|
unit_tests.root_module.addImport("smithy", smithy_dep.module("smithy"));
|
||||||
unit_tests.step.dependOn(gen_step);
|
unit_tests.step.dependOn(gen_step);
|
||||||
|
|
||||||
const run_unit_tests = b.addRunArtifact(unit_tests);
|
const run_unit_tests = b.addRunArtifact(unit_tests);
|
||||||
|
|
|
@ -1,11 +1,14 @@
|
||||||
.{
|
.{
|
||||||
.name = "aws-zig",
|
.name = "aws-zig",
|
||||||
.version = "0.0.1",
|
.version = "0.0.1",
|
||||||
|
.paths = .{""},
|
||||||
|
|
||||||
.dependencies = .{
|
.dependencies = .{
|
||||||
.smithy = .{
|
.smithy = .{
|
||||||
.url = "https://git.lerch.org/lobo/smithy/archive/41b61745d25a65817209dd5dddbb5f9b66896a99.tar.gz",
|
.url = "https://github.com/melhindi/smithy/archive/refs/heads/master.tar.gz",
|
||||||
.hash = "122087deb0ae309b2258d59b40d82fe5921fdfc35b420bb59033244851f7f276fa34",
|
.hash = "12209568cc5c0bad2fcf6a62dbf6792fbbf3727214da98b21c390fc399f6dd743d3a",
|
||||||
|
// .url = "https://git.lerch.org/lobo/smithy/archive/41b61745d25a65817209dd5dddbb5f9b66896a99.tar.gz",
|
||||||
|
// .hash = "122087deb0ae309b2258d59b40d82fe5921fdfc35b420bb59033244851f7f276fa34",
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
|
@ -77,13 +77,13 @@ pub fn hex64(x: u64) [16]u8 {
|
||||||
return result;
|
return result;
|
||||||
}
|
}
|
||||||
|
|
||||||
pub const walkerFn = *const fn (std.fs.IterableDir.Walker.WalkerEntry) bool;
|
pub const walkerFn = *const fn (std.fs.Dir.Walker.WalkerEntry) bool;
|
||||||
|
|
||||||
fn included(entry: std.fs.IterableDir.Walker.WalkerEntry) bool {
|
fn included(entry: std.fs.Dir.Walker.WalkerEntry) bool {
|
||||||
_ = entry;
|
_ = entry;
|
||||||
return true;
|
return true;
|
||||||
}
|
}
|
||||||
fn excluded(entry: std.fs.IterableDir.Walker.WalkerEntry) bool {
|
fn excluded(entry: std.fs.Dir.Walker.WalkerEntry) bool {
|
||||||
_ = entry;
|
_ = entry;
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
@ -96,7 +96,7 @@ pub const ComputeDirectoryOptions = struct {
|
||||||
|
|
||||||
pub fn computeDirectoryHash(
|
pub fn computeDirectoryHash(
|
||||||
thread_pool: *std.Thread.Pool,
|
thread_pool: *std.Thread.Pool,
|
||||||
dir: std.fs.IterableDir,
|
dir: std.fs.Dir,
|
||||||
options: *ComputeDirectoryOptions,
|
options: *ComputeDirectoryOptions,
|
||||||
) ![Hash.digest_length]u8 {
|
) ![Hash.digest_length]u8 {
|
||||||
const gpa = thread_pool.allocator;
|
const gpa = thread_pool.allocator;
|
||||||
|
@ -138,7 +138,7 @@ pub fn computeDirectoryHash(
|
||||||
.failure = undefined, // to be populated by the worker
|
.failure = undefined, // to be populated by the worker
|
||||||
};
|
};
|
||||||
wait_group.start();
|
wait_group.start();
|
||||||
try thread_pool.spawn(workerHashFile, .{ dir.dir, hashed_file, &wait_group });
|
try thread_pool.spawn(workerHashFile, .{ dir, hashed_file, &wait_group });
|
||||||
|
|
||||||
try all_files.append(hashed_file);
|
try all_files.append(hashed_file);
|
||||||
}
|
}
|
||||||
|
|
|
@ -17,7 +17,7 @@ pub fn main() anyerror!void {
|
||||||
|
|
||||||
var output_dir = std.fs.cwd();
|
var output_dir = std.fs.cwd();
|
||||||
defer if (output_dir.fd > 0) output_dir.close();
|
defer if (output_dir.fd > 0) output_dir.close();
|
||||||
var models_dir: ?std.fs.IterableDir = null;
|
var models_dir: ?std.fs.Dir = null;
|
||||||
defer if (models_dir) |*m| m.close();
|
defer if (models_dir) |*m| m.close();
|
||||||
for (args, 0..) |arg, i| {
|
for (args, 0..) |arg, i| {
|
||||||
if (std.mem.eql(u8, "--help", arg) or
|
if (std.mem.eql(u8, "--help", arg) or
|
||||||
|
@ -31,13 +31,13 @@ pub fn main() anyerror!void {
|
||||||
if (std.mem.eql(u8, "--output", arg))
|
if (std.mem.eql(u8, "--output", arg))
|
||||||
output_dir = try output_dir.makeOpenPath(args[i + 1], .{});
|
output_dir = try output_dir.makeOpenPath(args[i + 1], .{});
|
||||||
if (std.mem.eql(u8, "--models", arg))
|
if (std.mem.eql(u8, "--models", arg))
|
||||||
models_dir = try std.fs.cwd().openIterableDir(args[i + 1], .{});
|
models_dir = try std.fs.cwd().openDir(args[i + 1], .{ .iterate = true });
|
||||||
}
|
}
|
||||||
// TODO: Seems like we should remove this in favor of a package
|
// TODO: Seems like we should remove this in favor of a package
|
||||||
try output_dir.writeFile("json.zig", json_zig);
|
try output_dir.writeFile("json.zig", json_zig);
|
||||||
|
|
||||||
// TODO: We need a different way to handle this file...
|
// TODO: We need a different way to handle this file...
|
||||||
var manifest_file_started = false;
|
const manifest_file_started = false;
|
||||||
var manifest_file: std.fs.File = undefined;
|
var manifest_file: std.fs.File = undefined;
|
||||||
defer if (manifest_file_started) manifest_file.close();
|
defer if (manifest_file_started) manifest_file.close();
|
||||||
var manifest: std.fs.File.Writer = undefined;
|
var manifest: std.fs.File.Writer = undefined;
|
||||||
|
@ -71,11 +71,11 @@ pub fn main() anyerror!void {
|
||||||
// this is our normal mode of operation and where initial optimizations
|
// this is our normal mode of operation and where initial optimizations
|
||||||
// can be made
|
// can be made
|
||||||
if (models_dir) |m| {
|
if (models_dir) |m| {
|
||||||
var cwd = try std.fs.cwd().openDir(".", .{});
|
var cwd = try std.fs.cwd().openDir(".", .{ .iterate = true });
|
||||||
defer cwd.close();
|
defer cwd.close();
|
||||||
defer cwd.setAsCwd() catch unreachable;
|
defer cwd.setAsCwd() catch unreachable;
|
||||||
|
|
||||||
try m.dir.setAsCwd();
|
try m.setAsCwd();
|
||||||
try processDirectories(m, output_dir);
|
try processDirectories(m, output_dir);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -87,7 +87,7 @@ const OutputManifest = struct {
|
||||||
model_dir_hash_digest: [Hasher.hex_multihash_len]u8,
|
model_dir_hash_digest: [Hasher.hex_multihash_len]u8,
|
||||||
output_dir_hash_digest: [Hasher.hex_multihash_len]u8,
|
output_dir_hash_digest: [Hasher.hex_multihash_len]u8,
|
||||||
};
|
};
|
||||||
fn processDirectories(models_dir: std.fs.IterableDir, output_dir: std.fs.Dir) !void {
|
fn processDirectories(models_dir: std.fs.Dir, output_dir: std.fs.Dir) !void {
|
||||||
// Let's get ready to hash!!
|
// Let's get ready to hash!!
|
||||||
var arena = std.heap.ArenaAllocator.init(std.heap.page_allocator);
|
var arena = std.heap.ArenaAllocator.init(std.heap.page_allocator);
|
||||||
defer arena.deinit();
|
defer arena.deinit();
|
||||||
|
@ -131,15 +131,15 @@ fn processDirectories(models_dir: std.fs.IterableDir, output_dir: std.fs.Dir) !v
|
||||||
}
|
}
|
||||||
|
|
||||||
var model_digest: ?[Hasher.hex_multihash_len]u8 = null;
|
var model_digest: ?[Hasher.hex_multihash_len]u8 = null;
|
||||||
fn calculateDigests(models_dir: std.fs.IterableDir, output_dir: std.fs.Dir, thread_pool: *std.Thread.Pool) !OutputManifest {
|
fn calculateDigests(models_dir: std.fs.Dir, output_dir: std.fs.Dir, thread_pool: *std.Thread.Pool) !OutputManifest {
|
||||||
const model_hash = if (model_digest) |m| m[0..Hasher.digest_len].* else try Hasher.computeDirectoryHash(thread_pool, models_dir, @constCast(&Hasher.ComputeDirectoryOptions{
|
const model_hash = if (model_digest) |m| m[0..Hasher.digest_len].* else try Hasher.computeDirectoryHash(thread_pool, models_dir, @constCast(&Hasher.ComputeDirectoryOptions{
|
||||||
.isIncluded = struct {
|
.isIncluded = struct {
|
||||||
pub fn include(entry: std.fs.IterableDir.Walker.WalkerEntry) bool {
|
pub fn include(entry: std.fs.Dir.Walker.WalkerEntry) bool {
|
||||||
return std.mem.endsWith(u8, entry.basename, ".json");
|
return std.mem.endsWith(u8, entry.basename, ".json");
|
||||||
}
|
}
|
||||||
}.include,
|
}.include,
|
||||||
.isExcluded = struct {
|
.isExcluded = struct {
|
||||||
pub fn exclude(entry: std.fs.IterableDir.Walker.WalkerEntry) bool {
|
pub fn exclude(entry: std.fs.Dir.Walker.WalkerEntry) bool {
|
||||||
_ = entry;
|
_ = entry;
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
@ -148,14 +148,14 @@ fn calculateDigests(models_dir: std.fs.IterableDir, output_dir: std.fs.Dir, thre
|
||||||
}));
|
}));
|
||||||
if (verbose) std.log.info("Model directory hash: {s}", .{model_digest orelse Hasher.hexDigest(model_hash)});
|
if (verbose) std.log.info("Model directory hash: {s}", .{model_digest orelse Hasher.hexDigest(model_hash)});
|
||||||
|
|
||||||
const output_hash = try Hasher.computeDirectoryHash(thread_pool, try output_dir.openIterableDir(".", .{}), @constCast(&Hasher.ComputeDirectoryOptions{
|
const output_hash = try Hasher.computeDirectoryHash(thread_pool, try output_dir.openDir(".", .{ .iterate = true }), @constCast(&Hasher.ComputeDirectoryOptions{
|
||||||
.isIncluded = struct {
|
.isIncluded = struct {
|
||||||
pub fn include(entry: std.fs.IterableDir.Walker.WalkerEntry) bool {
|
pub fn include(entry: std.fs.Dir.Walker.WalkerEntry) bool {
|
||||||
return std.mem.endsWith(u8, entry.basename, ".zig");
|
return std.mem.endsWith(u8, entry.basename, ".zig");
|
||||||
}
|
}
|
||||||
}.include,
|
}.include,
|
||||||
.isExcluded = struct {
|
.isExcluded = struct {
|
||||||
pub fn exclude(entry: std.fs.IterableDir.Walker.WalkerEntry) bool {
|
pub fn exclude(entry: std.fs.Dir.Walker.WalkerEntry) bool {
|
||||||
_ = entry;
|
_ = entry;
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
@ -201,7 +201,7 @@ fn processFile(file_name: []const u8, output_dir: std.fs.Dir, manifest: anytype)
|
||||||
defer allocator.free(output_file_name);
|
defer allocator.free(output_file_name);
|
||||||
for (service_names) |name| {
|
for (service_names) |name| {
|
||||||
const seperator = if (output_file_name.len > 0) "-" else "";
|
const seperator = if (output_file_name.len > 0) "-" else "";
|
||||||
var new_output_file_name = try std.fmt.allocPrint(
|
const new_output_file_name = try std.fmt.allocPrint(
|
||||||
allocator,
|
allocator,
|
||||||
"{s}{s}{s}",
|
"{s}{s}{s}",
|
||||||
.{ output_file_name, seperator, name },
|
.{ output_file_name, seperator, name },
|
||||||
|
@ -211,7 +211,7 @@ fn processFile(file_name: []const u8, output_dir: std.fs.Dir, manifest: anytype)
|
||||||
}
|
}
|
||||||
{
|
{
|
||||||
// append .zig on to the file name
|
// append .zig on to the file name
|
||||||
var new_output_file_name = try std.fmt.allocPrint(
|
const new_output_file_name = try std.fmt.allocPrint(
|
||||||
allocator,
|
allocator,
|
||||||
"{s}.zig",
|
"{s}.zig",
|
||||||
.{output_file_name},
|
.{output_file_name},
|
||||||
|
@ -337,7 +337,7 @@ fn generateServices(allocator: std.mem.Allocator, comptime _: []const u8, file:
|
||||||
var generated = std.StringHashMap(void).init(allocator);
|
var generated = std.StringHashMap(void).init(allocator);
|
||||||
defer generated.deinit();
|
defer generated.deinit();
|
||||||
|
|
||||||
var state = FileGenerationState{
|
const state = FileGenerationState{
|
||||||
.shape_references = shape_references,
|
.shape_references = shape_references,
|
||||||
.additional_types_to_generate = &unresolved,
|
.additional_types_to_generate = &unresolved,
|
||||||
.additional_types_generated = &generated,
|
.additional_types_generated = &generated,
|
||||||
|
@ -345,8 +345,8 @@ fn generateServices(allocator: std.mem.Allocator, comptime _: []const u8, file:
|
||||||
};
|
};
|
||||||
for (services.items) |service| {
|
for (services.items) |service| {
|
||||||
var sdk_id: []const u8 = undefined;
|
var sdk_id: []const u8 = undefined;
|
||||||
var version: []const u8 = service.shape.service.version;
|
const version: []const u8 = service.shape.service.version;
|
||||||
var name: []const u8 = service.name;
|
const name: []const u8 = service.name;
|
||||||
var arn_namespace: []const u8 = undefined;
|
var arn_namespace: []const u8 = undefined;
|
||||||
var sigv4_name: []const u8 = undefined;
|
var sigv4_name: []const u8 = undefined;
|
||||||
var endpoint_prefix: []const u8 = undefined;
|
var endpoint_prefix: []const u8 = undefined;
|
||||||
|
|
|
@ -31,7 +31,7 @@ pub const services = servicemodel.services;
|
||||||
pub const Services = servicemodel.Services;
|
pub const Services = servicemodel.Services;
|
||||||
|
|
||||||
pub const ClientOptions = struct {
|
pub const ClientOptions = struct {
|
||||||
proxy: ?std.http.Client.HttpProxy = null,
|
proxy: ?std.http.Client.Proxy = null,
|
||||||
};
|
};
|
||||||
pub const Client = struct {
|
pub const Client = struct {
|
||||||
allocator: std.mem.Allocator,
|
allocator: std.mem.Allocator,
|
||||||
|
@ -365,7 +365,7 @@ pub fn Request(comptime request_action: anytype) type {
|
||||||
.raw_parsed = .{ .raw = .{} },
|
.raw_parsed = .{ .raw = .{} },
|
||||||
.allocator = options.client.allocator,
|
.allocator = options.client.allocator,
|
||||||
};
|
};
|
||||||
var body_field = @field(rc.response, action.Response.http_payload);
|
const body_field = @field(rc.response, action.Response.http_payload);
|
||||||
const BodyField = @TypeOf(body_field);
|
const BodyField = @TypeOf(body_field);
|
||||||
if (BodyField == []const u8 or BodyField == ?[]const u8) {
|
if (BodyField == []const u8 or BodyField == ?[]const u8) {
|
||||||
expected_body_field_len = 0;
|
expected_body_field_len = 0;
|
||||||
|
@ -875,7 +875,7 @@ fn FullResponse(comptime action: anytype) type {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if (@hasDecl(Response, "http_payload")) {
|
if (@hasDecl(Response, "http_payload")) {
|
||||||
var body_field = @field(self.response, Response.http_payload);
|
const body_field = @field(self.response, Response.http_payload);
|
||||||
const BodyField = @TypeOf(body_field);
|
const BodyField = @TypeOf(body_field);
|
||||||
if (BodyField == []const u8) {
|
if (BodyField == []const u8) {
|
||||||
self.allocator.free(body_field);
|
self.allocator.free(body_field);
|
||||||
|
@ -1465,7 +1465,7 @@ fn processRequest(options: *TestOptions, server: *std.http.Server) !void {
|
||||||
else
|
else
|
||||||
res.transfer_encoding = .chunked;
|
res.transfer_encoding = .chunked;
|
||||||
|
|
||||||
try res.do();
|
try res.send();
|
||||||
_ = try res.writer().writeAll(response_bytes);
|
_ = try res.writer().writeAll(response_bytes);
|
||||||
try res.finish();
|
try res.finish();
|
||||||
log.debug(
|
log.debug(
|
||||||
|
|
|
@ -126,9 +126,9 @@ fn getContainerCredentials(allocator: std.mem.Allocator) !?auth.Credentials {
|
||||||
defer empty_headers.deinit();
|
defer empty_headers.deinit();
|
||||||
var cl = std.http.Client{ .allocator = allocator };
|
var cl = std.http.Client{ .allocator = allocator };
|
||||||
defer cl.deinit(); // I don't belive connection pooling would help much here as it's non-ssl and local
|
defer cl.deinit(); // I don't belive connection pooling would help much here as it's non-ssl and local
|
||||||
var req = try cl.request(.GET, try std.Uri.parse(container_uri), empty_headers, .{});
|
var req = try cl.open(.GET, try std.Uri.parse(container_uri), empty_headers, .{});
|
||||||
defer req.deinit();
|
defer req.deinit();
|
||||||
try req.start();
|
try req.send(.{});
|
||||||
try req.wait();
|
try req.wait();
|
||||||
if (req.response.status != .ok and req.response.status != .not_found) {
|
if (req.response.status != .ok and req.response.status != .not_found) {
|
||||||
log.warn("Bad status code received from container credentials endpoint: {}", .{@intFromEnum(req.response.status)});
|
log.warn("Bad status code received from container credentials endpoint: {}", .{@intFromEnum(req.response.status)});
|
||||||
|
@ -140,7 +140,7 @@ fn getContainerCredentials(allocator: std.mem.Allocator) !?auth.Credentials {
|
||||||
var resp_payload = try std.ArrayList(u8).initCapacity(allocator, @intCast(req.response.content_length.?));
|
var resp_payload = try std.ArrayList(u8).initCapacity(allocator, @intCast(req.response.content_length.?));
|
||||||
defer resp_payload.deinit();
|
defer resp_payload.deinit();
|
||||||
try resp_payload.resize(@intCast(req.response.content_length.?));
|
try resp_payload.resize(@intCast(req.response.content_length.?));
|
||||||
var response_data = try resp_payload.toOwnedSlice();
|
const response_data = try resp_payload.toOwnedSlice();
|
||||||
defer allocator.free(response_data);
|
defer allocator.free(response_data);
|
||||||
_ = try req.readAll(response_data);
|
_ = try req.readAll(response_data);
|
||||||
log.debug("Read {d} bytes from container credentials endpoint", .{response_data.len});
|
log.debug("Read {d} bytes from container credentials endpoint", .{response_data.len});
|
||||||
|
@ -185,9 +185,9 @@ fn getImdsv2Credentials(allocator: std.mem.Allocator) !?auth.Credentials {
|
||||||
var headers = std.http.Headers.init(allocator);
|
var headers = std.http.Headers.init(allocator);
|
||||||
defer headers.deinit();
|
defer headers.deinit();
|
||||||
try headers.append("X-aws-ec2-metadata-token-ttl-seconds", "21600");
|
try headers.append("X-aws-ec2-metadata-token-ttl-seconds", "21600");
|
||||||
var req = try cl.request(.PUT, try std.Uri.parse("http://169.254.169.254/latest/api/token"), headers, .{});
|
var req = try cl.open(.PUT, try std.Uri.parse("http://169.254.169.254/latest/api/token"), headers, .{});
|
||||||
defer req.deinit();
|
defer req.deinit();
|
||||||
try req.start();
|
try req.send(.{});
|
||||||
try req.wait();
|
try req.wait();
|
||||||
if (req.response.status != .ok) {
|
if (req.response.status != .ok) {
|
||||||
log.warn("Bad status code received from IMDS v2: {}", .{@intFromEnum(req.response.status)});
|
log.warn("Bad status code received from IMDS v2: {}", .{@intFromEnum(req.response.status)});
|
||||||
|
@ -228,10 +228,10 @@ fn getImdsRoleName(allocator: std.mem.Allocator, client: *std.http.Client, imds_
|
||||||
defer headers.deinit();
|
defer headers.deinit();
|
||||||
try headers.append("X-aws-ec2-metadata-token", imds_token);
|
try headers.append("X-aws-ec2-metadata-token", imds_token);
|
||||||
|
|
||||||
var req = try client.request(.GET, try std.Uri.parse("http://169.254.169.254/latest/meta-data/iam/info"), headers, .{});
|
var req = try client.open(.GET, try std.Uri.parse("http://169.254.169.254/latest/meta-data/iam/info"), headers, .{});
|
||||||
defer req.deinit();
|
defer req.deinit();
|
||||||
|
|
||||||
try req.start();
|
try req.send(.{});
|
||||||
try req.wait();
|
try req.wait();
|
||||||
|
|
||||||
if (req.response.status != .ok and req.response.status != .not_found) {
|
if (req.response.status != .ok and req.response.status != .not_found) {
|
||||||
|
@ -243,7 +243,7 @@ fn getImdsRoleName(allocator: std.mem.Allocator, client: *std.http.Client, imds_
|
||||||
log.warn("Unexpected empty response from IMDS endpoint post token", .{});
|
log.warn("Unexpected empty response from IMDS endpoint post token", .{});
|
||||||
return null;
|
return null;
|
||||||
}
|
}
|
||||||
var resp = try allocator.alloc(u8, @intCast(req.response.content_length.?));
|
const resp = try allocator.alloc(u8, @intCast(req.response.content_length.?));
|
||||||
defer allocator.free(resp);
|
defer allocator.free(resp);
|
||||||
_ = try req.readAll(resp);
|
_ = try req.readAll(resp);
|
||||||
|
|
||||||
|
@ -281,10 +281,10 @@ fn getImdsCredentials(allocator: std.mem.Allocator, client: *std.http.Client, ro
|
||||||
const url = try std.fmt.allocPrint(allocator, "http://169.254.169.254/latest/meta-data/iam/security-credentials/{s}/", .{role_name});
|
const url = try std.fmt.allocPrint(allocator, "http://169.254.169.254/latest/meta-data/iam/security-credentials/{s}/", .{role_name});
|
||||||
defer allocator.free(url);
|
defer allocator.free(url);
|
||||||
|
|
||||||
var req = try client.request(.GET, try std.Uri.parse(url), headers, .{});
|
var req = try client.open(.GET, try std.Uri.parse(url), headers, .{});
|
||||||
defer req.deinit();
|
defer req.deinit();
|
||||||
|
|
||||||
try req.start();
|
try req.send(.{});
|
||||||
try req.wait();
|
try req.wait();
|
||||||
|
|
||||||
if (req.response.status != .ok and req.response.status != .not_found) {
|
if (req.response.status != .ok and req.response.status != .not_found) {
|
||||||
|
@ -296,7 +296,7 @@ fn getImdsCredentials(allocator: std.mem.Allocator, client: *std.http.Client, ro
|
||||||
log.warn("Unexpected empty response from IMDS role endpoint", .{});
|
log.warn("Unexpected empty response from IMDS role endpoint", .{});
|
||||||
return null;
|
return null;
|
||||||
}
|
}
|
||||||
var resp = try allocator.alloc(u8, @intCast(req.response.content_length.?));
|
const resp = try allocator.alloc(u8, @intCast(req.response.content_length.?));
|
||||||
defer allocator.free(resp);
|
defer allocator.free(resp);
|
||||||
_ = try req.readAll(resp);
|
_ = try req.readAll(resp);
|
||||||
|
|
||||||
|
@ -455,7 +455,7 @@ const LineIterator = struct {
|
||||||
pub fn next(self: *Self) ?[]const u8 {
|
pub fn next(self: *Self) ?[]const u8 {
|
||||||
if (self.inx >= self.text.len) return null;
|
if (self.inx >= self.text.len) return null;
|
||||||
var current = self.inx;
|
var current = self.inx;
|
||||||
var start = self.inx;
|
const start = self.inx;
|
||||||
for (self.text[self.inx..], 0..) |c, i| {
|
for (self.text[self.inx..], 0..) |c, i| {
|
||||||
if (c == '\n') {
|
if (c == '\n') {
|
||||||
// log.debug("got \\n: {d}", .{i});
|
// log.debug("got \\n: {d}", .{i});
|
||||||
|
@ -571,7 +571,7 @@ const EvaluatedPath = struct {
|
||||||
evaluated_path: []const u8,
|
evaluated_path: []const u8,
|
||||||
};
|
};
|
||||||
fn getDefaultPath(allocator: std.mem.Allocator, home_dir: ?[]const u8, dir: []const u8, file: []const u8) !EvaluatedPath {
|
fn getDefaultPath(allocator: std.mem.Allocator, home_dir: ?[]const u8, dir: []const u8, file: []const u8) !EvaluatedPath {
|
||||||
var home = home_dir orelse try getHomeDir(allocator);
|
const home = home_dir orelse try getHomeDir(allocator);
|
||||||
log.debug("Home directory: {s}", .{home});
|
log.debug("Home directory: {s}", .{home});
|
||||||
const rc = try std.fs.path.join(allocator, &[_][]const u8{ home, dir, file });
|
const rc = try std.fs.path.join(allocator, &[_][]const u8{ home, dir, file });
|
||||||
log.debug("Path evaluated as: {s}", .{rc});
|
log.debug("Path evaluated as: {s}", .{rc});
|
||||||
|
@ -581,29 +581,18 @@ fn getDefaultPath(allocator: std.mem.Allocator, home_dir: ?[]const u8, dir: []co
|
||||||
fn getHomeDir(allocator: std.mem.Allocator) ![]const u8 {
|
fn getHomeDir(allocator: std.mem.Allocator) ![]const u8 {
|
||||||
switch (builtin.os.tag) {
|
switch (builtin.os.tag) {
|
||||||
.windows => {
|
.windows => {
|
||||||
var dir_path_ptr: [*:0]u16 = undefined;
|
|
||||||
// https://docs.microsoft.com/en-us/windows/win32/shell/knownfolderid
|
// https://docs.microsoft.com/en-us/windows/win32/shell/knownfolderid
|
||||||
const FOLDERID_Profile = std.os.windows.GUID.parse("{5E6C858F-0E22-4760-9AFE-EA3317B67173}");
|
//const FOLDERID_Profile = std.os.windows.GUID.parse("{5E6C858F-0E22-4760-9AFE-EA3317B67173}");
|
||||||
switch (std.os.windows.shell32.SHGetKnownFolderPath(
|
|
||||||
&FOLDERID_Profile,
|
const local_app_data_dir = std.process.getEnvVarOwned(allocator, "LOCALAPPDATA") catch |err| switch (err) {
|
||||||
std.os.windows.KF_FLAG_CREATE,
|
error.InvalidUtf8 => return error.HomeDirUnavailable,
|
||||||
null,
|
error.EnvironmentVariableNotFound => return error.HomeDirUnavailable,
|
||||||
&dir_path_ptr,
|
error.OutOfMemory => return error.OutOfMemory,
|
||||||
)) {
|
};
|
||||||
std.os.windows.S_OK => {
|
|
||||||
defer std.os.windows.ole32.CoTaskMemFree(@as(*anyopaque, @ptrCast(dir_path_ptr)));
|
//defer allocator.free(local_app_data_dir);
|
||||||
const global_dir = std.unicode.utf16leToUtf8Alloc(allocator, std.mem.sliceTo(dir_path_ptr, 0)) catch |err| switch (err) {
|
//return std.fs.path.join(allocator, &[_][]const u8{ local_app_data_dir, appname });
|
||||||
error.UnexpectedSecondSurrogateHalf => return error.HomeDirUnavailable,
|
return local_app_data_dir;
|
||||||
error.ExpectedSecondSurrogateHalf => return error.HomeDirUnavailable,
|
|
||||||
error.DanglingSurrogateHalf => return error.HomeDirUnavailable,
|
|
||||||
error.OutOfMemory => return error.OutOfMemory,
|
|
||||||
};
|
|
||||||
return global_dir;
|
|
||||||
// defer allocator.free(global_dir);
|
|
||||||
},
|
|
||||||
std.os.windows.E_OUTOFMEMORY => return error.OutOfMemory,
|
|
||||||
else => return error.HomeDirUnavailable,
|
|
||||||
}
|
|
||||||
},
|
},
|
||||||
.macos, .linux, .freebsd, .netbsd, .dragonfly, .openbsd, .solaris => {
|
.macos, .linux, .freebsd, .netbsd, .dragonfly, .openbsd, .solaris => {
|
||||||
const home_dir = std.os.getenv("HOME") orelse {
|
const home_dir = std.os.getenv("HOME") orelse {
|
||||||
|
|
|
@ -64,11 +64,11 @@ const EndPoint = struct {
|
||||||
};
|
};
|
||||||
pub const AwsHttp = struct {
|
pub const AwsHttp = struct {
|
||||||
allocator: std.mem.Allocator,
|
allocator: std.mem.Allocator,
|
||||||
proxy: ?std.http.Client.HttpProxy,
|
proxy: ?std.http.Client.Proxy,
|
||||||
|
|
||||||
const Self = @This();
|
const Self = @This();
|
||||||
|
|
||||||
pub fn init(allocator: std.mem.Allocator, proxy: ?std.http.Client.HttpProxy) Self {
|
pub fn init(allocator: std.mem.Allocator, proxy: ?std.http.Client.Proxy) Self {
|
||||||
return Self{
|
return Self{
|
||||||
.allocator = allocator,
|
.allocator = allocator,
|
||||||
.proxy = proxy,
|
.proxy = proxy,
|
||||||
|
@ -175,7 +175,7 @@ pub const AwsHttp = struct {
|
||||||
const url = try std.fmt.allocPrint(self.allocator, "{s}{s}{s}", .{ endpoint.uri, request_cp.path, request_cp.query });
|
const url = try std.fmt.allocPrint(self.allocator, "{s}{s}{s}", .{ endpoint.uri, request_cp.path, request_cp.query });
|
||||||
defer self.allocator.free(url);
|
defer self.allocator.free(url);
|
||||||
log.debug("Request url: {s}", .{url});
|
log.debug("Request url: {s}", .{url});
|
||||||
var cl = std.http.Client{ .allocator = self.allocator, .proxy = self.proxy };
|
var cl = std.http.Client{ .allocator = self.allocator, .http_proxy = self.proxy };
|
||||||
defer cl.deinit(); // TODO: Connection pooling
|
defer cl.deinit(); // TODO: Connection pooling
|
||||||
//
|
//
|
||||||
// var req = try zfetch.Request.init(self.allocator, url, self.trust_chain);
|
// var req = try zfetch.Request.init(self.allocator, url, self.trust_chain);
|
||||||
|
@ -209,11 +209,11 @@ pub const AwsHttp = struct {
|
||||||
//
|
//
|
||||||
// const unescaped_url = try std.Uri.unescapeString(self.allocator, url);
|
// const unescaped_url = try std.Uri.unescapeString(self.allocator, url);
|
||||||
// defer self.allocator.free(unescaped_url);
|
// defer self.allocator.free(unescaped_url);
|
||||||
var req = try cl.request(method, try std.Uri.parse(url), headers, .{});
|
var req = try cl.open(method, try std.Uri.parse(url), headers, .{});
|
||||||
defer req.deinit();
|
defer req.deinit();
|
||||||
if (request_cp.body.len > 0)
|
if (request_cp.body.len > 0)
|
||||||
req.transfer_encoding = .{ .content_length = request_cp.body.len };
|
req.transfer_encoding = .{ .content_length = request_cp.body.len };
|
||||||
try @import("http_client_17015_issue.zig").start(&req);
|
try @import("http_client_17015_issue.zig").send(&req);
|
||||||
// try req.start();
|
// try req.start();
|
||||||
if (request_cp.body.len > 0) {
|
if (request_cp.body.len > 0) {
|
||||||
// Workaround for https://github.com/ziglang/zig/issues/15626
|
// Workaround for https://github.com/ziglang/zig/issues/15626
|
||||||
|
@ -250,12 +250,12 @@ pub const AwsHttp = struct {
|
||||||
content_length = std.fmt.parseInt(usize, h.value, 10) catch 0;
|
content_length = std.fmt.parseInt(usize, h.value, 10) catch 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
var response_data: []u8 =
|
const response_data: []u8 =
|
||||||
if (req.response.transfer_encoding) |_| // the only value here is "chunked"
|
if (req.response.transfer_encoding == .none) // the only value here is "chunked"
|
||||||
try req.reader().readAllAlloc(self.allocator, std.math.maxInt(usize))
|
try req.reader().readAllAlloc(self.allocator, std.math.maxInt(usize))
|
||||||
else blk: {
|
else blk: {
|
||||||
// content length
|
// content length
|
||||||
var tmp_data = try self.allocator.alloc(u8, content_length);
|
const tmp_data = try self.allocator.alloc(u8, content_length);
|
||||||
errdefer self.allocator.free(tmp_data);
|
errdefer self.allocator.free(tmp_data);
|
||||||
_ = try req.readAll(tmp_data);
|
_ = try req.readAll(tmp_data);
|
||||||
break :blk tmp_data;
|
break :blk tmp_data;
|
||||||
|
|
|
@ -177,7 +177,7 @@ pub fn signRequest(allocator: std.mem.Allocator, request: base.Request, config:
|
||||||
};
|
};
|
||||||
}
|
}
|
||||||
errdefer freeSignedRequest(allocator, &rc, config);
|
errdefer freeSignedRequest(allocator, &rc, config);
|
||||||
std.mem.copy(base.Header, newheaders, oldheaders);
|
@memcpy(newheaders[0..oldheaders.len], oldheaders);
|
||||||
newheaders[newheaders.len - 3] = base.Header{
|
newheaders[newheaders.len - 3] = base.Header{
|
||||||
.name = "X-Amz-Date",
|
.name = "X-Amz-Date",
|
||||||
.value = signing_iso8601,
|
.value = signing_iso8601,
|
||||||
|
@ -364,7 +364,7 @@ fn verifyParsedAuthorization(
|
||||||
const service = credential_iterator.next().?;
|
const service = credential_iterator.next().?;
|
||||||
const aws4_request = credential_iterator.next().?;
|
const aws4_request = credential_iterator.next().?;
|
||||||
if (!std.mem.eql(u8, aws4_request, "aws4_request")) return error.UnexpectedCredentialValue;
|
if (!std.mem.eql(u8, aws4_request, "aws4_request")) return error.UnexpectedCredentialValue;
|
||||||
var config = Config{
|
const config = Config{
|
||||||
.service = service,
|
.service = service,
|
||||||
.credentials = credentials,
|
.credentials = credentials,
|
||||||
.region = region,
|
.region = region,
|
||||||
|
@ -444,7 +444,7 @@ fn getSigningKey(allocator: std.mem.Allocator, signing_date: []const u8, config:
|
||||||
\\ region: {s}
|
\\ region: {s}
|
||||||
\\ service: {s}
|
\\ service: {s}
|
||||||
, .{ signing_date, config.region, config.service });
|
, .{ signing_date, config.region, config.service });
|
||||||
var secret = try std.fmt.allocPrint(allocator, "AWS4{s}", .{config.credentials.secret_key});
|
const secret = try std.fmt.allocPrint(allocator, "AWS4{s}", .{config.credentials.secret_key});
|
||||||
defer {
|
defer {
|
||||||
// secureZero avoids compiler optimizations that may say
|
// secureZero avoids compiler optimizations that may say
|
||||||
// "WTF are you doing this thing? Looks like nothing to me. It's silly and we will remove it"
|
// "WTF are you doing this thing? Looks like nothing to me. It's silly and we will remove it"
|
||||||
|
@ -716,7 +716,7 @@ fn canonicalQueryString(allocator: std.mem.Allocator, path: []const u8) ![]const
|
||||||
for (sort_me.items) |i| {
|
for (sort_me.items) |i| {
|
||||||
if (!first) try normalized.append('&');
|
if (!first) try normalized.append('&');
|
||||||
first = false;
|
first = false;
|
||||||
var first_equals = std.mem.indexOf(u8, i, "=");
|
const first_equals = std.mem.indexOf(u8, i, "=");
|
||||||
if (first_equals == null) {
|
if (first_equals == null) {
|
||||||
// Rare. This is "foo="
|
// Rare. This is "foo="
|
||||||
const normed_item = try encodeUri(allocator, i);
|
const normed_item = try encodeUri(allocator, i);
|
||||||
|
@ -744,7 +744,7 @@ fn canonicalQueryString(allocator: std.mem.Allocator, path: []const u8) ![]const
|
||||||
}
|
}
|
||||||
|
|
||||||
fn replace(allocator: std.mem.Allocator, haystack: []const u8, needle: []const u8, replacement_value: []const u8) ![]const u8 {
|
fn replace(allocator: std.mem.Allocator, haystack: []const u8, needle: []const u8, replacement_value: []const u8) ![]const u8 {
|
||||||
var buffer = try allocator.alloc(u8, std.mem.replacementSize(u8, haystack, needle, replacement_value));
|
const buffer = try allocator.alloc(u8, std.mem.replacementSize(u8, haystack, needle, replacement_value));
|
||||||
_ = std.mem.replace(u8, haystack, needle, replacement_value, buffer);
|
_ = std.mem.replace(u8, haystack, needle, replacement_value, buffer);
|
||||||
return buffer;
|
return buffer;
|
||||||
}
|
}
|
||||||
|
@ -837,7 +837,7 @@ fn canonicalHeaders(allocator: std.mem.Allocator, headers: []base.Header, servic
|
||||||
|
|
||||||
fn canonicalHeaderValue(allocator: std.mem.Allocator, value: []const u8) ![]const u8 {
|
fn canonicalHeaderValue(allocator: std.mem.Allocator, value: []const u8) ![]const u8 {
|
||||||
var started = false;
|
var started = false;
|
||||||
var in_quote = false;
|
const in_quote = false;
|
||||||
var start: usize = 0;
|
var start: usize = 0;
|
||||||
const rc = try allocator.alloc(u8, value.len);
|
const rc = try allocator.alloc(u8, value.len);
|
||||||
var rc_inx: usize = 0;
|
var rc_inx: usize = 0;
|
||||||
|
@ -1002,7 +1002,7 @@ test "can sign" {
|
||||||
try headers.append(.{ .name = "Content-Type", .value = "application/x-www-form-urlencoded; charset=utf-8" });
|
try headers.append(.{ .name = "Content-Type", .value = "application/x-www-form-urlencoded; charset=utf-8" });
|
||||||
try headers.append(.{ .name = "Content-Length", .value = "13" });
|
try headers.append(.{ .name = "Content-Length", .value = "13" });
|
||||||
try headers.append(.{ .name = "Host", .value = "example.amazonaws.com" });
|
try headers.append(.{ .name = "Host", .value = "example.amazonaws.com" });
|
||||||
var req = base.Request{
|
const req = base.Request{
|
||||||
.path = "/",
|
.path = "/",
|
||||||
.query = "",
|
.query = "",
|
||||||
.body = "Param1=value1",
|
.body = "Param1=value1",
|
||||||
|
@ -1071,7 +1071,7 @@ test "can verify server request" {
|
||||||
|
|
||||||
var buf = "bar".*;
|
var buf = "bar".*;
|
||||||
var fis = std.io.fixedBufferStream(&buf);
|
var fis = std.io.fixedBufferStream(&buf);
|
||||||
var request = std.http.Server.Request{
|
const request = std.http.Server.Request{
|
||||||
.method = std.http.Method.PUT,
|
.method = std.http.Method.PUT,
|
||||||
.target = "/mysfitszj3t6webstack-hostingbucketa91a61fe-1ep3ezkgwpxr0/i/am/a/teapot/foo?x-id=PutObject",
|
.target = "/mysfitszj3t6webstack-hostingbucketa91a61fe-1ep3ezkgwpxr0/i/am/a/teapot/foo?x-id=PutObject",
|
||||||
.version = .@"HTTP/1.1",
|
.version = .@"HTTP/1.1",
|
||||||
|
|
16
src/date.zig
16
src/date.zig
|
@ -21,7 +21,7 @@ pub fn timestampToDateTime(timestamp: i64) DateTime {
|
||||||
const DAY_NUMBER_ADJUSTED_1970_01_01 = 719468; //* Day number relates to March 1st */
|
const DAY_NUMBER_ADJUSTED_1970_01_01 = 719468; //* Day number relates to March 1st */
|
||||||
|
|
||||||
var dayN: u64 = DAY_NUMBER_ADJUSTED_1970_01_01 + unixtime / SECONDS_PER_DAY;
|
var dayN: u64 = DAY_NUMBER_ADJUSTED_1970_01_01 + unixtime / SECONDS_PER_DAY;
|
||||||
var seconds_since_midnight: u64 = unixtime % SECONDS_PER_DAY;
|
const seconds_since_midnight: u64 = unixtime % SECONDS_PER_DAY;
|
||||||
var temp: u64 = 0;
|
var temp: u64 = 0;
|
||||||
|
|
||||||
// Leap year rules for Gregorian Calendars
|
// Leap year rules for Gregorian Calendars
|
||||||
|
@ -37,7 +37,7 @@ pub fn timestampToDateTime(timestamp: i64) DateTime {
|
||||||
|
|
||||||
// dayN calculates the days of the year in relation to March 1
|
// dayN calculates the days of the year in relation to March 1
|
||||||
var month = @as(u8, @intCast((5 * dayN + 2) / 153));
|
var month = @as(u8, @intCast((5 * dayN + 2) / 153));
|
||||||
var day = @as(u8, @intCast(dayN - (@as(u64, @intCast(month)) * 153 + 2) / 5 + 1));
|
const day = @as(u8, @intCast(dayN - (@as(u64, @intCast(month)) * 153 + 2) / 5 + 1));
|
||||||
// 153 = 31+30+31+30+31 Days for the 5 months from March through July
|
// 153 = 31+30+31+30+31 Days for the 5 months from March through July
|
||||||
// 153 = 31+30+31+30+31 Days for the 5 months from August through December
|
// 153 = 31+30+31+30+31 Days for the 5 months from August through December
|
||||||
// 31+28 Days for January and February (see below)
|
// 31+28 Days for January and February (see below)
|
||||||
|
@ -50,9 +50,9 @@ pub fn timestampToDateTime(timestamp: i64) DateTime {
|
||||||
year += 1;
|
year += 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
var hours = @as(u8, @intCast(seconds_since_midnight / 3600));
|
const hours = @as(u8, @intCast(seconds_since_midnight / 3600));
|
||||||
var minutes = @as(u8, @intCast(seconds_since_midnight % 3600 / 60));
|
const minutes = @as(u8, @intCast(seconds_since_midnight % 3600 / 60));
|
||||||
var seconds = @as(u8, @intCast(seconds_since_midnight % 60));
|
const seconds = @as(u8, @intCast(seconds_since_midnight % 60));
|
||||||
|
|
||||||
return DateTime{ .day = day, .month = month, .year = year, .hour = hours, .minute = minutes, .second = seconds };
|
return DateTime{ .day = day, .month = month, .year = year, .hour = hours, .minute = minutes, .second = seconds };
|
||||||
}
|
}
|
||||||
|
@ -275,10 +275,10 @@ fn secondsBetween(start: DateTime, end: DateTime) DateTimeToTimestampError!i64 {
|
||||||
return (try secondsBetween(new_start, end)) - seconds_into_start_year;
|
return (try secondsBetween(new_start, end)) - seconds_into_start_year;
|
||||||
}
|
}
|
||||||
const leap_years_between = leapYearsBetween(start.year, end.year);
|
const leap_years_between = leapYearsBetween(start.year, end.year);
|
||||||
var add_days: u1 = 0;
|
const add_days: u1 = 0;
|
||||||
const years_diff = end.year - start.year;
|
const years_diff = end.year - start.year;
|
||||||
// log.debug("Years from epoch: {d}, Leap years: {d}", .{ years_diff, leap_years_between });
|
// log.debug("Years from epoch: {d}, Leap years: {d}", .{ years_diff, leap_years_between });
|
||||||
var days_diff: i32 = (years_diff * DAYS_PER_YEAR) + leap_years_between + add_days;
|
const days_diff: i32 = (years_diff * DAYS_PER_YEAR) + leap_years_between + add_days;
|
||||||
// log.debug("Days with leap year, without month: {d}", .{days_diff});
|
// log.debug("Days with leap year, without month: {d}", .{days_diff});
|
||||||
|
|
||||||
const seconds_into_year = secondsFromBeginningOfYear(
|
const seconds_into_year = secondsFromBeginningOfYear(
|
||||||
|
@ -306,7 +306,7 @@ fn secondsFromBeginningOfYear(year: u16, month: u8, day: u8, hour: u8, minute: u
|
||||||
const normal_days_per_month: [12]u5 = .{ 31, 28, 31, 30, 31, 30, 31, 31, 30, 31, 30, 31 };
|
const normal_days_per_month: [12]u5 = .{ 31, 28, 31, 30, 31, 30, 31, 31, 30, 31, 30, 31 };
|
||||||
const days_per_month = if (current_year_is_leap_year) leap_year_days_per_month else normal_days_per_month;
|
const days_per_month = if (current_year_is_leap_year) leap_year_days_per_month else normal_days_per_month;
|
||||||
var current_month: usize = 1;
|
var current_month: usize = 1;
|
||||||
var end_month = month;
|
const end_month = month;
|
||||||
var days_diff: u32 = 0;
|
var days_diff: u32 = 0;
|
||||||
while (current_month != end_month) {
|
while (current_month != end_month) {
|
||||||
days_diff += days_per_month[current_month - 1]; // months are 1-based vs array is 0-based
|
days_diff += days_per_month[current_month - 1]; // months are 1-based vs array is 0-based
|
||||||
|
|
|
@ -10,8 +10,8 @@ const Uri = std.Uri;
|
||||||
/// only the two w.print lines for req.uri 16 and 18 lines down from this comment
|
/// only the two w.print lines for req.uri 16 and 18 lines down from this comment
|
||||||
///////////////////////////////////////////////////////////////////////////
|
///////////////////////////////////////////////////////////////////////////
|
||||||
/// Send the request to the server.
|
/// Send the request to the server.
|
||||||
pub fn start(req: *std.http.Client.Request) std.http.Client.Request.StartError!void {
|
pub fn send(req: *std.http.Client.Request) std.http.Client.Request.SendError!void {
|
||||||
var buffered = std.io.bufferedWriter(req.connection.?.data.writer());
|
var buffered = std.io.bufferedWriter(req.connection.?.writer());
|
||||||
const w = buffered.writer();
|
const w = buffered.writer();
|
||||||
|
|
||||||
try w.writeAll(@tagName(req.method));
|
try w.writeAll(@tagName(req.method));
|
||||||
|
@ -21,7 +21,7 @@ pub fn start(req: *std.http.Client.Request) std.http.Client.Request.StartError!v
|
||||||
try w.writeAll(req.uri.host.?);
|
try w.writeAll(req.uri.host.?);
|
||||||
try w.writeByte(':');
|
try w.writeByte(':');
|
||||||
try w.print("{}", .{req.uri.port.?});
|
try w.print("{}", .{req.uri.port.?});
|
||||||
} else if (req.connection.?.data.proxied) {
|
} else if (req.connection.?.proxied) {
|
||||||
// proxied connections require the full uri
|
// proxied connections require the full uri
|
||||||
try format(req.uri, "+/", .{}, w);
|
try format(req.uri, "+/", .{}, w);
|
||||||
} else {
|
} else {
|
||||||
|
|
18
src/json.zig
18
src/json.zig
|
@ -1762,7 +1762,7 @@ fn parseInternal(comptime T: type, token: Token, tokens: *TokenStream, options:
|
||||||
var r: T = undefined;
|
var r: T = undefined;
|
||||||
const source_slice = stringToken.slice(tokens.slice, tokens.i - 1);
|
const source_slice = stringToken.slice(tokens.slice, tokens.i - 1);
|
||||||
switch (stringToken.escapes) {
|
switch (stringToken.escapes) {
|
||||||
.None => mem.copy(u8, &r, source_slice),
|
.None => @memcpy(&r, source_slice),
|
||||||
.Some => try unescapeValidString(&r, source_slice),
|
.Some => try unescapeValidString(&r, source_slice),
|
||||||
}
|
}
|
||||||
return r;
|
return r;
|
||||||
|
@ -2019,7 +2019,7 @@ test "parse into tagged union" {
|
||||||
}
|
}
|
||||||
|
|
||||||
{ // failing allocations should be bubbled up instantly without trying next member
|
{ // failing allocations should be bubbled up instantly without trying next member
|
||||||
var fail_alloc = testing.FailingAllocator.init(testing.allocator, 0);
|
var fail_alloc = testing.FailingAllocator.init(testing.allocator, .{});
|
||||||
const options = ParseOptions{ .allocator = fail_alloc.allocator() };
|
const options = ParseOptions{ .allocator = fail_alloc.allocator() };
|
||||||
const T = union(enum) {
|
const T = union(enum) {
|
||||||
// both fields here match the input
|
// both fields here match the input
|
||||||
|
@ -2067,7 +2067,7 @@ test "parse union bubbles up AllocatorRequired" {
|
||||||
}
|
}
|
||||||
|
|
||||||
test "parseFree descends into tagged union" {
|
test "parseFree descends into tagged union" {
|
||||||
var fail_alloc = testing.FailingAllocator.init(testing.allocator, 1);
|
var fail_alloc = testing.FailingAllocator.init(testing.allocator, .{});
|
||||||
const options = ParseOptions{ .allocator = fail_alloc.allocator() };
|
const options = ParseOptions{ .allocator = fail_alloc.allocator() };
|
||||||
const T = union(enum) {
|
const T = union(enum) {
|
||||||
int: i32,
|
int: i32,
|
||||||
|
@ -2299,7 +2299,7 @@ pub const Parser = struct {
|
||||||
},
|
},
|
||||||
.ObjectValue => {
|
.ObjectValue => {
|
||||||
var object = &p.stack.items[p.stack.items.len - 2].Object;
|
var object = &p.stack.items[p.stack.items.len - 2].Object;
|
||||||
var key = p.stack.items[p.stack.items.len - 1].String;
|
const key = p.stack.items[p.stack.items.len - 1].String;
|
||||||
|
|
||||||
switch (token) {
|
switch (token) {
|
||||||
.ObjectBegin => {
|
.ObjectBegin => {
|
||||||
|
@ -2827,14 +2827,14 @@ pub fn stringify(
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
.Enum => {
|
.Enum => {
|
||||||
if (comptime std.meta.trait.hasFn("jsonStringify")(T)) {
|
if (comptime std.meta.hasFn(T, "jsonStringify")) {
|
||||||
return value.jsonStringify(options, out_stream);
|
return value.jsonStringify(options, out_stream);
|
||||||
}
|
}
|
||||||
|
|
||||||
@compileError("Unable to stringify enum '" ++ @typeName(T) ++ "'");
|
@compileError("Unable to stringify enum '" ++ @typeName(T) ++ "'");
|
||||||
},
|
},
|
||||||
.Union => {
|
.Union => {
|
||||||
if (comptime std.meta.trait.hasFn("jsonStringify")(T)) {
|
if (comptime std.meta.hasFn(T, "jsonStringify")) {
|
||||||
return value.jsonStringify(options, out_stream);
|
return value.jsonStringify(options, out_stream);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -2850,7 +2850,7 @@ pub fn stringify(
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
.Struct => |S| {
|
.Struct => |S| {
|
||||||
if (comptime std.meta.trait.hasFn("jsonStringify")(T)) {
|
if (comptime std.meta.hasFn(T, "jsonStringify")) {
|
||||||
return value.jsonStringify(options, out_stream);
|
return value.jsonStringify(options, out_stream);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -2874,11 +2874,11 @@ pub fn stringify(
|
||||||
try child_whitespace.outputIndent(out_stream);
|
try child_whitespace.outputIndent(out_stream);
|
||||||
}
|
}
|
||||||
var field_written = false;
|
var field_written = false;
|
||||||
if (comptime std.meta.trait.hasFn("jsonStringifyField")(T))
|
if (comptime std.meta.hasFn(T, "jsonStringifyField"))
|
||||||
field_written = try value.jsonStringifyField(Field.name, child_options, out_stream);
|
field_written = try value.jsonStringifyField(Field.name, child_options, out_stream);
|
||||||
|
|
||||||
if (!field_written) {
|
if (!field_written) {
|
||||||
if (comptime std.meta.trait.hasFn("fieldNameFor")(T)) {
|
if (comptime std.meta.hasFn(T, "fieldNameFor")) {
|
||||||
const name = value.fieldNameFor(Field.name);
|
const name = value.fieldNameFor(Field.name);
|
||||||
try stringify(name, options, out_stream);
|
try stringify(name, options, out_stream);
|
||||||
} else {
|
} else {
|
||||||
|
|
11
src/main.zig
11
src/main.zig
|
@ -71,7 +71,7 @@ pub fn main() anyerror!void {
|
||||||
defer bw.flush() catch unreachable;
|
defer bw.flush() catch unreachable;
|
||||||
const stdout = bw.writer();
|
const stdout = bw.writer();
|
||||||
var arg0: ?[]const u8 = null;
|
var arg0: ?[]const u8 = null;
|
||||||
var proxy: ?std.http.Client.HttpProxy = null;
|
var proxy: ?std.http.Client.Proxy = null;
|
||||||
while (args.next()) |arg| {
|
while (args.next()) |arg| {
|
||||||
if (arg0 == null) arg0 = arg;
|
if (arg0 == null) arg0 = arg;
|
||||||
if (std.mem.eql(u8, "-h", arg) or std.mem.eql(u8, "--help", arg)) {
|
if (std.mem.eql(u8, "-h", arg) or std.mem.eql(u8, "--help", arg)) {
|
||||||
|
@ -87,7 +87,7 @@ pub fn main() anyerror!void {
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
if (std.mem.eql(u8, "-x", arg) or std.mem.eql(u8, "--proxy", arg)) {
|
if (std.mem.eql(u8, "-x", arg) or std.mem.eql(u8, "--proxy", arg)) {
|
||||||
proxy = try proxyFromString(args.next().?); // parse stuff
|
proxy = try proxyFromString(allocator, args.next().?); // parse stuff
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
if (std.mem.startsWith(u8, arg, "-v")) {
|
if (std.mem.startsWith(u8, arg, "-v")) {
|
||||||
|
@ -353,10 +353,13 @@ pub fn main() anyerror!void {
|
||||||
std.log.info("===== Tests complete =====", .{});
|
std.log.info("===== Tests complete =====", .{});
|
||||||
}
|
}
|
||||||
|
|
||||||
fn proxyFromString(string: []const u8) !std.http.Client.HttpProxy {
|
fn proxyFromString(allocator: std.mem.Allocator, string: []const u8) !std.http.Client.Proxy {
|
||||||
var rc = std.http.Client.HttpProxy{
|
var rc = std.http.Client.Proxy{
|
||||||
|
.headers = std.http.Headers{ .allocator = allocator },
|
||||||
|
.allocator = allocator,
|
||||||
.protocol = undefined,
|
.protocol = undefined,
|
||||||
.host = undefined,
|
.host = undefined,
|
||||||
|
.port = undefined,
|
||||||
};
|
};
|
||||||
var remaining: []const u8 = string;
|
var remaining: []const u8 = string;
|
||||||
if (std.mem.startsWith(u8, string, "http://")) {
|
if (std.mem.startsWith(u8, string, "http://")) {
|
||||||
|
|
|
@ -260,7 +260,7 @@ const ParseContext = struct {
|
||||||
begin = prev_nl + 1;
|
begin = prev_nl + 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
var end = mem.indexOfScalarPos(u8, self.source, self.offset, '\n') orelse self.source.len;
|
const end = mem.indexOfScalarPos(u8, self.source, self.offset, '\n') orelse self.source.len;
|
||||||
return self.source[begin..end];
|
return self.source[begin..end];
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
|
|
|
@ -214,9 +214,9 @@ fn parseInternal(comptime T: type, element: *xml.Element, options: ParseOptions)
|
||||||
|
|
||||||
log.debug("Processing fields in struct: {s}", .{@typeName(T)});
|
log.debug("Processing fields in struct: {s}", .{@typeName(T)});
|
||||||
inline for (struct_info.fields, 0..) |field, i| {
|
inline for (struct_info.fields, 0..) |field, i| {
|
||||||
var name = field.name;
|
var name: []const u8 = field.name;
|
||||||
var found_value = false;
|
var found_value = false;
|
||||||
if (comptime std.meta.trait.hasFn("fieldNameFor")(T))
|
if (comptime std.meta.hasFn(T, "fieldNameFor"))
|
||||||
name = r.fieldNameFor(field.name);
|
name = r.fieldNameFor(field.name);
|
||||||
log.debug("Field name: {s}, Element: {s}, Adjusted field name: {s}", .{ field.name, element.tag, name });
|
log.debug("Field name: {s}, Element: {s}, Adjusted field name: {s}", .{ field.name, element.tag, name });
|
||||||
var iterator = element.findChildrenByTag(name);
|
var iterator = element.findChildrenByTag(name);
|
||||||
|
|
Loading…
Reference in New Issue
Block a user