Merge branch 'main' into more-mcp-tools

This commit is contained in:
Adrià Arrufat
2026-03-17 13:26:44 +09:00
48 changed files with 685 additions and 389 deletions

View File

@@ -13,7 +13,7 @@ inputs:
zig-v8: zig-v8:
description: 'zig v8 version to install' description: 'zig v8 version to install'
required: false required: false
default: 'v0.3.3' default: 'v0.3.4'
v8: v8:
description: 'v8 version to install' description: 'v8 version to install'
required: false required: false
@@ -46,7 +46,7 @@ runs:
- name: Cache v8 - name: Cache v8
id: cache-v8 id: cache-v8
uses: actions/cache@v4 uses: actions/cache@v5
env: env:
cache-name: cache-v8 cache-name: cache-v8
with: with:

View File

@@ -72,7 +72,7 @@ jobs:
timeout-minutes: 20 timeout-minutes: 20
steps: steps:
- uses: actions/checkout@v4 - uses: actions/checkout@v6
with: with:
fetch-depth: 0 fetch-depth: 0
# fetch submodules recusively, to get zig-js-runtime submodules also. # fetch submodules recusively, to get zig-js-runtime submodules also.
@@ -116,7 +116,7 @@ jobs:
timeout-minutes: 20 timeout-minutes: 20
steps: steps:
- uses: actions/checkout@v4 - uses: actions/checkout@v6
with: with:
fetch-depth: 0 fetch-depth: 0
# fetch submodules recusively, to get zig-js-runtime submodules also. # fetch submodules recusively, to get zig-js-runtime submodules also.
@@ -158,7 +158,7 @@ jobs:
timeout-minutes: 20 timeout-minutes: 20
steps: steps:
- uses: actions/checkout@v4 - uses: actions/checkout@v6
with: with:
fetch-depth: 0 fetch-depth: 0
# fetch submodules recusively, to get zig-js-runtime submodules also. # fetch submodules recusively, to get zig-js-runtime submodules also.

View File

@@ -20,7 +20,7 @@ jobs:
if: github.event.pull_request.draft == false if: github.event.pull_request.draft == false
steps: steps:
- uses: actions/checkout@v4 - uses: actions/checkout@v6
with: with:
fetch-depth: 0 fetch-depth: 0
# fetch submodules recusively, to get zig-js-runtime submodules also. # fetch submodules recusively, to get zig-js-runtime submodules also.
@@ -32,7 +32,7 @@ jobs:
run: zig build -Dprebuilt_v8_path=v8/libc_v8.a -Doptimize=ReleaseFast -Dcpu=x86_64 -Dgit_commit=$(git rev-parse --short ${{ github.sha }}) run: zig build -Dprebuilt_v8_path=v8/libc_v8.a -Doptimize=ReleaseFast -Dcpu=x86_64 -Dgit_commit=$(git rev-parse --short ${{ github.sha }})
- name: upload artifact - name: upload artifact
uses: actions/upload-artifact@v4 uses: actions/upload-artifact@v7
with: with:
name: lightpanda-build-release name: lightpanda-build-release
path: | path: |
@@ -47,7 +47,7 @@ jobs:
timeout-minutes: 15 timeout-minutes: 15
steps: steps:
- uses: actions/checkout@v4 - uses: actions/checkout@v6
with: with:
repository: 'lightpanda-io/demo' repository: 'lightpanda-io/demo'
fetch-depth: 0 fetch-depth: 0
@@ -55,7 +55,7 @@ jobs:
- run: npm install - run: npm install
- name: download artifact - name: download artifact
uses: actions/download-artifact@v4 uses: actions/download-artifact@v8
with: with:
name: lightpanda-build-release name: lightpanda-build-release

View File

@@ -61,7 +61,7 @@ jobs:
run: zig build -Dprebuilt_v8_path=v8/libc_v8.a -Doptimize=ReleaseFast -Dcpu=x86_64 -Dgit_commit=$(git rev-parse --short ${{ github.sha }}) run: zig build -Dprebuilt_v8_path=v8/libc_v8.a -Doptimize=ReleaseFast -Dcpu=x86_64 -Dgit_commit=$(git rev-parse --short ${{ github.sha }})
- name: upload artifact - name: upload artifact
uses: actions/upload-artifact@v4 uses: actions/upload-artifact@v7
with: with:
name: lightpanda-build-release name: lightpanda-build-release
path: | path: |
@@ -76,7 +76,7 @@ jobs:
timeout-minutes: 15 timeout-minutes: 15
steps: steps:
- uses: actions/checkout@v4 - uses: actions/checkout@v6
with: with:
repository: 'lightpanda-io/demo' repository: 'lightpanda-io/demo'
fetch-depth: 0 fetch-depth: 0
@@ -84,7 +84,7 @@ jobs:
- run: npm install - run: npm install
- name: download artifact - name: download artifact
uses: actions/download-artifact@v4 uses: actions/download-artifact@v8
with: with:
name: lightpanda-build-release name: lightpanda-build-release
@@ -126,7 +126,7 @@ jobs:
timeout-minutes: 15 timeout-minutes: 15
steps: steps:
- uses: actions/checkout@v4 - uses: actions/checkout@v6
with: with:
repository: 'lightpanda-io/demo' repository: 'lightpanda-io/demo'
fetch-depth: 0 fetch-depth: 0
@@ -134,7 +134,7 @@ jobs:
- run: npm install - run: npm install
- name: download artifact - name: download artifact
uses: actions/download-artifact@v4 uses: actions/download-artifact@v8
with: with:
name: lightpanda-build-release name: lightpanda-build-release
@@ -189,32 +189,35 @@ jobs:
timeout-minutes: 5 timeout-minutes: 5
steps: steps:
- uses: actions/checkout@v4 - uses: actions/checkout@v6
with: with:
repository: 'lightpanda-io/demo' repository: 'lightpanda-io/demo'
fetch-depth: 0 fetch-depth: 0
- run: echo "${{ secrets.WBA_PRIVATE_KEY_PEM }}" > private_key.pem
- name: download artifact - name: download artifact
uses: actions/download-artifact@v4 uses: actions/download-artifact@v8
with: with:
name: lightpanda-build-release name: lightpanda-build-release
- run: chmod a+x ./lightpanda - run: chmod a+x ./lightpanda
- name: run wba test - name: run wba test
shell: bash
run: | run: |
node webbotauth/validator.js & node webbotauth/validator.js &
VALIDATOR_PID=$! VALIDATOR_PID=$!
sleep 2 sleep 2
./lightpanda fetch http://127.0.0.1:8989/ \ exec 3<<< "${{ secrets.WBA_PRIVATE_KEY_PEM }}"
--web_bot_auth_key_file private_key.pem \
./lightpanda fetch --dump http://127.0.0.1:8989/ \
--web_bot_auth_key_file /proc/self/fd/3 \
--web_bot_auth_keyid ${{ vars.WBA_KEY_ID }} \ --web_bot_auth_keyid ${{ vars.WBA_KEY_ID }} \
--web_bot_auth_domain ${{ vars.WBA_DOMAIN }} --web_bot_auth_domain ${{ vars.WBA_DOMAIN }}
wait $VALIDATOR_PID wait $VALIDATOR_PID
exec 3>&-
cdp-and-hyperfine-bench: cdp-and-hyperfine-bench:
name: cdp-and-hyperfine-bench name: cdp-and-hyperfine-bench
@@ -239,7 +242,7 @@ jobs:
timeout-minutes: 15 timeout-minutes: 15
steps: steps:
- uses: actions/checkout@v4 - uses: actions/checkout@v6
with: with:
repository: 'lightpanda-io/demo' repository: 'lightpanda-io/demo'
fetch-depth: 0 fetch-depth: 0
@@ -247,7 +250,7 @@ jobs:
- run: npm install - run: npm install
- name: download artifact - name: download artifact
uses: actions/download-artifact@v4 uses: actions/download-artifact@v8
with: with:
name: lightpanda-build-release name: lightpanda-build-release
@@ -333,7 +336,7 @@ jobs:
echo "${{github.sha}}" > commit.txt echo "${{github.sha}}" > commit.txt
- name: upload artifact - name: upload artifact
uses: actions/upload-artifact@v4 uses: actions/upload-artifact@v7
with: with:
name: bench-results name: bench-results
path: | path: |
@@ -361,7 +364,7 @@ jobs:
steps: steps:
- name: download artifact - name: download artifact
uses: actions/download-artifact@v4 uses: actions/download-artifact@v8
with: with:
name: bench-results name: bench-results
@@ -379,7 +382,7 @@ jobs:
steps: steps:
- name: download artifact - name: download artifact
uses: actions/download-artifact@v4 uses: actions/download-artifact@v8
with: with:
name: lightpanda-build-release name: lightpanda-build-release

View File

@@ -35,7 +35,7 @@ jobs:
run: zig build -Dprebuilt_v8_path=v8/libc_v8.a -Doptimize=ReleaseFast -Dcpu=x86_64 -Dgit_commit=$(git rev-parse --short ${{ github.sha }}) run: zig build -Dprebuilt_v8_path=v8/libc_v8.a -Doptimize=ReleaseFast -Dcpu=x86_64 -Dgit_commit=$(git rev-parse --short ${{ github.sha }})
- name: upload artifact - name: upload artifact
uses: actions/upload-artifact@v4 uses: actions/upload-artifact@v7
with: with:
name: lightpanda-build-release name: lightpanda-build-release
path: | path: |
@@ -59,7 +59,7 @@ jobs:
CGO_ENABLED=0 go build CGO_ENABLED=0 go build
- name: upload artifact - name: upload artifact
uses: actions/upload-artifact@v4 uses: actions/upload-artifact@v7
with: with:
name: wptrunner name: wptrunner
path: | path: |
@@ -91,14 +91,14 @@ jobs:
run: ./wpt manifest run: ./wpt manifest
- name: download lightpanda release - name: download lightpanda release
uses: actions/download-artifact@v4 uses: actions/download-artifact@v8
with: with:
name: lightpanda-build-release name: lightpanda-build-release
- run: chmod a+x ./lightpanda - run: chmod a+x ./lightpanda
- name: download wptrunner - name: download wptrunner
uses: actions/download-artifact@v4 uses: actions/download-artifact@v8
with: with:
name: wptrunner name: wptrunner
@@ -116,7 +116,7 @@ jobs:
echo "${{github.sha}}" > commit.txt echo "${{github.sha}}" > commit.txt
- name: upload artifact - name: upload artifact
uses: actions/upload-artifact@v4 uses: actions/upload-artifact@v7
with: with:
name: wpt-results name: wpt-results
path: | path: |
@@ -139,7 +139,7 @@ jobs:
steps: steps:
- name: download artifact - name: download artifact
uses: actions/download-artifact@v4 uses: actions/download-artifact@v8
with: with:
name: wpt-results name: wpt-results

View File

@@ -44,7 +44,7 @@ jobs:
runs-on: ubuntu-latest runs-on: ubuntu-latest
steps: steps:
- uses: actions/checkout@v4 - uses: actions/checkout@v6
with: with:
fetch-depth: 0 fetch-depth: 0
# fetch submodules recusively, to get zig-js-runtime submodules also. # fetch submodules recusively, to get zig-js-runtime submodules also.
@@ -67,7 +67,7 @@ jobs:
runs-on: ubuntu-latest runs-on: ubuntu-latest
steps: steps:
- uses: actions/checkout@v4 - uses: actions/checkout@v6
with: with:
fetch-depth: 0 fetch-depth: 0
# fetch submodules recusively, to get zig-js-runtime submodules also. # fetch submodules recusively, to get zig-js-runtime submodules also.
@@ -83,7 +83,7 @@ jobs:
echo "${{github.sha}}" > commit.txt echo "${{github.sha}}" > commit.txt
- name: upload artifact - name: upload artifact
uses: actions/upload-artifact@v4 uses: actions/upload-artifact@v7
with: with:
name: bench-results name: bench-results
path: | path: |
@@ -109,7 +109,7 @@ jobs:
steps: steps:
- name: download artifact - name: download artifact
uses: actions/download-artifact@v4 uses: actions/download-artifact@v8
with: with:
name: bench-results name: bench-results

View File

@@ -3,7 +3,7 @@ FROM debian:stable-slim
ARG MINISIG=0.12 ARG MINISIG=0.12
ARG ZIG_MINISIG=RWSGOq2NVecA2UPNdBUZykf1CCb147pkmdtYxgb3Ti+JO/wCYvhbAb/U ARG ZIG_MINISIG=RWSGOq2NVecA2UPNdBUZykf1CCb147pkmdtYxgb3Ti+JO/wCYvhbAb/U
ARG V8=14.0.365.4 ARG V8=14.0.365.4
ARG ZIG_V8=v0.3.3 ARG ZIG_V8=v0.3.4
ARG TARGETPLATFORM ARG TARGETPLATFORM
RUN apt-get update -yq && \ RUN apt-get update -yq && \

View File

@@ -5,8 +5,8 @@
.minimum_zig_version = "0.15.2", .minimum_zig_version = "0.15.2",
.dependencies = .{ .dependencies = .{
.v8 = .{ .v8 = .{
.url = "https://github.com/lightpanda-io/zig-v8-fork/archive/refs/tags/v0.3.3.tar.gz", .url = "https://github.com/lightpanda-io/zig-v8-fork/archive/refs/tags/v0.3.4.tar.gz",
.hash = "v8-0.0.0-xddH6yx3BAAGD9jSoq_ttt_bk9MectTU44s_HZxxE5LD", .hash = "v8-0.0.0-xddH6_F3BAAiFvKY6R1H-gkuQlk19BkDQ0--uZuTrSup",
}, },
// .v8 = .{ .path = "../zig-v8-fork" }, // .v8 = .{ .path = "../zig-v8-fork" },
.brotli = .{ .brotli = .{

View File

@@ -91,25 +91,32 @@ pub fn runMicrotasks(self: *Browser) void {
self.env.runMicrotasks(); self.env.runMicrotasks();
} }
pub fn runMacrotasks(self: *Browser) !?u64 { pub fn runMacrotasks(self: *Browser) !void {
const env = &self.env; const env = &self.env;
const time_to_next = try self.env.runMacrotasks(); try self.env.runMacrotasks();
env.pumpMessageLoop(); env.pumpMessageLoop();
// either of the above could have queued more microtasks // either of the above could have queued more microtasks
env.runMicrotasks(); env.runMicrotasks();
return time_to_next;
} }
pub fn hasBackgroundTasks(self: *Browser) bool { pub fn hasBackgroundTasks(self: *Browser) bool {
return self.env.hasBackgroundTasks(); return self.env.hasBackgroundTasks();
} }
pub fn waitForBackgroundTasks(self: *Browser) void { pub fn waitForBackgroundTasks(self: *Browser) void {
self.env.waitForBackgroundTasks(); self.env.waitForBackgroundTasks();
} }
pub fn msToNextMacrotask(self: *Browser) ?u64 {
return self.env.msToNextMacrotask();
}
pub fn msTo(self: *Browser) bool {
return self.env.hasBackgroundTasks();
}
pub fn runIdleTasks(self: *const Browser) void { pub fn runIdleTasks(self: *const Browser) void {
self.env.runIdleTasks(); self.env.runIdleTasks();
} }

View File

@@ -110,6 +110,8 @@ use_proxy: bool,
// Current TLS verification state, applied per-connection in makeRequest. // Current TLS verification state, applied per-connection in makeRequest.
tls_verify: bool = true, tls_verify: bool = true,
obey_robots: bool,
cdp_client: ?CDPClient = null, cdp_client: ?CDPClient = null,
// libcurl can monitor arbitrary sockets, this lets us use libcurl to poll // libcurl can monitor arbitrary sockets, this lets us use libcurl to poll
@@ -154,6 +156,7 @@ pub fn init(allocator: Allocator, network: *Network) !*Client {
.http_proxy = http_proxy, .http_proxy = http_proxy,
.use_proxy = http_proxy != null, .use_proxy = http_proxy != null,
.tls_verify = network.config.tlsVerifyHost(), .tls_verify = network.config.tlsVerifyHost(),
.obey_robots = network.config.obeyRobots(),
.transfer_pool = transfer_pool, .transfer_pool = transfer_pool,
}; };
@@ -257,34 +260,33 @@ pub fn tick(self: *Client, timeout_ms: u32) !PerformStatus {
} }
pub fn request(self: *Client, req: Request) !void { pub fn request(self: *Client, req: Request) !void {
if (self.network.config.obeyRobots()) { if (self.obey_robots == false) {
const robots_url = try URL.getRobotsUrl(self.allocator, req.url); return self.processRequest(req);
errdefer self.allocator.free(robots_url);
// If we have this robots cached, we can take a fast path.
if (self.network.robot_store.get(robots_url)) |robot_entry| {
defer self.allocator.free(robots_url);
switch (robot_entry) {
// If we have a found robots entry, we check it.
.present => |robots| {
const path = URL.getPathname(req.url);
if (!robots.isAllowed(path)) {
req.error_callback(req.ctx, error.RobotsBlocked);
return;
}
},
// Otherwise, we assume we won't find it again.
.absent => {},
}
return self.processRequest(req);
}
return self.fetchRobotsThenProcessRequest(robots_url, req);
} }
return self.processRequest(req); const robots_url = try URL.getRobotsUrl(self.allocator, req.url);
errdefer self.allocator.free(robots_url);
// If we have this robots cached, we can take a fast path.
if (self.network.robot_store.get(robots_url)) |robot_entry| {
defer self.allocator.free(robots_url);
switch (robot_entry) {
// If we have a found robots entry, we check it.
.present => |robots| {
const path = URL.getPathname(req.url);
if (!robots.isAllowed(path)) {
req.error_callback(req.ctx, error.RobotsBlocked);
return;
}
},
// Otherwise, we assume we won't find it again.
.absent => {},
}
return self.processRequest(req);
}
return self.fetchRobotsThenProcessRequest(robots_url, req);
} }
fn processRequest(self: *Client, req: Request) !void { fn processRequest(self: *Client, req: Request) !void {

View File

@@ -25,6 +25,7 @@ params: []const u8 = "",
// We keep 41 for null-termination since HTML parser expects in this format. // We keep 41 for null-termination since HTML parser expects in this format.
charset: [41]u8 = default_charset, charset: [41]u8 = default_charset,
charset_len: usize = default_charset_len, charset_len: usize = default_charset_len,
is_default_charset: bool = true,
/// String "UTF-8" continued by null characters. /// String "UTF-8" continued by null characters.
const default_charset = .{ 'U', 'T', 'F', '-', '8' } ++ .{0} ** 36; const default_charset = .{ 'U', 'T', 'F', '-', '8' } ++ .{0} ** 36;
@@ -130,6 +131,7 @@ pub fn parse(input: []u8) !Mime {
var charset: [41]u8 = default_charset; var charset: [41]u8 = default_charset;
var charset_len: usize = default_charset_len; var charset_len: usize = default_charset_len;
var has_explicit_charset = false;
var it = std.mem.splitScalar(u8, params, ';'); var it = std.mem.splitScalar(u8, params, ';');
while (it.next()) |attr| { while (it.next()) |attr| {
@@ -156,6 +158,7 @@ pub fn parse(input: []u8) !Mime {
// Null-terminate right after attribute value. // Null-terminate right after attribute value.
charset[attribute_value.len] = 0; charset[attribute_value.len] = 0;
charset_len = attribute_value.len; charset_len = attribute_value.len;
has_explicit_charset = true;
}, },
} }
} }
@@ -165,9 +168,137 @@ pub fn parse(input: []u8) !Mime {
.charset = charset, .charset = charset,
.charset_len = charset_len, .charset_len = charset_len,
.content_type = content_type, .content_type = content_type,
.is_default_charset = !has_explicit_charset,
}; };
} }
/// Prescan the first 1024 bytes of an HTML document for a charset declaration.
/// Looks for `<meta charset="X">` and `<meta http-equiv="Content-Type" content="...;charset=X">`.
/// Returns the charset value or null if none found.
/// See: https://www.w3.org/International/questions/qa-html-encoding-declarations
pub fn prescanCharset(html: []const u8) ?[]const u8 {
const limit = @min(html.len, 1024);
const data = html[0..limit];
// Scan for <meta tags
var pos: usize = 0;
while (pos < data.len) {
// Find next '<'
pos = std.mem.indexOfScalarPos(u8, data, pos, '<') orelse return null;
pos += 1;
if (pos >= data.len) return null;
// Check for "meta" (case-insensitive)
if (pos + 4 >= data.len) return null;
var tag_buf: [4]u8 = undefined;
_ = std.ascii.lowerString(&tag_buf, data[pos..][0..4]);
if (!std.mem.eql(u8, &tag_buf, "meta")) {
continue;
}
pos += 4;
// Must be followed by whitespace or end of tag
if (pos >= data.len) return null;
if (data[pos] != ' ' and data[pos] != '\t' and data[pos] != '\n' and
data[pos] != '\r' and data[pos] != '/')
{
continue;
}
// Scan attributes within this meta tag
const tag_end = std.mem.indexOfScalarPos(u8, data, pos, '>') orelse return null;
const attrs = data[pos..tag_end];
// Look for charset= attribute directly
if (findAttrValue(attrs, "charset")) |charset| {
if (charset.len > 0 and charset.len <= 40) return charset;
}
// Look for http-equiv="content-type" with content="...;charset=X"
if (findAttrValue(attrs, "http-equiv")) |he| {
if (std.ascii.eqlIgnoreCase(he, "content-type")) {
if (findAttrValue(attrs, "content")) |content| {
if (extractCharsetFromContentType(content)) |charset| {
return charset;
}
}
}
}
pos = tag_end + 1;
}
return null;
}
fn findAttrValue(attrs: []const u8, name: []const u8) ?[]const u8 {
var pos: usize = 0;
while (pos < attrs.len) {
// Skip whitespace
while (pos < attrs.len and (attrs[pos] == ' ' or attrs[pos] == '\t' or
attrs[pos] == '\n' or attrs[pos] == '\r'))
{
pos += 1;
}
if (pos >= attrs.len) return null;
// Read attribute name
const attr_start = pos;
while (pos < attrs.len and attrs[pos] != '=' and attrs[pos] != ' ' and
attrs[pos] != '\t' and attrs[pos] != '>' and attrs[pos] != '/')
{
pos += 1;
}
const attr_name = attrs[attr_start..pos];
// Skip whitespace around =
while (pos < attrs.len and (attrs[pos] == ' ' or attrs[pos] == '\t')) pos += 1;
if (pos >= attrs.len or attrs[pos] != '=') {
// No '=' found - skip this token. Advance at least one byte to avoid infinite loop.
if (pos == attr_start) pos += 1;
continue;
}
pos += 1; // skip '='
while (pos < attrs.len and (attrs[pos] == ' ' or attrs[pos] == '\t')) pos += 1;
if (pos >= attrs.len) return null;
// Read attribute value
const value = blk: {
if (attrs[pos] == '"' or attrs[pos] == '\'') {
const quote = attrs[pos];
pos += 1;
const val_start = pos;
while (pos < attrs.len and attrs[pos] != quote) pos += 1;
const val = attrs[val_start..pos];
if (pos < attrs.len) pos += 1; // skip closing quote
break :blk val;
} else {
const val_start = pos;
while (pos < attrs.len and attrs[pos] != ' ' and attrs[pos] != '\t' and
attrs[pos] != '>' and attrs[pos] != '/')
{
pos += 1;
}
break :blk attrs[val_start..pos];
}
};
if (std.ascii.eqlIgnoreCase(attr_name, name)) return value;
}
return null;
}
fn extractCharsetFromContentType(content: []const u8) ?[]const u8 {
var it = std.mem.splitScalar(u8, content, ';');
while (it.next()) |part| {
const trimmed = std.mem.trimLeft(u8, part, &.{ ' ', '\t' });
if (trimmed.len > 8 and std.ascii.eqlIgnoreCase(trimmed[0..8], "charset=")) {
const val = std.mem.trim(u8, trimmed[8..], &.{ ' ', '\t', '"', '\'' });
if (val.len > 0 and val.len <= 40) return val;
}
}
return null;
}
pub fn sniff(body: []const u8) ?Mime { pub fn sniff(body: []const u8) ?Mime {
// 0x0C is form feed // 0x0C is form feed
const content = std.mem.trimLeft(u8, body, &.{ ' ', '\t', '\n', '\r', 0x0C }); const content = std.mem.trimLeft(u8, body, &.{ ' ', '\t', '\n', '\r', 0x0C });
@@ -178,15 +309,30 @@ pub fn sniff(body: []const u8) ?Mime {
if (content[0] != '<') { if (content[0] != '<') {
if (std.mem.startsWith(u8, content, &.{ 0xEF, 0xBB, 0xBF })) { if (std.mem.startsWith(u8, content, &.{ 0xEF, 0xBB, 0xBF })) {
// UTF-8 BOM // UTF-8 BOM
return .{ .content_type = .{ .text_plain = {} } }; return .{
.content_type = .{ .text_plain = {} },
.charset = default_charset,
.charset_len = default_charset_len,
.is_default_charset = false,
};
} }
if (std.mem.startsWith(u8, content, &.{ 0xFE, 0xFF })) { if (std.mem.startsWith(u8, content, &.{ 0xFE, 0xFF })) {
// UTF-16 big-endian BOM // UTF-16 big-endian BOM
return .{ .content_type = .{ .text_plain = {} } }; return .{
.content_type = .{ .text_plain = {} },
.charset = .{ 'U', 'T', 'F', '-', '1', '6', 'B', 'E' } ++ .{0} ** 33,
.charset_len = 8,
.is_default_charset = false,
};
} }
if (std.mem.startsWith(u8, content, &.{ 0xFF, 0xFE })) { if (std.mem.startsWith(u8, content, &.{ 0xFF, 0xFE })) {
// UTF-16 little-endian BOM // UTF-16 little-endian BOM
return .{ .content_type = .{ .text_plain = {} } }; return .{
.content_type = .{ .text_plain = {} },
.charset = .{ 'U', 'T', 'F', '-', '1', '6', 'L', 'E' } ++ .{0} ** 33,
.charset_len = 8,
.is_default_charset = false,
};
} }
return null; return null;
} }
@@ -540,6 +686,24 @@ test "Mime: sniff" {
try expectHTML("<!-->"); try expectHTML("<!-->");
try expectHTML(" \n\t <!-->"); try expectHTML(" \n\t <!-->");
{
const mime = Mime.sniff(&.{ 0xEF, 0xBB, 0xBF }).?;
try testing.expectEqual(.text_plain, std.meta.activeTag(mime.content_type));
try testing.expectEqual("UTF-8", mime.charsetString());
}
{
const mime = Mime.sniff(&.{ 0xFE, 0xFF }).?;
try testing.expectEqual(.text_plain, std.meta.activeTag(mime.content_type));
try testing.expectEqual("UTF-16BE", mime.charsetString());
}
{
const mime = Mime.sniff(&.{ 0xFF, 0xFE }).?;
try testing.expectEqual(.text_plain, std.meta.activeTag(mime.content_type));
try testing.expectEqual("UTF-16LE", mime.charsetString());
}
} }
const Expectation = struct { const Expectation = struct {
@@ -576,3 +740,35 @@ fn expect(expected: Expectation, input: []const u8) !void {
try testing.expectEqual(m.charsetStringZ(), actual.charsetStringZ()); try testing.expectEqual(m.charsetStringZ(), actual.charsetStringZ());
} }
} }
test "Mime: prescanCharset" {
// <meta charset="X">
try testing.expectEqual("utf-8", Mime.prescanCharset("<html><head><meta charset=\"utf-8\">").?);
try testing.expectEqual("iso-8859-1", Mime.prescanCharset("<html><head><meta charset=\"iso-8859-1\">").?);
try testing.expectEqual("shift_jis", Mime.prescanCharset("<meta charset='shift_jis'>").?);
// Case-insensitive tag matching
try testing.expectEqual("utf-8", Mime.prescanCharset("<META charset=\"utf-8\">").?);
try testing.expectEqual("utf-8", Mime.prescanCharset("<Meta charset=\"utf-8\">").?);
// <meta http-equiv="Content-Type" content="text/html; charset=X">
try testing.expectEqual(
"iso-8859-1",
Mime.prescanCharset("<meta http-equiv=\"Content-Type\" content=\"text/html; charset=iso-8859-1\">").?,
);
// No charset found
try testing.expectEqual(null, Mime.prescanCharset("<html><head><title>Test</title>"));
try testing.expectEqual(null, Mime.prescanCharset(""));
try testing.expectEqual(null, Mime.prescanCharset("no html here"));
// Self-closing meta without charset must not loop forever
try testing.expectEqual(null, Mime.prescanCharset("<meta foo=\"bar\"/>"));
// Charset after 1024 bytes should not be found
var long_html: [1100]u8 = undefined;
@memset(&long_html, ' ');
const suffix = "<meta charset=\"windows-1252\">";
@memcpy(long_html[1050 .. 1050 + suffix.len], suffix);
try testing.expectEqual(null, Mime.prescanCharset(&long_html));
}

View File

@@ -710,11 +710,14 @@ pub fn scriptsCompletedLoading(self: *Page) void {
} }
pub fn iframeCompletedLoading(self: *Page, iframe: *IFrame) void { pub fn iframeCompletedLoading(self: *Page, iframe: *IFrame) void {
blk: { var ls: JS.Local.Scope = undefined;
var ls: JS.Local.Scope = undefined; self.js.localScope(&ls);
self.js.localScope(&ls); defer ls.deinit();
defer ls.deinit();
const entered = self.js.enter(&ls.handle_scope);
defer entered.exit();
blk: {
const event = Event.initTrusted(comptime .wrap("load"), .{}, self) catch |err| { const event = Event.initTrusted(comptime .wrap("load"), .{}, self) catch |err| {
log.err(.page, "iframe event init", .{ .err = err, .url = iframe._src }); log.err(.page, "iframe event init", .{ .err = err, .url = iframe._src });
break :blk; break :blk;
@@ -723,6 +726,7 @@ pub fn iframeCompletedLoading(self: *Page, iframe: *IFrame) void {
log.warn(.js, "iframe onload", .{ .err = err, .url = iframe._src }); log.warn(.js, "iframe onload", .{ .err = err, .url = iframe._src });
}; };
} }
self.pendingLoadCompleted(); self.pendingLoadCompleted();
} }
@@ -849,13 +853,25 @@ fn pageDataCallback(transfer: *HttpClient.Transfer, data: []const u8) !void {
if (self._parse_state == .pre) { if (self._parse_state == .pre) {
// we lazily do this, because we might need the first chunk of data // we lazily do this, because we might need the first chunk of data
// to sniff the content type // to sniff the content type
const mime: Mime = blk: { var mime: Mime = blk: {
if (transfer.response_header.?.contentType()) |ct| { if (transfer.response_header.?.contentType()) |ct| {
break :blk try Mime.parse(ct); break :blk try Mime.parse(ct);
} }
break :blk Mime.sniff(data); break :blk Mime.sniff(data);
} orelse .unknown; } orelse .unknown;
// If the HTTP Content-Type header didn't specify a charset and this is HTML,
// prescan the first 1024 bytes for a <meta charset> declaration.
if (mime.content_type == .text_html and mime.is_default_charset) {
if (Mime.prescanCharset(data)) |charset| {
if (charset.len <= 40) {
@memcpy(mime.charset[0..charset.len], charset);
mime.charset[charset.len] = 0;
mime.charset_len = charset.len;
}
}
}
if (comptime IS_DEBUG) { if (comptime IS_DEBUG) {
log.debug(.page, "navigate first chunk", .{ log.debug(.page, "navigate first chunk", .{
.content_type = mime.content_type, .content_type = mime.content_type,

View File

@@ -63,9 +63,6 @@ shutdown: bool = false,
client: *HttpClient, client: *HttpClient,
allocator: Allocator, allocator: Allocator,
buffer_pool: BufferPool,
script_pool: std.heap.MemoryPool(Script),
// We can download multiple sync modules in parallel, but we want to process // We can download multiple sync modules in parallel, but we want to process
// them in order. We can't use an std.DoublyLinkedList, like the other script types, // them in order. We can't use an std.DoublyLinkedList, like the other script types,
@@ -101,18 +98,14 @@ pub fn init(allocator: Allocator, http_client: *HttpClient, page: *Page) ScriptM
.imported_modules = .empty, .imported_modules = .empty,
.client = http_client, .client = http_client,
.static_scripts_done = false, .static_scripts_done = false,
.buffer_pool = BufferPool.init(allocator, 5),
.page_notified_of_completion = false, .page_notified_of_completion = false,
.script_pool = std.heap.MemoryPool(Script).init(allocator),
}; };
} }
pub fn deinit(self: *ScriptManager) void { pub fn deinit(self: *ScriptManager) void {
// necessary to free any buffers scripts may be referencing // necessary to free any arenas scripts may be referencing
self.reset(); self.reset();
self.buffer_pool.deinit();
self.script_pool.deinit();
self.imported_modules.deinit(self.allocator); self.imported_modules.deinit(self.allocator);
// we don't deinit self.importmap b/c we use the page's arena for its // we don't deinit self.importmap b/c we use the page's arena for its
// allocations. // allocations.
@@ -121,7 +114,10 @@ pub fn deinit(self: *ScriptManager) void {
pub fn reset(self: *ScriptManager) void { pub fn reset(self: *ScriptManager) void {
var it = self.imported_modules.valueIterator(); var it = self.imported_modules.valueIterator();
while (it.next()) |value_ptr| { while (it.next()) |value_ptr| {
self.buffer_pool.release(value_ptr.buffer); switch (value_ptr.state) {
.done => |script| script.deinit(),
else => {},
}
} }
self.imported_modules.clearRetainingCapacity(); self.imported_modules.clearRetainingCapacity();
@@ -138,13 +134,13 @@ pub fn reset(self: *ScriptManager) void {
fn clearList(list: *std.DoublyLinkedList) void { fn clearList(list: *std.DoublyLinkedList) void {
while (list.popFirst()) |n| { while (list.popFirst()) |n| {
const script: *Script = @fieldParentPtr("node", n); const script: *Script = @fieldParentPtr("node", n);
script.deinit(true); script.deinit();
} }
} }
pub fn getHeaders(self: *ScriptManager, url: [:0]const u8) !net_http.Headers { fn getHeaders(self: *ScriptManager, arena: Allocator, url: [:0]const u8) !net_http.Headers {
var headers = try self.client.newHeaders(); var headers = try self.client.newHeaders();
try self.page.headersForRequest(self.page.arena, url, &headers); try self.page.headersForRequest(arena, url, &headers);
return headers; return headers;
} }
@@ -191,19 +187,26 @@ pub fn addFromElement(self: *ScriptManager, comptime from_parser: bool, script_e
return; return;
}; };
var handover = false;
const page = self.page; const page = self.page;
const arena = try page.getArena(.{ .debug = "addFromElement" });
errdefer if (!handover) {
page.releaseArena(arena);
};
var source: Script.Source = undefined; var source: Script.Source = undefined;
var remote_url: ?[:0]const u8 = null; var remote_url: ?[:0]const u8 = null;
const base_url = page.base(); const base_url = page.base();
if (element.getAttributeSafe(comptime .wrap("src"))) |src| { if (element.getAttributeSafe(comptime .wrap("src"))) |src| {
if (try parseDataURI(page.arena, src)) |data_uri| { if (try parseDataURI(arena, src)) |data_uri| {
source = .{ .@"inline" = data_uri }; source = .{ .@"inline" = data_uri };
} else { } else {
remote_url = try URL.resolve(page.arena, base_url, src, .{}); remote_url = try URL.resolve(arena, base_url, src, .{});
source = .{ .remote = .{} }; source = .{ .remote = .{} };
} }
} else { } else {
var buf = std.Io.Writer.Allocating.init(page.arena); var buf = std.Io.Writer.Allocating.init(arena);
try element.asNode().getChildTextContent(&buf.writer); try element.asNode().getChildTextContent(&buf.writer);
try buf.writer.writeByte(0); try buf.writer.writeByte(0);
const data = buf.written(); const data = buf.written();
@@ -218,15 +221,13 @@ pub fn addFromElement(self: *ScriptManager, comptime from_parser: bool, script_e
// Only set _executed (already-started) when we actually have content to execute // Only set _executed (already-started) when we actually have content to execute
script_element._executed = true; script_element._executed = true;
const script = try self.script_pool.create();
errdefer self.script_pool.destroy(script);
const is_inline = source == .@"inline"; const is_inline = source == .@"inline";
const script = try arena.create(Script);
script.* = .{ script.* = .{
.kind = kind, .kind = kind,
.node = .{}, .node = .{},
.arena = arena,
.manager = self, .manager = self,
.source = source, .source = source,
.script_element = script_element, .script_element = script_element,
@@ -270,7 +271,7 @@ pub fn addFromElement(self: *ScriptManager, comptime from_parser: bool, script_e
if (is_blocking == false) { if (is_blocking == false) {
self.scriptList(script).remove(&script.node); self.scriptList(script).remove(&script.node);
} }
script.deinit(true); script.deinit();
} }
try self.client.request(.{ try self.client.request(.{
@@ -278,7 +279,7 @@ pub fn addFromElement(self: *ScriptManager, comptime from_parser: bool, script_e
.ctx = script, .ctx = script,
.method = .GET, .method = .GET,
.frame_id = page._frame_id, .frame_id = page._frame_id,
.headers = try self.getHeaders(url), .headers = try self.getHeaders(arena, url),
.blocking = is_blocking, .blocking = is_blocking,
.cookie_jar = &page._session.cookie_jar, .cookie_jar = &page._session.cookie_jar,
.resource_type = .script, .resource_type = .script,
@@ -289,6 +290,7 @@ pub fn addFromElement(self: *ScriptManager, comptime from_parser: bool, script_e
.done_callback = Script.doneCallback, .done_callback = Script.doneCallback,
.error_callback = Script.errorCallback, .error_callback = Script.errorCallback,
}); });
handover = true;
if (comptime IS_DEBUG) { if (comptime IS_DEBUG) {
var ls: js.Local.Scope = undefined; var ls: js.Local.Scope = undefined;
@@ -318,7 +320,7 @@ pub fn addFromElement(self: *ScriptManager, comptime from_parser: bool, script_e
} }
if (script.status == 0) { if (script.status == 0) {
// an error (that we already logged) // an error (that we already logged)
script.deinit(true); script.deinit();
return; return;
} }
@@ -327,7 +329,7 @@ pub fn addFromElement(self: *ScriptManager, comptime from_parser: bool, script_e
self.is_evaluating = true; self.is_evaluating = true;
defer { defer {
self.is_evaluating = was_evaluating; self.is_evaluating = was_evaluating;
script.deinit(true); script.deinit();
} }
return script.eval(page); return script.eval(page);
} }
@@ -359,11 +361,14 @@ pub fn preloadImport(self: *ScriptManager, url: [:0]const u8, referrer: []const
} }
errdefer _ = self.imported_modules.remove(url); errdefer _ = self.imported_modules.remove(url);
const script = try self.script_pool.create(); const page = self.page;
errdefer self.script_pool.destroy(script); const arena = try page.getArena(.{ .debug = "preloadImport" });
errdefer page.releaseArena(arena);
const script = try arena.create(Script);
script.* = .{ script.* = .{
.kind = .module, .kind = .module,
.arena = arena,
.url = url, .url = url,
.node = .{}, .node = .{},
.manager = self, .manager = self,
@@ -373,11 +378,7 @@ pub fn preloadImport(self: *ScriptManager, url: [:0]const u8, referrer: []const
.mode = .import, .mode = .import,
}; };
gop.value_ptr.* = ImportedModule{ gop.value_ptr.* = ImportedModule{};
.manager = self,
};
const page = self.page;
if (comptime IS_DEBUG) { if (comptime IS_DEBUG) {
var ls: js.Local.Scope = undefined; var ls: js.Local.Scope = undefined;
@@ -392,12 +393,18 @@ pub fn preloadImport(self: *ScriptManager, url: [:0]const u8, referrer: []const
}); });
} }
try self.client.request(.{ // This seems wrong since we're not dealing with an async import (unlike
// getAsyncModule below), but all we're trying to do here is pre-load the
// script for execution at some point in the future (when waitForImport is
// called).
self.async_scripts.append(&script.node);
self.client.request(.{
.url = url, .url = url,
.ctx = script, .ctx = script,
.method = .GET, .method = .GET,
.frame_id = page._frame_id, .frame_id = page._frame_id,
.headers = try self.getHeaders(url), .headers = try self.getHeaders(arena, url),
.cookie_jar = &page._session.cookie_jar, .cookie_jar = &page._session.cookie_jar,
.resource_type = .script, .resource_type = .script,
.notification = page._session.notification, .notification = page._session.notification,
@@ -406,13 +413,10 @@ pub fn preloadImport(self: *ScriptManager, url: [:0]const u8, referrer: []const
.data_callback = Script.dataCallback, .data_callback = Script.dataCallback,
.done_callback = Script.doneCallback, .done_callback = Script.doneCallback,
.error_callback = Script.errorCallback, .error_callback = Script.errorCallback,
}); }) catch |err| {
self.async_scripts.remove(&script.node);
// This seems wrong since we're not dealing with an async import (unlike return err;
// getAsyncModule below), but all we're trying to do here is pre-load the };
// script for execution at some point in the future (when waitForImport is
// called).
self.async_scripts.append(&script.node);
} }
pub fn waitForImport(self: *ScriptManager, url: [:0]const u8) !ModuleSource { pub fn waitForImport(self: *ScriptManager, url: [:0]const u8) !ModuleSource {
@@ -433,12 +437,12 @@ pub fn waitForImport(self: *ScriptManager, url: [:0]const u8) !ModuleSource {
_ = try client.tick(200); _ = try client.tick(200);
continue; continue;
}, },
.done => { .done => |script| {
var shared = false; var shared = false;
const buffer = entry.value_ptr.buffer; const buffer = entry.value_ptr.buffer;
const waiters = entry.value_ptr.waiters; const waiters = entry.value_ptr.waiters;
if (waiters == 0) { if (waiters == 1) {
self.imported_modules.removeByPtr(entry.key_ptr); self.imported_modules.removeByPtr(entry.key_ptr);
} else { } else {
shared = true; shared = true;
@@ -447,7 +451,7 @@ pub fn waitForImport(self: *ScriptManager, url: [:0]const u8) !ModuleSource {
return .{ return .{
.buffer = buffer, .buffer = buffer,
.shared = shared, .shared = shared,
.buffer_pool = &self.buffer_pool, .script = script,
}; };
}, },
.err => return error.Failed, .err => return error.Failed,
@@ -456,11 +460,14 @@ pub fn waitForImport(self: *ScriptManager, url: [:0]const u8) !ModuleSource {
} }
pub fn getAsyncImport(self: *ScriptManager, url: [:0]const u8, cb: ImportAsync.Callback, cb_data: *anyopaque, referrer: []const u8) !void { pub fn getAsyncImport(self: *ScriptManager, url: [:0]const u8, cb: ImportAsync.Callback, cb_data: *anyopaque, referrer: []const u8) !void {
const script = try self.script_pool.create(); const page = self.page;
errdefer self.script_pool.destroy(script); const arena = try page.getArena(.{ .debug = "getAsyncImport" });
errdefer page.releaseArena(arena);
const script = try arena.create(Script);
script.* = .{ script.* = .{
.kind = .module, .kind = .module,
.arena = arena,
.url = url, .url = url,
.node = .{}, .node = .{},
.manager = self, .manager = self,
@@ -473,7 +480,6 @@ pub fn getAsyncImport(self: *ScriptManager, url: [:0]const u8, cb: ImportAsync.C
} }, } },
}; };
const page = self.page;
if (comptime IS_DEBUG) { if (comptime IS_DEBUG) {
var ls: js.Local.Scope = undefined; var ls: js.Local.Scope = undefined;
page.js.localScope(&ls); page.js.localScope(&ls);
@@ -496,11 +502,12 @@ pub fn getAsyncImport(self: *ScriptManager, url: [:0]const u8, cb: ImportAsync.C
self.is_evaluating = true; self.is_evaluating = true;
defer self.is_evaluating = was_evaluating; defer self.is_evaluating = was_evaluating;
try self.client.request(.{ self.async_scripts.append(&script.node);
self.client.request(.{
.url = url, .url = url,
.method = .GET, .method = .GET,
.frame_id = page._frame_id, .frame_id = page._frame_id,
.headers = try self.getHeaders(url), .headers = try self.getHeaders(arena, url),
.ctx = script, .ctx = script,
.resource_type = .script, .resource_type = .script,
.cookie_jar = &page._session.cookie_jar, .cookie_jar = &page._session.cookie_jar,
@@ -510,9 +517,10 @@ pub fn getAsyncImport(self: *ScriptManager, url: [:0]const u8, cb: ImportAsync.C
.data_callback = Script.dataCallback, .data_callback = Script.dataCallback,
.done_callback = Script.doneCallback, .done_callback = Script.doneCallback,
.error_callback = Script.errorCallback, .error_callback = Script.errorCallback,
}); }) catch |err| {
self.async_scripts.remove(&script.node);
self.async_scripts.append(&script.node); return err;
};
} }
// Called from the Page to let us know it's done parsing the HTML. Necessary that // Called from the Page to let us know it's done parsing the HTML. Necessary that
@@ -537,18 +545,18 @@ fn evaluate(self: *ScriptManager) void {
var script: *Script = @fieldParentPtr("node", n); var script: *Script = @fieldParentPtr("node", n);
switch (script.mode) { switch (script.mode) {
.async => { .async => {
defer script.deinit(true); defer script.deinit();
script.eval(page); script.eval(page);
}, },
.import_async => |ia| { .import_async => |ia| {
defer script.deinit(false);
if (script.status < 200 or script.status > 299) { if (script.status < 200 or script.status > 299) {
script.deinit();
ia.callback(ia.data, error.FailedToLoad); ia.callback(ia.data, error.FailedToLoad);
} else { } else {
ia.callback(ia.data, .{ ia.callback(ia.data, .{
.shared = false, .shared = false,
.script = script,
.buffer = script.source.remote, .buffer = script.source.remote,
.buffer_pool = &self.buffer_pool,
}); });
} }
}, },
@@ -574,7 +582,7 @@ fn evaluate(self: *ScriptManager) void {
} }
defer { defer {
_ = self.defer_scripts.popFirst(); _ = self.defer_scripts.popFirst();
script.deinit(true); script.deinit();
} }
script.eval(page); script.eval(page);
} }
@@ -625,11 +633,12 @@ fn parseImportmap(self: *ScriptManager, script: *const Script) !void {
} }
pub const Script = struct { pub const Script = struct {
complete: bool,
kind: Kind, kind: Kind,
complete: bool,
status: u16 = 0, status: u16 = 0,
source: Source, source: Source,
url: []const u8, url: []const u8,
arena: Allocator,
mode: ExecutionMode, mode: ExecutionMode,
node: std.DoublyLinkedList.Node, node: std.DoublyLinkedList.Node,
script_element: ?*Element.Html.Script, script_element: ?*Element.Html.Script,
@@ -680,11 +689,8 @@ pub const Script = struct {
import_async: ImportAsync, import_async: ImportAsync,
}; };
fn deinit(self: *Script, comptime release_buffer: bool) void { fn deinit(self: *Script) void {
if ((comptime release_buffer) and self.source == .remote) { self.manager.page.releaseArena(self.arena);
self.manager.buffer_pool.release(self.source.remote);
}
self.manager.script_pool.destroy(self);
} }
fn startCallback(transfer: *HttpClient.Transfer) !void { fn startCallback(transfer: *HttpClient.Transfer) !void {
@@ -750,9 +756,9 @@ pub const Script = struct {
} }
lp.assert(self.source.remote.capacity == 0, "ScriptManager.Header buffer", .{ .capacity = self.source.remote.capacity }); lp.assert(self.source.remote.capacity == 0, "ScriptManager.Header buffer", .{ .capacity = self.source.remote.capacity });
var buffer = self.manager.buffer_pool.get(); var buffer: std.ArrayList(u8) = .empty;
if (transfer.getContentLength()) |cl| { if (transfer.getContentLength()) |cl| {
try buffer.ensureTotalCapacity(self.manager.allocator, cl); try buffer.ensureTotalCapacity(self.arena, cl);
} }
self.source = .{ .remote = buffer }; self.source = .{ .remote = buffer };
return true; return true;
@@ -766,7 +772,7 @@ pub const Script = struct {
}; };
} }
fn _dataCallback(self: *Script, _: *HttpClient.Transfer, data: []const u8) !void { fn _dataCallback(self: *Script, _: *HttpClient.Transfer, data: []const u8) !void {
try self.source.remote.appendSlice(self.manager.allocator, data); try self.source.remote.appendSlice(self.arena, data);
} }
fn doneCallback(ctx: *anyopaque) !void { fn doneCallback(ctx: *anyopaque) !void {
@@ -783,9 +789,8 @@ pub const Script = struct {
} else if (self.mode == .import) { } else if (self.mode == .import) {
manager.async_scripts.remove(&self.node); manager.async_scripts.remove(&self.node);
const entry = manager.imported_modules.getPtr(self.url).?; const entry = manager.imported_modules.getPtr(self.url).?;
entry.state = .done; entry.state = .{ .done = self };
entry.buffer = self.source.remote; entry.buffer = self.source.remote;
self.deinit(false);
} }
manager.evaluate(); manager.evaluate();
} }
@@ -811,7 +816,7 @@ pub const Script = struct {
const manager = self.manager; const manager = self.manager;
manager.scriptList(self).remove(&self.node); manager.scriptList(self).remove(&self.node);
if (manager.shutdown) { if (manager.shutdown) {
self.deinit(true); self.deinit();
return; return;
} }
@@ -823,7 +828,7 @@ pub const Script = struct {
}, },
else => {}, else => {},
} }
self.deinit(true); self.deinit();
manager.evaluate(); manager.evaluate();
} }
@@ -951,76 +956,6 @@ pub const Script = struct {
} }
}; };
const BufferPool = struct {
count: usize,
available: List = .{},
allocator: Allocator,
max_concurrent_transfers: u8,
mem_pool: std.heap.MemoryPool(Container),
const List = std.SinglyLinkedList;
const Container = struct {
node: List.Node,
buf: std.ArrayList(u8),
};
fn init(allocator: Allocator, max_concurrent_transfers: u8) BufferPool {
return .{
.available = .{},
.count = 0,
.allocator = allocator,
.max_concurrent_transfers = max_concurrent_transfers,
.mem_pool = std.heap.MemoryPool(Container).init(allocator),
};
}
fn deinit(self: *BufferPool) void {
const allocator = self.allocator;
var node = self.available.first;
while (node) |n| {
const container: *Container = @fieldParentPtr("node", n);
container.buf.deinit(allocator);
node = n.next;
}
self.mem_pool.deinit();
}
fn get(self: *BufferPool) std.ArrayList(u8) {
const node = self.available.popFirst() orelse {
// return a new buffer
return .{};
};
self.count -= 1;
const container: *Container = @fieldParentPtr("node", node);
defer self.mem_pool.destroy(container);
return container.buf;
}
fn release(self: *BufferPool, buffer: ArrayList(u8)) void {
// create mutable copy
var b = buffer;
if (self.count == self.max_concurrent_transfers) {
b.deinit(self.allocator);
return;
}
const container = self.mem_pool.create() catch |err| {
b.deinit(self.allocator);
log.err(.http, "SM BufferPool release", .{ .err = err });
return;
};
b.clearRetainingCapacity();
container.* = .{ .buf = b, .node = .{} };
self.count += 1;
self.available.prepend(&container.node);
}
};
const ImportAsync = struct { const ImportAsync = struct {
data: *anyopaque, data: *anyopaque,
callback: ImportAsync.Callback, callback: ImportAsync.Callback,
@@ -1030,12 +965,12 @@ const ImportAsync = struct {
pub const ModuleSource = struct { pub const ModuleSource = struct {
shared: bool, shared: bool,
buffer_pool: *BufferPool, script: *Script,
buffer: std.ArrayList(u8), buffer: std.ArrayList(u8),
pub fn deinit(self: *ModuleSource) void { pub fn deinit(self: *ModuleSource) void {
if (self.shared == false) { if (self.shared == false) {
self.buffer_pool.release(self.buffer); self.script.deinit();
} }
} }
@@ -1045,15 +980,14 @@ pub const ModuleSource = struct {
}; };
const ImportedModule = struct { const ImportedModule = struct {
manager: *ScriptManager, waiters: u16 = 1,
state: State = .loading, state: State = .loading,
buffer: std.ArrayList(u8) = .{}, buffer: std.ArrayList(u8) = .{},
waiters: u16 = 1,
const State = enum { const State = union(enum) {
err, err,
done,
loading, loading,
done: *Script,
}; };
}; };

View File

@@ -401,7 +401,7 @@ fn _wait(self: *Session, page: *Page, wait_ms: u32) !WaitResult {
// scheduler.run could trigger new http transfers, so do not // scheduler.run could trigger new http transfers, so do not
// store http_client.active BEFORE this call and then use // store http_client.active BEFORE this call and then use
// it AFTER. // it AFTER.
const ms_to_next_task = try browser.runMacrotasks(); try browser.runMacrotasks();
// Each call to this runs scheduled load events. // Each call to this runs scheduled load events.
try page.dispatchLoad(); try page.dispatchLoad();
@@ -423,16 +423,16 @@ fn _wait(self: *Session, page: *Page, wait_ms: u32) !WaitResult {
std.debug.assert(http_client.intercepted == 0); std.debug.assert(http_client.intercepted == 0);
} }
var ms: u64 = ms_to_next_task orelse blk: { var ms = blk: {
if (wait_ms - ms_remaining < 100) { // if (wait_ms - ms_remaining < 100) {
if (comptime builtin.is_test) { // if (comptime builtin.is_test) {
return .done; // return .done;
} // }
// Look, we want to exit ASAP, but we don't want // // Look, we want to exit ASAP, but we don't want
// to exit so fast that we've run none of the // // to exit so fast that we've run none of the
// background jobs. // // background jobs.
break :blk 50; // break :blk 50;
} // }
if (browser.hasBackgroundTasks()) { if (browser.hasBackgroundTasks()) {
// _we_ have nothing to run, but v8 is working on // _we_ have nothing to run, but v8 is working on
@@ -441,9 +441,7 @@ fn _wait(self: *Session, page: *Page, wait_ms: u32) !WaitResult {
break :blk 20; break :blk 20;
} }
// No http transfers, no cdp extra socket, no break :blk browser.msToNextMacrotask() orelse return .done;
// scheduled tasks, we're done.
return .done;
}; };
if (ms > ms_remaining) { if (ms > ms_remaining) {
@@ -470,9 +468,9 @@ fn _wait(self: *Session, page: *Page, wait_ms: u32) !WaitResult {
// We're here because we either have active HTTP // We're here because we either have active HTTP
// connections, or exit_when_done == false (aka, there's // connections, or exit_when_done == false (aka, there's
// an cdp_socket registered with the http client). // an cdp_socket registered with the http client).
// We should continue to run lowPriority tasks, so we // We should continue to run tasks, so we minimize how long
// minimize how long we'll poll for network I/O. // we'll poll for network I/O.
var ms_to_wait = @min(200, ms_to_next_task orelse 200); var ms_to_wait = @min(200, browser.msToNextMacrotask() orelse 200);
if (ms_to_wait > 10 and browser.hasBackgroundTasks()) { if (ms_to_wait > 10 and browser.hasBackgroundTasks()) {
// if we have background tasks, we don't want to wait too // if we have background tasks, we don't want to wait too
// long for a message from the client. We want to go back // long for a message from the client. We want to go back

View File

@@ -167,12 +167,11 @@ pub fn setOrigin(self: *Context, key: ?[]const u8) !void {
const env = self.env; const env = self.env;
const isolate = env.isolate; const isolate = env.isolate;
lp.assert(self.origin.rc == 1, "Ref opaque origin", .{ .rc = self.origin.rc });
const origin = try self.session.getOrCreateOrigin(key); const origin = try self.session.getOrCreateOrigin(key);
errdefer self.session.releaseOrigin(origin); errdefer self.session.releaseOrigin(origin);
try origin.takeover(self.origin);
try self.origin.transferTo(origin);
lp.assert(self.origin.rc == 1, "Ref opaque origin", .{ .rc = self.origin.rc });
self.origin.deinit(env.app);
self.origin = origin; self.origin = origin;
@@ -255,6 +254,10 @@ pub fn toLocal(self: *Context, global: anytype) js.Local.ToLocalReturnType(@Type
return l.toLocal(global); return l.toLocal(global);
} }
pub fn getIncumbent(self: *Context) *Page {
return fromC(v8.v8__Isolate__GetIncumbentContext(self.env.isolate.handle).?).page;
}
pub fn stringToPersistedFunction( pub fn stringToPersistedFunction(
self: *Context, self: *Context,
function_body: []const u8, function_body: []const u8,
@@ -306,15 +309,15 @@ pub fn module(self: *Context, comptime want_result: bool, local: *const js.Local
} }
const owned_url = try arena.dupeZ(u8, url); const owned_url = try arena.dupeZ(u8, url);
if (cacheable and !gop.found_existing) {
gop.key_ptr.* = owned_url;
}
const m = try compileModule(local, src, owned_url); const m = try compileModule(local, src, owned_url);
if (cacheable) { if (cacheable) {
// compileModule is synchronous - nothing can modify the cache during compilation // compileModule is synchronous - nothing can modify the cache during compilation
lp.assert(gop.value_ptr.module == null, "Context.module has module", .{}); lp.assert(gop.value_ptr.module == null, "Context.module has module", .{});
gop.value_ptr.module = try m.persist(); gop.value_ptr.module = try m.persist();
if (!gop.found_existing) {
gop.key_ptr.* = owned_url;
}
} }
break :blk .{ m, owned_url }; break :blk .{ m, owned_url };

View File

@@ -382,8 +382,7 @@ pub fn runMicrotasks(self: *Env) void {
} }
} }
pub fn runMacrotasks(self: *Env) !?u64 { pub fn runMacrotasks(self: *Env) !void {
var ms_to_next_task: ?u64 = null;
for (self.contexts[0..self.context_count]) |ctx| { for (self.contexts[0..self.context_count]) |ctx| {
if (comptime builtin.is_test == false) { if (comptime builtin.is_test == false) {
// I hate this comptime check as much as you do. But we have tests // I hate this comptime check as much as you do. But we have tests
@@ -398,13 +397,17 @@ pub fn runMacrotasks(self: *Env) !?u64 {
var hs: js.HandleScope = undefined; var hs: js.HandleScope = undefined;
const entered = ctx.enter(&hs); const entered = ctx.enter(&hs);
defer entered.exit(); defer entered.exit();
try ctx.scheduler.run();
const ms = (try ctx.scheduler.run()) orelse continue;
if (ms_to_next_task == null or ms < ms_to_next_task.?) {
ms_to_next_task = ms;
}
} }
return ms_to_next_task; }
pub fn msToNextMacrotask(self: *Env) ?u64 {
var next_task: u64 = std.math.maxInt(u64);
for (self.contexts[0..self.context_count]) |ctx| {
const candidate = ctx.scheduler.msToNextHigh() orelse continue;
next_task = @min(candidate, next_task);
}
return if (next_task == std.math.maxInt(u64)) null else next_task;
} }
pub fn pumpMessageLoop(self: *const Env) void { pub fn pumpMessageLoop(self: *const Env) void {

View File

@@ -68,6 +68,8 @@ temps: std.AutoHashMapUnmanaged(usize, v8.Global) = .empty,
// if v8 hasn't called the finalizer directly itself. // if v8 hasn't called the finalizer directly itself.
finalizer_callbacks: std.AutoHashMapUnmanaged(usize, *FinalizerCallback) = .empty, finalizer_callbacks: std.AutoHashMapUnmanaged(usize, *FinalizerCallback) = .empty,
taken_over: std.ArrayList(*Origin),
pub fn init(app: *App, isolate: js.Isolate, key: []const u8) !*Origin { pub fn init(app: *App, isolate: js.Isolate, key: []const u8) !*Origin {
const arena = try app.arena_pool.acquire(); const arena = try app.arena_pool.acquire();
errdefer app.arena_pool.release(arena); errdefer app.arena_pool.release(arena);
@@ -86,14 +88,19 @@ pub fn init(app: *App, isolate: js.Isolate, key: []const u8) !*Origin {
.rc = 1, .rc = 1,
.arena = arena, .arena = arena,
.key = owned_key, .key = owned_key,
.globals = .empty,
.temps = .empty, .temps = .empty,
.globals = .empty,
.taken_over = .empty,
.security_token = token_global, .security_token = token_global,
}; };
return self; return self;
} }
pub fn deinit(self: *Origin, app: *App) void { pub fn deinit(self: *Origin, app: *App) void {
for (self.taken_over.items) |o| {
o.deinit(app);
}
// Call finalizers before releasing anything // Call finalizers before releasing anything
{ {
var it = self.finalizer_callbacks.valueIterator(); var it = self.finalizer_callbacks.valueIterator();
@@ -196,42 +203,44 @@ pub fn createFinalizerCallback(
return fc; return fc;
} }
pub fn transferTo(self: *Origin, dest: *Origin) !void { pub fn takeover(self: *Origin, original: *Origin) !void {
const arena = dest.arena; const arena = self.arena;
try dest.globals.ensureUnusedCapacity(arena, self.globals.items.len); try self.globals.ensureUnusedCapacity(arena, self.globals.items.len);
for (self.globals.items) |obj| { for (original.globals.items) |obj| {
dest.globals.appendAssumeCapacity(obj); self.globals.appendAssumeCapacity(obj);
} }
self.globals.clearRetainingCapacity(); original.globals.clearRetainingCapacity();
{ {
try dest.temps.ensureUnusedCapacity(arena, self.temps.count()); try self.temps.ensureUnusedCapacity(arena, original.temps.count());
var it = self.temps.iterator(); var it = original.temps.iterator();
while (it.next()) |kv| { while (it.next()) |kv| {
try dest.temps.put(arena, kv.key_ptr.*, kv.value_ptr.*); try self.temps.put(arena, kv.key_ptr.*, kv.value_ptr.*);
} }
self.temps.clearRetainingCapacity(); original.temps.clearRetainingCapacity();
} }
{ {
try dest.finalizer_callbacks.ensureUnusedCapacity(arena, self.finalizer_callbacks.count()); try self.finalizer_callbacks.ensureUnusedCapacity(arena, original.finalizer_callbacks.count());
var it = self.finalizer_callbacks.iterator(); var it = original.finalizer_callbacks.iterator();
while (it.next()) |kv| { while (it.next()) |kv| {
kv.value_ptr.*.origin = dest; kv.value_ptr.*.origin = self;
try dest.finalizer_callbacks.put(arena, kv.key_ptr.*, kv.value_ptr.*); try self.finalizer_callbacks.put(arena, kv.key_ptr.*, kv.value_ptr.*);
} }
self.finalizer_callbacks.clearRetainingCapacity(); original.finalizer_callbacks.clearRetainingCapacity();
} }
{ {
try dest.identity_map.ensureUnusedCapacity(arena, self.identity_map.count()); try self.identity_map.ensureUnusedCapacity(arena, original.identity_map.count());
var it = self.identity_map.iterator(); var it = original.identity_map.iterator();
while (it.next()) |kv| { while (it.next()) |kv| {
try dest.identity_map.put(arena, kv.key_ptr.*, kv.value_ptr.*); try self.identity_map.put(arena, kv.key_ptr.*, kv.value_ptr.*);
} }
self.identity_map.clearRetainingCapacity(); original.identity_map.clearRetainingCapacity();
} }
try self.taken_over.append(self.arena, original);
} }
// A type that has a finalizer can have its finalizer called one of two ways. // A type that has a finalizer can have its finalizer called one of two ways.

View File

@@ -74,9 +74,10 @@ pub fn add(self: *Scheduler, ctx: *anyopaque, cb: Callback, run_in_ms: u32, opts
}); });
} }
pub fn run(self: *Scheduler) !?u64 { pub fn run(self: *Scheduler) !void {
_ = try self.runQueue(&self.low_priority); const now = milliTimestamp(.monotonic);
return self.runQueue(&self.high_priority); try self.runQueue(&self.low_priority, now);
try self.runQueue(&self.high_priority, now);
} }
pub fn hasReadyTasks(self: *Scheduler) bool { pub fn hasReadyTasks(self: *Scheduler) bool {
@@ -84,16 +85,23 @@ pub fn hasReadyTasks(self: *Scheduler) bool {
return queueuHasReadyTask(&self.low_priority, now) or queueuHasReadyTask(&self.high_priority, now); return queueuHasReadyTask(&self.low_priority, now) or queueuHasReadyTask(&self.high_priority, now);
} }
fn runQueue(self: *Scheduler, queue: *Queue) !?u64 { pub fn msToNextHigh(self: *Scheduler) ?u64 {
if (queue.count() == 0) { const task = self.high_priority.peek() orelse return null;
return null;
}
const now = milliTimestamp(.monotonic); const now = milliTimestamp(.monotonic);
if (task.run_at <= now) {
return 0;
}
return @intCast(task.run_at - now);
}
fn runQueue(self: *Scheduler, queue: *Queue, now: u64) !void {
if (queue.count() == 0) {
return;
}
while (queue.peek()) |*task_| { while (queue.peek()) |*task_| {
if (task_.run_at > now) { if (task_.run_at > now) {
return @intCast(task_.run_at - now); return;
} }
var task = queue.remove(); var task = queue.remove();
if (comptime IS_DEBUG) { if (comptime IS_DEBUG) {
@@ -114,7 +122,7 @@ fn runQueue(self: *Scheduler, queue: *Queue) !?u64 {
try self.low_priority.add(task); try self.low_priority.add(task);
} }
} }
return null; return;
} }
fn queueuHasReadyTask(queue: *Queue, now: u64) bool { fn queueuHasReadyTask(queue: *Queue, now: u64) bool {

View File

@@ -24,11 +24,10 @@
<script id=byId name="test1"> <script id=byId name="test1">
testing.expectEqual(1, document.querySelector.length); testing.expectEqual(1, document.querySelector.length);
testing.expectError("SyntaxError: Syntax Error", () => document.querySelector('')); testing.expectError("SyntaxError", () => document.querySelector(''));
testing.withError((err) => { testing.withError((err) => {
testing.expectEqual(12, err.code); testing.expectEqual(12, err.code);
testing.expectEqual("SyntaxError", err.name); testing.expectEqual("SyntaxError", err.name);
testing.expectEqual("Syntax Error", err.message);
}, () => document.querySelector('')); }, () => document.querySelector(''));
testing.expectEqual('test1', document.querySelector('#byId').getAttribute('name')); testing.expectEqual('test1', document.querySelector('#byId').getAttribute('name'));

View File

@@ -34,11 +34,10 @@
</script> </script>
<script id=script1 name="test1"> <script id=script1 name="test1">
testing.expectError("SyntaxError: Syntax Error", () => document.querySelectorAll('')); testing.expectError("SyntaxError", () => document.querySelectorAll(''));
testing.withError((err) => { testing.withError((err) => {
testing.expectEqual(12, err.code); testing.expectEqual(12, err.code);
testing.expectEqual("SyntaxError", err.name); testing.expectEqual("SyntaxError", err.name);
testing.expectEqual("Syntax Error", err.message);
}, () => document.querySelectorAll('')); }, () => document.querySelectorAll(''));
</script> </script>

View File

@@ -127,7 +127,7 @@
testing.withError((err) => { testing.withError((err) => {
testing.expectEqual(3, err.code); testing.expectEqual(3, err.code);
testing.expectEqual('Hierarchy Error', err.message); testing.expectEqual('HierarchyRequestError', err.name);
testing.expectEqual(true, err instanceof DOMException); testing.expectEqual(true, err instanceof DOMException);
testing.expectEqual(true, err instanceof Error); testing.expectEqual(true, err instanceof Error);
}, () => link.appendChild(content)); }, () => link.appendChild(content));

View File

@@ -36,7 +36,6 @@
testing.withError((err) => { testing.withError((err) => {
testing.expectEqual(8, err.code); testing.expectEqual(8, err.code);
testing.expectEqual("NotFoundError", err.name); testing.expectEqual("NotFoundError", err.name);
testing.expectEqual("Not Found", err.message);
}, () => el1.removeAttributeNode(script_id_node)); }, () => el1.removeAttributeNode(script_id_node));
testing.expectEqual(an1, el1.removeAttributeNode(an1)); testing.expectEqual(an1, el1.removeAttributeNode(an1));

View File

@@ -12,7 +12,7 @@
testing.expectEqual('', $('#a0').href); testing.expectEqual('', $('#a0').href);
testing.expectEqual(testing.BASE_URL + 'element/anchor1.html', $('#a1').href); testing.expectEqual(testing.BASE_URL + 'element/anchor1.html', $('#a1').href);
testing.expectEqual(testing.ORIGIN + 'hello/world/anchor2.html', $('#a2').href); testing.expectEqual(testing.ORIGIN + '/hello/world/anchor2.html', $('#a2').href);
testing.expectEqual('https://www.openmymind.net/Elixirs-With-Statement/', $('#a3').href); testing.expectEqual('https://www.openmymind.net/Elixirs-With-Statement/', $('#a3').href);
testing.expectEqual(testing.BASE_URL + 'element/html/foo', $('#link').href); testing.expectEqual(testing.BASE_URL + 'element/html/foo', $('#link').href);

View File

@@ -32,7 +32,7 @@
testing.expectEqual(testing.BASE_URL + 'element/html/hello', form.action) testing.expectEqual(testing.BASE_URL + 'element/html/hello', form.action)
form.action = '/hello'; form.action = '/hello';
testing.expectEqual(testing.ORIGIN + 'hello', form.action) testing.expectEqual(testing.ORIGIN + '/hello', form.action)
form.action = 'https://lightpanda.io/hello'; form.action = 'https://lightpanda.io/hello';
testing.expectEqual('https://lightpanda.io/hello', form.action) testing.expectEqual('https://lightpanda.io/hello', form.action)

View File

@@ -37,7 +37,7 @@
testing.expectEqual('test.png', img.getAttribute('src')); testing.expectEqual('test.png', img.getAttribute('src'));
img.src = '/absolute/path.png'; img.src = '/absolute/path.png';
testing.expectEqual(testing.ORIGIN + 'absolute/path.png', img.src); testing.expectEqual(testing.ORIGIN + '/absolute/path.png', img.src);
testing.expectEqual('/absolute/path.png', img.getAttribute('src')); testing.expectEqual('/absolute/path.png', img.getAttribute('src'));
img.src = 'https://example.com/image.png'; img.src = 'https://example.com/image.png';

View File

@@ -8,7 +8,7 @@
testing.expectEqual('https://lightpanda.io/opensource-browser/15', l2.href); testing.expectEqual('https://lightpanda.io/opensource-browser/15', l2.href);
l2.href = '/over/9000'; l2.href = '/over/9000';
testing.expectEqual(testing.ORIGIN + 'over/9000', l2.href); testing.expectEqual(testing.ORIGIN + '/over/9000', l2.href);
l2.crossOrigin = 'nope'; l2.crossOrigin = 'nope';
testing.expectEqual('anonymous', l2.crossOrigin); testing.expectEqual('anonymous', l2.crossOrigin);

View File

@@ -66,11 +66,10 @@
{ {
const container = $('#test-container'); const container = $('#test-container');
testing.expectError("SyntaxError: Syntax Error", () => container.matches('')); testing.expectError("SyntaxError", () => container.matches(''));
testing.withError((err) => { testing.withError((err) => {
testing.expectEqual(12, err.code); testing.expectEqual(12, err.code);
testing.expectEqual("SyntaxError", err.name); testing.expectEqual("SyntaxError", err.name);
testing.expectEqual("Syntax Error", err.message);
}, () => container.matches('')); }, () => container.matches(''));
} }
</script> </script>

View File

@@ -12,11 +12,10 @@
const p1 = $('#p1'); const p1 = $('#p1');
testing.expectEqual(null, p1.querySelector('#p1')); testing.expectEqual(null, p1.querySelector('#p1'));
testing.expectError("SyntaxError: Syntax Error", () => p1.querySelector('')); testing.expectError("SyntaxError", () => p1.querySelector(''));
testing.withError((err) => { testing.withError((err) => {
testing.expectEqual(12, err.code); testing.expectEqual(12, err.code);
testing.expectEqual("SyntaxError", err.name); testing.expectEqual("SyntaxError", err.name);
testing.expectEqual("Syntax Error", err.message);
}, () => p1.querySelector('')); }, () => p1.querySelector(''));
testing.expectEqual($('#c2'), p1.querySelector('#c2')); testing.expectEqual($('#c2'), p1.querySelector('#c2'));

View File

@@ -24,11 +24,10 @@
<script id=errors> <script id=errors>
{ {
const root = $('#root'); const root = $('#root');
testing.expectError("SyntaxError: Syntax Error", () => root.querySelectorAll('')); testing.expectError("SyntaxError", () => root.querySelectorAll(''));
testing.withError((err) => { testing.withError((err) => {
testing.expectEqual(12, err.code); testing.expectEqual(12, err.code);
testing.expectEqual("SyntaxError", err.name); testing.expectEqual("SyntaxError", err.name);
testing.expectEqual("Syntax Error", err.message);
}, () => root.querySelectorAll('')); }, () => root.querySelectorAll(''));
} }
</script> </script>

View File

@@ -43,8 +43,8 @@
const container = $('#container'); const container = $('#container');
// Empty selectors // Empty selectors
testing.expectError("SyntaxError: Syntax Error", () => container.querySelector('')); testing.expectError("SyntaxError", () => container.querySelector(''));
testing.expectError("SyntaxError: Syntax Error", () => document.querySelectorAll('')); testing.expectError("SyntaxError", () => document.querySelectorAll(''));
} }
</script> </script>

View File

@@ -0,0 +1,25 @@
<!DOCTYPE html>
<script src="../testing.js"></script>
<iframe id="receiver"></iframe>
<script id="messages">
{
let reply = null;
window.addEventListener('message', (e) => {
console.warn('reply')
reply = e.data;
});
const iframe = $('#receiver');
iframe.src = 'support/message_receiver.html';
iframe.addEventListener('load', () => {
iframe.contentWindow.postMessage('ping', '*');
});
testing.eventually(() => {
testing.expectEqual('pong', reply.data);
testing.expectEqual(testing.ORIGIN, reply.origin);
});
}
</script>

View File

@@ -0,0 +1,9 @@
<!DOCTYPE html>
<script>
window.addEventListener('message', (e) => {
console.warn('Frame Message', e.data);
if (e.data === 'ping') {
window.top.postMessage({data: 'pong', origin: e.origin}, '*');
}
});
</script>

View File

@@ -2,37 +2,17 @@
<script src="testing.js"></script> <script src="testing.js"></script>
<script id=history> <script id=history>
testing.expectEqual('auto', history.scrollRestoration); // This test is a bit wonky. But it's trying to test navigation, which is
// something we can't do in the main page (we can't navigate away from this
history.scrollRestoration = 'manual'; // page and still assertOk in the test runner).
testing.expectEqual('manual', history.scrollRestoration); // If support/history.html has a failed assertion, it'll log the error and
// stop the script. If it succeeds, it'll set support_history_completed
history.scrollRestoration = 'auto'; // which we can use here to assume everything passed.
testing.expectEqual('auto', history.scrollRestoration);
testing.expectEqual(null, history.state)
history.pushState({ testInProgress: true }, null, 'http://127.0.0.1:9582/src/browser/tests/history_after_nav.skip.html');
testing.expectEqual({ testInProgress: true }, history.state);
history.pushState({ testInProgress: false }, null, 'http://127.0.0.1:9582/xhr/json');
history.replaceState({ "new": "field", testComplete: true }, null);
let state = { "new": "field", testComplete: true };
testing.expectEqual(state, history.state);
let popstateEventFired = false;
let popstateEventState = null;
window.addEventListener('popstate', (event) => {
popstateEventFired = true;
popstateEventState = event.state;
});
testing.eventually(() => { testing.eventually(() => {
testing.expectEqual(true, popstateEventFired); testing.expectEqual(true, window.support_history_completed);
testing.expectEqual({testInProgress: true }, popstateEventState); testing.expectEqual(true, window.support_history_popstateEventFired);
}) testing.expectEqual({testInProgress: true }, window.support_history_popstateEventState);
});
history.back();
</script> </script>
<iframe id=frame src="support/history.html"></iframe>

View File

@@ -19,7 +19,6 @@
testing.withError((err) => { testing.withError((err) => {
testing.expectEqual(8, err.code); testing.expectEqual(8, err.code);
testing.expectEqual("NotFoundError", err.name); testing.expectEqual("NotFoundError", err.name);
testing.expectEqual("Not Found", err.message);
}, () => d1.insertBefore(document.createElement('div'), d2)); }, () => d1.insertBefore(document.createElement('div'), d2));
let c1 = document.createElement('div'); let c1 = document.createElement('div');

View File

@@ -7,7 +7,6 @@
testing.withError((err) => { testing.withError((err) => {
testing.expectEqual(8, err.code); testing.expectEqual(8, err.code);
testing.expectEqual("NotFoundError", err.name); testing.expectEqual("NotFoundError", err.name);
testing.expectEqual("Not Found", err.message);
}, () => $('#d1').removeChild($('#p1'))); }, () => $('#d1').removeChild($('#p1')));
const p1 = $('#p1'); const p1 = $('#p1');

View File

@@ -25,7 +25,6 @@
testing.withError((err) => { testing.withError((err) => {
testing.expectEqual(3, err.code); testing.expectEqual(3, err.code);
testing.expectEqual("HierarchyRequestError", err.name); testing.expectEqual("HierarchyRequestError", err.name);
testing.expectEqual("Hierarchy Error", err.message);
}, () => d1.replaceChild(c4, c3)); }, () => d1.replaceChild(c4, c3));
testing.expectEqual(c2, d1.replaceChild(c4, c2)); testing.expectEqual(c2, d1.replaceChild(c4, c2));

View File

@@ -451,12 +451,12 @@
const p1 = $('#p1'); const p1 = $('#p1');
// Test setStart with offset beyond node length // Test setStart with offset beyond node length
testing.expectError('IndexSizeError: Index or size is negative or greater than the allowed amount', () => { testing.expectError('IndexSizeError:', () => {
range.setStart(p1, 999); range.setStart(p1, 999);
}); });
// Test with negative offset (wraps to large u32) // Test with negative offset (wraps to large u32)
testing.expectError('IndexSizeError: Index or size is negative or greater than the allowed amount', () => { testing.expectError('IndexSizeError:', () => {
range.setStart(p1.firstChild, -1); range.setStart(p1.firstChild, -1);
}); });
} }
@@ -468,12 +468,12 @@
const p1 = $('#p1'); const p1 = $('#p1');
// Test setEnd with offset beyond node length // Test setEnd with offset beyond node length
testing.expectError('IndexSizeError: Index or size is negative or greater than the allowed amount', () => { testing.expectError('IndexSizeError:', () => {
range.setEnd(p1, 999); range.setEnd(p1, 999);
}); });
// Test with text node // Test with text node
testing.expectError('IndexSizeError: Index or size is negative or greater than the allowed amount', () => { testing.expectError('IndexSizeError:', () => {
range.setEnd(p1.firstChild, 9999); range.setEnd(p1.firstChild, 9999);
}); });
} }
@@ -525,11 +525,11 @@
range.setEnd(p1, 1); range.setEnd(p1, 1);
// Test comparePoint with invalid offset // Test comparePoint with invalid offset
testing.expectError('IndexSizeError: Index or size is negative or greater than the allowed amount', () => { testing.expectError('IndexSizeError:', () => {
range.comparePoint(p1, 20); range.comparePoint(p1, 20);
}); });
testing.expectError('IndexSizeError: Index or size is negative or greater than the allowed amount', () => { testing.expectError('IndexSizeError:', () => {
range.comparePoint(p1.firstChild, -1); range.comparePoint(p1.firstChild, -1);
}); });
} }
@@ -650,11 +650,11 @@
range.setEnd(p1, 1); range.setEnd(p1, 1);
// Invalid offset should throw IndexSizeError // Invalid offset should throw IndexSizeError
testing.expectError('IndexSizeError: Index or size is negative or greater than the allowed amount', () => { testing.expectError('IndexSizeError:', () => {
range.isPointInRange(p1, 999); range.isPointInRange(p1, 999);
}); });
testing.expectError('IndexSizeError: Index or size is negative or greater than the allowed amount', () => { testing.expectError('IndexSizeError:', () => {
range.isPointInRange(p1.firstChild, 9999); range.isPointInRange(p1.firstChild, 9999);
}); });
} }
@@ -854,11 +854,11 @@
range2.setStart(p, 0); range2.setStart(p, 0);
// Invalid how parameter should throw NotSupportedError // Invalid how parameter should throw NotSupportedError
testing.expectError('NotSupportedError: Not Supported', () => { testing.expectError('NotSupportedError:', () => {
range1.compareBoundaryPoints(4, range2); range1.compareBoundaryPoints(4, range2);
}); });
testing.expectError('NotSupportedError: Not Supported', () => { testing.expectError('NotSupportedError:', () => {
range1.compareBoundaryPoints(99, range2); range1.compareBoundaryPoints(99, range2);
}); });
} }
@@ -883,7 +883,7 @@
range2.setEnd(foreignP, 1); range2.setEnd(foreignP, 1);
// Comparing ranges in different documents should throw WrongDocumentError // Comparing ranges in different documents should throw WrongDocumentError
testing.expectError('WrongDocumentError: wrong_document_error', () => { testing.expectError('WrongDocumentError:', () => {
range1.compareBoundaryPoints(Range.START_TO_START, range2); range1.compareBoundaryPoints(Range.START_TO_START, range2);
}); });
} }

View File

@@ -0,0 +1,33 @@
<!DOCTYPE html>
<script src="../testing.js"></script>
<script id=history>
testing.expectEqual('auto', history.scrollRestoration);
history.scrollRestoration = 'manual';
testing.expectEqual('manual', history.scrollRestoration);
history.scrollRestoration = 'auto';
testing.expectEqual('auto', history.scrollRestoration);
testing.expectEqual(null, history.state)
history.pushState({ testInProgress: true }, null, testing.BASE_URL + 'history_after_nav.skip.html');
testing.expectEqual({ testInProgress: true }, history.state);
history.pushState({ testInProgress: false }, null, testing.ORIGIN + '/xhr/json');
history.replaceState({ "new": "field", testComplete: true }, null);
let state = { "new": "field", testComplete: true };
testing.expectEqual(state, history.state);
let popstateEventFired = false;
let popstateEventState = null;
window.top.support_history_completed = true;
window.addEventListener('popstate', (event) => {
window.top.window.support_history_popstateEventFired = true;
window.top.window.support_history_popstateEventState = event.state;
});
history.back();
</script>

View File

@@ -99,8 +99,7 @@
} }
} }
// our test runner sets this to true const IS_TEST_RUNNER = window.navigator.userAgent.startsWith("Lightpanda/");
const IS_TEST_RUNNER = window._lightpanda_skip_auto_assert === true;
window.testing = { window.testing = {
fail: fail, fail: fail,
@@ -114,17 +113,17 @@
eventually: eventually, eventually: eventually,
IS_TEST_RUNNER: IS_TEST_RUNNER, IS_TEST_RUNNER: IS_TEST_RUNNER,
HOST: '127.0.0.1', HOST: '127.0.0.1',
ORIGIN: 'http://127.0.0.1:9582/', ORIGIN: 'http://127.0.0.1:9582',
BASE_URL: 'http://127.0.0.1:9582/src/browser/tests/', BASE_URL: 'http://127.0.0.1:9582/src/browser/tests/',
}; };
if (window.navigator.userAgent.startsWith("Lightpanda/") == false) { if (IS_TEST_RUNNER === false) {
// The page is running in a different browser. Probably a developer making sure // The page is running in a different browser. Probably a developer making sure
// a test is correct. There are a few tweaks we need to do to make this a // a test is correct. There are a few tweaks we need to do to make this a
// seemless, namely around adapting paths/urls. // seemless, namely around adapting paths/urls.
console.warn(`The page is not being executed in the test runner, certain behavior has been adjusted`); console.warn(`The page is not being executed in the test runner, certain behavior has been adjusted`);
window.testing.HOST = location.hostname; window.testing.HOST = location.hostname;
window.testing.ORIGIN = location.origin + '/'; window.testing.ORIGIN = location.origin;
window.testing.BASE_URL = location.origin + '/src/browser/tests/'; window.testing.BASE_URL = location.origin + '/src/browser/tests/';
window.addEventListener('load', testing.assertOk); window.addEventListener('load', testing.assertOk);
} }

View File

@@ -82,7 +82,7 @@
testing.expectEqual('ceil', atob('Y2VpbA')); // 6 chars, len%4==2, needs '==' testing.expectEqual('ceil', atob('Y2VpbA')); // 6 chars, len%4==2, needs '=='
// length % 4 == 1 must still throw // length % 4 == 1 must still throw
testing.expectError('InvalidCharacterError: Invalid Character', () => { testing.expectError('InvalidCharacterError', () => {
atob('Y'); atob('Y');
}); });
</script> </script>

View File

@@ -104,13 +104,27 @@ pub fn getMessage(self: *const DOMException) []const u8 {
} }
return switch (self._code) { return switch (self._code) {
.none => "", .none => "",
.invalid_character_error => "Invalid Character",
.index_size_error => "Index or size is negative or greater than the allowed amount", .index_size_error => "Index or size is negative or greater than the allowed amount",
.syntax_error => "Syntax Error", .hierarchy_error => "The operation would yield an incorrect node tree",
.not_supported => "Not Supported", .wrong_document_error => "The object is in the wrong document",
.not_found => "Not Found", .invalid_character_error => "The string contains invalid characters",
.hierarchy_error => "Hierarchy Error", .no_modification_allowed_error => "The object can not be modified",
else => @tagName(self._code), .not_found => "The object can not be found here",
.not_supported => "The operation is not supported",
.inuse_attribute_error => "The attribute already in use",
.invalid_state_error => "The object is in an invalid state",
.syntax_error => "The string did not match the expected pattern",
.invalid_modification_error => "The object can not be modified in this way",
.namespace_error => "The operation is not allowed by Namespaces in XML",
.invalid_access_error => "The object does not support the operation or argument",
.security_error => "The operation is insecure",
.network_error => "A network error occurred",
.abort_error => "The operation was aborted",
.url_mismatch_error => "The given URL does not match another URL",
.quota_exceeded_error => "The quota has been exceeded",
.timeout_error => "The operation timed out",
.invalid_node_type_error => "The supplied node is incorrect or has an incorrect ancestor for this operation",
.data_clone_error => "The object can not be cloned",
}; };
} }

View File

@@ -37,6 +37,7 @@ pub fn registerTypes() []const type {
const IntersectionObserver = @This(); const IntersectionObserver = @This();
_rc: u8 = 0,
_arena: Allocator, _arena: Allocator,
_callback: js.Function.Temp, _callback: js.Function.Temp,
_observing: std.ArrayList(*Element) = .{}, _observing: std.ArrayList(*Element) = .{},
@@ -93,12 +94,24 @@ pub fn init(callback: js.Function.Temp, options: ?ObserverInit, page: *Page) !*I
} }
pub fn deinit(self: *IntersectionObserver, shutdown: bool, session: *Session) void { pub fn deinit(self: *IntersectionObserver, shutdown: bool, session: *Session) void {
self._callback.release(); const rc = self._rc;
if ((comptime IS_DEBUG) and !shutdown) { if (comptime IS_DEBUG) {
std.debug.assert(self._observing.items.len == 0); std.debug.assert(rc != 0);
} }
session.releaseArena(self._arena); if (rc == 1 or shutdown) {
self._callback.release();
if ((comptime IS_DEBUG) and !shutdown) {
std.debug.assert(self._observing.items.len == 0);
}
session.releaseArena(self._arena);
} else {
self._rc = rc - 1;
}
}
pub fn acquireRef(self: *IntersectionObserver) void {
self._rc += 1;
} }
pub fn observe(self: *IntersectionObserver, target: *Element, page: *Page) !void { pub fn observe(self: *IntersectionObserver, target: *Element, page: *Page) !void {
@@ -111,7 +124,7 @@ pub fn observe(self: *IntersectionObserver, target: *Element, page: *Page) !void
// Register with page if this is our first observation // Register with page if this is our first observation
if (self._observing.items.len == 0) { if (self._observing.items.len == 0) {
page.js.strongRef(self); self._rc += 1;
try page.registerIntersectionObserver(self); try page.registerIntersectionObserver(self);
} }
@@ -148,20 +161,26 @@ pub fn unobserve(self: *IntersectionObserver, target: *Element, page: *Page) voi
} }
if (self._observing.items.len == 0) { if (self._observing.items.len == 0) {
page.js.safeWeakRef(self); self.deinit(false, page._session);
} }
} }
pub fn disconnect(self: *IntersectionObserver, page: *Page) void { pub fn disconnect(self: *IntersectionObserver, page: *Page) void {
page.unregisterIntersectionObserver(self);
self._observing.clearRetainingCapacity();
self._previous_states.clearRetainingCapacity(); self._previous_states.clearRetainingCapacity();
for (self._pending_entries.items) |entry| { for (self._pending_entries.items) |entry| {
entry.deinit(false, page._session); entry.deinit(false, page._session);
} }
self._pending_entries.clearRetainingCapacity(); self._pending_entries.clearRetainingCapacity();
page.js.safeWeakRef(self);
const observing_count = self._observing.items.len;
self._observing.clearRetainingCapacity();
if (observing_count > 0) {
self.deinit(false, page._session);
}
page.unregisterIntersectionObserver(self);
} }
pub fn takeRecords(self: *IntersectionObserver, page: *Page) ![]*IntersectionObserverEntry { pub fn takeRecords(self: *IntersectionObserver, page: *Page) ![]*IntersectionObserverEntry {

View File

@@ -39,6 +39,7 @@ pub fn registerTypes() []const type {
const MutationObserver = @This(); const MutationObserver = @This();
_rc: u8 = 0,
_arena: Allocator, _arena: Allocator,
_callback: js.Function.Temp, _callback: js.Function.Temp,
_observing: std.ArrayList(Observing) = .{}, _observing: std.ArrayList(Observing) = .{},
@@ -86,12 +87,24 @@ pub fn init(callback: js.Function.Temp, page: *Page) !*MutationObserver {
} }
pub fn deinit(self: *MutationObserver, shutdown: bool, session: *Session) void { pub fn deinit(self: *MutationObserver, shutdown: bool, session: *Session) void {
self._callback.release(); const rc = self._rc;
if ((comptime IS_DEBUG) and !shutdown) { if (comptime IS_DEBUG) {
std.debug.assert(self._observing.items.len == 0); std.debug.assert(rc != 0);
} }
session.releaseArena(self._arena); if (rc == 1 or shutdown) {
self._callback.release();
if ((comptime IS_DEBUG) and !shutdown) {
std.debug.assert(self._observing.items.len == 0);
}
session.releaseArena(self._arena);
} else {
self._rc = rc - 1;
}
}
pub fn acquireRef(self: *MutationObserver) void {
self._rc += 1;
} }
pub fn observe(self: *MutationObserver, target: *Node, options: ObserveOptions, page: *Page) !void { pub fn observe(self: *MutationObserver, target: *Node, options: ObserveOptions, page: *Page) !void {
@@ -158,7 +171,7 @@ pub fn observe(self: *MutationObserver, target: *Node, options: ObserveOptions,
// Register with page if this is our first observation // Register with page if this is our first observation
if (self._observing.items.len == 0) { if (self._observing.items.len == 0) {
page.js.strongRef(self); self._rc += 1;
try page.registerMutationObserver(self); try page.registerMutationObserver(self);
} }
@@ -169,13 +182,18 @@ pub fn observe(self: *MutationObserver, target: *Node, options: ObserveOptions,
} }
pub fn disconnect(self: *MutationObserver, page: *Page) void { pub fn disconnect(self: *MutationObserver, page: *Page) void {
page.unregisterMutationObserver(self);
self._observing.clearRetainingCapacity();
for (self._pending_records.items) |record| { for (self._pending_records.items) |record| {
record.deinit(false, page._session); record.deinit(false, page._session);
} }
self._pending_records.clearRetainingCapacity(); self._pending_records.clearRetainingCapacity();
page.js.safeWeakRef(self);
const observing_count = self._observing.items.len;
self._observing.clearRetainingCapacity();
if (observing_count > 0) {
self.deinit(false, page._session);
}
page.unregisterMutationObserver(self);
} }
pub fn takeRecords(self: *MutationObserver, page: *Page) ![]*MutationRecord { pub fn takeRecords(self: *MutationObserver, page: *Page) ![]*MutationRecord {

View File

@@ -66,6 +66,7 @@ _on_load: ?js.Function.Global = null,
_on_pageshow: ?js.Function.Global = null, _on_pageshow: ?js.Function.Global = null,
_on_popstate: ?js.Function.Global = null, _on_popstate: ?js.Function.Global = null,
_on_error: ?js.Function.Global = null, _on_error: ?js.Function.Global = null,
_on_message: ?js.Function.Global = null,
_on_unhandled_rejection: ?js.Function.Global = null, // TODO: invoke on error _on_unhandled_rejection: ?js.Function.Global = null, // TODO: invoke on error
_location: *Location, _location: *Location,
_timer_id: u30 = 0, _timer_id: u30 = 0,
@@ -208,6 +209,14 @@ pub fn setOnError(self: *Window, setter: ?FunctionSetter) void {
self._on_error = getFunctionFromSetter(setter); self._on_error = getFunctionFromSetter(setter);
} }
pub fn getOnMessage(self: *const Window) ?js.Function.Global {
return self._on_message;
}
pub fn setOnMessage(self: *Window, setter: ?FunctionSetter) void {
self._on_message = getFunctionFromSetter(setter);
}
pub fn getOnUnhandledRejection(self: *const Window) ?js.Function.Global { pub fn getOnUnhandledRejection(self: *const Window) ?js.Function.Global {
return self._on_unhandled_rejection; return self._on_unhandled_rejection;
} }
@@ -369,19 +378,26 @@ pub fn postMessage(self: *Window, message: js.Value.Temp, target_origin: ?[]cons
// In a full implementation, we would validate the origin // In a full implementation, we would validate the origin
_ = target_origin; _ = target_origin;
// postMessage queues a task (not a microtask), so use the scheduler // self = the window that will get the message
const arena = try page.getArena(.{ .debug = "Window.schedule" }); // page = the context calling postMessage
errdefer page.releaseArena(arena); const target_page = self._page;
const source_window = target_page.js.getIncumbent().window;
const origin = try self._location.getOrigin(page); const arena = try target_page.getArena(.{ .debug = "Window.postMessage" });
errdefer target_page.releaseArena(arena);
// Origin should be the source window's origin (where the message came from)
const origin = try source_window._location.getOrigin(page);
const callback = try arena.create(PostMessageCallback); const callback = try arena.create(PostMessageCallback);
callback.* = .{ callback.* = .{
.page = page,
.arena = arena, .arena = arena,
.message = message, .message = message,
.page = target_page,
.source = source_window,
.origin = try arena.dupe(u8, origin), .origin = try arena.dupe(u8, origin),
}; };
try page.js.scheduler.add(callback, PostMessageCallback.run, 0, .{
try target_page.js.scheduler.add(callback, PostMessageCallback.run, 0, .{
.name = "postMessage", .name = "postMessage",
.low_priority = false, .low_priority = false,
.finalizer = PostMessageCallback.cancelled, .finalizer = PostMessageCallback.cancelled,
@@ -702,6 +718,7 @@ const ScheduleCallback = struct {
const PostMessageCallback = struct { const PostMessageCallback = struct {
page: *Page, page: *Page,
source: *Window,
arena: Allocator, arena: Allocator,
origin: []const u8, origin: []const u8,
message: js.Value.Temp, message: js.Value.Temp,
@@ -712,7 +729,7 @@ const PostMessageCallback = struct {
fn cancelled(ctx: *anyopaque) void { fn cancelled(ctx: *anyopaque) void {
const self: *PostMessageCallback = @ptrCast(@alignCast(ctx)); const self: *PostMessageCallback = @ptrCast(@alignCast(ctx));
self.page.releaseArena(self.arena); self.deinit();
} }
fn run(ctx: *anyopaque) !?u32 { fn run(ctx: *anyopaque) !?u32 {
@@ -722,14 +739,17 @@ const PostMessageCallback = struct {
const page = self.page; const page = self.page;
const window = page.window; const window = page.window;
const event = (try MessageEvent.initTrusted(comptime .wrap("message"), .{ const event_target = window.asEventTarget();
.data = self.message, if (page._event_manager.hasDirectListeners(event_target, "message", window._on_message)) {
.origin = self.origin, const event = (try MessageEvent.initTrusted(comptime .wrap("message"), .{
.source = window, .data = self.message,
.bubbles = false, .origin = self.origin,
.cancelable = false, .source = self.source,
}, page)).asEvent(); .bubbles = false,
try page._event_manager.dispatch(window.asEventTarget(), event); .cancelable = false,
}, page)).asEvent();
try page._event_manager.dispatchDirect(event_target, event, window._on_message, .{ .context = "window.postMessage" });
}
return null; return null;
} }
@@ -783,6 +803,7 @@ pub const JsApi = struct {
pub const onpageshow = bridge.accessor(Window.getOnPageShow, Window.setOnPageShow, .{}); pub const onpageshow = bridge.accessor(Window.getOnPageShow, Window.setOnPageShow, .{});
pub const onpopstate = bridge.accessor(Window.getOnPopState, Window.setOnPopState, .{}); pub const onpopstate = bridge.accessor(Window.getOnPopState, Window.setOnPopState, .{});
pub const onerror = bridge.accessor(Window.getOnError, Window.setOnError, .{}); pub const onerror = bridge.accessor(Window.getOnError, Window.setOnError, .{});
pub const onmessage = bridge.accessor(Window.getOnMessage, Window.setOnMessage, .{});
pub const onunhandledrejection = bridge.accessor(Window.getOnUnhandledRejection, Window.setOnUnhandledRejection, .{}); pub const onunhandledrejection = bridge.accessor(Window.getOnUnhandledRejection, Window.setOnUnhandledRejection, .{});
pub const fetch = bridge.function(Window.fetch, .{}); pub const fetch = bridge.function(Window.fetch, .{});
pub const queueMicrotask = bridge.function(Window.queueMicrotask, .{}); pub const queueMicrotask = bridge.function(Window.queueMicrotask, .{});

View File

@@ -28,6 +28,8 @@ const EventTarget = @import("../EventTarget.zig");
const UIEvent = @import("UIEvent.zig"); const UIEvent = @import("UIEvent.zig");
const PointerEvent = @import("PointerEvent.zig"); const PointerEvent = @import("PointerEvent.zig");
const Allocator = std.mem.Allocator;
const MouseEvent = @This(); const MouseEvent = @This();
pub const MouseButton = enum(u8) { pub const MouseButton = enum(u8) {
@@ -92,7 +94,7 @@ pub fn initTrusted(typ: String, _opts: ?Options, page: *Page) !*MouseEvent {
return initWithTrusted(arena, typ, _opts, true, page); return initWithTrusted(arena, typ, _opts, true, page);
} }
fn initWithTrusted(arena: std.mem.Allocator, typ: String, _opts: ?Options, trusted: bool, page: *Page) !*MouseEvent { fn initWithTrusted(arena: Allocator, typ: String, _opts: ?Options, trusted: bool, page: *Page) !*MouseEvent {
const opts = _opts orelse Options{}; const opts = _opts orelse Options{};
const event = try page._factory.uiEvent( const event = try page._factory.uiEvent(

View File

@@ -21,9 +21,11 @@ const std = @import("std");
pub fn processMessage(cmd: anytype) !void { pub fn processMessage(cmd: anytype) !void {
const action = std.meta.stringToEnum(enum { const action = std.meta.stringToEnum(enum {
enable, enable,
disable,
}, cmd.input.action) orelse return error.UnknownMethod; }, cmd.input.action) orelse return error.UnknownMethod;
switch (action) { switch (action) {
.enable => return cmd.sendResult(null, .{}), .enable => return cmd.sendResult(null, .{}),
.disable => return cmd.sendResult(null, .{}),
} }
} }

View File

@@ -21,9 +21,11 @@ const std = @import("std");
pub fn processMessage(cmd: anytype) !void { pub fn processMessage(cmd: anytype) !void {
const action = std.meta.stringToEnum(enum { const action = std.meta.stringToEnum(enum {
enable, enable,
disable,
}, cmd.input.action) orelse return error.UnknownMethod; }, cmd.input.action) orelse return error.UnknownMethod;
switch (action) { switch (action) {
.enable => return cmd.sendResult(null, .{}), .enable => return cmd.sendResult(null, .{}),
.disable => return cmd.sendResult(null, .{}),
} }
} }

View File

@@ -21,11 +21,13 @@ const std = @import("std");
pub fn processMessage(cmd: anytype) !void { pub fn processMessage(cmd: anytype) !void {
const action = std.meta.stringToEnum(enum { const action = std.meta.stringToEnum(enum {
enable, enable,
disable,
setIgnoreCertificateErrors, setIgnoreCertificateErrors,
}, cmd.input.action) orelse return error.UnknownMethod; }, cmd.input.action) orelse return error.UnknownMethod;
switch (action) { switch (action) {
.enable => return cmd.sendResult(null, .{}), .enable => return cmd.sendResult(null, .{}),
.disable => return cmd.sendResult(null, .{}),
.setIgnoreCertificateErrors => return setIgnoreCertificateErrors(cmd), .setIgnoreCertificateErrors => return setIgnoreCertificateErrors(cmd),
} }
} }