- Delete old Vite+Svelte frontend - Initialize new SvelteKit project with TypeScript - Configure Tailwind CSS v4 + DaisyUI - Implement JWT authentication with auto-refresh - Create login page with form validation (Zod) - Add protected route guards - Update Docker configuration for single-stage build - Add E2E tests with Playwright (6/11 passing) - Fix Svelte 5 reactivity with $state() runes Known issues: - 5 E2E tests failing (timing/async issues) - Token refresh implementation needs debugging - Validation error display timing
771 lines
25 KiB
JavaScript
771 lines
25 KiB
JavaScript
import { json, text } from "@sveltejs/kit";
|
|
import { SvelteKitError, HttpError } from "@sveltejs/kit/internal";
|
|
import { with_request_store } from "@sveltejs/kit/internal/server";
|
|
import * as devalue from "devalue";
|
|
import { t as text_decoder, b as base64_encode, c as base64_decode } from "./utils.js";
|
|
const SVELTE_KIT_ASSETS = "/_svelte_kit_assets";
|
|
const ENDPOINT_METHODS = ["GET", "POST", "PUT", "PATCH", "DELETE", "OPTIONS", "HEAD"];
|
|
const PAGE_METHODS = ["GET", "POST", "HEAD"];
|
|
function set_nested_value(object, path_string, value) {
|
|
if (path_string.startsWith("n:")) {
|
|
path_string = path_string.slice(2);
|
|
value = value === "" ? void 0 : parseFloat(value);
|
|
} else if (path_string.startsWith("b:")) {
|
|
path_string = path_string.slice(2);
|
|
value = value === "on";
|
|
}
|
|
deep_set(object, split_path(path_string), value);
|
|
}
|
|
function convert_formdata(data) {
|
|
const result = {};
|
|
for (let key of data.keys()) {
|
|
const is_array = key.endsWith("[]");
|
|
let values = data.getAll(key);
|
|
if (is_array) key = key.slice(0, -2);
|
|
if (values.length > 1 && !is_array) {
|
|
throw new Error(`Form cannot contain duplicated keys — "${key}" has ${values.length} values`);
|
|
}
|
|
values = values.filter(
|
|
(entry) => typeof entry === "string" || entry.name !== "" || entry.size > 0
|
|
);
|
|
if (key.startsWith("n:")) {
|
|
key = key.slice(2);
|
|
values = values.map((v) => v === "" ? void 0 : parseFloat(
|
|
/** @type {string} */
|
|
v
|
|
));
|
|
} else if (key.startsWith("b:")) {
|
|
key = key.slice(2);
|
|
values = values.map((v) => v === "on");
|
|
}
|
|
set_nested_value(result, key, is_array ? values : values[0]);
|
|
}
|
|
return result;
|
|
}
|
|
const BINARY_FORM_CONTENT_TYPE = "application/x-sveltekit-formdata";
|
|
const BINARY_FORM_VERSION = 0;
|
|
const HEADER_BYTES = 1 + 4 + 2;
|
|
async function deserialize_binary_form(request) {
|
|
if (request.headers.get("content-type") !== BINARY_FORM_CONTENT_TYPE) {
|
|
const form_data = await request.formData();
|
|
return { data: convert_formdata(form_data), meta: {}, form_data };
|
|
}
|
|
if (!request.body) {
|
|
throw deserialize_error("no body");
|
|
}
|
|
const content_length = parseInt(request.headers.get("content-length") ?? "");
|
|
if (Number.isNaN(content_length)) {
|
|
throw deserialize_error("invalid Content-Length header");
|
|
}
|
|
const reader = request.body.getReader();
|
|
const chunks = [];
|
|
function get_chunk(index) {
|
|
if (index in chunks) return chunks[index];
|
|
let i = chunks.length;
|
|
while (i <= index) {
|
|
chunks[i] = reader.read().then((chunk) => chunk.value);
|
|
i++;
|
|
}
|
|
return chunks[index];
|
|
}
|
|
async function get_buffer(offset, length) {
|
|
let start_chunk;
|
|
let chunk_start = 0;
|
|
let chunk_index;
|
|
for (chunk_index = 0; ; chunk_index++) {
|
|
const chunk = await get_chunk(chunk_index);
|
|
if (!chunk) return null;
|
|
const chunk_end = chunk_start + chunk.byteLength;
|
|
if (offset >= chunk_start && offset < chunk_end) {
|
|
start_chunk = chunk;
|
|
break;
|
|
}
|
|
chunk_start = chunk_end;
|
|
}
|
|
if (offset + length <= chunk_start + start_chunk.byteLength) {
|
|
return start_chunk.subarray(offset - chunk_start, offset + length - chunk_start);
|
|
}
|
|
const chunks2 = [start_chunk.subarray(offset - chunk_start)];
|
|
let cursor = start_chunk.byteLength - offset + chunk_start;
|
|
while (cursor < length) {
|
|
chunk_index++;
|
|
let chunk = await get_chunk(chunk_index);
|
|
if (!chunk) return null;
|
|
if (chunk.byteLength > length - cursor) {
|
|
chunk = chunk.subarray(0, length - cursor);
|
|
}
|
|
chunks2.push(chunk);
|
|
cursor += chunk.byteLength;
|
|
}
|
|
const buffer = new Uint8Array(length);
|
|
cursor = 0;
|
|
for (const chunk of chunks2) {
|
|
buffer.set(chunk, cursor);
|
|
cursor += chunk.byteLength;
|
|
}
|
|
return buffer;
|
|
}
|
|
const header = await get_buffer(0, HEADER_BYTES);
|
|
if (!header) throw deserialize_error("too short");
|
|
if (header[0] !== BINARY_FORM_VERSION) {
|
|
throw deserialize_error(`got version ${header[0]}, expected version ${BINARY_FORM_VERSION}`);
|
|
}
|
|
const header_view = new DataView(header.buffer, header.byteOffset, header.byteLength);
|
|
const data_length = header_view.getUint32(1, true);
|
|
if (HEADER_BYTES + data_length > content_length) {
|
|
throw deserialize_error("data overflow");
|
|
}
|
|
const file_offsets_length = header_view.getUint16(5, true);
|
|
if (HEADER_BYTES + data_length + file_offsets_length > content_length) {
|
|
throw deserialize_error("file offset table overflow");
|
|
}
|
|
const data_buffer = await get_buffer(HEADER_BYTES, data_length);
|
|
if (!data_buffer) throw deserialize_error("data too short");
|
|
let file_offsets;
|
|
let files_start_offset;
|
|
if (file_offsets_length > 0) {
|
|
const file_offsets_buffer = await get_buffer(HEADER_BYTES + data_length, file_offsets_length);
|
|
if (!file_offsets_buffer) throw deserialize_error("file offset table too short");
|
|
file_offsets = /** @type {Array<number>} */
|
|
JSON.parse(text_decoder.decode(file_offsets_buffer));
|
|
files_start_offset = HEADER_BYTES + data_length + file_offsets_length;
|
|
}
|
|
const [data, meta] = devalue.parse(text_decoder.decode(data_buffer), {
|
|
File: ([name, type, size, last_modified, index]) => {
|
|
if (files_start_offset + file_offsets[index] + size > content_length) {
|
|
throw deserialize_error("file data overflow");
|
|
}
|
|
return new Proxy(
|
|
new LazyFile(
|
|
name,
|
|
type,
|
|
size,
|
|
last_modified,
|
|
get_chunk,
|
|
files_start_offset + file_offsets[index]
|
|
),
|
|
{
|
|
getPrototypeOf() {
|
|
return File.prototype;
|
|
}
|
|
}
|
|
);
|
|
}
|
|
});
|
|
void (async () => {
|
|
let has_more = true;
|
|
while (has_more) {
|
|
const chunk = await get_chunk(chunks.length);
|
|
has_more = !!chunk;
|
|
}
|
|
})();
|
|
return { data, meta, form_data: null };
|
|
}
|
|
function deserialize_error(message) {
|
|
return new SvelteKitError(400, "Bad Request", `Could not deserialize binary form: ${message}`);
|
|
}
|
|
class LazyFile {
|
|
/** @type {(index: number) => Promise<Uint8Array<ArrayBuffer> | undefined>} */
|
|
#get_chunk;
|
|
/** @type {number} */
|
|
#offset;
|
|
/**
|
|
* @param {string} name
|
|
* @param {string} type
|
|
* @param {number} size
|
|
* @param {number} last_modified
|
|
* @param {(index: number) => Promise<Uint8Array<ArrayBuffer> | undefined>} get_chunk
|
|
* @param {number} offset
|
|
*/
|
|
constructor(name, type, size, last_modified, get_chunk, offset) {
|
|
this.name = name;
|
|
this.type = type;
|
|
this.size = size;
|
|
this.lastModified = last_modified;
|
|
this.webkitRelativePath = "";
|
|
this.#get_chunk = get_chunk;
|
|
this.#offset = offset;
|
|
this.arrayBuffer = this.arrayBuffer.bind(this);
|
|
this.bytes = this.bytes.bind(this);
|
|
this.slice = this.slice.bind(this);
|
|
this.stream = this.stream.bind(this);
|
|
this.text = this.text.bind(this);
|
|
}
|
|
/** @type {ArrayBuffer | undefined} */
|
|
#buffer;
|
|
async arrayBuffer() {
|
|
this.#buffer ??= await new Response(this.stream()).arrayBuffer();
|
|
return this.#buffer;
|
|
}
|
|
async bytes() {
|
|
return new Uint8Array(await this.arrayBuffer());
|
|
}
|
|
/**
|
|
* @param {number=} start
|
|
* @param {number=} end
|
|
* @param {string=} contentType
|
|
*/
|
|
slice(start = 0, end = this.size, contentType = this.type) {
|
|
if (start < 0) {
|
|
start = Math.max(this.size + start, 0);
|
|
} else {
|
|
start = Math.min(start, this.size);
|
|
}
|
|
if (end < 0) {
|
|
end = Math.max(this.size + end, 0);
|
|
} else {
|
|
end = Math.min(end, this.size);
|
|
}
|
|
const size = Math.max(end - start, 0);
|
|
const file = new LazyFile(
|
|
this.name,
|
|
contentType,
|
|
size,
|
|
this.lastModified,
|
|
this.#get_chunk,
|
|
this.#offset + start
|
|
);
|
|
return file;
|
|
}
|
|
stream() {
|
|
let cursor = 0;
|
|
let chunk_index = 0;
|
|
return new ReadableStream({
|
|
start: async (controller) => {
|
|
let chunk_start = 0;
|
|
let start_chunk = null;
|
|
for (chunk_index = 0; ; chunk_index++) {
|
|
const chunk = await this.#get_chunk(chunk_index);
|
|
if (!chunk) return null;
|
|
const chunk_end = chunk_start + chunk.byteLength;
|
|
if (this.#offset >= chunk_start && this.#offset < chunk_end) {
|
|
start_chunk = chunk;
|
|
break;
|
|
}
|
|
chunk_start = chunk_end;
|
|
}
|
|
if (this.#offset + this.size <= chunk_start + start_chunk.byteLength) {
|
|
controller.enqueue(
|
|
start_chunk.subarray(this.#offset - chunk_start, this.#offset + this.size - chunk_start)
|
|
);
|
|
controller.close();
|
|
} else {
|
|
controller.enqueue(start_chunk.subarray(this.#offset - chunk_start));
|
|
cursor = start_chunk.byteLength - this.#offset + chunk_start;
|
|
}
|
|
},
|
|
pull: async (controller) => {
|
|
chunk_index++;
|
|
let chunk = await this.#get_chunk(chunk_index);
|
|
if (!chunk) {
|
|
controller.error("incomplete file data");
|
|
controller.close();
|
|
return;
|
|
}
|
|
if (chunk.byteLength > this.size - cursor) {
|
|
chunk = chunk.subarray(0, this.size - cursor);
|
|
}
|
|
controller.enqueue(chunk);
|
|
cursor += chunk.byteLength;
|
|
if (cursor >= this.size) {
|
|
controller.close();
|
|
}
|
|
}
|
|
});
|
|
}
|
|
async text() {
|
|
return text_decoder.decode(await this.arrayBuffer());
|
|
}
|
|
}
|
|
const path_regex = /^[a-zA-Z_$]\w*(\.[a-zA-Z_$]\w*|\[\d+\])*$/;
|
|
function split_path(path) {
|
|
if (!path_regex.test(path)) {
|
|
throw new Error(`Invalid path ${path}`);
|
|
}
|
|
return path.split(/\.|\[|\]/).filter(Boolean);
|
|
}
|
|
function check_prototype_pollution(key) {
|
|
if (key === "__proto__" || key === "constructor" || key === "prototype") {
|
|
throw new Error(
|
|
`Invalid key "${key}"`
|
|
);
|
|
}
|
|
}
|
|
function deep_set(object, keys, value) {
|
|
let current = object;
|
|
for (let i = 0; i < keys.length - 1; i += 1) {
|
|
const key = keys[i];
|
|
check_prototype_pollution(key);
|
|
const is_array = /^\d+$/.test(keys[i + 1]);
|
|
const exists = Object.hasOwn(current, key);
|
|
const inner = current[key];
|
|
if (exists && is_array !== Array.isArray(inner)) {
|
|
throw new Error(`Invalid array key ${keys[i + 1]}`);
|
|
}
|
|
if (!exists) {
|
|
current[key] = is_array ? [] : {};
|
|
}
|
|
current = current[key];
|
|
}
|
|
const final_key = keys[keys.length - 1];
|
|
check_prototype_pollution(final_key);
|
|
current[final_key] = value;
|
|
}
|
|
function normalize_issue(issue, server = false) {
|
|
const normalized = { name: "", path: [], message: issue.message, server };
|
|
if (issue.path !== void 0) {
|
|
let name = "";
|
|
for (const segment of issue.path) {
|
|
const key = (
|
|
/** @type {string | number} */
|
|
typeof segment === "object" ? segment.key : segment
|
|
);
|
|
normalized.path.push(key);
|
|
if (typeof key === "number") {
|
|
name += `[${key}]`;
|
|
} else if (typeof key === "string") {
|
|
name += name === "" ? key : "." + key;
|
|
}
|
|
}
|
|
normalized.name = name;
|
|
}
|
|
return normalized;
|
|
}
|
|
function flatten_issues(issues) {
|
|
const result = {};
|
|
for (const issue of issues) {
|
|
(result.$ ??= []).push(issue);
|
|
let name = "";
|
|
if (issue.path !== void 0) {
|
|
for (const key of issue.path) {
|
|
if (typeof key === "number") {
|
|
name += `[${key}]`;
|
|
} else if (typeof key === "string") {
|
|
name += name === "" ? key : "." + key;
|
|
}
|
|
(result[name] ??= []).push(issue);
|
|
}
|
|
}
|
|
}
|
|
return result;
|
|
}
|
|
function deep_get(object, path) {
|
|
let current = object;
|
|
for (const key of path) {
|
|
if (current == null || typeof current !== "object") {
|
|
return current;
|
|
}
|
|
current = current[key];
|
|
}
|
|
return current;
|
|
}
|
|
function create_field_proxy(target, get_input, set_input, get_issues, path = []) {
|
|
const get_value = () => {
|
|
return deep_get(get_input(), path);
|
|
};
|
|
return new Proxy(target, {
|
|
get(target2, prop) {
|
|
if (typeof prop === "symbol") return target2[prop];
|
|
if (/^\d+$/.test(prop)) {
|
|
return create_field_proxy({}, get_input, set_input, get_issues, [
|
|
...path,
|
|
parseInt(prop, 10)
|
|
]);
|
|
}
|
|
const key = build_path_string(path);
|
|
if (prop === "set") {
|
|
const set_func = function(newValue) {
|
|
set_input(path, newValue);
|
|
return newValue;
|
|
};
|
|
return create_field_proxy(set_func, get_input, set_input, get_issues, [...path, prop]);
|
|
}
|
|
if (prop === "value") {
|
|
return create_field_proxy(get_value, get_input, set_input, get_issues, [...path, prop]);
|
|
}
|
|
if (prop === "issues" || prop === "allIssues") {
|
|
const issues_func = () => {
|
|
const all_issues = get_issues()[key === "" ? "$" : key];
|
|
if (prop === "allIssues") {
|
|
return all_issues?.map((issue) => ({
|
|
path: issue.path,
|
|
message: issue.message
|
|
}));
|
|
}
|
|
return all_issues?.filter((issue) => issue.name === key)?.map((issue) => ({
|
|
path: issue.path,
|
|
message: issue.message
|
|
}));
|
|
};
|
|
return create_field_proxy(issues_func, get_input, set_input, get_issues, [...path, prop]);
|
|
}
|
|
if (prop === "as") {
|
|
const as_func = (type, input_value) => {
|
|
const is_array = type === "file multiple" || type === "select multiple" || type === "checkbox" && typeof input_value === "string";
|
|
const prefix = type === "number" || type === "range" ? "n:" : type === "checkbox" && !is_array ? "b:" : "";
|
|
const base_props = {
|
|
name: prefix + key + (is_array ? "[]" : ""),
|
|
get "aria-invalid"() {
|
|
const issues = get_issues();
|
|
return key in issues ? "true" : void 0;
|
|
}
|
|
};
|
|
if (type !== "text" && type !== "select" && type !== "select multiple") {
|
|
base_props.type = type === "file multiple" ? "file" : type;
|
|
}
|
|
if (type === "submit" || type === "hidden") {
|
|
return Object.defineProperties(base_props, {
|
|
value: { value: input_value, enumerable: true }
|
|
});
|
|
}
|
|
if (type === "select" || type === "select multiple") {
|
|
return Object.defineProperties(base_props, {
|
|
multiple: { value: is_array, enumerable: true },
|
|
value: {
|
|
enumerable: true,
|
|
get() {
|
|
return get_value();
|
|
}
|
|
}
|
|
});
|
|
}
|
|
if (type === "checkbox" || type === "radio") {
|
|
return Object.defineProperties(base_props, {
|
|
value: { value: input_value ?? "on", enumerable: true },
|
|
checked: {
|
|
enumerable: true,
|
|
get() {
|
|
const value = get_value();
|
|
if (type === "radio") {
|
|
return value === input_value;
|
|
}
|
|
if (is_array) {
|
|
return (value ?? []).includes(input_value);
|
|
}
|
|
return value;
|
|
}
|
|
}
|
|
});
|
|
}
|
|
if (type === "file" || type === "file multiple") {
|
|
return Object.defineProperties(base_props, {
|
|
multiple: { value: is_array, enumerable: true },
|
|
files: {
|
|
enumerable: true,
|
|
get() {
|
|
const value = get_value();
|
|
if (value instanceof File) {
|
|
if (typeof DataTransfer !== "undefined") {
|
|
const fileList = new DataTransfer();
|
|
fileList.items.add(value);
|
|
return fileList.files;
|
|
}
|
|
return { 0: value, length: 1 };
|
|
}
|
|
if (Array.isArray(value) && value.every((f) => f instanceof File)) {
|
|
if (typeof DataTransfer !== "undefined") {
|
|
const fileList = new DataTransfer();
|
|
value.forEach((file) => fileList.items.add(file));
|
|
return fileList.files;
|
|
}
|
|
const fileListLike = { length: value.length };
|
|
value.forEach((file, index) => {
|
|
fileListLike[index] = file;
|
|
});
|
|
return fileListLike;
|
|
}
|
|
return null;
|
|
}
|
|
}
|
|
});
|
|
}
|
|
return Object.defineProperties(base_props, {
|
|
value: {
|
|
enumerable: true,
|
|
get() {
|
|
const value = get_value();
|
|
return value != null ? String(value) : "";
|
|
}
|
|
}
|
|
});
|
|
};
|
|
return create_field_proxy(as_func, get_input, set_input, get_issues, [...path, "as"]);
|
|
}
|
|
return create_field_proxy({}, get_input, set_input, get_issues, [...path, prop]);
|
|
}
|
|
});
|
|
}
|
|
function build_path_string(path) {
|
|
let result = "";
|
|
for (const segment of path) {
|
|
if (typeof segment === "number") {
|
|
result += `[${segment}]`;
|
|
} else {
|
|
result += result === "" ? segment : "." + segment;
|
|
}
|
|
}
|
|
return result;
|
|
}
|
|
function negotiate(accept, types) {
|
|
const parts = [];
|
|
accept.split(",").forEach((str, i) => {
|
|
const match = /([^/ \t]+)\/([^; \t]+)[ \t]*(?:;[ \t]*q=([0-9.]+))?/.exec(str);
|
|
if (match) {
|
|
const [, type, subtype, q = "1"] = match;
|
|
parts.push({ type, subtype, q: +q, i });
|
|
}
|
|
});
|
|
parts.sort((a, b) => {
|
|
if (a.q !== b.q) {
|
|
return b.q - a.q;
|
|
}
|
|
if (a.subtype === "*" !== (b.subtype === "*")) {
|
|
return a.subtype === "*" ? 1 : -1;
|
|
}
|
|
if (a.type === "*" !== (b.type === "*")) {
|
|
return a.type === "*" ? 1 : -1;
|
|
}
|
|
return a.i - b.i;
|
|
});
|
|
let accepted;
|
|
let min_priority = Infinity;
|
|
for (const mimetype of types) {
|
|
const [type, subtype] = mimetype.split("/");
|
|
const priority = parts.findIndex(
|
|
(part) => (part.type === type || part.type === "*") && (part.subtype === subtype || part.subtype === "*")
|
|
);
|
|
if (priority !== -1 && priority < min_priority) {
|
|
accepted = mimetype;
|
|
min_priority = priority;
|
|
}
|
|
}
|
|
return accepted;
|
|
}
|
|
function is_content_type(request, ...types) {
|
|
const type = request.headers.get("content-type")?.split(";", 1)[0].trim() ?? "";
|
|
return types.includes(type.toLowerCase());
|
|
}
|
|
function is_form_content_type(request) {
|
|
return is_content_type(
|
|
request,
|
|
"application/x-www-form-urlencoded",
|
|
"multipart/form-data",
|
|
"text/plain",
|
|
BINARY_FORM_CONTENT_TYPE
|
|
);
|
|
}
|
|
function coalesce_to_error(err) {
|
|
return err instanceof Error || err && /** @type {any} */
|
|
err.name && /** @type {any} */
|
|
err.message ? (
|
|
/** @type {Error} */
|
|
err
|
|
) : new Error(JSON.stringify(err));
|
|
}
|
|
function normalize_error(error) {
|
|
return (
|
|
/** @type {import('../exports/internal/index.js').Redirect | HttpError | SvelteKitError | Error} */
|
|
error
|
|
);
|
|
}
|
|
function get_status(error) {
|
|
return error instanceof HttpError || error instanceof SvelteKitError ? error.status : 500;
|
|
}
|
|
function get_message(error) {
|
|
return error instanceof SvelteKitError ? error.text : "Internal Error";
|
|
}
|
|
const escape_html_attr_dict = {
|
|
"&": "&",
|
|
'"': """
|
|
// Svelte also escapes < because the escape function could be called inside a `noscript` there
|
|
// https://github.com/sveltejs/svelte/security/advisories/GHSA-8266-84wp-wv5c
|
|
// However, that doesn't apply in SvelteKit
|
|
};
|
|
const escape_html_dict = {
|
|
"&": "&",
|
|
"<": "<"
|
|
};
|
|
const surrogates = (
|
|
// high surrogate without paired low surrogate
|
|
"[\\ud800-\\udbff](?![\\udc00-\\udfff])|[\\ud800-\\udbff][\\udc00-\\udfff]|[\\udc00-\\udfff]"
|
|
);
|
|
const escape_html_attr_regex = new RegExp(
|
|
`[${Object.keys(escape_html_attr_dict).join("")}]|` + surrogates,
|
|
"g"
|
|
);
|
|
const escape_html_regex = new RegExp(
|
|
`[${Object.keys(escape_html_dict).join("")}]|` + surrogates,
|
|
"g"
|
|
);
|
|
function escape_html(str, is_attr) {
|
|
const dict = is_attr ? escape_html_attr_dict : escape_html_dict;
|
|
const escaped_str = str.replace(is_attr ? escape_html_attr_regex : escape_html_regex, (match) => {
|
|
if (match.length === 2) {
|
|
return match;
|
|
}
|
|
return dict[match] ?? `&#${match.charCodeAt(0)};`;
|
|
});
|
|
return escaped_str;
|
|
}
|
|
function method_not_allowed(mod, method) {
|
|
return text(`${method} method not allowed`, {
|
|
status: 405,
|
|
headers: {
|
|
// https://developer.mozilla.org/en-US/docs/Web/HTTP/Status/405
|
|
// "The server must generate an Allow header field in a 405 status code response"
|
|
allow: allowed_methods(mod).join(", ")
|
|
}
|
|
});
|
|
}
|
|
function allowed_methods(mod) {
|
|
const allowed = ENDPOINT_METHODS.filter((method) => method in mod);
|
|
if ("GET" in mod && !("HEAD" in mod)) {
|
|
allowed.push("HEAD");
|
|
}
|
|
return allowed;
|
|
}
|
|
function get_global_name(options) {
|
|
return `__sveltekit_${options.version_hash}`;
|
|
}
|
|
function static_error_page(options, status, message) {
|
|
let page = options.templates.error({ status, message: escape_html(message) });
|
|
return text(page, {
|
|
headers: { "content-type": "text/html; charset=utf-8" },
|
|
status
|
|
});
|
|
}
|
|
async function handle_fatal_error(event, state, options, error) {
|
|
error = error instanceof HttpError ? error : coalesce_to_error(error);
|
|
const status = get_status(error);
|
|
const body = await handle_error_and_jsonify(event, state, options, error);
|
|
const type = negotiate(event.request.headers.get("accept") || "text/html", [
|
|
"application/json",
|
|
"text/html"
|
|
]);
|
|
if (event.isDataRequest || type === "application/json") {
|
|
return json(body, {
|
|
status
|
|
});
|
|
}
|
|
return static_error_page(options, status, body.message);
|
|
}
|
|
async function handle_error_and_jsonify(event, state, options, error) {
|
|
if (error instanceof HttpError) {
|
|
return { message: "Unknown Error", ...error.body };
|
|
}
|
|
const status = get_status(error);
|
|
const message = get_message(error);
|
|
return await with_request_store(
|
|
{ event, state },
|
|
() => options.hooks.handleError({ error, event, status, message })
|
|
) ?? { message };
|
|
}
|
|
function redirect_response(status, location) {
|
|
const response = new Response(void 0, {
|
|
status,
|
|
headers: { location }
|
|
});
|
|
return response;
|
|
}
|
|
function clarify_devalue_error(event, error) {
|
|
if (error.path) {
|
|
return `Data returned from \`load\` while rendering ${event.route.id} is not serializable: ${error.message} (${error.path}). If you need to serialize/deserialize custom types, use transport hooks: https://svelte.dev/docs/kit/hooks#Universal-hooks-transport.`;
|
|
}
|
|
if (error.path === "") {
|
|
return `Data returned from \`load\` while rendering ${event.route.id} is not a plain object`;
|
|
}
|
|
return error.message;
|
|
}
|
|
function serialize_uses(node) {
|
|
const uses = {};
|
|
if (node.uses && node.uses.dependencies.size > 0) {
|
|
uses.dependencies = Array.from(node.uses.dependencies);
|
|
}
|
|
if (node.uses && node.uses.search_params.size > 0) {
|
|
uses.search_params = Array.from(node.uses.search_params);
|
|
}
|
|
if (node.uses && node.uses.params.size > 0) {
|
|
uses.params = Array.from(node.uses.params);
|
|
}
|
|
if (node.uses?.parent) uses.parent = 1;
|
|
if (node.uses?.route) uses.route = 1;
|
|
if (node.uses?.url) uses.url = 1;
|
|
return uses;
|
|
}
|
|
function has_prerendered_path(manifest, pathname) {
|
|
return manifest._.prerendered_routes.has(pathname) || pathname.at(-1) === "/" && manifest._.prerendered_routes.has(pathname.slice(0, -1));
|
|
}
|
|
function format_server_error(status, error, event) {
|
|
const formatted_text = `
|
|
\x1B[1;31m[${status}] ${event.request.method} ${event.url.pathname}\x1B[0m`;
|
|
if (status === 404) {
|
|
return formatted_text;
|
|
}
|
|
return `${formatted_text}
|
|
${error.stack}`;
|
|
}
|
|
function get_node_type(node_id) {
|
|
const parts = node_id?.split("/");
|
|
const filename = parts?.at(-1);
|
|
if (!filename) return "unknown";
|
|
const dot_parts = filename.split(".");
|
|
return dot_parts.slice(0, -1).join(".");
|
|
}
|
|
const INVALIDATED_PARAM = "x-sveltekit-invalidated";
|
|
const TRAILING_SLASH_PARAM = "x-sveltekit-trailing-slash";
|
|
function stringify(data, transport) {
|
|
const encoders = Object.fromEntries(Object.entries(transport).map(([k, v]) => [k, v.encode]));
|
|
return devalue.stringify(data, encoders);
|
|
}
|
|
function stringify_remote_arg(value, transport) {
|
|
if (value === void 0) return "";
|
|
const json_string = stringify(value, transport);
|
|
const bytes = new TextEncoder().encode(json_string);
|
|
return base64_encode(bytes).replaceAll("=", "").replaceAll("+", "-").replaceAll("/", "_");
|
|
}
|
|
function parse_remote_arg(string, transport) {
|
|
if (!string) return void 0;
|
|
const json_string = text_decoder.decode(
|
|
// no need to add back `=` characters, atob can handle it
|
|
base64_decode(string.replaceAll("-", "+").replaceAll("_", "/"))
|
|
);
|
|
const decoders = Object.fromEntries(Object.entries(transport).map(([k, v]) => [k, v.decode]));
|
|
return devalue.parse(json_string, decoders);
|
|
}
|
|
function create_remote_key(id, payload) {
|
|
return id + "/" + payload;
|
|
}
|
|
export {
|
|
ENDPOINT_METHODS as E,
|
|
INVALIDATED_PARAM as I,
|
|
PAGE_METHODS as P,
|
|
SVELTE_KIT_ASSETS as S,
|
|
TRAILING_SLASH_PARAM as T,
|
|
normalize_error as a,
|
|
get_global_name as b,
|
|
clarify_devalue_error as c,
|
|
get_node_type as d,
|
|
escape_html as e,
|
|
create_remote_key as f,
|
|
get_status as g,
|
|
handle_error_and_jsonify as h,
|
|
is_form_content_type as i,
|
|
static_error_page as j,
|
|
stringify as k,
|
|
deserialize_binary_form as l,
|
|
method_not_allowed as m,
|
|
negotiate as n,
|
|
has_prerendered_path as o,
|
|
parse_remote_arg as p,
|
|
handle_fatal_error as q,
|
|
redirect_response as r,
|
|
serialize_uses as s,
|
|
format_server_error as t,
|
|
stringify_remote_arg as u,
|
|
flatten_issues as v,
|
|
create_field_proxy as w,
|
|
normalize_issue as x,
|
|
set_nested_value as y,
|
|
deep_set as z
|
|
};
|