mirror of
https://github.com/MercuryWorkshop/scramjet.git
synced 2025-05-14 15:00:01 -04:00
refactor: split swfetch function
This commit is contained in:
parent
41acba634d
commit
bd08461dbe
1 changed files with 139 additions and 125 deletions
|
@ -4,15 +4,17 @@ import { ParseResultType } from "parse-domain";
|
||||||
import { ScramjetServiceWorker } from ".";
|
import { ScramjetServiceWorker } from ".";
|
||||||
import { renderError } from "./error";
|
import { renderError } from "./error";
|
||||||
|
|
||||||
|
const { encodeUrl, decodeUrl } = self.$scramjet.shared.url;
|
||||||
|
const { rewriteHeaders, rewriteHtml, rewriteJs, rewriteCss, rewriteWorkers } =
|
||||||
|
self.$scramjet.shared.rewrite;
|
||||||
|
const { parseDomain } = self.$scramjet.shared.util;
|
||||||
|
|
||||||
export async function swfetch(
|
export async function swfetch(
|
||||||
this: ScramjetServiceWorker,
|
this: ScramjetServiceWorker,
|
||||||
{ request }: FetchEvent
|
{ request }: FetchEvent
|
||||||
) {
|
) {
|
||||||
const urlParam = new URLSearchParams(new URL(request.url).search);
|
const urlParam = new URLSearchParams(new URL(request.url).search);
|
||||||
const { encodeUrl, decodeUrl } = self.$scramjet.shared.url;
|
|
||||||
const { rewriteHeaders, rewriteHtml, rewriteJs, rewriteCss, rewriteWorkers } =
|
|
||||||
self.$scramjet.shared.rewrite;
|
|
||||||
const { parseDomain } = self.$scramjet.shared.util;
|
|
||||||
|
|
||||||
if (urlParam.has("url")) {
|
if (urlParam.has("url")) {
|
||||||
return Response.redirect(
|
return Response.redirect(
|
||||||
|
@ -23,11 +25,6 @@ export async function swfetch(
|
||||||
try {
|
try {
|
||||||
const url = new URL(decodeUrl(request.url));
|
const url = new URL(decodeUrl(request.url));
|
||||||
|
|
||||||
const cookieStore = new IDBMap(url.host, {
|
|
||||||
durability: "relaxed",
|
|
||||||
prefix: "Cookies",
|
|
||||||
});
|
|
||||||
|
|
||||||
const response: BareResponseFetch = await this.client.fetch(url, {
|
const response: BareResponseFetch = await this.client.fetch(url, {
|
||||||
method: request.method,
|
method: request.method,
|
||||||
body: request.body,
|
body: request.body,
|
||||||
|
@ -40,10 +37,101 @@ export async function swfetch(
|
||||||
duplex: "half",
|
duplex: "half",
|
||||||
});
|
});
|
||||||
|
|
||||||
let responseBody;
|
|
||||||
|
return await handleResponse(url, request.destination, response);
|
||||||
|
} catch (err) {
|
||||||
|
console.error("ERROR FROM SERVICE WORKER FETCH", err);
|
||||||
|
if (!["document", "iframe"].includes(request.destination))
|
||||||
|
return new Response(undefined, { status: 500 });
|
||||||
|
|
||||||
|
return renderError(err, decodeUrl(request.url));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
async function handleResponse(url: URL, destination: RequestDestination, response: BareResponseFetch): Promise<Response> {
|
||||||
|
let responseBody: string | ArrayBuffer | ReadableStream;
|
||||||
const responseHeaders = rewriteHeaders(response.rawHeaders, url);
|
const responseHeaders = rewriteHeaders(response.rawHeaders, url);
|
||||||
|
|
||||||
for (const cookie of (responseHeaders["set-cookie"] || []) as string[]) {
|
|
||||||
|
await handleCookies(url, (responseHeaders["set-cookie"] || []) as string[])
|
||||||
|
|
||||||
|
for (const header in responseHeaders) {
|
||||||
|
// flatten everything past here
|
||||||
|
if (Array.isArray(responseHeaders[header]))
|
||||||
|
responseHeaders[header] = responseHeaders[header][0];
|
||||||
|
}
|
||||||
|
|
||||||
|
if (response.body) {
|
||||||
|
switch (destination) {
|
||||||
|
case "iframe":
|
||||||
|
case "document":
|
||||||
|
if (
|
||||||
|
responseHeaders["content-type"]?.toString()?.startsWith("text/html")
|
||||||
|
) {
|
||||||
|
responseBody = rewriteHtml(await response.text(), url);
|
||||||
|
} else {
|
||||||
|
responseBody = response.body;
|
||||||
|
}
|
||||||
|
break;
|
||||||
|
case "script":
|
||||||
|
responseBody = rewriteJs(await response.arrayBuffer(), url);
|
||||||
|
// Disable threading for now, it's causing issues.
|
||||||
|
// responseBody = await this.threadpool.rewriteJs(await responseBody.arrayBuffer(), url.toString());
|
||||||
|
break;
|
||||||
|
case "style":
|
||||||
|
responseBody = rewriteCss(await response.text(), url);
|
||||||
|
break;
|
||||||
|
case "sharedworker":
|
||||||
|
case "worker":
|
||||||
|
responseBody = rewriteWorkers(await response.text(), url);
|
||||||
|
break;
|
||||||
|
default:
|
||||||
|
responseBody = response.body;
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
// downloads
|
||||||
|
if (["document", "iframe"].includes(destination)) {
|
||||||
|
const header = responseHeaders["content-disposition"];
|
||||||
|
|
||||||
|
// validate header and test for filename
|
||||||
|
if (!/\s*?((inline|attachment);\s*?)filename=/i.test(header)) {
|
||||||
|
// if filename= wasn"t specified then maybe the remote specified to download this as an attachment?
|
||||||
|
// if it"s invalid then we can still possibly test for the attachment/inline type
|
||||||
|
const type = /^\s*?attachment/i.test(header) ? "attachment" : "inline";
|
||||||
|
|
||||||
|
// set the filename
|
||||||
|
const [filename] = new URL(response.finalURL).pathname
|
||||||
|
.split("/")
|
||||||
|
.slice(-1);
|
||||||
|
|
||||||
|
responseHeaders["content-disposition"] =
|
||||||
|
`${type}; filename=${JSON.stringify(filename)}`;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if (responseHeaders["accept"] === "text/event-stream") {
|
||||||
|
responseHeaders["content-type"] = "text/event-stream";
|
||||||
|
}
|
||||||
|
if (crossOriginIsolated) {
|
||||||
|
responseHeaders["Cross-Origin-Embedder-Policy"] = "require-corp";
|
||||||
|
}
|
||||||
|
|
||||||
|
return new Response(responseBody, {
|
||||||
|
headers: responseHeaders as HeadersInit,
|
||||||
|
status: response.status,
|
||||||
|
statusText: response.statusText,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
async function handleCookies(url: URL, headers: string[]) {
|
||||||
|
const cookieStore = new IDBMap(url.host, {
|
||||||
|
durability: "relaxed",
|
||||||
|
prefix: "Cookies",
|
||||||
|
});
|
||||||
|
|
||||||
|
for (const cookie of headers) {
|
||||||
let cookieParsed = cookie.split(";").map((x) => x.trim().split("="));
|
let cookieParsed = cookie.split(";").map((x) => x.trim().split("="));
|
||||||
|
|
||||||
let [key, value] = cookieParsed.shift();
|
let [key, value] = cookieParsed.shift();
|
||||||
|
@ -90,78 +178,4 @@ export async function swfetch(
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
for (const header in responseHeaders) {
|
|
||||||
// flatten everything past here
|
|
||||||
if (Array.isArray(responseHeaders[header]))
|
|
||||||
responseHeaders[header] = responseHeaders[header][0];
|
|
||||||
}
|
|
||||||
|
|
||||||
if (response.body) {
|
|
||||||
switch (request.destination) {
|
|
||||||
case "iframe":
|
|
||||||
case "document":
|
|
||||||
if (
|
|
||||||
responseHeaders["content-type"]?.toString()?.startsWith("text/html")
|
|
||||||
) {
|
|
||||||
responseBody = rewriteHtml(await response.text(), url);
|
|
||||||
} else {
|
|
||||||
responseBody = response.body;
|
|
||||||
}
|
|
||||||
break;
|
|
||||||
case "script":
|
|
||||||
responseBody = rewriteJs(await response.arrayBuffer(), url);
|
|
||||||
// Disable threading for now, it's causing issues.
|
|
||||||
// responseBody = await this.threadpool.rewriteJs(await responseBody.arrayBuffer(), url.toString());
|
|
||||||
break;
|
|
||||||
case "style":
|
|
||||||
responseBody = rewriteCss(await response.text(), url);
|
|
||||||
break;
|
|
||||||
case "sharedworker":
|
|
||||||
case "worker":
|
|
||||||
responseBody = rewriteWorkers(await response.text(), url);
|
|
||||||
break;
|
|
||||||
default:
|
|
||||||
responseBody = response.body;
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
// downloads
|
|
||||||
if (["document", "iframe"].includes(request.destination)) {
|
|
||||||
const header = responseHeaders["content-disposition"];
|
|
||||||
|
|
||||||
// validate header and test for filename
|
|
||||||
if (!/\s*?((inline|attachment);\s*?)filename=/i.test(header)) {
|
|
||||||
// if filename= wasn"t specified then maybe the remote specified to download this as an attachment?
|
|
||||||
// if it"s invalid then we can still possibly test for the attachment/inline type
|
|
||||||
const type = /^\s*?attachment/i.test(header) ? "attachment" : "inline";
|
|
||||||
|
|
||||||
// set the filename
|
|
||||||
const [filename] = new URL(response.finalURL).pathname
|
|
||||||
.split("/")
|
|
||||||
.slice(-1);
|
|
||||||
|
|
||||||
responseHeaders["content-disposition"] =
|
|
||||||
`${type}; filename=${JSON.stringify(filename)}`;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
if (responseHeaders["accept"] === "text/event-stream") {
|
|
||||||
responseHeaders["content-type"] = "text/event-stream";
|
|
||||||
}
|
|
||||||
if (crossOriginIsolated) {
|
|
||||||
responseHeaders["Cross-Origin-Embedder-Policy"] = "require-corp";
|
|
||||||
}
|
|
||||||
|
|
||||||
return new Response(responseBody, {
|
|
||||||
headers: responseHeaders as HeadersInit,
|
|
||||||
status: response.status,
|
|
||||||
statusText: response.statusText,
|
|
||||||
});
|
|
||||||
} catch (err) {
|
|
||||||
console.error("ERROR FROM SERVICE WORKER FETCH", err);
|
|
||||||
if (!["document", "iframe"].includes(request.destination))
|
|
||||||
return new Response(undefined, { status: 500 });
|
|
||||||
|
|
||||||
return renderError(err, decodeUrl(request.url));
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue