mirror of
https://github.com/therealaleph/MasterHttpRelayVPN-RUST.git
synced 2026-05-17 21:24:48 +03:00
f4257a5be8
Fix the v1.9.28 Code.gs JSON parse regression by keeping normal relay responses wrapped and using req.r only for redirect handling.\n\nTests:\n- node --check /tmp/Code-1265-fix.js\n- cargo test --lib
799 lines
28 KiB
JavaScript
799 lines
28 KiB
JavaScript
/**
|
||
* DomainFront Relay — Google Apps Script
|
||
*
|
||
* TWO modes:
|
||
* 1. Single: POST { k, m, u, h, b, ct, r } → { s, h, b }
|
||
* 2. Batch: POST { k, q: [{m,u,h,b,ct,r}, ...] } → { q: [{s,h,b}, ...] }
|
||
* Uses UrlFetchApp.fetchAll() — all URLs fetched IN PARALLEL.
|
||
*
|
||
* OPTIONAL SPREADSHEET-BACKED RESPONSE CACHE:
|
||
* Set CACHE_SPREADSHEET_ID to a valid Google Sheet ID (must be owned by
|
||
* the same account). When enabled, public GET requests are stored in the
|
||
* sheet and served from there on repeat visits, reducing UrlFetchApp
|
||
* quota consumption. Bodies are gzipped before base64 storage so larger
|
||
* responses fit under the per-cell character limit, and persistent
|
||
* 4xx (404/410/451) get a short negative-cache TTL so buggy clients
|
||
* that hammer dead URLs cost zero quota; 5xx is never cached so a
|
||
* flapping upstream cannot poison a 24h slot with a transient outage.
|
||
* The cache is Vary-aware (Accept-Encoding and Accept-Language are
|
||
* hashed into the compound cache key). Leave CACHE_SPREADSHEET_ID as-is
|
||
* to disable caching entirely — zero overhead.
|
||
*
|
||
* DEPLOYMENT:
|
||
* 1. Go to https://script.google.com → New project
|
||
* 2. Delete the default code, paste THIS entire file
|
||
* 3. Change AUTH_KEY below to your own secret
|
||
* 4. (Optional) Set CACHE_SPREADSHEET_ID to enable caching
|
||
* 5. Click Deploy → New deployment
|
||
* 6. Type: Web app | Execute as: Me | Who has access: Anyone
|
||
* 7. Copy the Deployment ID into config.json as "script_id"
|
||
*
|
||
* CHANGE THE AUTH KEY BELOW TO YOUR OWN SECRET!
|
||
*/
|
||
|
||
const AUTH_KEY = "CHANGE_ME_TO_A_STRONG_SECRET";
|
||
|
||
// Active-probing defense. When false (production default), bad AUTH_KEY
|
||
// requests get a decoy HTML page that looks like a placeholder Apps
|
||
// Script web app instead of the JSON `{"e":"unauthorized"}` body. This
|
||
// makes the deployment indistinguishable from a forgotten-but-public
|
||
// Apps Script project to active scanners that POST malformed payloads
|
||
// looking for proxy endpoints.
|
||
//
|
||
// Set to `true` during initial setup if a misconfigured client is
|
||
// hitting "unauthorized" and you want the explicit JSON error to debug
|
||
// — then flip back to false before the deployment is widely shared.
|
||
// (Inspired by #365 Section 3, mhrv-rs v1.8.0+.)
|
||
const DIAGNOSTIC_MODE = false;
|
||
|
||
// ── Optional Spreadsheet Cache ──────────────────────────────
|
||
// Set to a valid Spreadsheet ID to enable response caching.
|
||
// Leave as-is to disable caching entirely (zero overhead).
|
||
const CACHE_SPREADSHEET_ID = "CHANGE_ME_TO_CACHE_SPREADSHEET_ID";
|
||
const CACHE_SHEET_NAME = "RelayCache";
|
||
const CACHE_META_SHEET_NAME = "RelayMeta";
|
||
const CACHE_META_CURSOR_CELL = "A1";
|
||
|
||
// ── Cache Tuning ────────────────────────────────────────────
|
||
const CACHE_MAX_ROWS = 5000; // circular buffer capacity
|
||
const CACHE_MAX_BODY_BYTES = 35000; // skip responses larger than ~35 KB
|
||
const CACHE_DEFAULT_TTL_SECONDS = 86400; // 24-hour fallback when no Cache-Control
|
||
|
||
// ── Negative Caching ────────────────────────────────────────
|
||
// Persistent 4xx errors get a short TTL when the upstream is silent on
|
||
// Cache-Control. Buggy clients hammer dead URLs (favicons, telemetry
|
||
// pixels, dev-tools probes); a 5-minute floor absorbs the storm at
|
||
// zero quota cost while letting transient 404s self-heal quickly.
|
||
// 5xx is never cached — see _fetchAndCache.
|
||
const NEGATIVE_CACHE_STATUSES = { 404: 1, 410: 1, 451: 1 };
|
||
const NEGATIVE_CACHE_TTL_SECONDS = 300;
|
||
|
||
// ── Body Compression ────────────────────────────────────────
|
||
// Bodies are gzipped before base64 storage when worthwhile. Gzip has
|
||
// ~20 bytes of header overhead, so very small payloads can bloat;
|
||
// skip below this threshold. Already-encoded responses (gzip/br/etc.)
|
||
// are stored as-is to avoid double-compression.
|
||
const GZIP_MIN_BYTES = 256;
|
||
|
||
// ── Vary-Aware Cache Key ────────────────────────────────────
|
||
// These request headers are hashed into the compound cache key
|
||
// alongside the URL so that responses with different encodings
|
||
// or languages never collide in the cache. Covers ~95 % of
|
||
// real-world Vary usage without inspecting the response.
|
||
const VARY_KEY_HEADERS = ["accept-encoding", "accept-language"];
|
||
|
||
// Connection-level + IP-leak request headers we strip before forwarding
|
||
// to the destination. Browser capability headers (sec-ch-ua*, sec-fetch-*)
|
||
// stay intact — modern apps like Google Meet use them for browser gating.
|
||
// We also drop the `X-Forwarded-*` / `Forwarded` / `Via` family so a
|
||
// misconfigured upstream proxy on the user side can't leak the user's
|
||
// real IP through the relay path. Mirrors upstream
|
||
// `masterking32/MasterHttpRelayVPN@3094288`.
|
||
const SKIP_HEADERS = {
|
||
host: 1, connection: 1, "content-length": 1,
|
||
"transfer-encoding": 1, "proxy-connection": 1, "proxy-authorization": 1,
|
||
"priority": 1, te: 1,
|
||
"x-forwarded-for": 1, "x-forwarded-host": 1, "x-forwarded-proto": 1,
|
||
"x-forwarded-port": 1, "x-real-ip": 1, "forwarded": 1, "via": 1,
|
||
};
|
||
|
||
// Methods we consider safe to replay if `UrlFetchApp.fetchAll()` raises.
|
||
// GET/HEAD/OPTIONS are idempotent per RFC 9110; POST/PUT/PATCH/DELETE
|
||
// can have side-effects so we surface the error instead of silently
|
||
// re-firing them.
|
||
const SAFE_REPLAY_METHODS = { GET: 1, HEAD: 1, OPTIONS: 1 };
|
||
|
||
// Headers that disqualify a request from the cache path.
|
||
const CACHE_BUSTING_HEADERS = {
|
||
authorization: 1, cookie: 1, "x-api-key": 1,
|
||
"proxy-authorization": 1, "set-cookie": 1,
|
||
};
|
||
|
||
// HTML body for the bad-auth decoy. Mimics a minimal Apps Script-style
|
||
// placeholder page — no proxy-shaped JSON, nothing distinctive enough
|
||
// for a probe to fingerprint as a tunnel endpoint.
|
||
const DECOY_HTML =
|
||
'<!DOCTYPE html><html><head><title>Web App</title></head>' +
|
||
'<body><p>The script completed but did not return anything.</p>' +
|
||
'</body></html>';
|
||
|
||
// ── Request Handlers ────────────────────────────────────────
|
||
|
||
function _decoyOrError(jsonBody) {
|
||
if (DIAGNOSTIC_MODE) return _json(jsonBody);
|
||
return ContentService
|
||
.createTextOutput(DECOY_HTML)
|
||
.setMimeType(ContentService.MimeType.HTML);
|
||
}
|
||
|
||
function doPost(e) {
|
||
try {
|
||
var req = JSON.parse(e.postData.contents);
|
||
if (req.k !== AUTH_KEY) return _decoyOrError({ e: "unauthorized" });
|
||
|
||
// Batch mode: { k, q: [...] }
|
||
if (Array.isArray(req.q)) return _doBatch(req.q);
|
||
|
||
// Single mode
|
||
return _doSingle(req);
|
||
} catch (err) {
|
||
// Parse failures of the request body are also probe-shaped — a real
|
||
// mhrv-rs client never sends invalid JSON. Decoy for the same reason.
|
||
return _decoyOrError({ e: String(err) });
|
||
}
|
||
}
|
||
|
||
// `doGet` is what active scanners hit first (HTTP GET probes are cheaper
|
||
// than POSTs). Apps Script defaults to a "Script function not found" page
|
||
// here which is a fine-enough decoy on its own, but explicitly returning
|
||
// the same harmless placeholder makes the response identical to the
|
||
// bad-auth POST decoy — one less fingerprint vector.
|
||
function doGet(e) {
|
||
return ContentService
|
||
.createTextOutput(DECOY_HTML)
|
||
.setMimeType(ContentService.MimeType.HTML);
|
||
}
|
||
|
||
// ── Single Request ─────────────────────────────────────────
|
||
|
||
function _doSingle(req) {
|
||
if (!req.u || typeof req.u !== "string" || !req.u.match(/^https?:\/\//i)) {
|
||
return _json({ e: "bad url" });
|
||
}
|
||
|
||
// ── Optional cache path ────────────────────────────────
|
||
// Only entered when CACHE_SPREADSHEET_ID is configured and
|
||
// the request qualifies as a public, cachable GET.
|
||
if (_canUseCache(req)) {
|
||
var cached = _getFromCache(req.u, req.h);
|
||
if (cached) {
|
||
return _json({
|
||
s: cached.status,
|
||
h: JSON.parse(cached.headers),
|
||
b: cached.body,
|
||
cached: true,
|
||
});
|
||
}
|
||
|
||
var fetchResult = _fetchAndCache(req.u, req.h);
|
||
if (fetchResult) {
|
||
return _json({
|
||
s: fetchResult.status,
|
||
h: JSON.parse(fetchResult.headers),
|
||
b: fetchResult.body,
|
||
cached: false,
|
||
});
|
||
}
|
||
// If _fetchAndCache returns null (spreadsheet unavailable),
|
||
// fall through to the normal relay path below.
|
||
}
|
||
|
||
// ── Normal relay (cache disabled or unavailable) ────────
|
||
// Wrap the fetch + body encode in try/catch so any failure surfaces as
|
||
// a JSON error envelope the Rust client can parse. Without this, throws
|
||
// from UrlFetchApp.fetch (URL too long, payload too large, quota
|
||
// exhausted, 6-minute execution timeout) or from base64Encode (response
|
||
// body near Apps Script's ~50 MB ceiling can blow the V8 heap during
|
||
// encode) propagate unhandled, and Apps Script serves its default
|
||
// `<title>Web App</title>` HTML error page — which the client then
|
||
// reports as "Relay failed: bad response: no json in: <title>Web App>..."
|
||
// and the user has no signal as to the actual cause. Mirrors the
|
||
// per-item try/catch in _doBatch below.
|
||
try {
|
||
var opts = _buildOpts(req);
|
||
var resp = UrlFetchApp.fetch(req.u, opts);
|
||
|
||
return _json({
|
||
s: resp.getResponseCode(),
|
||
h: _respHeaders(resp),
|
||
b: Utilities.base64Encode(resp.getContent()),
|
||
});
|
||
} catch (err) {
|
||
return _json({ e: "fetch failed: " + String(err) });
|
||
}
|
||
}
|
||
|
||
// ── Batch Request ──────────────────────────────────────────
|
||
|
||
function _doBatch(items) {
|
||
var fetchArgs = [];
|
||
var fetchIndex = [];
|
||
var fetchMethods = [];
|
||
var errorMap = {};
|
||
|
||
for (var i = 0; i < items.length; i++) {
|
||
var item = items[i];
|
||
if (!item || typeof item !== "object") {
|
||
errorMap[i] = "bad item";
|
||
continue;
|
||
}
|
||
if (!item.u || typeof item.u !== "string" || !item.u.match(/^https?:\/\//i)) {
|
||
errorMap[i] = "bad url";
|
||
continue;
|
||
}
|
||
try {
|
||
var opts = _buildOpts(item);
|
||
opts.url = item.u;
|
||
fetchArgs.push(opts);
|
||
fetchIndex.push(i);
|
||
fetchMethods.push(String(item.m || "GET").toUpperCase());
|
||
} catch (buildErr) {
|
||
errorMap[i] = String(buildErr);
|
||
}
|
||
}
|
||
|
||
// fetchAll() processes all requests in parallel inside Google. If it
|
||
// throws as a whole (e.g. one URL violates UrlFetchApp limits and
|
||
// poisons the whole batch), degrade to per-item fetch on safe methods
|
||
// so a single bad request does not zero out every response in the
|
||
// batch. Mirrors upstream `masterking32/MasterHttpRelayVPN@3094288`.
|
||
var responses = [];
|
||
if (fetchArgs.length > 0) {
|
||
try {
|
||
responses = UrlFetchApp.fetchAll(fetchArgs);
|
||
} catch (fetchAllErr) {
|
||
responses = [];
|
||
for (var j = 0; j < fetchArgs.length; j++) {
|
||
try {
|
||
if (!SAFE_REPLAY_METHODS[fetchMethods[j]]) {
|
||
errorMap[fetchIndex[j]] =
|
||
"batch fetchAll failed; unsafe method not replayed";
|
||
responses[j] = null;
|
||
continue;
|
||
}
|
||
var fallbackReq = fetchArgs[j];
|
||
var fallbackUrl = fallbackReq.url;
|
||
var fallbackOpts = {};
|
||
for (var key in fallbackReq) {
|
||
if (
|
||
Object.prototype.hasOwnProperty.call(fallbackReq, key) &&
|
||
key !== "url"
|
||
) {
|
||
fallbackOpts[key] = fallbackReq[key];
|
||
}
|
||
}
|
||
responses[j] = UrlFetchApp.fetch(fallbackUrl, fallbackOpts);
|
||
} catch (singleErr) {
|
||
errorMap[fetchIndex[j]] = String(singleErr);
|
||
responses[j] = null;
|
||
}
|
||
}
|
||
}
|
||
}
|
||
|
||
var results = [];
|
||
var rIdx = 0;
|
||
for (var i = 0; i < items.length; i++) {
|
||
if (Object.prototype.hasOwnProperty.call(errorMap, i)) {
|
||
results.push({ e: errorMap[i] });
|
||
} else {
|
||
var resp = responses[rIdx++];
|
||
if (!resp) {
|
||
results.push({ e: "fetch failed" });
|
||
} else {
|
||
results.push({
|
||
s: resp.getResponseCode(),
|
||
h: _respHeaders(resp),
|
||
b: Utilities.base64Encode(resp.getContent()),
|
||
});
|
||
}
|
||
}
|
||
}
|
||
return _json({ q: results });
|
||
}
|
||
|
||
// ── Request Building ───────────────────────────────────────
|
||
|
||
function _buildOpts(req) {
|
||
var opts = {
|
||
method: (req.m || "GET").toLowerCase(),
|
||
muteHttpExceptions: true,
|
||
followRedirects: req.r !== false,
|
||
validateHttpsCertificates: true,
|
||
escaping: false,
|
||
};
|
||
if (req.h && typeof req.h === "object") {
|
||
var headers = {};
|
||
for (var k in req.h) {
|
||
if (req.h.hasOwnProperty(k) && !SKIP_HEADERS[k.toLowerCase()]) {
|
||
headers[k] = req.h[k];
|
||
}
|
||
}
|
||
opts.headers = headers;
|
||
}
|
||
if (req.b) {
|
||
opts.payload = Utilities.base64Decode(req.b);
|
||
if (req.ct) opts.contentType = req.ct;
|
||
}
|
||
return opts;
|
||
}
|
||
|
||
function _respHeaders(resp) {
|
||
try {
|
||
if (typeof resp.getAllHeaders === "function") {
|
||
return resp.getAllHeaders();
|
||
}
|
||
} catch (err) {}
|
||
return resp.getHeaders();
|
||
}
|
||
|
||
function _json(obj) {
|
||
return ContentService.createTextOutput(JSON.stringify(obj)).setMimeType(
|
||
ContentService.MimeType.JSON
|
||
);
|
||
}
|
||
|
||
// ═══════════════════════════════════════════════════════════
|
||
// SPREADSHEET CACHE — SHEET MANAGEMENT
|
||
// ═══════════════════════════════════════════════════════════
|
||
|
||
function _initCacheSheet() {
|
||
if (CACHE_SPREADSHEET_ID === "CHANGE_ME_TO_CACHE_SPREADSHEET_ID") {
|
||
return null;
|
||
}
|
||
try {
|
||
var ss = SpreadsheetApp.openById(CACHE_SPREADSHEET_ID);
|
||
var sheet = ss.getSheetByName(CACHE_SHEET_NAME);
|
||
if (!sheet) {
|
||
sheet = ss.insertSheet(CACHE_SHEET_NAME);
|
||
// Schema: URL_Hash | URL | Status | Headers | Body | Timestamp | Expires_At | Z
|
||
// Z is 1 when Body is base64(gzip(rawBytes)), 0/empty when base64(rawBytes).
|
||
// Legacy 7-column rows from older deployments read back as Z=undefined,
|
||
// which the cache hit path treats as "not gzipped" — fully compatible.
|
||
sheet.getRange(1, 1, 1, 8).setValues([[
|
||
"URL_Hash", "URL", "Status", "Headers", "Body", "Timestamp", "Expires_At", "Z"
|
||
]]);
|
||
}
|
||
return sheet;
|
||
} catch (e) {
|
||
return null;
|
||
}
|
||
}
|
||
|
||
function _getMetaSheet() {
|
||
if (CACHE_SPREADSHEET_ID === "CHANGE_ME_TO_CACHE_SPREADSHEET_ID") {
|
||
return null;
|
||
}
|
||
try {
|
||
var ss = SpreadsheetApp.openById(CACHE_SPREADSHEET_ID);
|
||
var sheet = ss.getSheetByName(CACHE_META_SHEET_NAME);
|
||
if (!sheet) {
|
||
sheet = ss.insertSheet(CACHE_META_SHEET_NAME);
|
||
sheet.getRange(CACHE_META_CURSOR_CELL).setValue(2);
|
||
sheet.hideSheet();
|
||
}
|
||
return sheet;
|
||
} catch (e) {
|
||
return null;
|
||
}
|
||
}
|
||
|
||
function _getNextCursor(sheet, metaSheet) {
|
||
var cursorRange = metaSheet.getRange(CACHE_META_CURSOR_CELL);
|
||
var cursor = cursorRange.getValue();
|
||
if (typeof cursor !== "number" || cursor < 2) cursor = 2;
|
||
|
||
var totalRows = sheet.getDataRange().getNumRows();
|
||
|
||
if (totalRows < CACHE_MAX_ROWS + 1) {
|
||
return totalRows + 1;
|
||
}
|
||
|
||
return cursor;
|
||
}
|
||
|
||
function _advanceCursor(metaSheet, currentRow) {
|
||
var nextRow = currentRow + 1;
|
||
if (nextRow > CACHE_MAX_ROWS + 1) nextRow = 2;
|
||
metaSheet.getRange(CACHE_META_CURSOR_CELL).setValue(nextRow);
|
||
}
|
||
|
||
function _ensureRowsAllocated(sheet) {
|
||
var totalRows = sheet.getDataRange().getNumRows();
|
||
if (totalRows < CACHE_MAX_ROWS + 1) {
|
||
var needed = CACHE_MAX_ROWS + 1 - totalRows;
|
||
sheet.insertRowsAfter(totalRows, needed);
|
||
}
|
||
}
|
||
|
||
// ═══════════════════════════════════════════════════════════
|
||
// SPREADSHEET CACHE — VARY-AWARE COMPOUND KEY
|
||
// ═══════════════════════════════════════════════════════════
|
||
|
||
/**
|
||
* Case-insensitive header lookup.
|
||
* HTTP header names are case-insensitive per RFC 7230 § 3.2.
|
||
*/
|
||
function _getHeaderCaseInsensitive(headers, targetKey) {
|
||
var target = targetKey.toLowerCase();
|
||
for (var k in headers) {
|
||
if (headers.hasOwnProperty(k) && k.toLowerCase() === target) {
|
||
return headers[k];
|
||
}
|
||
}
|
||
return null;
|
||
}
|
||
|
||
/**
|
||
* Compute a compound cache key:
|
||
* MD5(URL | header1:value1 | header2:value2 | ...)
|
||
*
|
||
* Instead of reading the response Vary header (which would require
|
||
* fetching first — circular), we preemptively include the request
|
||
* headers that are known to cause response variation. This handles
|
||
* Vary: Accept-Encoding and Vary: Accept-Language without ever
|
||
* inspecting the response.
|
||
*
|
||
* Values are lowercased and whitespace-stripped so semantically
|
||
* identical requests from different clients produce the same hash.
|
||
* Missing and empty headers both map to "<none>" (same semantic).
|
||
*/
|
||
function _getCacheKey(url, reqHeaders) {
|
||
var parts = [url];
|
||
|
||
if (reqHeaders && typeof reqHeaders === "object") {
|
||
for (var i = 0; i < VARY_KEY_HEADERS.length; i++) {
|
||
var headerName = VARY_KEY_HEADERS[i];
|
||
var rawValue = _getHeaderCaseInsensitive(reqHeaders, headerName);
|
||
|
||
if (rawValue && String(rawValue).trim() !== "") {
|
||
parts.push(headerName + ":" + rawValue.toLowerCase().replace(/\s/g, ""));
|
||
} else {
|
||
parts.push(headerName + ":<none>");
|
||
}
|
||
}
|
||
} else {
|
||
for (var j = 0; j < VARY_KEY_HEADERS.length; j++) {
|
||
parts.push(VARY_KEY_HEADERS[j] + ":<none>");
|
||
}
|
||
}
|
||
|
||
var compoundKey = parts.join("|");
|
||
return _md5Hex(compoundKey);
|
||
}
|
||
|
||
function _md5Hex(input) {
|
||
var rawHash = Utilities.computeDigest(Utilities.DigestAlgorithm.MD5, input);
|
||
return rawHash
|
||
.map(function (byte) {
|
||
var v = (byte < 0) ? 256 + byte : byte;
|
||
return ("0" + v.toString(16)).slice(-2);
|
||
})
|
||
.join("");
|
||
}
|
||
|
||
// ═══════════════════════════════════════════════════════════
|
||
// SPREADSHEET CACHE — CORE LOGIC
|
||
// ═══════════════════════════════════════════════════════════
|
||
|
||
/**
|
||
* Returns true if the request is eligible for the cache path:
|
||
* public GET, no body, no auth/cookie headers, cache configured.
|
||
*/
|
||
function _canUseCache(req) {
|
||
if ((req.m || "GET") !== "GET") return false;
|
||
if (req.b) return false;
|
||
if (!req.u || !req.u.match(/^https?:\/\//i)) return false;
|
||
if (CACHE_SPREADSHEET_ID === "CHANGE_ME_TO_CACHE_SPREADSHEET_ID") return false;
|
||
|
||
if (req.h && typeof req.h === "object") {
|
||
for (var k in req.h) {
|
||
if (req.h.hasOwnProperty(k) && CACHE_BUSTING_HEADERS[k.toLowerCase()]) {
|
||
return false;
|
||
}
|
||
}
|
||
}
|
||
|
||
return true;
|
||
}
|
||
|
||
/**
|
||
* Extract max-age (seconds) from a Cache-Control header value.
|
||
* Returns 0 if the directive forbids caching (no-cache / no-store /
|
||
* private). Falls back to CACHE_DEFAULT_TTL_SECONDS when no header
|
||
* is present. Clamped to [60, 2592000] (1 min – 30 days).
|
||
*/
|
||
function _parseMaxAge(cacheControlHeader) {
|
||
if (!cacheControlHeader) return CACHE_DEFAULT_TTL_SECONDS;
|
||
|
||
var lower = cacheControlHeader.toLowerCase();
|
||
|
||
if (
|
||
lower.indexOf("no-cache") !== -1 ||
|
||
lower.indexOf("no-store") !== -1 ||
|
||
lower.indexOf("private") !== -1
|
||
) {
|
||
return 0;
|
||
}
|
||
|
||
var match = lower.match(/max-age=(\d+)/);
|
||
if (match) {
|
||
var ttl = parseInt(match[1], 10);
|
||
return Math.max(60, Math.min(ttl, 2592000));
|
||
}
|
||
|
||
return CACHE_DEFAULT_TTL_SECONDS;
|
||
}
|
||
|
||
/**
|
||
* Rewrite time-sensitive headers so the client sees accurate
|
||
* Date, Age, and Cache-Control values reflecting cache age.
|
||
*/
|
||
function _refreshCachedHeaders(headersJson, timestamp) {
|
||
var headers = JSON.parse(headersJson);
|
||
var cachedAt = new Date(timestamp);
|
||
var now = new Date();
|
||
var ageSeconds = Math.floor((now.getTime() - cachedAt.getTime()) / 1000);
|
||
|
||
if (ageSeconds < 0) ageSeconds = 0;
|
||
|
||
headers["Date"] = now.toUTCString();
|
||
headers["Age"] = String(ageSeconds);
|
||
|
||
var originalCc = headers["Cache-Control"] || headers["cache-control"];
|
||
if (originalCc) {
|
||
headers["X-Original-Cache-Control"] = originalCc;
|
||
}
|
||
|
||
var remainingMaxAge = Math.max(0, _parseMaxAge(originalCc) - ageSeconds);
|
||
headers["Cache-Control"] = "public, max-age=" + remainingMaxAge;
|
||
|
||
headers["X-Cache"] = "HIT from relay-spreadsheet";
|
||
headers["X-Cached-At"] = cachedAt.toUTCString();
|
||
|
||
return JSON.stringify(headers);
|
||
}
|
||
|
||
/**
|
||
* Retrieve a cached response by compound cache key.
|
||
* Uses TextFinder for O(log n) lookup. Skips expired entries.
|
||
* Returns null on miss, expired entry, or unavailable sheet.
|
||
*/
|
||
function _getFromCache(url, reqHeaders) {
|
||
var sheet = _initCacheSheet();
|
||
if (!sheet) return null;
|
||
|
||
var hash = _getCacheKey(url, reqHeaders);
|
||
var finder = sheet.createTextFinder(hash).matchEntireCell(true);
|
||
var found = finder.findNext();
|
||
|
||
if (found) {
|
||
// 8-column read. Legacy 7-column rows return undefined for the Z slot,
|
||
// which is falsy and falls through the not-gzipped branch below — fully
|
||
// compatible with caches written before the gzip-storage change.
|
||
var row = sheet.getRange(found.getRow(), 1, 1, 8).getValues()[0];
|
||
|
||
var expiresAt = row[6];
|
||
if (expiresAt && expiresAt instanceof Date && expiresAt < new Date()) {
|
||
return null;
|
||
}
|
||
|
||
var storedBody = row[4];
|
||
var body;
|
||
if (row[7]) {
|
||
// Stored as base64(gzip(rawBytes)). The relay protocol's `b` field
|
||
// is base64(rawBytes), so decompress and re-encode for the wire.
|
||
var gzipped = Utilities.base64Decode(storedBody);
|
||
var raw = Utilities
|
||
.ungzip(Utilities.newBlob(gzipped, "application/x-gzip"))
|
||
.getBytes();
|
||
body = Utilities.base64Encode(raw);
|
||
} else {
|
||
body = storedBody;
|
||
}
|
||
|
||
return {
|
||
status: row[2],
|
||
headers: _refreshCachedHeaders(row[3], row[5]),
|
||
body: body,
|
||
};
|
||
}
|
||
return null;
|
||
}
|
||
|
||
/**
|
||
* Fetch a URL and store the response in the spreadsheet cache
|
||
* using a circular buffer (O(1) writes). Skips storage on 5xx
|
||
* (transient outages must not poison a 24h slot), when Cache-Control
|
||
* forbids caching, or when the post-compression body exceeds
|
||
* CACHE_MAX_BODY_BYTES. Always returns the fetch result so the caller
|
||
* can serve the live response even when the cache write is skipped.
|
||
*/
|
||
function _fetchAndCache(url, reqHeaders) {
|
||
var sheet = _initCacheSheet();
|
||
if (!sheet) return null;
|
||
|
||
try {
|
||
var response = UrlFetchApp.fetch(url, { muteHttpExceptions: true });
|
||
var status = response.getResponseCode();
|
||
var headers = _respHeaders(response);
|
||
var bodyBytes = response.getContent();
|
||
var rawB64 = Utilities.base64Encode(bodyBytes);
|
||
var headersJson = JSON.stringify(headers);
|
||
var liveResult = { status: status, headers: headersJson, body: rawB64 };
|
||
|
||
// 5xx never enters the cache. A flapping upstream returning 503 once
|
||
// would otherwise pin that response for 24h and break the URL for
|
||
// every subsequent client until expiry.
|
||
if (status >= 500) return liveResult;
|
||
|
||
var cacheControl =
|
||
headers["Cache-Control"] || headers["cache-control"] || null;
|
||
var ttlSeconds = _parseMaxAge(cacheControl);
|
||
|
||
if (ttlSeconds === 0) return liveResult;
|
||
|
||
// Negative caching: cap TTL on persistent 4xx when upstream is silent.
|
||
// If they explicitly stated a max-age for the 404, we honor it instead
|
||
// — the origin knows best when it spoke up.
|
||
if (NEGATIVE_CACHE_STATUSES[status] && !cacheControl) {
|
||
ttlSeconds = NEGATIVE_CACHE_TTL_SECONDS;
|
||
}
|
||
|
||
// Decide whether to gzip-store. Skip when upstream is already encoded
|
||
// (avoids double-compressing gzip/br/zstd payloads) and when the body
|
||
// is too small to overcome gzip's header overhead.
|
||
var contentEncoding = String(
|
||
headers["Content-Encoding"] || headers["content-encoding"] || ""
|
||
).toLowerCase();
|
||
var alreadyEncoded = contentEncoding && contentEncoding !== "identity";
|
||
var storedBody;
|
||
var storedZ;
|
||
if (alreadyEncoded || bodyBytes.length < GZIP_MIN_BYTES) {
|
||
storedBody = rawB64;
|
||
storedZ = 0;
|
||
} else {
|
||
storedBody = Utilities.base64Encode(
|
||
Utilities.gzip(Utilities.newBlob(bodyBytes)).getBytes()
|
||
);
|
||
storedZ = 1;
|
||
}
|
||
|
||
// Cell-size safety gate, applied after compression so that a 100 KB
|
||
// text body that gzips to ~15 KB now fits where it previously bailed.
|
||
if (storedBody.length > CACHE_MAX_BODY_BYTES) return liveResult;
|
||
|
||
var hash = _getCacheKey(url, reqHeaders);
|
||
var timestamp = new Date();
|
||
var expiresAt = new Date(timestamp.getTime() + ttlSeconds * 1000);
|
||
|
||
// Safety: fallback if Date math produces invalid result
|
||
if (isNaN(expiresAt.getTime())) {
|
||
expiresAt = new Date(timestamp.getTime() + CACHE_DEFAULT_TTL_SECONDS * 1000);
|
||
}
|
||
|
||
var rowData = [
|
||
hash,
|
||
url,
|
||
status,
|
||
headersJson,
|
||
storedBody,
|
||
timestamp.toISOString(),
|
||
expiresAt,
|
||
storedZ,
|
||
];
|
||
|
||
// Circular buffer write (O(1))
|
||
var metaSheet = _getMetaSheet();
|
||
if (metaSheet) {
|
||
_ensureRowsAllocated(sheet);
|
||
var writeRow = _getNextCursor(sheet, metaSheet);
|
||
sheet.getRange(writeRow, 1, 1, 8).setValues([rowData]);
|
||
_advanceCursor(metaSheet, writeRow);
|
||
} else {
|
||
// Fallback: simple append if meta sheet is unavailable
|
||
sheet.appendRow(rowData);
|
||
}
|
||
|
||
return liveResult;
|
||
} catch (e) {
|
||
return null;
|
||
}
|
||
}
|
||
|
||
// ═══════════════════════════════════════════════════════════
|
||
// SPREADSHEET CACHE — DIAGNOSTICS
|
||
// ═══════════════════════════════════════════════════════════
|
||
|
||
function getCacheStats() {
|
||
var sheet = _initCacheSheet();
|
||
if (!sheet) {
|
||
console.log("Cache is not enabled or spreadsheet unavailable.");
|
||
return;
|
||
}
|
||
|
||
var data = sheet.getDataRange().getValues();
|
||
var totalEntries = data.length - 1;
|
||
var now = new Date();
|
||
var expiredCount = 0;
|
||
|
||
for (var i = 1; i < data.length; i++) {
|
||
var expiresAt = data[i][6];
|
||
if (expiresAt && expiresAt instanceof Date && expiresAt < now) {
|
||
expiredCount++;
|
||
}
|
||
}
|
||
|
||
var metaSheet = _getMetaSheet();
|
||
var cursorInfo = "N/A";
|
||
if (metaSheet) {
|
||
cursorInfo = String(metaSheet.getRange(CACHE_META_CURSOR_CELL).getValue());
|
||
}
|
||
|
||
console.log("=== CACHE STATS ===");
|
||
console.log("Total rows used: " + totalEntries + " / " + CACHE_MAX_ROWS);
|
||
console.log("Active entries: " + (totalEntries - expiredCount));
|
||
console.log("Expired entries: " + expiredCount);
|
||
console.log("Cursor position: " + cursorInfo);
|
||
console.log("Max body size: " + CACHE_MAX_BODY_BYTES + " chars");
|
||
console.log("Default TTL: " + CACHE_DEFAULT_TTL_SECONDS + " sec");
|
||
console.log("Vary key headers: " + VARY_KEY_HEADERS.join(", "));
|
||
if (totalEntries > 0) {
|
||
console.log("Oldest entry: " + data[1][5]);
|
||
console.log("Newest entry: " + data[data.length - 1][5]);
|
||
}
|
||
}
|
||
|
||
function clearExpiredCache() {
|
||
var sheet = _initCacheSheet();
|
||
if (!sheet) {
|
||
console.log("Cache is not enabled.");
|
||
return;
|
||
}
|
||
|
||
var data = sheet.getDataRange().getValues();
|
||
var now = new Date();
|
||
var rowsToClear = [];
|
||
|
||
for (var i = 1; i < data.length; i++) {
|
||
var expiresAt = data[i][6];
|
||
if (expiresAt && expiresAt instanceof Date && expiresAt < now) {
|
||
rowsToClear.push(i + 1);
|
||
}
|
||
}
|
||
|
||
for (var j = 0; j < rowsToClear.length; j++) {
|
||
sheet.getRange(rowsToClear[j], 1, 1, 8).clearContent();
|
||
}
|
||
|
||
console.log("Cleared " + rowsToClear.length + " expired entries (" +
|
||
(data.length - 1 - rowsToClear.length) + " remaining).");
|
||
}
|
||
|
||
function clearEntireCache() {
|
||
var sheet = _initCacheSheet();
|
||
if (sheet) {
|
||
var totalRows = sheet.getDataRange().getNumRows();
|
||
if (totalRows > 1) {
|
||
sheet.getRange(2, 1, totalRows - 1, 8).clearContent();
|
||
}
|
||
}
|
||
|
||
var metaSheet = _getMetaSheet();
|
||
if (metaSheet) {
|
||
metaSheet.getRange(CACHE_META_CURSOR_CELL).setValue(2);
|
||
}
|
||
|
||
console.log("Cache wiped. Cursor reset to row 2.");
|
||
}
|