Compare commits
78 Commits
1.13.0-rc.
...
1.13.0
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
5d6ee45125 | ||
|
|
fceaedfcd8 | ||
|
|
181612ce25 | ||
|
|
224b78fc64 | ||
|
|
757e540be6 | ||
|
|
bf1675686c | ||
|
|
f81909489a | ||
|
|
963468d7fa | ||
|
|
f67f4f8834 | ||
|
|
4c819d264b | ||
|
|
cbcb23ccea | ||
|
|
d8b27de5ac | ||
|
|
01f7842fd5 | ||
|
|
d409e58186 | ||
|
|
c9e1c4da1c | ||
|
|
9c38f65ad4 | ||
|
|
2316462721 | ||
|
|
7cc990107a | ||
|
|
9917a569ac | ||
|
|
c56574e431 | ||
|
|
f9c0e0ec3d | ||
|
|
85986dcccb | ||
|
|
c9779254c3 | ||
|
|
5b620469c7 | ||
|
|
df4b9de334 | ||
|
|
d490cab48c | ||
|
|
b68c0962c6 | ||
|
|
ee2a438602 | ||
|
|
74dd3fdc9f | ||
|
|
314da3ee3e | ||
|
|
68cfc84249 | ||
|
|
0bcf5c2b42 | ||
|
|
9210e005e9 | ||
|
|
f245632371 | ||
|
|
6453b070bb | ||
|
|
8c4db93a93 | ||
|
|
f9b03943c3 | ||
|
|
fa839a811f | ||
|
|
88d2c2eac8 | ||
|
|
c84cc1815b | ||
|
|
2c23ffd178 | ||
|
|
da3f7ae404 | ||
|
|
f460559a4b | ||
|
|
0c9deeb2d7 | ||
|
|
1289b99f14 | ||
|
|
1a7a6e5b6f | ||
|
|
f56135eed3 | ||
|
|
23e9a61f3e | ||
|
|
5428ad1009 | ||
|
|
bba28bc5f2 | ||
|
|
18498a32ce | ||
|
|
887af85db1 | ||
|
|
a306aa971b | ||
|
|
0a9b19ecfc | ||
|
|
e011580b96 | ||
|
|
048ce850a8 | ||
|
|
2ca1f15add | ||
|
|
05ebd547b5 | ||
|
|
5a8b1383a4 | ||
|
|
ede51bebb5 | ||
|
|
fd29071d57 | ||
|
|
8e1af79dc4 | ||
|
|
dc8c28626d | ||
|
|
9db2feff77 | ||
|
|
adf76bfb53 | ||
|
|
e0a79b7d4d | ||
|
|
9ea3914a93 | ||
|
|
1aeb31be04 | ||
|
|
64120ea878 | ||
|
|
0003ec021b | ||
|
|
c9a1da210f | ||
|
|
ace402af2d | ||
|
|
e60dce25c9 | ||
|
|
ccfff030e5 | ||
|
|
00765c1faf | ||
|
|
f6bbdeadb9 | ||
|
|
9cf520574a | ||
|
|
f8ab5b7af7 |
@@ -1,6 +1,3 @@
|
|||||||
{
|
{
|
||||||
"extends": [
|
"extends": ["next/core-web-vitals", "next/typescript"]
|
||||||
"next/core-web-vitals",
|
|
||||||
"next/typescript"
|
|
||||||
]
|
|
||||||
}
|
}
|
||||||
|
|||||||
2
.github/workflows/cicd.yml
vendored
@@ -36,7 +36,7 @@ jobs:
|
|||||||
|
|
||||||
steps:
|
steps:
|
||||||
- name: Checkout code
|
- name: Checkout code
|
||||||
uses: actions/checkout@1af3b93b6815bc44a9784bd300feb67ff0d1eeb3 # v6.0.0
|
uses: actions/checkout@8e8c483db84b4bee98b60c0593521ed34d9990e8 # v6.0.1
|
||||||
|
|
||||||
- name: Set up QEMU
|
- name: Set up QEMU
|
||||||
uses: docker/setup-qemu-action@c7c53464625b32c7a7e944ae62b3e17d2b600130 # v3.7.0
|
uses: docker/setup-qemu-action@c7c53464625b32c7a7e944ae62b3e17d2b600130 # v3.7.0
|
||||||
|
|||||||
4
.github/workflows/linting.yml
vendored
@@ -21,10 +21,10 @@ jobs:
|
|||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
steps:
|
steps:
|
||||||
- name: Checkout code
|
- name: Checkout code
|
||||||
uses: actions/checkout@1af3b93b6815bc44a9784bd300feb67ff0d1eeb3 # v6.0.0
|
uses: actions/checkout@8e8c483db84b4bee98b60c0593521ed34d9990e8 # v6.0.1
|
||||||
|
|
||||||
- name: Set up Node.js
|
- name: Set up Node.js
|
||||||
uses: actions/setup-node@2028fbc5c25fe9cf00d9f06a71cc4710d4507903 # v6.0.0
|
uses: actions/setup-node@395ad3262231945c25e8478fd5baf05154b1d79f # v6.1.0
|
||||||
with:
|
with:
|
||||||
node-version: '22'
|
node-version: '22'
|
||||||
|
|
||||||
|
|||||||
2
.github/workflows/stale-bot.yml
vendored
@@ -14,7 +14,7 @@ jobs:
|
|||||||
stale:
|
stale:
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/stale@5f858e3efba33a5ca4407a664cc011ad407f2008 # v10.1.0
|
- uses: actions/stale@997185467fa4f803885201cee163a9f38240193d # v10.1.1
|
||||||
with:
|
with:
|
||||||
days-before-stale: 14
|
days-before-stale: 14
|
||||||
days-before-close: 14
|
days-before-close: 14
|
||||||
|
|||||||
4
.github/workflows/test.yml
vendored
@@ -14,9 +14,9 @@ jobs:
|
|||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
|
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@1af3b93b6815bc44a9784bd300feb67ff0d1eeb3 # v6.0.0
|
- uses: actions/checkout@8e8c483db84b4bee98b60c0593521ed34d9990e8 # v6.0.1
|
||||||
|
|
||||||
- uses: actions/setup-node@2028fbc5c25fe9cf00d9f06a71cc4710d4507903 # v6.0.0
|
- uses: actions/setup-node@395ad3262231945c25e8478fd5baf05154b1d79f # v6.1.0
|
||||||
with:
|
with:
|
||||||
node-version: '22'
|
node-version: '22'
|
||||||
|
|
||||||
|
|||||||
12
.prettierignore
Normal file
@@ -0,0 +1,12 @@
|
|||||||
|
.github/
|
||||||
|
bruno/
|
||||||
|
cli/
|
||||||
|
config/
|
||||||
|
messages/
|
||||||
|
next.config.mjs/
|
||||||
|
public/
|
||||||
|
tailwind.config.js/
|
||||||
|
test/
|
||||||
|
**/*.yml
|
||||||
|
**/*.yaml
|
||||||
|
**/*.md
|
||||||
3
.vscode/extensions.json
vendored
Normal file
@@ -0,0 +1,3 @@
|
|||||||
|
{
|
||||||
|
"recommendations": ["esbenp.prettier-vscode"]
|
||||||
|
}
|
||||||
22
.vscode/settings.json
vendored
Normal file
@@ -0,0 +1,22 @@
|
|||||||
|
{
|
||||||
|
"editor.codeActionsOnSave": {
|
||||||
|
"source.addMissingImports.ts": "always"
|
||||||
|
},
|
||||||
|
"editor.defaultFormatter": "esbenp.prettier-vscode",
|
||||||
|
"[jsonc]": {
|
||||||
|
"editor.defaultFormatter": "esbenp.prettier-vscode"
|
||||||
|
},
|
||||||
|
"[javascript]": {
|
||||||
|
"editor.defaultFormatter": "esbenp.prettier-vscode"
|
||||||
|
},
|
||||||
|
"[typescript]": {
|
||||||
|
"editor.defaultFormatter": "esbenp.prettier-vscode"
|
||||||
|
},
|
||||||
|
"[typescriptreact]": {
|
||||||
|
"editor.defaultFormatter": "esbenp.prettier-vscode"
|
||||||
|
},
|
||||||
|
"[json]": {
|
||||||
|
"editor.defaultFormatter": "esbenp.prettier-vscode"
|
||||||
|
},
|
||||||
|
"editor.formatOnSave": true
|
||||||
|
}
|
||||||
20
README.md
@@ -41,7 +41,7 @@
|
|||||||
</strong>
|
</strong>
|
||||||
</p>
|
</p>
|
||||||
|
|
||||||
Pangolin is a self-hosted tunneled reverse proxy server with identity and context aware access control, designed to easily expose and protect applications running anywhere. Pangolin acts as a central hub and connects isolated networks — even those behind restrictive firewalls — through encrypted tunnels, enabling easy access to remote services without opening ports or requiring a VPN.
|
Pangolin is an open-source, identity-based remote access platform built on WireGuard that enables secure, seamless connectivity to private and public resources. Pangolin combines reverse proxy and VPN capabilities into one platform, providing browser-based access to web applications and client-based access to any private resources, all with zero-trust security and granular access control.
|
||||||
|
|
||||||
## Installation
|
## Installation
|
||||||
|
|
||||||
@@ -60,14 +60,20 @@ Pangolin is a self-hosted tunneled reverse proxy server with identity and contex
|
|||||||
|
|
||||||
## Key Features
|
## Key Features
|
||||||
|
|
||||||
Pangolin packages everything you need for seamless application access and exposure into one cohesive platform.
|
|
||||||
|
|
||||||
| <img width=500 /> | <img width=500 /> |
|
| <img width=500 /> | <img width=500 /> |
|
||||||
|----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|--------------------------------------------------------------------|
|
|----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|--------------------------------------------------------------------|
|
||||||
| **Manage applications in one place**<br /><br /> Pangolin provides a unified dashboard where you can monitor, configure, and secure all of your services regardless of where they are hosted. | <img src="public/screenshots/hero.png" width=500 /><tr></tr> |
|
| **Connect remote networks with sites**<br /><br />Pangolin's lightweight site connectors create secure tunnels from remote networks without requiring public IP addresses or open ports. Sites make any network anywhere available for authorized access. | <img src="public/screenshots/sites.png" width=500 /><tr></tr> |
|
||||||
| **Reverse proxy across networks anywhere**<br /><br />Route traffic via tunnels to any private network. Pangolin works like a reverse proxy that spans multiple networks and handles routing, load balancing, health checking, and more to the right services on the other end. | <img src="public/screenshots/sites.png" width=500 /><tr></tr> |
|
| **Browser-based reverse proxy access**<br /><br />Expose web applications through identity and context-aware tunneled reverse proxies. Pangolin handles routing, load balancing, health checking, and automatic SSL certificates without exposing your network directly to the internet. Users access applications through any web browser with authentication and granular access control. | <img src="public/clip.gif" width=500 /><tr></tr> |
|
||||||
| **Enforce identity and context aware rules**<br /><br />Protect your applications with identity and context aware rules such as SSO, OIDC, PIN, password, temporary share links, geolocation, IP, and more. | <img src="public/auth-diagram1.png" width=500 /><tr></tr> |
|
| **Client-based private resource access**<br /><br />Access private resources like SSH servers, databases, RDP, and entire network ranges through Pangolin clients. Intelligent NAT traversal enables connections even through restrictive firewalls, while DNS aliases provide friendly names and fast connections to resources across all your sites. | <img src="public/screenshots/private-resources.png" width=500 /><tr></tr> |
|
||||||
| **Quickly connect Pangolin sites**<br /><br />Pangolin's lightweight [Newt](https://github.com/fosrl/newt) client runs in userspace and can run anywhere. Use it as a site connector to route traffic to backends across all of your environments. | <img src="public/clip.gif" width=500 /><tr></tr> |
|
| **Zero-trust granular access**<br /><br />Grant users access to specific resources, not entire networks. Unlike traditional VPNs that expose full network access, Pangolin's zero-trust model ensures users can only reach the applications and services you explicitly define, reducing security risk and attack surface. | <img src="public/screenshots/user-devices.png" width=500 /><tr></tr> |
|
||||||
|
|
||||||
|
## Download Clients
|
||||||
|
|
||||||
|
Download the Pangolin client for your platform:
|
||||||
|
|
||||||
|
- [Mac](https://pangolin.net/downloads/mac)
|
||||||
|
- [Windows](https://pangolin.net/downloads/windows)
|
||||||
|
- [Linux](https://pangolin.net/downloads/linux)
|
||||||
|
|
||||||
## Get Started
|
## Get Started
|
||||||
|
|
||||||
|
|||||||
@@ -17,4 +17,4 @@
|
|||||||
"lib": "@/lib",
|
"lib": "@/lib",
|
||||||
"hooks": "@/hooks"
|
"hooks": "@/hooks"
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,9 +1,7 @@
|
|||||||
import { defineConfig } from "drizzle-kit";
|
import { defineConfig } from "drizzle-kit";
|
||||||
import path from "path";
|
import path from "path";
|
||||||
|
|
||||||
const schema = [
|
const schema = [path.join("server", "db", "pg", "schema")];
|
||||||
path.join("server", "db", "pg", "schema"),
|
|
||||||
];
|
|
||||||
|
|
||||||
export default defineConfig({
|
export default defineConfig({
|
||||||
dialect: "postgresql",
|
dialect: "postgresql",
|
||||||
|
|||||||
@@ -2,9 +2,7 @@ import { APP_PATH } from "@server/lib/consts";
|
|||||||
import { defineConfig } from "drizzle-kit";
|
import { defineConfig } from "drizzle-kit";
|
||||||
import path from "path";
|
import path from "path";
|
||||||
|
|
||||||
const schema = [
|
const schema = [path.join("server", "db", "sqlite", "schema")];
|
||||||
path.join("server", "db", "sqlite", "schema"),
|
|
||||||
];
|
|
||||||
|
|
||||||
export default defineConfig({
|
export default defineConfig({
|
||||||
dialect: "sqlite",
|
dialect: "sqlite",
|
||||||
|
|||||||
95
esbuild.mjs
@@ -24,20 +24,20 @@ const argv = yargs(hideBin(process.argv))
|
|||||||
alias: "e",
|
alias: "e",
|
||||||
describe: "Entry point file",
|
describe: "Entry point file",
|
||||||
type: "string",
|
type: "string",
|
||||||
demandOption: true,
|
demandOption: true
|
||||||
})
|
})
|
||||||
.option("out", {
|
.option("out", {
|
||||||
alias: "o",
|
alias: "o",
|
||||||
describe: "Output file path",
|
describe: "Output file path",
|
||||||
type: "string",
|
type: "string",
|
||||||
demandOption: true,
|
demandOption: true
|
||||||
})
|
})
|
||||||
.option("build", {
|
.option("build", {
|
||||||
alias: "b",
|
alias: "b",
|
||||||
describe: "Build type (oss, saas, enterprise)",
|
describe: "Build type (oss, saas, enterprise)",
|
||||||
type: "string",
|
type: "string",
|
||||||
choices: ["oss", "saas", "enterprise"],
|
choices: ["oss", "saas", "enterprise"],
|
||||||
default: "oss",
|
default: "oss"
|
||||||
})
|
})
|
||||||
.help()
|
.help()
|
||||||
.alias("help", "h").argv;
|
.alias("help", "h").argv;
|
||||||
@@ -66,7 +66,9 @@ function privateImportGuardPlugin() {
|
|||||||
|
|
||||||
// Check if the importing file is NOT in server/private
|
// Check if the importing file is NOT in server/private
|
||||||
const normalizedImporter = path.normalize(importingFile);
|
const normalizedImporter = path.normalize(importingFile);
|
||||||
const isInServerPrivate = normalizedImporter.includes(path.normalize("server/private"));
|
const isInServerPrivate = normalizedImporter.includes(
|
||||||
|
path.normalize("server/private")
|
||||||
|
);
|
||||||
|
|
||||||
if (!isInServerPrivate) {
|
if (!isInServerPrivate) {
|
||||||
const violation = {
|
const violation = {
|
||||||
@@ -79,8 +81,8 @@ function privateImportGuardPlugin() {
|
|||||||
console.log(`PRIVATE IMPORT VIOLATION:`);
|
console.log(`PRIVATE IMPORT VIOLATION:`);
|
||||||
console.log(` File: ${importingFile}`);
|
console.log(` File: ${importingFile}`);
|
||||||
console.log(` Import: ${args.path}`);
|
console.log(` Import: ${args.path}`);
|
||||||
console.log(` Resolve dir: ${args.resolveDir || 'N/A'}`);
|
console.log(` Resolve dir: ${args.resolveDir || "N/A"}`);
|
||||||
console.log('');
|
console.log("");
|
||||||
}
|
}
|
||||||
|
|
||||||
// Return null to let the default resolver handle it
|
// Return null to let the default resolver handle it
|
||||||
@@ -89,16 +91,20 @@ function privateImportGuardPlugin() {
|
|||||||
|
|
||||||
build.onEnd((result) => {
|
build.onEnd((result) => {
|
||||||
if (violations.length > 0) {
|
if (violations.length > 0) {
|
||||||
console.log(`\nSUMMARY: Found ${violations.length} private import violation(s):`);
|
console.log(
|
||||||
|
`\nSUMMARY: Found ${violations.length} private import violation(s):`
|
||||||
|
);
|
||||||
violations.forEach((v, i) => {
|
violations.forEach((v, i) => {
|
||||||
console.log(` ${i + 1}. ${path.relative(process.cwd(), v.file)} imports ${v.importPath}`);
|
console.log(
|
||||||
|
` ${i + 1}. ${path.relative(process.cwd(), v.file)} imports ${v.importPath}`
|
||||||
|
);
|
||||||
});
|
});
|
||||||
console.log('');
|
console.log("");
|
||||||
|
|
||||||
result.errors.push({
|
result.errors.push({
|
||||||
text: `Private import violations detected: ${violations.length} violation(s) found`,
|
text: `Private import violations detected: ${violations.length} violation(s) found`,
|
||||||
location: null,
|
location: null,
|
||||||
notes: violations.map(v => ({
|
notes: violations.map((v) => ({
|
||||||
text: `${path.relative(process.cwd(), v.file)} imports ${v.importPath}`,
|
text: `${path.relative(process.cwd(), v.file)} imports ${v.importPath}`,
|
||||||
location: null
|
location: null
|
||||||
}))
|
}))
|
||||||
@@ -121,7 +127,9 @@ function dynamicImportGuardPlugin() {
|
|||||||
|
|
||||||
// Check if the importing file is NOT in server/private
|
// Check if the importing file is NOT in server/private
|
||||||
const normalizedImporter = path.normalize(importingFile);
|
const normalizedImporter = path.normalize(importingFile);
|
||||||
const isInServerPrivate = normalizedImporter.includes(path.normalize("server/private"));
|
const isInServerPrivate = normalizedImporter.includes(
|
||||||
|
path.normalize("server/private")
|
||||||
|
);
|
||||||
|
|
||||||
if (isInServerPrivate) {
|
if (isInServerPrivate) {
|
||||||
const violation = {
|
const violation = {
|
||||||
@@ -134,8 +142,8 @@ function dynamicImportGuardPlugin() {
|
|||||||
console.log(`DYNAMIC IMPORT VIOLATION:`);
|
console.log(`DYNAMIC IMPORT VIOLATION:`);
|
||||||
console.log(` File: ${importingFile}`);
|
console.log(` File: ${importingFile}`);
|
||||||
console.log(` Import: ${args.path}`);
|
console.log(` Import: ${args.path}`);
|
||||||
console.log(` Resolve dir: ${args.resolveDir || 'N/A'}`);
|
console.log(` Resolve dir: ${args.resolveDir || "N/A"}`);
|
||||||
console.log('');
|
console.log("");
|
||||||
}
|
}
|
||||||
|
|
||||||
// Return null to let the default resolver handle it
|
// Return null to let the default resolver handle it
|
||||||
@@ -144,16 +152,20 @@ function dynamicImportGuardPlugin() {
|
|||||||
|
|
||||||
build.onEnd((result) => {
|
build.onEnd((result) => {
|
||||||
if (violations.length > 0) {
|
if (violations.length > 0) {
|
||||||
console.log(`\nSUMMARY: Found ${violations.length} dynamic import violation(s):`);
|
console.log(
|
||||||
|
`\nSUMMARY: Found ${violations.length} dynamic import violation(s):`
|
||||||
|
);
|
||||||
violations.forEach((v, i) => {
|
violations.forEach((v, i) => {
|
||||||
console.log(` ${i + 1}. ${path.relative(process.cwd(), v.file)} imports ${v.importPath}`);
|
console.log(
|
||||||
|
` ${i + 1}. ${path.relative(process.cwd(), v.file)} imports ${v.importPath}`
|
||||||
|
);
|
||||||
});
|
});
|
||||||
console.log('');
|
console.log("");
|
||||||
|
|
||||||
result.errors.push({
|
result.errors.push({
|
||||||
text: `Dynamic import violations detected: ${violations.length} violation(s) found`,
|
text: `Dynamic import violations detected: ${violations.length} violation(s) found`,
|
||||||
location: null,
|
location: null,
|
||||||
notes: violations.map(v => ({
|
notes: violations.map((v) => ({
|
||||||
text: `${path.relative(process.cwd(), v.file)} imports ${v.importPath}`,
|
text: `${path.relative(process.cwd(), v.file)} imports ${v.importPath}`,
|
||||||
location: null
|
location: null
|
||||||
}))
|
}))
|
||||||
@@ -172,21 +184,28 @@ function dynamicImportSwitcherPlugin(buildValue) {
|
|||||||
const switches = [];
|
const switches = [];
|
||||||
|
|
||||||
build.onStart(() => {
|
build.onStart(() => {
|
||||||
console.log(`Dynamic import switcher using build type: ${buildValue}`);
|
console.log(
|
||||||
|
`Dynamic import switcher using build type: ${buildValue}`
|
||||||
|
);
|
||||||
});
|
});
|
||||||
|
|
||||||
build.onResolve({ filter: /^#dynamic\// }, (args) => {
|
build.onResolve({ filter: /^#dynamic\// }, (args) => {
|
||||||
// Extract the path after #dynamic/
|
// Extract the path after #dynamic/
|
||||||
const dynamicPath = args.path.replace(/^#dynamic\//, '');
|
const dynamicPath = args.path.replace(/^#dynamic\//, "");
|
||||||
|
|
||||||
// Determine the replacement based on build type
|
// Determine the replacement based on build type
|
||||||
let replacement;
|
let replacement;
|
||||||
if (buildValue === "oss") {
|
if (buildValue === "oss") {
|
||||||
replacement = `#open/${dynamicPath}`;
|
replacement = `#open/${dynamicPath}`;
|
||||||
} else if (buildValue === "saas" || buildValue === "enterprise") {
|
} else if (
|
||||||
|
buildValue === "saas" ||
|
||||||
|
buildValue === "enterprise"
|
||||||
|
) {
|
||||||
replacement = `#closed/${dynamicPath}`; // We use #closed here so that the route guards dont complain after its been changed but this is the same as #private
|
replacement = `#closed/${dynamicPath}`; // We use #closed here so that the route guards dont complain after its been changed but this is the same as #private
|
||||||
} else {
|
} else {
|
||||||
console.warn(`Unknown build type '${buildValue}', defaulting to #open/`);
|
console.warn(
|
||||||
|
`Unknown build type '${buildValue}', defaulting to #open/`
|
||||||
|
);
|
||||||
replacement = `#open/${dynamicPath}`;
|
replacement = `#open/${dynamicPath}`;
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -201,8 +220,10 @@ function dynamicImportSwitcherPlugin(buildValue) {
|
|||||||
console.log(`DYNAMIC IMPORT SWITCH:`);
|
console.log(`DYNAMIC IMPORT SWITCH:`);
|
||||||
console.log(` File: ${args.importer}`);
|
console.log(` File: ${args.importer}`);
|
||||||
console.log(` Original: ${args.path}`);
|
console.log(` Original: ${args.path}`);
|
||||||
console.log(` Switched to: ${replacement} (build: ${buildValue})`);
|
console.log(
|
||||||
console.log('');
|
` Switched to: ${replacement} (build: ${buildValue})`
|
||||||
|
);
|
||||||
|
console.log("");
|
||||||
|
|
||||||
// Rewrite the import path and let the normal resolution continue
|
// Rewrite the import path and let the normal resolution continue
|
||||||
return build.resolve(replacement, {
|
return build.resolve(replacement, {
|
||||||
@@ -215,12 +236,18 @@ function dynamicImportSwitcherPlugin(buildValue) {
|
|||||||
|
|
||||||
build.onEnd((result) => {
|
build.onEnd((result) => {
|
||||||
if (switches.length > 0) {
|
if (switches.length > 0) {
|
||||||
console.log(`\nDYNAMIC IMPORT SUMMARY: Switched ${switches.length} import(s) for build type '${buildValue}':`);
|
console.log(
|
||||||
|
`\nDYNAMIC IMPORT SUMMARY: Switched ${switches.length} import(s) for build type '${buildValue}':`
|
||||||
|
);
|
||||||
switches.forEach((s, i) => {
|
switches.forEach((s, i) => {
|
||||||
console.log(` ${i + 1}. ${path.relative(process.cwd(), s.file)}`);
|
console.log(
|
||||||
console.log(` ${s.originalPath} → ${s.replacementPath}`);
|
` ${i + 1}. ${path.relative(process.cwd(), s.file)}`
|
||||||
|
);
|
||||||
|
console.log(
|
||||||
|
` ${s.originalPath} → ${s.replacementPath}`
|
||||||
|
);
|
||||||
});
|
});
|
||||||
console.log('');
|
console.log("");
|
||||||
}
|
}
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
@@ -235,7 +262,7 @@ esbuild
|
|||||||
format: "esm",
|
format: "esm",
|
||||||
minify: false,
|
minify: false,
|
||||||
banner: {
|
banner: {
|
||||||
js: banner,
|
js: banner
|
||||||
},
|
},
|
||||||
platform: "node",
|
platform: "node",
|
||||||
external: ["body-parser"],
|
external: ["body-parser"],
|
||||||
@@ -244,20 +271,22 @@ esbuild
|
|||||||
dynamicImportGuardPlugin(),
|
dynamicImportGuardPlugin(),
|
||||||
dynamicImportSwitcherPlugin(argv.build),
|
dynamicImportSwitcherPlugin(argv.build),
|
||||||
nodeExternalsPlugin({
|
nodeExternalsPlugin({
|
||||||
packagePath: getPackagePaths(),
|
packagePath: getPackagePaths()
|
||||||
}),
|
})
|
||||||
],
|
],
|
||||||
sourcemap: "inline",
|
sourcemap: "inline",
|
||||||
target: "node22",
|
target: "node22"
|
||||||
})
|
})
|
||||||
.then((result) => {
|
.then((result) => {
|
||||||
// Check if there were any errors in the build result
|
// Check if there were any errors in the build result
|
||||||
if (result.errors && result.errors.length > 0) {
|
if (result.errors && result.errors.length > 0) {
|
||||||
console.error(`Build failed with ${result.errors.length} error(s):`);
|
console.error(
|
||||||
|
`Build failed with ${result.errors.length} error(s):`
|
||||||
|
);
|
||||||
result.errors.forEach((error, i) => {
|
result.errors.forEach((error, i) => {
|
||||||
console.error(`${i + 1}. ${error.text}`);
|
console.error(`${i + 1}. ${error.text}`);
|
||||||
if (error.notes) {
|
if (error.notes) {
|
||||||
error.notes.forEach(note => {
|
error.notes.forEach((note) => {
|
||||||
console.error(` - ${note.text}`);
|
console.error(` - ${note.text}`);
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,19 +1,19 @@
|
|||||||
import tseslint from 'typescript-eslint';
|
import tseslint from "typescript-eslint";
|
||||||
|
|
||||||
export default tseslint.config({
|
export default tseslint.config({
|
||||||
files: ["**/*.{ts,tsx,js,jsx}"],
|
files: ["**/*.{ts,tsx,js,jsx}"],
|
||||||
languageOptions: {
|
languageOptions: {
|
||||||
parser: tseslint.parser,
|
parser: tseslint.parser,
|
||||||
parserOptions: {
|
parserOptions: {
|
||||||
ecmaVersion: "latest",
|
ecmaVersion: "latest",
|
||||||
sourceType: "module",
|
sourceType: "module",
|
||||||
ecmaFeatures: {
|
ecmaFeatures: {
|
||||||
jsx: true
|
jsx: true
|
||||||
}
|
}
|
||||||
|
}
|
||||||
|
},
|
||||||
|
rules: {
|
||||||
|
semi: "error",
|
||||||
|
"prefer-const": "warn"
|
||||||
}
|
}
|
||||||
},
|
});
|
||||||
rules: {
|
|
||||||
"semi": "error",
|
|
||||||
"prefer-const": "warn"
|
|
||||||
}
|
|
||||||
});
|
|
||||||
|
|||||||
@@ -3,8 +3,8 @@ module installer
|
|||||||
go 1.24.0
|
go 1.24.0
|
||||||
|
|
||||||
require (
|
require (
|
||||||
golang.org/x/term v0.37.0
|
golang.org/x/term v0.38.0
|
||||||
gopkg.in/yaml.v3 v3.0.1
|
gopkg.in/yaml.v3 v3.0.1
|
||||||
)
|
)
|
||||||
|
|
||||||
require golang.org/x/sys v0.38.0 // indirect
|
require golang.org/x/sys v0.39.0 // indirect
|
||||||
|
|||||||
@@ -1,7 +1,7 @@
|
|||||||
golang.org/x/sys v0.38.0 h1:3yZWxaJjBmCWXqhN1qh02AkOnCQ1poK6oF+a7xWL6Gc=
|
golang.org/x/sys v0.39.0 h1:CvCKL8MeisomCi6qNZ+wbb0DN9E5AATixKsvNtMoMFk=
|
||||||
golang.org/x/sys v0.38.0/go.mod h1:OgkHotnGiDImocRcuBABYBEXf8A9a87e/uXjp9XT3ks=
|
golang.org/x/sys v0.39.0/go.mod h1:OgkHotnGiDImocRcuBABYBEXf8A9a87e/uXjp9XT3ks=
|
||||||
golang.org/x/term v0.37.0 h1:8EGAD0qCmHYZg6J17DvsMy9/wJ7/D/4pV/wfnld5lTU=
|
golang.org/x/term v0.38.0 h1:PQ5pkm/rLO6HnxFR7N2lJHOZX6Kez5Y1gDSJla6jo7Q=
|
||||||
golang.org/x/term v0.37.0/go.mod h1:5pB4lxRNYYVZuTLmy8oR2BH8dflOR+IbTYFD8fi3254=
|
golang.org/x/term v0.38.0/go.mod h1:bSEAKrOT1W+VSu9TSCMtoGEOUcKxOKgl3LE5QEF/xVg=
|
||||||
gopkg.in/check.v1 v0.0.0-20161208181325-20d25e280405 h1:yhCVgyC4o1eVCa2tZl7eS0r+SDo693bJlVdllGtEeKM=
|
gopkg.in/check.v1 v0.0.0-20161208181325-20d25e280405 h1:yhCVgyC4o1eVCa2tZl7eS0r+SDo693bJlVdllGtEeKM=
|
||||||
gopkg.in/check.v1 v0.0.0-20161208181325-20d25e280405/go.mod h1:Co6ibVJAznAaIkqp8huTwlJQCZ016jof/cbN4VW5Yz0=
|
gopkg.in/check.v1 v0.0.0-20161208181325-20d25e280405/go.mod h1:Co6ibVJAznAaIkqp8huTwlJQCZ016jof/cbN4VW5Yz0=
|
||||||
gopkg.in/yaml.v3 v3.0.1 h1:fxVm/GzAzEWqLHuvctI91KS9hhNmmWOoWu0XTYJS7CA=
|
gopkg.in/yaml.v3 v3.0.1 h1:fxVm/GzAzEWqLHuvctI91KS9hhNmmWOoWu0XTYJS7CA=
|
||||||
|
|||||||
@@ -1043,7 +1043,7 @@
|
|||||||
"actionDeleteSite": "Standort löschen",
|
"actionDeleteSite": "Standort löschen",
|
||||||
"actionGetSite": "Standort abrufen",
|
"actionGetSite": "Standort abrufen",
|
||||||
"actionListSites": "Standorte auflisten",
|
"actionListSites": "Standorte auflisten",
|
||||||
"actionApplyBlueprint": "Blaupause anwenden",
|
"actionApplyBlueprint": "Blueprint anwenden",
|
||||||
"setupToken": "Setup-Token",
|
"setupToken": "Setup-Token",
|
||||||
"setupTokenDescription": "Geben Sie das Setup-Token von der Serverkonsole ein.",
|
"setupTokenDescription": "Geben Sie das Setup-Token von der Serverkonsole ein.",
|
||||||
"setupTokenRequired": "Setup-Token ist erforderlich",
|
"setupTokenRequired": "Setup-Token ist erforderlich",
|
||||||
@@ -1102,7 +1102,7 @@
|
|||||||
"actionDeleteIdpOrg": "IDP-Organisationsrichtlinie löschen",
|
"actionDeleteIdpOrg": "IDP-Organisationsrichtlinie löschen",
|
||||||
"actionListIdpOrgs": "IDP-Organisationen auflisten",
|
"actionListIdpOrgs": "IDP-Organisationen auflisten",
|
||||||
"actionUpdateIdpOrg": "IDP-Organisation aktualisieren",
|
"actionUpdateIdpOrg": "IDP-Organisation aktualisieren",
|
||||||
"actionCreateClient": "Endgerät anlegen",
|
"actionCreateClient": "Client erstellen",
|
||||||
"actionDeleteClient": "Client löschen",
|
"actionDeleteClient": "Client löschen",
|
||||||
"actionUpdateClient": "Client aktualisieren",
|
"actionUpdateClient": "Client aktualisieren",
|
||||||
"actionListClients": "Clients auflisten",
|
"actionListClients": "Clients auflisten",
|
||||||
@@ -1201,24 +1201,24 @@
|
|||||||
"sidebarLogsAnalytics": "Analytik",
|
"sidebarLogsAnalytics": "Analytik",
|
||||||
"blueprints": "Baupläne",
|
"blueprints": "Baupläne",
|
||||||
"blueprintsDescription": "Deklarative Konfigurationen anwenden und vorherige Abläufe anzeigen",
|
"blueprintsDescription": "Deklarative Konfigurationen anwenden und vorherige Abläufe anzeigen",
|
||||||
"blueprintAdd": "Blaupause hinzufügen",
|
"blueprintAdd": "Blueprint hinzufügen",
|
||||||
"blueprintGoBack": "Alle Blaupausen ansehen",
|
"blueprintGoBack": "Alle Blueprints ansehen",
|
||||||
"blueprintCreate": "Blaupause erstellen",
|
"blueprintCreate": "Blueprint erstellen",
|
||||||
"blueprintCreateDescription2": "Folge den Schritten unten, um eine neue Blaupause zu erstellen und anzuwenden",
|
"blueprintCreateDescription2": "Folge den unten aufgeführten Schritten, um einen neuen Blueprint zu erstellen und anzuwenden",
|
||||||
"blueprintDetails": "Blaupausendetails",
|
"blueprintDetails": "Blueprint Detailinformationen",
|
||||||
"blueprintDetailsDescription": "Siehe das Ergebnis der angewendeten Blaupause und alle aufgetretenen Fehler",
|
"blueprintDetailsDescription": "Siehe das Ergebnis des angewendeten Blueprints und alle aufgetretenen Fehler",
|
||||||
"blueprintInfo": "Blaupauseninformation",
|
"blueprintInfo": "Blueprint Informationen",
|
||||||
"message": "Nachricht",
|
"message": "Nachricht",
|
||||||
"blueprintContentsDescription": "Den YAML-Inhalt definieren, der die Infrastruktur beschreibt",
|
"blueprintContentsDescription": "Den YAML-Inhalt definieren, der die Infrastruktur beschreibt",
|
||||||
"blueprintErrorCreateDescription": "Fehler beim Anwenden der Blaupause",
|
"blueprintErrorCreateDescription": "Fehler beim Anwenden des Blueprints",
|
||||||
"blueprintErrorCreate": "Fehler beim Erstellen der Blaupause",
|
"blueprintErrorCreate": "Fehler beim Erstellen des Blueprints",
|
||||||
"searchBlueprintProgress": "Blaupausen suchen...",
|
"searchBlueprintProgress": "Blueprints suchen...",
|
||||||
"appliedAt": "Angewandt am",
|
"appliedAt": "Angewandt am",
|
||||||
"source": "Quelle",
|
"source": "Quelle",
|
||||||
"contents": "Inhalt",
|
"contents": "Inhalt",
|
||||||
"parsedContents": "Analysierte Inhalte (Nur lesen)",
|
"parsedContents": "Analysierte Inhalte (Nur lesen)",
|
||||||
"enableDockerSocket": "Docker Blaupause aktivieren",
|
"enableDockerSocket": "Docker Blueprint aktivieren",
|
||||||
"enableDockerSocketDescription": "Aktiviere Docker-Socket-Label-Scraping für Blaupausenbeschriftungen. Der Socket-Pfad muss neu angegeben werden.",
|
"enableDockerSocketDescription": "Aktiviere Docker-Socket-Label-Scraping für Blueprintbeschriftungen. Der Socket-Pfad muss neu angegeben werden.",
|
||||||
"enableDockerSocketLink": "Mehr erfahren",
|
"enableDockerSocketLink": "Mehr erfahren",
|
||||||
"viewDockerContainers": "Docker Container anzeigen",
|
"viewDockerContainers": "Docker Container anzeigen",
|
||||||
"containersIn": "Container in {siteName}",
|
"containersIn": "Container in {siteName}",
|
||||||
@@ -1543,7 +1543,7 @@
|
|||||||
"healthCheckPathRequired": "Gesundheits-Check-Pfad ist erforderlich",
|
"healthCheckPathRequired": "Gesundheits-Check-Pfad ist erforderlich",
|
||||||
"healthCheckMethodRequired": "HTTP-Methode ist erforderlich",
|
"healthCheckMethodRequired": "HTTP-Methode ist erforderlich",
|
||||||
"healthCheckIntervalMin": "Prüfintervall muss mindestens 5 Sekunden betragen",
|
"healthCheckIntervalMin": "Prüfintervall muss mindestens 5 Sekunden betragen",
|
||||||
"healthCheckTimeoutMin": "Timeout muss mindestens 1 Sekunde betragen",
|
"healthCheckTimeoutMin": "Zeitüberschreitung muss mindestens 1 Sekunde betragen",
|
||||||
"healthCheckRetryMin": "Wiederholungsversuche müssen mindestens 1 betragen",
|
"healthCheckRetryMin": "Wiederholungsversuche müssen mindestens 1 betragen",
|
||||||
"httpMethod": "HTTP-Methode",
|
"httpMethod": "HTTP-Methode",
|
||||||
"selectHttpMethod": "HTTP-Methode auswählen",
|
"selectHttpMethod": "HTTP-Methode auswählen",
|
||||||
|
|||||||
@@ -2067,6 +2067,8 @@
|
|||||||
"timestamp": "Timestamp",
|
"timestamp": "Timestamp",
|
||||||
"accessLogs": "Access Logs",
|
"accessLogs": "Access Logs",
|
||||||
"exportCsv": "Export CSV",
|
"exportCsv": "Export CSV",
|
||||||
|
"exportError": "Unknown error when exporting CSV",
|
||||||
|
"exportCsvTooltip": "Within Time Range",
|
||||||
"actorId": "Actor ID",
|
"actorId": "Actor ID",
|
||||||
"allowedByRule": "Allowed by Rule",
|
"allowedByRule": "Allowed by Rule",
|
||||||
"allowedNoAuth": "Allowed No Auth",
|
"allowedNoAuth": "Allowed No Auth",
|
||||||
|
|||||||
4397
package-lock.json
generated
157
package.json
@@ -29,16 +29,17 @@
|
|||||||
"build:pg": "mkdir -p dist && next build && node esbuild.mjs -e server/index.ts -o dist/server.mjs && node esbuild.mjs -e server/setup/migrationsPg.ts -o dist/migrations.mjs",
|
"build:pg": "mkdir -p dist && next build && node esbuild.mjs -e server/index.ts -o dist/server.mjs && node esbuild.mjs -e server/setup/migrationsPg.ts -o dist/migrations.mjs",
|
||||||
"start": "ENVIRONMENT=prod node dist/migrations.mjs && ENVIRONMENT=prod NODE_ENV=development node --enable-source-maps dist/server.mjs",
|
"start": "ENVIRONMENT=prod node dist/migrations.mjs && ENVIRONMENT=prod NODE_ENV=development node --enable-source-maps dist/server.mjs",
|
||||||
"email": "email dev --dir server/emails/templates --port 3005",
|
"email": "email dev --dir server/emails/templates --port 3005",
|
||||||
"build:cli": "node esbuild.mjs -e cli/index.ts -o dist/cli.mjs"
|
"build:cli": "node esbuild.mjs -e cli/index.ts -o dist/cli.mjs",
|
||||||
|
"format": "prettier --write ."
|
||||||
},
|
},
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"@asteasolutions/zod-to-openapi": "8.1.0",
|
"@asteasolutions/zod-to-openapi": "8.2.0",
|
||||||
"@faker-js/faker": "^10.1.0",
|
"@aws-sdk/client-s3": "3.948.0",
|
||||||
"@headlessui/react": "^2.2.9",
|
"@faker-js/faker": "10.1.0",
|
||||||
"@aws-sdk/client-s3": "3.943.0",
|
"@headlessui/react": "2.2.9",
|
||||||
"@hookform/resolvers": "5.2.2",
|
"@hookform/resolvers": "5.2.2",
|
||||||
"@monaco-editor/react": "^4.7.0",
|
"@monaco-editor/react": "4.7.0",
|
||||||
"@node-rs/argon2": "^2.0.2",
|
"@node-rs/argon2": "2.0.2",
|
||||||
"@oslojs/crypto": "1.0.1",
|
"@oslojs/crypto": "1.0.1",
|
||||||
"@oslojs/encoding": "1.1.0",
|
"@oslojs/encoding": "1.1.0",
|
||||||
"@radix-ui/react-avatar": "1.1.11",
|
"@radix-ui/react-avatar": "1.1.11",
|
||||||
@@ -49,138 +50,132 @@
|
|||||||
"@radix-ui/react-icons": "1.3.2",
|
"@radix-ui/react-icons": "1.3.2",
|
||||||
"@radix-ui/react-label": "2.1.8",
|
"@radix-ui/react-label": "2.1.8",
|
||||||
"@radix-ui/react-popover": "1.1.15",
|
"@radix-ui/react-popover": "1.1.15",
|
||||||
"@radix-ui/react-progress": "^1.1.8",
|
"@radix-ui/react-progress": "1.1.8",
|
||||||
"@radix-ui/react-radio-group": "1.3.8",
|
"@radix-ui/react-radio-group": "1.3.8",
|
||||||
"@radix-ui/react-scroll-area": "^1.2.10",
|
"@radix-ui/react-scroll-area": "1.2.10",
|
||||||
"@radix-ui/react-select": "2.2.6",
|
"@radix-ui/react-select": "2.2.6",
|
||||||
"@radix-ui/react-separator": "1.1.8",
|
"@radix-ui/react-separator": "1.1.8",
|
||||||
"@radix-ui/react-slot": "1.2.4",
|
"@radix-ui/react-slot": "1.2.4",
|
||||||
"@radix-ui/react-switch": "1.2.6",
|
"@radix-ui/react-switch": "1.2.6",
|
||||||
"@radix-ui/react-tabs": "1.1.13",
|
"@radix-ui/react-tabs": "1.1.13",
|
||||||
"@radix-ui/react-toast": "1.2.15",
|
"@radix-ui/react-toast": "1.2.15",
|
||||||
"@radix-ui/react-tooltip": "^1.2.8",
|
"@radix-ui/react-tooltip": "1.2.8",
|
||||||
"@react-email/components": "0.5.7",
|
"@react-email/components": "1.0.1",
|
||||||
"@react-email/render": "^1.3.2",
|
"@react-email/render": "2.0.0",
|
||||||
"@react-email/tailwind": "1.2.2",
|
"@react-email/tailwind": "2.0.1",
|
||||||
"@simplewebauthn/browser": "^13.2.2",
|
"@simplewebauthn/browser": "13.2.2",
|
||||||
"@simplewebauthn/server": "^13.2.2",
|
"@simplewebauthn/server": "13.2.2",
|
||||||
"@tailwindcss/forms": "^0.5.10",
|
"@tailwindcss/forms": "0.5.10",
|
||||||
"@tanstack/react-query": "^5.90.6",
|
"@tanstack/react-query": "5.90.12",
|
||||||
"@tanstack/react-table": "8.21.3",
|
"@tanstack/react-table": "8.21.3",
|
||||||
"arctic": "^3.7.0",
|
"arctic": "3.7.0",
|
||||||
"axios": "^1.13.2",
|
"axios": "1.13.2",
|
||||||
"better-sqlite3": "11.7.0",
|
"better-sqlite3": "11.9.1",
|
||||||
"canvas-confetti": "1.9.4",
|
"canvas-confetti": "1.9.4",
|
||||||
"class-variance-authority": "^0.7.1",
|
"class-variance-authority": "0.7.1",
|
||||||
"clsx": "2.1.1",
|
"clsx": "2.1.1",
|
||||||
"cmdk": "1.1.1",
|
"cmdk": "1.1.1",
|
||||||
"cookie": "^1.0.2",
|
"cookie": "1.1.1",
|
||||||
"cookie-parser": "1.4.7",
|
"cookie-parser": "1.4.7",
|
||||||
"cookies": "^0.9.1",
|
"cookies": "0.9.1",
|
||||||
"cors": "2.8.5",
|
"cors": "2.8.5",
|
||||||
"crypto-js": "^4.2.0",
|
"crypto-js": "4.2.0",
|
||||||
"d3": "^7.9.0",
|
"d3": "7.9.0",
|
||||||
"date-fns": "4.1.0",
|
"date-fns": "4.1.0",
|
||||||
"drizzle-orm": "0.45.0",
|
"drizzle-orm": "0.45.0",
|
||||||
"eslint": "9.39.1",
|
"eslint": "9.39.1",
|
||||||
"eslint-config-next": "16.0.7",
|
"eslint-config-next": "16.0.8",
|
||||||
"express": "5.2.1",
|
"express": "5.2.1",
|
||||||
"express-rate-limit": "8.2.1",
|
"express-rate-limit": "8.2.1",
|
||||||
"glob": "11.1.0",
|
"glob": "13.0.0",
|
||||||
"helmet": "8.1.0",
|
"helmet": "8.1.0",
|
||||||
"http-errors": "2.0.1",
|
"http-errors": "2.0.1",
|
||||||
"i": "^0.3.7",
|
"i": "0.3.7",
|
||||||
"input-otp": "1.4.2",
|
"input-otp": "1.4.2",
|
||||||
"ioredis": "5.8.2",
|
"ioredis": "5.8.2",
|
||||||
"jmespath": "^0.16.0",
|
"jmespath": "0.16.0",
|
||||||
"js-yaml": "4.1.1",
|
"js-yaml": "4.1.1",
|
||||||
"jsonwebtoken": "^9.0.2",
|
"jsonwebtoken": "9.0.3",
|
||||||
"lucide-react": "^0.556.0",
|
"lucide-react": "0.559.0",
|
||||||
"maxmind": "5.0.1",
|
"maxmind": "5.0.1",
|
||||||
"moment": "2.30.1",
|
"moment": "2.30.1",
|
||||||
"next": "15.5.7",
|
"next": "15.5.7",
|
||||||
"next-intl": "^4.4.0",
|
"next-intl": "4.5.8",
|
||||||
"next-themes": "0.4.6",
|
"next-themes": "0.4.6",
|
||||||
"nextjs-toploader": "^3.9.17",
|
"nextjs-toploader": "3.9.17",
|
||||||
"node-cache": "5.1.2",
|
"node-cache": "5.1.2",
|
||||||
"node-fetch": "3.3.2",
|
"node-fetch": "3.3.2",
|
||||||
"nodemailer": "7.0.11",
|
"nodemailer": "7.0.11",
|
||||||
"npm": "^11.6.4",
|
"npm": "11.7.0",
|
||||||
"nprogress": "^0.2.0",
|
"nprogress": "0.2.0",
|
||||||
"oslo": "1.2.1",
|
"oslo": "1.2.1",
|
||||||
"pg": "^8.16.2",
|
"pg": "8.16.3",
|
||||||
"posthog-node": "^5.11.2",
|
"posthog-node": "5.17.2",
|
||||||
"qrcode.react": "4.2.0",
|
"qrcode.react": "4.2.0",
|
||||||
"react": "19.2.1",
|
"react": "19.2.1",
|
||||||
"react-day-picker": "9.11.3",
|
"react-day-picker": "9.12.0",
|
||||||
"react-dom": "19.2.1",
|
"react-dom": "19.2.1",
|
||||||
"react-easy-sort": "^1.8.0",
|
"react-easy-sort": "1.8.0",
|
||||||
"react-hook-form": "7.68.0",
|
"react-hook-form": "7.68.0",
|
||||||
"react-icons": "^5.5.0",
|
"react-icons": "5.5.0",
|
||||||
"rebuild": "0.1.2",
|
"rebuild": "0.1.2",
|
||||||
"recharts": "^2.15.4",
|
"recharts": "2.15.4",
|
||||||
"reodotdev": "^1.0.0",
|
"reodotdev": "1.0.0",
|
||||||
"resend": "^6.4.2",
|
"resend": "6.6.0",
|
||||||
"semver": "^7.7.3",
|
"semver": "7.7.3",
|
||||||
"stripe": "18.2.1",
|
"stripe": "20.0.0",
|
||||||
"swagger-ui-express": "^5.0.1",
|
"swagger-ui-express": "5.0.1",
|
||||||
"topojson-client": "^3.1.0",
|
|
||||||
"tailwind-merge": "3.4.0",
|
"tailwind-merge": "3.4.0",
|
||||||
"tw-animate-css": "^1.3.8",
|
"topojson-client": "3.1.0",
|
||||||
"uuid": "^13.0.0",
|
"tw-animate-css": "1.4.0",
|
||||||
|
"uuid": "13.0.0",
|
||||||
"vaul": "1.1.2",
|
"vaul": "1.1.2",
|
||||||
"visionscarto-world-atlas": "^1.0.0",
|
"visionscarto-world-atlas": "1.0.0",
|
||||||
"winston": "3.18.3",
|
"winston": "3.19.0",
|
||||||
"winston-daily-rotate-file": "5.0.0",
|
"winston-daily-rotate-file": "5.0.0",
|
||||||
"ws": "8.18.3",
|
"ws": "8.18.3",
|
||||||
"yaml": "^2.8.1",
|
"yaml": "2.8.2",
|
||||||
"yargs": "18.0.0",
|
"yargs": "18.0.0",
|
||||||
"zod": "4.1.12",
|
"zod": "4.1.13",
|
||||||
"zod-validation-error": "5.0.0"
|
"zod-validation-error": "5.0.0"
|
||||||
},
|
},
|
||||||
"devDependencies": {
|
"devDependencies": {
|
||||||
"@dotenvx/dotenvx": "1.51.1",
|
"@dotenvx/dotenvx": "1.51.1",
|
||||||
"@esbuild-plugins/tsconfig-paths": "0.1.2",
|
"@esbuild-plugins/tsconfig-paths": "0.1.2",
|
||||||
"@react-email/preview-server": "4.3.2",
|
"@tailwindcss/postcss": "4.1.17",
|
||||||
"@tailwindcss/postcss": "^4.1.17",
|
"@tanstack/react-query-devtools": "5.91.1",
|
||||||
"@tanstack/react-query-devtools": "^5.90.2",
|
"@types/better-sqlite3": "7.6.13",
|
||||||
"@types/better-sqlite3": "7.6.12",
|
|
||||||
"@types/cookie-parser": "1.4.10",
|
"@types/cookie-parser": "1.4.10",
|
||||||
"@types/cors": "2.8.19",
|
"@types/cors": "2.8.19",
|
||||||
"@types/crypto-js": "^4.2.2",
|
"@types/crypto-js": "4.2.2",
|
||||||
"@types/d3": "^7.4.3",
|
"@types/d3": "7.4.3",
|
||||||
"@types/express": "5.0.6",
|
"@types/express": "5.0.6",
|
||||||
"@types/express-session": "^1.18.2",
|
"@types/express-session": "1.18.2",
|
||||||
"@types/jmespath": "^0.15.2",
|
"@types/jmespath": "0.15.2",
|
||||||
"@types/js-yaml": "4.0.9",
|
"@types/jsonwebtoken": "9.0.10",
|
||||||
"@types/jsonwebtoken": "^9.0.10",
|
"@types/node": "24.10.2",
|
||||||
"@types/node": "24.10.1",
|
|
||||||
"@types/nprogress": "^0.2.3",
|
|
||||||
"@types/nodemailer": "7.0.4",
|
"@types/nodemailer": "7.0.4",
|
||||||
"@types/pg": "8.15.6",
|
"@types/nprogress": "0.2.3",
|
||||||
|
"@types/pg": "8.16.0",
|
||||||
"@types/react": "19.2.7",
|
"@types/react": "19.2.7",
|
||||||
"@types/react-dom": "19.2.3",
|
"@types/react-dom": "19.2.3",
|
||||||
"@types/semver": "^7.7.1",
|
"@types/semver": "7.7.1",
|
||||||
"@types/swagger-ui-express": "^4.1.8",
|
"@types/swagger-ui-express": "4.1.8",
|
||||||
"@types/topojson-client": "^3.1.5",
|
"@types/topojson-client": "3.1.5",
|
||||||
"@types/ws": "8.18.1",
|
"@types/ws": "8.18.1",
|
||||||
"babel-plugin-react-compiler": "^1.0.0",
|
|
||||||
"@types/yargs": "17.0.35",
|
"@types/yargs": "17.0.35",
|
||||||
|
"@types/js-yaml": "4.0.9",
|
||||||
|
"babel-plugin-react-compiler": "1.0.0",
|
||||||
"drizzle-kit": "0.31.8",
|
"drizzle-kit": "0.31.8",
|
||||||
"esbuild": "0.27.1",
|
"esbuild": "0.27.1",
|
||||||
"esbuild-node-externals": "1.20.1",
|
"esbuild-node-externals": "1.20.1",
|
||||||
"postcss": "^8",
|
"postcss": "8.5.6",
|
||||||
"react-email": "4.3.2",
|
"prettier": "3.7.4",
|
||||||
"tailwindcss": "^4.1.4",
|
"react-email": "5.0.7",
|
||||||
|
"tailwindcss": "4.1.17",
|
||||||
"tsc-alias": "1.8.16",
|
"tsc-alias": "1.8.16",
|
||||||
"tsx": "4.21.0",
|
"tsx": "4.21.0",
|
||||||
"typescript": "^5",
|
"typescript": "5.9.3",
|
||||||
"typescript-eslint": "^8.46.3"
|
"typescript-eslint": "8.49.0"
|
||||||
},
|
|
||||||
"overrides": {
|
|
||||||
"emblor": {
|
|
||||||
"react": "19.0.0",
|
|
||||||
"react-dom": "19.0.0"
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -1,8 +1,8 @@
|
|||||||
/** @type {import('postcss-load-config').Config} */
|
/** @type {import('postcss-load-config').Config} */
|
||||||
const config = {
|
const config = {
|
||||||
plugins: {
|
plugins: {
|
||||||
"@tailwindcss/postcss": {},
|
"@tailwindcss/postcss": {}
|
||||||
},
|
}
|
||||||
};
|
};
|
||||||
|
|
||||||
export default config;
|
export default config;
|
||||||
|
|||||||
|
Before Width: | Height: | Size: 687 KiB |
|
Before Width: | Height: | Size: 713 KiB After Width: | Height: | Size: 493 KiB |
|
Before Width: | Height: | Size: 636 KiB |
|
Before Width: | Height: | Size: 713 KiB After Width: | Height: | Size: 484 KiB |
BIN
public/screenshots/private-resources.png
Normal file
|
After Width: | Height: | Size: 421 KiB |
BIN
public/screenshots/public-resources.png
Normal file
|
After Width: | Height: | Size: 484 KiB |
|
Before Width: | Height: | Size: 713 KiB |
|
Before Width: | Height: | Size: 456 KiB |
|
Before Width: | Height: | Size: 674 KiB After Width: | Height: | Size: 396 KiB |
BIN
public/screenshots/user-devices.png
Normal file
|
After Width: | Height: | Size: 434 KiB |
@@ -2,13 +2,13 @@ import { hash, verify } from "@node-rs/argon2";
|
|||||||
|
|
||||||
export async function verifyPassword(
|
export async function verifyPassword(
|
||||||
password: string,
|
password: string,
|
||||||
hash: string,
|
hash: string
|
||||||
): Promise<boolean> {
|
): Promise<boolean> {
|
||||||
const validPassword = await verify(hash, password, {
|
const validPassword = await verify(hash, password, {
|
||||||
memoryCost: 19456,
|
memoryCost: 19456,
|
||||||
timeCost: 2,
|
timeCost: 2,
|
||||||
outputLen: 32,
|
outputLen: 32,
|
||||||
parallelism: 1,
|
parallelism: 1
|
||||||
});
|
});
|
||||||
return validPassword;
|
return validPassword;
|
||||||
}
|
}
|
||||||
@@ -18,7 +18,7 @@ export async function hashPassword(password: string): Promise<string> {
|
|||||||
memoryCost: 19456,
|
memoryCost: 19456,
|
||||||
timeCost: 2,
|
timeCost: 2,
|
||||||
outputLen: 32,
|
outputLen: 32,
|
||||||
parallelism: 1,
|
parallelism: 1
|
||||||
});
|
});
|
||||||
|
|
||||||
return passwordHash;
|
return passwordHash;
|
||||||
|
|||||||
@@ -4,10 +4,13 @@ export const passwordSchema = z
|
|||||||
.string()
|
.string()
|
||||||
.min(8, { message: "Password must be at least 8 characters long" })
|
.min(8, { message: "Password must be at least 8 characters long" })
|
||||||
.max(128, { message: "Password must be at most 128 characters long" })
|
.max(128, { message: "Password must be at most 128 characters long" })
|
||||||
.regex(/^(?=.*?[A-Z])(?=.*?[a-z])(?=.*?[0-9])(?=.*?[~!`@#$%^&*()_\-+={}[\]|\\:;"'<>,.\/?]).*$/, {
|
.regex(
|
||||||
message: `Your password must meet the following conditions:
|
/^(?=.*?[A-Z])(?=.*?[a-z])(?=.*?[0-9])(?=.*?[~!`@#$%^&*()_\-+={}[\]|\\:;"'<>,.\/?]).*$/,
|
||||||
|
{
|
||||||
|
message: `Your password must meet the following conditions:
|
||||||
at least one uppercase English letter,
|
at least one uppercase English letter,
|
||||||
at least one lowercase English letter,
|
at least one lowercase English letter,
|
||||||
at least one digit,
|
at least one digit,
|
||||||
at least one special character.`
|
at least one special character.`
|
||||||
});
|
}
|
||||||
|
);
|
||||||
|
|||||||
@@ -1,6 +1,4 @@
|
|||||||
import {
|
import { encodeHexLowerCase } from "@oslojs/encoding";
|
||||||
encodeHexLowerCase,
|
|
||||||
} from "@oslojs/encoding";
|
|
||||||
import { sha256 } from "@oslojs/crypto/sha2";
|
import { sha256 } from "@oslojs/crypto/sha2";
|
||||||
import { Newt, newts, newtSessions, NewtSession } from "@server/db";
|
import { Newt, newts, newtSessions, NewtSession } from "@server/db";
|
||||||
import { db } from "@server/db";
|
import { db } from "@server/db";
|
||||||
@@ -10,25 +8,25 @@ export const EXPIRES = 1000 * 60 * 60 * 24 * 30;
|
|||||||
|
|
||||||
export async function createNewtSession(
|
export async function createNewtSession(
|
||||||
token: string,
|
token: string,
|
||||||
newtId: string,
|
newtId: string
|
||||||
): Promise<NewtSession> {
|
): Promise<NewtSession> {
|
||||||
const sessionId = encodeHexLowerCase(
|
const sessionId = encodeHexLowerCase(
|
||||||
sha256(new TextEncoder().encode(token)),
|
sha256(new TextEncoder().encode(token))
|
||||||
);
|
);
|
||||||
const session: NewtSession = {
|
const session: NewtSession = {
|
||||||
sessionId: sessionId,
|
sessionId: sessionId,
|
||||||
newtId,
|
newtId,
|
||||||
expiresAt: new Date(Date.now() + EXPIRES).getTime(),
|
expiresAt: new Date(Date.now() + EXPIRES).getTime()
|
||||||
};
|
};
|
||||||
await db.insert(newtSessions).values(session);
|
await db.insert(newtSessions).values(session);
|
||||||
return session;
|
return session;
|
||||||
}
|
}
|
||||||
|
|
||||||
export async function validateNewtSessionToken(
|
export async function validateNewtSessionToken(
|
||||||
token: string,
|
token: string
|
||||||
): Promise<SessionValidationResult> {
|
): Promise<SessionValidationResult> {
|
||||||
const sessionId = encodeHexLowerCase(
|
const sessionId = encodeHexLowerCase(
|
||||||
sha256(new TextEncoder().encode(token)),
|
sha256(new TextEncoder().encode(token))
|
||||||
);
|
);
|
||||||
const result = await db
|
const result = await db
|
||||||
.select({ newt: newts, session: newtSessions })
|
.select({ newt: newts, session: newtSessions })
|
||||||
@@ -45,14 +43,12 @@ export async function validateNewtSessionToken(
|
|||||||
.where(eq(newtSessions.sessionId, session.sessionId));
|
.where(eq(newtSessions.sessionId, session.sessionId));
|
||||||
return { session: null, newt: null };
|
return { session: null, newt: null };
|
||||||
}
|
}
|
||||||
if (Date.now() >= session.expiresAt - (EXPIRES / 2)) {
|
if (Date.now() >= session.expiresAt - EXPIRES / 2) {
|
||||||
session.expiresAt = new Date(
|
session.expiresAt = new Date(Date.now() + EXPIRES).getTime();
|
||||||
Date.now() + EXPIRES,
|
|
||||||
).getTime();
|
|
||||||
await db
|
await db
|
||||||
.update(newtSessions)
|
.update(newtSessions)
|
||||||
.set({
|
.set({
|
||||||
expiresAt: session.expiresAt,
|
expiresAt: session.expiresAt
|
||||||
})
|
})
|
||||||
.where(eq(newtSessions.sessionId, session.sessionId));
|
.where(eq(newtSessions.sessionId, session.sessionId));
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,6 +1,4 @@
|
|||||||
import {
|
import { encodeHexLowerCase } from "@oslojs/encoding";
|
||||||
encodeHexLowerCase,
|
|
||||||
} from "@oslojs/encoding";
|
|
||||||
import { sha256 } from "@oslojs/crypto/sha2";
|
import { sha256 } from "@oslojs/crypto/sha2";
|
||||||
import { Olm, olms, olmSessions, OlmSession } from "@server/db";
|
import { Olm, olms, olmSessions, OlmSession } from "@server/db";
|
||||||
import { db } from "@server/db";
|
import { db } from "@server/db";
|
||||||
@@ -10,25 +8,25 @@ export const EXPIRES = 1000 * 60 * 60 * 24 * 30;
|
|||||||
|
|
||||||
export async function createOlmSession(
|
export async function createOlmSession(
|
||||||
token: string,
|
token: string,
|
||||||
olmId: string,
|
olmId: string
|
||||||
): Promise<OlmSession> {
|
): Promise<OlmSession> {
|
||||||
const sessionId = encodeHexLowerCase(
|
const sessionId = encodeHexLowerCase(
|
||||||
sha256(new TextEncoder().encode(token)),
|
sha256(new TextEncoder().encode(token))
|
||||||
);
|
);
|
||||||
const session: OlmSession = {
|
const session: OlmSession = {
|
||||||
sessionId: sessionId,
|
sessionId: sessionId,
|
||||||
olmId,
|
olmId,
|
||||||
expiresAt: new Date(Date.now() + EXPIRES).getTime(),
|
expiresAt: new Date(Date.now() + EXPIRES).getTime()
|
||||||
};
|
};
|
||||||
await db.insert(olmSessions).values(session);
|
await db.insert(olmSessions).values(session);
|
||||||
return session;
|
return session;
|
||||||
}
|
}
|
||||||
|
|
||||||
export async function validateOlmSessionToken(
|
export async function validateOlmSessionToken(
|
||||||
token: string,
|
token: string
|
||||||
): Promise<SessionValidationResult> {
|
): Promise<SessionValidationResult> {
|
||||||
const sessionId = encodeHexLowerCase(
|
const sessionId = encodeHexLowerCase(
|
||||||
sha256(new TextEncoder().encode(token)),
|
sha256(new TextEncoder().encode(token))
|
||||||
);
|
);
|
||||||
const result = await db
|
const result = await db
|
||||||
.select({ olm: olms, session: olmSessions })
|
.select({ olm: olms, session: olmSessions })
|
||||||
@@ -45,14 +43,12 @@ export async function validateOlmSessionToken(
|
|||||||
.where(eq(olmSessions.sessionId, session.sessionId));
|
.where(eq(olmSessions.sessionId, session.sessionId));
|
||||||
return { session: null, olm: null };
|
return { session: null, olm: null };
|
||||||
}
|
}
|
||||||
if (Date.now() >= session.expiresAt - (EXPIRES / 2)) {
|
if (Date.now() >= session.expiresAt - EXPIRES / 2) {
|
||||||
session.expiresAt = new Date(
|
session.expiresAt = new Date(Date.now() + EXPIRES).getTime();
|
||||||
Date.now() + EXPIRES,
|
|
||||||
).getTime();
|
|
||||||
await db
|
await db
|
||||||
.update(olmSessions)
|
.update(olmSessions)
|
||||||
.set({
|
.set({
|
||||||
expiresAt: session.expiresAt,
|
expiresAt: session.expiresAt
|
||||||
})
|
})
|
||||||
.where(eq(olmSessions.sessionId, session.sessionId));
|
.where(eq(olmSessions.sessionId, session.sessionId));
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -10,4 +10,4 @@ export async function initCleanup() {
|
|||||||
// Handle process termination
|
// Handle process termination
|
||||||
process.on("SIGTERM", () => cleanup());
|
process.on("SIGTERM", () => cleanup());
|
||||||
process.on("SIGINT", () => cleanup());
|
process.on("SIGINT", () => cleanup());
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1708,4 +1708,4 @@
|
|||||||
"Desert Box Turtle",
|
"Desert Box Turtle",
|
||||||
"African Striped Weasel"
|
"African Striped Weasel"
|
||||||
]
|
]
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -215,42 +215,56 @@ export const sessionTransferToken = pgTable("sessionTransferToken", {
|
|||||||
expiresAt: bigint("expiresAt", { mode: "number" }).notNull()
|
expiresAt: bigint("expiresAt", { mode: "number" }).notNull()
|
||||||
});
|
});
|
||||||
|
|
||||||
export const actionAuditLog = pgTable("actionAuditLog", {
|
export const actionAuditLog = pgTable(
|
||||||
id: serial("id").primaryKey(),
|
"actionAuditLog",
|
||||||
timestamp: bigint("timestamp", { mode: "number" }).notNull(), // this is EPOCH time in seconds
|
{
|
||||||
orgId: varchar("orgId")
|
id: serial("id").primaryKey(),
|
||||||
.notNull()
|
timestamp: bigint("timestamp", { mode: "number" }).notNull(), // this is EPOCH time in seconds
|
||||||
.references(() => orgs.orgId, { onDelete: "cascade" }),
|
orgId: varchar("orgId")
|
||||||
actorType: varchar("actorType", { length: 50 }).notNull(),
|
.notNull()
|
||||||
actor: varchar("actor", { length: 255 }).notNull(),
|
.references(() => orgs.orgId, { onDelete: "cascade" }),
|
||||||
actorId: varchar("actorId", { length: 255 }).notNull(),
|
actorType: varchar("actorType", { length: 50 }).notNull(),
|
||||||
action: varchar("action", { length: 100 }).notNull(),
|
actor: varchar("actor", { length: 255 }).notNull(),
|
||||||
metadata: text("metadata")
|
actorId: varchar("actorId", { length: 255 }).notNull(),
|
||||||
}, (table) => ([
|
action: varchar("action", { length: 100 }).notNull(),
|
||||||
index("idx_actionAuditLog_timestamp").on(table.timestamp),
|
metadata: text("metadata")
|
||||||
index("idx_actionAuditLog_org_timestamp").on(table.orgId, table.timestamp)
|
},
|
||||||
]));
|
(table) => [
|
||||||
|
index("idx_actionAuditLog_timestamp").on(table.timestamp),
|
||||||
|
index("idx_actionAuditLog_org_timestamp").on(
|
||||||
|
table.orgId,
|
||||||
|
table.timestamp
|
||||||
|
)
|
||||||
|
]
|
||||||
|
);
|
||||||
|
|
||||||
export const accessAuditLog = pgTable("accessAuditLog", {
|
export const accessAuditLog = pgTable(
|
||||||
id: serial("id").primaryKey(),
|
"accessAuditLog",
|
||||||
timestamp: bigint("timestamp", { mode: "number" }).notNull(), // this is EPOCH time in seconds
|
{
|
||||||
orgId: varchar("orgId")
|
id: serial("id").primaryKey(),
|
||||||
.notNull()
|
timestamp: bigint("timestamp", { mode: "number" }).notNull(), // this is EPOCH time in seconds
|
||||||
.references(() => orgs.orgId, { onDelete: "cascade" }),
|
orgId: varchar("orgId")
|
||||||
actorType: varchar("actorType", { length: 50 }),
|
.notNull()
|
||||||
actor: varchar("actor", { length: 255 }),
|
.references(() => orgs.orgId, { onDelete: "cascade" }),
|
||||||
actorId: varchar("actorId", { length: 255 }),
|
actorType: varchar("actorType", { length: 50 }),
|
||||||
resourceId: integer("resourceId"),
|
actor: varchar("actor", { length: 255 }),
|
||||||
ip: varchar("ip", { length: 45 }),
|
actorId: varchar("actorId", { length: 255 }),
|
||||||
type: varchar("type", { length: 100 }).notNull(),
|
resourceId: integer("resourceId"),
|
||||||
action: boolean("action").notNull(),
|
ip: varchar("ip", { length: 45 }),
|
||||||
location: text("location"),
|
type: varchar("type", { length: 100 }).notNull(),
|
||||||
userAgent: text("userAgent"),
|
action: boolean("action").notNull(),
|
||||||
metadata: text("metadata")
|
location: text("location"),
|
||||||
}, (table) => ([
|
userAgent: text("userAgent"),
|
||||||
index("idx_identityAuditLog_timestamp").on(table.timestamp),
|
metadata: text("metadata")
|
||||||
index("idx_identityAuditLog_org_timestamp").on(table.orgId, table.timestamp)
|
},
|
||||||
]));
|
(table) => [
|
||||||
|
index("idx_identityAuditLog_timestamp").on(table.timestamp),
|
||||||
|
index("idx_identityAuditLog_org_timestamp").on(
|
||||||
|
table.orgId,
|
||||||
|
table.timestamp
|
||||||
|
)
|
||||||
|
]
|
||||||
|
);
|
||||||
|
|
||||||
export type Limit = InferSelectModel<typeof limits>;
|
export type Limit = InferSelectModel<typeof limits>;
|
||||||
export type Account = InferSelectModel<typeof account>;
|
export type Account = InferSelectModel<typeof account>;
|
||||||
@@ -270,4 +284,4 @@ export type RemoteExitNodeSession = InferSelectModel<
|
|||||||
export type ExitNodeOrg = InferSelectModel<typeof exitNodeOrgs>;
|
export type ExitNodeOrg = InferSelectModel<typeof exitNodeOrgs>;
|
||||||
export type LoginPage = InferSelectModel<typeof loginPage>;
|
export type LoginPage = InferSelectModel<typeof loginPage>;
|
||||||
export type ActionAuditLog = InferSelectModel<typeof actionAuditLog>;
|
export type ActionAuditLog = InferSelectModel<typeof actionAuditLog>;
|
||||||
export type AccessAuditLog = InferSelectModel<typeof accessAuditLog>;
|
export type AccessAuditLog = InferSelectModel<typeof accessAuditLog>;
|
||||||
|
|||||||
@@ -177,7 +177,7 @@ export const targetHealthCheck = pgTable("targetHealthCheck", {
|
|||||||
hcMethod: varchar("hcMethod").default("GET"),
|
hcMethod: varchar("hcMethod").default("GET"),
|
||||||
hcStatus: integer("hcStatus"), // http code
|
hcStatus: integer("hcStatus"), // http code
|
||||||
hcHealth: text("hcHealth").default("unknown"), // "unknown", "healthy", "unhealthy"
|
hcHealth: text("hcHealth").default("unknown"), // "unknown", "healthy", "unhealthy"
|
||||||
hcTlsServerName: text("hcTlsServerName"),
|
hcTlsServerName: text("hcTlsServerName")
|
||||||
});
|
});
|
||||||
|
|
||||||
export const exitNodes = pgTable("exitNodes", {
|
export const exitNodes = pgTable("exitNodes", {
|
||||||
|
|||||||
@@ -52,10 +52,7 @@ export async function getResourceByDomain(
|
|||||||
resourceHeaderAuth,
|
resourceHeaderAuth,
|
||||||
eq(resourceHeaderAuth.resourceId, resources.resourceId)
|
eq(resourceHeaderAuth.resourceId, resources.resourceId)
|
||||||
)
|
)
|
||||||
.innerJoin(
|
.innerJoin(orgs, eq(orgs.orgId, resources.orgId))
|
||||||
orgs,
|
|
||||||
eq(orgs.orgId, resources.orgId)
|
|
||||||
)
|
|
||||||
.where(eq(resources.fullDomain, domain))
|
.where(eq(resources.fullDomain, domain))
|
||||||
.limit(1);
|
.limit(1);
|
||||||
|
|
||||||
|
|||||||
@@ -8,7 +8,7 @@ const runMigrations = async () => {
|
|||||||
console.log("Running migrations...");
|
console.log("Running migrations...");
|
||||||
try {
|
try {
|
||||||
migrate(db as any, {
|
migrate(db as any, {
|
||||||
migrationsFolder: migrationsFolder,
|
migrationsFolder: migrationsFolder
|
||||||
});
|
});
|
||||||
console.log("Migrations completed successfully.");
|
console.log("Migrations completed successfully.");
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
|
|||||||
@@ -29,7 +29,9 @@ export const certificates = sqliteTable("certificates", {
|
|||||||
});
|
});
|
||||||
|
|
||||||
export const dnsChallenge = sqliteTable("dnsChallenges", {
|
export const dnsChallenge = sqliteTable("dnsChallenges", {
|
||||||
dnsChallengeId: integer("dnsChallengeId").primaryKey({ autoIncrement: true }),
|
dnsChallengeId: integer("dnsChallengeId").primaryKey({
|
||||||
|
autoIncrement: true
|
||||||
|
}),
|
||||||
domain: text("domain").notNull(),
|
domain: text("domain").notNull(),
|
||||||
token: text("token").notNull(),
|
token: text("token").notNull(),
|
||||||
keyAuthorization: text("keyAuthorization").notNull(),
|
keyAuthorization: text("keyAuthorization").notNull(),
|
||||||
@@ -61,9 +63,7 @@ export const customers = sqliteTable("customers", {
|
|||||||
});
|
});
|
||||||
|
|
||||||
export const subscriptions = sqliteTable("subscriptions", {
|
export const subscriptions = sqliteTable("subscriptions", {
|
||||||
subscriptionId: text("subscriptionId")
|
subscriptionId: text("subscriptionId").primaryKey().notNull(),
|
||||||
.primaryKey()
|
|
||||||
.notNull(),
|
|
||||||
customerId: text("customerId")
|
customerId: text("customerId")
|
||||||
.notNull()
|
.notNull()
|
||||||
.references(() => customers.customerId, { onDelete: "cascade" }),
|
.references(() => customers.customerId, { onDelete: "cascade" }),
|
||||||
@@ -75,7 +75,9 @@ export const subscriptions = sqliteTable("subscriptions", {
|
|||||||
});
|
});
|
||||||
|
|
||||||
export const subscriptionItems = sqliteTable("subscriptionItems", {
|
export const subscriptionItems = sqliteTable("subscriptionItems", {
|
||||||
subscriptionItemId: integer("subscriptionItemId").primaryKey({ autoIncrement: true }),
|
subscriptionItemId: integer("subscriptionItemId").primaryKey({
|
||||||
|
autoIncrement: true
|
||||||
|
}),
|
||||||
subscriptionId: text("subscriptionId")
|
subscriptionId: text("subscriptionId")
|
||||||
.notNull()
|
.notNull()
|
||||||
.references(() => subscriptions.subscriptionId, {
|
.references(() => subscriptions.subscriptionId, {
|
||||||
@@ -129,7 +131,9 @@ export const limits = sqliteTable("limits", {
|
|||||||
});
|
});
|
||||||
|
|
||||||
export const usageNotifications = sqliteTable("usageNotifications", {
|
export const usageNotifications = sqliteTable("usageNotifications", {
|
||||||
notificationId: integer("notificationId").primaryKey({ autoIncrement: true }),
|
notificationId: integer("notificationId").primaryKey({
|
||||||
|
autoIncrement: true
|
||||||
|
}),
|
||||||
orgId: text("orgId")
|
orgId: text("orgId")
|
||||||
.notNull()
|
.notNull()
|
||||||
.references(() => orgs.orgId, { onDelete: "cascade" }),
|
.references(() => orgs.orgId, { onDelete: "cascade" }),
|
||||||
@@ -210,42 +214,56 @@ export const sessionTransferToken = sqliteTable("sessionTransferToken", {
|
|||||||
expiresAt: integer("expiresAt").notNull()
|
expiresAt: integer("expiresAt").notNull()
|
||||||
});
|
});
|
||||||
|
|
||||||
export const actionAuditLog = sqliteTable("actionAuditLog", {
|
export const actionAuditLog = sqliteTable(
|
||||||
id: integer("id").primaryKey({ autoIncrement: true }),
|
"actionAuditLog",
|
||||||
timestamp: integer("timestamp").notNull(), // this is EPOCH time in seconds
|
{
|
||||||
orgId: text("orgId")
|
id: integer("id").primaryKey({ autoIncrement: true }),
|
||||||
.notNull()
|
timestamp: integer("timestamp").notNull(), // this is EPOCH time in seconds
|
||||||
.references(() => orgs.orgId, { onDelete: "cascade" }),
|
orgId: text("orgId")
|
||||||
actorType: text("actorType").notNull(),
|
.notNull()
|
||||||
actor: text("actor").notNull(),
|
.references(() => orgs.orgId, { onDelete: "cascade" }),
|
||||||
actorId: text("actorId").notNull(),
|
actorType: text("actorType").notNull(),
|
||||||
action: text("action").notNull(),
|
actor: text("actor").notNull(),
|
||||||
metadata: text("metadata")
|
actorId: text("actorId").notNull(),
|
||||||
}, (table) => ([
|
action: text("action").notNull(),
|
||||||
index("idx_actionAuditLog_timestamp").on(table.timestamp),
|
metadata: text("metadata")
|
||||||
index("idx_actionAuditLog_org_timestamp").on(table.orgId, table.timestamp)
|
},
|
||||||
]));
|
(table) => [
|
||||||
|
index("idx_actionAuditLog_timestamp").on(table.timestamp),
|
||||||
|
index("idx_actionAuditLog_org_timestamp").on(
|
||||||
|
table.orgId,
|
||||||
|
table.timestamp
|
||||||
|
)
|
||||||
|
]
|
||||||
|
);
|
||||||
|
|
||||||
export const accessAuditLog = sqliteTable("accessAuditLog", {
|
export const accessAuditLog = sqliteTable(
|
||||||
id: integer("id").primaryKey({ autoIncrement: true }),
|
"accessAuditLog",
|
||||||
timestamp: integer("timestamp").notNull(), // this is EPOCH time in seconds
|
{
|
||||||
orgId: text("orgId")
|
id: integer("id").primaryKey({ autoIncrement: true }),
|
||||||
.notNull()
|
timestamp: integer("timestamp").notNull(), // this is EPOCH time in seconds
|
||||||
.references(() => orgs.orgId, { onDelete: "cascade" }),
|
orgId: text("orgId")
|
||||||
actorType: text("actorType"),
|
.notNull()
|
||||||
actor: text("actor"),
|
.references(() => orgs.orgId, { onDelete: "cascade" }),
|
||||||
actorId: text("actorId"),
|
actorType: text("actorType"),
|
||||||
resourceId: integer("resourceId"),
|
actor: text("actor"),
|
||||||
ip: text("ip"),
|
actorId: text("actorId"),
|
||||||
location: text("location"),
|
resourceId: integer("resourceId"),
|
||||||
type: text("type").notNull(),
|
ip: text("ip"),
|
||||||
action: integer("action", { mode: "boolean" }).notNull(),
|
location: text("location"),
|
||||||
userAgent: text("userAgent"),
|
type: text("type").notNull(),
|
||||||
metadata: text("metadata")
|
action: integer("action", { mode: "boolean" }).notNull(),
|
||||||
}, (table) => ([
|
userAgent: text("userAgent"),
|
||||||
index("idx_identityAuditLog_timestamp").on(table.timestamp),
|
metadata: text("metadata")
|
||||||
index("idx_identityAuditLog_org_timestamp").on(table.orgId, table.timestamp)
|
},
|
||||||
]));
|
(table) => [
|
||||||
|
index("idx_identityAuditLog_timestamp").on(table.timestamp),
|
||||||
|
index("idx_identityAuditLog_org_timestamp").on(
|
||||||
|
table.orgId,
|
||||||
|
table.timestamp
|
||||||
|
)
|
||||||
|
]
|
||||||
|
);
|
||||||
|
|
||||||
export type Limit = InferSelectModel<typeof limits>;
|
export type Limit = InferSelectModel<typeof limits>;
|
||||||
export type Account = InferSelectModel<typeof account>;
|
export type Account = InferSelectModel<typeof account>;
|
||||||
@@ -265,4 +283,4 @@ export type RemoteExitNodeSession = InferSelectModel<
|
|||||||
export type ExitNodeOrg = InferSelectModel<typeof exitNodeOrgs>;
|
export type ExitNodeOrg = InferSelectModel<typeof exitNodeOrgs>;
|
||||||
export type LoginPage = InferSelectModel<typeof loginPage>;
|
export type LoginPage = InferSelectModel<typeof loginPage>;
|
||||||
export type ActionAuditLog = InferSelectModel<typeof actionAuditLog>;
|
export type ActionAuditLog = InferSelectModel<typeof actionAuditLog>;
|
||||||
export type AccessAuditLog = InferSelectModel<typeof accessAuditLog>;
|
export type AccessAuditLog = InferSelectModel<typeof accessAuditLog>;
|
||||||
|
|||||||
@@ -18,10 +18,13 @@ function createEmailClient() {
|
|||||||
host: emailConfig.smtp_host,
|
host: emailConfig.smtp_host,
|
||||||
port: emailConfig.smtp_port,
|
port: emailConfig.smtp_port,
|
||||||
secure: emailConfig.smtp_secure || false,
|
secure: emailConfig.smtp_secure || false,
|
||||||
auth: (emailConfig.smtp_user && emailConfig.smtp_pass) ? {
|
auth:
|
||||||
user: emailConfig.smtp_user,
|
emailConfig.smtp_user && emailConfig.smtp_pass
|
||||||
pass: emailConfig.smtp_pass
|
? {
|
||||||
} : null
|
user: emailConfig.smtp_user,
|
||||||
|
pass: emailConfig.smtp_pass
|
||||||
|
}
|
||||||
|
: null
|
||||||
} as SMTPTransport.Options;
|
} as SMTPTransport.Options;
|
||||||
|
|
||||||
if (emailConfig.smtp_tls_reject_unauthorized !== undefined) {
|
if (emailConfig.smtp_tls_reject_unauthorized !== undefined) {
|
||||||
|
|||||||
@@ -19,7 +19,13 @@ interface Props {
|
|||||||
billingLink: string; // Link to billing page
|
billingLink: string; // Link to billing page
|
||||||
}
|
}
|
||||||
|
|
||||||
export const NotifyUsageLimitApproaching = ({ email, limitName, currentUsage, usageLimit, billingLink }: Props) => {
|
export const NotifyUsageLimitApproaching = ({
|
||||||
|
email,
|
||||||
|
limitName,
|
||||||
|
currentUsage,
|
||||||
|
usageLimit,
|
||||||
|
billingLink
|
||||||
|
}: Props) => {
|
||||||
const previewText = `Your usage for ${limitName} is approaching the limit.`;
|
const previewText = `Your usage for ${limitName} is approaching the limit.`;
|
||||||
const usagePercentage = Math.round((currentUsage / usageLimit) * 100);
|
const usagePercentage = Math.round((currentUsage / usageLimit) * 100);
|
||||||
|
|
||||||
@@ -37,23 +43,32 @@ export const NotifyUsageLimitApproaching = ({ email, limitName, currentUsage, us
|
|||||||
<EmailGreeting>Hi there,</EmailGreeting>
|
<EmailGreeting>Hi there,</EmailGreeting>
|
||||||
|
|
||||||
<EmailText>
|
<EmailText>
|
||||||
We wanted to let you know that your usage for <strong>{limitName}</strong> is approaching your plan limit.
|
We wanted to let you know that your usage for{" "}
|
||||||
|
<strong>{limitName}</strong> is approaching your
|
||||||
|
plan limit.
|
||||||
</EmailText>
|
</EmailText>
|
||||||
|
|
||||||
<EmailText>
|
<EmailText>
|
||||||
<strong>Current Usage:</strong> {currentUsage} of {usageLimit} ({usagePercentage}%)
|
<strong>Current Usage:</strong> {currentUsage} of{" "}
|
||||||
|
{usageLimit} ({usagePercentage}%)
|
||||||
</EmailText>
|
</EmailText>
|
||||||
|
|
||||||
<EmailText>
|
<EmailText>
|
||||||
Once you reach your limit, some functionality may be restricted or your sites may disconnect until you upgrade your plan or your usage resets.
|
Once you reach your limit, some functionality may be
|
||||||
|
restricted or your sites may disconnect until you
|
||||||
|
upgrade your plan or your usage resets.
|
||||||
</EmailText>
|
</EmailText>
|
||||||
|
|
||||||
<EmailText>
|
<EmailText>
|
||||||
To avoid any interruption to your service, we recommend upgrading your plan or monitoring your usage closely. You can <a href={billingLink}>upgrade your plan here</a>.
|
To avoid any interruption to your service, we
|
||||||
|
recommend upgrading your plan or monitoring your
|
||||||
|
usage closely. You can{" "}
|
||||||
|
<a href={billingLink}>upgrade your plan here</a>.
|
||||||
</EmailText>
|
</EmailText>
|
||||||
|
|
||||||
<EmailText>
|
<EmailText>
|
||||||
If you have any questions or need assistance, please don't hesitate to reach out to our support team.
|
If you have any questions or need assistance, please
|
||||||
|
don't hesitate to reach out to our support team.
|
||||||
</EmailText>
|
</EmailText>
|
||||||
|
|
||||||
<EmailFooter>
|
<EmailFooter>
|
||||||
|
|||||||
@@ -19,7 +19,13 @@ interface Props {
|
|||||||
billingLink: string; // Link to billing page
|
billingLink: string; // Link to billing page
|
||||||
}
|
}
|
||||||
|
|
||||||
export const NotifyUsageLimitReached = ({ email, limitName, currentUsage, usageLimit, billingLink }: Props) => {
|
export const NotifyUsageLimitReached = ({
|
||||||
|
email,
|
||||||
|
limitName,
|
||||||
|
currentUsage,
|
||||||
|
usageLimit,
|
||||||
|
billingLink
|
||||||
|
}: Props) => {
|
||||||
const previewText = `You've reached your ${limitName} usage limit - Action required`;
|
const previewText = `You've reached your ${limitName} usage limit - Action required`;
|
||||||
const usagePercentage = Math.round((currentUsage / usageLimit) * 100);
|
const usagePercentage = Math.round((currentUsage / usageLimit) * 100);
|
||||||
|
|
||||||
@@ -32,30 +38,48 @@ export const NotifyUsageLimitReached = ({ email, limitName, currentUsage, usageL
|
|||||||
<EmailContainer>
|
<EmailContainer>
|
||||||
<EmailLetterHead />
|
<EmailLetterHead />
|
||||||
|
|
||||||
<EmailHeading>Usage Limit Reached - Action Required</EmailHeading>
|
<EmailHeading>
|
||||||
|
Usage Limit Reached - Action Required
|
||||||
|
</EmailHeading>
|
||||||
|
|
||||||
<EmailGreeting>Hi there,</EmailGreeting>
|
<EmailGreeting>Hi there,</EmailGreeting>
|
||||||
|
|
||||||
<EmailText>
|
<EmailText>
|
||||||
You have reached your usage limit for <strong>{limitName}</strong>.
|
You have reached your usage limit for{" "}
|
||||||
|
<strong>{limitName}</strong>.
|
||||||
</EmailText>
|
</EmailText>
|
||||||
|
|
||||||
<EmailText>
|
<EmailText>
|
||||||
<strong>Current Usage:</strong> {currentUsage} of {usageLimit} ({usagePercentage}%)
|
<strong>Current Usage:</strong> {currentUsage} of{" "}
|
||||||
|
{usageLimit} ({usagePercentage}%)
|
||||||
</EmailText>
|
</EmailText>
|
||||||
|
|
||||||
<EmailText>
|
<EmailText>
|
||||||
<strong>Important:</strong> Your functionality may now be restricted and your sites may disconnect until you either upgrade your plan or your usage resets. To prevent any service interruption, immediate action is recommended.
|
<strong>Important:</strong> Your functionality may
|
||||||
|
now be restricted and your sites may disconnect
|
||||||
|
until you either upgrade your plan or your usage
|
||||||
|
resets. To prevent any service interruption,
|
||||||
|
immediate action is recommended.
|
||||||
</EmailText>
|
</EmailText>
|
||||||
|
|
||||||
<EmailText>
|
<EmailText>
|
||||||
<strong>What you can do:</strong>
|
<strong>What you can do:</strong>
|
||||||
<br />• <a href={billingLink} style={{ color: '#2563eb', fontWeight: 'bold' }}>Upgrade your plan immediately</a> to restore full functionality
|
<br />•{" "}
|
||||||
<br />• Monitor your usage to stay within limits in the future
|
<a
|
||||||
|
href={billingLink}
|
||||||
|
style={{ color: "#2563eb", fontWeight: "bold" }}
|
||||||
|
>
|
||||||
|
Upgrade your plan immediately
|
||||||
|
</a>{" "}
|
||||||
|
to restore full functionality
|
||||||
|
<br />• Monitor your usage to stay within limits in
|
||||||
|
the future
|
||||||
</EmailText>
|
</EmailText>
|
||||||
|
|
||||||
<EmailText>
|
<EmailText>
|
||||||
If you have any questions or need immediate assistance, please contact our support team right away.
|
If you have any questions or need immediate
|
||||||
|
assistance, please contact our support team right
|
||||||
|
away.
|
||||||
</EmailText>
|
</EmailText>
|
||||||
|
|
||||||
<EmailFooter>
|
<EmailFooter>
|
||||||
|
|||||||
@@ -5,7 +5,7 @@ import config from "@server/lib/config";
|
|||||||
import logger from "@server/logger";
|
import logger from "@server/logger";
|
||||||
import {
|
import {
|
||||||
errorHandlerMiddleware,
|
errorHandlerMiddleware,
|
||||||
notFoundMiddleware,
|
notFoundMiddleware
|
||||||
} from "@server/middlewares";
|
} from "@server/middlewares";
|
||||||
import { authenticated, unauthenticated } from "#dynamic/routers/integration";
|
import { authenticated, unauthenticated } from "#dynamic/routers/integration";
|
||||||
import { logIncomingMiddleware } from "./middlewares/logIncoming";
|
import { logIncomingMiddleware } from "./middlewares/logIncoming";
|
||||||
|
|||||||
@@ -25,16 +25,22 @@ export const FeatureMeterIdsSandbox: Record<FeatureId, string> = {
|
|||||||
};
|
};
|
||||||
|
|
||||||
export function getFeatureMeterId(featureId: FeatureId): string {
|
export function getFeatureMeterId(featureId: FeatureId): string {
|
||||||
if (process.env.ENVIRONMENT == "prod" && process.env.SANDBOX_MODE !== "true") {
|
if (
|
||||||
|
process.env.ENVIRONMENT == "prod" &&
|
||||||
|
process.env.SANDBOX_MODE !== "true"
|
||||||
|
) {
|
||||||
return FeatureMeterIds[featureId];
|
return FeatureMeterIds[featureId];
|
||||||
} else {
|
} else {
|
||||||
return FeatureMeterIdsSandbox[featureId];
|
return FeatureMeterIdsSandbox[featureId];
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
export function getFeatureIdByMetricId(metricId: string): FeatureId | undefined {
|
export function getFeatureIdByMetricId(
|
||||||
return (Object.entries(FeatureMeterIds) as [FeatureId, string][])
|
metricId: string
|
||||||
.find(([_, v]) => v === metricId)?.[0];
|
): FeatureId | undefined {
|
||||||
|
return (Object.entries(FeatureMeterIds) as [FeatureId, string][]).find(
|
||||||
|
([_, v]) => v === metricId
|
||||||
|
)?.[0];
|
||||||
}
|
}
|
||||||
|
|
||||||
export type FeaturePriceSet = {
|
export type FeaturePriceSet = {
|
||||||
@@ -43,7 +49,8 @@ export type FeaturePriceSet = {
|
|||||||
[FeatureId.DOMAINS]?: string; // Optional since domains are not billed
|
[FeatureId.DOMAINS]?: string; // Optional since domains are not billed
|
||||||
};
|
};
|
||||||
|
|
||||||
export const standardFeaturePriceSet: FeaturePriceSet = { // Free tier matches the freeLimitSet
|
export const standardFeaturePriceSet: FeaturePriceSet = {
|
||||||
|
// Free tier matches the freeLimitSet
|
||||||
[FeatureId.SITE_UPTIME]: "price_1RrQc4D3Ee2Ir7WmaJGZ3MtF",
|
[FeatureId.SITE_UPTIME]: "price_1RrQc4D3Ee2Ir7WmaJGZ3MtF",
|
||||||
[FeatureId.USERS]: "price_1RrQeJD3Ee2Ir7WmgveP3xea",
|
[FeatureId.USERS]: "price_1RrQeJD3Ee2Ir7WmgveP3xea",
|
||||||
[FeatureId.EGRESS_DATA_MB]: "price_1RrQXFD3Ee2Ir7WmvGDlgxQk",
|
[FeatureId.EGRESS_DATA_MB]: "price_1RrQXFD3Ee2Ir7WmvGDlgxQk",
|
||||||
@@ -51,7 +58,8 @@ export const standardFeaturePriceSet: FeaturePriceSet = { // Free tier matches t
|
|||||||
[FeatureId.REMOTE_EXIT_NODES]: "price_1S46weD3Ee2Ir7Wm94KEHI4h"
|
[FeatureId.REMOTE_EXIT_NODES]: "price_1S46weD3Ee2Ir7Wm94KEHI4h"
|
||||||
};
|
};
|
||||||
|
|
||||||
export const standardFeaturePriceSetSandbox: FeaturePriceSet = { // Free tier matches the freeLimitSet
|
export const standardFeaturePriceSetSandbox: FeaturePriceSet = {
|
||||||
|
// Free tier matches the freeLimitSet
|
||||||
[FeatureId.SITE_UPTIME]: "price_1RefFBDCpkOb237BPrKZ8IEU",
|
[FeatureId.SITE_UPTIME]: "price_1RefFBDCpkOb237BPrKZ8IEU",
|
||||||
[FeatureId.USERS]: "price_1ReNa4DCpkOb237Bc67G5muF",
|
[FeatureId.USERS]: "price_1ReNa4DCpkOb237Bc67G5muF",
|
||||||
[FeatureId.EGRESS_DATA_MB]: "price_1Rfp9LDCpkOb237BwuN5Oiu0",
|
[FeatureId.EGRESS_DATA_MB]: "price_1Rfp9LDCpkOb237BwuN5Oiu0",
|
||||||
@@ -60,15 +68,20 @@ export const standardFeaturePriceSetSandbox: FeaturePriceSet = { // Free tier ma
|
|||||||
};
|
};
|
||||||
|
|
||||||
export function getStandardFeaturePriceSet(): FeaturePriceSet {
|
export function getStandardFeaturePriceSet(): FeaturePriceSet {
|
||||||
if (process.env.ENVIRONMENT == "prod" && process.env.SANDBOX_MODE !== "true") {
|
if (
|
||||||
|
process.env.ENVIRONMENT == "prod" &&
|
||||||
|
process.env.SANDBOX_MODE !== "true"
|
||||||
|
) {
|
||||||
return standardFeaturePriceSet;
|
return standardFeaturePriceSet;
|
||||||
} else {
|
} else {
|
||||||
return standardFeaturePriceSetSandbox;
|
return standardFeaturePriceSetSandbox;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
export function getLineItems(featurePriceSet: FeaturePriceSet): Stripe.Checkout.SessionCreateParams.LineItem[] {
|
export function getLineItems(
|
||||||
|
featurePriceSet: FeaturePriceSet
|
||||||
|
): Stripe.Checkout.SessionCreateParams.LineItem[] {
|
||||||
return Object.entries(featurePriceSet).map(([featureId, priceId]) => ({
|
return Object.entries(featurePriceSet).map(([featureId, priceId]) => ({
|
||||||
price: priceId,
|
price: priceId
|
||||||
}));
|
}));
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -2,4 +2,4 @@ export * from "./limitSet";
|
|||||||
export * from "./features";
|
export * from "./features";
|
||||||
export * from "./limitsService";
|
export * from "./limitsService";
|
||||||
export * from "./getOrgTierData";
|
export * from "./getOrgTierData";
|
||||||
export * from "./createCustomer";
|
export * from "./createCustomer";
|
||||||
|
|||||||
@@ -12,7 +12,7 @@ export const sandboxLimitSet: LimitSet = {
|
|||||||
[FeatureId.USERS]: { value: 1, description: "Sandbox limit" },
|
[FeatureId.USERS]: { value: 1, description: "Sandbox limit" },
|
||||||
[FeatureId.EGRESS_DATA_MB]: { value: 1000, description: "Sandbox limit" }, // 1 GB
|
[FeatureId.EGRESS_DATA_MB]: { value: 1000, description: "Sandbox limit" }, // 1 GB
|
||||||
[FeatureId.DOMAINS]: { value: 0, description: "Sandbox limit" },
|
[FeatureId.DOMAINS]: { value: 0, description: "Sandbox limit" },
|
||||||
[FeatureId.REMOTE_EXIT_NODES]: { value: 0, description: "Sandbox limit" },
|
[FeatureId.REMOTE_EXIT_NODES]: { value: 0, description: "Sandbox limit" }
|
||||||
};
|
};
|
||||||
|
|
||||||
export const freeLimitSet: LimitSet = {
|
export const freeLimitSet: LimitSet = {
|
||||||
@@ -29,7 +29,7 @@ export const freeLimitSet: LimitSet = {
|
|||||||
export const subscribedLimitSet: LimitSet = {
|
export const subscribedLimitSet: LimitSet = {
|
||||||
[FeatureId.SITE_UPTIME]: {
|
[FeatureId.SITE_UPTIME]: {
|
||||||
value: 2232000,
|
value: 2232000,
|
||||||
description: "Contact us to increase soft limit.",
|
description: "Contact us to increase soft limit."
|
||||||
}, // 50 sites up for 31 days
|
}, // 50 sites up for 31 days
|
||||||
[FeatureId.USERS]: {
|
[FeatureId.USERS]: {
|
||||||
value: 150,
|
value: 150,
|
||||||
@@ -38,7 +38,7 @@ export const subscribedLimitSet: LimitSet = {
|
|||||||
[FeatureId.EGRESS_DATA_MB]: {
|
[FeatureId.EGRESS_DATA_MB]: {
|
||||||
value: 12000000,
|
value: 12000000,
|
||||||
description: "Contact us to increase soft limit."
|
description: "Contact us to increase soft limit."
|
||||||
}, // 12000 GB
|
}, // 12000 GB
|
||||||
[FeatureId.DOMAINS]: {
|
[FeatureId.DOMAINS]: {
|
||||||
value: 25,
|
value: 25,
|
||||||
description: "Contact us to increase soft limit."
|
description: "Contact us to increase soft limit."
|
||||||
|
|||||||
@@ -1,22 +1,32 @@
|
|||||||
export enum TierId {
|
export enum TierId {
|
||||||
STANDARD = "standard",
|
STANDARD = "standard"
|
||||||
}
|
}
|
||||||
|
|
||||||
export type TierPriceSet = {
|
export type TierPriceSet = {
|
||||||
[key in TierId]: string;
|
[key in TierId]: string;
|
||||||
};
|
};
|
||||||
|
|
||||||
export const tierPriceSet: TierPriceSet = { // Free tier matches the freeLimitSet
|
export const tierPriceSet: TierPriceSet = {
|
||||||
[TierId.STANDARD]: "price_1RrQ9cD3Ee2Ir7Wmqdy3KBa0",
|
// Free tier matches the freeLimitSet
|
||||||
|
[TierId.STANDARD]: "price_1RrQ9cD3Ee2Ir7Wmqdy3KBa0"
|
||||||
};
|
};
|
||||||
|
|
||||||
export const tierPriceSetSandbox: TierPriceSet = { // Free tier matches the freeLimitSet
|
export const tierPriceSetSandbox: TierPriceSet = {
|
||||||
|
// Free tier matches the freeLimitSet
|
||||||
// when matching tier the keys closer to 0 index are matched first so list the tiers in descending order of value
|
// when matching tier the keys closer to 0 index are matched first so list the tiers in descending order of value
|
||||||
[TierId.STANDARD]: "price_1RrAYJDCpkOb237By2s1P32m",
|
[TierId.STANDARD]: "price_1RrAYJDCpkOb237By2s1P32m"
|
||||||
};
|
};
|
||||||
|
|
||||||
export function getTierPriceSet(environment?: string, sandbox_mode?: boolean): TierPriceSet {
|
export function getTierPriceSet(
|
||||||
if ((process.env.ENVIRONMENT == "prod" && process.env.SANDBOX_MODE !== "true") || (environment === "prod" && sandbox_mode !== true)) { // THIS GETS LOADED CLIENT SIDE AND SERVER SIDE
|
environment?: string,
|
||||||
|
sandbox_mode?: boolean
|
||||||
|
): TierPriceSet {
|
||||||
|
if (
|
||||||
|
(process.env.ENVIRONMENT == "prod" &&
|
||||||
|
process.env.SANDBOX_MODE !== "true") ||
|
||||||
|
(environment === "prod" && sandbox_mode !== true)
|
||||||
|
) {
|
||||||
|
// THIS GETS LOADED CLIENT SIDE AND SERVER SIDE
|
||||||
return tierPriceSet;
|
return tierPriceSet;
|
||||||
} else {
|
} else {
|
||||||
return tierPriceSetSandbox;
|
return tierPriceSetSandbox;
|
||||||
|
|||||||
@@ -19,7 +19,7 @@ import logger from "@server/logger";
|
|||||||
import { sendToClient } from "#dynamic/routers/ws";
|
import { sendToClient } from "#dynamic/routers/ws";
|
||||||
import { build } from "@server/build";
|
import { build } from "@server/build";
|
||||||
import { s3Client } from "@server/lib/s3";
|
import { s3Client } from "@server/lib/s3";
|
||||||
import cache from "@server/lib/cache";
|
import cache from "@server/lib/cache";
|
||||||
|
|
||||||
interface StripeEvent {
|
interface StripeEvent {
|
||||||
identifier?: string;
|
identifier?: string;
|
||||||
|
|||||||
@@ -34,7 +34,10 @@ export async function applyNewtDockerBlueprint(
|
|||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
if (isEmptyObject(blueprint["proxy-resources"]) && isEmptyObject(blueprint["client-resources"])) {
|
if (
|
||||||
|
isEmptyObject(blueprint["proxy-resources"]) &&
|
||||||
|
isEmptyObject(blueprint["client-resources"])
|
||||||
|
) {
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -84,12 +84,20 @@ export function processContainerLabels(containers: Container[]): {
|
|||||||
|
|
||||||
// Process proxy resources
|
// Process proxy resources
|
||||||
if (Object.keys(proxyResourceLabels).length > 0) {
|
if (Object.keys(proxyResourceLabels).length > 0) {
|
||||||
processResourceLabels(proxyResourceLabels, container, result["proxy-resources"]);
|
processResourceLabels(
|
||||||
|
proxyResourceLabels,
|
||||||
|
container,
|
||||||
|
result["proxy-resources"]
|
||||||
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
// Process client resources
|
// Process client resources
|
||||||
if (Object.keys(clientResourceLabels).length > 0) {
|
if (Object.keys(clientResourceLabels).length > 0) {
|
||||||
processResourceLabels(clientResourceLabels, container, result["client-resources"]);
|
processResourceLabels(
|
||||||
|
clientResourceLabels,
|
||||||
|
container,
|
||||||
|
result["client-resources"]
|
||||||
|
);
|
||||||
}
|
}
|
||||||
});
|
});
|
||||||
|
|
||||||
@@ -161,8 +169,7 @@ function processResourceLabels(
|
|||||||
const finalTarget = { ...target };
|
const finalTarget = { ...target };
|
||||||
if (!finalTarget.hostname) {
|
if (!finalTarget.hostname) {
|
||||||
finalTarget.hostname =
|
finalTarget.hostname =
|
||||||
container.name ||
|
container.name || container.hostname;
|
||||||
container.hostname;
|
|
||||||
}
|
}
|
||||||
if (!finalTarget.port) {
|
if (!finalTarget.port) {
|
||||||
const containerPort =
|
const containerPort =
|
||||||
|
|||||||
@@ -1086,10 +1086,8 @@ async function getDomainId(
|
|||||||
|
|
||||||
// remove the base domain of the domain
|
// remove the base domain of the domain
|
||||||
let subdomain = null;
|
let subdomain = null;
|
||||||
if (domainSelection.type == "ns" || domainSelection.type == "wildcard") {
|
if (fullDomain != baseDomain) {
|
||||||
if (fullDomain != baseDomain) {
|
subdomain = fullDomain.replace(`.${baseDomain}`, "");
|
||||||
subdomain = fullDomain.replace(`.${baseDomain}`, "");
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// Return the first valid domain
|
// Return the first valid domain
|
||||||
|
|||||||
@@ -312,7 +312,7 @@ export const ConfigSchema = z
|
|||||||
};
|
};
|
||||||
delete (data as any)["public-resources"];
|
delete (data as any)["public-resources"];
|
||||||
}
|
}
|
||||||
|
|
||||||
// Merge private-resources into client-resources
|
// Merge private-resources into client-resources
|
||||||
if (data["private-resources"]) {
|
if (data["private-resources"]) {
|
||||||
data["client-resources"] = {
|
data["client-resources"] = {
|
||||||
@@ -321,10 +321,13 @@ export const ConfigSchema = z
|
|||||||
};
|
};
|
||||||
delete (data as any)["private-resources"];
|
delete (data as any)["private-resources"];
|
||||||
}
|
}
|
||||||
|
|
||||||
return data as {
|
return data as {
|
||||||
"proxy-resources": Record<string, z.infer<typeof ResourceSchema>>;
|
"proxy-resources": Record<string, z.infer<typeof ResourceSchema>>;
|
||||||
"client-resources": Record<string, z.infer<typeof ClientResourceSchema>>;
|
"client-resources": Record<
|
||||||
|
string,
|
||||||
|
z.infer<typeof ClientResourceSchema>
|
||||||
|
>;
|
||||||
sites: Record<string, z.infer<typeof SiteSchema>>;
|
sites: Record<string, z.infer<typeof SiteSchema>>;
|
||||||
};
|
};
|
||||||
})
|
})
|
||||||
|
|||||||
@@ -2,4 +2,4 @@ import NodeCache from "node-cache";
|
|||||||
|
|
||||||
export const cache = new NodeCache({ stdTTL: 3600, checkperiod: 120 });
|
export const cache = new NodeCache({ stdTTL: 3600, checkperiod: 120 });
|
||||||
|
|
||||||
export default cache;
|
export default cache;
|
||||||
|
|||||||
@@ -166,7 +166,10 @@ export async function calculateUserClientsForOrgs(
|
|||||||
];
|
];
|
||||||
|
|
||||||
// Get next available subnet
|
// Get next available subnet
|
||||||
const newSubnet = await getNextAvailableClientSubnet(orgId);
|
const newSubnet = await getNextAvailableClientSubnet(
|
||||||
|
orgId,
|
||||||
|
transaction
|
||||||
|
);
|
||||||
if (!newSubnet) {
|
if (!newSubnet) {
|
||||||
logger.warn(
|
logger.warn(
|
||||||
`Skipping org ${orgId} for OLM ${olm.olmId} (user ${userId}): no available subnet found`
|
`Skipping org ${orgId} for OLM ${olm.olmId} (user ${userId}): no available subnet found`
|
||||||
|
|||||||
@@ -1,4 +1,6 @@
|
|||||||
export async function getValidCertificatesForDomains(domains: Set<string>): Promise<
|
export async function getValidCertificatesForDomains(
|
||||||
|
domains: Set<string>
|
||||||
|
): Promise<
|
||||||
Array<{
|
Array<{
|
||||||
id: number;
|
id: number;
|
||||||
domain: string;
|
domain: string;
|
||||||
@@ -10,4 +12,4 @@ export async function getValidCertificatesForDomains(domains: Set<string>): Prom
|
|||||||
}>
|
}>
|
||||||
> {
|
> {
|
||||||
return []; // stub
|
return []; // stub
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -7,7 +7,10 @@ function dateToTimestamp(dateStr: string): number {
|
|||||||
|
|
||||||
// Testable version of calculateCutoffTimestamp that accepts a "now" timestamp
|
// Testable version of calculateCutoffTimestamp that accepts a "now" timestamp
|
||||||
// This matches the logic in cleanupLogs.ts but allows injecting the current time
|
// This matches the logic in cleanupLogs.ts but allows injecting the current time
|
||||||
function calculateCutoffTimestampWithNow(retentionDays: number, nowTimestamp: number): number {
|
function calculateCutoffTimestampWithNow(
|
||||||
|
retentionDays: number,
|
||||||
|
nowTimestamp: number
|
||||||
|
): number {
|
||||||
if (retentionDays === 9001) {
|
if (retentionDays === 9001) {
|
||||||
// Special case: data is erased at the end of the year following the year it was generated
|
// Special case: data is erased at the end of the year following the year it was generated
|
||||||
// This means we delete logs from 2 years ago or older (logs from year Y are deleted after Dec 31 of year Y+1)
|
// This means we delete logs from 2 years ago or older (logs from year Y are deleted after Dec 31 of year Y+1)
|
||||||
@@ -28,7 +31,7 @@ function testCalculateCutoffTimestamp() {
|
|||||||
{
|
{
|
||||||
const now = dateToTimestamp("2025-12-06T12:00:00Z");
|
const now = dateToTimestamp("2025-12-06T12:00:00Z");
|
||||||
const result = calculateCutoffTimestampWithNow(30, now);
|
const result = calculateCutoffTimestampWithNow(30, now);
|
||||||
const expected = now - (30 * 24 * 60 * 60);
|
const expected = now - 30 * 24 * 60 * 60;
|
||||||
assertEquals(result, expected, "30 days retention calculation failed");
|
assertEquals(result, expected, "30 days retention calculation failed");
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -36,7 +39,7 @@ function testCalculateCutoffTimestamp() {
|
|||||||
{
|
{
|
||||||
const now = dateToTimestamp("2025-06-15T00:00:00Z");
|
const now = dateToTimestamp("2025-06-15T00:00:00Z");
|
||||||
const result = calculateCutoffTimestampWithNow(90, now);
|
const result = calculateCutoffTimestampWithNow(90, now);
|
||||||
const expected = now - (90 * 24 * 60 * 60);
|
const expected = now - 90 * 24 * 60 * 60;
|
||||||
assertEquals(result, expected, "90 days retention calculation failed");
|
assertEquals(result, expected, "90 days retention calculation failed");
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -48,7 +51,11 @@ function testCalculateCutoffTimestamp() {
|
|||||||
const now = dateToTimestamp("2025-12-06T12:00:00Z");
|
const now = dateToTimestamp("2025-12-06T12:00:00Z");
|
||||||
const result = calculateCutoffTimestampWithNow(9001, now);
|
const result = calculateCutoffTimestampWithNow(9001, now);
|
||||||
const expected = dateToTimestamp("2024-01-01T00:00:00Z");
|
const expected = dateToTimestamp("2024-01-01T00:00:00Z");
|
||||||
assertEquals(result, expected, "9001 retention (Dec 2025) - should cutoff at Jan 1, 2024");
|
assertEquals(
|
||||||
|
result,
|
||||||
|
expected,
|
||||||
|
"9001 retention (Dec 2025) - should cutoff at Jan 1, 2024"
|
||||||
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
// Test 4: Special case 9001 - January 2026
|
// Test 4: Special case 9001 - January 2026
|
||||||
@@ -58,7 +65,11 @@ function testCalculateCutoffTimestamp() {
|
|||||||
const now = dateToTimestamp("2026-01-15T12:00:00Z");
|
const now = dateToTimestamp("2026-01-15T12:00:00Z");
|
||||||
const result = calculateCutoffTimestampWithNow(9001, now);
|
const result = calculateCutoffTimestampWithNow(9001, now);
|
||||||
const expected = dateToTimestamp("2025-01-01T00:00:00Z");
|
const expected = dateToTimestamp("2025-01-01T00:00:00Z");
|
||||||
assertEquals(result, expected, "9001 retention (Jan 2026) - should cutoff at Jan 1, 2025");
|
assertEquals(
|
||||||
|
result,
|
||||||
|
expected,
|
||||||
|
"9001 retention (Jan 2026) - should cutoff at Jan 1, 2025"
|
||||||
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
// Test 5: Special case 9001 - December 31, 2025 at 23:59:59 UTC
|
// Test 5: Special case 9001 - December 31, 2025 at 23:59:59 UTC
|
||||||
@@ -68,7 +79,11 @@ function testCalculateCutoffTimestamp() {
|
|||||||
const now = dateToTimestamp("2025-12-31T23:59:59Z");
|
const now = dateToTimestamp("2025-12-31T23:59:59Z");
|
||||||
const result = calculateCutoffTimestampWithNow(9001, now);
|
const result = calculateCutoffTimestampWithNow(9001, now);
|
||||||
const expected = dateToTimestamp("2024-01-01T00:00:00Z");
|
const expected = dateToTimestamp("2024-01-01T00:00:00Z");
|
||||||
assertEquals(result, expected, "9001 retention (Dec 31, 2025 23:59:59) - should cutoff at Jan 1, 2024");
|
assertEquals(
|
||||||
|
result,
|
||||||
|
expected,
|
||||||
|
"9001 retention (Dec 31, 2025 23:59:59) - should cutoff at Jan 1, 2024"
|
||||||
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
// Test 6: Special case 9001 - January 1, 2026 at 00:00:01 UTC
|
// Test 6: Special case 9001 - January 1, 2026 at 00:00:01 UTC
|
||||||
@@ -78,7 +93,11 @@ function testCalculateCutoffTimestamp() {
|
|||||||
const now = dateToTimestamp("2026-01-01T00:00:01Z");
|
const now = dateToTimestamp("2026-01-01T00:00:01Z");
|
||||||
const result = calculateCutoffTimestampWithNow(9001, now);
|
const result = calculateCutoffTimestampWithNow(9001, now);
|
||||||
const expected = dateToTimestamp("2025-01-01T00:00:00Z");
|
const expected = dateToTimestamp("2025-01-01T00:00:00Z");
|
||||||
assertEquals(result, expected, "9001 retention (Jan 1, 2026 00:00:01) - should cutoff at Jan 1, 2025");
|
assertEquals(
|
||||||
|
result,
|
||||||
|
expected,
|
||||||
|
"9001 retention (Jan 1, 2026 00:00:01) - should cutoff at Jan 1, 2025"
|
||||||
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
// Test 7: Special case 9001 - Mid year 2025
|
// Test 7: Special case 9001 - Mid year 2025
|
||||||
@@ -87,7 +106,11 @@ function testCalculateCutoffTimestamp() {
|
|||||||
const now = dateToTimestamp("2025-06-15T12:00:00Z");
|
const now = dateToTimestamp("2025-06-15T12:00:00Z");
|
||||||
const result = calculateCutoffTimestampWithNow(9001, now);
|
const result = calculateCutoffTimestampWithNow(9001, now);
|
||||||
const expected = dateToTimestamp("2024-01-01T00:00:00Z");
|
const expected = dateToTimestamp("2024-01-01T00:00:00Z");
|
||||||
assertEquals(result, expected, "9001 retention (mid 2025) - should cutoff at Jan 1, 2024");
|
assertEquals(
|
||||||
|
result,
|
||||||
|
expected,
|
||||||
|
"9001 retention (mid 2025) - should cutoff at Jan 1, 2024"
|
||||||
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
// Test 8: Special case 9001 - Early 2024
|
// Test 8: Special case 9001 - Early 2024
|
||||||
@@ -96,14 +119,18 @@ function testCalculateCutoffTimestamp() {
|
|||||||
const now = dateToTimestamp("2024-02-01T12:00:00Z");
|
const now = dateToTimestamp("2024-02-01T12:00:00Z");
|
||||||
const result = calculateCutoffTimestampWithNow(9001, now);
|
const result = calculateCutoffTimestampWithNow(9001, now);
|
||||||
const expected = dateToTimestamp("2023-01-01T00:00:00Z");
|
const expected = dateToTimestamp("2023-01-01T00:00:00Z");
|
||||||
assertEquals(result, expected, "9001 retention (early 2024) - should cutoff at Jan 1, 2023");
|
assertEquals(
|
||||||
|
result,
|
||||||
|
expected,
|
||||||
|
"9001 retention (early 2024) - should cutoff at Jan 1, 2023"
|
||||||
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
// Test 9: 1 day retention
|
// Test 9: 1 day retention
|
||||||
{
|
{
|
||||||
const now = dateToTimestamp("2025-12-06T12:00:00Z");
|
const now = dateToTimestamp("2025-12-06T12:00:00Z");
|
||||||
const result = calculateCutoffTimestampWithNow(1, now);
|
const result = calculateCutoffTimestampWithNow(1, now);
|
||||||
const expected = now - (1 * 24 * 60 * 60);
|
const expected = now - 1 * 24 * 60 * 60;
|
||||||
assertEquals(result, expected, "1 day retention calculation failed");
|
assertEquals(result, expected, "1 day retention calculation failed");
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -111,7 +138,7 @@ function testCalculateCutoffTimestamp() {
|
|||||||
{
|
{
|
||||||
const now = dateToTimestamp("2025-12-06T12:00:00Z");
|
const now = dateToTimestamp("2025-12-06T12:00:00Z");
|
||||||
const result = calculateCutoffTimestampWithNow(365, now);
|
const result = calculateCutoffTimestampWithNow(365, now);
|
||||||
const expected = now - (365 * 24 * 60 * 60);
|
const expected = now - 365 * 24 * 60 * 60;
|
||||||
assertEquals(result, expected, "365 days retention calculation failed");
|
assertEquals(result, expected, "365 days retention calculation failed");
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -123,11 +150,19 @@ function testCalculateCutoffTimestamp() {
|
|||||||
const cutoff = calculateCutoffTimestampWithNow(9001, now);
|
const cutoff = calculateCutoffTimestampWithNow(9001, now);
|
||||||
const logFromDec2023 = dateToTimestamp("2023-12-31T23:59:59Z");
|
const logFromDec2023 = dateToTimestamp("2023-12-31T23:59:59Z");
|
||||||
const logFromJan2024 = dateToTimestamp("2024-01-01T00:00:00Z");
|
const logFromJan2024 = dateToTimestamp("2024-01-01T00:00:00Z");
|
||||||
|
|
||||||
// Log from Dec 2023 should be before cutoff (deleted)
|
// Log from Dec 2023 should be before cutoff (deleted)
|
||||||
assertEquals(logFromDec2023 < cutoff, true, "Log from Dec 2023 should be deleted");
|
assertEquals(
|
||||||
|
logFromDec2023 < cutoff,
|
||||||
|
true,
|
||||||
|
"Log from Dec 2023 should be deleted"
|
||||||
|
);
|
||||||
// Log from Jan 2024 should be at or after cutoff (kept)
|
// Log from Jan 2024 should be at or after cutoff (kept)
|
||||||
assertEquals(logFromJan2024 >= cutoff, true, "Log from Jan 2024 should be kept");
|
assertEquals(
|
||||||
|
logFromJan2024 >= cutoff,
|
||||||
|
true,
|
||||||
|
"Log from Jan 2024 should be kept"
|
||||||
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
// Test 12: Verify 9001 in 2026 - logs from 2024 should now be deleted
|
// Test 12: Verify 9001 in 2026 - logs from 2024 should now be deleted
|
||||||
@@ -136,11 +171,19 @@ function testCalculateCutoffTimestamp() {
|
|||||||
const cutoff = calculateCutoffTimestampWithNow(9001, now);
|
const cutoff = calculateCutoffTimestampWithNow(9001, now);
|
||||||
const logFromDec2024 = dateToTimestamp("2024-12-31T23:59:59Z");
|
const logFromDec2024 = dateToTimestamp("2024-12-31T23:59:59Z");
|
||||||
const logFromJan2025 = dateToTimestamp("2025-01-01T00:00:00Z");
|
const logFromJan2025 = dateToTimestamp("2025-01-01T00:00:00Z");
|
||||||
|
|
||||||
// Log from Dec 2024 should be before cutoff (deleted)
|
// Log from Dec 2024 should be before cutoff (deleted)
|
||||||
assertEquals(logFromDec2024 < cutoff, true, "Log from Dec 2024 should be deleted in 2026");
|
assertEquals(
|
||||||
|
logFromDec2024 < cutoff,
|
||||||
|
true,
|
||||||
|
"Log from Dec 2024 should be deleted in 2026"
|
||||||
|
);
|
||||||
// Log from Jan 2025 should be at or after cutoff (kept)
|
// Log from Jan 2025 should be at or after cutoff (kept)
|
||||||
assertEquals(logFromJan2025 >= cutoff, true, "Log from Jan 2025 should be kept in 2026");
|
assertEquals(
|
||||||
|
logFromJan2025 >= cutoff,
|
||||||
|
true,
|
||||||
|
"Log from Jan 2025 should be kept in 2026"
|
||||||
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
// Test 13: Edge case - exactly at year boundary for 9001
|
// Test 13: Edge case - exactly at year boundary for 9001
|
||||||
@@ -149,7 +192,11 @@ function testCalculateCutoffTimestamp() {
|
|||||||
const now = dateToTimestamp("2025-01-01T00:00:00Z");
|
const now = dateToTimestamp("2025-01-01T00:00:00Z");
|
||||||
const result = calculateCutoffTimestampWithNow(9001, now);
|
const result = calculateCutoffTimestampWithNow(9001, now);
|
||||||
const expected = dateToTimestamp("2024-01-01T00:00:00Z");
|
const expected = dateToTimestamp("2024-01-01T00:00:00Z");
|
||||||
assertEquals(result, expected, "9001 retention (Jan 1, 2025 00:00:00) - should cutoff at Jan 1, 2024");
|
assertEquals(
|
||||||
|
result,
|
||||||
|
expected,
|
||||||
|
"9001 retention (Jan 1, 2025 00:00:00) - should cutoff at Jan 1, 2024"
|
||||||
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
// Test 14: Verify data from 2024 is kept throughout 2025 when using 9001
|
// Test 14: Verify data from 2024 is kept throughout 2025 when using 9001
|
||||||
@@ -157,18 +204,29 @@ function testCalculateCutoffTimestamp() {
|
|||||||
{
|
{
|
||||||
// Running in June 2025
|
// Running in June 2025
|
||||||
const nowJune2025 = dateToTimestamp("2025-06-15T12:00:00Z");
|
const nowJune2025 = dateToTimestamp("2025-06-15T12:00:00Z");
|
||||||
const cutoffJune2025 = calculateCutoffTimestampWithNow(9001, nowJune2025);
|
const cutoffJune2025 = calculateCutoffTimestampWithNow(
|
||||||
|
9001,
|
||||||
|
nowJune2025
|
||||||
|
);
|
||||||
const logFromJuly2024 = dateToTimestamp("2024-07-15T12:00:00Z");
|
const logFromJuly2024 = dateToTimestamp("2024-07-15T12:00:00Z");
|
||||||
|
|
||||||
// Log from July 2024 should be KEPT in June 2025
|
// Log from July 2024 should be KEPT in June 2025
|
||||||
assertEquals(logFromJuly2024 >= cutoffJune2025, true, "Log from July 2024 should be kept in June 2025");
|
assertEquals(
|
||||||
|
logFromJuly2024 >= cutoffJune2025,
|
||||||
|
true,
|
||||||
|
"Log from July 2024 should be kept in June 2025"
|
||||||
|
);
|
||||||
|
|
||||||
// Running in January 2026
|
// Running in January 2026
|
||||||
const nowJan2026 = dateToTimestamp("2026-01-15T12:00:00Z");
|
const nowJan2026 = dateToTimestamp("2026-01-15T12:00:00Z");
|
||||||
const cutoffJan2026 = calculateCutoffTimestampWithNow(9001, nowJan2026);
|
const cutoffJan2026 = calculateCutoffTimestampWithNow(9001, nowJan2026);
|
||||||
|
|
||||||
// Log from July 2024 should be DELETED in January 2026
|
// Log from July 2024 should be DELETED in January 2026
|
||||||
assertEquals(logFromJuly2024 < cutoffJan2026, true, "Log from July 2024 should be deleted in Jan 2026");
|
assertEquals(
|
||||||
|
logFromJuly2024 < cutoffJan2026,
|
||||||
|
true,
|
||||||
|
"Log from July 2024 should be deleted in Jan 2026"
|
||||||
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
// Test 15: Verify the exact requirement - data from 2024 must be purged on December 31, 2025
|
// Test 15: Verify the exact requirement - data from 2024 must be purged on December 31, 2025
|
||||||
@@ -176,16 +234,27 @@ function testCalculateCutoffTimestamp() {
|
|||||||
// On Jan 1, 2026 (now 2026), data from 2024 can be deleted
|
// On Jan 1, 2026 (now 2026), data from 2024 can be deleted
|
||||||
{
|
{
|
||||||
const logFromMid2024 = dateToTimestamp("2024-06-15T12:00:00Z");
|
const logFromMid2024 = dateToTimestamp("2024-06-15T12:00:00Z");
|
||||||
|
|
||||||
// Dec 31, 2025 23:59:59 - still 2025, log should be kept
|
// Dec 31, 2025 23:59:59 - still 2025, log should be kept
|
||||||
const nowDec31_2025 = dateToTimestamp("2025-12-31T23:59:59Z");
|
const nowDec31_2025 = dateToTimestamp("2025-12-31T23:59:59Z");
|
||||||
const cutoffDec31 = calculateCutoffTimestampWithNow(9001, nowDec31_2025);
|
const cutoffDec31 = calculateCutoffTimestampWithNow(
|
||||||
assertEquals(logFromMid2024 >= cutoffDec31, true, "Log from mid-2024 should be kept on Dec 31, 2025");
|
9001,
|
||||||
|
nowDec31_2025
|
||||||
|
);
|
||||||
|
assertEquals(
|
||||||
|
logFromMid2024 >= cutoffDec31,
|
||||||
|
true,
|
||||||
|
"Log from mid-2024 should be kept on Dec 31, 2025"
|
||||||
|
);
|
||||||
|
|
||||||
// Jan 1, 2026 00:00:00 - now 2026, log can be deleted
|
// Jan 1, 2026 00:00:00 - now 2026, log can be deleted
|
||||||
const nowJan1_2026 = dateToTimestamp("2026-01-01T00:00:00Z");
|
const nowJan1_2026 = dateToTimestamp("2026-01-01T00:00:00Z");
|
||||||
const cutoffJan1 = calculateCutoffTimestampWithNow(9001, nowJan1_2026);
|
const cutoffJan1 = calculateCutoffTimestampWithNow(9001, nowJan1_2026);
|
||||||
assertEquals(logFromMid2024 < cutoffJan1, true, "Log from mid-2024 should be deleted on Jan 1, 2026");
|
assertEquals(
|
||||||
|
logFromMid2024 < cutoffJan1,
|
||||||
|
true,
|
||||||
|
"Log from mid-2024 should be deleted on Jan 1, 2026"
|
||||||
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
console.log("All calculateCutoffTimestamp tests passed!");
|
console.log("All calculateCutoffTimestamp tests passed!");
|
||||||
|
|||||||
@@ -2,7 +2,7 @@ import path from "path";
|
|||||||
import { fileURLToPath } from "url";
|
import { fileURLToPath } from "url";
|
||||||
|
|
||||||
// This is a placeholder value replaced by the build process
|
// This is a placeholder value replaced by the build process
|
||||||
export const APP_VERSION = "1.13.0-rc.0";
|
export const APP_VERSION = "1.13.0";
|
||||||
|
|
||||||
export const __FILENAME = fileURLToPath(import.meta.url);
|
export const __FILENAME = fileURLToPath(import.meta.url);
|
||||||
export const __DIRNAME = path.dirname(__FILENAME);
|
export const __DIRNAME = path.dirname(__FILENAME);
|
||||||
|
|||||||
@@ -4,18 +4,20 @@ import { eq, and } from "drizzle-orm";
|
|||||||
import { subdomainSchema } from "@server/lib/schemas";
|
import { subdomainSchema } from "@server/lib/schemas";
|
||||||
import { fromError } from "zod-validation-error";
|
import { fromError } from "zod-validation-error";
|
||||||
|
|
||||||
export type DomainValidationResult = {
|
export type DomainValidationResult =
|
||||||
success: true;
|
| {
|
||||||
fullDomain: string;
|
success: true;
|
||||||
subdomain: string | null;
|
fullDomain: string;
|
||||||
} | {
|
subdomain: string | null;
|
||||||
success: false;
|
}
|
||||||
error: string;
|
| {
|
||||||
};
|
success: false;
|
||||||
|
error: string;
|
||||||
|
};
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Validates a domain and constructs the full domain based on domain type and subdomain.
|
* Validates a domain and constructs the full domain based on domain type and subdomain.
|
||||||
*
|
*
|
||||||
* @param domainId - The ID of the domain to validate
|
* @param domainId - The ID of the domain to validate
|
||||||
* @param orgId - The organization ID to check domain access
|
* @param orgId - The organization ID to check domain access
|
||||||
* @param subdomain - Optional subdomain to append (for ns and wildcard domains)
|
* @param subdomain - Optional subdomain to append (for ns and wildcard domains)
|
||||||
@@ -34,7 +36,10 @@ export async function validateAndConstructDomain(
|
|||||||
.where(eq(domains.domainId, domainId))
|
.where(eq(domains.domainId, domainId))
|
||||||
.leftJoin(
|
.leftJoin(
|
||||||
orgDomains,
|
orgDomains,
|
||||||
and(eq(orgDomains.orgId, orgId), eq(orgDomains.domainId, domainId))
|
and(
|
||||||
|
eq(orgDomains.orgId, orgId),
|
||||||
|
eq(orgDomains.domainId, domainId)
|
||||||
|
)
|
||||||
);
|
);
|
||||||
|
|
||||||
// Check if domain exists
|
// Check if domain exists
|
||||||
@@ -106,7 +111,7 @@ export async function validateAndConstructDomain(
|
|||||||
} catch (error) {
|
} catch (error) {
|
||||||
return {
|
return {
|
||||||
success: false,
|
success: false,
|
||||||
error: `An error occurred while validating domain: ${error instanceof Error ? error.message : 'Unknown error'}`
|
error: `An error occurred while validating domain: ${error instanceof Error ? error.message : "Unknown error"}`
|
||||||
};
|
};
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,39 +1,39 @@
|
|||||||
import crypto from 'crypto';
|
import crypto from "crypto";
|
||||||
|
|
||||||
export function encryptData(data: string, key: Buffer): string {
|
export function encryptData(data: string, key: Buffer): string {
|
||||||
const algorithm = 'aes-256-gcm';
|
const algorithm = "aes-256-gcm";
|
||||||
const iv = crypto.randomBytes(16);
|
const iv = crypto.randomBytes(16);
|
||||||
const cipher = crypto.createCipheriv(algorithm, key, iv);
|
const cipher = crypto.createCipheriv(algorithm, key, iv);
|
||||||
|
|
||||||
let encrypted = cipher.update(data, 'utf8', 'hex');
|
let encrypted = cipher.update(data, "utf8", "hex");
|
||||||
encrypted += cipher.final('hex');
|
encrypted += cipher.final("hex");
|
||||||
|
|
||||||
const authTag = cipher.getAuthTag();
|
const authTag = cipher.getAuthTag();
|
||||||
|
|
||||||
// Combine IV, auth tag, and encrypted data
|
// Combine IV, auth tag, and encrypted data
|
||||||
return iv.toString('hex') + ':' + authTag.toString('hex') + ':' + encrypted;
|
return iv.toString("hex") + ":" + authTag.toString("hex") + ":" + encrypted;
|
||||||
}
|
}
|
||||||
|
|
||||||
// Helper function to decrypt data (you'll need this to read certificates)
|
// Helper function to decrypt data (you'll need this to read certificates)
|
||||||
export function decryptData(encryptedData: string, key: Buffer): string {
|
export function decryptData(encryptedData: string, key: Buffer): string {
|
||||||
const algorithm = 'aes-256-gcm';
|
const algorithm = "aes-256-gcm";
|
||||||
const parts = encryptedData.split(':');
|
const parts = encryptedData.split(":");
|
||||||
|
|
||||||
if (parts.length !== 3) {
|
if (parts.length !== 3) {
|
||||||
throw new Error('Invalid encrypted data format');
|
throw new Error("Invalid encrypted data format");
|
||||||
}
|
}
|
||||||
|
|
||||||
const iv = Buffer.from(parts[0], 'hex');
|
const iv = Buffer.from(parts[0], "hex");
|
||||||
const authTag = Buffer.from(parts[1], 'hex');
|
const authTag = Buffer.from(parts[1], "hex");
|
||||||
const encrypted = parts[2];
|
const encrypted = parts[2];
|
||||||
|
|
||||||
const decipher = crypto.createDecipheriv(algorithm, key, iv);
|
const decipher = crypto.createDecipheriv(algorithm, key, iv);
|
||||||
decipher.setAuthTag(authTag);
|
decipher.setAuthTag(authTag);
|
||||||
|
|
||||||
let decrypted = decipher.update(encrypted, 'hex', 'utf8');
|
let decrypted = decipher.update(encrypted, "hex", "utf8");
|
||||||
decrypted += decipher.final('utf8');
|
decrypted += decipher.final("utf8");
|
||||||
|
|
||||||
return decrypted;
|
return decrypted;
|
||||||
}
|
}
|
||||||
|
|
||||||
// openssl rand -hex 32 > config/encryption.key
|
// openssl rand -hex 32 > config/encryption.key
|
||||||
|
|||||||
@@ -30,4 +30,4 @@ export async function getCurrentExitNodeId(): Promise<number> {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
return currentExitNodeId;
|
return currentExitNodeId;
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
export * from "./exitNodes";
|
export * from "./exitNodes";
|
||||||
export * from "./exitNodeComms";
|
export * from "./exitNodeComms";
|
||||||
export * from "./subnet";
|
export * from "./subnet";
|
||||||
export * from "./getCurrentExitNodeId";
|
export * from "./getCurrentExitNodeId";
|
||||||
|
|||||||
@@ -27,4 +27,4 @@ export async function getNextAvailableSubnet(): Promise<string> {
|
|||||||
"/" +
|
"/" +
|
||||||
subnet.split("/")[1];
|
subnet.split("/")[1];
|
||||||
return subnet;
|
return subnet;
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -30,4 +30,4 @@ export async function getCountryCodeForIp(
|
|||||||
}
|
}
|
||||||
|
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -33,7 +33,11 @@ export async function generateOidcRedirectUrl(
|
|||||||
)
|
)
|
||||||
.limit(1);
|
.limit(1);
|
||||||
|
|
||||||
if (res?.loginPage && res.loginPage.domainId && res.loginPage.fullDomain) {
|
if (
|
||||||
|
res?.loginPage &&
|
||||||
|
res.loginPage.domainId &&
|
||||||
|
res.loginPage.fullDomain
|
||||||
|
) {
|
||||||
baseUrl = `${method}://${res.loginPage.fullDomain}`;
|
baseUrl = `${method}://${res.loginPage.fullDomain}`;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -4,7 +4,7 @@ import { assertEquals } from "@test/assert";
|
|||||||
// Test cases
|
// Test cases
|
||||||
function testFindNextAvailableCidr() {
|
function testFindNextAvailableCidr() {
|
||||||
console.log("Running findNextAvailableCidr tests...");
|
console.log("Running findNextAvailableCidr tests...");
|
||||||
|
|
||||||
// Test 0: Basic IPv4 allocation with a subnet in the wrong range
|
// Test 0: Basic IPv4 allocation with a subnet in the wrong range
|
||||||
{
|
{
|
||||||
const existing = ["100.90.130.1/30", "100.90.128.4/30"];
|
const existing = ["100.90.130.1/30", "100.90.128.4/30"];
|
||||||
@@ -23,7 +23,11 @@ function testFindNextAvailableCidr() {
|
|||||||
{
|
{
|
||||||
const existing = ["10.0.0.0/16", "10.2.0.0/16"];
|
const existing = ["10.0.0.0/16", "10.2.0.0/16"];
|
||||||
const result = findNextAvailableCidr(existing, 16, "10.0.0.0/8");
|
const result = findNextAvailableCidr(existing, 16, "10.0.0.0/8");
|
||||||
assertEquals(result, "10.1.0.0/16", "Finding gap between allocations failed");
|
assertEquals(
|
||||||
|
result,
|
||||||
|
"10.1.0.0/16",
|
||||||
|
"Finding gap between allocations failed"
|
||||||
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
// Test 3: No available space
|
// Test 3: No available space
|
||||||
@@ -33,7 +37,7 @@ function testFindNextAvailableCidr() {
|
|||||||
assertEquals(result, null, "No available space test failed");
|
assertEquals(result, null, "No available space test failed");
|
||||||
}
|
}
|
||||||
|
|
||||||
// Test 4: Empty existing
|
// Test 4: Empty existing
|
||||||
{
|
{
|
||||||
const existing: string[] = [];
|
const existing: string[] = [];
|
||||||
const result = findNextAvailableCidr(existing, 30, "10.0.0.0/8");
|
const result = findNextAvailableCidr(existing, 30, "10.0.0.0/8");
|
||||||
@@ -137,4 +141,4 @@ try {
|
|||||||
} catch (error) {
|
} catch (error) {
|
||||||
console.error("Test failed:", error);
|
console.error("Test failed:", error);
|
||||||
process.exit(1);
|
process.exit(1);
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -116,6 +116,68 @@ function bigIntToIp(num: bigint, version: IPVersion): string {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Parses an endpoint string (ip:port) handling both IPv4 and IPv6 addresses.
|
||||||
|
* IPv6 addresses may be bracketed like [::1]:8080 or unbracketed like ::1:8080.
|
||||||
|
* For unbracketed IPv6, the last colon-separated segment is treated as the port.
|
||||||
|
*
|
||||||
|
* @param endpoint The endpoint string to parse (e.g., "192.168.1.1:8080" or "[::1]:8080" or "2607:fea8::1:8080")
|
||||||
|
* @returns An object with ip and port, or null if parsing fails
|
||||||
|
*/
|
||||||
|
export function parseEndpoint(endpoint: string): { ip: string; port: number } | null {
|
||||||
|
if (!endpoint) return null;
|
||||||
|
|
||||||
|
// Check for bracketed IPv6 format: [ip]:port
|
||||||
|
const bracketedMatch = endpoint.match(/^\[([^\]]+)\]:(\d+)$/);
|
||||||
|
if (bracketedMatch) {
|
||||||
|
const ip = bracketedMatch[1];
|
||||||
|
const port = parseInt(bracketedMatch[2], 10);
|
||||||
|
if (isNaN(port)) return null;
|
||||||
|
return { ip, port };
|
||||||
|
}
|
||||||
|
|
||||||
|
// Check if this looks like IPv6 (contains multiple colons)
|
||||||
|
const colonCount = (endpoint.match(/:/g) || []).length;
|
||||||
|
|
||||||
|
if (colonCount > 1) {
|
||||||
|
// This is IPv6 - the port is after the last colon
|
||||||
|
const lastColonIndex = endpoint.lastIndexOf(":");
|
||||||
|
const ip = endpoint.substring(0, lastColonIndex);
|
||||||
|
const portStr = endpoint.substring(lastColonIndex + 1);
|
||||||
|
const port = parseInt(portStr, 10);
|
||||||
|
if (isNaN(port)) return null;
|
||||||
|
return { ip, port };
|
||||||
|
}
|
||||||
|
|
||||||
|
// IPv4 format: ip:port
|
||||||
|
if (colonCount === 1) {
|
||||||
|
const [ip, portStr] = endpoint.split(":");
|
||||||
|
const port = parseInt(portStr, 10);
|
||||||
|
if (isNaN(port)) return null;
|
||||||
|
return { ip, port };
|
||||||
|
}
|
||||||
|
|
||||||
|
return null;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Formats an IP and port into a consistent endpoint string.
|
||||||
|
* IPv6 addresses are wrapped in brackets for proper parsing.
|
||||||
|
*
|
||||||
|
* @param ip The IP address (IPv4 or IPv6)
|
||||||
|
* @param port The port number
|
||||||
|
* @returns Formatted endpoint string
|
||||||
|
*/
|
||||||
|
export function formatEndpoint(ip: string, port: number): string {
|
||||||
|
// Check if this is IPv6 (contains colons)
|
||||||
|
if (ip.includes(":")) {
|
||||||
|
// Remove brackets if already present
|
||||||
|
const cleanIp = ip.replace(/^\[|\]$/g, "");
|
||||||
|
return `[${cleanIp}]:${port}`;
|
||||||
|
}
|
||||||
|
return `${ip}:${port}`;
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Converts CIDR to IP range
|
* Converts CIDR to IP range
|
||||||
*/
|
*/
|
||||||
@@ -244,9 +306,13 @@ export function isIpInCidr(ip: string, cidr: string): boolean {
|
|||||||
}
|
}
|
||||||
|
|
||||||
export async function getNextAvailableClientSubnet(
|
export async function getNextAvailableClientSubnet(
|
||||||
orgId: string
|
orgId: string,
|
||||||
|
transaction: Transaction | typeof db = db
|
||||||
): Promise<string> {
|
): Promise<string> {
|
||||||
const [org] = await db.select().from(orgs).where(eq(orgs.orgId, orgId));
|
const [org] = await transaction
|
||||||
|
.select()
|
||||||
|
.from(orgs)
|
||||||
|
.where(eq(orgs.orgId, orgId));
|
||||||
|
|
||||||
if (!org) {
|
if (!org) {
|
||||||
throw new Error(`Organization with ID ${orgId} not found`);
|
throw new Error(`Organization with ID ${orgId} not found`);
|
||||||
@@ -256,14 +322,14 @@ export async function getNextAvailableClientSubnet(
|
|||||||
throw new Error(`Organization with ID ${orgId} has no subnet defined`);
|
throw new Error(`Organization with ID ${orgId} has no subnet defined`);
|
||||||
}
|
}
|
||||||
|
|
||||||
const existingAddressesSites = await db
|
const existingAddressesSites = await transaction
|
||||||
.select({
|
.select({
|
||||||
address: sites.address
|
address: sites.address
|
||||||
})
|
})
|
||||||
.from(sites)
|
.from(sites)
|
||||||
.where(and(isNotNull(sites.address), eq(sites.orgId, orgId)));
|
.where(and(isNotNull(sites.address), eq(sites.orgId, orgId)));
|
||||||
|
|
||||||
const existingAddressesClients = await db
|
const existingAddressesClients = await transaction
|
||||||
.select({
|
.select({
|
||||||
address: clients.subnet
|
address: clients.subnet
|
||||||
})
|
})
|
||||||
@@ -359,7 +425,9 @@ export async function getNextAvailableOrgSubnet(): Promise<string> {
|
|||||||
return subnet;
|
return subnet;
|
||||||
}
|
}
|
||||||
|
|
||||||
export function generateRemoteSubnets(allSiteResources: SiteResource[]): string[] {
|
export function generateRemoteSubnets(
|
||||||
|
allSiteResources: SiteResource[]
|
||||||
|
): string[] {
|
||||||
const remoteSubnets = allSiteResources
|
const remoteSubnets = allSiteResources
|
||||||
.filter((sr) => {
|
.filter((sr) => {
|
||||||
if (sr.mode === "cidr") return true;
|
if (sr.mode === "cidr") return true;
|
||||||
|
|||||||
@@ -14,4 +14,4 @@ export async function logAccessAudit(data: {
|
|||||||
requestIp?: string;
|
requestIp?: string;
|
||||||
}) {
|
}) {
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -14,7 +14,8 @@ export const configSchema = z
|
|||||||
.object({
|
.object({
|
||||||
app: z
|
app: z
|
||||||
.object({
|
.object({
|
||||||
dashboard_url: z.url()
|
dashboard_url: z
|
||||||
|
.url()
|
||||||
.pipe(z.url())
|
.pipe(z.url())
|
||||||
.transform((url) => url.toLowerCase())
|
.transform((url) => url.toLowerCase())
|
||||||
.optional(),
|
.optional(),
|
||||||
@@ -255,7 +256,10 @@ export const configSchema = z
|
|||||||
.object({
|
.object({
|
||||||
block_size: z.number().positive().gt(0).optional().default(24),
|
block_size: z.number().positive().gt(0).optional().default(24),
|
||||||
subnet_group: z.string().optional().default("100.90.128.0/24"),
|
subnet_group: z.string().optional().default("100.90.128.0/24"),
|
||||||
utility_subnet_group: z.string().optional().default("100.96.128.0/24") //just hardcode this for now as well
|
utility_subnet_group: z
|
||||||
|
.string()
|
||||||
|
.optional()
|
||||||
|
.default("100.96.128.0/24") //just hardcode this for now as well
|
||||||
})
|
})
|
||||||
.optional()
|
.optional()
|
||||||
.default({
|
.default({
|
||||||
|
|||||||
@@ -24,7 +24,7 @@ import {
|
|||||||
deletePeer as newtDeletePeer
|
deletePeer as newtDeletePeer
|
||||||
} from "@server/routers/newt/peers";
|
} from "@server/routers/newt/peers";
|
||||||
import {
|
import {
|
||||||
initPeerAddHandshake as holepunchSiteAdd,
|
initPeerAddHandshake,
|
||||||
deletePeer as olmDeletePeer
|
deletePeer as olmDeletePeer
|
||||||
} from "@server/routers/olm/peers";
|
} from "@server/routers/olm/peers";
|
||||||
import { sendToExitNode } from "#dynamic/lib/exitNodes";
|
import { sendToExitNode } from "#dynamic/lib/exitNodes";
|
||||||
@@ -33,6 +33,8 @@ import {
|
|||||||
generateAliasConfig,
|
generateAliasConfig,
|
||||||
generateRemoteSubnets,
|
generateRemoteSubnets,
|
||||||
generateSubnetProxyTargets,
|
generateSubnetProxyTargets,
|
||||||
|
parseEndpoint,
|
||||||
|
formatEndpoint
|
||||||
} from "@server/lib/ip";
|
} from "@server/lib/ip";
|
||||||
import {
|
import {
|
||||||
addPeerData,
|
addPeerData,
|
||||||
@@ -109,21 +111,22 @@ export async function getClientSiteResourceAccess(
|
|||||||
const directClientIds = allClientSiteResources.map((row) => row.clientId);
|
const directClientIds = allClientSiteResources.map((row) => row.clientId);
|
||||||
|
|
||||||
// Get full client details for directly associated clients
|
// Get full client details for directly associated clients
|
||||||
const directClients = directClientIds.length > 0
|
const directClients =
|
||||||
? await trx
|
directClientIds.length > 0
|
||||||
.select({
|
? await trx
|
||||||
clientId: clients.clientId,
|
.select({
|
||||||
pubKey: clients.pubKey,
|
clientId: clients.clientId,
|
||||||
subnet: clients.subnet
|
pubKey: clients.pubKey,
|
||||||
})
|
subnet: clients.subnet
|
||||||
.from(clients)
|
})
|
||||||
.where(
|
.from(clients)
|
||||||
and(
|
.where(
|
||||||
inArray(clients.clientId, directClientIds),
|
and(
|
||||||
eq(clients.orgId, siteResource.orgId) // filter by org to prevent cross-org associations
|
inArray(clients.clientId, directClientIds),
|
||||||
|
eq(clients.orgId, siteResource.orgId) // filter by org to prevent cross-org associations
|
||||||
|
)
|
||||||
)
|
)
|
||||||
)
|
: [];
|
||||||
: [];
|
|
||||||
|
|
||||||
// Merge user-based clients with directly associated clients
|
// Merge user-based clients with directly associated clients
|
||||||
const allClientsMap = new Map(
|
const allClientsMap = new Map(
|
||||||
@@ -474,7 +477,7 @@ async function handleMessagesForSiteClients(
|
|||||||
}
|
}
|
||||||
|
|
||||||
if (isAdd) {
|
if (isAdd) {
|
||||||
await holepunchSiteAdd(
|
await initPeerAddHandshake(
|
||||||
// this will kick off the add peer process for the client
|
// this will kick off the add peer process for the client
|
||||||
client.clientId,
|
client.clientId,
|
||||||
{
|
{
|
||||||
@@ -541,6 +544,17 @@ export async function updateClientSiteDestinations(
|
|||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Parse the endpoint properly for both IPv4 and IPv6
|
||||||
|
const parsedEndpoint = parseEndpoint(
|
||||||
|
site.clientSitesAssociationsCache.endpoint
|
||||||
|
);
|
||||||
|
if (!parsedEndpoint) {
|
||||||
|
logger.warn(
|
||||||
|
`Failed to parse endpoint ${site.clientSitesAssociationsCache.endpoint}, skipping`
|
||||||
|
);
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
// find the destinations in the array
|
// find the destinations in the array
|
||||||
let destinations = exitNodeDestinations.find(
|
let destinations = exitNodeDestinations.find(
|
||||||
(d) => d.reachableAt === site.exitNodes?.reachableAt
|
(d) => d.reachableAt === site.exitNodes?.reachableAt
|
||||||
@@ -552,13 +566,8 @@ export async function updateClientSiteDestinations(
|
|||||||
exitNodeId: site.exitNodes?.exitNodeId || 0,
|
exitNodeId: site.exitNodes?.exitNodeId || 0,
|
||||||
type: site.exitNodes?.type || "",
|
type: site.exitNodes?.type || "",
|
||||||
name: site.exitNodes?.name || "",
|
name: site.exitNodes?.name || "",
|
||||||
sourceIp:
|
sourceIp: parsedEndpoint.ip,
|
||||||
site.clientSitesAssociationsCache.endpoint.split(":")[0] ||
|
sourcePort: parsedEndpoint.port,
|
||||||
"",
|
|
||||||
sourcePort:
|
|
||||||
parseInt(
|
|
||||||
site.clientSitesAssociationsCache.endpoint.split(":")[1]
|
|
||||||
) || 0,
|
|
||||||
destinations: [
|
destinations: [
|
||||||
{
|
{
|
||||||
destinationIP: site.sites.subnet.split("/")[0],
|
destinationIP: site.sites.subnet.split("/")[0],
|
||||||
@@ -701,11 +710,46 @@ async function handleSubnetProxyTargetUpdates(
|
|||||||
}
|
}
|
||||||
|
|
||||||
for (const client of removedClients) {
|
for (const client of removedClients) {
|
||||||
|
// Check if this client still has access to another resource on this site with the same destination
|
||||||
|
const destinationStillInUse = await trx
|
||||||
|
.select()
|
||||||
|
.from(siteResources)
|
||||||
|
.innerJoin(
|
||||||
|
clientSiteResourcesAssociationsCache,
|
||||||
|
eq(
|
||||||
|
clientSiteResourcesAssociationsCache.siteResourceId,
|
||||||
|
siteResources.siteResourceId
|
||||||
|
)
|
||||||
|
)
|
||||||
|
.where(
|
||||||
|
and(
|
||||||
|
eq(
|
||||||
|
clientSiteResourcesAssociationsCache.clientId,
|
||||||
|
client.clientId
|
||||||
|
),
|
||||||
|
eq(siteResources.siteId, siteResource.siteId),
|
||||||
|
eq(
|
||||||
|
siteResources.destination,
|
||||||
|
siteResource.destination
|
||||||
|
),
|
||||||
|
ne(
|
||||||
|
siteResources.siteResourceId,
|
||||||
|
siteResource.siteResourceId
|
||||||
|
)
|
||||||
|
)
|
||||||
|
);
|
||||||
|
|
||||||
|
// Only remove remote subnet if no other resource uses the same destination
|
||||||
|
const remoteSubnetsToRemove =
|
||||||
|
destinationStillInUse.length > 0
|
||||||
|
? []
|
||||||
|
: generateRemoteSubnets([siteResource]);
|
||||||
|
|
||||||
olmJobs.push(
|
olmJobs.push(
|
||||||
removePeerData(
|
removePeerData(
|
||||||
client.clientId,
|
client.clientId,
|
||||||
siteResource.siteId,
|
siteResource.siteId,
|
||||||
generateRemoteSubnets([siteResource]),
|
remoteSubnetsToRemove,
|
||||||
generateAliasConfig([siteResource])
|
generateAliasConfig([siteResource])
|
||||||
)
|
)
|
||||||
);
|
);
|
||||||
@@ -783,7 +827,10 @@ export async function rebuildClientAssociationsFromClient(
|
|||||||
.from(roleSiteResources)
|
.from(roleSiteResources)
|
||||||
.innerJoin(
|
.innerJoin(
|
||||||
siteResources,
|
siteResources,
|
||||||
eq(siteResources.siteResourceId, roleSiteResources.siteResourceId)
|
eq(
|
||||||
|
siteResources.siteResourceId,
|
||||||
|
roleSiteResources.siteResourceId
|
||||||
|
)
|
||||||
)
|
)
|
||||||
.where(
|
.where(
|
||||||
and(
|
and(
|
||||||
@@ -1038,7 +1085,7 @@ async function handleMessagesForClientSites(
|
|||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
|
|
||||||
await holepunchSiteAdd(
|
await initPeerAddHandshake(
|
||||||
// this will kick off the add peer process for the client
|
// this will kick off the add peer process for the client
|
||||||
client.clientId,
|
client.clientId,
|
||||||
{
|
{
|
||||||
@@ -1213,12 +1260,47 @@ async function handleMessagesForClientResources(
|
|||||||
}
|
}
|
||||||
|
|
||||||
try {
|
try {
|
||||||
|
// Check if this client still has access to another resource on this site with the same destination
|
||||||
|
const destinationStillInUse = await trx
|
||||||
|
.select()
|
||||||
|
.from(siteResources)
|
||||||
|
.innerJoin(
|
||||||
|
clientSiteResourcesAssociationsCache,
|
||||||
|
eq(
|
||||||
|
clientSiteResourcesAssociationsCache.siteResourceId,
|
||||||
|
siteResources.siteResourceId
|
||||||
|
)
|
||||||
|
)
|
||||||
|
.where(
|
||||||
|
and(
|
||||||
|
eq(
|
||||||
|
clientSiteResourcesAssociationsCache.clientId,
|
||||||
|
client.clientId
|
||||||
|
),
|
||||||
|
eq(siteResources.siteId, resource.siteId),
|
||||||
|
eq(
|
||||||
|
siteResources.destination,
|
||||||
|
resource.destination
|
||||||
|
),
|
||||||
|
ne(
|
||||||
|
siteResources.siteResourceId,
|
||||||
|
resource.siteResourceId
|
||||||
|
)
|
||||||
|
)
|
||||||
|
);
|
||||||
|
|
||||||
|
// Only remove remote subnet if no other resource uses the same destination
|
||||||
|
const remoteSubnetsToRemove =
|
||||||
|
destinationStillInUse.length > 0
|
||||||
|
? []
|
||||||
|
: generateRemoteSubnets([resource]);
|
||||||
|
|
||||||
// Remove peer data from olm
|
// Remove peer data from olm
|
||||||
olmJobs.push(
|
olmJobs.push(
|
||||||
removePeerData(
|
removePeerData(
|
||||||
client.clientId,
|
client.clientId,
|
||||||
resource.siteId,
|
resource.siteId,
|
||||||
generateRemoteSubnets([resource]),
|
remoteSubnetsToRemove,
|
||||||
generateAliasConfig([resource])
|
generateAliasConfig([resource])
|
||||||
)
|
)
|
||||||
);
|
);
|
||||||
|
|||||||
@@ -1,8 +1,8 @@
|
|||||||
export enum AudienceIds {
|
export enum AudienceIds {
|
||||||
SignUps = "",
|
SignUps = "",
|
||||||
Subscribed = "",
|
Subscribed = "",
|
||||||
Churned = "",
|
Churned = "",
|
||||||
Newsletter = ""
|
Newsletter = ""
|
||||||
}
|
}
|
||||||
|
|
||||||
let resend;
|
let resend;
|
||||||
|
|||||||
@@ -3,14 +3,14 @@ import { Response } from "express";
|
|||||||
|
|
||||||
export const response = <T>(
|
export const response = <T>(
|
||||||
res: Response,
|
res: Response,
|
||||||
{ data, success, error, message, status }: ResponseT<T>,
|
{ data, success, error, message, status }: ResponseT<T>
|
||||||
) => {
|
) => {
|
||||||
return res.status(status).send({
|
return res.status(status).send({
|
||||||
data,
|
data,
|
||||||
success,
|
success,
|
||||||
error,
|
error,
|
||||||
message,
|
message,
|
||||||
status,
|
status
|
||||||
});
|
});
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|||||||
@@ -1,5 +1,5 @@
|
|||||||
import { S3Client } from "@aws-sdk/client-s3";
|
import { S3Client } from "@aws-sdk/client-s3";
|
||||||
|
|
||||||
export const s3Client = new S3Client({
|
export const s3Client = new S3Client({
|
||||||
region: process.env.S3_REGION || "us-east-1",
|
region: process.env.S3_REGION || "us-east-1"
|
||||||
});
|
});
|
||||||
|
|||||||
@@ -6,7 +6,7 @@ let serverIp: string | null = null;
|
|||||||
const services = [
|
const services = [
|
||||||
"https://checkip.amazonaws.com",
|
"https://checkip.amazonaws.com",
|
||||||
"https://ifconfig.io/ip",
|
"https://ifconfig.io/ip",
|
||||||
"https://api.ipify.org",
|
"https://api.ipify.org"
|
||||||
];
|
];
|
||||||
|
|
||||||
export async function fetchServerIp() {
|
export async function fetchServerIp() {
|
||||||
@@ -17,7 +17,9 @@ export async function fetchServerIp() {
|
|||||||
logger.debug("Detected public IP: " + serverIp);
|
logger.debug("Detected public IP: " + serverIp);
|
||||||
return;
|
return;
|
||||||
} catch (err: any) {
|
} catch (err: any) {
|
||||||
console.warn(`Failed to fetch server IP from ${url}: ${err.message || err.code}`);
|
console.warn(
|
||||||
|
`Failed to fetch server IP from ${url}: ${err.message || err.code}`
|
||||||
|
);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -1,8 +1,7 @@
|
|||||||
export default function stoi(val: any) {
|
export default function stoi(val: any) {
|
||||||
if (typeof val === "string") {
|
if (typeof val === "string") {
|
||||||
return parseInt(val);
|
return parseInt(val);
|
||||||
|
} else {
|
||||||
|
return val;
|
||||||
}
|
}
|
||||||
else {
|
}
|
||||||
return val;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|||||||
@@ -2,9 +2,9 @@ import { PostHog } from "posthog-node";
|
|||||||
import config from "./config";
|
import config from "./config";
|
||||||
import { getHostMeta } from "./hostMeta";
|
import { getHostMeta } from "./hostMeta";
|
||||||
import logger from "@server/logger";
|
import logger from "@server/logger";
|
||||||
import { apiKeys, db, roles } from "@server/db";
|
import { apiKeys, db, roles, siteResources } from "@server/db";
|
||||||
import { sites, users, orgs, resources, clients, idp } from "@server/db";
|
import { sites, users, orgs, resources, clients, idp } from "@server/db";
|
||||||
import { eq, count, notInArray, and } from "drizzle-orm";
|
import { eq, count, notInArray, and, isNotNull, isNull } from "drizzle-orm";
|
||||||
import { APP_VERSION } from "./consts";
|
import { APP_VERSION } from "./consts";
|
||||||
import crypto from "crypto";
|
import crypto from "crypto";
|
||||||
import { UserType } from "@server/types/UserTypes";
|
import { UserType } from "@server/types/UserTypes";
|
||||||
@@ -25,7 +25,7 @@ class TelemetryClient {
|
|||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
if (build !== "oss") {
|
if (build === "saas") {
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -41,14 +41,18 @@ class TelemetryClient {
|
|||||||
this.client?.shutdown();
|
this.client?.shutdown();
|
||||||
});
|
});
|
||||||
|
|
||||||
this.sendStartupEvents().catch((err) => {
|
this.sendStartupEvents()
|
||||||
logger.error("Failed to send startup telemetry:", err);
|
.catch((err) => {
|
||||||
});
|
logger.error("Failed to send startup telemetry:", err);
|
||||||
|
})
|
||||||
|
.then(() => {
|
||||||
|
logger.debug("Successfully sent startup telemetry data");
|
||||||
|
});
|
||||||
|
|
||||||
this.startAnalyticsInterval();
|
this.startAnalyticsInterval();
|
||||||
|
|
||||||
logger.info(
|
logger.info(
|
||||||
"Pangolin now gathers anonymous usage data to help us better understand how the software is used and guide future improvements and feature development. You can find more details, including instructions for opting out of this anonymous data collection, at: https://docs.pangolin.net/telemetry"
|
"Pangolin gathers anonymous usage data to help us better understand how the software is used and guide future improvements and feature development. You can find more details, including instructions for opting out of this anonymous data collection, at: https://docs.pangolin.net/telemetry"
|
||||||
);
|
);
|
||||||
} else if (!this.enabled) {
|
} else if (!this.enabled) {
|
||||||
logger.info(
|
logger.info(
|
||||||
@@ -60,9 +64,13 @@ class TelemetryClient {
|
|||||||
private startAnalyticsInterval() {
|
private startAnalyticsInterval() {
|
||||||
this.intervalId = setInterval(
|
this.intervalId = setInterval(
|
||||||
() => {
|
() => {
|
||||||
this.collectAndSendAnalytics().catch((err) => {
|
this.collectAndSendAnalytics()
|
||||||
logger.error("Failed to collect analytics:", err);
|
.catch((err) => {
|
||||||
});
|
logger.error("Failed to collect analytics:", err);
|
||||||
|
})
|
||||||
|
.then(() => {
|
||||||
|
logger.debug("Successfully sent analytics data");
|
||||||
|
});
|
||||||
},
|
},
|
||||||
48 * 60 * 60 * 1000
|
48 * 60 * 60 * 1000
|
||||||
);
|
);
|
||||||
@@ -99,9 +107,14 @@ class TelemetryClient {
|
|||||||
const [resourcesCount] = await db
|
const [resourcesCount] = await db
|
||||||
.select({ count: count() })
|
.select({ count: count() })
|
||||||
.from(resources);
|
.from(resources);
|
||||||
const [clientsCount] = await db
|
const [userDevicesCount] = await db
|
||||||
.select({ count: count() })
|
.select({ count: count() })
|
||||||
.from(clients);
|
.from(clients)
|
||||||
|
.where(isNotNull(clients.userId));
|
||||||
|
const [machineClients] = await db
|
||||||
|
.select({ count: count() })
|
||||||
|
.from(clients)
|
||||||
|
.where(isNull(clients.userId));
|
||||||
const [idpCount] = await db.select({ count: count() }).from(idp);
|
const [idpCount] = await db.select({ count: count() }).from(idp);
|
||||||
const [onlineSitesCount] = await db
|
const [onlineSitesCount] = await db
|
||||||
.select({ count: count() })
|
.select({ count: count() })
|
||||||
@@ -146,6 +159,24 @@ class TelemetryClient {
|
|||||||
|
|
||||||
const supporterKey = config.getSupporterData();
|
const supporterKey = config.getSupporterData();
|
||||||
|
|
||||||
|
const allPrivateResources = await db.select().from(siteResources);
|
||||||
|
|
||||||
|
const numPrivResources = allPrivateResources.length;
|
||||||
|
let numPrivResourceAliases = 0;
|
||||||
|
let numPrivResourceHosts = 0;
|
||||||
|
let numPrivResourceCidr = 0;
|
||||||
|
for (const res of allPrivateResources) {
|
||||||
|
if (res.mode === "host") {
|
||||||
|
numPrivResourceHosts += 1;
|
||||||
|
} else if (res.mode === "cidr") {
|
||||||
|
numPrivResourceCidr += 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (res.alias) {
|
||||||
|
numPrivResourceAliases += 1;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
return {
|
return {
|
||||||
numSites: sitesCount.count,
|
numSites: sitesCount.count,
|
||||||
numUsers: usersCount.count,
|
numUsers: usersCount.count,
|
||||||
@@ -153,7 +184,11 @@ class TelemetryClient {
|
|||||||
numUsersOidc: usersOidcCount.count,
|
numUsersOidc: usersOidcCount.count,
|
||||||
numOrganizations: orgsCount.count,
|
numOrganizations: orgsCount.count,
|
||||||
numResources: resourcesCount.count,
|
numResources: resourcesCount.count,
|
||||||
numClients: clientsCount.count,
|
numPrivateResources: numPrivResources,
|
||||||
|
numPrivateResourceAliases: numPrivResourceAliases,
|
||||||
|
numPrivateResourceHosts: numPrivResourceHosts,
|
||||||
|
numUserDevices: userDevicesCount.count,
|
||||||
|
numMachineClients: machineClients.count,
|
||||||
numIdentityProviders: idpCount.count,
|
numIdentityProviders: idpCount.count,
|
||||||
numSitesOnline: onlineSitesCount.count,
|
numSitesOnline: onlineSitesCount.count,
|
||||||
resources: resourceDetails,
|
resources: resourceDetails,
|
||||||
@@ -196,7 +231,7 @@ class TelemetryClient {
|
|||||||
logger.debug("Sending enterprise startup telemetry payload:", {
|
logger.debug("Sending enterprise startup telemetry payload:", {
|
||||||
payload
|
payload
|
||||||
});
|
});
|
||||||
// this.client.capture(payload);
|
this.client.capture(payload);
|
||||||
}
|
}
|
||||||
|
|
||||||
if (build === "oss") {
|
if (build === "oss") {
|
||||||
@@ -246,7 +281,12 @@ class TelemetryClient {
|
|||||||
num_users_oidc: stats.numUsersOidc,
|
num_users_oidc: stats.numUsersOidc,
|
||||||
num_organizations: stats.numOrganizations,
|
num_organizations: stats.numOrganizations,
|
||||||
num_resources: stats.numResources,
|
num_resources: stats.numResources,
|
||||||
num_clients: stats.numClients,
|
num_private_resources: stats.numPrivateResources,
|
||||||
|
num_private_resource_aliases:
|
||||||
|
stats.numPrivateResourceAliases,
|
||||||
|
num_private_resource_hosts: stats.numPrivateResourceHosts,
|
||||||
|
num_user_devices: stats.numUserDevices,
|
||||||
|
num_machine_clients: stats.numMachineClients,
|
||||||
num_identity_providers: stats.numIdentityProviders,
|
num_identity_providers: stats.numIdentityProviders,
|
||||||
num_sites_online: stats.numSitesOnline,
|
num_sites_online: stats.numSitesOnline,
|
||||||
num_resources_sso_enabled: stats.resources.filter(
|
num_resources_sso_enabled: stats.resources.filter(
|
||||||
|
|||||||
@@ -195,7 +195,9 @@ export class TraefikConfigManager {
|
|||||||
|
|
||||||
state.set(domain, {
|
state.set(domain, {
|
||||||
exists: certExists && keyExists,
|
exists: certExists && keyExists,
|
||||||
lastModified: lastModified ? Math.floor(lastModified.getTime() / 1000) : null,
|
lastModified: lastModified
|
||||||
|
? Math.floor(lastModified.getTime() / 1000)
|
||||||
|
: null,
|
||||||
expiresAt,
|
expiresAt,
|
||||||
wildcard
|
wildcard
|
||||||
});
|
});
|
||||||
@@ -464,7 +466,9 @@ export class TraefikConfigManager {
|
|||||||
config.getRawConfig().traefik.site_types,
|
config.getRawConfig().traefik.site_types,
|
||||||
build == "oss", // filter out the namespace domains in open source
|
build == "oss", // filter out the namespace domains in open source
|
||||||
build != "oss", // generate the login pages on the cloud and hybrid,
|
build != "oss", // generate the login pages on the cloud and hybrid,
|
||||||
build == "saas" ? false : config.getRawConfig().traefik.allow_raw_resources // dont allow raw resources on saas otherwise use config
|
build == "saas"
|
||||||
|
? false
|
||||||
|
: config.getRawConfig().traefik.allow_raw_resources // dont allow raw resources on saas otherwise use config
|
||||||
);
|
);
|
||||||
|
|
||||||
const domains = new Set<string>();
|
const domains = new Set<string>();
|
||||||
@@ -786,29 +790,30 @@ export class TraefikConfigManager {
|
|||||||
"utf8"
|
"utf8"
|
||||||
);
|
);
|
||||||
|
|
||||||
// Store the certificate expiry time
|
|
||||||
if (cert.expiresAt) {
|
|
||||||
const expiresAtPath = path.join(domainDir, ".expires_at");
|
|
||||||
fs.writeFileSync(
|
|
||||||
expiresAtPath,
|
|
||||||
cert.expiresAt.toString(),
|
|
||||||
"utf8"
|
|
||||||
);
|
|
||||||
}
|
|
||||||
|
|
||||||
logger.info(
|
logger.info(
|
||||||
`Certificate updated for domain: ${cert.domain}${cert.wildcard ? " (wildcard)" : ""}`
|
`Certificate updated for domain: ${cert.domain}${cert.wildcard ? " (wildcard)" : ""}`
|
||||||
);
|
);
|
||||||
|
|
||||||
// Update local state tracking
|
|
||||||
this.lastLocalCertificateState.set(cert.domain, {
|
|
||||||
exists: true,
|
|
||||||
lastModified: Math.floor(Date.now() / 1000),
|
|
||||||
expiresAt: cert.expiresAt,
|
|
||||||
wildcard: cert.wildcard
|
|
||||||
});
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Always update expiry tracking when we fetch a certificate,
|
||||||
|
// even if the cert content didn't change
|
||||||
|
if (cert.expiresAt) {
|
||||||
|
const expiresAtPath = path.join(domainDir, ".expires_at");
|
||||||
|
fs.writeFileSync(
|
||||||
|
expiresAtPath,
|
||||||
|
cert.expiresAt.toString(),
|
||||||
|
"utf8"
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
// Update local state tracking
|
||||||
|
this.lastLocalCertificateState.set(cert.domain, {
|
||||||
|
exists: true,
|
||||||
|
lastModified: Math.floor(Date.now() / 1000),
|
||||||
|
expiresAt: cert.expiresAt,
|
||||||
|
wildcard: cert.wildcard
|
||||||
|
});
|
||||||
|
|
||||||
// Always ensure the config entry exists and is up to date
|
// Always ensure the config entry exists and is up to date
|
||||||
const certEntry = {
|
const certEntry = {
|
||||||
certFile: certPath,
|
certFile: certPath,
|
||||||
|
|||||||
@@ -1 +1 @@
|
|||||||
export * from "./getTraefikConfig";
|
export * from "./getTraefikConfig";
|
||||||
|
|||||||
@@ -2,234 +2,249 @@ import { assertEquals } from "@test/assert";
|
|||||||
import { isDomainCoveredByWildcard } from "./TraefikConfigManager";
|
import { isDomainCoveredByWildcard } from "./TraefikConfigManager";
|
||||||
|
|
||||||
function runTests() {
|
function runTests() {
|
||||||
console.log('Running wildcard domain coverage tests...');
|
console.log("Running wildcard domain coverage tests...");
|
||||||
|
|
||||||
// Test case 1: Basic wildcard certificate at example.com
|
// Test case 1: Basic wildcard certificate at example.com
|
||||||
const basicWildcardCerts = new Map([
|
const basicWildcardCerts = new Map([
|
||||||
['example.com', { exists: true, wildcard: true }]
|
["example.com", { exists: true, wildcard: true }]
|
||||||
]);
|
]);
|
||||||
|
|
||||||
// Should match first-level subdomains
|
// Should match first-level subdomains
|
||||||
assertEquals(
|
assertEquals(
|
||||||
isDomainCoveredByWildcard('level1.example.com', basicWildcardCerts),
|
isDomainCoveredByWildcard("level1.example.com", basicWildcardCerts),
|
||||||
true,
|
true,
|
||||||
'Wildcard cert at example.com should match level1.example.com'
|
"Wildcard cert at example.com should match level1.example.com"
|
||||||
);
|
);
|
||||||
|
|
||||||
assertEquals(
|
assertEquals(
|
||||||
isDomainCoveredByWildcard('api.example.com', basicWildcardCerts),
|
isDomainCoveredByWildcard("api.example.com", basicWildcardCerts),
|
||||||
true,
|
true,
|
||||||
'Wildcard cert at example.com should match api.example.com'
|
"Wildcard cert at example.com should match api.example.com"
|
||||||
);
|
);
|
||||||
|
|
||||||
assertEquals(
|
assertEquals(
|
||||||
isDomainCoveredByWildcard('www.example.com', basicWildcardCerts),
|
isDomainCoveredByWildcard("www.example.com", basicWildcardCerts),
|
||||||
true,
|
true,
|
||||||
'Wildcard cert at example.com should match www.example.com'
|
"Wildcard cert at example.com should match www.example.com"
|
||||||
);
|
);
|
||||||
|
|
||||||
// Should match the root domain (exact match)
|
// Should match the root domain (exact match)
|
||||||
assertEquals(
|
assertEquals(
|
||||||
isDomainCoveredByWildcard('example.com', basicWildcardCerts),
|
isDomainCoveredByWildcard("example.com", basicWildcardCerts),
|
||||||
true,
|
true,
|
||||||
'Wildcard cert at example.com should match example.com itself'
|
"Wildcard cert at example.com should match example.com itself"
|
||||||
);
|
);
|
||||||
|
|
||||||
// Should NOT match second-level subdomains
|
// Should NOT match second-level subdomains
|
||||||
assertEquals(
|
assertEquals(
|
||||||
isDomainCoveredByWildcard('level2.level1.example.com', basicWildcardCerts),
|
isDomainCoveredByWildcard(
|
||||||
|
"level2.level1.example.com",
|
||||||
|
basicWildcardCerts
|
||||||
|
),
|
||||||
false,
|
false,
|
||||||
'Wildcard cert at example.com should NOT match level2.level1.example.com'
|
"Wildcard cert at example.com should NOT match level2.level1.example.com"
|
||||||
);
|
);
|
||||||
|
|
||||||
assertEquals(
|
assertEquals(
|
||||||
isDomainCoveredByWildcard('deep.nested.subdomain.example.com', basicWildcardCerts),
|
isDomainCoveredByWildcard(
|
||||||
|
"deep.nested.subdomain.example.com",
|
||||||
|
basicWildcardCerts
|
||||||
|
),
|
||||||
false,
|
false,
|
||||||
'Wildcard cert at example.com should NOT match deep.nested.subdomain.example.com'
|
"Wildcard cert at example.com should NOT match deep.nested.subdomain.example.com"
|
||||||
);
|
);
|
||||||
|
|
||||||
// Should NOT match different domains
|
// Should NOT match different domains
|
||||||
assertEquals(
|
assertEquals(
|
||||||
isDomainCoveredByWildcard('test.otherdomain.com', basicWildcardCerts),
|
isDomainCoveredByWildcard("test.otherdomain.com", basicWildcardCerts),
|
||||||
false,
|
false,
|
||||||
'Wildcard cert at example.com should NOT match test.otherdomain.com'
|
"Wildcard cert at example.com should NOT match test.otherdomain.com"
|
||||||
);
|
);
|
||||||
|
|
||||||
assertEquals(
|
assertEquals(
|
||||||
isDomainCoveredByWildcard('notexample.com', basicWildcardCerts),
|
isDomainCoveredByWildcard("notexample.com", basicWildcardCerts),
|
||||||
false,
|
false,
|
||||||
'Wildcard cert at example.com should NOT match notexample.com'
|
"Wildcard cert at example.com should NOT match notexample.com"
|
||||||
);
|
);
|
||||||
|
|
||||||
// Test case 2: Multiple wildcard certificates
|
// Test case 2: Multiple wildcard certificates
|
||||||
const multipleWildcardCerts = new Map([
|
const multipleWildcardCerts = new Map([
|
||||||
['example.com', { exists: true, wildcard: true }],
|
["example.com", { exists: true, wildcard: true }],
|
||||||
['test.org', { exists: true, wildcard: true }],
|
["test.org", { exists: true, wildcard: true }],
|
||||||
['api.service.net', { exists: true, wildcard: true }]
|
["api.service.net", { exists: true, wildcard: true }]
|
||||||
]);
|
]);
|
||||||
|
|
||||||
assertEquals(
|
assertEquals(
|
||||||
isDomainCoveredByWildcard('app.example.com', multipleWildcardCerts),
|
isDomainCoveredByWildcard("app.example.com", multipleWildcardCerts),
|
||||||
true,
|
true,
|
||||||
'Should match subdomain of first wildcard cert'
|
"Should match subdomain of first wildcard cert"
|
||||||
);
|
);
|
||||||
|
|
||||||
assertEquals(
|
assertEquals(
|
||||||
isDomainCoveredByWildcard('staging.test.org', multipleWildcardCerts),
|
isDomainCoveredByWildcard("staging.test.org", multipleWildcardCerts),
|
||||||
true,
|
true,
|
||||||
'Should match subdomain of second wildcard cert'
|
"Should match subdomain of second wildcard cert"
|
||||||
);
|
);
|
||||||
|
|
||||||
assertEquals(
|
assertEquals(
|
||||||
isDomainCoveredByWildcard('v1.api.service.net', multipleWildcardCerts),
|
isDomainCoveredByWildcard("v1.api.service.net", multipleWildcardCerts),
|
||||||
true,
|
true,
|
||||||
'Should match subdomain of third wildcard cert'
|
"Should match subdomain of third wildcard cert"
|
||||||
);
|
);
|
||||||
|
|
||||||
assertEquals(
|
assertEquals(
|
||||||
isDomainCoveredByWildcard('deep.nested.api.service.net', multipleWildcardCerts),
|
isDomainCoveredByWildcard(
|
||||||
|
"deep.nested.api.service.net",
|
||||||
|
multipleWildcardCerts
|
||||||
|
),
|
||||||
false,
|
false,
|
||||||
'Should NOT match multi-level subdomain of third wildcard cert'
|
"Should NOT match multi-level subdomain of third wildcard cert"
|
||||||
);
|
);
|
||||||
|
|
||||||
// Test exact domain matches for multiple certs
|
// Test exact domain matches for multiple certs
|
||||||
assertEquals(
|
assertEquals(
|
||||||
isDomainCoveredByWildcard('example.com', multipleWildcardCerts),
|
isDomainCoveredByWildcard("example.com", multipleWildcardCerts),
|
||||||
true,
|
true,
|
||||||
'Should match exact domain of first wildcard cert'
|
"Should match exact domain of first wildcard cert"
|
||||||
);
|
);
|
||||||
|
|
||||||
assertEquals(
|
assertEquals(
|
||||||
isDomainCoveredByWildcard('test.org', multipleWildcardCerts),
|
isDomainCoveredByWildcard("test.org", multipleWildcardCerts),
|
||||||
true,
|
true,
|
||||||
'Should match exact domain of second wildcard cert'
|
"Should match exact domain of second wildcard cert"
|
||||||
);
|
);
|
||||||
|
|
||||||
assertEquals(
|
assertEquals(
|
||||||
isDomainCoveredByWildcard('api.service.net', multipleWildcardCerts),
|
isDomainCoveredByWildcard("api.service.net", multipleWildcardCerts),
|
||||||
true,
|
true,
|
||||||
'Should match exact domain of third wildcard cert'
|
"Should match exact domain of third wildcard cert"
|
||||||
);
|
);
|
||||||
|
|
||||||
// Test case 3: Non-wildcard certificates (should not match anything)
|
// Test case 3: Non-wildcard certificates (should not match anything)
|
||||||
const nonWildcardCerts = new Map([
|
const nonWildcardCerts = new Map([
|
||||||
['example.com', { exists: true, wildcard: false }],
|
["example.com", { exists: true, wildcard: false }],
|
||||||
['specific.domain.com', { exists: true, wildcard: false }]
|
["specific.domain.com", { exists: true, wildcard: false }]
|
||||||
]);
|
]);
|
||||||
|
|
||||||
assertEquals(
|
assertEquals(
|
||||||
isDomainCoveredByWildcard('sub.example.com', nonWildcardCerts),
|
isDomainCoveredByWildcard("sub.example.com", nonWildcardCerts),
|
||||||
false,
|
false,
|
||||||
'Non-wildcard cert should not match subdomains'
|
"Non-wildcard cert should not match subdomains"
|
||||||
);
|
);
|
||||||
|
|
||||||
assertEquals(
|
assertEquals(
|
||||||
isDomainCoveredByWildcard('example.com', nonWildcardCerts),
|
isDomainCoveredByWildcard("example.com", nonWildcardCerts),
|
||||||
false,
|
false,
|
||||||
'Non-wildcard cert should not match even exact domain via this function'
|
"Non-wildcard cert should not match even exact domain via this function"
|
||||||
);
|
);
|
||||||
|
|
||||||
// Test case 4: Non-existent certificates (should not match)
|
// Test case 4: Non-existent certificates (should not match)
|
||||||
const nonExistentCerts = new Map([
|
const nonExistentCerts = new Map([
|
||||||
['example.com', { exists: false, wildcard: true }],
|
["example.com", { exists: false, wildcard: true }],
|
||||||
['missing.com', { exists: false, wildcard: true }]
|
["missing.com", { exists: false, wildcard: true }]
|
||||||
]);
|
]);
|
||||||
|
|
||||||
assertEquals(
|
assertEquals(
|
||||||
isDomainCoveredByWildcard('sub.example.com', nonExistentCerts),
|
isDomainCoveredByWildcard("sub.example.com", nonExistentCerts),
|
||||||
false,
|
false,
|
||||||
'Non-existent wildcard cert should not match'
|
"Non-existent wildcard cert should not match"
|
||||||
);
|
);
|
||||||
|
|
||||||
// Test case 5: Edge cases with special domain names
|
// Test case 5: Edge cases with special domain names
|
||||||
const specialDomainCerts = new Map([
|
const specialDomainCerts = new Map([
|
||||||
['localhost', { exists: true, wildcard: true }],
|
["localhost", { exists: true, wildcard: true }],
|
||||||
['127-0-0-1.nip.io', { exists: true, wildcard: true }],
|
["127-0-0-1.nip.io", { exists: true, wildcard: true }],
|
||||||
['xn--e1afmkfd.xn--p1ai', { exists: true, wildcard: true }] // IDN domain
|
["xn--e1afmkfd.xn--p1ai", { exists: true, wildcard: true }] // IDN domain
|
||||||
]);
|
]);
|
||||||
|
|
||||||
assertEquals(
|
assertEquals(
|
||||||
isDomainCoveredByWildcard('app.localhost', specialDomainCerts),
|
isDomainCoveredByWildcard("app.localhost", specialDomainCerts),
|
||||||
true,
|
true,
|
||||||
'Should match subdomain of localhost wildcard'
|
"Should match subdomain of localhost wildcard"
|
||||||
);
|
);
|
||||||
|
|
||||||
assertEquals(
|
assertEquals(
|
||||||
isDomainCoveredByWildcard('test.127-0-0-1.nip.io', specialDomainCerts),
|
isDomainCoveredByWildcard("test.127-0-0-1.nip.io", specialDomainCerts),
|
||||||
true,
|
true,
|
||||||
'Should match subdomain of nip.io wildcard'
|
"Should match subdomain of nip.io wildcard"
|
||||||
);
|
);
|
||||||
|
|
||||||
assertEquals(
|
assertEquals(
|
||||||
isDomainCoveredByWildcard('sub.xn--e1afmkfd.xn--p1ai', specialDomainCerts),
|
isDomainCoveredByWildcard(
|
||||||
|
"sub.xn--e1afmkfd.xn--p1ai",
|
||||||
|
specialDomainCerts
|
||||||
|
),
|
||||||
true,
|
true,
|
||||||
'Should match subdomain of IDN wildcard'
|
"Should match subdomain of IDN wildcard"
|
||||||
);
|
);
|
||||||
|
|
||||||
// Test case 6: Empty input and edge cases
|
// Test case 6: Empty input and edge cases
|
||||||
const emptyCerts = new Map();
|
const emptyCerts = new Map();
|
||||||
|
|
||||||
assertEquals(
|
assertEquals(
|
||||||
isDomainCoveredByWildcard('any.domain.com', emptyCerts),
|
isDomainCoveredByWildcard("any.domain.com", emptyCerts),
|
||||||
false,
|
false,
|
||||||
'Empty certificate map should not match any domain'
|
"Empty certificate map should not match any domain"
|
||||||
);
|
);
|
||||||
|
|
||||||
// Test case 7: Domains with single character components
|
// Test case 7: Domains with single character components
|
||||||
const singleCharCerts = new Map([
|
const singleCharCerts = new Map([
|
||||||
['a.com', { exists: true, wildcard: true }],
|
["a.com", { exists: true, wildcard: true }],
|
||||||
['x.y.z', { exists: true, wildcard: true }]
|
["x.y.z", { exists: true, wildcard: true }]
|
||||||
]);
|
]);
|
||||||
|
|
||||||
assertEquals(
|
assertEquals(
|
||||||
isDomainCoveredByWildcard('b.a.com', singleCharCerts),
|
isDomainCoveredByWildcard("b.a.com", singleCharCerts),
|
||||||
true,
|
true,
|
||||||
'Should match single character subdomain'
|
"Should match single character subdomain"
|
||||||
);
|
);
|
||||||
|
|
||||||
assertEquals(
|
assertEquals(
|
||||||
isDomainCoveredByWildcard('w.x.y.z', singleCharCerts),
|
isDomainCoveredByWildcard("w.x.y.z", singleCharCerts),
|
||||||
true,
|
true,
|
||||||
'Should match single character subdomain of multi-part domain'
|
"Should match single character subdomain of multi-part domain"
|
||||||
);
|
);
|
||||||
|
|
||||||
assertEquals(
|
assertEquals(
|
||||||
isDomainCoveredByWildcard('v.w.x.y.z', singleCharCerts),
|
isDomainCoveredByWildcard("v.w.x.y.z", singleCharCerts),
|
||||||
false,
|
false,
|
||||||
'Should NOT match multi-level subdomain of single char domain'
|
"Should NOT match multi-level subdomain of single char domain"
|
||||||
);
|
);
|
||||||
|
|
||||||
// Test case 8: Domains with numbers and hyphens
|
// Test case 8: Domains with numbers and hyphens
|
||||||
const numericCerts = new Map([
|
const numericCerts = new Map([
|
||||||
['api-v2.service-1.com', { exists: true, wildcard: true }],
|
["api-v2.service-1.com", { exists: true, wildcard: true }],
|
||||||
['123.456.net', { exists: true, wildcard: true }]
|
["123.456.net", { exists: true, wildcard: true }]
|
||||||
]);
|
]);
|
||||||
|
|
||||||
assertEquals(
|
assertEquals(
|
||||||
isDomainCoveredByWildcard('staging.api-v2.service-1.com', numericCerts),
|
isDomainCoveredByWildcard("staging.api-v2.service-1.com", numericCerts),
|
||||||
true,
|
true,
|
||||||
'Should match subdomain with hyphens and numbers'
|
"Should match subdomain with hyphens and numbers"
|
||||||
);
|
);
|
||||||
|
|
||||||
assertEquals(
|
assertEquals(
|
||||||
isDomainCoveredByWildcard('test.123.456.net', numericCerts),
|
isDomainCoveredByWildcard("test.123.456.net", numericCerts),
|
||||||
true,
|
true,
|
||||||
'Should match subdomain with numeric components'
|
"Should match subdomain with numeric components"
|
||||||
);
|
);
|
||||||
|
|
||||||
assertEquals(
|
assertEquals(
|
||||||
isDomainCoveredByWildcard('deep.staging.api-v2.service-1.com', numericCerts),
|
isDomainCoveredByWildcard(
|
||||||
|
"deep.staging.api-v2.service-1.com",
|
||||||
|
numericCerts
|
||||||
|
),
|
||||||
false,
|
false,
|
||||||
'Should NOT match multi-level subdomain with hyphens and numbers'
|
"Should NOT match multi-level subdomain with hyphens and numbers"
|
||||||
);
|
);
|
||||||
|
|
||||||
console.log('All wildcard domain coverage tests passed!');
|
console.log("All wildcard domain coverage tests passed!");
|
||||||
}
|
}
|
||||||
|
|
||||||
// Run all tests
|
// Run all tests
|
||||||
try {
|
try {
|
||||||
runTests();
|
runTests();
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
console.error('Test failed:', error);
|
console.error("Test failed:", error);
|
||||||
process.exit(1);
|
process.exit(1);
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -31,12 +31,17 @@ export function validatePathRewriteConfig(
|
|||||||
}
|
}
|
||||||
|
|
||||||
if (rewritePathType !== "stripPrefix") {
|
if (rewritePathType !== "stripPrefix") {
|
||||||
if ((rewritePath && !rewritePathType) || (!rewritePath && rewritePathType)) {
|
if (
|
||||||
return { isValid: false, error: "Both rewritePath and rewritePathType must be specified together" };
|
(rewritePath && !rewritePathType) ||
|
||||||
|
(!rewritePath && rewritePathType)
|
||||||
|
) {
|
||||||
|
return {
|
||||||
|
isValid: false,
|
||||||
|
error: "Both rewritePath and rewritePathType must be specified together"
|
||||||
|
};
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
if (!rewritePath || !rewritePathType) {
|
if (!rewritePath || !rewritePathType) {
|
||||||
return { isValid: true };
|
return { isValid: true };
|
||||||
}
|
}
|
||||||
@@ -68,14 +73,14 @@ export function validatePathRewriteConfig(
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
// Additional validation for stripPrefix
|
// Additional validation for stripPrefix
|
||||||
if (rewritePathType === "stripPrefix") {
|
if (rewritePathType === "stripPrefix") {
|
||||||
if (pathMatchType !== "prefix") {
|
if (pathMatchType !== "prefix") {
|
||||||
logger.warn(`stripPrefix rewrite type is most effective with prefix path matching. Current match type: ${pathMatchType}`);
|
logger.warn(
|
||||||
|
`stripPrefix rewrite type is most effective with prefix path matching. Current match type: ${pathMatchType}`
|
||||||
|
);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
return { isValid: true };
|
return { isValid: true };
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -1,71 +1,247 @@
|
|||||||
import { isValidUrlGlobPattern } from "./validators";
|
import { isValidUrlGlobPattern } from "./validators";
|
||||||
import { assertEquals } from "@test/assert";
|
import { assertEquals } from "@test/assert";
|
||||||
|
|
||||||
function runTests() {
|
function runTests() {
|
||||||
console.log('Running URL pattern validation tests...');
|
console.log("Running URL pattern validation tests...");
|
||||||
|
|
||||||
// Test valid patterns
|
// Test valid patterns
|
||||||
assertEquals(isValidUrlGlobPattern('simple'), true, 'Simple path segment should be valid');
|
assertEquals(
|
||||||
assertEquals(isValidUrlGlobPattern('simple/path'), true, 'Simple path with slash should be valid');
|
isValidUrlGlobPattern("simple"),
|
||||||
assertEquals(isValidUrlGlobPattern('/leading/slash'), true, 'Path with leading slash should be valid');
|
true,
|
||||||
assertEquals(isValidUrlGlobPattern('path/'), true, 'Path with trailing slash should be valid');
|
"Simple path segment should be valid"
|
||||||
assertEquals(isValidUrlGlobPattern('path/*'), true, 'Path with wildcard segment should be valid');
|
);
|
||||||
assertEquals(isValidUrlGlobPattern('*'), true, 'Single wildcard should be valid');
|
assertEquals(
|
||||||
assertEquals(isValidUrlGlobPattern('*/subpath'), true, 'Wildcard with subpath should be valid');
|
isValidUrlGlobPattern("simple/path"),
|
||||||
assertEquals(isValidUrlGlobPattern('path/*/more'), true, 'Path with wildcard in the middle should be valid');
|
true,
|
||||||
|
"Simple path with slash should be valid"
|
||||||
|
);
|
||||||
|
assertEquals(
|
||||||
|
isValidUrlGlobPattern("/leading/slash"),
|
||||||
|
true,
|
||||||
|
"Path with leading slash should be valid"
|
||||||
|
);
|
||||||
|
assertEquals(
|
||||||
|
isValidUrlGlobPattern("path/"),
|
||||||
|
true,
|
||||||
|
"Path with trailing slash should be valid"
|
||||||
|
);
|
||||||
|
assertEquals(
|
||||||
|
isValidUrlGlobPattern("path/*"),
|
||||||
|
true,
|
||||||
|
"Path with wildcard segment should be valid"
|
||||||
|
);
|
||||||
|
assertEquals(
|
||||||
|
isValidUrlGlobPattern("*"),
|
||||||
|
true,
|
||||||
|
"Single wildcard should be valid"
|
||||||
|
);
|
||||||
|
assertEquals(
|
||||||
|
isValidUrlGlobPattern("*/subpath"),
|
||||||
|
true,
|
||||||
|
"Wildcard with subpath should be valid"
|
||||||
|
);
|
||||||
|
assertEquals(
|
||||||
|
isValidUrlGlobPattern("path/*/more"),
|
||||||
|
true,
|
||||||
|
"Path with wildcard in the middle should be valid"
|
||||||
|
);
|
||||||
|
|
||||||
// Test with special characters
|
// Test with special characters
|
||||||
assertEquals(isValidUrlGlobPattern('path-with-dash'), true, 'Path with dash should be valid');
|
assertEquals(
|
||||||
assertEquals(isValidUrlGlobPattern('path_with_underscore'), true, 'Path with underscore should be valid');
|
isValidUrlGlobPattern("path-with-dash"),
|
||||||
assertEquals(isValidUrlGlobPattern('path.with.dots'), true, 'Path with dots should be valid');
|
true,
|
||||||
assertEquals(isValidUrlGlobPattern('path~with~tilde'), true, 'Path with tilde should be valid');
|
"Path with dash should be valid"
|
||||||
assertEquals(isValidUrlGlobPattern('path!with!exclamation'), true, 'Path with exclamation should be valid');
|
);
|
||||||
assertEquals(isValidUrlGlobPattern('path$with$dollar'), true, 'Path with dollar should be valid');
|
assertEquals(
|
||||||
assertEquals(isValidUrlGlobPattern('path&with&ersand'), true, 'Path with ampersand should be valid');
|
isValidUrlGlobPattern("path_with_underscore"),
|
||||||
assertEquals(isValidUrlGlobPattern("path'with'quote"), true, "Path with quote should be valid");
|
true,
|
||||||
assertEquals(isValidUrlGlobPattern('path(with)parentheses'), true, 'Path with parentheses should be valid');
|
"Path with underscore should be valid"
|
||||||
assertEquals(isValidUrlGlobPattern('path+with+plus'), true, 'Path with plus should be valid');
|
);
|
||||||
assertEquals(isValidUrlGlobPattern('path,with,comma'), true, 'Path with comma should be valid');
|
assertEquals(
|
||||||
assertEquals(isValidUrlGlobPattern('path;with;semicolon'), true, 'Path with semicolon should be valid');
|
isValidUrlGlobPattern("path.with.dots"),
|
||||||
assertEquals(isValidUrlGlobPattern('path=with=equals'), true, 'Path with equals should be valid');
|
true,
|
||||||
assertEquals(isValidUrlGlobPattern('path:with:colon'), true, 'Path with colon should be valid');
|
"Path with dots should be valid"
|
||||||
assertEquals(isValidUrlGlobPattern('path@with@at'), true, 'Path with at should be valid');
|
);
|
||||||
|
assertEquals(
|
||||||
|
isValidUrlGlobPattern("path~with~tilde"),
|
||||||
|
true,
|
||||||
|
"Path with tilde should be valid"
|
||||||
|
);
|
||||||
|
assertEquals(
|
||||||
|
isValidUrlGlobPattern("path!with!exclamation"),
|
||||||
|
true,
|
||||||
|
"Path with exclamation should be valid"
|
||||||
|
);
|
||||||
|
assertEquals(
|
||||||
|
isValidUrlGlobPattern("path$with$dollar"),
|
||||||
|
true,
|
||||||
|
"Path with dollar should be valid"
|
||||||
|
);
|
||||||
|
assertEquals(
|
||||||
|
isValidUrlGlobPattern("path&with&ersand"),
|
||||||
|
true,
|
||||||
|
"Path with ampersand should be valid"
|
||||||
|
);
|
||||||
|
assertEquals(
|
||||||
|
isValidUrlGlobPattern("path'with'quote"),
|
||||||
|
true,
|
||||||
|
"Path with quote should be valid"
|
||||||
|
);
|
||||||
|
assertEquals(
|
||||||
|
isValidUrlGlobPattern("path(with)parentheses"),
|
||||||
|
true,
|
||||||
|
"Path with parentheses should be valid"
|
||||||
|
);
|
||||||
|
assertEquals(
|
||||||
|
isValidUrlGlobPattern("path+with+plus"),
|
||||||
|
true,
|
||||||
|
"Path with plus should be valid"
|
||||||
|
);
|
||||||
|
assertEquals(
|
||||||
|
isValidUrlGlobPattern("path,with,comma"),
|
||||||
|
true,
|
||||||
|
"Path with comma should be valid"
|
||||||
|
);
|
||||||
|
assertEquals(
|
||||||
|
isValidUrlGlobPattern("path;with;semicolon"),
|
||||||
|
true,
|
||||||
|
"Path with semicolon should be valid"
|
||||||
|
);
|
||||||
|
assertEquals(
|
||||||
|
isValidUrlGlobPattern("path=with=equals"),
|
||||||
|
true,
|
||||||
|
"Path with equals should be valid"
|
||||||
|
);
|
||||||
|
assertEquals(
|
||||||
|
isValidUrlGlobPattern("path:with:colon"),
|
||||||
|
true,
|
||||||
|
"Path with colon should be valid"
|
||||||
|
);
|
||||||
|
assertEquals(
|
||||||
|
isValidUrlGlobPattern("path@with@at"),
|
||||||
|
true,
|
||||||
|
"Path with at should be valid"
|
||||||
|
);
|
||||||
|
|
||||||
// Test with percent encoding
|
// Test with percent encoding
|
||||||
assertEquals(isValidUrlGlobPattern('path%20with%20spaces'), true, 'Path with percent-encoded spaces should be valid');
|
assertEquals(
|
||||||
assertEquals(isValidUrlGlobPattern('path%2Fwith%2Fencoded%2Fslashes'), true, 'Path with percent-encoded slashes should be valid');
|
isValidUrlGlobPattern("path%20with%20spaces"),
|
||||||
|
true,
|
||||||
|
"Path with percent-encoded spaces should be valid"
|
||||||
|
);
|
||||||
|
assertEquals(
|
||||||
|
isValidUrlGlobPattern("path%2Fwith%2Fencoded%2Fslashes"),
|
||||||
|
true,
|
||||||
|
"Path with percent-encoded slashes should be valid"
|
||||||
|
);
|
||||||
|
|
||||||
// Test with wildcards in segments (the fixed functionality)
|
// Test with wildcards in segments (the fixed functionality)
|
||||||
assertEquals(isValidUrlGlobPattern('padbootstrap*'), true, 'Path with wildcard at the end of segment should be valid');
|
assertEquals(
|
||||||
assertEquals(isValidUrlGlobPattern('pad*bootstrap'), true, 'Path with wildcard in the middle of segment should be valid');
|
isValidUrlGlobPattern("padbootstrap*"),
|
||||||
assertEquals(isValidUrlGlobPattern('*bootstrap'), true, 'Path with wildcard at the start of segment should be valid');
|
true,
|
||||||
assertEquals(isValidUrlGlobPattern('multiple*wildcards*in*segment'), true, 'Path with multiple wildcards in segment should be valid');
|
"Path with wildcard at the end of segment should be valid"
|
||||||
assertEquals(isValidUrlGlobPattern('wild*/cards/in*/different/seg*ments'), true, 'Path with wildcards in different segments should be valid');
|
);
|
||||||
|
assertEquals(
|
||||||
|
isValidUrlGlobPattern("pad*bootstrap"),
|
||||||
|
true,
|
||||||
|
"Path with wildcard in the middle of segment should be valid"
|
||||||
|
);
|
||||||
|
assertEquals(
|
||||||
|
isValidUrlGlobPattern("*bootstrap"),
|
||||||
|
true,
|
||||||
|
"Path with wildcard at the start of segment should be valid"
|
||||||
|
);
|
||||||
|
assertEquals(
|
||||||
|
isValidUrlGlobPattern("multiple*wildcards*in*segment"),
|
||||||
|
true,
|
||||||
|
"Path with multiple wildcards in segment should be valid"
|
||||||
|
);
|
||||||
|
assertEquals(
|
||||||
|
isValidUrlGlobPattern("wild*/cards/in*/different/seg*ments"),
|
||||||
|
true,
|
||||||
|
"Path with wildcards in different segments should be valid"
|
||||||
|
);
|
||||||
|
|
||||||
// Test invalid patterns
|
// Test invalid patterns
|
||||||
assertEquals(isValidUrlGlobPattern(''), false, 'Empty string should be invalid');
|
assertEquals(
|
||||||
assertEquals(isValidUrlGlobPattern('//double/slash'), false, 'Path with double slash should be invalid');
|
isValidUrlGlobPattern(""),
|
||||||
assertEquals(isValidUrlGlobPattern('path//end'), false, 'Path with double slash in the middle should be invalid');
|
false,
|
||||||
assertEquals(isValidUrlGlobPattern('invalid<char>'), false, 'Path with invalid characters should be invalid');
|
"Empty string should be invalid"
|
||||||
assertEquals(isValidUrlGlobPattern('invalid|char'), false, 'Path with invalid pipe character should be invalid');
|
);
|
||||||
assertEquals(isValidUrlGlobPattern('invalid"char'), false, 'Path with invalid quote character should be invalid');
|
assertEquals(
|
||||||
assertEquals(isValidUrlGlobPattern('invalid`char'), false, 'Path with invalid backtick character should be invalid');
|
isValidUrlGlobPattern("//double/slash"),
|
||||||
assertEquals(isValidUrlGlobPattern('invalid^char'), false, 'Path with invalid caret character should be invalid');
|
false,
|
||||||
assertEquals(isValidUrlGlobPattern('invalid\\char'), false, 'Path with invalid backslash character should be invalid');
|
"Path with double slash should be invalid"
|
||||||
assertEquals(isValidUrlGlobPattern('invalid[char]'), false, 'Path with invalid square brackets should be invalid');
|
);
|
||||||
assertEquals(isValidUrlGlobPattern('invalid{char}'), false, 'Path with invalid curly braces should be invalid');
|
assertEquals(
|
||||||
|
isValidUrlGlobPattern("path//end"),
|
||||||
|
false,
|
||||||
|
"Path with double slash in the middle should be invalid"
|
||||||
|
);
|
||||||
|
assertEquals(
|
||||||
|
isValidUrlGlobPattern("invalid<char>"),
|
||||||
|
false,
|
||||||
|
"Path with invalid characters should be invalid"
|
||||||
|
);
|
||||||
|
assertEquals(
|
||||||
|
isValidUrlGlobPattern("invalid|char"),
|
||||||
|
false,
|
||||||
|
"Path with invalid pipe character should be invalid"
|
||||||
|
);
|
||||||
|
assertEquals(
|
||||||
|
isValidUrlGlobPattern('invalid"char'),
|
||||||
|
false,
|
||||||
|
"Path with invalid quote character should be invalid"
|
||||||
|
);
|
||||||
|
assertEquals(
|
||||||
|
isValidUrlGlobPattern("invalid`char"),
|
||||||
|
false,
|
||||||
|
"Path with invalid backtick character should be invalid"
|
||||||
|
);
|
||||||
|
assertEquals(
|
||||||
|
isValidUrlGlobPattern("invalid^char"),
|
||||||
|
false,
|
||||||
|
"Path with invalid caret character should be invalid"
|
||||||
|
);
|
||||||
|
assertEquals(
|
||||||
|
isValidUrlGlobPattern("invalid\\char"),
|
||||||
|
false,
|
||||||
|
"Path with invalid backslash character should be invalid"
|
||||||
|
);
|
||||||
|
assertEquals(
|
||||||
|
isValidUrlGlobPattern("invalid[char]"),
|
||||||
|
false,
|
||||||
|
"Path with invalid square brackets should be invalid"
|
||||||
|
);
|
||||||
|
assertEquals(
|
||||||
|
isValidUrlGlobPattern("invalid{char}"),
|
||||||
|
false,
|
||||||
|
"Path with invalid curly braces should be invalid"
|
||||||
|
);
|
||||||
|
|
||||||
// Test invalid percent encoding
|
// Test invalid percent encoding
|
||||||
assertEquals(isValidUrlGlobPattern('invalid%2'), false, 'Path with incomplete percent encoding should be invalid');
|
assertEquals(
|
||||||
assertEquals(isValidUrlGlobPattern('invalid%GZ'), false, 'Path with invalid hex in percent encoding should be invalid');
|
isValidUrlGlobPattern("invalid%2"),
|
||||||
assertEquals(isValidUrlGlobPattern('invalid%'), false, 'Path with isolated percent sign should be invalid');
|
false,
|
||||||
|
"Path with incomplete percent encoding should be invalid"
|
||||||
console.log('All tests passed!');
|
);
|
||||||
|
assertEquals(
|
||||||
|
isValidUrlGlobPattern("invalid%GZ"),
|
||||||
|
false,
|
||||||
|
"Path with invalid hex in percent encoding should be invalid"
|
||||||
|
);
|
||||||
|
assertEquals(
|
||||||
|
isValidUrlGlobPattern("invalid%"),
|
||||||
|
false,
|
||||||
|
"Path with isolated percent sign should be invalid"
|
||||||
|
);
|
||||||
|
|
||||||
|
console.log("All tests passed!");
|
||||||
}
|
}
|
||||||
|
|
||||||
// Run all tests
|
// Run all tests
|
||||||
try {
|
try {
|
||||||
runTests();
|
runTests();
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
console.error('Test failed:', error);
|
console.error("Test failed:", error);
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -2,7 +2,9 @@ import z from "zod";
|
|||||||
import ipaddr from "ipaddr.js";
|
import ipaddr from "ipaddr.js";
|
||||||
|
|
||||||
export function isValidCIDR(cidr: string): boolean {
|
export function isValidCIDR(cidr: string): boolean {
|
||||||
return z.cidrv4().safeParse(cidr).success || z.cidrv6().safeParse(cidr).success;
|
return (
|
||||||
|
z.cidrv4().safeParse(cidr).success || z.cidrv6().safeParse(cidr).success
|
||||||
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
export function isValidIP(ip: string): boolean {
|
export function isValidIP(ip: string): boolean {
|
||||||
@@ -69,11 +71,11 @@ export function isUrlValid(url: string | undefined) {
|
|||||||
if (!url) return true; // the link is optional in the schema so if it's empty it's valid
|
if (!url) return true; // the link is optional in the schema so if it's empty it's valid
|
||||||
var pattern = new RegExp(
|
var pattern = new RegExp(
|
||||||
"^(https?:\\/\\/)?" + // protocol
|
"^(https?:\\/\\/)?" + // protocol
|
||||||
"((([a-z\\d]([a-z\\d-]*[a-z\\d])*)\\.)+[a-z]{2,}|" + // domain name
|
"((([a-z\\d]([a-z\\d-]*[a-z\\d])*)\\.)+[a-z]{2,}|" + // domain name
|
||||||
"((\\d{1,3}\\.){3}\\d{1,3}))" + // OR ip (v4) address
|
"((\\d{1,3}\\.){3}\\d{1,3}))" + // OR ip (v4) address
|
||||||
"(\\:\\d+)?(\\/[-a-z\\d%_.~+]*)*" + // port and path
|
"(\\:\\d+)?(\\/[-a-z\\d%_.~+]*)*" + // port and path
|
||||||
"(\\?[;&a-z\\d%_.~+=-]*)?" + // query string
|
"(\\?[;&a-z\\d%_.~+=-]*)?" + // query string
|
||||||
"(\\#[-a-z\\d_]*)?$",
|
"(\\#[-a-z\\d_]*)?$",
|
||||||
"i"
|
"i"
|
||||||
);
|
);
|
||||||
return !!pattern.test(url);
|
return !!pattern.test(url);
|
||||||
@@ -168,14 +170,14 @@ export function validateHeaders(headers: string): boolean {
|
|||||||
}
|
}
|
||||||
|
|
||||||
export function isSecondLevelDomain(domain: string): boolean {
|
export function isSecondLevelDomain(domain: string): boolean {
|
||||||
if (!domain || typeof domain !== 'string') {
|
if (!domain || typeof domain !== "string") {
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
const trimmedDomain = domain.trim().toLowerCase();
|
const trimmedDomain = domain.trim().toLowerCase();
|
||||||
|
|
||||||
// Split into parts
|
// Split into parts
|
||||||
const parts = trimmedDomain.split('.');
|
const parts = trimmedDomain.split(".");
|
||||||
|
|
||||||
// Should have exactly 2 parts for a second-level domain (e.g., "example.com")
|
// Should have exactly 2 parts for a second-level domain (e.g., "example.com")
|
||||||
if (parts.length !== 2) {
|
if (parts.length !== 2) {
|
||||||
|
|||||||
@@ -20,6 +20,6 @@ export const errorHandlerMiddleware: ErrorRequestHandler = (
|
|||||||
error: true,
|
error: true,
|
||||||
message: error.message || "Internal Server Error",
|
message: error.message || "Internal Server Error",
|
||||||
status: statusCode,
|
status: statusCode,
|
||||||
stack: process.env.ENVIRONMENT === "prod" ? null : error.stack,
|
stack: process.env.ENVIRONMENT === "prod" ? null : error.stack
|
||||||
});
|
});
|
||||||
};
|
};
|
||||||
|
|||||||
@@ -8,13 +8,13 @@ import HttpCode from "@server/types/HttpCode";
|
|||||||
export async function getUserOrgs(
|
export async function getUserOrgs(
|
||||||
req: Request,
|
req: Request,
|
||||||
res: Response,
|
res: Response,
|
||||||
next: NextFunction,
|
next: NextFunction
|
||||||
) {
|
) {
|
||||||
const userId = req.user?.userId; // Assuming you have user information in the request
|
const userId = req.user?.userId; // Assuming you have user information in the request
|
||||||
|
|
||||||
if (!userId) {
|
if (!userId) {
|
||||||
return next(
|
return next(
|
||||||
createHttpError(HttpCode.UNAUTHORIZED, "User not authenticated"),
|
createHttpError(HttpCode.UNAUTHORIZED, "User not authenticated")
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -22,7 +22,7 @@ export async function getUserOrgs(
|
|||||||
const userOrganizations = await db
|
const userOrganizations = await db
|
||||||
.select({
|
.select({
|
||||||
orgId: userOrgs.orgId,
|
orgId: userOrgs.orgId,
|
||||||
roleId: userOrgs.roleId,
|
roleId: userOrgs.roleId
|
||||||
})
|
})
|
||||||
.from(userOrgs)
|
.from(userOrgs)
|
||||||
.where(eq(userOrgs.userId, userId));
|
.where(eq(userOrgs.userId, userId));
|
||||||
@@ -38,8 +38,8 @@ export async function getUserOrgs(
|
|||||||
next(
|
next(
|
||||||
createHttpError(
|
createHttpError(
|
||||||
HttpCode.INTERNAL_SERVER_ERROR,
|
HttpCode.INTERNAL_SERVER_ERROR,
|
||||||
"Error retrieving user organizations",
|
"Error retrieving user organizations"
|
||||||
),
|
)
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -12,4 +12,4 @@ export * from "./verifyAccessTokenAccess";
|
|||||||
export * from "./verifyApiKeyIsRoot";
|
export * from "./verifyApiKeyIsRoot";
|
||||||
export * from "./verifyApiKeyApiKeyAccess";
|
export * from "./verifyApiKeyApiKeyAccess";
|
||||||
export * from "./verifyApiKeyClientAccess";
|
export * from "./verifyApiKeyClientAccess";
|
||||||
export * from "./verifyApiKeySiteResourceAccess";
|
export * from "./verifyApiKeySiteResourceAccess";
|
||||||
|
|||||||
@@ -97,7 +97,6 @@ export async function verifyApiKeyAccessTokenAccess(
|
|||||||
);
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
return next();
|
return next();
|
||||||
} catch (e) {
|
} catch (e) {
|
||||||
return next(
|
return next(
|
||||||
|
|||||||
@@ -11,7 +11,7 @@ export async function verifyApiKeyApiKeyAccess(
|
|||||||
next: NextFunction
|
next: NextFunction
|
||||||
) {
|
) {
|
||||||
try {
|
try {
|
||||||
const {apiKey: callerApiKey } = req;
|
const { apiKey: callerApiKey } = req;
|
||||||
|
|
||||||
const apiKeyId =
|
const apiKeyId =
|
||||||
req.params.apiKeyId || req.body.apiKeyId || req.query.apiKeyId;
|
req.params.apiKeyId || req.body.apiKeyId || req.query.apiKeyId;
|
||||||
@@ -44,7 +44,10 @@ export async function verifyApiKeyApiKeyAccess(
|
|||||||
.select()
|
.select()
|
||||||
.from(apiKeyOrg)
|
.from(apiKeyOrg)
|
||||||
.where(
|
.where(
|
||||||
and(eq(apiKeys.apiKeyId, callerApiKey.apiKeyId), eq(apiKeyOrg.orgId, orgId))
|
and(
|
||||||
|
eq(apiKeys.apiKeyId, callerApiKey.apiKeyId),
|
||||||
|
eq(apiKeyOrg.orgId, orgId)
|
||||||
|
)
|
||||||
)
|
)
|
||||||
.limit(1);
|
.limit(1);
|
||||||
|
|
||||||
|
|||||||
@@ -11,9 +11,12 @@ export async function verifyApiKeySetResourceClients(
|
|||||||
next: NextFunction
|
next: NextFunction
|
||||||
) {
|
) {
|
||||||
const apiKey = req.apiKey;
|
const apiKey = req.apiKey;
|
||||||
const singleClientId = req.params.clientId || req.body.clientId || req.query.clientId;
|
const singleClientId =
|
||||||
|
req.params.clientId || req.body.clientId || req.query.clientId;
|
||||||
const { clientIds } = req.body;
|
const { clientIds } = req.body;
|
||||||
const allClientIds = clientIds || (singleClientId ? [parseInt(singleClientId as string)] : []);
|
const allClientIds =
|
||||||
|
clientIds ||
|
||||||
|
(singleClientId ? [parseInt(singleClientId as string)] : []);
|
||||||
|
|
||||||
if (!apiKey) {
|
if (!apiKey) {
|
||||||
return next(
|
return next(
|
||||||
@@ -70,4 +73,3 @@ export async function verifyApiKeySetResourceClients(
|
|||||||
);
|
);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -11,7 +11,8 @@ export async function verifyApiKeySetResourceUsers(
|
|||||||
next: NextFunction
|
next: NextFunction
|
||||||
) {
|
) {
|
||||||
const apiKey = req.apiKey;
|
const apiKey = req.apiKey;
|
||||||
const singleUserId = req.params.userId || req.body.userId || req.query.userId;
|
const singleUserId =
|
||||||
|
req.params.userId || req.body.userId || req.query.userId;
|
||||||
const { userIds } = req.body;
|
const { userIds } = req.body;
|
||||||
const allUserIds = userIds || (singleUserId ? [singleUserId] : []);
|
const allUserIds = userIds || (singleUserId ? [singleUserId] : []);
|
||||||
|
|
||||||
|
|||||||
@@ -38,17 +38,12 @@ export async function verifyApiKeySiteResourceAccess(
|
|||||||
const [siteResource] = await db
|
const [siteResource] = await db
|
||||||
.select()
|
.select()
|
||||||
.from(siteResources)
|
.from(siteResources)
|
||||||
.where(and(
|
.where(and(eq(siteResources.siteResourceId, siteResourceId)))
|
||||||
eq(siteResources.siteResourceId, siteResourceId)
|
|
||||||
))
|
|
||||||
.limit(1);
|
.limit(1);
|
||||||
|
|
||||||
if (!siteResource) {
|
if (!siteResource) {
|
||||||
return next(
|
return next(
|
||||||
createHttpError(
|
createHttpError(HttpCode.NOT_FOUND, "Site resource not found")
|
||||||
HttpCode.NOT_FOUND,
|
|
||||||
"Site resource not found"
|
|
||||||
)
|
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -5,7 +5,7 @@ import HttpCode from "@server/types/HttpCode";
|
|||||||
export function notFoundMiddleware(
|
export function notFoundMiddleware(
|
||||||
req: Request,
|
req: Request,
|
||||||
res: Response,
|
res: Response,
|
||||||
next: NextFunction,
|
next: NextFunction
|
||||||
) {
|
) {
|
||||||
if (req.path.startsWith("/api")) {
|
if (req.path.startsWith("/api")) {
|
||||||
const message = `The requests url is not found - ${req.originalUrl}`;
|
const message = `The requests url is not found - ${req.originalUrl}`;
|
||||||
|
|||||||
@@ -1,30 +1,32 @@
|
|||||||
import { Request, Response, NextFunction } from 'express';
|
import { Request, Response, NextFunction } from "express";
|
||||||
import logger from '@server/logger';
|
import logger from "@server/logger";
|
||||||
import createHttpError from 'http-errors';
|
import createHttpError from "http-errors";
|
||||||
import HttpCode from '@server/types/HttpCode';
|
import HttpCode from "@server/types/HttpCode";
|
||||||
|
|
||||||
export function requestTimeoutMiddleware(timeoutMs: number = 30000) {
|
export function requestTimeoutMiddleware(timeoutMs: number = 30000) {
|
||||||
return (req: Request, res: Response, next: NextFunction) => {
|
return (req: Request, res: Response, next: NextFunction) => {
|
||||||
// Set a timeout for the request
|
// Set a timeout for the request
|
||||||
const timeout = setTimeout(() => {
|
const timeout = setTimeout(() => {
|
||||||
if (!res.headersSent) {
|
if (!res.headersSent) {
|
||||||
logger.error(`Request timeout: ${req.method} ${req.url} from ${req.ip}`);
|
logger.error(
|
||||||
|
`Request timeout: ${req.method} ${req.url} from ${req.ip}`
|
||||||
|
);
|
||||||
return next(
|
return next(
|
||||||
createHttpError(
|
createHttpError(
|
||||||
HttpCode.REQUEST_TIMEOUT,
|
HttpCode.REQUEST_TIMEOUT,
|
||||||
'Request timeout - operation took too long to complete'
|
"Request timeout - operation took too long to complete"
|
||||||
)
|
)
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
}, timeoutMs);
|
}, timeoutMs);
|
||||||
|
|
||||||
// Clear timeout when response finishes
|
// Clear timeout when response finishes
|
||||||
res.on('finish', () => {
|
res.on("finish", () => {
|
||||||
clearTimeout(timeout);
|
clearTimeout(timeout);
|
||||||
});
|
});
|
||||||
|
|
||||||
// Clear timeout when response closes
|
// Clear timeout when response closes
|
||||||
res.on('close', () => {
|
res.on("close", () => {
|
||||||
clearTimeout(timeout);
|
clearTimeout(timeout);
|
||||||
});
|
});
|
||||||
|
|
||||||
|
|||||||
@@ -76,7 +76,10 @@ export async function verifySiteAccess(
|
|||||||
.select()
|
.select()
|
||||||
.from(userOrgs)
|
.from(userOrgs)
|
||||||
.where(
|
.where(
|
||||||
and(eq(userOrgs.userId, userId), eq(userOrgs.orgId, site.orgId))
|
and(
|
||||||
|
eq(userOrgs.userId, userId),
|
||||||
|
eq(userOrgs.orgId, site.orgId)
|
||||||
|
)
|
||||||
)
|
)
|
||||||
.limit(1);
|
.limit(1);
|
||||||
req.userOrg = userOrgRole[0];
|
req.userOrg = userOrgRole[0];
|
||||||
|
|||||||
@@ -9,7 +9,10 @@ const nextPort = config.getRawConfig().server.next_port;
|
|||||||
|
|
||||||
export async function createNextServer() {
|
export async function createNextServer() {
|
||||||
// const app = next({ dev });
|
// const app = next({ dev });
|
||||||
const app = next({ dev: process.env.ENVIRONMENT !== "prod", turbopack: true });
|
const app = next({
|
||||||
|
dev: process.env.ENVIRONMENT !== "prod",
|
||||||
|
turbopack: true
|
||||||
|
});
|
||||||
const handle = app.getRequestHandler();
|
const handle = app.getRequestHandler();
|
||||||
|
|
||||||
await app.prepare();
|
await app.prepare();
|
||||||
|
|||||||
@@ -11,11 +11,14 @@
|
|||||||
* This file is not licensed under the AGPLv3.
|
* This file is not licensed under the AGPLv3.
|
||||||
*/
|
*/
|
||||||
|
|
||||||
import {
|
import { encodeHexLowerCase } from "@oslojs/encoding";
|
||||||
encodeHexLowerCase,
|
|
||||||
} from "@oslojs/encoding";
|
|
||||||
import { sha256 } from "@oslojs/crypto/sha2";
|
import { sha256 } from "@oslojs/crypto/sha2";
|
||||||
import { RemoteExitNode, remoteExitNodes, remoteExitNodeSessions, RemoteExitNodeSession } from "@server/db";
|
import {
|
||||||
|
RemoteExitNode,
|
||||||
|
remoteExitNodes,
|
||||||
|
remoteExitNodeSessions,
|
||||||
|
RemoteExitNodeSession
|
||||||
|
} from "@server/db";
|
||||||
import { db } from "@server/db";
|
import { db } from "@server/db";
|
||||||
import { eq } from "drizzle-orm";
|
import { eq } from "drizzle-orm";
|
||||||
|
|
||||||
@@ -23,30 +26,39 @@ export const EXPIRES = 1000 * 60 * 60 * 24 * 30;
|
|||||||
|
|
||||||
export async function createRemoteExitNodeSession(
|
export async function createRemoteExitNodeSession(
|
||||||
token: string,
|
token: string,
|
||||||
remoteExitNodeId: string,
|
remoteExitNodeId: string
|
||||||
): Promise<RemoteExitNodeSession> {
|
): Promise<RemoteExitNodeSession> {
|
||||||
const sessionId = encodeHexLowerCase(
|
const sessionId = encodeHexLowerCase(
|
||||||
sha256(new TextEncoder().encode(token)),
|
sha256(new TextEncoder().encode(token))
|
||||||
);
|
);
|
||||||
const session: RemoteExitNodeSession = {
|
const session: RemoteExitNodeSession = {
|
||||||
sessionId: sessionId,
|
sessionId: sessionId,
|
||||||
remoteExitNodeId,
|
remoteExitNodeId,
|
||||||
expiresAt: new Date(Date.now() + EXPIRES).getTime(),
|
expiresAt: new Date(Date.now() + EXPIRES).getTime()
|
||||||
};
|
};
|
||||||
await db.insert(remoteExitNodeSessions).values(session);
|
await db.insert(remoteExitNodeSessions).values(session);
|
||||||
return session;
|
return session;
|
||||||
}
|
}
|
||||||
|
|
||||||
export async function validateRemoteExitNodeSessionToken(
|
export async function validateRemoteExitNodeSessionToken(
|
||||||
token: string,
|
token: string
|
||||||
): Promise<SessionValidationResult> {
|
): Promise<SessionValidationResult> {
|
||||||
const sessionId = encodeHexLowerCase(
|
const sessionId = encodeHexLowerCase(
|
||||||
sha256(new TextEncoder().encode(token)),
|
sha256(new TextEncoder().encode(token))
|
||||||
);
|
);
|
||||||
const result = await db
|
const result = await db
|
||||||
.select({ remoteExitNode: remoteExitNodes, session: remoteExitNodeSessions })
|
.select({
|
||||||
|
remoteExitNode: remoteExitNodes,
|
||||||
|
session: remoteExitNodeSessions
|
||||||
|
})
|
||||||
.from(remoteExitNodeSessions)
|
.from(remoteExitNodeSessions)
|
||||||
.innerJoin(remoteExitNodes, eq(remoteExitNodeSessions.remoteExitNodeId, remoteExitNodes.remoteExitNodeId))
|
.innerJoin(
|
||||||
|
remoteExitNodes,
|
||||||
|
eq(
|
||||||
|
remoteExitNodeSessions.remoteExitNodeId,
|
||||||
|
remoteExitNodes.remoteExitNodeId
|
||||||
|
)
|
||||||
|
)
|
||||||
.where(eq(remoteExitNodeSessions.sessionId, sessionId));
|
.where(eq(remoteExitNodeSessions.sessionId, sessionId));
|
||||||
if (result.length < 1) {
|
if (result.length < 1) {
|
||||||
return { session: null, remoteExitNode: null };
|
return { session: null, remoteExitNode: null };
|
||||||
@@ -58,26 +70,32 @@ export async function validateRemoteExitNodeSessionToken(
|
|||||||
.where(eq(remoteExitNodeSessions.sessionId, session.sessionId));
|
.where(eq(remoteExitNodeSessions.sessionId, session.sessionId));
|
||||||
return { session: null, remoteExitNode: null };
|
return { session: null, remoteExitNode: null };
|
||||||
}
|
}
|
||||||
if (Date.now() >= session.expiresAt - (EXPIRES / 2)) {
|
if (Date.now() >= session.expiresAt - EXPIRES / 2) {
|
||||||
session.expiresAt = new Date(
|
session.expiresAt = new Date(Date.now() + EXPIRES).getTime();
|
||||||
Date.now() + EXPIRES,
|
|
||||||
).getTime();
|
|
||||||
await db
|
await db
|
||||||
.update(remoteExitNodeSessions)
|
.update(remoteExitNodeSessions)
|
||||||
.set({
|
.set({
|
||||||
expiresAt: session.expiresAt,
|
expiresAt: session.expiresAt
|
||||||
})
|
})
|
||||||
.where(eq(remoteExitNodeSessions.sessionId, session.sessionId));
|
.where(eq(remoteExitNodeSessions.sessionId, session.sessionId));
|
||||||
}
|
}
|
||||||
return { session, remoteExitNode };
|
return { session, remoteExitNode };
|
||||||
}
|
}
|
||||||
|
|
||||||
export async function invalidateRemoteExitNodeSession(sessionId: string): Promise<void> {
|
export async function invalidateRemoteExitNodeSession(
|
||||||
await db.delete(remoteExitNodeSessions).where(eq(remoteExitNodeSessions.sessionId, sessionId));
|
sessionId: string
|
||||||
|
): Promise<void> {
|
||||||
|
await db
|
||||||
|
.delete(remoteExitNodeSessions)
|
||||||
|
.where(eq(remoteExitNodeSessions.sessionId, sessionId));
|
||||||
}
|
}
|
||||||
|
|
||||||
export async function invalidateAllRemoteExitNodeSessions(remoteExitNodeId: string): Promise<void> {
|
export async function invalidateAllRemoteExitNodeSessions(
|
||||||
await db.delete(remoteExitNodeSessions).where(eq(remoteExitNodeSessions.remoteExitNodeId, remoteExitNodeId));
|
remoteExitNodeId: string
|
||||||
|
): Promise<void> {
|
||||||
|
await db
|
||||||
|
.delete(remoteExitNodeSessions)
|
||||||
|
.where(eq(remoteExitNodeSessions.remoteExitNodeId, remoteExitNodeId));
|
||||||
}
|
}
|
||||||
|
|
||||||
export type SessionValidationResult =
|
export type SessionValidationResult =
|
||||||
|
|||||||
@@ -25,4 +25,4 @@ export async function initCleanup() {
|
|||||||
// Handle process termination
|
// Handle process termination
|
||||||
process.on("SIGTERM", () => cleanup());
|
process.on("SIGTERM", () => cleanup());
|
||||||
process.on("SIGINT", () => cleanup());
|
process.on("SIGINT", () => cleanup());
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -12,4 +12,4 @@
|
|||||||
*/
|
*/
|
||||||
|
|
||||||
export * from "./getOrgTierData";
|
export * from "./getOrgTierData";
|
||||||
export * from "./createCustomer";
|
export * from "./createCustomer";
|
||||||
|
|||||||