mirror of
https://github.com/fosrl/pangolin.git
synced 2026-02-27 15:26:41 +00:00
Compare commits
12 Commits
multi-role
...
1.15.4-s.9
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
49a326cde7 | ||
|
|
63e208f4ec | ||
|
|
f50d1549b0 | ||
|
|
55e24df671 | ||
|
|
b37e1d0cc0 | ||
|
|
afa26c0dd4 | ||
|
|
d6fe04ec4e | ||
|
|
b8a364af6a | ||
|
|
5ef808d4a2 | ||
|
|
a502780c9b | ||
|
|
418e099804 | ||
|
|
b622aca221 |
@@ -28,9 +28,9 @@ LICENSE
|
|||||||
CONTRIBUTING.md
|
CONTRIBUTING.md
|
||||||
dist
|
dist
|
||||||
.git
|
.git
|
||||||
migrations/
|
server/migrations/
|
||||||
config/
|
config/
|
||||||
build.ts
|
build.ts
|
||||||
tsconfig.json
|
tsconfig.json
|
||||||
Dockerfile*
|
Dockerfile*
|
||||||
migrations/
|
drizzle.config.ts
|
||||||
|
|||||||
@@ -1,8 +1,8 @@
|
|||||||
FROM node:24-alpine AS base
|
FROM node:24-slim AS base
|
||||||
|
|
||||||
WORKDIR /app
|
WORKDIR /app
|
||||||
|
|
||||||
RUN apk add --no-cache python3 make g++
|
RUN apt-get update && apt-get install -y python3 make g++ && rm -rf /var/lib/apt/lists/*
|
||||||
|
|
||||||
COPY package*.json ./
|
COPY package*.json ./
|
||||||
|
|
||||||
@@ -27,11 +27,11 @@ FROM base AS builder
|
|||||||
|
|
||||||
RUN npm ci --omit=dev
|
RUN npm ci --omit=dev
|
||||||
|
|
||||||
FROM node:24-alpine AS runner
|
FROM node:24-slim AS runner
|
||||||
|
|
||||||
WORKDIR /app
|
WORKDIR /app
|
||||||
|
|
||||||
RUN apk add --no-cache curl tzdata
|
RUN apt-get update && apt-get install -y curl tzdata && rm -rf /var/lib/apt/lists/*
|
||||||
|
|
||||||
COPY --from=builder /app/node_modules ./node_modules
|
COPY --from=builder /app/node_modules ./node_modules
|
||||||
COPY --from=builder /app/package.json ./package.json
|
COPY --from=builder /app/package.json ./package.json
|
||||||
|
|||||||
@@ -1,4 +1,5 @@
|
|||||||
export * from "./driver";
|
export * from "./driver";
|
||||||
|
export * from "./logsDriver";
|
||||||
export * from "./safeRead";
|
export * from "./safeRead";
|
||||||
export * from "./schema/schema";
|
export * from "./schema/schema";
|
||||||
export * from "./schema/privateSchema";
|
export * from "./schema/privateSchema";
|
||||||
|
|||||||
89
server/db/pg/logsDriver.ts
Normal file
89
server/db/pg/logsDriver.ts
Normal file
@@ -0,0 +1,89 @@
|
|||||||
|
import { drizzle as DrizzlePostgres } from "drizzle-orm/node-postgres";
|
||||||
|
import { Pool } from "pg";
|
||||||
|
import { readConfigFile } from "@server/lib/readConfigFile";
|
||||||
|
import { readPrivateConfigFile } from "@server/private/lib/readConfigFile";
|
||||||
|
import { withReplicas } from "drizzle-orm/pg-core";
|
||||||
|
import { build } from "@server/build";
|
||||||
|
import { db as mainDb, primaryDb as mainPrimaryDb } from "./driver";
|
||||||
|
|
||||||
|
function createLogsDb() {
|
||||||
|
// Only use separate logs database in SaaS builds
|
||||||
|
if (build !== "saas") {
|
||||||
|
return mainDb;
|
||||||
|
}
|
||||||
|
|
||||||
|
const config = readConfigFile();
|
||||||
|
const privateConfig = readPrivateConfigFile();
|
||||||
|
|
||||||
|
// Merge configs, prioritizing private config
|
||||||
|
const logsConfig = privateConfig.postgres_logs || config.postgres_logs;
|
||||||
|
|
||||||
|
// Check environment variable first
|
||||||
|
let connectionString = process.env.POSTGRES_LOGS_CONNECTION_STRING;
|
||||||
|
let replicaConnections: Array<{ connection_string: string }> = [];
|
||||||
|
|
||||||
|
if (!connectionString && logsConfig) {
|
||||||
|
connectionString = logsConfig.connection_string;
|
||||||
|
replicaConnections = logsConfig.replicas || [];
|
||||||
|
}
|
||||||
|
|
||||||
|
// If POSTGRES_LOGS_REPLICA_CONNECTION_STRINGS is set, use it
|
||||||
|
if (process.env.POSTGRES_LOGS_REPLICA_CONNECTION_STRINGS) {
|
||||||
|
replicaConnections =
|
||||||
|
process.env.POSTGRES_LOGS_REPLICA_CONNECTION_STRINGS.split(",").map(
|
||||||
|
(conn) => ({
|
||||||
|
connection_string: conn.trim()
|
||||||
|
})
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
// If no logs database is configured, fall back to main database
|
||||||
|
if (!connectionString) {
|
||||||
|
return mainDb;
|
||||||
|
}
|
||||||
|
|
||||||
|
// Create separate connection pool for logs database
|
||||||
|
const poolConfig = logsConfig?.pool || config.postgres?.pool;
|
||||||
|
const primaryPool = new Pool({
|
||||||
|
connectionString,
|
||||||
|
max: poolConfig?.max_connections || 20,
|
||||||
|
idleTimeoutMillis: poolConfig?.idle_timeout_ms || 30000,
|
||||||
|
connectionTimeoutMillis: poolConfig?.connection_timeout_ms || 5000
|
||||||
|
});
|
||||||
|
|
||||||
|
const replicas = [];
|
||||||
|
|
||||||
|
if (!replicaConnections.length) {
|
||||||
|
replicas.push(
|
||||||
|
DrizzlePostgres(primaryPool, {
|
||||||
|
logger: process.env.QUERY_LOGGING == "true"
|
||||||
|
})
|
||||||
|
);
|
||||||
|
} else {
|
||||||
|
for (const conn of replicaConnections) {
|
||||||
|
const replicaPool = new Pool({
|
||||||
|
connectionString: conn.connection_string,
|
||||||
|
max: poolConfig?.max_replica_connections || 20,
|
||||||
|
idleTimeoutMillis: poolConfig?.idle_timeout_ms || 30000,
|
||||||
|
connectionTimeoutMillis:
|
||||||
|
poolConfig?.connection_timeout_ms || 5000
|
||||||
|
});
|
||||||
|
replicas.push(
|
||||||
|
DrizzlePostgres(replicaPool, {
|
||||||
|
logger: process.env.QUERY_LOGGING == "true"
|
||||||
|
})
|
||||||
|
);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return withReplicas(
|
||||||
|
DrizzlePostgres(primaryPool, {
|
||||||
|
logger: process.env.QUERY_LOGGING == "true"
|
||||||
|
}),
|
||||||
|
replicas as any
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
export const logsDb = createLogsDb();
|
||||||
|
export default logsDb;
|
||||||
|
export const primaryLogsDb = logsDb.$primary;
|
||||||
@@ -1,4 +1,5 @@
|
|||||||
export * from "./driver";
|
export * from "./driver";
|
||||||
|
export * from "./logsDriver";
|
||||||
export * from "./safeRead";
|
export * from "./safeRead";
|
||||||
export * from "./schema/schema";
|
export * from "./schema/schema";
|
||||||
export * from "./schema/privateSchema";
|
export * from "./schema/privateSchema";
|
||||||
|
|||||||
7
server/db/sqlite/logsDriver.ts
Normal file
7
server/db/sqlite/logsDriver.ts
Normal file
@@ -0,0 +1,7 @@
|
|||||||
|
import { db as mainDb } from "./driver";
|
||||||
|
|
||||||
|
// SQLite doesn't support separate databases for logs in the same way as Postgres
|
||||||
|
// Always use the main database connection for SQLite
|
||||||
|
export const logsDb = mainDb;
|
||||||
|
export default logsDb;
|
||||||
|
export const primaryLogsDb = logsDb;
|
||||||
@@ -16,6 +16,11 @@ const internalPort = config.getRawConfig().server.internal_port;
|
|||||||
export function createInternalServer() {
|
export function createInternalServer() {
|
||||||
const internalServer = express();
|
const internalServer = express();
|
||||||
|
|
||||||
|
const trustProxy = config.getRawConfig().server.trust_proxy;
|
||||||
|
if (trustProxy) {
|
||||||
|
internalServer.set("trust proxy", trustProxy);
|
||||||
|
}
|
||||||
|
|
||||||
internalServer.use(helmet());
|
internalServer.use(helmet());
|
||||||
internalServer.use(cors());
|
internalServer.use(cors());
|
||||||
internalServer.use(stripDuplicateSesions);
|
internalServer.use(stripDuplicateSesions);
|
||||||
|
|||||||
@@ -230,7 +230,7 @@ export class UsageService {
|
|||||||
const orgIdToUse = await this.getBillingOrg(orgId);
|
const orgIdToUse = await this.getBillingOrg(orgId);
|
||||||
|
|
||||||
const cacheKey = `customer_${orgIdToUse}_${featureId}`;
|
const cacheKey = `customer_${orgIdToUse}_${featureId}`;
|
||||||
const cached = cache.get<string>(cacheKey);
|
const cached = await cache.get<string>(cacheKey);
|
||||||
|
|
||||||
if (cached) {
|
if (cached) {
|
||||||
return cached;
|
return cached;
|
||||||
@@ -253,7 +253,7 @@ export class UsageService {
|
|||||||
const customerId = customer.customerId;
|
const customerId = customer.customerId;
|
||||||
|
|
||||||
// Cache the result
|
// Cache the result
|
||||||
cache.set(cacheKey, customerId, 300); // 5 minute TTL
|
await cache.set(cacheKey, customerId, 300); // 5 minute TTL
|
||||||
|
|
||||||
return customerId;
|
return customerId;
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
|
|||||||
@@ -11,7 +11,7 @@ import {
|
|||||||
userSiteResources
|
userSiteResources
|
||||||
} from "@server/db";
|
} from "@server/db";
|
||||||
import { sites } from "@server/db";
|
import { sites } from "@server/db";
|
||||||
import { eq, and, ne, inArray } from "drizzle-orm";
|
import { eq, and, ne, inArray, or } from "drizzle-orm";
|
||||||
import { Config } from "./types";
|
import { Config } from "./types";
|
||||||
import logger from "@server/logger";
|
import logger from "@server/logger";
|
||||||
import { getNextAvailableAliasAddress } from "../ip";
|
import { getNextAvailableAliasAddress } from "../ip";
|
||||||
@@ -142,7 +142,10 @@ export async function updateClientResources(
|
|||||||
.innerJoin(userOrgs, eq(users.userId, userOrgs.userId))
|
.innerJoin(userOrgs, eq(users.userId, userOrgs.userId))
|
||||||
.where(
|
.where(
|
||||||
and(
|
and(
|
||||||
|
or(
|
||||||
inArray(users.username, resourceData.users),
|
inArray(users.username, resourceData.users),
|
||||||
|
inArray(users.email, resourceData.users)
|
||||||
|
),
|
||||||
eq(userOrgs.orgId, orgId)
|
eq(userOrgs.orgId, orgId)
|
||||||
)
|
)
|
||||||
);
|
);
|
||||||
@@ -276,7 +279,10 @@ export async function updateClientResources(
|
|||||||
.innerJoin(userOrgs, eq(users.userId, userOrgs.userId))
|
.innerJoin(userOrgs, eq(users.userId, userOrgs.userId))
|
||||||
.where(
|
.where(
|
||||||
and(
|
and(
|
||||||
|
or(
|
||||||
inArray(users.username, resourceData.users),
|
inArray(users.username, resourceData.users),
|
||||||
|
inArray(users.email, resourceData.users)
|
||||||
|
),
|
||||||
eq(userOrgs.orgId, orgId)
|
eq(userOrgs.orgId, orgId)
|
||||||
)
|
)
|
||||||
);
|
);
|
||||||
|
|||||||
@@ -212,7 +212,10 @@ export async function updateProxyResources(
|
|||||||
} else {
|
} else {
|
||||||
// Update existing resource
|
// Update existing resource
|
||||||
|
|
||||||
const isLicensed = await isLicensedOrSubscribed(orgId, tierMatrix.maintencePage);
|
const isLicensed = await isLicensedOrSubscribed(
|
||||||
|
orgId,
|
||||||
|
tierMatrix.maintencePage
|
||||||
|
);
|
||||||
if (!isLicensed) {
|
if (!isLicensed) {
|
||||||
resourceData.maintenance = undefined;
|
resourceData.maintenance = undefined;
|
||||||
}
|
}
|
||||||
@@ -590,7 +593,10 @@ export async function updateProxyResources(
|
|||||||
existingRule.action !== getRuleAction(rule.action) ||
|
existingRule.action !== getRuleAction(rule.action) ||
|
||||||
existingRule.match !== rule.match.toUpperCase() ||
|
existingRule.match !== rule.match.toUpperCase() ||
|
||||||
existingRule.value !==
|
existingRule.value !==
|
||||||
getRuleValue(rule.match.toUpperCase(), rule.value) ||
|
getRuleValue(
|
||||||
|
rule.match.toUpperCase(),
|
||||||
|
rule.value
|
||||||
|
) ||
|
||||||
existingRule.priority !== intendedPriority
|
existingRule.priority !== intendedPriority
|
||||||
) {
|
) {
|
||||||
validateRule(rule);
|
validateRule(rule);
|
||||||
@@ -648,7 +654,10 @@ export async function updateProxyResources(
|
|||||||
);
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
const isLicensed = await isLicensedOrSubscribed(orgId, tierMatrix.maintencePage);
|
const isLicensed = await isLicensedOrSubscribed(
|
||||||
|
orgId,
|
||||||
|
tierMatrix.maintencePage
|
||||||
|
);
|
||||||
if (!isLicensed) {
|
if (!isLicensed) {
|
||||||
resourceData.maintenance = undefined;
|
resourceData.maintenance = undefined;
|
||||||
}
|
}
|
||||||
@@ -935,7 +944,12 @@ async function syncUserResources(
|
|||||||
.select()
|
.select()
|
||||||
.from(users)
|
.from(users)
|
||||||
.innerJoin(userOrgs, eq(users.userId, userOrgs.userId))
|
.innerJoin(userOrgs, eq(users.userId, userOrgs.userId))
|
||||||
.where(and(eq(users.username, username), eq(userOrgs.orgId, orgId)))
|
.where(
|
||||||
|
and(
|
||||||
|
or(eq(users.username, username), eq(users.email, username)),
|
||||||
|
eq(userOrgs.orgId, orgId)
|
||||||
|
)
|
||||||
|
)
|
||||||
.limit(1);
|
.limit(1);
|
||||||
|
|
||||||
if (!user) {
|
if (!user) {
|
||||||
|
|||||||
@@ -69,7 +69,7 @@ export const AuthSchema = z.object({
|
|||||||
.refine((roles) => !roles.includes("Admin"), {
|
.refine((roles) => !roles.includes("Admin"), {
|
||||||
error: "Admin role cannot be included in sso-roles"
|
error: "Admin role cannot be included in sso-roles"
|
||||||
}),
|
}),
|
||||||
"sso-users": z.array(z.email()).optional().default([]),
|
"sso-users": z.array(z.string()).optional().default([]),
|
||||||
"whitelist-users": z.array(z.email()).optional().default([]),
|
"whitelist-users": z.array(z.email()).optional().default([]),
|
||||||
"auto-login-idp": z.int().positive().optional()
|
"auto-login-idp": z.int().positive().optional()
|
||||||
});
|
});
|
||||||
@@ -335,7 +335,7 @@ export const ClientResourceSchema = z
|
|||||||
.refine((roles) => !roles.includes("Admin"), {
|
.refine((roles) => !roles.includes("Admin"), {
|
||||||
error: "Admin role cannot be included in roles"
|
error: "Admin role cannot be included in roles"
|
||||||
}),
|
}),
|
||||||
users: z.array(z.email()).optional().default([]),
|
users: z.array(z.string()).optional().default([]),
|
||||||
machines: z.array(z.string()).optional().default([])
|
machines: z.array(z.string()).optional().default([])
|
||||||
})
|
})
|
||||||
.refine(
|
.refine(
|
||||||
|
|||||||
@@ -1,9 +1,10 @@
|
|||||||
import NodeCache from "node-cache";
|
import NodeCache from "node-cache";
|
||||||
import logger from "@server/logger";
|
import logger from "@server/logger";
|
||||||
|
import { redisManager } from "@server/private/lib/redis";
|
||||||
|
|
||||||
// Create cache with maxKeys limit to prevent memory leaks
|
// Create local cache with maxKeys limit to prevent memory leaks
|
||||||
// With ~10k requests/day and 5min TTL, 10k keys should be more than sufficient
|
// With ~10k requests/day and 5min TTL, 10k keys should be more than sufficient
|
||||||
export const cache = new NodeCache({
|
export const localCache = new NodeCache({
|
||||||
stdTTL: 3600,
|
stdTTL: 3600,
|
||||||
checkperiod: 120,
|
checkperiod: 120,
|
||||||
maxKeys: 10000
|
maxKeys: 10000
|
||||||
@@ -11,10 +12,255 @@ export const cache = new NodeCache({
|
|||||||
|
|
||||||
// Log cache statistics periodically for monitoring
|
// Log cache statistics periodically for monitoring
|
||||||
setInterval(() => {
|
setInterval(() => {
|
||||||
const stats = cache.getStats();
|
const stats = localCache.getStats();
|
||||||
logger.debug(
|
logger.debug(
|
||||||
`Cache stats - Keys: ${stats.keys}, Hits: ${stats.hits}, Misses: ${stats.misses}, Hit rate: ${stats.hits > 0 ? ((stats.hits / (stats.hits + stats.misses)) * 100).toFixed(2) : 0}%`
|
`Local cache stats - Keys: ${stats.keys}, Hits: ${stats.hits}, Misses: ${stats.misses}, Hit rate: ${stats.hits > 0 ? ((stats.hits / (stats.hits + stats.misses)) * 100).toFixed(2) : 0}%`
|
||||||
);
|
);
|
||||||
}, 300000); // Every 5 minutes
|
}, 300000); // Every 5 minutes
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Adaptive cache that uses Redis when available in multi-node environments,
|
||||||
|
* otherwise falls back to local memory cache for single-node deployments.
|
||||||
|
*/
|
||||||
|
class AdaptiveCache {
|
||||||
|
private useRedis(): boolean {
|
||||||
|
return redisManager.isRedisEnabled() && redisManager.getHealthStatus().isHealthy;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Set a value in the cache
|
||||||
|
* @param key - Cache key
|
||||||
|
* @param value - Value to cache (will be JSON stringified for Redis)
|
||||||
|
* @param ttl - Time to live in seconds (0 = no expiration)
|
||||||
|
* @returns boolean indicating success
|
||||||
|
*/
|
||||||
|
async set(key: string, value: any, ttl?: number): Promise<boolean> {
|
||||||
|
const effectiveTtl = ttl === 0 ? undefined : ttl;
|
||||||
|
|
||||||
|
if (this.useRedis()) {
|
||||||
|
try {
|
||||||
|
const serialized = JSON.stringify(value);
|
||||||
|
const success = await redisManager.set(key, serialized, effectiveTtl);
|
||||||
|
|
||||||
|
if (success) {
|
||||||
|
logger.debug(`Set key in Redis: ${key}`);
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
|
||||||
|
// Redis failed, fall through to local cache
|
||||||
|
logger.debug(`Redis set failed for key ${key}, falling back to local cache`);
|
||||||
|
} catch (error) {
|
||||||
|
logger.error(`Redis set error for key ${key}:`, error);
|
||||||
|
// Fall through to local cache
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Use local cache as fallback or primary
|
||||||
|
const success = localCache.set(key, value, effectiveTtl || 0);
|
||||||
|
if (success) {
|
||||||
|
logger.debug(`Set key in local cache: ${key}`);
|
||||||
|
}
|
||||||
|
return success;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Get a value from the cache
|
||||||
|
* @param key - Cache key
|
||||||
|
* @returns The cached value or undefined if not found
|
||||||
|
*/
|
||||||
|
async get<T = any>(key: string): Promise<T | undefined> {
|
||||||
|
if (this.useRedis()) {
|
||||||
|
try {
|
||||||
|
const value = await redisManager.get(key);
|
||||||
|
|
||||||
|
if (value !== null) {
|
||||||
|
logger.debug(`Cache hit in Redis: ${key}`);
|
||||||
|
return JSON.parse(value) as T;
|
||||||
|
}
|
||||||
|
|
||||||
|
logger.debug(`Cache miss in Redis: ${key}`);
|
||||||
|
return undefined;
|
||||||
|
} catch (error) {
|
||||||
|
logger.error(`Redis get error for key ${key}:`, error);
|
||||||
|
// Fall through to local cache
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Use local cache as fallback or primary
|
||||||
|
const value = localCache.get<T>(key);
|
||||||
|
if (value !== undefined) {
|
||||||
|
logger.debug(`Cache hit in local cache: ${key}`);
|
||||||
|
} else {
|
||||||
|
logger.debug(`Cache miss in local cache: ${key}`);
|
||||||
|
}
|
||||||
|
return value;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Delete a value from the cache
|
||||||
|
* @param key - Cache key or array of keys
|
||||||
|
* @returns Number of deleted entries
|
||||||
|
*/
|
||||||
|
async del(key: string | string[]): Promise<number> {
|
||||||
|
const keys = Array.isArray(key) ? key : [key];
|
||||||
|
let deletedCount = 0;
|
||||||
|
|
||||||
|
if (this.useRedis()) {
|
||||||
|
try {
|
||||||
|
for (const k of keys) {
|
||||||
|
const success = await redisManager.del(k);
|
||||||
|
if (success) {
|
||||||
|
deletedCount++;
|
||||||
|
logger.debug(`Deleted key from Redis: ${k}`);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if (deletedCount === keys.length) {
|
||||||
|
return deletedCount;
|
||||||
|
}
|
||||||
|
|
||||||
|
// Some Redis deletes failed, fall through to local cache
|
||||||
|
logger.debug(`Some Redis deletes failed, falling back to local cache`);
|
||||||
|
} catch (error) {
|
||||||
|
logger.error(`Redis del error for keys ${keys.join(", ")}:`, error);
|
||||||
|
// Fall through to local cache
|
||||||
|
deletedCount = 0;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Use local cache as fallback or primary
|
||||||
|
for (const k of keys) {
|
||||||
|
const success = localCache.del(k);
|
||||||
|
if (success > 0) {
|
||||||
|
deletedCount++;
|
||||||
|
logger.debug(`Deleted key from local cache: ${k}`);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return deletedCount;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Check if a key exists in the cache
|
||||||
|
* @param key - Cache key
|
||||||
|
* @returns boolean indicating if key exists
|
||||||
|
*/
|
||||||
|
async has(key: string): Promise<boolean> {
|
||||||
|
if (this.useRedis()) {
|
||||||
|
try {
|
||||||
|
const value = await redisManager.get(key);
|
||||||
|
return value !== null;
|
||||||
|
} catch (error) {
|
||||||
|
logger.error(`Redis has error for key ${key}:`, error);
|
||||||
|
// Fall through to local cache
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Use local cache as fallback or primary
|
||||||
|
return localCache.has(key);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Get multiple values from the cache
|
||||||
|
* @param keys - Array of cache keys
|
||||||
|
* @returns Array of values (undefined for missing keys)
|
||||||
|
*/
|
||||||
|
async mget<T = any>(keys: string[]): Promise<(T | undefined)[]> {
|
||||||
|
if (this.useRedis()) {
|
||||||
|
try {
|
||||||
|
const results: (T | undefined)[] = [];
|
||||||
|
|
||||||
|
for (const key of keys) {
|
||||||
|
const value = await redisManager.get(key);
|
||||||
|
if (value !== null) {
|
||||||
|
results.push(JSON.parse(value) as T);
|
||||||
|
} else {
|
||||||
|
results.push(undefined);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return results;
|
||||||
|
} catch (error) {
|
||||||
|
logger.error(`Redis mget error:`, error);
|
||||||
|
// Fall through to local cache
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Use local cache as fallback or primary
|
||||||
|
return keys.map((key) => localCache.get<T>(key));
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Flush all keys from the cache
|
||||||
|
*/
|
||||||
|
async flushAll(): Promise<void> {
|
||||||
|
if (this.useRedis()) {
|
||||||
|
logger.warn("Adaptive cache flushAll called - Redis flush not implemented, only local cache will be flushed");
|
||||||
|
}
|
||||||
|
|
||||||
|
localCache.flushAll();
|
||||||
|
logger.debug("Flushed local cache");
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Get cache statistics
|
||||||
|
* Note: Only returns local cache stats, Redis stats are not included
|
||||||
|
*/
|
||||||
|
getStats() {
|
||||||
|
return localCache.getStats();
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Get the current cache backend being used
|
||||||
|
* @returns "redis" if Redis is available and healthy, "local" otherwise
|
||||||
|
*/
|
||||||
|
getCurrentBackend(): "redis" | "local" {
|
||||||
|
return this.useRedis() ? "redis" : "local";
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Take a key from the cache and delete it
|
||||||
|
* @param key - Cache key
|
||||||
|
* @returns The value or undefined if not found
|
||||||
|
*/
|
||||||
|
async take<T = any>(key: string): Promise<T | undefined> {
|
||||||
|
const value = await this.get<T>(key);
|
||||||
|
if (value !== undefined) {
|
||||||
|
await this.del(key);
|
||||||
|
}
|
||||||
|
return value;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Get TTL (time to live) for a key
|
||||||
|
* @param key - Cache key
|
||||||
|
* @returns TTL in seconds, 0 if no expiration, -1 if key doesn't exist
|
||||||
|
*/
|
||||||
|
getTtl(key: string): number {
|
||||||
|
// Note: This only works for local cache, Redis TTL is not supported
|
||||||
|
if (this.useRedis()) {
|
||||||
|
logger.warn(`getTtl called for key ${key} but Redis TTL lookup is not implemented`);
|
||||||
|
}
|
||||||
|
|
||||||
|
const ttl = localCache.getTtl(key);
|
||||||
|
if (ttl === undefined) {
|
||||||
|
return -1;
|
||||||
|
}
|
||||||
|
return Math.max(0, Math.floor((ttl - Date.now()) / 1000));
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Get all keys from the cache
|
||||||
|
* Note: Only returns local cache keys, Redis keys are not included
|
||||||
|
*/
|
||||||
|
keys(): string[] {
|
||||||
|
if (this.useRedis()) {
|
||||||
|
logger.warn("keys() called but Redis keys are not included, only local cache keys returned");
|
||||||
|
}
|
||||||
|
return localCache.keys();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Export singleton instance
|
||||||
|
export const cache = new AdaptiveCache();
|
||||||
export default cache;
|
export default cache;
|
||||||
|
|||||||
@@ -189,6 +189,46 @@ export const configSchema = z
|
|||||||
.prefault({})
|
.prefault({})
|
||||||
})
|
})
|
||||||
.optional(),
|
.optional(),
|
||||||
|
postgres_logs: z
|
||||||
|
.object({
|
||||||
|
connection_string: z
|
||||||
|
.string()
|
||||||
|
.optional()
|
||||||
|
.transform(getEnvOrYaml("POSTGRES_LOGS_CONNECTION_STRING")),
|
||||||
|
replicas: z
|
||||||
|
.array(
|
||||||
|
z.object({
|
||||||
|
connection_string: z.string()
|
||||||
|
})
|
||||||
|
)
|
||||||
|
.optional(),
|
||||||
|
pool: z
|
||||||
|
.object({
|
||||||
|
max_connections: z
|
||||||
|
.number()
|
||||||
|
.positive()
|
||||||
|
.optional()
|
||||||
|
.default(20),
|
||||||
|
max_replica_connections: z
|
||||||
|
.number()
|
||||||
|
.positive()
|
||||||
|
.optional()
|
||||||
|
.default(10),
|
||||||
|
idle_timeout_ms: z
|
||||||
|
.number()
|
||||||
|
.positive()
|
||||||
|
.optional()
|
||||||
|
.default(30000),
|
||||||
|
connection_timeout_ms: z
|
||||||
|
.number()
|
||||||
|
.positive()
|
||||||
|
.optional()
|
||||||
|
.default(5000)
|
||||||
|
})
|
||||||
|
.optional()
|
||||||
|
.prefault({})
|
||||||
|
})
|
||||||
|
.optional(),
|
||||||
traefik: z
|
traefik: z
|
||||||
.object({
|
.object({
|
||||||
http_entrypoint: z.string().optional().default("web"),
|
http_entrypoint: z.string().optional().default("web"),
|
||||||
|
|||||||
@@ -55,7 +55,7 @@ export async function getValidCertificatesForDomains(
|
|||||||
if (useCache) {
|
if (useCache) {
|
||||||
for (const domain of domains) {
|
for (const domain of domains) {
|
||||||
const cacheKey = `cert:${domain}`;
|
const cacheKey = `cert:${domain}`;
|
||||||
const cachedCert = cache.get<CertificateResult>(cacheKey);
|
const cachedCert = await cache.get<CertificateResult>(cacheKey);
|
||||||
if (cachedCert) {
|
if (cachedCert) {
|
||||||
finalResults.push(cachedCert); // Valid cache hit
|
finalResults.push(cachedCert); // Valid cache hit
|
||||||
} else {
|
} else {
|
||||||
@@ -169,7 +169,7 @@ export async function getValidCertificatesForDomains(
|
|||||||
// Add to cache for future requests, using the *requested domain* as the key
|
// Add to cache for future requests, using the *requested domain* as the key
|
||||||
if (useCache) {
|
if (useCache) {
|
||||||
const cacheKey = `cert:${domain}`;
|
const cacheKey = `cert:${domain}`;
|
||||||
cache.set(cacheKey, resultCert, 180);
|
await cache.set(cacheKey, resultCert, 180);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -11,7 +11,7 @@
|
|||||||
* This file is not licensed under the AGPLv3.
|
* This file is not licensed under the AGPLv3.
|
||||||
*/
|
*/
|
||||||
|
|
||||||
import { accessAuditLog, db, orgs } from "@server/db";
|
import { accessAuditLog, logsDb, db, orgs } from "@server/db";
|
||||||
import { getCountryCodeForIp } from "@server/lib/geoip";
|
import { getCountryCodeForIp } from "@server/lib/geoip";
|
||||||
import logger from "@server/logger";
|
import logger from "@server/logger";
|
||||||
import { and, eq, lt } from "drizzle-orm";
|
import { and, eq, lt } from "drizzle-orm";
|
||||||
@@ -21,7 +21,7 @@ import { stripPortFromHost } from "@server/lib/ip";
|
|||||||
|
|
||||||
async function getAccessDays(orgId: string): Promise<number> {
|
async function getAccessDays(orgId: string): Promise<number> {
|
||||||
// check cache first
|
// check cache first
|
||||||
const cached = cache.get<number>(`org_${orgId}_accessDays`);
|
const cached = await cache.get<number>(`org_${orgId}_accessDays`);
|
||||||
if (cached !== undefined) {
|
if (cached !== undefined) {
|
||||||
return cached;
|
return cached;
|
||||||
}
|
}
|
||||||
@@ -39,7 +39,7 @@ async function getAccessDays(orgId: string): Promise<number> {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// store the result in cache
|
// store the result in cache
|
||||||
cache.set(
|
await cache.set(
|
||||||
`org_${orgId}_accessDays`,
|
`org_${orgId}_accessDays`,
|
||||||
org.settingsLogRetentionDaysAction,
|
org.settingsLogRetentionDaysAction,
|
||||||
300
|
300
|
||||||
@@ -52,7 +52,7 @@ export async function cleanUpOldLogs(orgId: string, retentionDays: number) {
|
|||||||
const cutoffTimestamp = calculateCutoffTimestamp(retentionDays);
|
const cutoffTimestamp = calculateCutoffTimestamp(retentionDays);
|
||||||
|
|
||||||
try {
|
try {
|
||||||
await db
|
await logsDb
|
||||||
.delete(accessAuditLog)
|
.delete(accessAuditLog)
|
||||||
.where(
|
.where(
|
||||||
and(
|
and(
|
||||||
@@ -124,7 +124,7 @@ export async function logAccessAudit(data: {
|
|||||||
? await getCountryCodeFromIp(data.requestIp)
|
? await getCountryCodeFromIp(data.requestIp)
|
||||||
: undefined;
|
: undefined;
|
||||||
|
|
||||||
await db.insert(accessAuditLog).values({
|
await logsDb.insert(accessAuditLog).values({
|
||||||
timestamp: timestamp,
|
timestamp: timestamp,
|
||||||
orgId: data.orgId,
|
orgId: data.orgId,
|
||||||
actorType,
|
actorType,
|
||||||
@@ -146,14 +146,14 @@ export async function logAccessAudit(data: {
|
|||||||
async function getCountryCodeFromIp(ip: string): Promise<string | undefined> {
|
async function getCountryCodeFromIp(ip: string): Promise<string | undefined> {
|
||||||
const geoIpCacheKey = `geoip_access:${ip}`;
|
const geoIpCacheKey = `geoip_access:${ip}`;
|
||||||
|
|
||||||
let cachedCountryCode: string | undefined = cache.get(geoIpCacheKey);
|
let cachedCountryCode: string | undefined = await cache.get(geoIpCacheKey);
|
||||||
|
|
||||||
if (!cachedCountryCode) {
|
if (!cachedCountryCode) {
|
||||||
cachedCountryCode = await getCountryCodeForIp(ip); // do it locally
|
cachedCountryCode = await getCountryCodeForIp(ip); // do it locally
|
||||||
// Only cache successful lookups to avoid filling cache with undefined values
|
// Only cache successful lookups to avoid filling cache with undefined values
|
||||||
if (cachedCountryCode) {
|
if (cachedCountryCode) {
|
||||||
// Cache for longer since IP geolocation doesn't change frequently
|
// Cache for longer since IP geolocation doesn't change frequently
|
||||||
cache.set(geoIpCacheKey, cachedCountryCode, 300); // 5 minutes
|
await cache.set(geoIpCacheKey, cachedCountryCode, 300); // 5 minutes
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -83,6 +83,46 @@ export const privateConfigSchema = z.object({
|
|||||||
.optional()
|
.optional()
|
||||||
})
|
})
|
||||||
.optional(),
|
.optional(),
|
||||||
|
postgres_logs: z
|
||||||
|
.object({
|
||||||
|
connection_string: z
|
||||||
|
.string()
|
||||||
|
.optional()
|
||||||
|
.transform(getEnvOrYaml("POSTGRES_LOGS_CONNECTION_STRING")),
|
||||||
|
replicas: z
|
||||||
|
.array(
|
||||||
|
z.object({
|
||||||
|
connection_string: z.string()
|
||||||
|
})
|
||||||
|
)
|
||||||
|
.optional(),
|
||||||
|
pool: z
|
||||||
|
.object({
|
||||||
|
max_connections: z
|
||||||
|
.number()
|
||||||
|
.positive()
|
||||||
|
.optional()
|
||||||
|
.default(20),
|
||||||
|
max_replica_connections: z
|
||||||
|
.number()
|
||||||
|
.positive()
|
||||||
|
.optional()
|
||||||
|
.default(10),
|
||||||
|
idle_timeout_ms: z
|
||||||
|
.number()
|
||||||
|
.positive()
|
||||||
|
.optional()
|
||||||
|
.default(30000),
|
||||||
|
connection_timeout_ms: z
|
||||||
|
.number()
|
||||||
|
.positive()
|
||||||
|
.optional()
|
||||||
|
.default(5000)
|
||||||
|
})
|
||||||
|
.optional()
|
||||||
|
.prefault({})
|
||||||
|
})
|
||||||
|
.optional(),
|
||||||
gerbil: z
|
gerbil: z
|
||||||
.object({
|
.object({
|
||||||
local_exit_node_reachable_at: z
|
local_exit_node_reachable_at: z
|
||||||
|
|||||||
@@ -12,7 +12,7 @@
|
|||||||
*/
|
*/
|
||||||
|
|
||||||
import { ActionsEnum } from "@server/auth/actions";
|
import { ActionsEnum } from "@server/auth/actions";
|
||||||
import { actionAuditLog, db, orgs } from "@server/db";
|
import { actionAuditLog, logsDb, db, orgs } from "@server/db";
|
||||||
import logger from "@server/logger";
|
import logger from "@server/logger";
|
||||||
import HttpCode from "@server/types/HttpCode";
|
import HttpCode from "@server/types/HttpCode";
|
||||||
import { Request, Response, NextFunction } from "express";
|
import { Request, Response, NextFunction } from "express";
|
||||||
@@ -23,7 +23,7 @@ import { calculateCutoffTimestamp } from "@server/lib/cleanupLogs";
|
|||||||
|
|
||||||
async function getActionDays(orgId: string): Promise<number> {
|
async function getActionDays(orgId: string): Promise<number> {
|
||||||
// check cache first
|
// check cache first
|
||||||
const cached = cache.get<number>(`org_${orgId}_actionDays`);
|
const cached = await cache.get<number>(`org_${orgId}_actionDays`);
|
||||||
if (cached !== undefined) {
|
if (cached !== undefined) {
|
||||||
return cached;
|
return cached;
|
||||||
}
|
}
|
||||||
@@ -41,7 +41,7 @@ async function getActionDays(orgId: string): Promise<number> {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// store the result in cache
|
// store the result in cache
|
||||||
cache.set(
|
await cache.set(
|
||||||
`org_${orgId}_actionDays`,
|
`org_${orgId}_actionDays`,
|
||||||
org.settingsLogRetentionDaysAction,
|
org.settingsLogRetentionDaysAction,
|
||||||
300
|
300
|
||||||
@@ -54,7 +54,7 @@ export async function cleanUpOldLogs(orgId: string, retentionDays: number) {
|
|||||||
const cutoffTimestamp = calculateCutoffTimestamp(retentionDays);
|
const cutoffTimestamp = calculateCutoffTimestamp(retentionDays);
|
||||||
|
|
||||||
try {
|
try {
|
||||||
await db
|
await logsDb
|
||||||
.delete(actionAuditLog)
|
.delete(actionAuditLog)
|
||||||
.where(
|
.where(
|
||||||
and(
|
and(
|
||||||
@@ -123,7 +123,7 @@ export function logActionAudit(action: ActionsEnum) {
|
|||||||
metadata = JSON.stringify(req.params);
|
metadata = JSON.stringify(req.params);
|
||||||
}
|
}
|
||||||
|
|
||||||
await db.insert(actionAuditLog).values({
|
await logsDb.insert(actionAuditLog).values({
|
||||||
timestamp,
|
timestamp,
|
||||||
orgId,
|
orgId,
|
||||||
actorType,
|
actorType,
|
||||||
|
|||||||
@@ -11,11 +11,11 @@
|
|||||||
* This file is not licensed under the AGPLv3.
|
* This file is not licensed under the AGPLv3.
|
||||||
*/
|
*/
|
||||||
|
|
||||||
import { accessAuditLog, db, resources } from "@server/db";
|
import { accessAuditLog, logsDb, resources, db, primaryDb } from "@server/db";
|
||||||
import { registry } from "@server/openApi";
|
import { registry } from "@server/openApi";
|
||||||
import { NextFunction } from "express";
|
import { NextFunction } from "express";
|
||||||
import { Request, Response } from "express";
|
import { Request, Response } from "express";
|
||||||
import { eq, gt, lt, and, count, desc } from "drizzle-orm";
|
import { eq, gt, lt, and, count, desc, inArray } from "drizzle-orm";
|
||||||
import { OpenAPITags } from "@server/openApi";
|
import { OpenAPITags } from "@server/openApi";
|
||||||
import { z } from "zod";
|
import { z } from "zod";
|
||||||
import createHttpError from "http-errors";
|
import createHttpError from "http-errors";
|
||||||
@@ -115,15 +115,13 @@ function getWhere(data: Q) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
export function queryAccess(data: Q) {
|
export function queryAccess(data: Q) {
|
||||||
return db
|
return logsDb
|
||||||
.select({
|
.select({
|
||||||
orgId: accessAuditLog.orgId,
|
orgId: accessAuditLog.orgId,
|
||||||
action: accessAuditLog.action,
|
action: accessAuditLog.action,
|
||||||
actorType: accessAuditLog.actorType,
|
actorType: accessAuditLog.actorType,
|
||||||
actorId: accessAuditLog.actorId,
|
actorId: accessAuditLog.actorId,
|
||||||
resourceId: accessAuditLog.resourceId,
|
resourceId: accessAuditLog.resourceId,
|
||||||
resourceName: resources.name,
|
|
||||||
resourceNiceId: resources.niceId,
|
|
||||||
ip: accessAuditLog.ip,
|
ip: accessAuditLog.ip,
|
||||||
location: accessAuditLog.location,
|
location: accessAuditLog.location,
|
||||||
userAgent: accessAuditLog.userAgent,
|
userAgent: accessAuditLog.userAgent,
|
||||||
@@ -133,16 +131,46 @@ export function queryAccess(data: Q) {
|
|||||||
actor: accessAuditLog.actor
|
actor: accessAuditLog.actor
|
||||||
})
|
})
|
||||||
.from(accessAuditLog)
|
.from(accessAuditLog)
|
||||||
.leftJoin(
|
|
||||||
resources,
|
|
||||||
eq(accessAuditLog.resourceId, resources.resourceId)
|
|
||||||
)
|
|
||||||
.where(getWhere(data))
|
.where(getWhere(data))
|
||||||
.orderBy(desc(accessAuditLog.timestamp), desc(accessAuditLog.id));
|
.orderBy(desc(accessAuditLog.timestamp), desc(accessAuditLog.id));
|
||||||
}
|
}
|
||||||
|
|
||||||
|
async function enrichWithResourceDetails(logs: Awaited<ReturnType<typeof queryAccess>>) {
|
||||||
|
// If logs database is the same as main database, we can do a join
|
||||||
|
// Otherwise, we need to fetch resource details separately
|
||||||
|
const resourceIds = logs
|
||||||
|
.map(log => log.resourceId)
|
||||||
|
.filter((id): id is number => id !== null && id !== undefined);
|
||||||
|
|
||||||
|
if (resourceIds.length === 0) {
|
||||||
|
return logs.map(log => ({ ...log, resourceName: null, resourceNiceId: null }));
|
||||||
|
}
|
||||||
|
|
||||||
|
// Fetch resource details from main database
|
||||||
|
const resourceDetails = await primaryDb
|
||||||
|
.select({
|
||||||
|
resourceId: resources.resourceId,
|
||||||
|
name: resources.name,
|
||||||
|
niceId: resources.niceId
|
||||||
|
})
|
||||||
|
.from(resources)
|
||||||
|
.where(inArray(resources.resourceId, resourceIds));
|
||||||
|
|
||||||
|
// Create a map for quick lookup
|
||||||
|
const resourceMap = new Map(
|
||||||
|
resourceDetails.map(r => [r.resourceId, { name: r.name, niceId: r.niceId }])
|
||||||
|
);
|
||||||
|
|
||||||
|
// Enrich logs with resource details
|
||||||
|
return logs.map(log => ({
|
||||||
|
...log,
|
||||||
|
resourceName: log.resourceId ? resourceMap.get(log.resourceId)?.name ?? null : null,
|
||||||
|
resourceNiceId: log.resourceId ? resourceMap.get(log.resourceId)?.niceId ?? null : null
|
||||||
|
}));
|
||||||
|
}
|
||||||
|
|
||||||
export function countAccessQuery(data: Q) {
|
export function countAccessQuery(data: Q) {
|
||||||
const countQuery = db
|
const countQuery = logsDb
|
||||||
.select({ count: count() })
|
.select({ count: count() })
|
||||||
.from(accessAuditLog)
|
.from(accessAuditLog)
|
||||||
.where(getWhere(data));
|
.where(getWhere(data));
|
||||||
@@ -161,7 +189,7 @@ async function queryUniqueFilterAttributes(
|
|||||||
);
|
);
|
||||||
|
|
||||||
// Get unique actors
|
// Get unique actors
|
||||||
const uniqueActors = await db
|
const uniqueActors = await logsDb
|
||||||
.selectDistinct({
|
.selectDistinct({
|
||||||
actor: accessAuditLog.actor
|
actor: accessAuditLog.actor
|
||||||
})
|
})
|
||||||
@@ -169,7 +197,7 @@ async function queryUniqueFilterAttributes(
|
|||||||
.where(baseConditions);
|
.where(baseConditions);
|
||||||
|
|
||||||
// Get unique locations
|
// Get unique locations
|
||||||
const uniqueLocations = await db
|
const uniqueLocations = await logsDb
|
||||||
.selectDistinct({
|
.selectDistinct({
|
||||||
locations: accessAuditLog.location
|
locations: accessAuditLog.location
|
||||||
})
|
})
|
||||||
@@ -177,25 +205,40 @@ async function queryUniqueFilterAttributes(
|
|||||||
.where(baseConditions);
|
.where(baseConditions);
|
||||||
|
|
||||||
// Get unique resources with names
|
// Get unique resources with names
|
||||||
const uniqueResources = await db
|
const uniqueResources = await logsDb
|
||||||
.selectDistinct({
|
.selectDistinct({
|
||||||
id: accessAuditLog.resourceId,
|
id: accessAuditLog.resourceId
|
||||||
name: resources.name
|
|
||||||
})
|
})
|
||||||
.from(accessAuditLog)
|
.from(accessAuditLog)
|
||||||
.leftJoin(
|
|
||||||
resources,
|
|
||||||
eq(accessAuditLog.resourceId, resources.resourceId)
|
|
||||||
)
|
|
||||||
.where(baseConditions);
|
.where(baseConditions);
|
||||||
|
|
||||||
|
// Fetch resource names from main database for the unique resource IDs
|
||||||
|
const resourceIds = uniqueResources
|
||||||
|
.map(row => row.id)
|
||||||
|
.filter((id): id is number => id !== null);
|
||||||
|
|
||||||
|
let resourcesWithNames: Array<{ id: number; name: string | null }> = [];
|
||||||
|
|
||||||
|
if (resourceIds.length > 0) {
|
||||||
|
const resourceDetails = await primaryDb
|
||||||
|
.select({
|
||||||
|
resourceId: resources.resourceId,
|
||||||
|
name: resources.name
|
||||||
|
})
|
||||||
|
.from(resources)
|
||||||
|
.where(inArray(resources.resourceId, resourceIds));
|
||||||
|
|
||||||
|
resourcesWithNames = resourceDetails.map(r => ({
|
||||||
|
id: r.resourceId,
|
||||||
|
name: r.name
|
||||||
|
}));
|
||||||
|
}
|
||||||
|
|
||||||
return {
|
return {
|
||||||
actors: uniqueActors
|
actors: uniqueActors
|
||||||
.map((row) => row.actor)
|
.map((row) => row.actor)
|
||||||
.filter((actor): actor is string => actor !== null),
|
.filter((actor): actor is string => actor !== null),
|
||||||
resources: uniqueResources.filter(
|
resources: resourcesWithNames,
|
||||||
(row): row is { id: number; name: string | null } => row.id !== null
|
|
||||||
),
|
|
||||||
locations: uniqueLocations
|
locations: uniqueLocations
|
||||||
.map((row) => row.locations)
|
.map((row) => row.locations)
|
||||||
.filter((location): location is string => location !== null)
|
.filter((location): location is string => location !== null)
|
||||||
@@ -243,7 +286,10 @@ export async function queryAccessAuditLogs(
|
|||||||
|
|
||||||
const baseQuery = queryAccess(data);
|
const baseQuery = queryAccess(data);
|
||||||
|
|
||||||
const log = await baseQuery.limit(data.limit).offset(data.offset);
|
const logsRaw = await baseQuery.limit(data.limit).offset(data.offset);
|
||||||
|
|
||||||
|
// Enrich with resource details (handles cross-database scenario)
|
||||||
|
const log = await enrichWithResourceDetails(logsRaw);
|
||||||
|
|
||||||
const totalCountResult = await countAccessQuery(data);
|
const totalCountResult = await countAccessQuery(data);
|
||||||
const totalCount = totalCountResult[0].count;
|
const totalCount = totalCountResult[0].count;
|
||||||
|
|||||||
@@ -11,7 +11,7 @@
|
|||||||
* This file is not licensed under the AGPLv3.
|
* This file is not licensed under the AGPLv3.
|
||||||
*/
|
*/
|
||||||
|
|
||||||
import { actionAuditLog, db } from "@server/db";
|
import { actionAuditLog, logsDb } from "@server/db";
|
||||||
import { registry } from "@server/openApi";
|
import { registry } from "@server/openApi";
|
||||||
import { NextFunction } from "express";
|
import { NextFunction } from "express";
|
||||||
import { Request, Response } from "express";
|
import { Request, Response } from "express";
|
||||||
@@ -97,7 +97,7 @@ function getWhere(data: Q) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
export function queryAction(data: Q) {
|
export function queryAction(data: Q) {
|
||||||
return db
|
return logsDb
|
||||||
.select({
|
.select({
|
||||||
orgId: actionAuditLog.orgId,
|
orgId: actionAuditLog.orgId,
|
||||||
action: actionAuditLog.action,
|
action: actionAuditLog.action,
|
||||||
@@ -113,7 +113,7 @@ export function queryAction(data: Q) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
export function countActionQuery(data: Q) {
|
export function countActionQuery(data: Q) {
|
||||||
const countQuery = db
|
const countQuery = logsDb
|
||||||
.select({ count: count() })
|
.select({ count: count() })
|
||||||
.from(actionAuditLog)
|
.from(actionAuditLog)
|
||||||
.where(getWhere(data));
|
.where(getWhere(data));
|
||||||
@@ -132,14 +132,14 @@ async function queryUniqueFilterAttributes(
|
|||||||
);
|
);
|
||||||
|
|
||||||
// Get unique actors
|
// Get unique actors
|
||||||
const uniqueActors = await db
|
const uniqueActors = await logsDb
|
||||||
.selectDistinct({
|
.selectDistinct({
|
||||||
actor: actionAuditLog.actor
|
actor: actionAuditLog.actor
|
||||||
})
|
})
|
||||||
.from(actionAuditLog)
|
.from(actionAuditLog)
|
||||||
.where(baseConditions);
|
.where(baseConditions);
|
||||||
|
|
||||||
const uniqueActions = await db
|
const uniqueActions = await logsDb
|
||||||
.selectDistinct({
|
.selectDistinct({
|
||||||
action: actionAuditLog.action
|
action: actionAuditLog.action
|
||||||
})
|
})
|
||||||
|
|||||||
@@ -480,9 +480,9 @@ authenticated.get(
|
|||||||
|
|
||||||
authenticated.post(
|
authenticated.post(
|
||||||
"/re-key/:clientId/regenerate-client-secret",
|
"/re-key/:clientId/regenerate-client-secret",
|
||||||
|
verifyClientAccess, // this is first to set the org id
|
||||||
verifyValidLicense,
|
verifyValidLicense,
|
||||||
verifyValidSubscription(tierMatrix.rotateCredentials),
|
verifyValidSubscription(tierMatrix.rotateCredentials),
|
||||||
verifyClientAccess, // this is first to set the org id
|
|
||||||
verifyLimits,
|
verifyLimits,
|
||||||
verifyUserHasAction(ActionsEnum.reGenerateSecret),
|
verifyUserHasAction(ActionsEnum.reGenerateSecret),
|
||||||
reKey.reGenerateClientSecret
|
reKey.reGenerateClientSecret
|
||||||
@@ -490,9 +490,9 @@ authenticated.post(
|
|||||||
|
|
||||||
authenticated.post(
|
authenticated.post(
|
||||||
"/re-key/:siteId/regenerate-site-secret",
|
"/re-key/:siteId/regenerate-site-secret",
|
||||||
|
verifySiteAccess, // this is first to set the org id
|
||||||
verifyValidLicense,
|
verifyValidLicense,
|
||||||
verifyValidSubscription(tierMatrix.rotateCredentials),
|
verifyValidSubscription(tierMatrix.rotateCredentials),
|
||||||
verifySiteAccess, // this is first to set the org id
|
|
||||||
verifyLimits,
|
verifyLimits,
|
||||||
verifyUserHasAction(ActionsEnum.reGenerateSecret),
|
verifyUserHasAction(ActionsEnum.reGenerateSecret),
|
||||||
reKey.reGenerateSiteSecret
|
reKey.reGenerateSiteSecret
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
import { db, requestAuditLog, driver, primaryDb } from "@server/db";
|
import { logsDb, requestAuditLog, driver, primaryLogsDb } from "@server/db";
|
||||||
import { registry } from "@server/openApi";
|
import { registry } from "@server/openApi";
|
||||||
import { NextFunction } from "express";
|
import { NextFunction } from "express";
|
||||||
import { Request, Response } from "express";
|
import { Request, Response } from "express";
|
||||||
@@ -74,12 +74,12 @@ async function query(query: Q) {
|
|||||||
);
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
const [all] = await primaryDb
|
const [all] = await primaryLogsDb
|
||||||
.select({ total: count() })
|
.select({ total: count() })
|
||||||
.from(requestAuditLog)
|
.from(requestAuditLog)
|
||||||
.where(baseConditions);
|
.where(baseConditions);
|
||||||
|
|
||||||
const [blocked] = await primaryDb
|
const [blocked] = await primaryLogsDb
|
||||||
.select({ total: count() })
|
.select({ total: count() })
|
||||||
.from(requestAuditLog)
|
.from(requestAuditLog)
|
||||||
.where(and(baseConditions, eq(requestAuditLog.action, false)));
|
.where(and(baseConditions, eq(requestAuditLog.action, false)));
|
||||||
@@ -90,7 +90,7 @@ async function query(query: Q) {
|
|||||||
|
|
||||||
const DISTINCT_LIMIT = 500;
|
const DISTINCT_LIMIT = 500;
|
||||||
|
|
||||||
const requestsPerCountry = await primaryDb
|
const requestsPerCountry = await primaryLogsDb
|
||||||
.selectDistinct({
|
.selectDistinct({
|
||||||
code: requestAuditLog.location,
|
code: requestAuditLog.location,
|
||||||
count: totalQ
|
count: totalQ
|
||||||
@@ -118,7 +118,7 @@ async function query(query: Q) {
|
|||||||
const booleanTrue = driver === "pg" ? sql`true` : sql`1`;
|
const booleanTrue = driver === "pg" ? sql`true` : sql`1`;
|
||||||
const booleanFalse = driver === "pg" ? sql`false` : sql`0`;
|
const booleanFalse = driver === "pg" ? sql`false` : sql`0`;
|
||||||
|
|
||||||
const requestsPerDay = await primaryDb
|
const requestsPerDay = await primaryLogsDb
|
||||||
.select({
|
.select({
|
||||||
day: groupByDayFunction.as("day"),
|
day: groupByDayFunction.as("day"),
|
||||||
allowedCount:
|
allowedCount:
|
||||||
|
|||||||
@@ -1,8 +1,8 @@
|
|||||||
import { db, primaryDb, requestAuditLog, resources } from "@server/db";
|
import { logsDb, primaryLogsDb, requestAuditLog, resources, db, primaryDb } from "@server/db";
|
||||||
import { registry } from "@server/openApi";
|
import { registry } from "@server/openApi";
|
||||||
import { NextFunction } from "express";
|
import { NextFunction } from "express";
|
||||||
import { Request, Response } from "express";
|
import { Request, Response } from "express";
|
||||||
import { eq, gt, lt, and, count, desc } from "drizzle-orm";
|
import { eq, gt, lt, and, count, desc, inArray } from "drizzle-orm";
|
||||||
import { OpenAPITags } from "@server/openApi";
|
import { OpenAPITags } from "@server/openApi";
|
||||||
import { z } from "zod";
|
import { z } from "zod";
|
||||||
import createHttpError from "http-errors";
|
import createHttpError from "http-errors";
|
||||||
@@ -107,7 +107,7 @@ function getWhere(data: Q) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
export function queryRequest(data: Q) {
|
export function queryRequest(data: Q) {
|
||||||
return primaryDb
|
return primaryLogsDb
|
||||||
.select({
|
.select({
|
||||||
id: requestAuditLog.id,
|
id: requestAuditLog.id,
|
||||||
timestamp: requestAuditLog.timestamp,
|
timestamp: requestAuditLog.timestamp,
|
||||||
@@ -129,21 +129,49 @@ export function queryRequest(data: Q) {
|
|||||||
host: requestAuditLog.host,
|
host: requestAuditLog.host,
|
||||||
path: requestAuditLog.path,
|
path: requestAuditLog.path,
|
||||||
method: requestAuditLog.method,
|
method: requestAuditLog.method,
|
||||||
tls: requestAuditLog.tls,
|
tls: requestAuditLog.tls
|
||||||
resourceName: resources.name,
|
|
||||||
resourceNiceId: resources.niceId
|
|
||||||
})
|
})
|
||||||
.from(requestAuditLog)
|
.from(requestAuditLog)
|
||||||
.leftJoin(
|
|
||||||
resources,
|
|
||||||
eq(requestAuditLog.resourceId, resources.resourceId)
|
|
||||||
) // TODO: Is this efficient?
|
|
||||||
.where(getWhere(data))
|
.where(getWhere(data))
|
||||||
.orderBy(desc(requestAuditLog.timestamp));
|
.orderBy(desc(requestAuditLog.timestamp));
|
||||||
}
|
}
|
||||||
|
|
||||||
|
async function enrichWithResourceDetails(logs: Awaited<ReturnType<typeof queryRequest>>) {
|
||||||
|
// If logs database is the same as main database, we can do a join
|
||||||
|
// Otherwise, we need to fetch resource details separately
|
||||||
|
const resourceIds = logs
|
||||||
|
.map(log => log.resourceId)
|
||||||
|
.filter((id): id is number => id !== null && id !== undefined);
|
||||||
|
|
||||||
|
if (resourceIds.length === 0) {
|
||||||
|
return logs.map(log => ({ ...log, resourceName: null, resourceNiceId: null }));
|
||||||
|
}
|
||||||
|
|
||||||
|
// Fetch resource details from main database
|
||||||
|
const resourceDetails = await primaryDb
|
||||||
|
.select({
|
||||||
|
resourceId: resources.resourceId,
|
||||||
|
name: resources.name,
|
||||||
|
niceId: resources.niceId
|
||||||
|
})
|
||||||
|
.from(resources)
|
||||||
|
.where(inArray(resources.resourceId, resourceIds));
|
||||||
|
|
||||||
|
// Create a map for quick lookup
|
||||||
|
const resourceMap = new Map(
|
||||||
|
resourceDetails.map(r => [r.resourceId, { name: r.name, niceId: r.niceId }])
|
||||||
|
);
|
||||||
|
|
||||||
|
// Enrich logs with resource details
|
||||||
|
return logs.map(log => ({
|
||||||
|
...log,
|
||||||
|
resourceName: log.resourceId ? resourceMap.get(log.resourceId)?.name ?? null : null,
|
||||||
|
resourceNiceId: log.resourceId ? resourceMap.get(log.resourceId)?.niceId ?? null : null
|
||||||
|
}));
|
||||||
|
}
|
||||||
|
|
||||||
export function countRequestQuery(data: Q) {
|
export function countRequestQuery(data: Q) {
|
||||||
const countQuery = primaryDb
|
const countQuery = primaryLogsDb
|
||||||
.select({ count: count() })
|
.select({ count: count() })
|
||||||
.from(requestAuditLog)
|
.from(requestAuditLog)
|
||||||
.where(getWhere(data));
|
.where(getWhere(data));
|
||||||
@@ -185,36 +213,31 @@ async function queryUniqueFilterAttributes(
|
|||||||
uniquePaths,
|
uniquePaths,
|
||||||
uniqueResources
|
uniqueResources
|
||||||
] = await Promise.all([
|
] = await Promise.all([
|
||||||
primaryDb
|
primaryLogsDb
|
||||||
.selectDistinct({ actor: requestAuditLog.actor })
|
.selectDistinct({ actor: requestAuditLog.actor })
|
||||||
.from(requestAuditLog)
|
.from(requestAuditLog)
|
||||||
.where(baseConditions)
|
.where(baseConditions)
|
||||||
.limit(DISTINCT_LIMIT + 1),
|
.limit(DISTINCT_LIMIT + 1),
|
||||||
primaryDb
|
primaryLogsDb
|
||||||
.selectDistinct({ locations: requestAuditLog.location })
|
.selectDistinct({ locations: requestAuditLog.location })
|
||||||
.from(requestAuditLog)
|
.from(requestAuditLog)
|
||||||
.where(baseConditions)
|
.where(baseConditions)
|
||||||
.limit(DISTINCT_LIMIT + 1),
|
.limit(DISTINCT_LIMIT + 1),
|
||||||
primaryDb
|
primaryLogsDb
|
||||||
.selectDistinct({ hosts: requestAuditLog.host })
|
.selectDistinct({ hosts: requestAuditLog.host })
|
||||||
.from(requestAuditLog)
|
.from(requestAuditLog)
|
||||||
.where(baseConditions)
|
.where(baseConditions)
|
||||||
.limit(DISTINCT_LIMIT + 1),
|
.limit(DISTINCT_LIMIT + 1),
|
||||||
primaryDb
|
primaryLogsDb
|
||||||
.selectDistinct({ paths: requestAuditLog.path })
|
.selectDistinct({ paths: requestAuditLog.path })
|
||||||
.from(requestAuditLog)
|
.from(requestAuditLog)
|
||||||
.where(baseConditions)
|
.where(baseConditions)
|
||||||
.limit(DISTINCT_LIMIT + 1),
|
.limit(DISTINCT_LIMIT + 1),
|
||||||
primaryDb
|
primaryLogsDb
|
||||||
.selectDistinct({
|
.selectDistinct({
|
||||||
id: requestAuditLog.resourceId,
|
id: requestAuditLog.resourceId
|
||||||
name: resources.name
|
|
||||||
})
|
})
|
||||||
.from(requestAuditLog)
|
.from(requestAuditLog)
|
||||||
.leftJoin(
|
|
||||||
resources,
|
|
||||||
eq(requestAuditLog.resourceId, resources.resourceId)
|
|
||||||
)
|
|
||||||
.where(baseConditions)
|
.where(baseConditions)
|
||||||
.limit(DISTINCT_LIMIT + 1)
|
.limit(DISTINCT_LIMIT + 1)
|
||||||
]);
|
]);
|
||||||
@@ -231,13 +254,33 @@ async function queryUniqueFilterAttributes(
|
|||||||
// throw new Error("Too many distinct filter attributes to retrieve. Please refine your time range.");
|
// throw new Error("Too many distinct filter attributes to retrieve. Please refine your time range.");
|
||||||
// }
|
// }
|
||||||
|
|
||||||
|
// Fetch resource names from main database for the unique resource IDs
|
||||||
|
const resourceIds = uniqueResources
|
||||||
|
.map(row => row.id)
|
||||||
|
.filter((id): id is number => id !== null);
|
||||||
|
|
||||||
|
let resourcesWithNames: Array<{ id: number; name: string | null }> = [];
|
||||||
|
|
||||||
|
if (resourceIds.length > 0) {
|
||||||
|
const resourceDetails = await primaryDb
|
||||||
|
.select({
|
||||||
|
resourceId: resources.resourceId,
|
||||||
|
name: resources.name
|
||||||
|
})
|
||||||
|
.from(resources)
|
||||||
|
.where(inArray(resources.resourceId, resourceIds));
|
||||||
|
|
||||||
|
resourcesWithNames = resourceDetails.map(r => ({
|
||||||
|
id: r.resourceId,
|
||||||
|
name: r.name
|
||||||
|
}));
|
||||||
|
}
|
||||||
|
|
||||||
return {
|
return {
|
||||||
actors: uniqueActors
|
actors: uniqueActors
|
||||||
.map((row) => row.actor)
|
.map((row) => row.actor)
|
||||||
.filter((actor): actor is string => actor !== null),
|
.filter((actor): actor is string => actor !== null),
|
||||||
resources: uniqueResources.filter(
|
resources: resourcesWithNames,
|
||||||
(row): row is { id: number; name: string | null } => row.id !== null
|
|
||||||
),
|
|
||||||
locations: uniqueLocations
|
locations: uniqueLocations
|
||||||
.map((row) => row.locations)
|
.map((row) => row.locations)
|
||||||
.filter((location): location is string => location !== null),
|
.filter((location): location is string => location !== null),
|
||||||
@@ -280,7 +323,10 @@ export async function queryRequestAuditLogs(
|
|||||||
|
|
||||||
const baseQuery = queryRequest(data);
|
const baseQuery = queryRequest(data);
|
||||||
|
|
||||||
const log = await baseQuery.limit(data.limit).offset(data.offset);
|
const logsRaw = await baseQuery.limit(data.limit).offset(data.offset);
|
||||||
|
|
||||||
|
// Enrich with resource details (handles cross-database scenario)
|
||||||
|
const log = await enrichWithResourceDetails(logsRaw);
|
||||||
|
|
||||||
const totalCountResult = await countRequestQuery(data);
|
const totalCountResult = await countRequestQuery(data);
|
||||||
const totalCount = totalCountResult[0].count;
|
const totalCount = totalCountResult[0].count;
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
import { db, orgs, requestAuditLog } from "@server/db";
|
import { logsDb, primaryLogsDb, db, orgs, requestAuditLog } from "@server/db";
|
||||||
import logger from "@server/logger";
|
import logger from "@server/logger";
|
||||||
import { and, eq, lt, sql } from "drizzle-orm";
|
import { and, eq, lt, sql } from "drizzle-orm";
|
||||||
import cache from "@server/lib/cache";
|
import cache from "@server/lib/cache";
|
||||||
@@ -69,7 +69,7 @@ async function flushAuditLogs() {
|
|||||||
try {
|
try {
|
||||||
// Use a transaction to ensure all inserts succeed or fail together
|
// Use a transaction to ensure all inserts succeed or fail together
|
||||||
// This prevents index corruption from partial writes
|
// This prevents index corruption from partial writes
|
||||||
await db.transaction(async (tx) => {
|
await logsDb.transaction(async (tx) => {
|
||||||
// Batch insert logs in groups of 25 to avoid overwhelming the database
|
// Batch insert logs in groups of 25 to avoid overwhelming the database
|
||||||
const BATCH_DB_SIZE = 25;
|
const BATCH_DB_SIZE = 25;
|
||||||
for (let i = 0; i < logsToWrite.length; i += BATCH_DB_SIZE) {
|
for (let i = 0; i < logsToWrite.length; i += BATCH_DB_SIZE) {
|
||||||
@@ -130,7 +130,7 @@ export async function shutdownAuditLogger() {
|
|||||||
|
|
||||||
async function getRetentionDays(orgId: string): Promise<number> {
|
async function getRetentionDays(orgId: string): Promise<number> {
|
||||||
// check cache first
|
// check cache first
|
||||||
const cached = cache.get<number>(`org_${orgId}_retentionDays`);
|
const cached = await cache.get<number>(`org_${orgId}_retentionDays`);
|
||||||
if (cached !== undefined) {
|
if (cached !== undefined) {
|
||||||
return cached;
|
return cached;
|
||||||
}
|
}
|
||||||
@@ -149,7 +149,7 @@ async function getRetentionDays(orgId: string): Promise<number> {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// store the result in cache
|
// store the result in cache
|
||||||
cache.set(
|
await cache.set(
|
||||||
`org_${orgId}_retentionDays`,
|
`org_${orgId}_retentionDays`,
|
||||||
org.settingsLogRetentionDaysRequest,
|
org.settingsLogRetentionDaysRequest,
|
||||||
300
|
300
|
||||||
@@ -162,7 +162,7 @@ export async function cleanUpOldLogs(orgId: string, retentionDays: number) {
|
|||||||
const cutoffTimestamp = calculateCutoffTimestamp(retentionDays);
|
const cutoffTimestamp = calculateCutoffTimestamp(retentionDays);
|
||||||
|
|
||||||
try {
|
try {
|
||||||
await db
|
await logsDb
|
||||||
.delete(requestAuditLog)
|
.delete(requestAuditLog)
|
||||||
.where(
|
.where(
|
||||||
and(
|
and(
|
||||||
|
|||||||
@@ -37,7 +37,7 @@ import {
|
|||||||
enforceResourceSessionLength
|
enforceResourceSessionLength
|
||||||
} from "#dynamic/lib/checkOrgAccessPolicy";
|
} from "#dynamic/lib/checkOrgAccessPolicy";
|
||||||
import { logRequestAudit } from "./logRequestAudit";
|
import { logRequestAudit } from "./logRequestAudit";
|
||||||
import cache from "@server/lib/cache";
|
import { localCache } from "@server/lib/cache";
|
||||||
import { APP_VERSION } from "@server/lib/consts";
|
import { APP_VERSION } from "@server/lib/consts";
|
||||||
import { isSubscribed } from "#dynamic/lib/isSubscribed";
|
import { isSubscribed } from "#dynamic/lib/isSubscribed";
|
||||||
import { tierMatrix } from "@server/lib/billing/tierMatrix";
|
import { tierMatrix } from "@server/lib/billing/tierMatrix";
|
||||||
@@ -137,7 +137,7 @@ export async function verifyResourceSession(
|
|||||||
headerAuthExtendedCompatibility: ResourceHeaderAuthExtendedCompatibility | null;
|
headerAuthExtendedCompatibility: ResourceHeaderAuthExtendedCompatibility | null;
|
||||||
org: Org;
|
org: Org;
|
||||||
}
|
}
|
||||||
| undefined = cache.get(resourceCacheKey);
|
| undefined = localCache.get(resourceCacheKey);
|
||||||
|
|
||||||
if (!resourceData) {
|
if (!resourceData) {
|
||||||
const result = await getResourceByDomain(cleanHost);
|
const result = await getResourceByDomain(cleanHost);
|
||||||
@@ -161,7 +161,7 @@ export async function verifyResourceSession(
|
|||||||
}
|
}
|
||||||
|
|
||||||
resourceData = result;
|
resourceData = result;
|
||||||
cache.set(resourceCacheKey, resourceData, 5);
|
localCache.set(resourceCacheKey, resourceData, 5);
|
||||||
}
|
}
|
||||||
|
|
||||||
const {
|
const {
|
||||||
@@ -405,7 +405,7 @@ export async function verifyResourceSession(
|
|||||||
// check for HTTP Basic Auth header
|
// check for HTTP Basic Auth header
|
||||||
const clientHeaderAuthKey = `headerAuth:${clientHeaderAuth}`;
|
const clientHeaderAuthKey = `headerAuth:${clientHeaderAuth}`;
|
||||||
if (headerAuth && clientHeaderAuth) {
|
if (headerAuth && clientHeaderAuth) {
|
||||||
if (cache.get(clientHeaderAuthKey)) {
|
if (localCache.get(clientHeaderAuthKey)) {
|
||||||
logger.debug(
|
logger.debug(
|
||||||
"Resource allowed because header auth is valid (cached)"
|
"Resource allowed because header auth is valid (cached)"
|
||||||
);
|
);
|
||||||
@@ -428,7 +428,7 @@ export async function verifyResourceSession(
|
|||||||
headerAuth.headerAuthHash
|
headerAuth.headerAuthHash
|
||||||
)
|
)
|
||||||
) {
|
) {
|
||||||
cache.set(clientHeaderAuthKey, clientHeaderAuth, 5);
|
localCache.set(clientHeaderAuthKey, clientHeaderAuth, 5);
|
||||||
logger.debug("Resource allowed because header auth is valid");
|
logger.debug("Resource allowed because header auth is valid");
|
||||||
|
|
||||||
logRequestAudit(
|
logRequestAudit(
|
||||||
@@ -520,7 +520,7 @@ export async function verifyResourceSession(
|
|||||||
|
|
||||||
if (resourceSessionToken) {
|
if (resourceSessionToken) {
|
||||||
const sessionCacheKey = `session:${resourceSessionToken}`;
|
const sessionCacheKey = `session:${resourceSessionToken}`;
|
||||||
let resourceSession: any = cache.get(sessionCacheKey);
|
let resourceSession: any = localCache.get(sessionCacheKey);
|
||||||
|
|
||||||
if (!resourceSession) {
|
if (!resourceSession) {
|
||||||
const result = await validateResourceSessionToken(
|
const result = await validateResourceSessionToken(
|
||||||
@@ -529,7 +529,7 @@ export async function verifyResourceSession(
|
|||||||
);
|
);
|
||||||
|
|
||||||
resourceSession = result?.resourceSession;
|
resourceSession = result?.resourceSession;
|
||||||
cache.set(sessionCacheKey, resourceSession, 5);
|
localCache.set(sessionCacheKey, resourceSession, 5);
|
||||||
}
|
}
|
||||||
|
|
||||||
if (resourceSession?.isRequestToken) {
|
if (resourceSession?.isRequestToken) {
|
||||||
@@ -662,7 +662,7 @@ export async function verifyResourceSession(
|
|||||||
}:${resource.resourceId}`;
|
}:${resource.resourceId}`;
|
||||||
|
|
||||||
let allowedUserData: BasicUserData | null | undefined =
|
let allowedUserData: BasicUserData | null | undefined =
|
||||||
cache.get(userAccessCacheKey);
|
localCache.get(userAccessCacheKey);
|
||||||
|
|
||||||
if (allowedUserData === undefined) {
|
if (allowedUserData === undefined) {
|
||||||
allowedUserData = await isUserAllowedToAccessResource(
|
allowedUserData = await isUserAllowedToAccessResource(
|
||||||
@@ -671,7 +671,7 @@ export async function verifyResourceSession(
|
|||||||
resourceData.org
|
resourceData.org
|
||||||
);
|
);
|
||||||
|
|
||||||
cache.set(userAccessCacheKey, allowedUserData, 5);
|
localCache.set(userAccessCacheKey, allowedUserData, 5);
|
||||||
}
|
}
|
||||||
|
|
||||||
if (
|
if (
|
||||||
@@ -974,11 +974,11 @@ async function checkRules(
|
|||||||
): Promise<"ACCEPT" | "DROP" | "PASS" | undefined> {
|
): Promise<"ACCEPT" | "DROP" | "PASS" | undefined> {
|
||||||
const ruleCacheKey = `rules:${resourceId}`;
|
const ruleCacheKey = `rules:${resourceId}`;
|
||||||
|
|
||||||
let rules: ResourceRule[] | undefined = cache.get(ruleCacheKey);
|
let rules: ResourceRule[] | undefined = localCache.get(ruleCacheKey);
|
||||||
|
|
||||||
if (!rules) {
|
if (!rules) {
|
||||||
rules = await getResourceRules(resourceId);
|
rules = await getResourceRules(resourceId);
|
||||||
cache.set(ruleCacheKey, rules, 5);
|
localCache.set(ruleCacheKey, rules, 5);
|
||||||
}
|
}
|
||||||
|
|
||||||
if (rules.length === 0) {
|
if (rules.length === 0) {
|
||||||
@@ -1208,13 +1208,13 @@ async function isIpInAsn(
|
|||||||
async function getAsnFromIp(ip: string): Promise<number | undefined> {
|
async function getAsnFromIp(ip: string): Promise<number | undefined> {
|
||||||
const asnCacheKey = `asn:${ip}`;
|
const asnCacheKey = `asn:${ip}`;
|
||||||
|
|
||||||
let cachedAsn: number | undefined = cache.get(asnCacheKey);
|
let cachedAsn: number | undefined = localCache.get(asnCacheKey);
|
||||||
|
|
||||||
if (!cachedAsn) {
|
if (!cachedAsn) {
|
||||||
cachedAsn = await getAsnForIp(ip); // do it locally
|
cachedAsn = await getAsnForIp(ip); // do it locally
|
||||||
// Cache for longer since IP ASN doesn't change frequently
|
// Cache for longer since IP ASN doesn't change frequently
|
||||||
if (cachedAsn) {
|
if (cachedAsn) {
|
||||||
cache.set(asnCacheKey, cachedAsn, 300); // 5 minutes
|
localCache.set(asnCacheKey, cachedAsn, 300); // 5 minutes
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -1224,14 +1224,14 @@ async function getAsnFromIp(ip: string): Promise<number | undefined> {
|
|||||||
async function getCountryCodeFromIp(ip: string): Promise<string | undefined> {
|
async function getCountryCodeFromIp(ip: string): Promise<string | undefined> {
|
||||||
const geoIpCacheKey = `geoip:${ip}`;
|
const geoIpCacheKey = `geoip:${ip}`;
|
||||||
|
|
||||||
let cachedCountryCode: string | undefined = cache.get(geoIpCacheKey);
|
let cachedCountryCode: string | undefined = localCache.get(geoIpCacheKey);
|
||||||
|
|
||||||
if (!cachedCountryCode) {
|
if (!cachedCountryCode) {
|
||||||
cachedCountryCode = await getCountryCodeForIp(ip); // do it locally
|
cachedCountryCode = await getCountryCodeForIp(ip); // do it locally
|
||||||
// Only cache successful lookups to avoid filling cache with undefined values
|
// Only cache successful lookups to avoid filling cache with undefined values
|
||||||
if (cachedCountryCode) {
|
if (cachedCountryCode) {
|
||||||
// Cache for longer since IP geolocation doesn't change frequently
|
// Cache for longer since IP geolocation doesn't change frequently
|
||||||
cache.set(geoIpCacheKey, cachedCountryCode, 300); // 5 minutes
|
localCache.set(geoIpCacheKey, cachedCountryCode, 300); // 5 minutes
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -24,8 +24,8 @@ export const handleDockerStatusMessage: MessageHandler = async (context) => {
|
|||||||
|
|
||||||
if (available) {
|
if (available) {
|
||||||
logger.info(`Newt ${newt.newtId} has Docker socket access`);
|
logger.info(`Newt ${newt.newtId} has Docker socket access`);
|
||||||
cache.set(`${newt.newtId}:socketPath`, socketPath, 0);
|
await cache.set(`${newt.newtId}:socketPath`, socketPath, 0);
|
||||||
cache.set(`${newt.newtId}:isAvailable`, available, 0);
|
await cache.set(`${newt.newtId}:isAvailable`, available, 0);
|
||||||
} else {
|
} else {
|
||||||
logger.warn(`Newt ${newt.newtId} does not have Docker socket access`);
|
logger.warn(`Newt ${newt.newtId} does not have Docker socket access`);
|
||||||
}
|
}
|
||||||
@@ -54,7 +54,7 @@ export const handleDockerContainersMessage: MessageHandler = async (
|
|||||||
);
|
);
|
||||||
|
|
||||||
if (containers && containers.length > 0) {
|
if (containers && containers.length > 0) {
|
||||||
cache.set(`${newt.newtId}:dockerContainers`, containers, 0);
|
await cache.set(`${newt.newtId}:dockerContainers`, containers, 0);
|
||||||
} else {
|
} else {
|
||||||
logger.warn(`Newt ${newt.newtId} does not have Docker containers`);
|
logger.warn(`Newt ${newt.newtId} does not have Docker containers`);
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,4 +1,7 @@
|
|||||||
import { generateSessionToken } from "@server/auth/sessions/app";
|
import {
|
||||||
|
generateSessionToken,
|
||||||
|
validateSessionToken
|
||||||
|
} from "@server/auth/sessions/app";
|
||||||
import {
|
import {
|
||||||
clients,
|
clients,
|
||||||
db,
|
db,
|
||||||
@@ -26,8 +29,9 @@ import { APP_VERSION } from "@server/lib/consts";
|
|||||||
|
|
||||||
export const olmGetTokenBodySchema = z.object({
|
export const olmGetTokenBodySchema = z.object({
|
||||||
olmId: z.string(),
|
olmId: z.string(),
|
||||||
secret: z.string(),
|
secret: z.string().optional(),
|
||||||
token: z.string().optional(),
|
userToken: z.string().optional(),
|
||||||
|
token: z.string().optional(), // this is the olm token
|
||||||
orgId: z.string().optional()
|
orgId: z.string().optional()
|
||||||
});
|
});
|
||||||
|
|
||||||
@@ -49,7 +53,7 @@ export async function getOlmToken(
|
|||||||
);
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
const { olmId, secret, token, orgId } = parsedBody.data;
|
const { olmId, secret, token, orgId, userToken } = parsedBody.data;
|
||||||
|
|
||||||
try {
|
try {
|
||||||
if (token) {
|
if (token) {
|
||||||
@@ -84,6 +88,24 @@ export async function getOlmToken(
|
|||||||
);
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if (userToken) {
|
||||||
|
const { session: userSession, user } =
|
||||||
|
await validateSessionToken(userToken);
|
||||||
|
if (!userSession || !user) {
|
||||||
|
return next(
|
||||||
|
createHttpError(HttpCode.BAD_REQUEST, "Invalid user token")
|
||||||
|
);
|
||||||
|
}
|
||||||
|
if (user.userId !== existingOlm.userId) {
|
||||||
|
return next(
|
||||||
|
createHttpError(
|
||||||
|
HttpCode.BAD_REQUEST,
|
||||||
|
"User token does not match olm"
|
||||||
|
)
|
||||||
|
);
|
||||||
|
}
|
||||||
|
} else if (secret) {
|
||||||
|
// this is for backward compatibility, we want to move towards userToken but some old clients may still be using secret so we will support both for now
|
||||||
const validSecret = await verifyPassword(
|
const validSecret = await verifyPassword(
|
||||||
secret,
|
secret,
|
||||||
existingOlm.secretHash
|
existingOlm.secretHash
|
||||||
@@ -99,6 +121,14 @@ export async function getOlmToken(
|
|||||||
createHttpError(HttpCode.BAD_REQUEST, "Secret is incorrect")
|
createHttpError(HttpCode.BAD_REQUEST, "Secret is incorrect")
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
|
} else {
|
||||||
|
return next(
|
||||||
|
createHttpError(
|
||||||
|
HttpCode.BAD_REQUEST,
|
||||||
|
"Either secret or userToken is required"
|
||||||
|
)
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
logger.debug("Creating new olm session token");
|
logger.debug("Creating new olm session token");
|
||||||
|
|
||||||
|
|||||||
@@ -194,9 +194,9 @@ export async function updateOrg(
|
|||||||
}
|
}
|
||||||
|
|
||||||
// invalidate the cache for all of the orgs retention days
|
// invalidate the cache for all of the orgs retention days
|
||||||
cache.del(`org_${orgId}_retentionDays`);
|
await cache.del(`org_${orgId}_retentionDays`);
|
||||||
cache.del(`org_${orgId}_actionDays`);
|
await cache.del(`org_${orgId}_actionDays`);
|
||||||
cache.del(`org_${orgId}_accessDays`);
|
await cache.del(`org_${orgId}_accessDays`);
|
||||||
|
|
||||||
return response(res, {
|
return response(res, {
|
||||||
data: updatedOrg[0],
|
data: updatedOrg[0],
|
||||||
|
|||||||
@@ -23,7 +23,7 @@ import { fromError } from "zod-validation-error";
|
|||||||
|
|
||||||
async function getLatestNewtVersion(): Promise<string | null> {
|
async function getLatestNewtVersion(): Promise<string | null> {
|
||||||
try {
|
try {
|
||||||
const cachedVersion = cache.get<string>("latestNewtVersion");
|
const cachedVersion = await cache.get<string>("latestNewtVersion");
|
||||||
if (cachedVersion) {
|
if (cachedVersion) {
|
||||||
return cachedVersion;
|
return cachedVersion;
|
||||||
}
|
}
|
||||||
@@ -55,7 +55,7 @@ async function getLatestNewtVersion(): Promise<string | null> {
|
|||||||
tags = tags.filter((version) => !version.name.includes("rc"));
|
tags = tags.filter((version) => !version.name.includes("rc"));
|
||||||
const latestVersion = tags[0].name;
|
const latestVersion = tags[0].name;
|
||||||
|
|
||||||
cache.set("latestNewtVersion", latestVersion);
|
await cache.set("latestNewtVersion", latestVersion);
|
||||||
|
|
||||||
return latestVersion;
|
return latestVersion;
|
||||||
} catch (error: any) {
|
} catch (error: any) {
|
||||||
|
|||||||
@@ -150,7 +150,7 @@ async function triggerFetch(siteId: number) {
|
|||||||
|
|
||||||
// clear the cache for this Newt ID so that the site has to keep asking for the containers
|
// clear the cache for this Newt ID so that the site has to keep asking for the containers
|
||||||
// this is to ensure that the site always gets the latest data
|
// this is to ensure that the site always gets the latest data
|
||||||
cache.del(`${newt.newtId}:dockerContainers`);
|
await cache.del(`${newt.newtId}:dockerContainers`);
|
||||||
|
|
||||||
return { siteId, newtId: newt.newtId };
|
return { siteId, newtId: newt.newtId };
|
||||||
}
|
}
|
||||||
@@ -158,7 +158,7 @@ async function triggerFetch(siteId: number) {
|
|||||||
async function queryContainers(siteId: number) {
|
async function queryContainers(siteId: number) {
|
||||||
const { newt } = await getSiteAndNewt(siteId);
|
const { newt } = await getSiteAndNewt(siteId);
|
||||||
|
|
||||||
const result = cache.get(`${newt.newtId}:dockerContainers`) as Container[];
|
const result = await cache.get<Container[]>(`${newt.newtId}:dockerContainers`);
|
||||||
if (!result) {
|
if (!result) {
|
||||||
throw createHttpError(
|
throw createHttpError(
|
||||||
HttpCode.TOO_EARLY,
|
HttpCode.TOO_EARLY,
|
||||||
@@ -173,7 +173,7 @@ async function isDockerAvailable(siteId: number): Promise<boolean> {
|
|||||||
const { newt } = await getSiteAndNewt(siteId);
|
const { newt } = await getSiteAndNewt(siteId);
|
||||||
|
|
||||||
const key = `${newt.newtId}:isAvailable`;
|
const key = `${newt.newtId}:isAvailable`;
|
||||||
const isAvailable = cache.get(key);
|
const isAvailable = await cache.get(key);
|
||||||
|
|
||||||
return !!isAvailable;
|
return !!isAvailable;
|
||||||
}
|
}
|
||||||
@@ -186,9 +186,11 @@ async function getDockerStatus(
|
|||||||
const keys = ["isAvailable", "socketPath"];
|
const keys = ["isAvailable", "socketPath"];
|
||||||
const mappedKeys = keys.map((x) => `${newt.newtId}:${x}`);
|
const mappedKeys = keys.map((x) => `${newt.newtId}:${x}`);
|
||||||
|
|
||||||
|
const values = await cache.mget<boolean | string>(mappedKeys);
|
||||||
|
|
||||||
const result = {
|
const result = {
|
||||||
isAvailable: cache.get(mappedKeys[0]) as boolean,
|
isAvailable: values[0] as boolean,
|
||||||
socketPath: cache.get(mappedKeys[1]) as string | undefined
|
socketPath: values[1] as string | undefined
|
||||||
};
|
};
|
||||||
|
|
||||||
return result;
|
return result;
|
||||||
|
|||||||
@@ -191,7 +191,7 @@ export async function inviteUser(
|
|||||||
}
|
}
|
||||||
|
|
||||||
if (existingInvite.length) {
|
if (existingInvite.length) {
|
||||||
const attempts = cache.get<number>(email) || 0;
|
const attempts = (await cache.get<number>(email)) || 0;
|
||||||
if (attempts >= 3) {
|
if (attempts >= 3) {
|
||||||
return next(
|
return next(
|
||||||
createHttpError(
|
createHttpError(
|
||||||
@@ -201,7 +201,7 @@ export async function inviteUser(
|
|||||||
);
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
cache.set(email, attempts + 1);
|
await cache.set(email, attempts + 1);
|
||||||
|
|
||||||
const inviteId = existingInvite[0].inviteId; // Retrieve the original inviteId
|
const inviteId = existingInvite[0].inviteId; // Retrieve the original inviteId
|
||||||
const token = generateRandomString(
|
const token = generateRandomString(
|
||||||
|
|||||||
Reference in New Issue
Block a user