Skip to content
Closed
Changes from 1 commit
Commits
Show all changes
42 commits
Select commit Hold shift + click to select a range
4f91839
Bulldozer DB
N2D4 Mar 24, 2026
9694c33
declareGroupByTable
N2D4 Mar 24, 2026
09ba416
Fix Prisma schema
N2D4 Mar 24, 2026
4232a94
declareMapTable
N2D4 Mar 24, 2026
31b6ac6
Performance tests
N2D4 Mar 24, 2026
2f7f09a
Load tests
N2D4 Mar 24, 2026
863ee05
Interface updates
N2D4 Mar 24, 2026
109cf5d
Bulldozer Studio
N2D4 Mar 25, 2026
53f7302
Remove unnecessary table
N2D4 Mar 25, 2026
006cf5e
Add flat map interface
N2D4 Mar 25, 2026
4cdc057
FlatMap table
N2D4 Mar 25, 2026
49dc922
Flat map fuzz tests
N2D4 Mar 25, 2026
3eccc22
Build MapTable from FlatMapTable
N2D4 Mar 25, 2026
e041e7d
Filter tables
N2D4 Mar 25, 2026
e2b0d3d
Limit tables
N2D4 Mar 25, 2026
f7f21aa
Speed up fuzzing
N2D4 Mar 26, 2026
90f61ec
Concat table
N2D4 Mar 26, 2026
69b3d4f
Bulldozer Studio: Better node placing algorithm
N2D4 Mar 26, 2026
e306770
Add left-join table
N2D4 Mar 26, 2026
5036c2a
Sort table
N2D4 Mar 27, 2026
d55470b
LFold table
N2D4 Mar 27, 2026
43d7304
Left join table
N2D4 Mar 27, 2026
a33faa0
Improve left join performance
N2D4 Mar 27, 2026
aaeb7d1
Improved performance for most tables
N2D4 Mar 27, 2026
ef9915a
Refactor Bulldozer into individual files
N2D4 Mar 27, 2026
037d20f
Merge branch 'dev' into bulldozer-db
N2D4 Mar 27, 2026
2403c17
Update apps/backend/src/lib/bulldozer/db/tables/group-by-table.ts
N2D4 Mar 27, 2026
d3a2daa
Performance improvements
N2D4 Mar 28, 2026
b459240
PR comments
N2D4 Mar 29, 2026
c01d931
Some more perf changes
N2D4 Mar 30, 2026
2d49d23
Fix various comparison key bugs
N2D4 Mar 31, 2026
2bb89c2
Performance improvements
N2D4 Mar 31, 2026
199cdf2
Lint fixes
N2D4 Apr 1, 2026
1dc76dc
Merge remote-tracking branch 'origin/dev' into bulldozer-db
N2D4 Apr 1, 2026
e4a5221
More fixes...
N2D4 Apr 3, 2026
e353232
Various changes
N2D4 Apr 6, 2026
4b400e2
Comments from Aman
N2D4 Apr 8, 2026
5b69a18
Fix tests
N2D4 Apr 8, 2026
a7f999f
Improve perf tests
N2D4 Apr 10, 2026
19abfb3
TimeFold table
N2D4 Apr 11, 2026
9cb5f5b
Clean up Prisma schema
N2D4 Apr 11, 2026
e3c3865
Make migration warnings into errors
N2D4 Apr 13, 2026
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Prev Previous commit
Next Next commit
Load tests
  • Loading branch information
N2D4 committed Mar 24, 2026
commit 2f7f09afca8314985193094594fdef8023528658
266 changes: 252 additions & 14 deletions apps/backend/src/lib/bulldozer/db/index.perf.test.ts
Original file line number Diff line number Diff line change
@@ -1,3 +1,4 @@
import { getEnvBoolean } from "@stackframe/stack-shared/dist/utils/env";
import postgres from "postgres";
import { afterAll, beforeAll, beforeEach, describe, expect, it } from "vitest";
import { declareGroupByTable, declareMapTable, declareStoredTable, toExecutableSqlTransaction, toQueryableSqlQuery } from "./index";
Expand All @@ -14,6 +15,15 @@ type WorkloadOperation =
const TEST_DB_PREFIX = "stack_bulldozer_db_perf_test";
const DEFAULT_WARMUP_OPS = 80;
const DEFAULT_MEASURED_OPS = 500;
const DEFAULT_LOAD_ROW_COUNT = getEnvBoolean("CI") ? 200_000 : 20_000;
const LOAD_PREFILL_MAX_MS = 30_000;
const LOAD_COUNT_QUERY_MAX_MS = 5_000;
const LOAD_POINT_MUTATION_MAX_MS = 400;
const LOAD_SET_ROW_AVG_ITERATIONS = 10;
const LOAD_SET_ROW_AVG_MAX_MS = 50;
const LOAD_TABLE_DELETE_MAX_MS = 20_000;
const LOAD_DERIVED_INIT_MAX_MS = 90_000;
const LOAD_DERIVED_COUNT_QUERY_MAX_MS = 10_000;

function getTestDbUrls(): TestDb {
const env = Reflect.get(import.meta, "env");
Expand Down Expand Up @@ -77,15 +87,6 @@ function createWorkload(seed: number, operationCount: number): WorkloadOperation
return operations;
}

function operationCountFromEnv(varName: string, fallback: number): number {
const env = Reflect.get(import.meta, "env");
const raw = Reflect.get(env, varName);
if (typeof raw !== "string") return fallback;
const parsed = Number(raw);
if (!Number.isFinite(parsed) || parsed < 1) return fallback;
return Math.floor(parsed);
}

function logLine(message: string): void {
console.log(`${message}\n`);
}
Expand All @@ -104,6 +105,53 @@ describe.sequential("bulldozer db performance (real postgres)", () => {
return await sql.unsafe(toQueryableSqlQuery(query));
}

async function measureMs<T>(label: string, fn: () => Promise<T>): Promise<{ result: T, elapsedMs: number }> {
const startedAt = performance.now();
const result = await fn();
const elapsedMs = performance.now() - startedAt;
logLine(`[bulldozer-perf] ${label}: ${elapsedMs.toFixed(1)} ms`);
return { result, elapsedMs };
}

async function prefillStoredTableInSingleStatement(tableId: string, rowCount: number): Promise<void> {
const externalId = `external:${tableId}`;
await sql`
INSERT INTO "BulldozerStorageEngine" ("keyPath", "value")
SELECT "seedRows"."keyPath", "seedRows"."value"
FROM (
VALUES
(ARRAY[to_jsonb('table'::text), to_jsonb(${externalId}::text)]::jsonb[], 'null'::jsonb),
(ARRAY[to_jsonb('table'::text), to_jsonb(${externalId}::text), to_jsonb('storage'::text)]::jsonb[], 'null'::jsonb),
(ARRAY[to_jsonb('table'::text), to_jsonb(${externalId}::text), to_jsonb('storage'::text), to_jsonb('rows'::text)]::jsonb[], 'null'::jsonb),
(ARRAY[to_jsonb('table'::text), to_jsonb(${externalId}::text), to_jsonb('storage'::text), to_jsonb('metadata'::text)]::jsonb[], '{ "version": 1 }'::jsonb)
) AS "seedRows"("keyPath", "value")
UNION ALL
SELECT
ARRAY[
to_jsonb('table'::text),
to_jsonb(${externalId}::text),
to_jsonb('storage'::text),
to_jsonb('rows'::text),
to_jsonb(('seed-' || "n"::text)::text)
]::jsonb[],
jsonb_build_object(
'rowData',
jsonb_build_object(
'team',
CASE
WHEN "n" % 4 = 0 THEN 'null'::jsonb
WHEN "n" % 4 = 1 THEN to_jsonb('alpha'::text)
WHEN "n" % 4 = 2 THEN to_jsonb('beta'::text)
ELSE to_jsonb('gamma'::text)
END,
'value',
to_jsonb(("n" % 1000)::int)
)
)
FROM generate_series(1, ${rowCount}) AS "n"
`;
}

async function executeWorkload(
fromTable: ReturnType<typeof declareStoredTable<{ value: number, team: string | null }>>,
operations: WorkloadOperation[],
Expand Down Expand Up @@ -186,10 +234,8 @@ describe.sequential("bulldozer db performance (real postgres)", () => {
});

it("reports ops/sec for baseline and composed example setup", async () => {
const warmupCount = operationCountFromEnv("STACK_BULLDOZER_PERF_WARMUP_OPS", DEFAULT_WARMUP_OPS);
const measuredCount = operationCountFromEnv("STACK_BULLDOZER_PERF_MEASURED_OPS", DEFAULT_MEASURED_OPS);
const warmupOperations = createWorkload(111, warmupCount);
const measuredOperations = createWorkload(222, measuredCount);
const warmupOperations = createWorkload(111, DEFAULT_WARMUP_OPS);
const measuredOperations = createWorkload(222, DEFAULT_MEASURED_OPS);

const baseline = await benchmarkScenario({
name: "stored-table baseline",
Expand Down Expand Up @@ -266,10 +312,202 @@ describe.sequential("bulldozer db performance (real postgres)", () => {

const slowdownFactor = baseline.operationsPerSecond / composed.operationsPerSecond;
logLine(`[bulldozer-perf] slowdown factor (baseline/composed): ${slowdownFactor.toFixed(2)}x`);
logLine(`[bulldozer-perf] config: warmup=${warmupCount}, measured=${measuredCount}`);
logLine(`[bulldozer-perf] config: warmup=${DEFAULT_WARMUP_OPS}, measured=${DEFAULT_MEASURED_OPS}`);

expect(baseline.operationsPerSecond).toBeGreaterThan(0);
expect(composed.operationsPerSecond).toBeGreaterThan(0);
});

it("load test: prefilled stored table with hundreds of thousands of rows stays functional and fast", async () => {
const loadRowCount = DEFAULT_LOAD_ROW_COUNT;
const tableId = "load-prefilled-users";
const externalTableId = `external:${tableId}`;
const table = declareStoredTable<{ value: number, team: string | null }>({ tableId });

const prefill = await measureMs(`load prefill (${loadRowCount} rows)`, async () => {
await prefillStoredTableInSingleStatement(tableId, loadRowCount);
});
expect(prefill.elapsedMs).toBeLessThan(LOAD_PREFILL_MAX_MS);

const metadataInitializedRows = await sql`
SELECT EXISTS (
SELECT 1 FROM "BulldozerStorageEngine"
WHERE "keyPath" = ARRAY[
to_jsonb('table'::text),
to_jsonb(${externalTableId}::text),
to_jsonb('storage'::text),
to_jsonb('metadata'::text)
]::jsonb[]
) AS "initialized"
`;
expect(metadataInitializedRows[0].initialized).toBe(true);

const listRowsQuery = table.listRowsInGroup({
start: "start",
end: "end",
startInclusive: true,
endInclusive: true,
});
const countRows = await measureMs("load count via listRowsInGroup", async () => {
return await sql.unsafe(`
SELECT COUNT(*)::int AS "count"
FROM (${toQueryableSqlQuery(listRowsQuery)}) AS "rows"
`);
});
expect(countRows.elapsedMs).toBeLessThan(LOAD_COUNT_QUERY_MAX_MS);
expect(Number(countRows.result[0].count)).toBe(loadRowCount);

const setRowIterationTimes: number[] = [];
for (let i = 0; i < LOAD_SET_ROW_AVG_ITERATIONS; i++) {
const startedAt = performance.now();
await runStatements(table.setRow(
`seed-${Math.floor(loadRowCount / 2) + i}`,
expr(jsonbLiteral({ team: "beta", value: 777 + i })),
));
setRowIterationTimes.push(performance.now() - startedAt);
}
const setRowAverageMs = setRowIterationTimes.reduce((acc, value) => acc + value, 0) / setRowIterationTimes.length;
logLine(`[bulldozer-perf] load setRow average (${LOAD_SET_ROW_AVG_ITERATIONS} iterations): ${setRowAverageMs.toFixed(1)} ms`);
expect(setRowAverageMs).toBeLessThanOrEqual(LOAD_SET_ROW_AVG_MAX_MS);

const pointDelete = await measureMs("load point delete (deleteRow existing)", async () => {
await runStatements(table.deleteRow(`seed-${Math.floor(loadRowCount / 2) - 1}`));
});
expect(pointDelete.elapsedMs).toBeLessThan(LOAD_POINT_MUTATION_MAX_MS);

const countAfterDelete = await sql.unsafe(`
SELECT COUNT(*)::int AS "count"
FROM (${toQueryableSqlQuery(listRowsQuery)}) AS "rows"
`);
expect(Number(countAfterDelete[0].count)).toBe(loadRowCount - 1);

const groupedByTeam = declareGroupByTable({
tableId: "load-prefilled-users-by-team",
fromTable: table,
groupBy: { type: "mapper", sql: `"rowData"->'team' AS "groupKey"` },
});
const mappedByTeam = declareMapTable({
tableId: "load-prefilled-users-mapped",
fromTable: groupedByTeam,
mapper: { type: "mapper", sql: `
("rowData"->'team') AS "team",
(("rowData"->>'value')::int + 10) AS "valuePlusTen",
(
CASE
WHEN (("rowData"->>'value')::int + 10) >= 700 THEN 'high'
ELSE 'low'
END
) AS "bucket"
` },
});
const mappedTwice = declareMapTable({
tableId: "load-prefilled-users-mapped-twice",
fromTable: mappedByTeam,
mapper: { type: "mapper", sql: `
("rowData"->'team') AS "team",
("rowData"->'bucket') AS "bucket",
((("rowData"->>'valuePlusTen')::int * 2)) AS "valueScaled"
` },
});
const groupedByBucket = declareGroupByTable({
tableId: "load-prefilled-users-by-bucket",
fromTable: mappedTwice,
groupBy: { type: "mapper", sql: `"rowData"->'bucket' AS "groupKey"` },
});

const groupInit = await measureMs("load init groupedByTeam", async () => {
await runStatements(groupedByTeam.init());
});
expect(groupInit.elapsedMs).toBeLessThan(LOAD_DERIVED_INIT_MAX_MS);
const mapInit = await measureMs("load init mappedByTeam", async () => {
await runStatements(mappedByTeam.init());
});
expect(mapInit.elapsedMs).toBeLessThan(LOAD_DERIVED_INIT_MAX_MS);
const mapTwiceInit = await measureMs("load init mappedTwice", async () => {
await runStatements(mappedTwice.init());
});
expect(mapTwiceInit.elapsedMs).toBeLessThan(LOAD_DERIVED_INIT_MAX_MS);
const bucketInit = await measureMs("load init groupedByBucket", async () => {
await runStatements(groupedByBucket.init());
});
expect(bucketInit.elapsedMs).toBeLessThan(LOAD_DERIVED_INIT_MAX_MS);

const groupedCountQuery = groupedByTeam.listRowsInGroup({
start: "start",
end: "end",
startInclusive: true,
endInclusive: true,
});
const mappedCountQuery = mappedTwice.listRowsInGroup({
start: "start",
end: "end",
startInclusive: true,
endInclusive: true,
});
const bucketCountQuery = groupedByBucket.listRowsInGroup({
start: "start",
end: "end",
startInclusive: true,
endInclusive: true,
});
const derivedCounts = await measureMs("load count derived tables", async () => {
return await Promise.all([
sql.unsafe(`SELECT COUNT(*)::int AS "count" FROM (${toQueryableSqlQuery(groupedCountQuery)}) AS "rows"`),
sql.unsafe(`SELECT COUNT(*)::int AS "count" FROM (${toQueryableSqlQuery(mappedCountQuery)}) AS "rows"`),
sql.unsafe(`SELECT COUNT(*)::int AS "count" FROM (${toQueryableSqlQuery(bucketCountQuery)}) AS "rows"`),
]);
});
expect(derivedCounts.elapsedMs).toBeLessThan(LOAD_DERIVED_COUNT_QUERY_MAX_MS);
expect(Number(derivedCounts.result[0][0].count)).toBe(loadRowCount - 1);
expect(Number(derivedCounts.result[1][0].count)).toBe(loadRowCount - 1);
expect(Number(derivedCounts.result[2][0].count)).toBe(loadRowCount - 1);

await runStatements(table.setRow(
"seed-100000",
expr(jsonbLiteral({ team: "delta", value: 999 })),
));
const deltaGroupedRows = await readRows(groupedByTeam.listRowsInGroup({
groupKey: expr(`to_jsonb('delta'::text)`),
start: "start",
end: "end",
startInclusive: true,
endInclusive: true,
}));
expect(deltaGroupedRows.some((row) => row.rowidentifier === "seed-100000")).toBe(true);
const highBucketRows = await readRows(groupedByBucket.listRowsInGroup({
groupKey: expr(`to_jsonb('high'::text)`),
start: "start",
end: "end",
startInclusive: true,
endInclusive: true,
}));
const highBucketRow = highBucketRows.find((row) => row.rowidentifier === "seed-100000");
expect(highBucketRow).toBeDefined();
expect(highBucketRow?.rowdata).toEqual({
team: "delta",
bucket: "high",
valueScaled: 2018,
});

const bulkDelete = await measureMs("load full table delete", async () => {
await runStatements(table.delete());
});
expect(bulkDelete.elapsedMs).toBeLessThan(LOAD_TABLE_DELETE_MAX_MS);

const isInitializedRows = await sql`
SELECT EXISTS (
SELECT 1 FROM "BulldozerStorageEngine"
WHERE "keyPath" = ARRAY[
to_jsonb('table'::text),
to_jsonb(${externalTableId}::text),
to_jsonb('storage'::text),
to_jsonb('metadata'::text)
]::jsonb[]
) AS "initialized"
`;
expect(isInitializedRows[0].initialized).toBe(false);

logLine(`[bulldozer-perf] load thresholds(ms): prefill<=${LOAD_PREFILL_MAX_MS}, baseCount<=${LOAD_COUNT_QUERY_MAX_MS}, setRowAvg<=${LOAD_SET_ROW_AVG_MAX_MS} over ${LOAD_SET_ROW_AVG_ITERATIONS}, pointDelete<=${LOAD_POINT_MUTATION_MAX_MS}, derivedInit<=${LOAD_DERIVED_INIT_MAX_MS}, derivedCount<=${LOAD_DERIVED_COUNT_QUERY_MAX_MS}, tableDelete<=${LOAD_TABLE_DELETE_MAX_MS}`);
}, 180_000);
});

Loading