perf(test): speed up memory suites

This commit is contained in:
Peter Steinberger
2026-02-14 16:03:14 +00:00
parent a7142c6218
commit f2c56de955
5 changed files with 47 additions and 35 deletions

View File

@@ -89,7 +89,7 @@ describe("memory index", () => {
throw new Error("manager missing");
}
manager = result.manager;
await result.manager.sync({ force: true });
await result.manager.sync({ reason: "test" });
const results = await result.manager.search("alpha");
expect(results.length).toBeGreaterThan(0);
expect(results[0]?.path).toContain("memory/2026-01-12.md");
@@ -141,7 +141,7 @@ describe("memory index", () => {
if (!first.manager) {
throw new Error("manager missing");
}
await first.manager.sync({ force: true });
await first.manager.sync({ reason: "test" });
const callsAfterFirstSync = embedBatchCalls;
await first.manager.close();
@@ -234,7 +234,7 @@ describe("memory index", () => {
return;
}
await manager.sync({ force: true });
await manager.sync({ reason: "test" });
const results = await manager.search("zebra");
expect(results.length).toBeGreaterThan(0);
expect(results[0]?.path).toContain("memory/2026-01-12.md");

View File

@@ -172,7 +172,6 @@ describe("memory indexing with OpenAI batches", () => {
manager = result.manager;
const labels: string[] = [];
await manager.sync({
force: true,
progress: (update) => {
if (update.label) {
labels.push(update.label);
@@ -287,7 +286,7 @@ describe("memory indexing with OpenAI batches", () => {
throw new Error("manager missing");
}
manager = result.manager;
await manager.sync({ force: true });
await manager.sync({ reason: "test" });
const status = manager.status();
expect(status.chunks).toBeGreaterThan(0);
@@ -300,7 +299,16 @@ describe("memory indexing with OpenAI batches", () => {
it("tracks batch failures, resets on success, and disables after repeated failures", async () => {
const restoreTimeouts = useFastShortTimeouts();
const content = ["flaky", "batch"].join("\n\n");
await fs.writeFile(path.join(workspaceDir, "memory", "2026-01-09.md"), content);
const memoryFile = path.join(workspaceDir, "memory", "2026-01-09.md");
let mtimeMs = Date.now();
const touch = async () => {
mtimeMs += 1_000;
const date = new Date(mtimeMs);
await fs.utimes(memoryFile, date, date);
};
await fs.writeFile(memoryFile, content);
await touch();
let uploadedRequests: Array<{ custom_id?: string }> = [];
let mode: "fail" | "ok" = "fail";
@@ -395,20 +403,24 @@ describe("memory indexing with OpenAI batches", () => {
manager = result.manager;
// First failure: fallback to regular embeddings and increment failure count.
await manager.sync({ force: true });
await manager.sync({ reason: "test" });
expect(embedBatch).toHaveBeenCalled();
let status = manager.status();
expect(status.batch?.enabled).toBe(true);
expect(status.batch?.failures).toBe(1);
const markDirty = () => {
// `sync` only indexes when marked dirty (unless doing a full reindex).
(manager as unknown as { dirty: boolean }).dirty = true;
};
// Success should reset failure count.
embedBatch.mockClear();
mode = "ok";
await fs.writeFile(
path.join(workspaceDir, "memory", "2026-01-09.md"),
["flaky", "batch", "recovery"].join("\n\n"),
);
await manager.sync({ force: true });
await fs.writeFile(memoryFile, ["flaky", "batch", "recovery"].join("\n\n"));
await touch();
markDirty();
await manager.sync({ reason: "test" });
status = manager.status();
expect(status.batch?.enabled).toBe(true);
expect(status.batch?.failures).toBe(0);
@@ -416,20 +428,18 @@ describe("memory indexing with OpenAI batches", () => {
// Two more failures after reset should disable remote batching.
mode = "fail";
await fs.writeFile(
path.join(workspaceDir, "memory", "2026-01-09.md"),
["flaky", "batch", "fail-a"].join("\n\n"),
);
await manager.sync({ force: true });
await fs.writeFile(memoryFile, ["flaky", "batch", "fail-a"].join("\n\n"));
await touch();
markDirty();
await manager.sync({ reason: "test" });
status = manager.status();
expect(status.batch?.enabled).toBe(true);
expect(status.batch?.failures).toBe(1);
await fs.writeFile(
path.join(workspaceDir, "memory", "2026-01-09.md"),
["flaky", "batch", "fail-b"].join("\n\n"),
);
await manager.sync({ force: true });
await fs.writeFile(memoryFile, ["flaky", "batch", "fail-b"].join("\n\n"));
await touch();
markDirty();
await manager.sync({ reason: "test" });
status = manager.status();
expect(status.batch?.enabled).toBe(false);
expect(status.batch?.failures).toBeGreaterThanOrEqual(2);
@@ -437,11 +447,10 @@ describe("memory indexing with OpenAI batches", () => {
// Once disabled, batch endpoints are skipped and fallback embeddings run directly.
const fetchCalls = fetchMock.mock.calls.length;
embedBatch.mockClear();
await fs.writeFile(
path.join(workspaceDir, "memory", "2026-01-09.md"),
["flaky", "batch", "fallback"].join("\n\n"),
);
await manager.sync({ force: true });
await fs.writeFile(memoryFile, ["flaky", "batch", "fallback"].join("\n\n"));
await touch();
markDirty();
await manager.sync({ reason: "test" });
expect(fetchMock.mock.calls.length).toBe(fetchCalls);
expect(embedBatch).toHaveBeenCalled();
} finally {

View File

@@ -85,7 +85,6 @@ describe("memory embedding batches", () => {
manager = result.manager;
const updates: Array<{ completed: number; total: number; label?: string }> = [];
await manager.sync({
force: true,
progress: (update) => {
updates.push(update);
},
@@ -130,7 +129,7 @@ describe("memory embedding batches", () => {
throw new Error("manager missing");
}
manager = result.manager;
await manager.sync({ force: true });
await manager.sync({ reason: "test" });
expect(embedBatch.mock.calls.length).toBe(1);
});
@@ -191,7 +190,7 @@ describe("memory embedding batches", () => {
}
manager = result.manager;
try {
await manager.sync({ force: true });
await manager.sync({ reason: "test" });
} finally {
setTimeoutSpy.mockRestore();
}
@@ -224,7 +223,7 @@ describe("memory embedding batches", () => {
throw new Error("manager missing");
}
manager = result.manager;
await manager.sync({ force: true });
await manager.sync({ reason: "test" });
const inputs = embedBatch.mock.calls.flatMap((call) => call[0] ?? []);
expect(inputs).not.toContain("");

View File

@@ -74,7 +74,7 @@ describe("memory embedding token limits", () => {
throw new Error("manager missing");
}
manager = result.manager;
await manager.sync({ force: true });
await manager.sync({ reason: "test" });
const inputs = embedBatch.mock.calls.flatMap((call) => call[0] ?? []);
expect(inputs.length).toBeGreaterThan(1);
@@ -111,7 +111,7 @@ describe("memory embedding token limits", () => {
throw new Error("manager missing");
}
manager = result.manager;
await manager.sync({ force: true });
await manager.sync({ reason: "test" });
const batchSizes = embedBatch.mock.calls.map(
(call) => (call[0] as string[] | undefined)?.length ?? 0,

View File

@@ -1009,12 +1009,16 @@ describe("QmdMemoryManager", () => {
});
async function waitForCondition(check: () => boolean, timeoutMs: number): Promise<void> {
const deadline = Date.now() + timeoutMs;
while (Date.now() < deadline) {
// Tests only need to yield the event loop a few times; real-time sleeps slow the suite down.
const maxTicks = Math.max(10, Math.min(5000, timeoutMs * 5));
for (let tick = 0; tick < maxTicks; tick += 1) {
if (check()) {
return;
}
await new Promise<void>((resolve) => setImmediate(resolve));
}
if (check()) {
return;
}
throw new Error("condition was not met in time");
}