import fs from "node:fs/promises"; import { join } from "node:path"; import { afterEach, describe, expect, it, vi } from "vitest"; import { normalizeTestText } from "../../test/helpers/normalize-text.js"; import { withTempHome as withTempHomeBase } from "../../test/helpers/temp-home.js"; vi.mock("../agents/pi-embedded.js", () => ({ abortEmbeddedPiRun: vi.fn().mockReturnValue(false), compactEmbeddedPiSession: vi.fn(), runEmbeddedPiAgent: vi.fn(), queueEmbeddedPiMessage: vi.fn().mockReturnValue(false), resolveEmbeddedSessionLane: (key: string) => `session:${key.trim() || "main"}`, isEmbeddedPiRunActive: vi.fn().mockReturnValue(false), isEmbeddedPiRunStreaming: vi.fn().mockReturnValue(false), })); const usageMocks = vi.hoisted(() => ({ loadProviderUsageSummary: vi.fn().mockResolvedValue({ updatedAt: 0, providers: [], }), formatUsageSummaryLine: vi.fn().mockReturnValue("📊 Usage: Claude 80% left"), resolveUsageProviderId: vi.fn((provider: string) => provider.split("/")[0]), })); vi.mock("../infra/provider-usage.js", () => usageMocks); const modelCatalogMocks = vi.hoisted(() => ({ loadModelCatalog: vi.fn().mockResolvedValue([ { provider: "anthropic", id: "claude-opus-4-5", name: "Claude Opus 4.5", contextWindow: 200000, }, { provider: "openrouter", id: "anthropic/claude-opus-4-5", name: "Claude Opus 4.5 (OpenRouter)", contextWindow: 200000, }, { provider: "openai", id: "gpt-4.1-mini", name: "GPT-4.1 mini" }, { provider: "openai", id: "gpt-5.2", name: "GPT-5.2" }, { provider: "openai-codex", id: "gpt-5.2", name: "GPT-5.2 (Codex)" }, { provider: "minimax", id: "MiniMax-M2.1", name: "MiniMax M2.1" }, ]), resetModelCatalogCacheForTest: vi.fn(), })); vi.mock("../agents/model-catalog.js", () => modelCatalogMocks); import { abortEmbeddedPiRun, runEmbeddedPiAgent } from "../agents/pi-embedded.js"; import { loadSessionStore } from "../config/sessions.js"; import { getReplyFromConfig } from "./reply.js"; const _MAIN_SESSION_KEY = "agent:main:main"; const webMocks = vi.hoisted(() => ({ webAuthExists: vi.fn().mockResolvedValue(true), getWebAuthAgeMs: vi.fn().mockReturnValue(120_000), readWebSelfId: vi.fn().mockReturnValue({ e164: "+1999" }), })); vi.mock("../web/session.js", () => webMocks); async function withTempHome(fn: (home: string) => Promise): Promise { return withTempHomeBase( async (home) => { vi.mocked(runEmbeddedPiAgent).mockClear(); vi.mocked(abortEmbeddedPiRun).mockClear(); return await fn(home); }, { prefix: "clawdbot-triggers-" }, ); } function makeCfg(home: string) { return { agents: { defaults: { model: "anthropic/claude-opus-4-5", workspace: join(home, "clawd"), }, }, channels: { whatsapp: { allowFrom: ["*"], }, }, session: { store: join(home, "sessions.json") }, }; } afterEach(() => { vi.restoreAllMocks(); }); describe("trigger handling", () => { it("shows a quick /model picker grouped by model with providers", async () => { await withTempHome(async (home) => { const cfg = makeCfg(home); const res = await getReplyFromConfig( { Body: "/model", From: "telegram:111", To: "telegram:111", ChatType: "direct", Provider: "telegram", Surface: "telegram", SessionKey: "telegram:slash:111", }, {}, cfg, ); const text = Array.isArray(res) ? res[0]?.text : res?.text; const normalized = normalizeTestText(text ?? ""); expect(normalized).toContain("Pick: /model <#> or /model "); expect(normalized).toContain("1) claude-opus-4-5 — anthropic, openrouter"); expect(normalized).toContain("3) gpt-5.2 — openai, openai-codex"); expect(normalized).toContain("More: /model status"); expect(normalized).not.toContain("reasoning"); expect(normalized).not.toContain("image"); }); }); it("rejects invalid /model <#> selections", async () => { await withTempHome(async (home) => { const cfg = makeCfg(home); const sessionKey = "telegram:slash:111"; const res = await getReplyFromConfig( { Body: "/model 99", From: "telegram:111", To: "telegram:111", ChatType: "direct", Provider: "telegram", Surface: "telegram", SessionKey: sessionKey, }, {}, cfg, ); const text = Array.isArray(res) ? res[0]?.text : res?.text; expect(normalizeTestText(text ?? "")).toContain( 'Invalid model selection "99". Use /model to list.', ); const store = loadSessionStore(cfg.session.store); expect(store[sessionKey]?.providerOverride).toBeUndefined(); expect(store[sessionKey]?.modelOverride).toBeUndefined(); }); }); it("prefers the current provider when selecting /model <#>", async () => { await withTempHome(async (home) => { const cfg = makeCfg(home); const sessionKey = "telegram:slash:111"; await fs.writeFile( cfg.session.store, JSON.stringify( { [sessionKey]: { sessionId: "session-openrouter", updatedAt: Date.now(), providerOverride: "openrouter", modelOverride: "anthropic/claude-opus-4-5", }, }, null, 2, ), ); const res = await getReplyFromConfig( { Body: "/model 1", From: "telegram:111", To: "telegram:111", ChatType: "direct", Provider: "telegram", Surface: "telegram", SessionKey: sessionKey, }, {}, cfg, ); const text = Array.isArray(res) ? res[0]?.text : res?.text; expect(normalizeTestText(text ?? "")).toContain( "Model set to openrouter/anthropic/claude-opus-4-5", ); const store = loadSessionStore(cfg.session.store); expect(store[sessionKey]?.providerOverride).toBe("openrouter"); expect(store[sessionKey]?.modelOverride).toBe("anthropic/claude-opus-4-5"); }); }); it("selects a model by index via /model <#>", async () => { await withTempHome(async (home) => { const cfg = makeCfg(home); const sessionKey = "telegram:slash:111"; const res = await getReplyFromConfig( { Body: "/model 3", From: "telegram:111", To: "telegram:111", ChatType: "direct", Provider: "telegram", Surface: "telegram", SessionKey: sessionKey, }, {}, cfg, ); const text = Array.isArray(res) ? res[0]?.text : res?.text; expect(normalizeTestText(text ?? "")).toContain("Model set to openai/gpt-5.2"); const store = loadSessionStore(cfg.session.store); expect(store[sessionKey]?.providerOverride).toBe("openai"); expect(store[sessionKey]?.modelOverride).toBe("gpt-5.2"); }); }); });