import { join } from "node:path"; import { afterEach, describe, expect, it, vi } from "vitest"; import { normalizeTestText } from "../../test/helpers/normalize-text.js"; import { withTempHome as withTempHomeBase } from "../../test/helpers/temp-home.js"; vi.mock("../agents/pi-embedded.js", () => ({ abortEmbeddedPiRun: vi.fn().mockReturnValue(false), compactEmbeddedPiSession: vi.fn(), runEmbeddedPiAgent: vi.fn(), queueEmbeddedPiMessage: vi.fn().mockReturnValue(false), resolveEmbeddedSessionLane: (key: string) => `session:${key.trim() || "main"}`, isEmbeddedPiRunActive: vi.fn().mockReturnValue(false), isEmbeddedPiRunStreaming: vi.fn().mockReturnValue(false), })); const usageMocks = vi.hoisted(() => ({ loadProviderUsageSummary: vi.fn().mockResolvedValue({ updatedAt: 0, providers: [], }), formatUsageSummaryLine: vi.fn().mockReturnValue("📊 Usage: Claude 80% left"), resolveUsageProviderId: vi.fn((provider: string) => provider.split("/")[0]), })); vi.mock("../infra/provider-usage.js", () => usageMocks); const modelCatalogMocks = vi.hoisted(() => ({ loadModelCatalog: vi.fn().mockResolvedValue([ { provider: "anthropic", id: "claude-opus-4-5", name: "Claude Opus 4.5", contextWindow: 200000, }, { provider: "openrouter", id: "anthropic/claude-opus-4-5", name: "Claude Opus 4.5 (OpenRouter)", contextWindow: 200000, }, { provider: "openai", id: "gpt-4.1-mini", name: "GPT-4.1 mini" }, { provider: "openai", id: "gpt-5.2", name: "GPT-5.2" }, { provider: "openai-codex", id: "gpt-5.2", name: "GPT-5.2 (Codex)" }, { provider: "minimax", id: "MiniMax-M2.1", name: "MiniMax M2.1" }, ]), resetModelCatalogCacheForTest: vi.fn(), })); vi.mock("../agents/model-catalog.js", () => modelCatalogMocks); import { abortEmbeddedPiRun, runEmbeddedPiAgent } from "../agents/pi-embedded.js"; import { loadSessionStore } from "../config/sessions.js"; import { getReplyFromConfig } from "./reply.js"; const _MAIN_SESSION_KEY = "agent:main:main"; const webMocks = vi.hoisted(() => ({ webAuthExists: vi.fn().mockResolvedValue(true), getWebAuthAgeMs: vi.fn().mockReturnValue(120_000), readWebSelfId: vi.fn().mockReturnValue({ e164: "+1999" }), })); vi.mock("../web/session.js", () => webMocks); async function withTempHome(fn: (home: string) => Promise): Promise { return withTempHomeBase( async (home) => { vi.mocked(runEmbeddedPiAgent).mockClear(); vi.mocked(abortEmbeddedPiRun).mockClear(); return await fn(home); }, { prefix: "clawdbot-triggers-" }, ); } function makeCfg(home: string) { return { agents: { defaults: { model: "anthropic/claude-opus-4-5", workspace: join(home, "clawd"), }, }, channels: { whatsapp: { allowFrom: ["*"], }, }, session: { store: join(home, "sessions.json") }, }; } afterEach(() => { vi.restoreAllMocks(); }); describe("trigger handling", () => { it("shows a /model summary and points to /models", async () => { await withTempHome(async (home) => { const cfg = makeCfg(home); const res = await getReplyFromConfig( { Body: "/model", From: "telegram:111", To: "telegram:111", ChatType: "direct", Provider: "telegram", Surface: "telegram", SessionKey: "telegram:slash:111", CommandAuthorized: true, }, {}, cfg, ); const text = Array.isArray(res) ? res[0]?.text : res?.text; const normalized = normalizeTestText(text ?? ""); expect(normalized).toContain("Current: anthropic/claude-opus-4-5"); expect(normalized).toContain("Switch: /model "); expect(normalized).toContain("Browse: /models (providers) or /models (models)"); expect(normalized).toContain("More: /model status"); expect(normalized).not.toContain("reasoning"); expect(normalized).not.toContain("image"); }); }); it("aliases /model list to /models", async () => { await withTempHome(async (home) => { const cfg = makeCfg(home); const res = await getReplyFromConfig( { Body: "/model list", From: "telegram:111", To: "telegram:111", ChatType: "direct", Provider: "telegram", Surface: "telegram", SessionKey: "telegram:slash:111", CommandAuthorized: true, }, {}, cfg, ); const text = Array.isArray(res) ? res[0]?.text : res?.text; const normalized = normalizeTestText(text ?? ""); expect(normalized).toContain("Providers:"); expect(normalized).toContain("Use: /models "); expect(normalized).toContain("Switch: /model "); }); }); it("selects the exact provider/model pair for openrouter", async () => { await withTempHome(async (home) => { const cfg = makeCfg(home); const sessionKey = "telegram:slash:111"; const res = await getReplyFromConfig( { Body: "/model openrouter/anthropic/claude-opus-4-5", From: "telegram:111", To: "telegram:111", ChatType: "direct", Provider: "telegram", Surface: "telegram", SessionKey: sessionKey, CommandAuthorized: true, }, {}, cfg, ); const text = Array.isArray(res) ? res[0]?.text : res?.text; expect(normalizeTestText(text ?? "")).toContain( "Model set to openrouter/anthropic/claude-opus-4-5", ); const store = loadSessionStore(cfg.session.store); expect(store[sessionKey]?.providerOverride).toBe("openrouter"); expect(store[sessionKey]?.modelOverride).toBe("anthropic/claude-opus-4-5"); }); }); it("rejects invalid /model <#> selections", async () => { await withTempHome(async (home) => { const cfg = makeCfg(home); const sessionKey = "telegram:slash:111"; const res = await getReplyFromConfig( { Body: "/model 99", From: "telegram:111", To: "telegram:111", ChatType: "direct", Provider: "telegram", Surface: "telegram", SessionKey: sessionKey, CommandAuthorized: true, }, {}, cfg, ); const text = Array.isArray(res) ? res[0]?.text : res?.text; const normalized = normalizeTestText(text ?? ""); expect(normalized).toContain("Numeric model selection is not supported in chat."); expect(normalized).toContain("Browse: /models or /models "); expect(normalized).toContain("Switch: /model "); const store = loadSessionStore(cfg.session.store); expect(store[sessionKey]?.providerOverride).toBeUndefined(); expect(store[sessionKey]?.modelOverride).toBeUndefined(); }); }); it("resets to the default model via /model ", async () => { await withTempHome(async (home) => { const cfg = makeCfg(home); const sessionKey = "telegram:slash:111"; const res = await getReplyFromConfig( { Body: "/model anthropic/claude-opus-4-5", From: "telegram:111", To: "telegram:111", ChatType: "direct", Provider: "telegram", Surface: "telegram", SessionKey: sessionKey, CommandAuthorized: true, }, {}, cfg, ); const text = Array.isArray(res) ? res[0]?.text : res?.text; expect(normalizeTestText(text ?? "")).toContain( "Model reset to default (anthropic/claude-opus-4-5)", ); const store = loadSessionStore(cfg.session.store); // When selecting the default, overrides are cleared expect(store[sessionKey]?.providerOverride).toBeUndefined(); expect(store[sessionKey]?.modelOverride).toBeUndefined(); }); }); it("selects a model via /model ", async () => { await withTempHome(async (home) => { const cfg = makeCfg(home); const sessionKey = "telegram:slash:111"; const res = await getReplyFromConfig( { Body: "/model openai/gpt-5.2", From: "telegram:111", To: "telegram:111", ChatType: "direct", Provider: "telegram", Surface: "telegram", SessionKey: sessionKey, CommandAuthorized: true, }, {}, cfg, ); const text = Array.isArray(res) ? res[0]?.text : res?.text; expect(normalizeTestText(text ?? "")).toContain("Model set to openai/gpt-5.2"); const store = loadSessionStore(cfg.session.store); expect(store[sessionKey]?.providerOverride).toBe("openai"); expect(store[sessionKey]?.modelOverride).toBe("gpt-5.2"); }); }); });