When multiple Moltbot instances share a channel, they can now coordinate via Discord typing indicators to avoid response storms. New config options (messages.inbound): - peerBots: array of bot user IDs to watch - peerTypingDelayMs: backoff delay when peer is typing (default: 3000) - peerTypingMaxRetries: max retries before proceeding anyway (default: 3) Implementation: - PeerTypingListener tracks TYPING_START events from configured peers - onFlush checks peer typing state before processing - If peer is typing, message is re-queued with backoff delay - Debouncer now exposes requeue() for custom delay re-enqueueing Fixes race condition where multiple bots respond simultaneously.
147 lines
4.4 KiB
TypeScript
147 lines
4.4 KiB
TypeScript
import type { MoltbotConfig } from "../config/config.js";
|
|
import type { InboundDebounceByProvider } from "../config/types.messages.js";
|
|
|
|
const resolveMs = (value: unknown): number | undefined => {
|
|
if (typeof value !== "number" || !Number.isFinite(value)) return undefined;
|
|
return Math.max(0, Math.trunc(value));
|
|
};
|
|
|
|
const resolveChannelOverride = (params: {
|
|
byChannel?: InboundDebounceByProvider;
|
|
channel: string;
|
|
}): number | undefined => {
|
|
if (!params.byChannel) return undefined;
|
|
return resolveMs(params.byChannel[params.channel]);
|
|
};
|
|
|
|
export function resolveInboundDebounceMs(params: {
|
|
cfg: MoltbotConfig;
|
|
channel: string;
|
|
overrideMs?: number;
|
|
}): number {
|
|
const inbound = params.cfg.messages?.inbound;
|
|
const override = resolveMs(params.overrideMs);
|
|
const byChannel = resolveChannelOverride({
|
|
byChannel: inbound?.byChannel,
|
|
channel: params.channel,
|
|
});
|
|
const base = resolveMs(inbound?.debounceMs);
|
|
return override ?? byChannel ?? base ?? 0;
|
|
}
|
|
|
|
export function resolvePeerBots(params: { cfg: MoltbotConfig }): string[] {
|
|
const peerBots = params.cfg.messages?.inbound?.peerBots;
|
|
if (!Array.isArray(peerBots)) return [];
|
|
return peerBots.filter((id): id is string => typeof id === "string" && id.trim().length > 0);
|
|
}
|
|
|
|
export function resolvePeerTypingDelayMs(params: { cfg: MoltbotConfig }): number {
|
|
return resolveMs(params.cfg.messages?.inbound?.peerTypingDelayMs) ?? 3000;
|
|
}
|
|
|
|
export function resolvePeerTypingMaxRetries(params: { cfg: MoltbotConfig }): number {
|
|
return resolveMs(params.cfg.messages?.inbound?.peerTypingMaxRetries) ?? 3;
|
|
}
|
|
|
|
type DebounceBuffer<T> = {
|
|
items: T[];
|
|
timeout: ReturnType<typeof setTimeout> | null;
|
|
};
|
|
|
|
export type DebounceFlushContext<T> = {
|
|
/** Re-enqueue an item with a custom delay (e.g., for peer typing backoff) */
|
|
requeue: (item: T, delayMs: number) => void;
|
|
};
|
|
|
|
export function createInboundDebouncer<T>(params: {
|
|
debounceMs: number;
|
|
buildKey: (item: T) => string | null | undefined;
|
|
shouldDebounce?: (item: T) => boolean;
|
|
onFlush: (items: T[], ctx: DebounceFlushContext<T>) => Promise<void>;
|
|
onError?: (err: unknown, items: T[]) => void;
|
|
}) {
|
|
const buffers = new Map<string, DebounceBuffer<T>>();
|
|
const debounceMs = Math.max(0, Math.trunc(params.debounceMs));
|
|
|
|
const scheduleFlushWithDelay = (key: string, buffer: DebounceBuffer<T>, delayMs: number) => {
|
|
if (buffer.timeout) clearTimeout(buffer.timeout);
|
|
buffer.timeout = setTimeout(() => {
|
|
void flushBuffer(key, buffer);
|
|
}, delayMs);
|
|
buffer.timeout.unref?.();
|
|
};
|
|
|
|
const requeueItem = (item: T, delayMs: number) => {
|
|
const key = params.buildKey(item);
|
|
if (!key) return;
|
|
|
|
const existing = buffers.get(key);
|
|
if (existing) {
|
|
existing.items.push(item);
|
|
scheduleFlushWithDelay(key, existing, delayMs);
|
|
} else {
|
|
const buffer: DebounceBuffer<T> = { items: [item], timeout: null };
|
|
buffers.set(key, buffer);
|
|
scheduleFlushWithDelay(key, buffer, delayMs);
|
|
}
|
|
};
|
|
|
|
const flushContext: DebounceFlushContext<T> = {
|
|
requeue: requeueItem,
|
|
};
|
|
|
|
const flushBuffer = async (key: string, buffer: DebounceBuffer<T>) => {
|
|
buffers.delete(key);
|
|
if (buffer.timeout) {
|
|
clearTimeout(buffer.timeout);
|
|
buffer.timeout = null;
|
|
}
|
|
if (buffer.items.length === 0) return;
|
|
try {
|
|
await params.onFlush(buffer.items, flushContext);
|
|
} catch (err) {
|
|
params.onError?.(err, buffer.items);
|
|
}
|
|
};
|
|
|
|
const flushKey = async (key: string) => {
|
|
const buffer = buffers.get(key);
|
|
if (!buffer) return;
|
|
await flushBuffer(key, buffer);
|
|
};
|
|
|
|
const scheduleFlush = (key: string, buffer: DebounceBuffer<T>) => {
|
|
if (buffer.timeout) clearTimeout(buffer.timeout);
|
|
buffer.timeout = setTimeout(() => {
|
|
void flushBuffer(key, buffer);
|
|
}, debounceMs);
|
|
buffer.timeout.unref?.();
|
|
};
|
|
|
|
const enqueue = async (item: T) => {
|
|
const key = params.buildKey(item);
|
|
const canDebounce = debounceMs > 0 && (params.shouldDebounce?.(item) ?? true);
|
|
|
|
if (!canDebounce || !key) {
|
|
if (key && buffers.has(key)) {
|
|
await flushKey(key);
|
|
}
|
|
await params.onFlush([item], flushContext);
|
|
return;
|
|
}
|
|
|
|
const existing = buffers.get(key);
|
|
if (existing) {
|
|
existing.items.push(item);
|
|
scheduleFlush(key, existing);
|
|
return;
|
|
}
|
|
|
|
const buffer: DebounceBuffer<T> = { items: [item], timeout: null };
|
|
buffers.set(key, buffer);
|
|
scheduleFlush(key, buffer);
|
|
};
|
|
|
|
return { enqueue, flushKey };
|
|
}
|