fix: prefer luker host llm routing

This commit is contained in:
Youzini-afk
2026-04-16 00:52:21 +08:00
parent 3ded93351d
commit 3feb7be760
2 changed files with 92 additions and 3 deletions

View File

@@ -1962,10 +1962,20 @@ async function callDedicatedOpenAICompatible(
const transportMessages = buildTransportMessages(messages);
const config = getMemoryLLMConfig(taskType);
const hostRouting = resolveHostChatCompletionRouting(taskType, {
profileName: "",
profileName: config.requestedLlmPresetName || "",
});
const settings = extension_settings[MODULE_NAME] || {};
const hasDedicatedConfig = hasDedicatedLLMConfig(config);
const shouldPreferLukerHostRoute =
hostRouting.hostProfile === "luker" &&
(
config.llmConfigSource === "global" ||
(
String(config.llmConfigSource || "").startsWith("global-fallback-") &&
hostRouting.routeApplied === true
)
);
const hasDedicatedConfig =
hasDedicatedLLMConfig(config) && !shouldPreferLukerHostRoute;
if (taskType && config.llmPresetFallbackReason) {
debugWarn(
`[ST-BME] 任务 ${taskType} 指定的 API 模板不可用,已回退当前 API: ` +
@@ -2042,6 +2052,7 @@ async function callDedicatedOpenAICompatible(
hostRequestApi: hostRouting.requestApi,
hostRouteApplied: hostRouting.routeApplied,
hostRouteReason: hostRouting.routeReason,
preferHostRoute: shouldPreferLukerHostRoute,
apiSettingsOverride: hostRouting.apiSettingsOverride,
maxCompletionTokens,
...buildStreamDebugSnapshot(streamState),