added model list to proxy log in diagnostic server
This commit is contained in:
+10
-3
@@ -365,8 +365,14 @@ async function testProxy(prompt) {
|
||||
}
|
||||
|
||||
const modelsData = await modelsRes.json();
|
||||
const modelCount = modelsData.data?.length || 0;
|
||||
log("info", "proxy", `Proxy erreichbar — ${modelCount} Model(s) verfuegbar`);
|
||||
const models = (modelsData.data || []).map(m => m.id).filter(Boolean);
|
||||
log("info", "proxy", `Proxy erreichbar — ${models.length} Model(s) verfuegbar`);
|
||||
|
||||
// Modellnamen loggen + OpenClaw-Config Hinweis
|
||||
if (models.length > 0) {
|
||||
log("info", "proxy", `Modelle: ${models.join(", ")}`);
|
||||
log("info", "proxy", `Fuer docker-compose.yml (DEFAULT_MODEL): ${models.map(m => m.replace("openai/", "")).join(" | ")}`);
|
||||
}
|
||||
|
||||
// Schritt 2: Chat Completion testen (kurzer Prompt)
|
||||
const testPrompt = prompt || "Antworte mit genau einem Wort: Ping";
|
||||
@@ -397,8 +403,9 @@ async function testProxy(prompt) {
|
||||
|
||||
state.proxy.status = "connected";
|
||||
state.proxy.lastError = null;
|
||||
state.proxy.models = models;
|
||||
broadcastState();
|
||||
broadcast({ type: "proxy_result", ok: true, reply });
|
||||
broadcast({ type: "proxy_result", ok: true, reply, models });
|
||||
|
||||
} catch (err) {
|
||||
log("error", "proxy", `Fehler: ${err.message}`);
|
||||
|
||||
Reference in New Issue
Block a user