diff --git a/static/script.js b/static/script.js index 32f3a91..3a3aacc 100644 --- a/static/script.js +++ b/static/script.js @@ -181,15 +181,24 @@ function updateRunningModels(models) { return; } - container.innerHTML = models.map(model => ` + container.innerHTML = models.map(model => { + const gpuLayers = model.gpu_layers || 0; + const totalLayers = model.total_layers || 0; + const cpuLayers = totalLayers > 0 ? totalLayers - gpuLayers : 0; + + return `
${escapeHtml(model.name)}
VRAM: ${model.vram_gb.toFixed(2)} GB ${model.offload_pct > 0 ? ` | CPU: ${model.offload_pct.toFixed(1)}%` : ''}
+
+ GPU: ${gpuLayers} layers | CPU: ${cpuLayers} layers +
- `).join(''); + `; + }).join(''); } // ===== MODEL MANAGEMENT ===== diff --git a/static/style.css b/static/style.css index 14f2b9f..b16f4f3 100644 --- a/static/style.css +++ b/static/style.css @@ -110,8 +110,16 @@ body { } .running-model-stats { - font-size: 11px; + font-size: 0.85rem; color: var(--text-secondary); + margin-top: 0.25rem; +} + +.running-model-layers { + font-size: 0.85rem; + color: var(--accent); + margin-top: 0.25rem; + font-weight: 500; } .no-models { diff --git a/web_app.py b/web_app.py index 6049d73..0ecb296 100644 --- a/web_app.py +++ b/web_app.py @@ -398,12 +398,19 @@ def api_status(): size_total = model.get('size', 0) / (1024**3) # GB offload_pct = ((size_total - size_vram) / size_total * 100) if size_total > 0 else 0 + # Extract layer information from model details + details = model.get('details', {}) + total_layers = details.get('parameter_size', 0) + gpu_layers = details.get('quantization_level', 0) + running_models.append({ 'name': model.get('name', 'Unknown'), 'size_gb': size_total, 'vram_gb': size_vram, 'offload_pct': offload_pct, - 'expires_at': model.get('expires_at', '') + 'expires_at': model.get('expires_at', ''), + 'gpu_layers': gpu_layers, + 'total_layers': total_layers }) except Exception as e: print(f"Error getting running models: {e}")