diff --git a/static/script.js b/static/script.js
index 32f3a91..3a3aacc 100644
--- a/static/script.js
+++ b/static/script.js
@@ -181,15 +181,24 @@ function updateRunningModels(models) {
return;
}
- container.innerHTML = models.map(model => `
+ container.innerHTML = models.map(model => {
+ const gpuLayers = model.gpu_layers || 0;
+ const totalLayers = model.total_layers || 0;
+ const cpuLayers = totalLayers > 0 ? totalLayers - gpuLayers : 0;
+
+ return `
${escapeHtml(model.name)}
VRAM: ${model.vram_gb.toFixed(2)} GB
${model.offload_pct > 0 ? ` | CPU: ${model.offload_pct.toFixed(1)}%` : ''}
+
+ GPU: ${gpuLayers} layers | CPU: ${cpuLayers} layers
+
- `).join('');
+ `;
+ }).join('');
}
// ===== MODEL MANAGEMENT =====
diff --git a/static/style.css b/static/style.css
index 14f2b9f..b16f4f3 100644
--- a/static/style.css
+++ b/static/style.css
@@ -110,8 +110,16 @@ body {
}
.running-model-stats {
- font-size: 11px;
+ font-size: 0.85rem;
color: var(--text-secondary);
+ margin-top: 0.25rem;
+}
+
+.running-model-layers {
+ font-size: 0.85rem;
+ color: var(--accent);
+ margin-top: 0.25rem;
+ font-weight: 500;
}
.no-models {
diff --git a/web_app.py b/web_app.py
index 6049d73..0ecb296 100644
--- a/web_app.py
+++ b/web_app.py
@@ -398,12 +398,19 @@ def api_status():
size_total = model.get('size', 0) / (1024**3) # GB
offload_pct = ((size_total - size_vram) / size_total * 100) if size_total > 0 else 0
+ # Extract layer information from model details
+ details = model.get('details', {})
+ total_layers = details.get('parameter_size', 0)
+ gpu_layers = details.get('quantization_level', 0)
+
running_models.append({
'name': model.get('name', 'Unknown'),
'size_gb': size_total,
'vram_gb': size_vram,
'offload_pct': offload_pct,
- 'expires_at': model.get('expires_at', '')
+ 'expires_at': model.get('expires_at', ''),
+ 'gpu_layers': gpu_layers,
+ 'total_layers': total_layers
})
except Exception as e:
print(f"Error getting running models: {e}")