Compare commits
13 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
| 1656f20e96 | |||
| dcd350ec2c | |||
| a34b095857 | |||
| 7362464b46 | |||
| c8b79576fa | |||
| f53df5440b | |||
| d1b0e2c431 | |||
| b1719a60fc | |||
| d922e8d6f3 | |||
| 407bc9dbc2 | |||
| 3c278351c9 | |||
| 8da4522d85 | |||
| 5b1e39cfca |
6
Cargo.lock
generated
6
Cargo.lock
generated
@ -279,7 +279,7 @@ checksum = "a1d728cc89cf3aee9ff92b05e62b19ee65a02b5702cff7d5a377e32c6ae29d8d"
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "cm-dashboard"
|
name = "cm-dashboard"
|
||||||
version = "0.1.261"
|
version = "0.1.274"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"anyhow",
|
"anyhow",
|
||||||
"chrono",
|
"chrono",
|
||||||
@ -301,7 +301,7 @@ dependencies = [
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "cm-dashboard-agent"
|
name = "cm-dashboard-agent"
|
||||||
version = "0.1.261"
|
version = "0.1.274"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"anyhow",
|
"anyhow",
|
||||||
"async-trait",
|
"async-trait",
|
||||||
@ -325,7 +325,7 @@ dependencies = [
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "cm-dashboard-shared"
|
name = "cm-dashboard-shared"
|
||||||
version = "0.1.261"
|
version = "0.1.274"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"chrono",
|
"chrono",
|
||||||
"serde",
|
"serde",
|
||||||
|
|||||||
@ -1,6 +1,6 @@
|
|||||||
[package]
|
[package]
|
||||||
name = "cm-dashboard-agent"
|
name = "cm-dashboard-agent"
|
||||||
version = "0.1.262"
|
version = "0.1.275"
|
||||||
edition = "2021"
|
edition = "2021"
|
||||||
|
|
||||||
[dependencies]
|
[dependencies]
|
||||||
|
|||||||
@ -1,7 +1,7 @@
|
|||||||
use async_trait::async_trait;
|
use async_trait::async_trait;
|
||||||
use cm_dashboard_shared::{AgentData, BackupData, BackupDiskData, Status};
|
use cm_dashboard_shared::{AgentData, BackupData, BackupRepositoryData, Status};
|
||||||
use serde::{Deserialize, Serialize};
|
use serde::{Deserialize, Serialize};
|
||||||
use std::collections::{HashMap, HashSet};
|
use std::collections::HashMap;
|
||||||
use std::fs;
|
use std::fs;
|
||||||
use std::path::{Path, PathBuf};
|
use std::path::{Path, PathBuf};
|
||||||
use tracing::{debug, warn};
|
use tracing::{debug, warn};
|
||||||
@ -21,7 +21,7 @@ impl BackupCollector {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Scan directory for all backup status files
|
/// Scan directory for backup status file (nfs-backup.toml)
|
||||||
async fn scan_status_files(&self) -> Result<Vec<PathBuf>, CollectorError> {
|
async fn scan_status_files(&self) -> Result<Vec<PathBuf>, CollectorError> {
|
||||||
let status_path = Path::new(&self.status_dir);
|
let status_path = Path::new(&self.status_dir);
|
||||||
|
|
||||||
@ -30,30 +30,15 @@ impl BackupCollector {
|
|||||||
return Ok(Vec::new());
|
return Ok(Vec::new());
|
||||||
}
|
}
|
||||||
|
|
||||||
let mut status_files = Vec::new();
|
// Look for nfs-backup.toml (new NFS-based backup)
|
||||||
|
let nfs_backup_file = status_path.join("nfs-backup.toml");
|
||||||
match fs::read_dir(status_path) {
|
if nfs_backup_file.exists() {
|
||||||
Ok(entries) => {
|
return Ok(vec![nfs_backup_file]);
|
||||||
for entry in entries {
|
|
||||||
if let Ok(entry) = entry {
|
|
||||||
let path = entry.path();
|
|
||||||
if path.is_file() {
|
|
||||||
if let Some(filename) = path.file_name().and_then(|n| n.to_str()) {
|
|
||||||
if filename.starts_with("backup-status-") && filename.ends_with(".toml") {
|
|
||||||
status_files.push(path);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
Err(e) => {
|
|
||||||
warn!("Failed to read backup status directory: {}", e);
|
|
||||||
return Ok(Vec::new());
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
Ok(status_files)
|
// No backup status file found
|
||||||
|
debug!("No nfs-backup.toml found in {}", self.status_dir);
|
||||||
|
Ok(Vec::new())
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Read a single backup status file
|
/// Read a single backup status file
|
||||||
@ -76,24 +61,13 @@ impl BackupCollector {
|
|||||||
/// Calculate backup status from TOML status field
|
/// Calculate backup status from TOML status field
|
||||||
fn calculate_backup_status(status_str: &str) -> Status {
|
fn calculate_backup_status(status_str: &str) -> Status {
|
||||||
match status_str.to_lowercase().as_str() {
|
match status_str.to_lowercase().as_str() {
|
||||||
"success" => Status::Ok,
|
"success" | "completed" => Status::Ok,
|
||||||
"warning" => Status::Warning,
|
"warning" => Status::Warning,
|
||||||
"failed" | "error" => Status::Critical,
|
"failed" | "error" => Status::Critical,
|
||||||
_ => Status::Unknown,
|
_ => Status::Unknown,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Calculate usage status from disk usage percentage
|
|
||||||
fn calculate_usage_status(usage_percent: f32) -> Status {
|
|
||||||
if usage_percent < 80.0 {
|
|
||||||
Status::Ok
|
|
||||||
} else if usage_percent < 90.0 {
|
|
||||||
Status::Warning
|
|
||||||
} else {
|
|
||||||
Status::Critical
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Convert BackupStatusToml to BackupData and populate AgentData
|
/// Convert BackupStatusToml to BackupData and populate AgentData
|
||||||
async fn populate_backup_data(&self, agent_data: &mut AgentData) -> Result<(), CollectorError> {
|
async fn populate_backup_data(&self, agent_data: &mut AgentData) -> Result<(), CollectorError> {
|
||||||
let status_files = self.scan_status_files().await?;
|
let status_files = self.scan_status_files().await?;
|
||||||
@ -101,76 +75,47 @@ impl BackupCollector {
|
|||||||
if status_files.is_empty() {
|
if status_files.is_empty() {
|
||||||
debug!("No backup status files found");
|
debug!("No backup status files found");
|
||||||
agent_data.backup = BackupData {
|
agent_data.backup = BackupData {
|
||||||
|
last_backup_time: None,
|
||||||
|
backup_status: Status::Unknown,
|
||||||
repositories: Vec::new(),
|
repositories: Vec::new(),
|
||||||
repository_status: Status::Unknown,
|
|
||||||
disks: Vec::new(),
|
|
||||||
};
|
};
|
||||||
return Ok(());
|
return Ok(());
|
||||||
}
|
}
|
||||||
|
|
||||||
let mut all_repositories = HashSet::new();
|
// Aggregate repository data across all backup status files
|
||||||
let mut disks = Vec::new();
|
let mut repo_map: HashMap<String, BackupRepositoryData> = HashMap::new();
|
||||||
let mut worst_status = Status::Ok;
|
let mut worst_status = Status::Ok;
|
||||||
|
let mut latest_backup_time: Option<String> = None;
|
||||||
|
|
||||||
for status_file in status_files {
|
for status_file in status_files {
|
||||||
match self.read_status_file(&status_file).await {
|
match self.read_status_file(&status_file).await {
|
||||||
Ok(backup_status) => {
|
Ok(backup_status) => {
|
||||||
// Collect all service names
|
|
||||||
for service_name in backup_status.services.keys() {
|
|
||||||
all_repositories.insert(service_name.clone());
|
|
||||||
}
|
|
||||||
|
|
||||||
// Calculate backup status
|
// Calculate backup status
|
||||||
let backup_status_enum = Self::calculate_backup_status(&backup_status.status);
|
let backup_status_enum = Self::calculate_backup_status(&backup_status.status);
|
||||||
|
worst_status = worst_status.max(backup_status_enum);
|
||||||
|
|
||||||
// Calculate usage status from disk space
|
// Track latest backup time
|
||||||
let (usage_percent, used_gb, total_gb, usage_status) = if let Some(disk_space) = &backup_status.disk_space {
|
if latest_backup_time.is_none() || Some(&backup_status.start_time) > latest_backup_time.as_ref() {
|
||||||
let usage_pct = disk_space.usage_percent as f32;
|
latest_backup_time = Some(backup_status.start_time.clone());
|
||||||
(
|
}
|
||||||
usage_pct,
|
|
||||||
disk_space.used_gb as f32,
|
|
||||||
disk_space.total_gb as f32,
|
|
||||||
Self::calculate_usage_status(usage_pct),
|
|
||||||
)
|
|
||||||
} else {
|
|
||||||
(0.0, 0.0, 0.0, Status::Unknown)
|
|
||||||
};
|
|
||||||
|
|
||||||
// Update worst status
|
// Process each service in this backup
|
||||||
worst_status = worst_status.max(backup_status_enum).max(usage_status);
|
for (service_name, service_status) in backup_status.services {
|
||||||
|
// Convert bytes to GB
|
||||||
|
let repo_size_gb = service_status.repo_size_bytes as f32 / 1_073_741_824.0;
|
||||||
|
|
||||||
// Build service list for this disk
|
// Calculate service status
|
||||||
let services: Vec<String> = backup_status.services.keys().cloned().collect();
|
let service_status_enum = Self::calculate_backup_status(&service_status.status);
|
||||||
|
worst_status = worst_status.max(service_status_enum);
|
||||||
|
|
||||||
// Get min and max archive counts to detect inconsistencies
|
// Update or insert repository data
|
||||||
let archives_min: i64 = backup_status.services.values()
|
repo_map.insert(service_name.clone(), BackupRepositoryData {
|
||||||
.map(|service| service.archive_count)
|
name: service_name,
|
||||||
.min()
|
archive_count: service_status.archive_count,
|
||||||
.unwrap_or(0);
|
repo_size_gb,
|
||||||
|
status: service_status_enum,
|
||||||
let archives_max: i64 = backup_status.services.values()
|
});
|
||||||
.map(|service| service.archive_count)
|
}
|
||||||
.max()
|
|
||||||
.unwrap_or(0);
|
|
||||||
|
|
||||||
// Create disk data
|
|
||||||
let disk_data = BackupDiskData {
|
|
||||||
serial: backup_status.disk_serial_number.unwrap_or_else(|| "Unknown".to_string()),
|
|
||||||
product_name: backup_status.disk_product_name,
|
|
||||||
wear_percent: backup_status.disk_wear_percent,
|
|
||||||
temperature_celsius: None, // Not available in current TOML
|
|
||||||
last_backup_time: Some(backup_status.start_time),
|
|
||||||
backup_status: backup_status_enum,
|
|
||||||
disk_usage_percent: usage_percent,
|
|
||||||
disk_used_gb: used_gb,
|
|
||||||
disk_total_gb: total_gb,
|
|
||||||
usage_status,
|
|
||||||
services,
|
|
||||||
archives_min,
|
|
||||||
archives_max,
|
|
||||||
};
|
|
||||||
|
|
||||||
disks.push(disk_data);
|
|
||||||
}
|
}
|
||||||
Err(e) => {
|
Err(e) => {
|
||||||
warn!("Failed to read backup status file {:?}: {}", status_file, e);
|
warn!("Failed to read backup status file {:?}: {}", status_file, e);
|
||||||
@ -178,12 +123,14 @@ impl BackupCollector {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
let repositories: Vec<String> = all_repositories.into_iter().collect();
|
// Convert HashMap to sorted Vec
|
||||||
|
let mut repositories: Vec<BackupRepositoryData> = repo_map.into_values().collect();
|
||||||
|
repositories.sort_by(|a, b| a.name.cmp(&b.name));
|
||||||
|
|
||||||
agent_data.backup = BackupData {
|
agent_data.backup = BackupData {
|
||||||
|
last_backup_time: latest_backup_time,
|
||||||
|
backup_status: worst_status,
|
||||||
repositories,
|
repositories,
|
||||||
repository_status: worst_status,
|
|
||||||
disks,
|
|
||||||
};
|
};
|
||||||
|
|
||||||
Ok(())
|
Ok(())
|
||||||
|
|||||||
@ -114,7 +114,7 @@ impl DiskCollector {
|
|||||||
let mut cmd = TokioCommand::new("lsblk");
|
let mut cmd = TokioCommand::new("lsblk");
|
||||||
cmd.args(&["-rn", "-o", "NAME,MOUNTPOINT"]);
|
cmd.args(&["-rn", "-o", "NAME,MOUNTPOINT"]);
|
||||||
|
|
||||||
let output = run_command_with_timeout(cmd, 2).await
|
let output = run_command_with_timeout(cmd, 10).await
|
||||||
.map_err(|e| CollectorError::SystemRead {
|
.map_err(|e| CollectorError::SystemRead {
|
||||||
path: "block devices".to_string(),
|
path: "block devices".to_string(),
|
||||||
error: e.to_string(),
|
error: e.to_string(),
|
||||||
@ -184,7 +184,7 @@ impl DiskCollector {
|
|||||||
/// Get filesystem info for a single mount point
|
/// Get filesystem info for a single mount point
|
||||||
fn get_filesystem_info(&self, mount_point: &str) -> Result<(u64, u64), CollectorError> {
|
fn get_filesystem_info(&self, mount_point: &str) -> Result<(u64, u64), CollectorError> {
|
||||||
let output = StdCommand::new("timeout")
|
let output = StdCommand::new("timeout")
|
||||||
.args(&["2", "df", "--block-size=1", mount_point])
|
.args(&["10", "df", "--block-size=1", mount_point])
|
||||||
.output()
|
.output()
|
||||||
.map_err(|e| CollectorError::SystemRead {
|
.map_err(|e| CollectorError::SystemRead {
|
||||||
path: format!("df {}", mount_point),
|
path: format!("df {}", mount_point),
|
||||||
@ -433,7 +433,7 @@ impl DiskCollector {
|
|||||||
cmd.args(&["-a", &format!("/dev/{}", drive_name)]);
|
cmd.args(&["-a", &format!("/dev/{}", drive_name)]);
|
||||||
}
|
}
|
||||||
|
|
||||||
let output = run_command_with_timeout(cmd, 3).await
|
let output = run_command_with_timeout(cmd, 15).await
|
||||||
.map_err(|e| CollectorError::SystemRead {
|
.map_err(|e| CollectorError::SystemRead {
|
||||||
path: format!("SMART data for {}", drive_name),
|
path: format!("SMART data for {}", drive_name),
|
||||||
error: e.to_string(),
|
error: e.to_string(),
|
||||||
@ -772,7 +772,7 @@ impl DiskCollector {
|
|||||||
fn get_drive_info_for_path(&self, path: &str) -> anyhow::Result<PoolDrive> {
|
fn get_drive_info_for_path(&self, path: &str) -> anyhow::Result<PoolDrive> {
|
||||||
// Use lsblk to find the backing device with timeout
|
// Use lsblk to find the backing device with timeout
|
||||||
let output = StdCommand::new("timeout")
|
let output = StdCommand::new("timeout")
|
||||||
.args(&["2", "lsblk", "-rn", "-o", "NAME,MOUNTPOINT"])
|
.args(&["10", "lsblk", "-rn", "-o", "NAME,MOUNTPOINT"])
|
||||||
.output()
|
.output()
|
||||||
.map_err(|e| anyhow::anyhow!("Failed to run lsblk: {}", e))?;
|
.map_err(|e| anyhow::anyhow!("Failed to run lsblk: {}", e))?;
|
||||||
|
|
||||||
|
|||||||
@ -217,14 +217,47 @@ impl SystemdCollector {
|
|||||||
}
|
}
|
||||||
|
|
||||||
if service_name == "tailscaled" && status_info.active_state == "active" {
|
if service_name == "tailscaled" && status_info.active_state == "active" {
|
||||||
// Add Tailscale connection method as sub-service
|
// Add Tailscale peers with their connection methods as sub-services
|
||||||
if let Some(conn_method) = self.get_tailscale_connection_method() {
|
let peers = self.get_tailscale_peers();
|
||||||
|
for (peer_name, conn_method) in peers {
|
||||||
let metrics = Vec::new();
|
let metrics = Vec::new();
|
||||||
sub_services.push(SubServiceData {
|
sub_services.push(SubServiceData {
|
||||||
name: format!("Connection: {}", conn_method),
|
name: format!("{}: {}", peer_name, conn_method),
|
||||||
service_status: Status::Info,
|
service_status: Status::Info,
|
||||||
metrics,
|
metrics,
|
||||||
service_type: "tailscale_connection".to_string(),
|
service_type: "tailscale_peer".to_string(),
|
||||||
|
});
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if service_name == "nfs-server" && status_info.active_state == "active" {
|
||||||
|
// Add NFS exports as sub-services
|
||||||
|
let exports = self.get_nfs_exports();
|
||||||
|
for (export_path, info) in exports {
|
||||||
|
let display = if !info.is_empty() {
|
||||||
|
format!("{} {}", export_path, info)
|
||||||
|
} else {
|
||||||
|
export_path
|
||||||
|
};
|
||||||
|
sub_services.push(SubServiceData {
|
||||||
|
name: display,
|
||||||
|
service_status: Status::Info,
|
||||||
|
metrics: Vec::new(),
|
||||||
|
service_type: "nfs_export".to_string(),
|
||||||
|
});
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if (service_name == "smbd" || service_name == "samba-smbd") && status_info.active_state == "active" {
|
||||||
|
// Add SMB shares as sub-services
|
||||||
|
let shares = self.get_smb_shares();
|
||||||
|
for (share_name, share_path) in shares {
|
||||||
|
let metrics = Vec::new();
|
||||||
|
sub_services.push(SubServiceData {
|
||||||
|
name: format!("{}: {}", share_name, share_path),
|
||||||
|
service_status: Status::Info,
|
||||||
|
metrics,
|
||||||
|
service_type: "smb_share".to_string(),
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -936,50 +969,219 @@ impl SystemdCollector {
|
|||||||
None
|
None
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Get Tailscale connection method (direct, relay, or proxy)
|
/// Get Tailscale connected peers with their connection methods
|
||||||
fn get_tailscale_connection_method(&self) -> Option<String> {
|
/// Returns a list of (device_name, connection_method) tuples
|
||||||
|
fn get_tailscale_peers(&self) -> Vec<(String, String)> {
|
||||||
match Command::new("timeout")
|
match Command::new("timeout")
|
||||||
.args(["2", "tailscale", "status", "--json"])
|
.args(["2", "tailscale", "status"])
|
||||||
.output()
|
.output()
|
||||||
{
|
{
|
||||||
Ok(output) if output.status.success() => {
|
Ok(output) if output.status.success() => {
|
||||||
let json_str = String::from_utf8_lossy(&output.stdout);
|
let status_output = String::from_utf8_lossy(&output.stdout);
|
||||||
|
let mut peers = Vec::new();
|
||||||
|
|
||||||
if let Ok(json_data) = serde_json::from_str::<serde_json::Value>(&json_str) {
|
// Get current hostname to filter it out
|
||||||
// Look for the self peer (current node) in the peer list
|
let current_hostname = gethostname::gethostname()
|
||||||
if let Some(peers) = json_data["Peer"].as_object() {
|
.to_string_lossy()
|
||||||
// Find the first active peer connection to determine connection method
|
.to_string();
|
||||||
for (_peer_id, peer_data) in peers {
|
|
||||||
if peer_data["Active"].as_bool().unwrap_or(false) {
|
|
||||||
// Check if using relay
|
|
||||||
let relay_node = peer_data["Relay"].as_str().unwrap_or("");
|
|
||||||
if !relay_node.is_empty() {
|
|
||||||
return Some("relay".to_string());
|
|
||||||
}
|
|
||||||
|
|
||||||
// Check if using direct connection
|
// Parse tailscale status output
|
||||||
if let Some(endpoints) = peer_data["CurAddr"].as_str() {
|
// Format: IP hostname user os status
|
||||||
if !endpoints.is_empty() {
|
// Example: 100.110.98.3 wslbox cm@ linux active; direct 192.168.30.227:53757
|
||||||
return Some("direct".to_string());
|
// Note: First line is always the current host, skip it
|
||||||
}
|
for (idx, line) in status_output.lines().enumerate() {
|
||||||
}
|
if idx == 0 {
|
||||||
}
|
continue; // Skip first line (current host)
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// Check if using proxy from backend state
|
let parts: Vec<&str> = line.split_whitespace().collect();
|
||||||
if let Some(backend_state) = json_data["BackendState"].as_str() {
|
if parts.len() < 5 {
|
||||||
if backend_state == "Running" {
|
continue; // Skip invalid lines
|
||||||
// If we're running but have no direct or relay, might be proxy
|
}
|
||||||
// This is a fallback heuristic
|
|
||||||
return Some("unknown".to_string());
|
// parts[0] = IP
|
||||||
|
// parts[1] = hostname
|
||||||
|
// parts[2] = user
|
||||||
|
// parts[3] = OS
|
||||||
|
// parts[4+] = status (e.g., "active;", "direct", "192.168.30.227:53757" or "idle;" or "offline")
|
||||||
|
|
||||||
|
let hostname = parts[1];
|
||||||
|
|
||||||
|
// Skip if this is the current host (double-check in case format changes)
|
||||||
|
if hostname == current_hostname {
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
|
let status_parts = &parts[4..];
|
||||||
|
|
||||||
|
// Determine connection method from status
|
||||||
|
let connection_method = if status_parts.is_empty() {
|
||||||
|
continue; // Skip if no status
|
||||||
|
} else {
|
||||||
|
let status_str = status_parts.join(" ");
|
||||||
|
if status_str.contains("offline") {
|
||||||
|
continue; // Skip offline peers
|
||||||
|
} else if status_str.contains("direct") {
|
||||||
|
"direct"
|
||||||
|
} else if status_str.contains("relay") {
|
||||||
|
"relay"
|
||||||
|
} else if status_str.contains("idle") {
|
||||||
|
"idle"
|
||||||
|
} else if status_str.contains("active") {
|
||||||
|
"active"
|
||||||
|
} else {
|
||||||
|
continue; // Skip unknown status
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
peers.push((hostname.to_string(), connection_method.to_string()));
|
||||||
|
}
|
||||||
|
|
||||||
|
peers
|
||||||
|
}
|
||||||
|
_ => Vec::new(),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Get NFS exports from exportfs
|
||||||
|
/// Returns a list of (export_path, info_string) tuples
|
||||||
|
fn get_nfs_exports(&self) -> Vec<(String, String)> {
|
||||||
|
let output = match Command::new("timeout")
|
||||||
|
.args(["2", "exportfs", "-v"])
|
||||||
|
.output()
|
||||||
|
{
|
||||||
|
Ok(output) if output.status.success() => output,
|
||||||
|
_ => return Vec::new(),
|
||||||
|
};
|
||||||
|
|
||||||
|
let exports_output = String::from_utf8_lossy(&output.stdout);
|
||||||
|
let mut exports_map: std::collections::HashMap<String, Vec<(String, String)>> =
|
||||||
|
std::collections::HashMap::new();
|
||||||
|
let mut current_path: Option<String> = None;
|
||||||
|
|
||||||
|
for line in exports_output.lines() {
|
||||||
|
let trimmed = line.trim();
|
||||||
|
|
||||||
|
if trimmed.is_empty() || trimmed.starts_with('#') {
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
|
if trimmed.starts_with('/') {
|
||||||
|
// Export path line - may have network on same line or continuation
|
||||||
|
let parts: Vec<&str> = trimmed.splitn(2, char::is_whitespace).collect();
|
||||||
|
let path = parts[0].to_string();
|
||||||
|
current_path = Some(path.clone());
|
||||||
|
|
||||||
|
// Check if network info is on the same line
|
||||||
|
if parts.len() > 1 {
|
||||||
|
let rest = parts[1].trim();
|
||||||
|
if let Some(paren_pos) = rest.find('(') {
|
||||||
|
let network = rest[..paren_pos].trim();
|
||||||
|
|
||||||
|
if let Some(end_paren) = rest.find(')') {
|
||||||
|
let options = &rest[paren_pos+1..end_paren];
|
||||||
|
let mode = if options.contains(",rw,") || options.ends_with(",rw") {
|
||||||
|
"rw"
|
||||||
|
} else {
|
||||||
|
"ro"
|
||||||
|
};
|
||||||
|
|
||||||
|
exports_map.entry(path)
|
||||||
|
.or_insert_with(Vec::new)
|
||||||
|
.push((network.to_string(), mode.to_string()));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
} else if let Some(ref path) = current_path {
|
||||||
|
// Continuation line with network and options
|
||||||
|
if let Some(paren_pos) = trimmed.find('(') {
|
||||||
|
let network = trimmed[..paren_pos].trim();
|
||||||
|
|
||||||
|
if let Some(end_paren) = trimmed.find(')') {
|
||||||
|
let options = &trimmed[paren_pos+1..end_paren];
|
||||||
|
let mode = if options.contains(",rw,") || options.ends_with(",rw") {
|
||||||
|
"rw"
|
||||||
|
} else {
|
||||||
|
"ro"
|
||||||
|
};
|
||||||
|
|
||||||
|
exports_map.entry(path.clone())
|
||||||
|
.or_insert_with(Vec::new)
|
||||||
|
.push((network.to_string(), mode.to_string()));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Build display strings: "path: mode [networks]"
|
||||||
|
let mut exports: Vec<(String, String)> = exports_map
|
||||||
|
.into_iter()
|
||||||
|
.map(|(path, mut entries)| {
|
||||||
|
if entries.is_empty() {
|
||||||
|
return (path, String::new());
|
||||||
|
}
|
||||||
|
|
||||||
|
let mode = entries[0].1.clone();
|
||||||
|
let networks: Vec<String> = entries.drain(..).map(|(n, _)| n).collect();
|
||||||
|
let info = format!("{} [{}]", mode, networks.join(", "));
|
||||||
|
(path, info)
|
||||||
|
})
|
||||||
|
.collect();
|
||||||
|
|
||||||
|
exports.sort_by(|a, b| a.0.cmp(&b.0));
|
||||||
|
exports
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Get SMB shares from smb.conf
|
||||||
|
/// Returns a list of (share_name, share_path) tuples
|
||||||
|
fn get_smb_shares(&self) -> Vec<(String, String)> {
|
||||||
|
match std::fs::read_to_string("/etc/samba/smb.conf") {
|
||||||
|
Ok(config) => {
|
||||||
|
let mut shares = Vec::new();
|
||||||
|
let mut current_share: Option<String> = None;
|
||||||
|
let mut current_path: Option<String> = None;
|
||||||
|
|
||||||
|
for line in config.lines() {
|
||||||
|
let line = line.trim();
|
||||||
|
|
||||||
|
// Skip comments and empty lines
|
||||||
|
if line.is_empty() || line.starts_with('#') || line.starts_with(';') {
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
|
// Detect share section [sharename]
|
||||||
|
if line.starts_with('[') && line.ends_with(']') {
|
||||||
|
// Save previous share if we have both name and path
|
||||||
|
if let (Some(name), Some(path)) = (current_share.take(), current_path.take()) {
|
||||||
|
// Skip special sections
|
||||||
|
if name != "global" && name != "homes" && name != "printers" {
|
||||||
|
shares.push((name, path));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Start new share
|
||||||
|
let share_name = line[1..line.len()-1].trim().to_string();
|
||||||
|
current_share = Some(share_name);
|
||||||
|
current_path = None;
|
||||||
|
}
|
||||||
|
// Look for path = /some/path
|
||||||
|
else if line.starts_with("path") && line.contains('=') {
|
||||||
|
if let Some(path_value) = line.split('=').nth(1) {
|
||||||
|
current_path = Some(path_value.trim().to_string());
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
None
|
// Don't forget the last share
|
||||||
|
if let (Some(name), Some(path)) = (current_share, current_path) {
|
||||||
|
if name != "global" && name != "homes" && name != "printers" {
|
||||||
|
shares.push((name, path));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
shares
|
||||||
}
|
}
|
||||||
_ => None,
|
_ => Vec::new(),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@ -1,6 +1,6 @@
|
|||||||
[package]
|
[package]
|
||||||
name = "cm-dashboard"
|
name = "cm-dashboard"
|
||||||
version = "0.1.262"
|
version = "0.1.275"
|
||||||
edition = "2021"
|
edition = "2021"
|
||||||
|
|
||||||
[dependencies]
|
[dependencies]
|
||||||
|
|||||||
@ -138,11 +138,12 @@ impl Dashboard {
|
|||||||
let metrics_check_interval = Duration::from_millis(100); // Check for metrics every 100ms
|
let metrics_check_interval = Duration::from_millis(100); // Check for metrics every 100ms
|
||||||
let mut last_heartbeat_check = Instant::now();
|
let mut last_heartbeat_check = Instant::now();
|
||||||
let heartbeat_check_interval = Duration::from_secs(1); // Check for host connectivity every 1 second
|
let heartbeat_check_interval = Duration::from_secs(1); // Check for host connectivity every 1 second
|
||||||
|
let mut needs_render = true; // Track if we need to render
|
||||||
|
|
||||||
loop {
|
loop {
|
||||||
// Handle terminal events (keyboard and mouse input) only if not headless
|
// Handle terminal events (keyboard and mouse input) only if not headless
|
||||||
if !self.headless {
|
if !self.headless {
|
||||||
match event::poll(Duration::from_millis(50)) {
|
match event::poll(Duration::from_millis(200)) {
|
||||||
Ok(true) => {
|
Ok(true) => {
|
||||||
match event::read() {
|
match event::read() {
|
||||||
Ok(event) => {
|
Ok(event) => {
|
||||||
@ -152,6 +153,7 @@ impl Dashboard {
|
|||||||
// Handle keyboard input
|
// Handle keyboard input
|
||||||
match tui_app.handle_input(event) {
|
match tui_app.handle_input(event) {
|
||||||
Ok(_) => {
|
Ok(_) => {
|
||||||
|
needs_render = true;
|
||||||
// Check if we should quit
|
// Check if we should quit
|
||||||
if tui_app.should_quit() {
|
if tui_app.should_quit() {
|
||||||
info!("Quit requested, exiting dashboard");
|
info!("Quit requested, exiting dashboard");
|
||||||
@ -168,10 +170,11 @@ impl Dashboard {
|
|||||||
if let Err(e) = self.handle_mouse_event(mouse_event) {
|
if let Err(e) = self.handle_mouse_event(mouse_event) {
|
||||||
error!("Error handling mouse event: {}", e);
|
error!("Error handling mouse event: {}", e);
|
||||||
}
|
}
|
||||||
|
needs_render = true;
|
||||||
}
|
}
|
||||||
Event::Resize(_width, _height) => {
|
Event::Resize(_width, _height) => {
|
||||||
// Terminal was resized - just continue and re-render
|
// Terminal was resized - mark for re-render
|
||||||
// The next render will automatically use the new size
|
needs_render = true;
|
||||||
}
|
}
|
||||||
_ => {}
|
_ => {}
|
||||||
}
|
}
|
||||||
@ -189,38 +192,6 @@ impl Dashboard {
|
|||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Render UI immediately after handling input for responsive feedback
|
|
||||||
if let Some(ref mut terminal) = self.terminal {
|
|
||||||
if let Some(ref mut tui_app) = self.tui_app {
|
|
||||||
// Clear and autoresize terminal to handle any resize events
|
|
||||||
if let Err(e) = terminal.autoresize() {
|
|
||||||
warn!("Error autoresizing terminal: {}", e);
|
|
||||||
}
|
|
||||||
|
|
||||||
// Check minimum terminal size to prevent panics
|
|
||||||
let size = terminal.size().unwrap_or_default();
|
|
||||||
if size.width < 90 || size.height < 15 {
|
|
||||||
// Terminal too small, show error message
|
|
||||||
let msg_text = format!("Terminal too small\n\nMinimum: 90x15\nCurrent: {}x{}", size.width, size.height);
|
|
||||||
let _ = terminal.draw(|frame| {
|
|
||||||
use ratatui::widgets::{Paragraph, Block, Borders};
|
|
||||||
use ratatui::layout::Alignment;
|
|
||||||
let msg = Paragraph::new(msg_text.clone())
|
|
||||||
.alignment(Alignment::Center)
|
|
||||||
.block(Block::default().borders(Borders::ALL));
|
|
||||||
frame.render_widget(msg, frame.size());
|
|
||||||
});
|
|
||||||
} else if let Err(e) = terminal.draw(|frame| {
|
|
||||||
let (title_area, system_area, services_area) = tui_app.render(frame, &self.metric_store);
|
|
||||||
self.title_area = title_area;
|
|
||||||
self.system_area = system_area;
|
|
||||||
self.services_area = services_area;
|
|
||||||
}) {
|
|
||||||
error!("Error rendering TUI after input: {}", e);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// Check for new metrics
|
// Check for new metrics
|
||||||
@ -259,8 +230,10 @@ impl Dashboard {
|
|||||||
if let Some(ref mut tui_app) = self.tui_app {
|
if let Some(ref mut tui_app) = self.tui_app {
|
||||||
tui_app.update_metrics(&mut self.metric_store);
|
tui_app.update_metrics(&mut self.metric_store);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
needs_render = true; // New metrics received, need to render
|
||||||
}
|
}
|
||||||
|
|
||||||
// Also check for command output messages
|
// Also check for command output messages
|
||||||
if let Ok(Some(cmd_output)) = self.zmq_consumer.receive_command_output().await {
|
if let Ok(Some(cmd_output)) = self.zmq_consumer.receive_command_output().await {
|
||||||
debug!(
|
debug!(
|
||||||
@ -271,26 +244,27 @@ impl Dashboard {
|
|||||||
|
|
||||||
// Command output (terminal popup removed - output not displayed)
|
// Command output (terminal popup removed - output not displayed)
|
||||||
}
|
}
|
||||||
|
|
||||||
last_metrics_check = Instant::now();
|
last_metrics_check = Instant::now();
|
||||||
}
|
}
|
||||||
|
|
||||||
// Check for host connectivity changes (heartbeat timeouts) periodically
|
// Check for host connectivity changes (heartbeat timeouts) periodically
|
||||||
if last_heartbeat_check.elapsed() >= heartbeat_check_interval {
|
if last_heartbeat_check.elapsed() >= heartbeat_check_interval {
|
||||||
let timeout = Duration::from_secs(self.config.zmq.heartbeat_timeout_seconds);
|
let timeout = Duration::from_secs(self.config.zmq.heartbeat_timeout_seconds);
|
||||||
|
|
||||||
// Clean up metrics for offline hosts
|
// Clean up metrics for offline hosts
|
||||||
self.metric_store.cleanup_offline_hosts(timeout);
|
self.metric_store.cleanup_offline_hosts(timeout);
|
||||||
|
|
||||||
if let Some(ref mut tui_app) = self.tui_app {
|
if let Some(ref mut tui_app) = self.tui_app {
|
||||||
let connected_hosts = self.metric_store.get_connected_hosts(timeout);
|
let connected_hosts = self.metric_store.get_connected_hosts(timeout);
|
||||||
tui_app.update_hosts(connected_hosts);
|
tui_app.update_hosts(connected_hosts);
|
||||||
}
|
}
|
||||||
last_heartbeat_check = Instant::now();
|
last_heartbeat_check = Instant::now();
|
||||||
|
needs_render = true; // Heartbeat check happened, may have changed hosts
|
||||||
}
|
}
|
||||||
|
|
||||||
// Render TUI (only if not headless)
|
// Render TUI only when needed (not headless and something changed)
|
||||||
if !self.headless {
|
if !self.headless && needs_render {
|
||||||
if let Some(ref mut terminal) = self.terminal {
|
if let Some(ref mut terminal) = self.terminal {
|
||||||
if let Some(ref mut tui_app) = self.tui_app {
|
if let Some(ref mut tui_app) = self.tui_app {
|
||||||
// Clear and autoresize terminal to handle any resize events
|
// Clear and autoresize terminal to handle any resize events
|
||||||
@ -322,10 +296,8 @@ impl Dashboard {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
needs_render = false; // Reset flag after rendering
|
||||||
}
|
}
|
||||||
|
|
||||||
// Small sleep to prevent excessive CPU usage
|
|
||||||
tokio::time::sleep(Duration::from_millis(10)).await;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
info!("Dashboard main loop ended");
|
info!("Dashboard main loop ended");
|
||||||
|
|||||||
@ -44,9 +44,9 @@ pub struct SystemWidget {
|
|||||||
storage_pools: Vec<StoragePool>,
|
storage_pools: Vec<StoragePool>,
|
||||||
|
|
||||||
// Backup metrics
|
// Backup metrics
|
||||||
backup_repositories: Vec<String>,
|
backup_last_time: Option<String>,
|
||||||
backup_repository_status: Status,
|
backup_status: Status,
|
||||||
backup_disks: Vec<cm_dashboard_shared::BackupDiskData>,
|
backup_repositories: Vec<cm_dashboard_shared::BackupRepositoryData>,
|
||||||
|
|
||||||
// Overall status
|
// Overall status
|
||||||
has_data: bool,
|
has_data: bool,
|
||||||
@ -112,9 +112,9 @@ impl SystemWidget {
|
|||||||
tmp_status: Status::Unknown,
|
tmp_status: Status::Unknown,
|
||||||
tmpfs_mounts: Vec::new(),
|
tmpfs_mounts: Vec::new(),
|
||||||
storage_pools: Vec::new(),
|
storage_pools: Vec::new(),
|
||||||
|
backup_last_time: None,
|
||||||
|
backup_status: Status::Unknown,
|
||||||
backup_repositories: Vec::new(),
|
backup_repositories: Vec::new(),
|
||||||
backup_repository_status: Status::Unknown,
|
|
||||||
backup_disks: Vec::new(),
|
|
||||||
has_data: false,
|
has_data: false,
|
||||||
scroll_offset: 0,
|
scroll_offset: 0,
|
||||||
last_viewport_height: 0,
|
last_viewport_height: 0,
|
||||||
@ -221,9 +221,9 @@ impl Widget for SystemWidget {
|
|||||||
|
|
||||||
// Extract backup data
|
// Extract backup data
|
||||||
let backup = &agent_data.backup;
|
let backup = &agent_data.backup;
|
||||||
|
self.backup_last_time = backup.last_backup_time.clone();
|
||||||
|
self.backup_status = backup.backup_status;
|
||||||
self.backup_repositories = backup.repositories.clone();
|
self.backup_repositories = backup.repositories.clone();
|
||||||
self.backup_repository_status = backup.repository_status;
|
|
||||||
self.backup_disks = backup.disks.clone();
|
|
||||||
|
|
||||||
// Clamp scroll offset to valid range after update
|
// Clamp scroll offset to valid range after update
|
||||||
// This prevents scroll issues when switching between hosts
|
// This prevents scroll issues when switching between hosts
|
||||||
@ -533,79 +533,42 @@ impl SystemWidget {
|
|||||||
fn render_backup(&self) -> Vec<Line<'_>> {
|
fn render_backup(&self) -> Vec<Line<'_>> {
|
||||||
let mut lines = Vec::new();
|
let mut lines = Vec::new();
|
||||||
|
|
||||||
// First section: Repository status and list
|
if self.backup_repositories.is_empty() {
|
||||||
if !self.backup_repositories.is_empty() {
|
return lines;
|
||||||
let repo_text = format!("Repo: {}", self.backup_repositories.len());
|
|
||||||
let repo_spans = StatusIcons::create_status_spans(self.backup_repository_status, &repo_text);
|
|
||||||
lines.push(Line::from(repo_spans));
|
|
||||||
|
|
||||||
// List all repositories (sorted for consistent display)
|
|
||||||
let mut sorted_repos = self.backup_repositories.clone();
|
|
||||||
sorted_repos.sort();
|
|
||||||
let repo_count = sorted_repos.len();
|
|
||||||
for (idx, repo) in sorted_repos.iter().enumerate() {
|
|
||||||
let tree_char = if idx == repo_count - 1 { "└─" } else { "├─" };
|
|
||||||
lines.push(Line::from(vec![
|
|
||||||
Span::styled(format!(" {} ", tree_char), Typography::tree()),
|
|
||||||
Span::styled(repo.clone(), Typography::secondary()),
|
|
||||||
]));
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// Second section: Per-disk backup information (sorted by serial for consistent display)
|
// Format backup time (use complete timestamp)
|
||||||
let mut sorted_disks = self.backup_disks.clone();
|
let time_display = if let Some(ref time_str) = self.backup_last_time {
|
||||||
sorted_disks.sort_by(|a, b| a.serial.cmp(&b.serial));
|
time_str.clone()
|
||||||
for disk in &sorted_disks {
|
} else {
|
||||||
let truncated_serial = truncate_serial(&disk.serial);
|
"unknown".to_string()
|
||||||
let mut details = Vec::new();
|
};
|
||||||
|
|
||||||
if let Some(temp) = disk.temperature_celsius {
|
// Header: just the timestamp
|
||||||
details.push(format!("T: {}°C", temp as i32));
|
let repo_spans = StatusIcons::create_status_spans(self.backup_status, &time_display);
|
||||||
}
|
lines.push(Line::from(repo_spans));
|
||||||
if let Some(wear) = disk.wear_percent {
|
|
||||||
details.push(format!("W: {}%", wear as i32));
|
|
||||||
}
|
|
||||||
|
|
||||||
let disk_text = if !details.is_empty() {
|
// List all repositories with archive count and size
|
||||||
format!("{} {}", truncated_serial, details.join(" "))
|
let repo_count = self.backup_repositories.len();
|
||||||
|
for (idx, repo) in self.backup_repositories.iter().enumerate() {
|
||||||
|
let tree_char = if idx == repo_count - 1 { "└─" } else { "├─" };
|
||||||
|
|
||||||
|
// Format size: use kB for < 1MB, MB for < 1GB, otherwise GB
|
||||||
|
let size_display = if repo.repo_size_gb < 0.001 {
|
||||||
|
format!("{:.0}kB", repo.repo_size_gb * 1024.0 * 1024.0)
|
||||||
|
} else if repo.repo_size_gb < 1.0 {
|
||||||
|
format!("{:.0}MB", repo.repo_size_gb * 1024.0)
|
||||||
} else {
|
} else {
|
||||||
truncated_serial
|
format!("{:.1}GB", repo.repo_size_gb)
|
||||||
};
|
};
|
||||||
|
|
||||||
// Overall disk status (worst of backup and usage)
|
let repo_text = format!("{} ({}) {}", repo.name, repo.archive_count, size_display);
|
||||||
let disk_status = disk.backup_status.max(disk.usage_status);
|
|
||||||
let disk_spans = StatusIcons::create_status_spans(disk_status, &disk_text);
|
|
||||||
lines.push(Line::from(disk_spans));
|
|
||||||
|
|
||||||
// Show backup time with status
|
let mut repo_spans = vec![
|
||||||
if let Some(backup_time) = &disk.last_backup_time {
|
Span::styled(format!(" {} ", tree_char), Typography::tree()),
|
||||||
let time_text = format!("Backup: {}", backup_time);
|
|
||||||
let mut time_spans = vec![
|
|
||||||
Span::styled(" ├─ ", Typography::tree()),
|
|
||||||
];
|
|
||||||
time_spans.extend(StatusIcons::create_status_spans(disk.backup_status, &time_text));
|
|
||||||
lines.push(Line::from(time_spans));
|
|
||||||
}
|
|
||||||
|
|
||||||
// Show usage with status and archive count
|
|
||||||
let archive_display = if disk.archives_min == disk.archives_max {
|
|
||||||
format!("{}", disk.archives_min)
|
|
||||||
} else {
|
|
||||||
format!("{}-{}", disk.archives_min, disk.archives_max)
|
|
||||||
};
|
|
||||||
|
|
||||||
let usage_text = format!(
|
|
||||||
"Usage: ({}) {:.0}% {:.0}GB/{:.0}GB",
|
|
||||||
archive_display,
|
|
||||||
disk.disk_usage_percent,
|
|
||||||
disk.disk_used_gb,
|
|
||||||
disk.disk_total_gb
|
|
||||||
);
|
|
||||||
let mut usage_spans = vec![
|
|
||||||
Span::styled(" └─ ", Typography::tree()),
|
|
||||||
];
|
];
|
||||||
usage_spans.extend(StatusIcons::create_status_spans(disk.usage_status, &usage_text));
|
repo_spans.extend(StatusIcons::create_status_spans(repo.status, &repo_text));
|
||||||
lines.push(Line::from(usage_spans));
|
lines.push(Line::from(repo_spans));
|
||||||
}
|
}
|
||||||
|
|
||||||
lines
|
lines
|
||||||
@ -876,13 +839,10 @@ impl SystemWidget {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// Backup section
|
// Backup section
|
||||||
if !self.backup_repositories.is_empty() || !self.backup_disks.is_empty() {
|
if !self.backup_repositories.is_empty() {
|
||||||
count += 1; // Header
|
count += 1; // Header: "Backup:"
|
||||||
if !self.backup_repositories.is_empty() {
|
count += 1; // Repo count and timestamp header
|
||||||
count += 1; // Repo header
|
count += self.backup_repositories.len(); // Individual repos
|
||||||
count += self.backup_repositories.len();
|
|
||||||
}
|
|
||||||
count += self.backup_disks.len() * 3; // Each disk has 3 lines
|
|
||||||
}
|
}
|
||||||
|
|
||||||
count
|
count
|
||||||
@ -988,7 +948,7 @@ impl SystemWidget {
|
|||||||
lines.extend(storage_lines);
|
lines.extend(storage_lines);
|
||||||
|
|
||||||
// Backup section (if available)
|
// Backup section (if available)
|
||||||
if !self.backup_repositories.is_empty() || !self.backup_disks.is_empty() {
|
if !self.backup_repositories.is_empty() {
|
||||||
lines.push(Line::from(vec![
|
lines.push(Line::from(vec![
|
||||||
Span::styled("Backup:", Typography::widget_title())
|
Span::styled("Backup:", Typography::widget_title())
|
||||||
]));
|
]));
|
||||||
|
|||||||
@ -1,6 +1,6 @@
|
|||||||
[package]
|
[package]
|
||||||
name = "cm-dashboard-shared"
|
name = "cm-dashboard-shared"
|
||||||
version = "0.1.262"
|
version = "0.1.275"
|
||||||
edition = "2021"
|
edition = "2021"
|
||||||
|
|
||||||
[dependencies]
|
[dependencies]
|
||||||
|
|||||||
@ -182,27 +182,18 @@ pub struct SubServiceMetric {
|
|||||||
/// Backup system data
|
/// Backup system data
|
||||||
#[derive(Debug, Clone, Serialize, Deserialize)]
|
#[derive(Debug, Clone, Serialize, Deserialize)]
|
||||||
pub struct BackupData {
|
pub struct BackupData {
|
||||||
pub repositories: Vec<String>,
|
|
||||||
pub repository_status: Status,
|
|
||||||
pub disks: Vec<BackupDiskData>,
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Backup repository disk information
|
|
||||||
#[derive(Debug, Clone, Serialize, Deserialize)]
|
|
||||||
pub struct BackupDiskData {
|
|
||||||
pub serial: String,
|
|
||||||
pub product_name: Option<String>,
|
|
||||||
pub wear_percent: Option<f32>,
|
|
||||||
pub temperature_celsius: Option<f32>,
|
|
||||||
pub last_backup_time: Option<String>,
|
pub last_backup_time: Option<String>,
|
||||||
pub backup_status: Status,
|
pub backup_status: Status,
|
||||||
pub disk_usage_percent: f32,
|
pub repositories: Vec<BackupRepositoryData>,
|
||||||
pub disk_used_gb: f32,
|
}
|
||||||
pub disk_total_gb: f32,
|
|
||||||
pub usage_status: Status,
|
/// Individual backup repository information
|
||||||
pub services: Vec<String>,
|
#[derive(Debug, Clone, Serialize, Deserialize)]
|
||||||
pub archives_min: i64,
|
pub struct BackupRepositoryData {
|
||||||
pub archives_max: i64,
|
pub name: String,
|
||||||
|
pub archive_count: i64,
|
||||||
|
pub repo_size_gb: f32,
|
||||||
|
pub status: Status,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl AgentData {
|
impl AgentData {
|
||||||
@ -245,9 +236,9 @@ impl AgentData {
|
|||||||
},
|
},
|
||||||
services: Vec::new(),
|
services: Vec::new(),
|
||||||
backup: BackupData {
|
backup: BackupData {
|
||||||
|
last_backup_time: None,
|
||||||
|
backup_status: Status::Unknown,
|
||||||
repositories: Vec::new(),
|
repositories: Vec::new(),
|
||||||
repository_status: Status::Unknown,
|
|
||||||
disks: Vec::new(),
|
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user