1051 lines
28 KiB
Go
1051 lines
28 KiB
Go
package main
|
|
|
|
import (
|
|
"bufio"
|
|
"context"
|
|
"fmt"
|
|
"io"
|
|
"log"
|
|
"net/http"
|
|
"os"
|
|
"sort"
|
|
"strconv"
|
|
"strings"
|
|
"sync"
|
|
"time"
|
|
|
|
"github.com/gdamore/tcell/v2"
|
|
"github.com/pelletier/go-toml"
|
|
"github.com/prometheus/common/expfmt"
|
|
"github.com/rivo/tview"
|
|
"github.com/shirou/gopsutil/v3/cpu"
|
|
"github.com/shirou/gopsutil/v3/mem"
|
|
"github.com/shirou/gopsutil/v3/process"
|
|
)
|
|
|
|
var (
|
|
prometheusURL string
|
|
configFilePath string // Pfad der gefundenen Konfiguration
|
|
logFilePath string // Pfad der Logdatei aus der Konfiguration
|
|
metricsEnabled bool // Neue Variable für die Aktivierung von Metriken
|
|
bindIP string // Neue Variable für die gebundene IP-Adresse
|
|
)
|
|
|
|
func init() {
|
|
configPaths := []string{
|
|
"/etc/hmac-file-server/config.toml",
|
|
"../config.toml",
|
|
"./config.toml",
|
|
}
|
|
|
|
var config *toml.Tree
|
|
var err error
|
|
|
|
// Lade die config.toml aus den definierten Pfaden
|
|
for _, path := range configPaths {
|
|
config, err = toml.LoadFile(path)
|
|
if err == nil {
|
|
configFilePath = path
|
|
log.Printf("Using config file: %s", configFilePath)
|
|
break
|
|
}
|
|
}
|
|
|
|
if err != nil {
|
|
log.Fatalf("Error loading config file: %v\nPlease create a config.toml in one of the following locations:\n%v", err, configPaths)
|
|
}
|
|
|
|
// Metricsport auslesen
|
|
portValue := config.Get("server.metricsport")
|
|
if portValue == nil {
|
|
log.Println("Warning: 'server.metricsport' is missing in the configuration, using default port 9090")
|
|
portValue = int64(9090)
|
|
}
|
|
|
|
var port int64
|
|
switch v := portValue.(type) {
|
|
case int64:
|
|
port = v
|
|
case string:
|
|
parsedPort, err := strconv.ParseInt(v, 10, 64)
|
|
if err != nil {
|
|
log.Fatalf("Error parsing 'server.metricsport' as int64: %v", err)
|
|
}
|
|
port = parsedPort
|
|
default:
|
|
log.Fatalf("Error: 'server.metricsport' is not of type int64 or string, got %T", v)
|
|
}
|
|
|
|
// Lesen von 'metricsenabled' aus der Konfiguration
|
|
metricsEnabledValue := config.Get("server.metricsenabled")
|
|
if metricsEnabledValue == nil {
|
|
log.Println("Warning: 'server.metricsenabled' ist in der Konfiguration nicht gesetzt. Standardmäßig deaktiviert.")
|
|
metricsEnabled = false
|
|
} else {
|
|
var ok bool
|
|
metricsEnabled, ok = metricsEnabledValue.(bool)
|
|
if !ok {
|
|
log.Fatalf("Konfigurationsfehler: 'server.metricsenabled' sollte ein boolescher Wert sein, aber %T wurde gefunden.", metricsEnabledValue)
|
|
}
|
|
}
|
|
|
|
// Lesen von 'bind_ip' aus der Konfiguration
|
|
bindIPValue := config.Get("server.bind_ip")
|
|
if bindIPValue == nil {
|
|
log.Println("Warning: 'server.bind_ip' ist in der Konfiguration nicht gesetzt. Standardmäßig auf 'localhost' gesetzt.")
|
|
bindIP = "localhost"
|
|
} else {
|
|
var ok bool
|
|
bindIP, ok = bindIPValue.(string)
|
|
if !ok {
|
|
log.Fatalf("Konfigurationsfehler: 'server.bind_ip' sollte ein String sein, aber %T wurde gefunden.", bindIPValue)
|
|
}
|
|
}
|
|
|
|
// Konstruktion der prometheusURL basierend auf 'bind_ip' und 'metricsport'
|
|
prometheusURL = fmt.Sprintf("http://%s:%d/metrics", bindIP, port)
|
|
log.Printf("Metrics URL gesetzt auf: %s", prometheusURL)
|
|
|
|
// Log-Datei auslesen über server.logfile
|
|
logFileValue := config.Get("server.logfile")
|
|
if logFileValue == nil {
|
|
log.Println("Warning: 'server.logfile' is missing, using default '/var/log/hmac-file-server.log'")
|
|
logFilePath = "/var/log/hmac-file-server.log"
|
|
} else {
|
|
lf, ok := logFileValue.(string)
|
|
if !ok {
|
|
log.Fatalf("Error: 'server.logfile' is not of type string, got %T", logFileValue)
|
|
}
|
|
logFilePath = lf
|
|
}
|
|
}
|
|
|
|
// Thresholds for color coding
|
|
const (
|
|
HighUsage = 80.0
|
|
MediumUsage = 50.0
|
|
)
|
|
|
|
// ProcessInfo holds information about a process
|
|
type ProcessInfo struct {
|
|
PID int32
|
|
Name string
|
|
CPUPercent float64
|
|
MemPercent float32
|
|
CommandLine string
|
|
Uptime string // Neues Feld für die Uptime
|
|
Status string // Neues Feld für den Status
|
|
ErrorCount int // Neues Feld für die Anzahl der Fehler
|
|
TotalRequests int64 // Neues Feld für die Gesamtanzahl der Anfragen
|
|
ActiveConnections int // Neues Feld für aktive Verbindungen
|
|
AverageResponseTime float64 // Neues Feld für die durchschnittliche Antwortzeit in Millisekunden
|
|
}
|
|
|
|
// Optimized metrics fetching with timeout and connection reuse
|
|
func fetchMetrics() (map[string]float64, error) {
|
|
// Create HTTP client with timeout and connection reuse
|
|
client := &http.Client{
|
|
Timeout: 5 * time.Second,
|
|
Transport: &http.Transport{
|
|
MaxIdleConns: 10,
|
|
IdleConnTimeout: 30 * time.Second,
|
|
DisableCompression: true,
|
|
},
|
|
}
|
|
|
|
resp, err := client.Get(prometheusURL)
|
|
if err != nil {
|
|
return nil, fmt.Errorf("failed to fetch metrics: %w", err)
|
|
}
|
|
defer resp.Body.Close()
|
|
|
|
// Limit response body size to prevent memory issues
|
|
limitedReader := io.LimitReader(resp.Body, 1024*1024) // 1MB limit
|
|
|
|
parser := &expfmt.TextParser{}
|
|
metricFamilies, err := parser.TextToMetricFamilies(limitedReader)
|
|
if err != nil {
|
|
return nil, fmt.Errorf("failed to parse metrics: %w", err)
|
|
}
|
|
|
|
metrics := make(map[string]float64)
|
|
|
|
// More selective metric filtering to reduce processing
|
|
relevantPrefixes := []string{
|
|
"hmac_file_server_",
|
|
"memory_usage_bytes",
|
|
"cpu_usage_percent",
|
|
"active_connections_total",
|
|
"goroutines_count",
|
|
"total_requests",
|
|
"average_response_time_ms",
|
|
}
|
|
|
|
for name, mf := range metricFamilies {
|
|
// Quick prefix check to skip irrelevant metrics
|
|
relevant := false
|
|
for _, prefix := range relevantPrefixes {
|
|
if strings.HasPrefix(name, prefix) || name == prefix {
|
|
relevant = true
|
|
break
|
|
}
|
|
}
|
|
if !relevant {
|
|
continue
|
|
}
|
|
|
|
for _, m := range mf.GetMetric() {
|
|
var value float64
|
|
if m.GetGauge() != nil {
|
|
value = m.GetGauge().GetValue()
|
|
} else if m.GetCounter() != nil {
|
|
value = m.GetCounter().GetValue()
|
|
} else if m.GetUntyped() != nil {
|
|
value = m.GetUntyped().GetValue()
|
|
} else {
|
|
continue
|
|
}
|
|
|
|
// Simplified label handling
|
|
if len(m.GetLabel()) > 0 {
|
|
labels := make([]string, 0, len(m.GetLabel()))
|
|
for _, label := range m.GetLabel() {
|
|
labels = append(labels, fmt.Sprintf("%s=\"%s\"", label.GetName(), label.GetValue()))
|
|
}
|
|
metricKey := fmt.Sprintf("%s{%s}", name, strings.Join(labels, ","))
|
|
metrics[metricKey] = value
|
|
} else {
|
|
metrics[name] = value
|
|
}
|
|
}
|
|
}
|
|
|
|
return metrics, nil
|
|
}
|
|
|
|
// Function to fetch system data
|
|
func fetchSystemData() (float64, float64, int, error) {
|
|
v, err := mem.VirtualMemory()
|
|
if err != nil {
|
|
return 0, 0, 0, fmt.Errorf("failed to fetch memory data: %w", err)
|
|
}
|
|
|
|
c, err := cpu.Percent(0, false)
|
|
if err != nil {
|
|
return 0, 0, 0, fmt.Errorf("failed to fetch CPU data: %w", err)
|
|
}
|
|
|
|
cores, err := cpu.Counts(true)
|
|
if err != nil {
|
|
return 0, 0, 0, fmt.Errorf("failed to fetch CPU cores: %w", err)
|
|
}
|
|
|
|
cpuUsage := 0.0
|
|
if len(c) > 0 {
|
|
cpuUsage = c[0]
|
|
}
|
|
|
|
return v.UsedPercent, cpuUsage, cores, nil
|
|
}
|
|
|
|
// Optimized process list fetching with better resource management
|
|
func fetchProcessList() ([]ProcessInfo, error) {
|
|
processes, err := process.Processes()
|
|
if err != nil {
|
|
return nil, fmt.Errorf("failed to fetch processes: %w", err)
|
|
}
|
|
|
|
// Pre-allocate slice with reasonable capacity
|
|
processList := make([]ProcessInfo, 0, len(processes))
|
|
var mu sync.Mutex
|
|
var wg sync.WaitGroup
|
|
|
|
// Limit concurrent goroutines to prevent resource exhaustion
|
|
sem := make(chan struct{}, 5) // Reduced from 10 to 5
|
|
timeout := time.After(10 * time.Second) // Add timeout
|
|
|
|
// Process only a subset of processes to reduce load
|
|
maxProcesses := 200
|
|
if len(processes) > maxProcesses {
|
|
processes = processes[:maxProcesses]
|
|
}
|
|
|
|
for _, p := range processes {
|
|
select {
|
|
case <-timeout:
|
|
log.Printf("Process list fetch timeout, returning partial results")
|
|
return processList, nil
|
|
default:
|
|
}
|
|
|
|
wg.Add(1)
|
|
sem <- struct{}{} // Enter semaphore
|
|
|
|
go func(p *process.Process) {
|
|
defer wg.Done()
|
|
defer func() {
|
|
<-sem // Exit semaphore
|
|
// Recover from any panics in process info fetching
|
|
if r := recover(); r != nil {
|
|
log.Printf("Process info fetch panic: %v", r)
|
|
}
|
|
}()
|
|
|
|
// Set shorter timeout for individual process operations
|
|
ctx, cancel := context.WithTimeout(context.Background(), 1*time.Second)
|
|
defer cancel()
|
|
|
|
// Use context for process operations where possible
|
|
cpuPercent, err := p.CPUPercentWithContext(ctx)
|
|
if err != nil {
|
|
return
|
|
}
|
|
|
|
memPercent, err := p.MemoryPercentWithContext(ctx)
|
|
if err != nil {
|
|
return
|
|
}
|
|
|
|
name, err := p.NameWithContext(ctx)
|
|
if err != nil {
|
|
return
|
|
}
|
|
|
|
// Skip if CPU and memory usage are both very low to reduce noise
|
|
if cpuPercent < 0.1 && memPercent < 0.1 {
|
|
return
|
|
}
|
|
|
|
// Limit command line length to prevent memory bloat
|
|
cmdline, err := p.CmdlineWithContext(ctx)
|
|
if err != nil {
|
|
cmdline = ""
|
|
}
|
|
if len(cmdline) > 100 {
|
|
cmdline = cmdline[:100] + "..."
|
|
}
|
|
|
|
info := ProcessInfo{
|
|
PID: p.Pid,
|
|
Name: name,
|
|
CPUPercent: cpuPercent,
|
|
MemPercent: memPercent,
|
|
CommandLine: cmdline,
|
|
}
|
|
|
|
mu.Lock()
|
|
processList = append(processList, info)
|
|
mu.Unlock()
|
|
}(p)
|
|
}
|
|
|
|
// Wait with timeout
|
|
done := make(chan struct{})
|
|
go func() {
|
|
wg.Wait()
|
|
close(done)
|
|
}()
|
|
|
|
select {
|
|
case <-done:
|
|
// All goroutines completed
|
|
case <-time.After(15 * time.Second):
|
|
log.Printf("Process list fetch timeout after 15 seconds, returning partial results")
|
|
}
|
|
|
|
return processList, nil
|
|
}
|
|
|
|
// Function to fetch detailed information about hmac-file-server
|
|
func fetchHmacFileServerInfo() (*ProcessInfo, error) {
|
|
processes, err := process.Processes()
|
|
if err != nil {
|
|
return nil, fmt.Errorf("failed to fetch processes: %w", err)
|
|
}
|
|
|
|
for _, p := range processes {
|
|
name, err := p.Name()
|
|
if err != nil {
|
|
continue
|
|
}
|
|
|
|
if name == "hmac-file-server" {
|
|
cpuPercent, err := p.CPUPercent()
|
|
if err != nil {
|
|
cpuPercent = 0.0
|
|
}
|
|
|
|
memPercent, err := p.MemoryPercent()
|
|
if err != nil {
|
|
memPercent = 0.0
|
|
}
|
|
|
|
cmdline, err := p.Cmdline()
|
|
if err != nil {
|
|
cmdline = ""
|
|
}
|
|
|
|
createTime, err := p.CreateTime()
|
|
if err != nil {
|
|
return nil, fmt.Errorf("failed to get process start time: %w", err)
|
|
}
|
|
uptime := time.Since(time.Unix(0, createTime*int64(time.Millisecond)))
|
|
|
|
status := "Running" // Standardstatus
|
|
|
|
// Überprüfung, ob der Prozess aktiv ist
|
|
isRunning, err := p.IsRunning()
|
|
if err != nil || !isRunning {
|
|
status = "Stopped"
|
|
}
|
|
|
|
errorCount, err := countHmacErrors()
|
|
if err != nil {
|
|
errorCount = 0
|
|
}
|
|
|
|
metrics, err := fetchMetrics()
|
|
if err != nil {
|
|
return nil, fmt.Errorf("failed to fetch metrics: %w", err)
|
|
}
|
|
|
|
totalRequests, ok := metrics["total_requests"]
|
|
if !ok {
|
|
totalRequests = 0
|
|
}
|
|
|
|
activeConnections, ok := metrics["active_connections_total"]
|
|
if !ok {
|
|
activeConnections = 0
|
|
}
|
|
|
|
averageResponseTime, ok := metrics["average_response_time_ms"]
|
|
if !ok {
|
|
averageResponseTime = 0.0
|
|
}
|
|
|
|
return &ProcessInfo{
|
|
PID: p.Pid,
|
|
Name: name,
|
|
CPUPercent: cpuPercent,
|
|
MemPercent: memPercent,
|
|
CommandLine: cmdline,
|
|
Uptime: uptime.String(),
|
|
Status: status,
|
|
ErrorCount: errorCount,
|
|
TotalRequests: int64(totalRequests),
|
|
ActiveConnections: int(activeConnections),
|
|
AverageResponseTime: averageResponseTime,
|
|
}, nil
|
|
}
|
|
}
|
|
|
|
return nil, fmt.Errorf("hmac-file-server process not found")
|
|
}
|
|
|
|
// Optimized error counting with caching and limits
|
|
var (
|
|
errorCountCache int
|
|
errorCountCacheTime time.Time
|
|
errorCountMutex sync.RWMutex
|
|
)
|
|
|
|
func countHmacErrors() (int, error) {
|
|
// Use cached value if recent (within 30 seconds)
|
|
errorCountMutex.RLock()
|
|
if time.Since(errorCountCacheTime) < 30*time.Second {
|
|
count := errorCountCache
|
|
errorCountMutex.RUnlock()
|
|
return count, nil
|
|
}
|
|
errorCountMutex.RUnlock()
|
|
|
|
// Use the configured log file path
|
|
file, err := os.Open(logFilePath)
|
|
if err != nil {
|
|
return 0, err
|
|
}
|
|
defer file.Close()
|
|
|
|
// Get file size to limit reading for very large files
|
|
stat, err := file.Stat()
|
|
if err != nil {
|
|
return 0, err
|
|
}
|
|
|
|
// Limit to last 1MB for large log files
|
|
var startPos int64 = 0
|
|
if stat.Size() > 1024*1024 {
|
|
startPos = stat.Size() - 1024*1024
|
|
file.Seek(startPos, io.SeekStart)
|
|
}
|
|
|
|
scanner := bufio.NewScanner(file)
|
|
errorCount := 0
|
|
lineCount := 0
|
|
maxLines := 1000 // Limit lines scanned
|
|
|
|
for scanner.Scan() && lineCount < maxLines {
|
|
line := scanner.Text()
|
|
if strings.Contains(line, "level=error") {
|
|
errorCount++
|
|
}
|
|
lineCount++
|
|
}
|
|
|
|
if err := scanner.Err(); err != nil {
|
|
return 0, err
|
|
}
|
|
|
|
// Update cache
|
|
errorCountMutex.Lock()
|
|
errorCountCache = errorCount
|
|
errorCountCacheTime = time.Now()
|
|
errorCountMutex.Unlock()
|
|
|
|
return errorCount, nil
|
|
}
|
|
|
|
// Optimized data structure for caching
|
|
type cachedData struct {
|
|
systemData systemData
|
|
metrics map[string]float64
|
|
processes []ProcessInfo
|
|
hmacInfo *ProcessInfo
|
|
lastUpdate time.Time
|
|
mu sync.RWMutex
|
|
}
|
|
|
|
type systemData struct {
|
|
memUsage float64
|
|
cpuUsage float64
|
|
cores int
|
|
}
|
|
|
|
var cache = &cachedData{}
|
|
|
|
// Optimized updateUI with reduced frequency and better resource management
|
|
func updateUI(ctx context.Context, app *tview.Application, pages *tview.Pages, sysPage, hmacPage tview.Primitive) {
|
|
// Reduce update frequency significantly
|
|
fastTicker := time.NewTicker(5 * time.Second) // UI updates
|
|
slowTicker := time.NewTicker(15 * time.Second) // Process list updates
|
|
defer fastTicker.Stop()
|
|
defer slowTicker.Stop()
|
|
|
|
// Worker pool to limit concurrent operations
|
|
workerPool := make(chan struct{}, 3) // Max 3 concurrent operations
|
|
|
|
// Single goroutine for data collection
|
|
go func() {
|
|
defer func() {
|
|
if r := recover(); r != nil {
|
|
log.Printf("Data collection goroutine recovered from panic: %v", r)
|
|
}
|
|
}()
|
|
|
|
for {
|
|
select {
|
|
case <-ctx.Done():
|
|
return
|
|
case <-fastTicker.C:
|
|
// Only update system data and metrics (lightweight operations)
|
|
select {
|
|
case workerPool <- struct{}{}:
|
|
go func() {
|
|
defer func() { <-workerPool }()
|
|
updateSystemAndMetrics()
|
|
}()
|
|
default:
|
|
// Skip if worker pool is full
|
|
}
|
|
case <-slowTicker.C:
|
|
// Update process list less frequently (expensive operation)
|
|
select {
|
|
case workerPool <- struct{}{}:
|
|
go func() {
|
|
defer func() { <-workerPool }()
|
|
updateProcessData()
|
|
}()
|
|
default:
|
|
// Skip if worker pool is full
|
|
}
|
|
}
|
|
}
|
|
}()
|
|
|
|
// UI update loop
|
|
uiTicker := time.NewTicker(2 * time.Second)
|
|
defer uiTicker.Stop()
|
|
|
|
for {
|
|
select {
|
|
case <-ctx.Done():
|
|
return
|
|
case <-uiTicker.C:
|
|
app.QueueUpdateDraw(func() {
|
|
updateUIComponents(pages, sysPage, hmacPage)
|
|
})
|
|
}
|
|
}
|
|
}
|
|
|
|
// Separate function to update system data and metrics
|
|
func updateSystemAndMetrics() {
|
|
defer func() {
|
|
if r := recover(); r != nil {
|
|
log.Printf("updateSystemAndMetrics recovered from panic: %v", r)
|
|
}
|
|
}()
|
|
|
|
// Get system data
|
|
memUsage, cpuUsage, cores, err := fetchSystemData()
|
|
if err != nil {
|
|
log.Printf("Error fetching system data: %v", err)
|
|
return
|
|
}
|
|
|
|
// Get metrics if enabled
|
|
var metrics map[string]float64
|
|
if metricsEnabled {
|
|
metrics, err = fetchMetrics()
|
|
if err != nil {
|
|
log.Printf("Error fetching metrics: %v", err)
|
|
metrics = make(map[string]float64) // Use empty map on error
|
|
}
|
|
}
|
|
|
|
// Update cache
|
|
cache.mu.Lock()
|
|
cache.systemData = systemData{memUsage, cpuUsage, cores}
|
|
cache.metrics = metrics
|
|
cache.lastUpdate = time.Now()
|
|
cache.mu.Unlock()
|
|
}
|
|
|
|
// Separate function to update process data (expensive operation)
|
|
func updateProcessData() {
|
|
defer func() {
|
|
if r := recover(); r != nil {
|
|
log.Printf("updateProcessData recovered from panic: %v", r)
|
|
}
|
|
}()
|
|
|
|
// Get process list
|
|
processes, err := fetchProcessList()
|
|
if err != nil {
|
|
log.Printf("Error fetching process list: %v", err)
|
|
return
|
|
}
|
|
|
|
// Get HMAC info
|
|
hmacInfo, err := fetchHmacFileServerInfo()
|
|
if err != nil {
|
|
log.Printf("Error fetching HMAC info: %v", err)
|
|
}
|
|
|
|
// Update cache
|
|
cache.mu.Lock()
|
|
cache.processes = processes
|
|
cache.hmacInfo = hmacInfo
|
|
cache.mu.Unlock()
|
|
}
|
|
|
|
// Update UI components with cached data
|
|
func updateUIComponents(pages *tview.Pages, sysPage, hmacPage tview.Primitive) {
|
|
currentPage, _ := pages.GetFrontPage()
|
|
|
|
cache.mu.RLock()
|
|
defer cache.mu.RUnlock()
|
|
|
|
switch currentPage {
|
|
case "system":
|
|
sysFlex := sysPage.(*tview.Flex)
|
|
|
|
// Update system table
|
|
sysTable := sysFlex.GetItem(0).(*tview.Table)
|
|
updateSystemTable(sysTable, cache.systemData.memUsage, cache.systemData.cpuUsage, cache.systemData.cores)
|
|
|
|
// Update metrics table
|
|
if metricsEnabled && len(cache.metrics) > 0 {
|
|
metricsTable := sysFlex.GetItem(1).(*tview.Table)
|
|
updateMetricsTable(metricsTable, cache.metrics)
|
|
}
|
|
|
|
// Update process table
|
|
if len(cache.processes) > 0 {
|
|
processTable := sysFlex.GetItem(2).(*tview.Table)
|
|
updateProcessTable(processTable, cache.processes)
|
|
}
|
|
|
|
case "hmac":
|
|
if cache.hmacInfo != nil {
|
|
hmacFlex := hmacPage.(*tview.Flex)
|
|
hmacTable := hmacFlex.GetItem(0).(*tview.Table)
|
|
updateHmacTable(hmacTable, cache.hmacInfo, cache.metrics)
|
|
}
|
|
}
|
|
}
|
|
|
|
// Helper function to update system data table
|
|
func updateSystemTable(sysTable *tview.Table, memUsage, cpuUsage float64, cores int) {
|
|
sysTable.Clear()
|
|
sysTable.SetCell(0, 0, tview.NewTableCell("Metric").SetAttributes(tcell.AttrBold))
|
|
sysTable.SetCell(0, 1, tview.NewTableCell("Value").SetAttributes(tcell.AttrBold))
|
|
|
|
// CPU Usage Row
|
|
cpuUsageCell := tview.NewTableCell(fmt.Sprintf("%.2f%%", cpuUsage))
|
|
if cpuUsage > HighUsage {
|
|
cpuUsageCell.SetTextColor(tcell.ColorRed)
|
|
} else if cpuUsage > MediumUsage {
|
|
cpuUsageCell.SetTextColor(tcell.ColorYellow)
|
|
} else {
|
|
cpuUsageCell.SetTextColor(tcell.ColorGreen)
|
|
}
|
|
sysTable.SetCell(1, 0, tview.NewTableCell("CPU Usage"))
|
|
sysTable.SetCell(1, 1, cpuUsageCell)
|
|
|
|
// Memory Usage Row
|
|
memUsageCell := tview.NewTableCell(fmt.Sprintf("%.2f%%", memUsage))
|
|
if memUsage > HighUsage {
|
|
memUsageCell.SetTextColor(tcell.ColorRed)
|
|
} else if memUsage > MediumUsage {
|
|
memUsageCell.SetTextColor(tcell.ColorYellow)
|
|
} else {
|
|
memUsageCell.SetTextColor(tcell.ColorGreen)
|
|
}
|
|
sysTable.SetCell(2, 0, tview.NewTableCell("Memory Usage"))
|
|
sysTable.SetCell(2, 1, memUsageCell)
|
|
|
|
// CPU Cores Row
|
|
sysTable.SetCell(3, 0, tview.NewTableCell("CPU Cores"))
|
|
sysTable.SetCell(3, 1, tview.NewTableCell(fmt.Sprintf("%d", cores)))
|
|
}
|
|
|
|
// Helper function to update metrics table
|
|
func updateMetricsTable(metricsTable *tview.Table, metrics map[string]float64) {
|
|
metricsTable.Clear()
|
|
metricsTable.SetCell(0, 0, tview.NewTableCell("Metric").SetAttributes(tcell.AttrBold))
|
|
metricsTable.SetCell(0, 1, tview.NewTableCell("Value").SetAttributes(tcell.AttrBold))
|
|
|
|
row := 1
|
|
for key, value := range metrics {
|
|
metricsTable.SetCell(row, 0, tview.NewTableCell(key))
|
|
metricsTable.SetCell(row, 1, tview.NewTableCell(fmt.Sprintf("%.2f", value)))
|
|
row++
|
|
}
|
|
}
|
|
|
|
// Helper function to update process table
|
|
func updateProcessTable(processTable *tview.Table, processes []ProcessInfo) {
|
|
processTable.Clear()
|
|
processTable.SetCell(0, 0, tview.NewTableCell("PID").SetAttributes(tcell.AttrBold))
|
|
processTable.SetCell(0, 1, tview.NewTableCell("Name").SetAttributes(tcell.AttrBold))
|
|
processTable.SetCell(0, 2, tview.NewTableCell("CPU%").SetAttributes(tcell.AttrBold))
|
|
processTable.SetCell(0, 3, tview.NewTableCell("Mem%").SetAttributes(tcell.AttrBold))
|
|
processTable.SetCell(0, 4, tview.NewTableCell("Command").SetAttributes(tcell.AttrBold))
|
|
|
|
// Sort processes by CPU usage
|
|
sort.Slice(processes, func(i, j int) bool {
|
|
return processes[i].CPUPercent > processes[j].CPUPercent
|
|
})
|
|
|
|
// Limit to top 20 processes
|
|
maxRows := 20
|
|
if len(processes) < maxRows {
|
|
maxRows = len(processes)
|
|
}
|
|
|
|
for i := 0; i < maxRows; i++ {
|
|
p := processes[i]
|
|
processTable.SetCell(i+1, 0, tview.NewTableCell(fmt.Sprintf("%d", p.PID)))
|
|
processTable.SetCell(i+1, 1, tview.NewTableCell(p.Name))
|
|
processTable.SetCell(i+1, 2, tview.NewTableCell(fmt.Sprintf("%.2f", p.CPUPercent)))
|
|
processTable.SetCell(i+1, 3, tview.NewTableCell(fmt.Sprintf("%.2f", p.MemPercent)))
|
|
processTable.SetCell(i+1, 4, tview.NewTableCell(p.CommandLine))
|
|
}
|
|
}
|
|
|
|
// Helper function to update hmac-table
|
|
func updateHmacTable(hmacTable *tview.Table, hmacInfo *ProcessInfo, metrics map[string]float64) {
|
|
hmacTable.Clear()
|
|
hmacTable.SetCell(0, 0, tview.NewTableCell("Property").SetAttributes(tcell.AttrBold))
|
|
hmacTable.SetCell(0, 1, tview.NewTableCell("Value").SetAttributes(tcell.AttrBold))
|
|
|
|
// Process information
|
|
hmacTable.SetCell(1, 0, tview.NewTableCell("PID"))
|
|
hmacTable.SetCell(1, 1, tview.NewTableCell(fmt.Sprintf("%d", hmacInfo.PID)))
|
|
|
|
hmacTable.SetCell(2, 0, tview.NewTableCell("CPU%"))
|
|
hmacTable.SetCell(2, 1, tview.NewTableCell(fmt.Sprintf("%.2f", hmacInfo.CPUPercent)))
|
|
|
|
hmacTable.SetCell(3, 0, tview.NewTableCell("Mem%"))
|
|
hmacTable.SetCell(3, 1, tview.NewTableCell(fmt.Sprintf("%.2f", hmacInfo.MemPercent)))
|
|
|
|
hmacTable.SetCell(4, 0, tview.NewTableCell("Command"))
|
|
hmacTable.SetCell(4, 1, tview.NewTableCell(hmacInfo.CommandLine))
|
|
|
|
hmacTable.SetCell(5, 0, tview.NewTableCell("Uptime"))
|
|
hmacTable.SetCell(5, 1, tview.NewTableCell(hmacInfo.Uptime)) // Neue Zeile für Uptime
|
|
|
|
hmacTable.SetCell(6, 0, tview.NewTableCell("Status"))
|
|
hmacTable.SetCell(6, 1, tview.NewTableCell(hmacInfo.Status)) // Neue Zeile für Status
|
|
|
|
hmacTable.SetCell(7, 0, tview.NewTableCell("Error Count"))
|
|
hmacTable.SetCell(7, 1, tview.NewTableCell(fmt.Sprintf("%d", hmacInfo.ErrorCount))) // Neue Zeile für Error Count
|
|
|
|
hmacTable.SetCell(8, 0, tview.NewTableCell("Total Requests"))
|
|
hmacTable.SetCell(8, 1, tview.NewTableCell(fmt.Sprintf("%d", hmacInfo.TotalRequests))) // Neue Zeile für Total Requests
|
|
|
|
hmacTable.SetCell(9, 0, tview.NewTableCell("Active Connections"))
|
|
hmacTable.SetCell(9, 1, tview.NewTableCell(fmt.Sprintf("%d", hmacInfo.ActiveConnections))) // Neue Zeile für Active Connections
|
|
|
|
hmacTable.SetCell(10, 0, tview.NewTableCell("Avg. Response Time (ms)"))
|
|
hmacTable.SetCell(10, 1, tview.NewTableCell(fmt.Sprintf("%.2f", hmacInfo.AverageResponseTime))) // Neue Zeile für Average Response Time
|
|
|
|
// Metrics related to hmac-file-server
|
|
row := 12
|
|
hmacTable.SetCell(row, 0, tview.NewTableCell("Metric").SetAttributes(tcell.AttrBold))
|
|
hmacTable.SetCell(row, 1, tview.NewTableCell("Value").SetAttributes(tcell.AttrBold))
|
|
row++
|
|
|
|
for key, value := range metrics {
|
|
if strings.Contains(key, "hmac_file_server_") {
|
|
hmacTable.SetCell(row, 0, tview.NewTableCell(key))
|
|
hmacTable.SetCell(row, 1, tview.NewTableCell(fmt.Sprintf("%.2f", value)))
|
|
row++
|
|
}
|
|
}
|
|
}
|
|
|
|
func createSystemPage() tview.Primitive {
|
|
// Create system data table
|
|
sysTable := tview.NewTable().SetBorders(false)
|
|
sysTable.SetTitle(" [::b]System Data ").SetBorder(true)
|
|
|
|
// Create Prometheus metrics table
|
|
metricsTable := tview.NewTable().SetBorders(false)
|
|
metricsTable.SetTitle(" [::b]Prometheus Metrics ").SetBorder(true)
|
|
|
|
// Create process list table
|
|
processTable := tview.NewTable().SetBorders(false)
|
|
processTable.SetTitle(" [::b]Process List ").SetBorder(true)
|
|
|
|
// Create a flex layout to hold the tables
|
|
sysFlex := tview.NewFlex().
|
|
SetDirection(tview.FlexRow).
|
|
AddItem(sysTable, 7, 0, false).
|
|
AddItem(metricsTable, 0, 1, false).
|
|
AddItem(processTable, 0, 2, false)
|
|
|
|
return sysFlex
|
|
}
|
|
|
|
func createHmacPage() tview.Primitive {
|
|
hmacTable := tview.NewTable().SetBorders(false)
|
|
hmacTable.SetTitle(" [::b]hmac-file-server Details ").SetBorder(true)
|
|
|
|
hmacFlex := tview.NewFlex().
|
|
SetDirection(tview.FlexRow).
|
|
AddItem(hmacTable, 0, 1, false)
|
|
|
|
return hmacFlex
|
|
}
|
|
|
|
func createLogsPage(ctx context.Context, app *tview.Application, logFilePath string) tview.Primitive {
|
|
logsTextView := tview.NewTextView().
|
|
SetDynamicColors(true).
|
|
SetRegions(true).
|
|
SetWordWrap(true)
|
|
logsTextView.SetTitle(" [::b]Logs ").SetBorder(true)
|
|
|
|
const numLines = 50 // Reduced from 100 to 50 lines
|
|
|
|
// Cache for log content to avoid reading file too frequently
|
|
var lastLogUpdate time.Time
|
|
var logMutex sync.RWMutex
|
|
|
|
// Read logs less frequently and only when on logs page
|
|
go func() {
|
|
ticker := time.NewTicker(5 * time.Second) // Increased from 2 to 5 seconds
|
|
defer ticker.Stop()
|
|
|
|
for {
|
|
select {
|
|
case <-ctx.Done():
|
|
return
|
|
case <-ticker.C:
|
|
// Only update if we haven't updated recently
|
|
logMutex.RLock()
|
|
timeSinceUpdate := time.Since(lastLogUpdate)
|
|
logMutex.RUnlock()
|
|
|
|
if timeSinceUpdate < 4*time.Second {
|
|
continue
|
|
}
|
|
|
|
content, err := readLastNLines(logFilePath, numLines)
|
|
if err != nil {
|
|
app.QueueUpdateDraw(func() {
|
|
logsTextView.SetText(fmt.Sprintf("[red]Error reading log file: %v[white]", err))
|
|
})
|
|
continue
|
|
}
|
|
|
|
// Process the log content with color coding
|
|
lines := strings.Split(content, "\n")
|
|
var coloredLines []string
|
|
|
|
// Limit the number of lines processed
|
|
maxLines := min(len(lines), numLines)
|
|
coloredLines = make([]string, 0, maxLines)
|
|
|
|
for i := len(lines) - maxLines; i < len(lines); i++ {
|
|
if i < 0 {
|
|
continue
|
|
}
|
|
line := lines[i]
|
|
if strings.Contains(line, "level=info") {
|
|
coloredLines = append(coloredLines, "[green]"+line+"[white]")
|
|
} else if strings.Contains(line, "level=warn") {
|
|
coloredLines = append(coloredLines, "[yellow]"+line+"[white]")
|
|
} else if strings.Contains(line, "level=error") {
|
|
coloredLines = append(coloredLines, "[red]"+line+"[white]")
|
|
} else {
|
|
coloredLines = append(coloredLines, line)
|
|
}
|
|
}
|
|
|
|
logContent := strings.Join(coloredLines, "\n")
|
|
|
|
// Update cache
|
|
logMutex.Lock()
|
|
lastLogUpdate = time.Now()
|
|
logMutex.Unlock()
|
|
|
|
app.QueueUpdateDraw(func() {
|
|
logsTextView.SetText(logContent)
|
|
})
|
|
}
|
|
}
|
|
}()
|
|
|
|
return logsTextView
|
|
}
|
|
|
|
// Helper function for min
|
|
func min(a, b int) int {
|
|
if a < b {
|
|
return a
|
|
}
|
|
return b
|
|
}
|
|
|
|
// Optimized readLastNLines to handle large files efficiently
|
|
func readLastNLines(filePath string, n int) (string, error) {
|
|
file, err := os.Open(filePath)
|
|
if err != nil {
|
|
return "", err
|
|
}
|
|
defer file.Close()
|
|
|
|
const bufferSize = 1024
|
|
buffer := make([]byte, bufferSize)
|
|
var content []byte
|
|
var fileSize int64
|
|
|
|
fileInfo, err := file.Stat()
|
|
if err != nil {
|
|
return "", err
|
|
}
|
|
fileSize = fileInfo.Size()
|
|
|
|
var offset int64 = 0
|
|
for {
|
|
if fileSize-offset < bufferSize {
|
|
offset = fileSize
|
|
} else {
|
|
offset += bufferSize
|
|
}
|
|
|
|
_, err := file.Seek(-offset, io.SeekEnd)
|
|
if err != nil {
|
|
return "", err
|
|
}
|
|
|
|
bytesRead, err := file.Read(buffer)
|
|
if err != nil && err != io.EOF {
|
|
return "", err
|
|
}
|
|
|
|
content = append(buffer[:bytesRead], content...)
|
|
|
|
if bytesRead < bufferSize || len(strings.Split(string(content), "\n")) > n+1 {
|
|
break
|
|
}
|
|
|
|
if offset >= fileSize {
|
|
break
|
|
}
|
|
}
|
|
|
|
lines := strings.Split(string(content), "\n")
|
|
if len(lines) > n {
|
|
lines = lines[len(lines)-n:]
|
|
}
|
|
return strings.Join(lines, "\n"), nil
|
|
}
|
|
|
|
func main() {
|
|
app := tview.NewApplication()
|
|
|
|
// Create a cancellable context
|
|
ctx, cancel := context.WithCancel(context.Background())
|
|
defer cancel()
|
|
|
|
// Create pages
|
|
pages := tview.NewPages()
|
|
|
|
// System page
|
|
sysPage := createSystemPage()
|
|
pages.AddPage("system", sysPage, true, true)
|
|
|
|
// hmac-file-server page
|
|
hmacPage := createHmacPage()
|
|
pages.AddPage("hmac", hmacPage, true, false)
|
|
|
|
// Logs page mit dem gelesenen logFilePath
|
|
logsPage := createLogsPage(ctx, app, logFilePath)
|
|
pages.AddPage("logs", logsPage, true, false)
|
|
|
|
// Add key binding to switch views and handle exit
|
|
app.SetInputCapture(func(event *tcell.EventKey) *tcell.EventKey {
|
|
if event.Key() == tcell.KeyRune {
|
|
switch event.Rune() {
|
|
case 'q', 'Q':
|
|
cancel()
|
|
app.Stop()
|
|
return nil
|
|
case 's', 'S':
|
|
// Switch to system page
|
|
pages.SwitchToPage("system")
|
|
case 'h', 'H':
|
|
// Switch to hmac-file-server page
|
|
pages.SwitchToPage("hmac")
|
|
case 'l', 'L':
|
|
// Switch to logs page
|
|
pages.SwitchToPage("logs")
|
|
}
|
|
}
|
|
return event
|
|
})
|
|
|
|
// Start the UI update loop in a separate goroutine
|
|
go updateUI(ctx, app, pages, sysPage, hmacPage)
|
|
|
|
// Set the root and run the application
|
|
if err := app.SetRoot(pages, true).EnableMouse(true).Run(); err != nil {
|
|
log.Fatalf("Error running application: %v", err)
|
|
log.Fatalf("Error running application: %v", err)
|
|
}
|
|
}
|