mirror of
https://github.com/1Panel-dev/1Panel.git
synced 2024-12-15 17:29:00 +08:00
217 lines
6.0 KiB
Go
217 lines
6.0 KiB
Go
|
package service
|
||
|
|
||
|
import (
|
||
|
"context"
|
||
|
"fmt"
|
||
|
"strconv"
|
||
|
"time"
|
||
|
|
||
|
"github.com/1Panel-dev/1Panel/agent/app/model"
|
||
|
"github.com/1Panel-dev/1Panel/agent/global"
|
||
|
"github.com/robfig/cron/v3"
|
||
|
"github.com/shirou/gopsutil/v3/cpu"
|
||
|
"github.com/shirou/gopsutil/v3/disk"
|
||
|
"github.com/shirou/gopsutil/v3/load"
|
||
|
"github.com/shirou/gopsutil/v3/mem"
|
||
|
"github.com/shirou/gopsutil/v3/net"
|
||
|
)
|
||
|
|
||
|
type MonitorService struct {
|
||
|
DiskIO chan ([]disk.IOCountersStat)
|
||
|
NetIO chan ([]net.IOCountersStat)
|
||
|
}
|
||
|
|
||
|
var monitorCancel context.CancelFunc
|
||
|
|
||
|
type IMonitorService interface {
|
||
|
Run()
|
||
|
|
||
|
saveIODataToDB(ctx context.Context, interval float64)
|
||
|
saveNetDataToDB(ctx context.Context, interval float64)
|
||
|
}
|
||
|
|
||
|
func NewIMonitorService() IMonitorService {
|
||
|
return &MonitorService{
|
||
|
DiskIO: make(chan []disk.IOCountersStat, 2),
|
||
|
NetIO: make(chan []net.IOCountersStat, 2),
|
||
|
}
|
||
|
}
|
||
|
|
||
|
func (m *MonitorService) Run() {
|
||
|
var itemModel model.MonitorBase
|
||
|
totalPercent, _ := cpu.Percent(3*time.Second, false)
|
||
|
if len(totalPercent) == 1 {
|
||
|
itemModel.Cpu = totalPercent[0]
|
||
|
}
|
||
|
cpuCount, _ := cpu.Counts(false)
|
||
|
|
||
|
loadInfo, _ := load.Avg()
|
||
|
itemModel.CpuLoad1 = loadInfo.Load1
|
||
|
itemModel.CpuLoad5 = loadInfo.Load5
|
||
|
itemModel.CpuLoad15 = loadInfo.Load15
|
||
|
itemModel.LoadUsage = loadInfo.Load1 / (float64(cpuCount*2) * 0.75) * 100
|
||
|
|
||
|
memoryInfo, _ := mem.VirtualMemory()
|
||
|
itemModel.Memory = memoryInfo.UsedPercent
|
||
|
|
||
|
if err := settingRepo.CreateMonitorBase(itemModel); err != nil {
|
||
|
global.LOG.Errorf("Insert basic monitoring data failed, err: %v", err)
|
||
|
}
|
||
|
|
||
|
m.loadDiskIO()
|
||
|
m.loadNetIO()
|
||
|
|
||
|
MonitorStoreDays, err := settingRepo.Get(settingRepo.WithByKey("MonitorStoreDays"))
|
||
|
if err != nil {
|
||
|
return
|
||
|
}
|
||
|
storeDays, _ := strconv.Atoi(MonitorStoreDays.Value)
|
||
|
timeForDelete := time.Now().AddDate(0, 0, -storeDays)
|
||
|
_ = settingRepo.DelMonitorBase(timeForDelete)
|
||
|
_ = settingRepo.DelMonitorIO(timeForDelete)
|
||
|
_ = settingRepo.DelMonitorNet(timeForDelete)
|
||
|
}
|
||
|
|
||
|
func (m *MonitorService) loadDiskIO() {
|
||
|
ioStat, _ := disk.IOCounters()
|
||
|
var diskIOList []disk.IOCountersStat
|
||
|
for _, io := range ioStat {
|
||
|
diskIOList = append(diskIOList, io)
|
||
|
}
|
||
|
m.DiskIO <- diskIOList
|
||
|
}
|
||
|
|
||
|
func (m *MonitorService) loadNetIO() {
|
||
|
netStat, _ := net.IOCounters(true)
|
||
|
netStatAll, _ := net.IOCounters(false)
|
||
|
var netList []net.IOCountersStat
|
||
|
netList = append(netList, netStat...)
|
||
|
netList = append(netList, netStatAll...)
|
||
|
m.NetIO <- netList
|
||
|
}
|
||
|
|
||
|
func (m *MonitorService) saveIODataToDB(ctx context.Context, interval float64) {
|
||
|
defer close(m.DiskIO)
|
||
|
for {
|
||
|
select {
|
||
|
case <-ctx.Done():
|
||
|
return
|
||
|
case ioStat := <-m.DiskIO:
|
||
|
select {
|
||
|
case <-ctx.Done():
|
||
|
return
|
||
|
case ioStat2 := <-m.DiskIO:
|
||
|
var ioList []model.MonitorIO
|
||
|
for _, io2 := range ioStat2 {
|
||
|
for _, io1 := range ioStat {
|
||
|
if io2.Name == io1.Name {
|
||
|
var itemIO model.MonitorIO
|
||
|
itemIO.Name = io1.Name
|
||
|
if io2.ReadBytes != 0 && io1.ReadBytes != 0 && io2.ReadBytes > io1.ReadBytes {
|
||
|
itemIO.Read = uint64(float64(io2.ReadBytes-io1.ReadBytes) / interval / 60)
|
||
|
}
|
||
|
if io2.WriteBytes != 0 && io1.WriteBytes != 0 && io2.WriteBytes > io1.WriteBytes {
|
||
|
itemIO.Write = uint64(float64(io2.WriteBytes-io1.WriteBytes) / interval / 60)
|
||
|
}
|
||
|
|
||
|
if io2.ReadCount != 0 && io1.ReadCount != 0 && io2.ReadCount > io1.ReadCount {
|
||
|
itemIO.Count = uint64(float64(io2.ReadCount-io1.ReadCount) / interval / 60)
|
||
|
}
|
||
|
writeCount := uint64(0)
|
||
|
if io2.WriteCount != 0 && io1.WriteCount != 0 && io2.WriteCount > io1.WriteCount {
|
||
|
writeCount = uint64(float64(io2.WriteCount-io1.WriteCount) / interval * 60)
|
||
|
}
|
||
|
if writeCount > itemIO.Count {
|
||
|
itemIO.Count = writeCount
|
||
|
}
|
||
|
|
||
|
if io2.ReadTime != 0 && io1.ReadTime != 0 && io2.ReadTime > io1.ReadTime {
|
||
|
itemIO.Time = uint64(float64(io2.ReadTime-io1.ReadTime) / interval / 60)
|
||
|
}
|
||
|
writeTime := uint64(0)
|
||
|
if io2.WriteTime != 0 && io1.WriteTime != 0 && io2.WriteTime > io1.WriteTime {
|
||
|
writeTime = uint64(float64(io2.WriteTime-io1.WriteTime) / interval / 60)
|
||
|
}
|
||
|
if writeTime > itemIO.Time {
|
||
|
itemIO.Time = writeTime
|
||
|
}
|
||
|
ioList = append(ioList, itemIO)
|
||
|
break
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
if err := settingRepo.BatchCreateMonitorIO(ioList); err != nil {
|
||
|
global.LOG.Errorf("Insert io monitoring data failed, err: %v", err)
|
||
|
}
|
||
|
m.DiskIO <- ioStat2
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
|
||
|
func (m *MonitorService) saveNetDataToDB(ctx context.Context, interval float64) {
|
||
|
defer close(m.NetIO)
|
||
|
for {
|
||
|
select {
|
||
|
case <-ctx.Done():
|
||
|
return
|
||
|
case netStat := <-m.NetIO:
|
||
|
select {
|
||
|
case <-ctx.Done():
|
||
|
return
|
||
|
case netStat2 := <-m.NetIO:
|
||
|
var netList []model.MonitorNetwork
|
||
|
for _, net2 := range netStat2 {
|
||
|
for _, net1 := range netStat {
|
||
|
if net2.Name == net1.Name {
|
||
|
var itemNet model.MonitorNetwork
|
||
|
itemNet.Name = net1.Name
|
||
|
|
||
|
if net2.BytesSent != 0 && net1.BytesSent != 0 && net2.BytesSent > net1.BytesSent {
|
||
|
itemNet.Up = float64(net2.BytesSent-net1.BytesSent) / 1024 / interval / 60
|
||
|
}
|
||
|
if net2.BytesRecv != 0 && net1.BytesRecv != 0 && net2.BytesRecv > net1.BytesRecv {
|
||
|
itemNet.Down = float64(net2.BytesRecv-net1.BytesRecv) / 1024 / interval / 60
|
||
|
}
|
||
|
netList = append(netList, itemNet)
|
||
|
break
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
|
||
|
if err := settingRepo.BatchCreateMonitorNet(netList); err != nil {
|
||
|
global.LOG.Errorf("Insert network monitoring data failed, err: %v", err)
|
||
|
}
|
||
|
m.NetIO <- netStat2
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
|
||
|
func StartMonitor(removeBefore bool, interval string) error {
|
||
|
if removeBefore {
|
||
|
monitorCancel()
|
||
|
global.Cron.Remove(cron.EntryID(global.MonitorCronID))
|
||
|
}
|
||
|
intervalItem, err := strconv.Atoi(interval)
|
||
|
if err != nil {
|
||
|
return err
|
||
|
}
|
||
|
|
||
|
service := NewIMonitorService()
|
||
|
ctx, cancel := context.WithCancel(context.Background())
|
||
|
monitorCancel = cancel
|
||
|
monitorID, err := global.Cron.AddJob(fmt.Sprintf("@every %sm", interval), service)
|
||
|
if err != nil {
|
||
|
return err
|
||
|
}
|
||
|
|
||
|
service.Run()
|
||
|
|
||
|
go service.saveIODataToDB(ctx, float64(intervalItem))
|
||
|
go service.saveNetDataToDB(ctx, float64(intervalItem))
|
||
|
|
||
|
global.MonitorCronID = monitorID
|
||
|
return nil
|
||
|
}
|