mirror of
https://github.com/seaweedfs/seaweedfs.git
synced 2024-12-21 06:57:51 +08:00
35fd1e1c9a
1. unwrap the map to avoid extra map object creation 2. fix ec shard counting in UpdateEcShards
107 lines
3.1 KiB
Go
107 lines
3.1 KiB
Go
package topology
|
|
|
|
import (
|
|
"math/rand/v2"
|
|
"time"
|
|
|
|
"github.com/seaweedfs/seaweedfs/weed/stats"
|
|
"github.com/seaweedfs/seaweedfs/weed/storage/erasure_coding"
|
|
"github.com/seaweedfs/seaweedfs/weed/storage/types"
|
|
"google.golang.org/grpc"
|
|
|
|
"github.com/seaweedfs/seaweedfs/weed/glog"
|
|
"github.com/seaweedfs/seaweedfs/weed/storage"
|
|
)
|
|
|
|
func (t *Topology) StartRefreshWritableVolumes(grpcDialOption grpc.DialOption, garbageThreshold float64, concurrentVacuumLimitPerVolumeServer int, growThreshold float64, preallocate int64) {
|
|
go func() {
|
|
for {
|
|
if t.IsLeader() {
|
|
freshThreshHold := time.Now().Unix() - 3*t.pulse //3 times of sleep interval
|
|
t.CollectDeadNodeAndFullVolumes(freshThreshHold, t.volumeSizeLimit, growThreshold)
|
|
}
|
|
time.Sleep(time.Duration(float32(t.pulse*1e3)*(1+rand.Float32())) * time.Millisecond)
|
|
}
|
|
}()
|
|
go func(garbageThreshold float64) {
|
|
for {
|
|
if t.IsLeader() {
|
|
if !t.isDisableVacuum {
|
|
t.Vacuum(grpcDialOption, garbageThreshold, concurrentVacuumLimitPerVolumeServer, 0, "", preallocate)
|
|
}
|
|
} else {
|
|
stats.MasterReplicaPlacementMismatch.Reset()
|
|
}
|
|
time.Sleep(14*time.Minute + time.Duration(120*rand.Float32())*time.Second)
|
|
}
|
|
}(garbageThreshold)
|
|
go func() {
|
|
for {
|
|
select {
|
|
case fv := <-t.chanFullVolumes:
|
|
t.SetVolumeCapacityFull(fv)
|
|
case cv := <-t.chanCrowdedVolumes:
|
|
t.SetVolumeCrowded(cv)
|
|
}
|
|
}
|
|
}()
|
|
}
|
|
func (t *Topology) SetVolumeCapacityFull(volumeInfo storage.VolumeInfo) bool {
|
|
diskType := types.ToDiskType(volumeInfo.DiskType)
|
|
vl := t.GetVolumeLayout(volumeInfo.Collection, volumeInfo.ReplicaPlacement, volumeInfo.Ttl, diskType)
|
|
if !vl.SetVolumeCapacityFull(volumeInfo.Id) {
|
|
return false
|
|
}
|
|
|
|
vl.accessLock.RLock()
|
|
defer vl.accessLock.RUnlock()
|
|
|
|
vidLocations, found := vl.vid2location[volumeInfo.Id]
|
|
if !found {
|
|
return false
|
|
}
|
|
|
|
for _, dn := range vidLocations.list {
|
|
if !volumeInfo.ReadOnly {
|
|
|
|
disk := dn.getOrCreateDisk(volumeInfo.DiskType)
|
|
disk.UpAdjustDiskUsageDelta(types.ToDiskType(volumeInfo.DiskType), &DiskUsageCounts{
|
|
activeVolumeCount: -1,
|
|
})
|
|
|
|
}
|
|
}
|
|
return true
|
|
}
|
|
|
|
func (t *Topology) SetVolumeCrowded(volumeInfo storage.VolumeInfo) {
|
|
diskType := types.ToDiskType(volumeInfo.DiskType)
|
|
vl := t.GetVolumeLayout(volumeInfo.Collection, volumeInfo.ReplicaPlacement, volumeInfo.Ttl, diskType)
|
|
vl.SetVolumeCrowded(volumeInfo.Id)
|
|
}
|
|
|
|
func (t *Topology) UnRegisterDataNode(dn *DataNode) {
|
|
dn.IsTerminating = true
|
|
for _, v := range dn.GetVolumes() {
|
|
glog.V(0).Infoln("Removing Volume", v.Id, "from the dead volume server", dn.Id())
|
|
diskType := types.ToDiskType(v.DiskType)
|
|
vl := t.GetVolumeLayout(v.Collection, v.ReplicaPlacement, v.Ttl, diskType)
|
|
vl.SetVolumeUnavailable(dn, v.Id)
|
|
}
|
|
|
|
// unregister ec shards when volume server disconnected
|
|
for _, s := range dn.GetEcShards() {
|
|
t.UnRegisterEcShards(s, dn)
|
|
}
|
|
|
|
negativeUsages := dn.GetDiskUsages().negative()
|
|
for dt, du := range negativeUsages.usages {
|
|
dn.UpAdjustDiskUsageDelta(dt, du)
|
|
}
|
|
dn.DeltaUpdateVolumes([]storage.VolumeInfo{}, dn.GetVolumes())
|
|
dn.DeltaUpdateEcShards([]*erasure_coding.EcVolumeInfo{}, dn.GetEcShards())
|
|
if dn.Parent() != nil {
|
|
dn.Parent().UnlinkChildNode(dn.Id())
|
|
}
|
|
}
|