mirror of
https://github.com/seaweedfs/seaweedfs.git
synced 2024-12-28 03:47:57 +08:00
c57c79a0ab
* optimiz vacuuming volume * fix bugx * rename parameters * fix conflict * change copyDataBasedOnIndexFile to an instance method * close needlemap * optimiz commiting Vacuum volume for leveldb index * fix bugs * fix leveldb loading bugs * refactor * fix leveldb loading bug * add leveldb recovery * add test case for levelDB * modify test case to cover all the new branches * use one tmpNm instead of two instances * refactor * refactor * move setWatermark to the end * add test for watermark and updating leveldb * fix error logic * refactor, add test * check nil before close needlemapeer add test case fix metric bug * add tests, fix bugs * adjust log level remove wrong test case refactor * avoid duplicate updating metric for leveldb index
171 lines
4.1 KiB
Go
171 lines
4.1 KiB
Go
package storage
|
|
|
|
import (
|
|
"fmt"
|
|
"os"
|
|
"sync/atomic"
|
|
|
|
"github.com/seaweedfs/seaweedfs/weed/storage/idx"
|
|
. "github.com/seaweedfs/seaweedfs/weed/storage/types"
|
|
boom "github.com/tylertreat/BoomFilters"
|
|
)
|
|
|
|
type mapMetric struct {
|
|
DeletionCounter uint32 `json:"DeletionCounter"`
|
|
FileCounter uint32 `json:"FileCounter"`
|
|
DeletionByteCounter uint64 `json:"DeletionByteCounter"`
|
|
FileByteCounter uint64 `json:"FileByteCounter"`
|
|
MaximumFileKey uint64 `json:"MaxFileKey"`
|
|
}
|
|
|
|
func (mm *mapMetric) logDelete(deletedByteCount Size) {
|
|
if mm == nil {
|
|
return
|
|
}
|
|
mm.LogDeletionCounter(deletedByteCount)
|
|
}
|
|
|
|
func (mm *mapMetric) logPut(key NeedleId, oldSize Size, newSize Size) {
|
|
if mm == nil {
|
|
return
|
|
}
|
|
mm.MaybeSetMaxFileKey(key)
|
|
mm.LogFileCounter(newSize)
|
|
if oldSize > 0 && oldSize.IsValid() {
|
|
mm.LogDeletionCounter(oldSize)
|
|
}
|
|
}
|
|
func (mm *mapMetric) LogFileCounter(newSize Size) {
|
|
if mm == nil {
|
|
return
|
|
}
|
|
atomic.AddUint32(&mm.FileCounter, 1)
|
|
atomic.AddUint64(&mm.FileByteCounter, uint64(newSize))
|
|
}
|
|
func (mm *mapMetric) LogDeletionCounter(oldSize Size) {
|
|
if mm == nil {
|
|
return
|
|
}
|
|
if oldSize > 0 {
|
|
atomic.AddUint32(&mm.DeletionCounter, 1)
|
|
atomic.AddUint64(&mm.DeletionByteCounter, uint64(oldSize))
|
|
}
|
|
}
|
|
func (mm *mapMetric) ContentSize() uint64 {
|
|
if mm == nil {
|
|
return 0
|
|
}
|
|
return atomic.LoadUint64(&mm.FileByteCounter)
|
|
}
|
|
func (mm *mapMetric) DeletedSize() uint64 {
|
|
if mm == nil {
|
|
return 0
|
|
}
|
|
return atomic.LoadUint64(&mm.DeletionByteCounter)
|
|
}
|
|
func (mm *mapMetric) FileCount() int {
|
|
if mm == nil {
|
|
return 0
|
|
}
|
|
return int(atomic.LoadUint32(&mm.FileCounter))
|
|
}
|
|
func (mm *mapMetric) DeletedCount() int {
|
|
if mm == nil {
|
|
return 0
|
|
}
|
|
return int(atomic.LoadUint32(&mm.DeletionCounter))
|
|
}
|
|
func (mm *mapMetric) MaxFileKey() NeedleId {
|
|
if mm == nil {
|
|
return 0
|
|
}
|
|
t := uint64(mm.MaximumFileKey)
|
|
return Uint64ToNeedleId(t)
|
|
}
|
|
func (mm *mapMetric) MaybeSetMaxFileKey(key NeedleId) {
|
|
if mm == nil {
|
|
return
|
|
}
|
|
if key > mm.MaxFileKey() {
|
|
atomic.StoreUint64(&mm.MaximumFileKey, uint64(key))
|
|
}
|
|
}
|
|
|
|
func needleMapMetricFromIndexFile(r *os.File, mm *mapMetric) error {
|
|
var bf *boom.BloomFilter
|
|
buf := make([]byte, NeedleIdSize)
|
|
err := reverseWalkIndexFile(r, func(entryCount int64) {
|
|
bf = boom.NewBloomFilter(uint(entryCount), 0.001)
|
|
}, func(key NeedleId, offset Offset, size Size) error {
|
|
|
|
mm.MaybeSetMaxFileKey(key)
|
|
NeedleIdToBytes(buf, key)
|
|
if size.IsValid() {
|
|
mm.FileByteCounter += uint64(size)
|
|
}
|
|
|
|
mm.FileCounter++
|
|
if !bf.TestAndAdd(buf) {
|
|
// if !size.IsValid(), then this file is deleted already
|
|
if !size.IsValid() {
|
|
mm.DeletionCounter++
|
|
}
|
|
} else {
|
|
// deleted file
|
|
mm.DeletionCounter++
|
|
if size.IsValid() {
|
|
// previously already deleted file
|
|
mm.DeletionByteCounter += uint64(size)
|
|
}
|
|
}
|
|
return nil
|
|
})
|
|
return err
|
|
}
|
|
|
|
func newNeedleMapMetricFromIndexFile(r *os.File) (mm *mapMetric, err error) {
|
|
mm = &mapMetric{}
|
|
err = needleMapMetricFromIndexFile(r, mm)
|
|
return
|
|
}
|
|
|
|
func reverseWalkIndexFile(r *os.File, initFn func(entryCount int64), fn func(key NeedleId, offset Offset, size Size) error) error {
|
|
fi, err := r.Stat()
|
|
if err != nil {
|
|
return fmt.Errorf("file %s stat error: %v", r.Name(), err)
|
|
}
|
|
fileSize := fi.Size()
|
|
if fileSize%NeedleMapEntrySize != 0 {
|
|
return fmt.Errorf("unexpected file %s size: %d", r.Name(), fileSize)
|
|
}
|
|
|
|
entryCount := fileSize / NeedleMapEntrySize
|
|
initFn(entryCount)
|
|
|
|
batchSize := int64(1024 * 4)
|
|
|
|
bytes := make([]byte, NeedleMapEntrySize*batchSize)
|
|
nextBatchSize := entryCount % batchSize
|
|
if nextBatchSize == 0 {
|
|
nextBatchSize = batchSize
|
|
}
|
|
remainingCount := entryCount - nextBatchSize
|
|
|
|
for remainingCount >= 0 {
|
|
_, e := r.ReadAt(bytes[:NeedleMapEntrySize*nextBatchSize], NeedleMapEntrySize*remainingCount)
|
|
// glog.V(0).Infoln("file", r.Name(), "readerOffset", NeedleMapEntrySize*remainingCount, "count", count, "e", e)
|
|
if e != nil {
|
|
return e
|
|
}
|
|
for i := int(nextBatchSize) - 1; i >= 0; i-- {
|
|
key, offset, size := idx.IdxFileEntry(bytes[i*NeedleMapEntrySize : i*NeedleMapEntrySize+NeedleMapEntrySize])
|
|
if e = fn(key, offset, size); e != nil {
|
|
return e
|
|
}
|
|
}
|
|
nextBatchSize = batchSize
|
|
remainingCount -= nextBatchSize
|
|
}
|
|
return nil
|
|
}
|