2015-05-26 15:58:41 +08:00
|
|
|
package storage
|
|
|
|
|
|
|
|
import (
|
2018-10-16 12:44:41 +08:00
|
|
|
"context"
|
2015-05-26 15:58:41 +08:00
|
|
|
"fmt"
|
|
|
|
"os"
|
|
|
|
"sort"
|
|
|
|
|
2016-06-03 09:09:14 +08:00
|
|
|
"github.com/chrislusf/seaweedfs/weed/glog"
|
|
|
|
"github.com/chrislusf/seaweedfs/weed/operation"
|
2018-10-15 16:19:15 +08:00
|
|
|
"github.com/chrislusf/seaweedfs/weed/pb/volume_server_pb"
|
2017-05-27 13:51:25 +08:00
|
|
|
"github.com/chrislusf/seaweedfs/weed/storage/needle"
|
2018-07-08 17:28:04 +08:00
|
|
|
. "github.com/chrislusf/seaweedfs/weed/storage/types"
|
2015-05-26 15:58:41 +08:00
|
|
|
)
|
|
|
|
|
|
|
|
// The volume sync with a master volume via 2 steps:
|
|
|
|
// 1. The slave checks master side to find subscription checkpoint
|
|
|
|
// to setup the replication.
|
|
|
|
// 2. The slave receives the updates from master
|
|
|
|
|
|
|
|
/*
|
|
|
|
Assume the slave volume needs to follow the master volume.
|
|
|
|
|
|
|
|
The master volume could be compacted, and could be many files ahead of
|
|
|
|
slave volume.
|
|
|
|
|
|
|
|
Step 1:
|
|
|
|
The slave volume will ask the master volume for a snapshot
|
|
|
|
of (existing file entries, last offset, number of compacted times).
|
|
|
|
|
|
|
|
For each entry x in master existing file entries:
|
|
|
|
if x does not exist locally:
|
|
|
|
add x locally
|
|
|
|
|
|
|
|
For each entry y in local slave existing file entries:
|
|
|
|
if y does not exist on master:
|
|
|
|
delete y locally
|
|
|
|
|
|
|
|
Step 2:
|
|
|
|
After this, use the last offset and number of compacted times to request
|
|
|
|
the master volume to send a new file, and keep looping. If the number of
|
|
|
|
compacted times is changed, go back to step 1 (very likely this can be
|
|
|
|
optimized more later).
|
|
|
|
|
|
|
|
*/
|
|
|
|
|
|
|
|
func (v *Volume) Synchronize(volumeServer string) (err error) {
|
|
|
|
var lastCompactRevision uint16 = 0
|
|
|
|
var compactRevision uint16 = 0
|
2017-05-27 13:51:25 +08:00
|
|
|
var masterMap *needle.CompactMap
|
2015-05-26 15:58:41 +08:00
|
|
|
for i := 0; i < 3; i++ {
|
|
|
|
if masterMap, _, compactRevision, err = fetchVolumeFileEntries(volumeServer, v.Id); err != nil {
|
|
|
|
return fmt.Errorf("Failed to sync volume %d entries with %s: %v", v.Id, volumeServer, err)
|
|
|
|
}
|
|
|
|
if lastCompactRevision != compactRevision && lastCompactRevision != 0 {
|
2017-08-30 14:59:53 +08:00
|
|
|
if err = v.Compact(0); err != nil {
|
2015-05-26 15:58:41 +08:00
|
|
|
return fmt.Errorf("Compact Volume before synchronizing %v", err)
|
|
|
|
}
|
|
|
|
if err = v.commitCompact(); err != nil {
|
|
|
|
return fmt.Errorf("Commit Compact before synchronizing %v", err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
lastCompactRevision = compactRevision
|
|
|
|
if err = v.trySynchronizing(volumeServer, masterMap, compactRevision); err == nil {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
2017-05-27 13:51:25 +08:00
|
|
|
type ByOffset []needle.NeedleValue
|
2015-05-26 15:58:41 +08:00
|
|
|
|
|
|
|
func (a ByOffset) Len() int { return len(a) }
|
|
|
|
func (a ByOffset) Swap(i, j int) { a[i], a[j] = a[j], a[i] }
|
|
|
|
func (a ByOffset) Less(i, j int) bool { return a[i].Offset < a[j].Offset }
|
|
|
|
|
|
|
|
// trySynchronizing sync with remote volume server incrementally by
|
|
|
|
// make up the local and remote delta.
|
2017-05-27 13:51:25 +08:00
|
|
|
func (v *Volume) trySynchronizing(volumeServer string, masterMap *needle.CompactMap, compactRevision uint16) error {
|
2015-05-26 15:58:41 +08:00
|
|
|
slaveIdxFile, err := os.Open(v.nm.IndexFileName())
|
|
|
|
if err != nil {
|
|
|
|
return fmt.Errorf("Open volume %d index file: %v", v.Id, err)
|
|
|
|
}
|
|
|
|
defer slaveIdxFile.Close()
|
2017-05-27 13:51:25 +08:00
|
|
|
slaveMap, err := LoadBtreeNeedleMap(slaveIdxFile)
|
2015-05-26 15:58:41 +08:00
|
|
|
if err != nil {
|
|
|
|
return fmt.Errorf("Load volume %d index file: %v", v.Id, err)
|
|
|
|
}
|
2017-05-27 13:51:25 +08:00
|
|
|
var delta []needle.NeedleValue
|
|
|
|
if err := masterMap.Visit(func(needleValue needle.NeedleValue) error {
|
2018-08-01 14:25:26 +08:00
|
|
|
if needleValue.Key == NeedleIdEmpty {
|
2015-05-26 15:58:41 +08:00
|
|
|
return nil
|
|
|
|
}
|
2018-07-08 17:28:04 +08:00
|
|
|
if _, ok := slaveMap.Get(needleValue.Key); ok {
|
2015-05-26 15:58:41 +08:00
|
|
|
return nil // skip intersection
|
|
|
|
}
|
|
|
|
delta = append(delta, needleValue)
|
|
|
|
return nil
|
|
|
|
}); err != nil {
|
|
|
|
return fmt.Errorf("Add master entry: %v", err)
|
|
|
|
}
|
2017-05-27 13:51:25 +08:00
|
|
|
if err := slaveMap.m.Visit(func(needleValue needle.NeedleValue) error {
|
2018-08-01 14:25:26 +08:00
|
|
|
if needleValue.Key == NeedleIdEmpty {
|
2015-05-26 15:58:41 +08:00
|
|
|
return nil
|
|
|
|
}
|
|
|
|
if _, ok := masterMap.Get(needleValue.Key); ok {
|
|
|
|
return nil // skip intersection
|
|
|
|
}
|
|
|
|
needleValue.Size = 0
|
|
|
|
delta = append(delta, needleValue)
|
|
|
|
return nil
|
|
|
|
}); err != nil {
|
|
|
|
return fmt.Errorf("Remove local entry: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// simulate to same ordering of remote .dat file needle entries
|
|
|
|
sort.Sort(ByOffset(delta))
|
|
|
|
|
|
|
|
// make up the delta
|
|
|
|
fetchCount := 0
|
|
|
|
for _, needleValue := range delta {
|
|
|
|
if needleValue.Size == 0 {
|
|
|
|
// remove file entry from local
|
|
|
|
v.removeNeedle(needleValue.Key)
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
// add master file entry to local data file
|
2018-10-16 12:44:41 +08:00
|
|
|
if err := v.fetchNeedle(volumeServer, needleValue, compactRevision); err != nil {
|
2015-05-26 15:58:41 +08:00
|
|
|
glog.V(0).Infof("Fetch needle %v from %s: %v", needleValue, volumeServer, err)
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
fetchCount++
|
|
|
|
}
|
|
|
|
glog.V(1).Infof("Fetched %d needles from %s", fetchCount, volumeServer)
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2017-05-27 13:51:25 +08:00
|
|
|
func fetchVolumeFileEntries(volumeServer string, vid VolumeId) (m *needle.CompactMap, lastOffset uint64, compactRevision uint16, err error) {
|
|
|
|
m = needle.NewCompactMap()
|
2015-05-26 15:58:41 +08:00
|
|
|
|
2018-10-15 16:19:15 +08:00
|
|
|
syncStatus, err := operation.GetVolumeSyncStatus(volumeServer, uint32(vid))
|
2015-05-26 15:58:41 +08:00
|
|
|
if err != nil {
|
|
|
|
return m, 0, 0, err
|
|
|
|
}
|
|
|
|
|
|
|
|
total := 0
|
2018-10-16 12:44:41 +08:00
|
|
|
err = operation.GetVolumeIdxEntries(volumeServer, uint32(vid), func(key NeedleId, offset Offset, size uint32) {
|
2015-05-26 15:58:41 +08:00
|
|
|
// println("remote key", key, "offset", offset*NeedlePaddingSize, "size", size)
|
2017-01-07 02:22:20 +08:00
|
|
|
if offset > 0 && size != TombstoneFileSize {
|
2018-07-08 17:28:04 +08:00
|
|
|
m.Set(NeedleId(key), offset, size)
|
2015-05-26 15:58:41 +08:00
|
|
|
} else {
|
2018-07-08 17:28:04 +08:00
|
|
|
m.Delete(NeedleId(key))
|
2015-05-26 15:58:41 +08:00
|
|
|
}
|
|
|
|
total++
|
|
|
|
})
|
|
|
|
|
|
|
|
glog.V(2).Infof("server %s volume %d, entries %d, last offset %d, revision %d", volumeServer, vid, total, syncStatus.TailOffset, syncStatus.CompactRevision)
|
2018-10-15 16:19:15 +08:00
|
|
|
return m, syncStatus.TailOffset, uint16(syncStatus.CompactRevision), err
|
2015-05-26 15:58:41 +08:00
|
|
|
|
|
|
|
}
|
|
|
|
|
2018-10-15 16:19:15 +08:00
|
|
|
func (v *Volume) GetVolumeSyncStatus() *volume_server_pb.VolumeSyncStatusResponse {
|
|
|
|
var syncStatus = &volume_server_pb.VolumeSyncStatusResponse{}
|
2015-05-26 15:58:41 +08:00
|
|
|
if stat, err := v.dataFile.Stat(); err == nil {
|
|
|
|
syncStatus.TailOffset = uint64(stat.Size())
|
|
|
|
}
|
|
|
|
syncStatus.IdxFileSize = v.nm.IndexFileSize()
|
2018-10-15 16:19:15 +08:00
|
|
|
syncStatus.CompactRevision = uint32(v.SuperBlock.CompactRevision)
|
2015-05-26 15:58:41 +08:00
|
|
|
syncStatus.Ttl = v.SuperBlock.Ttl.String()
|
|
|
|
syncStatus.Replication = v.SuperBlock.ReplicaPlacement.String()
|
|
|
|
return syncStatus
|
|
|
|
}
|
|
|
|
|
|
|
|
func (v *Volume) IndexFileContent() ([]byte, error) {
|
|
|
|
return v.nm.IndexFileContent()
|
|
|
|
}
|
|
|
|
|
|
|
|
// removeNeedle removes one needle by needle key
|
2018-07-08 17:28:04 +08:00
|
|
|
func (v *Volume) removeNeedle(key NeedleId) {
|
2015-05-26 15:58:41 +08:00
|
|
|
n := new(Needle)
|
2018-07-08 17:28:04 +08:00
|
|
|
n.Id = key
|
2016-07-03 15:10:27 +08:00
|
|
|
v.deleteNeedle(n)
|
2015-05-26 15:58:41 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
// fetchNeedle fetches a remote volume needle by vid, id, offset
|
|
|
|
// The compact revision is checked first in case the remote volume
|
|
|
|
// is compacted and the offset is invalid any more.
|
2018-10-16 12:44:41 +08:00
|
|
|
func (v *Volume) fetchNeedle(volumeServer string, needleValue needle.NeedleValue, compactRevision uint16) error {
|
|
|
|
|
|
|
|
return operation.WithVolumeServerClient(volumeServer, func(client volume_server_pb.VolumeServerClient) error {
|
|
|
|
resp, err := client.VolumeSyncData(context.Background(), &volume_server_pb.VolumeSyncDataRequest{
|
|
|
|
VolumdId: uint32(v.Id),
|
|
|
|
Revision: uint32(compactRevision),
|
|
|
|
Offset: uint32(needleValue.Offset),
|
|
|
|
Size: uint32(needleValue.Size),
|
|
|
|
NeedleId: needleValue.Key.String(),
|
|
|
|
})
|
2015-05-26 15:58:41 +08:00
|
|
|
if err != nil {
|
2018-10-16 12:44:41 +08:00
|
|
|
return err
|
2015-05-26 15:58:41 +08:00
|
|
|
}
|
2018-10-16 12:44:41 +08:00
|
|
|
|
|
|
|
offset, err := v.AppendBlob(resp.FileContent)
|
2015-05-26 15:58:41 +08:00
|
|
|
if err != nil {
|
|
|
|
return fmt.Errorf("Appending volume %d error: %v", v.Id, err)
|
|
|
|
}
|
|
|
|
// println("add key", needleValue.Key, "offset", offset, "size", needleValue.Size)
|
2018-07-08 17:28:04 +08:00
|
|
|
v.nm.Put(needleValue.Key, Offset(offset/NeedlePaddingSize), needleValue.Size)
|
2015-05-26 15:58:41 +08:00
|
|
|
return nil
|
|
|
|
})
|
2018-10-16 12:44:41 +08:00
|
|
|
|
2015-05-26 15:58:41 +08:00
|
|
|
}
|