2015-05-26 07:58:41 +00:00
|
|
|
package storage
|
|
|
|
|
|
|
|
import (
|
|
|
|
"fmt"
|
|
|
|
"io"
|
|
|
|
"io/ioutil"
|
|
|
|
"net/url"
|
|
|
|
"os"
|
|
|
|
"sort"
|
|
|
|
"strconv"
|
|
|
|
|
2016-06-03 01:09:14 +00:00
|
|
|
"github.com/chrislusf/seaweedfs/weed/glog"
|
|
|
|
"github.com/chrislusf/seaweedfs/weed/operation"
|
2017-05-27 05:51:25 +00:00
|
|
|
"github.com/chrislusf/seaweedfs/weed/storage/needle"
|
2018-07-08 09:28:04 +00:00
|
|
|
. "github.com/chrislusf/seaweedfs/weed/storage/types"
|
2016-06-03 01:09:14 +00:00
|
|
|
"github.com/chrislusf/seaweedfs/weed/util"
|
2015-05-26 07:58:41 +00:00
|
|
|
)
|
|
|
|
|
|
|
|
// The volume sync with a master volume via 2 steps:
|
|
|
|
// 1. The slave checks master side to find subscription checkpoint
|
|
|
|
// to setup the replication.
|
|
|
|
// 2. The slave receives the updates from master
|
|
|
|
|
|
|
|
/*
|
|
|
|
Assume the slave volume needs to follow the master volume.
|
|
|
|
|
|
|
|
The master volume could be compacted, and could be many files ahead of
|
|
|
|
slave volume.
|
|
|
|
|
|
|
|
Step 1:
|
|
|
|
The slave volume will ask the master volume for a snapshot
|
|
|
|
of (existing file entries, last offset, number of compacted times).
|
|
|
|
|
|
|
|
For each entry x in master existing file entries:
|
|
|
|
if x does not exist locally:
|
|
|
|
add x locally
|
|
|
|
|
|
|
|
For each entry y in local slave existing file entries:
|
|
|
|
if y does not exist on master:
|
|
|
|
delete y locally
|
|
|
|
|
|
|
|
Step 2:
|
|
|
|
After this, use the last offset and number of compacted times to request
|
|
|
|
the master volume to send a new file, and keep looping. If the number of
|
|
|
|
compacted times is changed, go back to step 1 (very likely this can be
|
|
|
|
optimized more later).
|
|
|
|
|
|
|
|
*/
|
|
|
|
|
|
|
|
func (v *Volume) Synchronize(volumeServer string) (err error) {
|
|
|
|
var lastCompactRevision uint16 = 0
|
|
|
|
var compactRevision uint16 = 0
|
2017-05-27 05:51:25 +00:00
|
|
|
var masterMap *needle.CompactMap
|
2015-05-26 07:58:41 +00:00
|
|
|
for i := 0; i < 3; i++ {
|
|
|
|
if masterMap, _, compactRevision, err = fetchVolumeFileEntries(volumeServer, v.Id); err != nil {
|
|
|
|
return fmt.Errorf("Failed to sync volume %d entries with %s: %v", v.Id, volumeServer, err)
|
|
|
|
}
|
|
|
|
if lastCompactRevision != compactRevision && lastCompactRevision != 0 {
|
2017-08-30 06:59:53 +00:00
|
|
|
if err = v.Compact(0); err != nil {
|
2015-05-26 07:58:41 +00:00
|
|
|
return fmt.Errorf("Compact Volume before synchronizing %v", err)
|
|
|
|
}
|
|
|
|
if err = v.commitCompact(); err != nil {
|
|
|
|
return fmt.Errorf("Commit Compact before synchronizing %v", err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
lastCompactRevision = compactRevision
|
|
|
|
if err = v.trySynchronizing(volumeServer, masterMap, compactRevision); err == nil {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
2017-05-27 05:51:25 +00:00
|
|
|
type ByOffset []needle.NeedleValue
|
2015-05-26 07:58:41 +00:00
|
|
|
|
|
|
|
func (a ByOffset) Len() int { return len(a) }
|
|
|
|
func (a ByOffset) Swap(i, j int) { a[i], a[j] = a[j], a[i] }
|
|
|
|
func (a ByOffset) Less(i, j int) bool { return a[i].Offset < a[j].Offset }
|
|
|
|
|
|
|
|
// trySynchronizing sync with remote volume server incrementally by
|
|
|
|
// make up the local and remote delta.
|
2017-05-27 05:51:25 +00:00
|
|
|
func (v *Volume) trySynchronizing(volumeServer string, masterMap *needle.CompactMap, compactRevision uint16) error {
|
2015-05-26 07:58:41 +00:00
|
|
|
slaveIdxFile, err := os.Open(v.nm.IndexFileName())
|
|
|
|
if err != nil {
|
|
|
|
return fmt.Errorf("Open volume %d index file: %v", v.Id, err)
|
|
|
|
}
|
|
|
|
defer slaveIdxFile.Close()
|
2017-05-27 05:51:25 +00:00
|
|
|
slaveMap, err := LoadBtreeNeedleMap(slaveIdxFile)
|
2015-05-26 07:58:41 +00:00
|
|
|
if err != nil {
|
|
|
|
return fmt.Errorf("Load volume %d index file: %v", v.Id, err)
|
|
|
|
}
|
2017-05-27 05:51:25 +00:00
|
|
|
var delta []needle.NeedleValue
|
|
|
|
if err := masterMap.Visit(func(needleValue needle.NeedleValue) error {
|
2015-05-26 07:58:41 +00:00
|
|
|
if needleValue.Key == 0 {
|
|
|
|
return nil
|
|
|
|
}
|
2018-07-08 09:28:04 +00:00
|
|
|
if _, ok := slaveMap.Get(needleValue.Key); ok {
|
2015-05-26 07:58:41 +00:00
|
|
|
return nil // skip intersection
|
|
|
|
}
|
|
|
|
delta = append(delta, needleValue)
|
|
|
|
return nil
|
|
|
|
}); err != nil {
|
|
|
|
return fmt.Errorf("Add master entry: %v", err)
|
|
|
|
}
|
2017-05-27 05:51:25 +00:00
|
|
|
if err := slaveMap.m.Visit(func(needleValue needle.NeedleValue) error {
|
2015-05-26 07:58:41 +00:00
|
|
|
if needleValue.Key == 0 {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
if _, ok := masterMap.Get(needleValue.Key); ok {
|
|
|
|
return nil // skip intersection
|
|
|
|
}
|
|
|
|
needleValue.Size = 0
|
|
|
|
delta = append(delta, needleValue)
|
|
|
|
return nil
|
|
|
|
}); err != nil {
|
|
|
|
return fmt.Errorf("Remove local entry: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// simulate to same ordering of remote .dat file needle entries
|
|
|
|
sort.Sort(ByOffset(delta))
|
|
|
|
|
|
|
|
// make up the delta
|
|
|
|
fetchCount := 0
|
|
|
|
volumeDataContentHandlerUrl := "http://" + volumeServer + "/admin/sync/data"
|
|
|
|
for _, needleValue := range delta {
|
|
|
|
if needleValue.Size == 0 {
|
|
|
|
// remove file entry from local
|
|
|
|
v.removeNeedle(needleValue.Key)
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
// add master file entry to local data file
|
|
|
|
if err := v.fetchNeedle(volumeDataContentHandlerUrl, needleValue, compactRevision); err != nil {
|
|
|
|
glog.V(0).Infof("Fetch needle %v from %s: %v", needleValue, volumeServer, err)
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
fetchCount++
|
|
|
|
}
|
|
|
|
glog.V(1).Infof("Fetched %d needles from %s", fetchCount, volumeServer)
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2017-05-27 05:51:25 +00:00
|
|
|
func fetchVolumeFileEntries(volumeServer string, vid VolumeId) (m *needle.CompactMap, lastOffset uint64, compactRevision uint16, err error) {
|
|
|
|
m = needle.NewCompactMap()
|
2015-05-26 07:58:41 +00:00
|
|
|
|
|
|
|
syncStatus, err := operation.GetVolumeSyncStatus(volumeServer, vid.String())
|
|
|
|
if err != nil {
|
|
|
|
return m, 0, 0, err
|
|
|
|
}
|
|
|
|
|
|
|
|
total := 0
|
2018-07-08 09:28:04 +00:00
|
|
|
err = operation.GetVolumeIdxEntries(volumeServer, vid.String(), func(key NeedleId, offset Offset, size uint32) {
|
2015-05-26 07:58:41 +00:00
|
|
|
// println("remote key", key, "offset", offset*NeedlePaddingSize, "size", size)
|
2017-01-06 18:22:20 +00:00
|
|
|
if offset > 0 && size != TombstoneFileSize {
|
2018-07-08 09:28:04 +00:00
|
|
|
m.Set(NeedleId(key), offset, size)
|
2015-05-26 07:58:41 +00:00
|
|
|
} else {
|
2018-07-08 09:28:04 +00:00
|
|
|
m.Delete(NeedleId(key))
|
2015-05-26 07:58:41 +00:00
|
|
|
}
|
|
|
|
total++
|
|
|
|
})
|
|
|
|
|
|
|
|
glog.V(2).Infof("server %s volume %d, entries %d, last offset %d, revision %d", volumeServer, vid, total, syncStatus.TailOffset, syncStatus.CompactRevision)
|
|
|
|
return m, syncStatus.TailOffset, syncStatus.CompactRevision, err
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
func (v *Volume) GetVolumeSyncStatus() operation.SyncVolumeResponse {
|
|
|
|
var syncStatus = operation.SyncVolumeResponse{}
|
|
|
|
if stat, err := v.dataFile.Stat(); err == nil {
|
|
|
|
syncStatus.TailOffset = uint64(stat.Size())
|
|
|
|
}
|
|
|
|
syncStatus.IdxFileSize = v.nm.IndexFileSize()
|
|
|
|
syncStatus.CompactRevision = v.SuperBlock.CompactRevision
|
|
|
|
syncStatus.Ttl = v.SuperBlock.Ttl.String()
|
|
|
|
syncStatus.Replication = v.SuperBlock.ReplicaPlacement.String()
|
|
|
|
return syncStatus
|
|
|
|
}
|
|
|
|
|
|
|
|
func (v *Volume) IndexFileContent() ([]byte, error) {
|
|
|
|
return v.nm.IndexFileContent()
|
|
|
|
}
|
|
|
|
|
|
|
|
// removeNeedle removes one needle by needle key
|
2018-07-08 09:28:04 +00:00
|
|
|
func (v *Volume) removeNeedle(key NeedleId) {
|
2015-05-26 07:58:41 +00:00
|
|
|
n := new(Needle)
|
2018-07-08 09:28:04 +00:00
|
|
|
n.Id = key
|
2016-07-03 07:10:27 +00:00
|
|
|
v.deleteNeedle(n)
|
2015-05-26 07:58:41 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// fetchNeedle fetches a remote volume needle by vid, id, offset
|
|
|
|
// The compact revision is checked first in case the remote volume
|
|
|
|
// is compacted and the offset is invalid any more.
|
|
|
|
func (v *Volume) fetchNeedle(volumeDataContentHandlerUrl string,
|
2017-05-27 05:51:25 +00:00
|
|
|
needleValue needle.NeedleValue, compactRevision uint16) error {
|
2015-05-26 07:58:41 +00:00
|
|
|
// add master file entry to local data file
|
|
|
|
values := make(url.Values)
|
|
|
|
values.Add("revision", strconv.Itoa(int(compactRevision)))
|
|
|
|
values.Add("volume", v.Id.String())
|
|
|
|
values.Add("id", needleValue.Key.String())
|
|
|
|
values.Add("offset", strconv.FormatUint(uint64(needleValue.Offset), 10))
|
|
|
|
values.Add("size", strconv.FormatUint(uint64(needleValue.Size), 10))
|
|
|
|
glog.V(4).Infof("Fetch %+v", needleValue)
|
|
|
|
return util.GetUrlStream(volumeDataContentHandlerUrl, values, func(r io.Reader) error {
|
|
|
|
b, err := ioutil.ReadAll(r)
|
|
|
|
if err != nil {
|
|
|
|
return fmt.Errorf("Reading from %s error: %v", volumeDataContentHandlerUrl, err)
|
|
|
|
}
|
|
|
|
offset, err := v.AppendBlob(b)
|
|
|
|
if err != nil {
|
|
|
|
return fmt.Errorf("Appending volume %d error: %v", v.Id, err)
|
|
|
|
}
|
|
|
|
// println("add key", needleValue.Key, "offset", offset, "size", needleValue.Size)
|
2018-07-08 09:28:04 +00:00
|
|
|
v.nm.Put(needleValue.Key, Offset(offset/NeedlePaddingSize), needleValue.Size)
|
2015-05-26 07:58:41 +00:00
|
|
|
return nil
|
|
|
|
})
|
|
|
|
}
|