2019-05-22 05:41:20 +00:00
|
|
|
package storage
|
|
|
|
|
|
|
|
import (
|
2019-05-28 04:22:23 +00:00
|
|
|
"context"
|
2019-05-26 06:23:19 +00:00
|
|
|
"fmt"
|
2019-05-28 04:22:23 +00:00
|
|
|
"io"
|
2019-05-29 06:48:39 +00:00
|
|
|
"sync"
|
2019-05-28 05:54:58 +00:00
|
|
|
"time"
|
2019-05-26 06:23:19 +00:00
|
|
|
|
|
|
|
"github.com/chrislusf/seaweedfs/weed/glog"
|
2019-05-28 04:22:23 +00:00
|
|
|
"github.com/chrislusf/seaweedfs/weed/operation"
|
2019-05-22 05:41:20 +00:00
|
|
|
"github.com/chrislusf/seaweedfs/weed/pb/master_pb"
|
2019-05-28 04:22:23 +00:00
|
|
|
"github.com/chrislusf/seaweedfs/weed/pb/volume_server_pb"
|
2019-05-26 06:23:19 +00:00
|
|
|
"github.com/chrislusf/seaweedfs/weed/storage/erasure_coding"
|
|
|
|
"github.com/chrislusf/seaweedfs/weed/storage/needle"
|
2019-05-29 06:48:39 +00:00
|
|
|
"github.com/klauspost/reedsolomon"
|
2019-05-22 05:41:20 +00:00
|
|
|
)
|
|
|
|
|
|
|
|
func (s *Store) CollectErasureCodingHeartbeat() *master_pb.Heartbeat {
|
|
|
|
var ecShardMessages []*master_pb.VolumeEcShardInformationMessage
|
|
|
|
for _, location := range s.Locations {
|
2019-05-28 04:40:51 +00:00
|
|
|
location.ecVolumesLock.RLock()
|
|
|
|
for _, ecShards := range location.ecVolumes {
|
2019-05-24 20:28:44 +00:00
|
|
|
ecShardMessages = append(ecShardMessages, ecShards.ToVolumeEcShardInformationMessage()...)
|
2019-05-22 05:41:20 +00:00
|
|
|
}
|
2019-05-28 04:40:51 +00:00
|
|
|
location.ecVolumesLock.RUnlock()
|
2019-05-22 05:41:20 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
return &master_pb.Heartbeat{
|
|
|
|
EcShards: ecShardMessages,
|
|
|
|
}
|
|
|
|
|
|
|
|
}
|
2019-05-26 06:23:19 +00:00
|
|
|
|
|
|
|
func (s *Store) MountEcShards(collection string, vid needle.VolumeId, shardId erasure_coding.ShardId) error {
|
|
|
|
for _, location := range s.Locations {
|
|
|
|
if err := location.LoadEcShard(collection, vid, shardId); err == nil {
|
|
|
|
glog.V(0).Infof("MountEcShards %d.%d", vid, shardId)
|
|
|
|
|
|
|
|
var shardBits erasure_coding.ShardBits
|
|
|
|
|
|
|
|
s.NewEcShardsChan <- master_pb.VolumeEcShardInformationMessage{
|
2019-05-27 08:29:46 +00:00
|
|
|
Id: uint32(vid),
|
|
|
|
Collection: collection,
|
|
|
|
EcIndexBits: uint32(shardBits.AddShardId(shardId)),
|
2019-05-26 06:23:19 +00:00
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return fmt.Errorf("MountEcShards %d.%d not found on disk", vid, shardId)
|
|
|
|
}
|
|
|
|
|
|
|
|
func (s *Store) UnmountEcShards(vid needle.VolumeId, shardId erasure_coding.ShardId) error {
|
|
|
|
|
|
|
|
ecShard, found := s.findEcShard(vid, shardId)
|
|
|
|
if !found {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
var shardBits erasure_coding.ShardBits
|
|
|
|
message := master_pb.VolumeEcShardInformationMessage{
|
2019-05-27 08:29:46 +00:00
|
|
|
Id: uint32(vid),
|
|
|
|
Collection: ecShard.Collection,
|
|
|
|
EcIndexBits: uint32(shardBits.AddShardId(shardId)),
|
2019-05-26 06:23:19 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
for _, location := range s.Locations {
|
|
|
|
if deleted := location.UnloadEcShard(vid, shardId); deleted {
|
|
|
|
glog.V(0).Infof("UnmountEcShards %d.%d", vid, shardId)
|
|
|
|
s.DeletedEcShardsChan <- message
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return fmt.Errorf("UnmountEcShards %d.%d not found on disk", vid, shardId)
|
|
|
|
}
|
|
|
|
|
2019-05-27 08:29:46 +00:00
|
|
|
func (s *Store) findEcShard(vid needle.VolumeId, shardId erasure_coding.ShardId) (*erasure_coding.EcVolumeShard, bool) {
|
2019-05-26 06:23:19 +00:00
|
|
|
for _, location := range s.Locations {
|
|
|
|
if v, found := location.FindEcShard(vid, shardId); found {
|
|
|
|
return v, found
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return nil, false
|
|
|
|
}
|
2019-05-27 08:29:46 +00:00
|
|
|
|
2019-05-28 04:40:51 +00:00
|
|
|
func (s *Store) FindEcVolume(vid needle.VolumeId) (*erasure_coding.EcVolume, bool) {
|
2019-05-27 08:29:46 +00:00
|
|
|
for _, location := range s.Locations {
|
2019-05-28 04:40:51 +00:00
|
|
|
if s, found := location.FindEcVolume(vid); found {
|
2019-05-27 08:29:46 +00:00
|
|
|
return s, true
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return nil, false
|
|
|
|
}
|
|
|
|
|
2019-05-28 04:22:23 +00:00
|
|
|
func (s *Store) ReadEcShardNeedle(ctx context.Context, vid needle.VolumeId, n *needle.Needle) (int, error) {
|
2019-05-27 08:29:46 +00:00
|
|
|
for _, location := range s.Locations {
|
2019-05-28 04:40:51 +00:00
|
|
|
if localEcVolume, found := location.FindEcVolume(vid); found {
|
2019-05-27 18:59:03 +00:00
|
|
|
|
2019-05-29 04:29:07 +00:00
|
|
|
// TODO need to read the version
|
|
|
|
version := needle.CurrentVersion
|
|
|
|
|
|
|
|
offset, size, intervals, err := localEcVolume.LocateEcShardNeedle(n, version)
|
2019-05-27 18:59:03 +00:00
|
|
|
if err != nil {
|
|
|
|
return 0, err
|
|
|
|
}
|
|
|
|
|
2019-05-28 07:51:01 +00:00
|
|
|
glog.V(4).Infof("read ec volume %d offset %d size %d intervals:%+v", vid, offset.ToAcutalOffset(), size, intervals)
|
|
|
|
|
2019-05-29 04:29:07 +00:00
|
|
|
if len(intervals) > 1 {
|
|
|
|
glog.V(4).Infof("ReadEcShardNeedle needle id %s intervals:%+v", n.String(), intervals)
|
|
|
|
}
|
|
|
|
bytes, err := s.readEcShardIntervals(ctx, vid, localEcVolume, intervals)
|
2019-05-27 18:59:03 +00:00
|
|
|
if err != nil {
|
|
|
|
return 0, fmt.Errorf("ReadEcShardIntervals: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
err = n.ReadBytes(bytes, offset.ToAcutalOffset(), size, version)
|
|
|
|
if err != nil {
|
|
|
|
return 0, fmt.Errorf("readbytes: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
return len(bytes), nil
|
2019-05-27 08:29:46 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
return 0, fmt.Errorf("ec shard %d not found", vid)
|
|
|
|
}
|
2019-05-27 18:59:03 +00:00
|
|
|
|
2019-05-29 04:29:07 +00:00
|
|
|
func (s *Store) readEcShardIntervals(ctx context.Context, vid needle.VolumeId, ecVolume *erasure_coding.EcVolume, intervals []erasure_coding.Interval) (data []byte, err error) {
|
2019-05-28 05:54:58 +00:00
|
|
|
|
|
|
|
if err = s.cachedLookupEcShardLocations(ctx, ecVolume); err != nil {
|
|
|
|
return nil, fmt.Errorf("failed to locate shard via master grpc %s: %v", s.MasterAddress, err)
|
2019-05-28 04:22:23 +00:00
|
|
|
}
|
|
|
|
|
2019-05-27 18:59:03 +00:00
|
|
|
for i, interval := range intervals {
|
2019-05-29 04:29:07 +00:00
|
|
|
if d, e := s.readOneEcShardInterval(ctx, ecVolume, interval); e != nil {
|
2019-05-27 18:59:03 +00:00
|
|
|
return nil, e
|
|
|
|
} else {
|
|
|
|
if i == 0 {
|
|
|
|
data = d
|
|
|
|
} else {
|
|
|
|
data = append(data, d...)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
2019-05-29 04:29:07 +00:00
|
|
|
func (s *Store) readOneEcShardInterval(ctx context.Context, ecVolume *erasure_coding.EcVolume, interval erasure_coding.Interval) (data []byte, err error) {
|
2019-05-27 18:59:03 +00:00
|
|
|
shardId, actualOffset := interval.ToShardIdAndOffset(erasure_coding.ErasureCodingLargeBlockSize, erasure_coding.ErasureCodingSmallBlockSize)
|
2019-05-29 04:29:07 +00:00
|
|
|
data = make([]byte, interval.Size)
|
2019-05-28 04:40:51 +00:00
|
|
|
if shard, found := ecVolume.FindEcVolumeShard(shardId); found {
|
2019-05-27 18:59:03 +00:00
|
|
|
if _, err = shard.ReadAt(data, actualOffset); err != nil {
|
2019-05-28 07:51:01 +00:00
|
|
|
glog.V(0).Infof("read local ec shard %d.%d: %v", ecVolume.VolumeId, shardId, err)
|
2019-05-27 18:59:03 +00:00
|
|
|
return
|
|
|
|
}
|
|
|
|
} else {
|
2019-05-28 05:54:58 +00:00
|
|
|
ecVolume.ShardLocationsLock.RLock()
|
|
|
|
sourceDataNodes, found := ecVolume.ShardLocations[shardId]
|
|
|
|
ecVolume.ShardLocationsLock.RUnlock()
|
|
|
|
if !found || len(sourceDataNodes) == 0 {
|
|
|
|
return nil, fmt.Errorf("failed to find ec shard %d.%d", ecVolume.VolumeId, shardId)
|
|
|
|
}
|
2019-05-29 06:48:39 +00:00
|
|
|
|
|
|
|
// try reading directly
|
|
|
|
_, err = s.readRemoteEcShardInterval(ctx, sourceDataNodes, ecVolume.VolumeId, shardId, data, actualOffset)
|
|
|
|
if err == nil {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
// try reading by recovering from other shards
|
|
|
|
_, err = s.recoverOneRemoteEcShardInterval(ctx, ecVolume, shardId, data, actualOffset)
|
|
|
|
if err == nil {
|
|
|
|
return
|
2019-05-28 04:22:23 +00:00
|
|
|
}
|
2019-05-29 06:48:39 +00:00
|
|
|
glog.V(0).Infof("recover ec shard %d.%d : %v", ecVolume.VolumeId, shardId, err)
|
2019-05-27 18:59:03 +00:00
|
|
|
}
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
2019-05-28 05:54:58 +00:00
|
|
|
func (s *Store) cachedLookupEcShardLocations(ctx context.Context, ecVolume *erasure_coding.EcVolume) (err error) {
|
|
|
|
|
|
|
|
if ecVolume.ShardLocationsRefreshTime.Add(10 * time.Minute).After(time.Now()) {
|
|
|
|
// still fresh
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2019-05-28 07:51:01 +00:00
|
|
|
glog.V(3).Infof("lookup and cache ec volume %d locations", ecVolume.VolumeId)
|
2019-05-28 05:54:58 +00:00
|
|
|
|
|
|
|
err = operation.WithMasterServerClient(s.MasterAddress, s.grpcDialOption, func(masterClient master_pb.SeaweedClient) error {
|
2019-05-28 07:13:13 +00:00
|
|
|
req := &master_pb.LookupEcVolumeRequest{
|
|
|
|
VolumeId: uint32(ecVolume.VolumeId),
|
|
|
|
}
|
|
|
|
resp, err := masterClient.LookupEcVolume(ctx, req)
|
|
|
|
if err != nil {
|
|
|
|
return fmt.Errorf("lookup ec volume %d: %v", ecVolume.VolumeId, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
ecVolume.ShardLocationsLock.Lock()
|
|
|
|
for _, shardIdLocations := range resp.ShardIdLocations {
|
|
|
|
shardId := erasure_coding.ShardId(shardIdLocations.ShardId)
|
2019-05-28 07:51:01 +00:00
|
|
|
delete(ecVolume.ShardLocations, shardId)
|
2019-05-28 07:13:13 +00:00
|
|
|
for _, loc := range shardIdLocations.Locations {
|
|
|
|
ecVolume.ShardLocations[shardId] = append(ecVolume.ShardLocations[shardId], loc.Url)
|
|
|
|
}
|
|
|
|
}
|
2019-05-29 04:29:07 +00:00
|
|
|
ecVolume.ShardLocationsRefreshTime = time.Now()
|
2019-05-28 07:13:13 +00:00
|
|
|
ecVolume.ShardLocationsLock.Unlock()
|
|
|
|
|
2019-05-28 05:54:58 +00:00
|
|
|
return nil
|
|
|
|
})
|
2019-05-28 04:22:23 +00:00
|
|
|
return
|
|
|
|
}
|
|
|
|
|
2019-05-29 06:48:39 +00:00
|
|
|
func (s *Store) readRemoteEcShardInterval(ctx context.Context, sourceDataNodes []string, vid needle.VolumeId, shardId erasure_coding.ShardId, buf []byte, offset int64) (n int, err error) {
|
|
|
|
|
|
|
|
for _, sourceDataNode := range sourceDataNodes {
|
|
|
|
glog.V(4).Infof("read remote ec shard %d.%d from %s", vid, shardId, sourceDataNode)
|
|
|
|
n, err = s.doReadRemoteEcShardInterval(ctx, sourceDataNode, vid, shardId, buf, offset)
|
|
|
|
if err == nil {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
glog.V(1).Infof("read remote ec shard %d.%d from %s: %v", vid, shardId, sourceDataNode, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
func (s *Store) doReadRemoteEcShardInterval(ctx context.Context, sourceDataNode string, vid needle.VolumeId, shardId erasure_coding.ShardId, buf []byte, offset int64) (n int, err error) {
|
2019-05-28 04:22:23 +00:00
|
|
|
|
|
|
|
err = operation.WithVolumeServerClient(sourceDataNode, s.grpcDialOption, func(client volume_server_pb.VolumeServerClient) error {
|
|
|
|
|
|
|
|
// copy data slice
|
|
|
|
shardReadClient, err := client.VolumeEcShardRead(ctx, &volume_server_pb.VolumeEcShardReadRequest{
|
|
|
|
VolumeId: uint32(vid),
|
|
|
|
ShardId: uint32(shardId),
|
|
|
|
Offset: offset,
|
|
|
|
Size: int64(len(buf)),
|
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
return fmt.Errorf("failed to start reading ec shard %d.%d from %s: %v", vid, shardId, sourceDataNode, err)
|
|
|
|
}
|
2019-05-27 18:59:03 +00:00
|
|
|
|
2019-05-28 04:22:23 +00:00
|
|
|
for {
|
|
|
|
resp, receiveErr := shardReadClient.Recv()
|
|
|
|
if receiveErr == io.EOF {
|
|
|
|
break
|
|
|
|
}
|
|
|
|
if receiveErr != nil {
|
|
|
|
return fmt.Errorf("receiving ec shard %d.%d from %s: %v", vid, shardId, sourceDataNode, err)
|
|
|
|
}
|
|
|
|
copy(buf[n:n+len(resp.Data)], resp.Data)
|
|
|
|
n += len(resp.Data)
|
|
|
|
}
|
2019-05-27 18:59:03 +00:00
|
|
|
|
2019-05-28 04:22:23 +00:00
|
|
|
return nil
|
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
return 0, fmt.Errorf("read ec shard %d.%d from %s: %v", vid, shardId, sourceDataNode, err)
|
|
|
|
}
|
2019-05-27 18:59:03 +00:00
|
|
|
|
|
|
|
return
|
|
|
|
}
|
2019-05-28 04:22:23 +00:00
|
|
|
|
2019-05-29 06:48:39 +00:00
|
|
|
func (s *Store) recoverOneRemoteEcShardInterval(ctx context.Context, ecVolume *erasure_coding.EcVolume, shardIdToRecover erasure_coding.ShardId, buf []byte, offset int64) (n int, err error) {
|
|
|
|
glog.V(1).Infof("recover ec shard %d.%d from other locations", ecVolume.VolumeId, shardIdToRecover)
|
|
|
|
|
|
|
|
enc, err := reedsolomon.New(erasure_coding.DataShardsCount, erasure_coding.ParityShardsCount)
|
|
|
|
if err != nil {
|
|
|
|
return 0, fmt.Errorf("failed to create encoder: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
bufs := make([][]byte, erasure_coding.TotalShardsCount)
|
|
|
|
|
|
|
|
var wg sync.WaitGroup
|
|
|
|
ecVolume.ShardLocationsLock.RLock()
|
|
|
|
for shardId, locations := range ecVolume.ShardLocations {
|
|
|
|
|
|
|
|
// skip currnent shard or empty shard
|
|
|
|
if shardId == shardIdToRecover {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
if len(locations) == 0 {
|
|
|
|
glog.V(3).Infof("readRemoteEcShardInterval missing %d.%d from %+v", ecVolume.VolumeId, shardId, locations)
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
|
|
|
|
// read from remote locations
|
|
|
|
wg.Add(1)
|
|
|
|
go func(shardId erasure_coding.ShardId, locations []string) {
|
|
|
|
defer wg.Done()
|
|
|
|
data := make([]byte, len(buf))
|
|
|
|
n, err = s.readRemoteEcShardInterval(ctx, locations, ecVolume.VolumeId, shardId, data, offset)
|
|
|
|
if err != nil {
|
|
|
|
glog.V(3).Infof("readRemoteEcShardInterval %d.%d from %+v", ecVolume.VolumeId, shardId, locations)
|
|
|
|
}
|
|
|
|
if n == len(buf) {
|
|
|
|
bufs[shardId] = data
|
|
|
|
return
|
|
|
|
}
|
|
|
|
}(shardId, locations)
|
|
|
|
}
|
|
|
|
ecVolume.ShardLocationsLock.RUnlock()
|
|
|
|
|
|
|
|
wg.Wait()
|
|
|
|
|
|
|
|
if err = enc.ReconstructData(bufs); err != nil {
|
|
|
|
return 0, err
|
|
|
|
}
|
|
|
|
glog.V(3).Infof("recovered ec shard %d.%d from other locations", ecVolume.VolumeId, shardIdToRecover)
|
|
|
|
|
|
|
|
copy(buf, bufs[shardIdToRecover])
|
|
|
|
|
|
|
|
return len(buf), nil
|
2019-05-28 04:22:23 +00:00
|
|
|
}
|