seaweedfs/weed/topology/topology.go

244 lines
6.7 KiB
Go
Raw Normal View History

2012-08-24 05:33:37 +00:00
package topology
import (
"errors"
2019-04-21 17:14:17 +00:00
"fmt"
"math/rand"
2019-05-24 06:34:29 +00:00
"sync"
2015-05-03 19:37:49 +00:00
"github.com/chrislusf/raft"
"github.com/chrislusf/seaweedfs/weed/glog"
"github.com/chrislusf/seaweedfs/weed/pb/master_pb"
"github.com/chrislusf/seaweedfs/weed/sequence"
"github.com/chrislusf/seaweedfs/weed/storage"
2019-04-19 04:43:36 +00:00
"github.com/chrislusf/seaweedfs/weed/storage/needle"
"github.com/chrislusf/seaweedfs/weed/util"
2012-08-24 05:33:37 +00:00
)
type Topology struct {
NodeImpl
2019-05-24 06:34:29 +00:00
collectionMap *util.ConcurrentReadMap
ecShardMap map[needle.VolumeId]*EcShardLocations
ecShardMapLock sync.RWMutex
pulse int64
volumeSizeLimit uint64
Sequence sequence.Sequencer
2012-09-18 21:05:12 +00:00
2017-01-10 09:01:12 +00:00
chanFullVolumes chan storage.VolumeInfo
2017-01-10 09:01:12 +00:00
Configuration *Configuration
RaftServer raft.Server
2012-08-31 08:35:11 +00:00
}
2017-02-13 05:58:44 +00:00
func NewTopology(id string, seq sequence.Sequencer, volumeSizeLimit uint64, pulse int) *Topology {
t := &Topology{}
t.id = NodeId(id)
t.nodeType = "Topology"
t.NodeImpl.value = t
t.children = make(map[NodeId]Node)
t.collectionMap = util.NewConcurrentReadMap()
2019-05-24 06:34:29 +00:00
t.ecShardMap = make(map[needle.VolumeId]*EcShardLocations)
t.pulse = int64(pulse)
t.volumeSizeLimit = volumeSizeLimit
t.Sequence = seq
t.chanFullVolumes = make(chan storage.VolumeInfo)
2017-02-13 05:58:44 +00:00
t.Configuration = &Configuration{}
2017-02-13 05:58:44 +00:00
return t
}
func (t *Topology) IsLeader() bool {
2019-02-15 08:09:48 +00:00
if t.RaftServer != nil {
2019-01-28 18:36:16 +00:00
return t.RaftServer.State() == raft.Leader
}
return false
}
func (t *Topology) Leader() (string, error) {
l := ""
if t.RaftServer != nil {
l = t.RaftServer.Leader()
} else {
return "", errors.New("Raft Server not ready yet!")
}
if l == "" {
// We are a single node cluster, we are the leader
return t.RaftServer.Name(), errors.New("Raft Server not initialized!")
}
return l, nil
}
2019-06-06 06:20:26 +00:00
func (t *Topology) Lookup(collection string, vid needle.VolumeId) (dataNodes []*DataNode) {
2013-11-12 10:21:22 +00:00
//maybe an issue if lots of collections?
if collection == "" {
2016-05-30 19:30:26 +00:00
for _, c := range t.collectionMap.Items() {
if list := c.(*Collection).Lookup(vid); list != nil {
return list
}
}
2013-11-12 10:21:22 +00:00
} else {
2016-05-30 19:30:26 +00:00
if c, ok := t.collectionMap.Find(collection); ok {
return c.(*Collection).Lookup(vid)
2013-11-12 10:21:22 +00:00
}
}
2019-06-06 06:20:26 +00:00
if locations, found := t.LookupEcShards(vid); found {
for _, loc := range locations.Locations {
dataNodes = append(dataNodes, loc...)
}
return dataNodes
}
return nil
}
2019-04-19 04:43:36 +00:00
func (t *Topology) NextVolumeId() (needle.VolumeId, error) {
vid := t.GetMaxVolumeId()
next := vid.Next()
if _, err := t.RaftServer.Do(NewMaxVolumeIdCommand(next)); err != nil {
return 0, err
}
return next, nil
}
2014-10-26 06:45:31 +00:00
func (t *Topology) HasWritableVolume(option *VolumeGrowOption) bool {
vl := t.GetVolumeLayout(option.Collection, option.ReplicaPlacement, option.Ttl)
return vl.GetActiveVolumeCount(option) > 0
}
func (t *Topology) PickForWrite(count uint64, option *VolumeGrowOption) (string, uint64, *DataNode, error) {
vid, count, datanodes, err := t.GetVolumeLayout(option.Collection, option.ReplicaPlacement, option.Ttl).PickForWrite(count, option)
2019-04-21 17:14:17 +00:00
if err != nil {
return "", 0, nil, fmt.Errorf("failed to find writable volumes for collectio:%s replication:%s ttl:%s error: %v", option.Collection, option.ReplicaPlacement.String(), option.Ttl.String(), err)
}
if datanodes.Length() == 0 {
return "", 0, nil, fmt.Errorf("no writable volumes available for for collectio:%s replication:%s ttl:%s", option.Collection, option.ReplicaPlacement.String(), option.Ttl.String())
}
fileId, count := t.Sequence.NextFileId(count)
2019-04-19 04:43:36 +00:00
return needle.NewFileId(*vid, fileId, rand.Uint32()).String(), count, datanodes.Head(), nil
}
2019-04-19 04:43:36 +00:00
func (t *Topology) GetVolumeLayout(collectionName string, rp *storage.ReplicaPlacement, ttl *needle.TTL) *VolumeLayout {
return t.collectionMap.Get(collectionName, func() interface{} {
return NewCollection(collectionName, t.volumeSizeLimit)
}).(*Collection).GetOrCreateVolumeLayout(rp, ttl)
2012-09-14 08:17:13 +00:00
}
2019-05-30 16:27:23 +00:00
func (t *Topology) ListCollections(includeNormalVolumes, includeEcVolumes bool) (ret []string) {
2019-05-30 16:17:58 +00:00
mapOfCollections := make(map[string]bool)
for _, c := range t.collectionMap.Items() {
2019-05-30 16:17:58 +00:00
mapOfCollections[c.(*Collection).Name] = true
}
2019-05-30 16:27:23 +00:00
if includeEcVolumes {
t.ecShardMapLock.RLock()
for _, ecVolumeLocation := range t.ecShardMap {
mapOfCollections[ecVolumeLocation.Collection] = true
}
t.ecShardMapLock.RUnlock()
2019-05-30 16:17:58 +00:00
}
for k, _ := range mapOfCollections {
ret = append(ret, k)
}
return ret
}
2016-05-30 19:30:26 +00:00
func (t *Topology) FindCollection(collectionName string) (*Collection, bool) {
c, hasCollection := t.collectionMap.Find(collectionName)
if !hasCollection {
return nil, false
}
return c.(*Collection), hasCollection
}
func (t *Topology) DeleteCollection(collectionName string) {
2016-05-30 19:30:26 +00:00
t.collectionMap.Delete(collectionName)
}
func (t *Topology) RegisterVolumeLayout(v storage.VolumeInfo, dn *DataNode) {
t.GetVolumeLayout(v.Collection, v.ReplicaPlacement, v.Ttl).RegisterVolume(&v, dn)
}
func (t *Topology) UnRegisterVolumeLayout(v storage.VolumeInfo, dn *DataNode) {
glog.Infof("removing volume info:%+v", v)
volumeLayout := t.GetVolumeLayout(v.Collection, v.ReplicaPlacement, v.Ttl)
volumeLayout.UnRegisterVolume(&v, dn)
if volumeLayout.isEmpty() {
t.DeleteCollection(v.Collection)
}
}
func (t *Topology) GetOrCreateDataCenter(dcName string) *DataCenter {
for _, c := range t.Children() {
dc := c.(*DataCenter)
if string(dc.Id()) == dcName {
return dc
}
}
dc := NewDataCenter(dcName)
t.LinkChildNode(dc)
return dc
}
2018-06-25 07:01:53 +00:00
func (t *Topology) SyncDataNodeRegistration(volumes []*master_pb.VolumeInformationMessage, dn *DataNode) (newVolumes, deletedVolumes []storage.VolumeInfo) {
2019-05-23 07:04:24 +00:00
// convert into in memory struct storage.VolumeInfo
2018-06-25 07:01:53 +00:00
var volumeInfos []storage.VolumeInfo
for _, v := range volumes {
if vi, err := storage.NewVolumeInfo(v); err == nil {
volumeInfos = append(volumeInfos, vi)
} else {
glog.V(0).Infof("Fail to convert joined volume information: %v", err)
}
}
2019-05-23 07:04:24 +00:00
// find out the delta volumes
newVolumes, deletedVolumes = dn.UpdateVolumes(volumeInfos)
2019-04-20 18:35:20 +00:00
for _, v := range newVolumes {
2018-06-25 07:01:53 +00:00
t.RegisterVolumeLayout(v, dn)
}
for _, v := range deletedVolumes {
t.UnRegisterVolumeLayout(v, dn)
}
return
2018-06-25 07:01:53 +00:00
}
2019-04-20 18:35:20 +00:00
func (t *Topology) IncrementalSyncDataNodeRegistration(newVolumes, deletedVolumes []*master_pb.VolumeShortInformationMessage, dn *DataNode) {
var newVis, oldVis []storage.VolumeInfo
for _, v := range newVolumes {
vi, err := storage.NewVolumeInfoFromShort(v)
if err != nil {
glog.V(0).Infof("NewVolumeInfoFromShort %v: %v", v, err)
continue
}
newVis = append(newVis, vi)
}
for _, v := range deletedVolumes {
vi, err := storage.NewVolumeInfoFromShort(v)
if err != nil {
glog.V(0).Infof("NewVolumeInfoFromShort %v: %v", v, err)
continue
}
oldVis = append(oldVis, vi)
}
dn.DeltaUpdateVolumes(newVis, oldVis)
2019-04-21 06:53:37 +00:00
for _, vi := range newVis {
t.RegisterVolumeLayout(vi, dn)
}
for _, vi := range oldVis {
t.UnRegisterVolumeLayout(vi, dn)
}
2019-04-20 18:35:20 +00:00
return
}