mirror of
https://github.com/seaweedfs/seaweedfs.git
synced 2024-01-19 02:48:24 +00:00
fix tests
This commit is contained in:
parent
7403cd43c5
commit
68775d29e3
|
@ -127,6 +127,7 @@ func newEcNode(dc string, rack string, dataNodeId string, freeEcSlot int) *EcNod
|
||||||
return &EcNode{
|
return &EcNode{
|
||||||
info: &master_pb.DataNodeInfo{
|
info: &master_pb.DataNodeInfo{
|
||||||
Id: dataNodeId,
|
Id: dataNodeId,
|
||||||
|
DiskInfos: make(map[string]*master_pb.DiskInfo),
|
||||||
},
|
},
|
||||||
dc: dc,
|
dc: dc,
|
||||||
rack: RackId(rack),
|
rack: RackId(rack),
|
||||||
|
|
|
@ -14,11 +14,11 @@ import (
|
||||||
func TestRemoveDataCenter(t *testing.T) {
|
func TestRemoveDataCenter(t *testing.T) {
|
||||||
topo := setup(topologyLayout)
|
topo := setup(topologyLayout)
|
||||||
topo.UnlinkChildNode(NodeId("dc2"))
|
topo.UnlinkChildNode(NodeId("dc2"))
|
||||||
if topo.GetActiveVolumeCount() != 15 {
|
if topo.diskUsages.usages[types.HardDriveType].activeVolumeCount != 15 {
|
||||||
t.Fail()
|
t.Fail()
|
||||||
}
|
}
|
||||||
topo.UnlinkChildNode(NodeId("dc3"))
|
topo.UnlinkChildNode(NodeId("dc3"))
|
||||||
if topo.GetActiveVolumeCount() != 12 {
|
if topo.diskUsages.usages[types.HardDriveType].activeVolumeCount != 12 {
|
||||||
t.Fail()
|
t.Fail()
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -28,7 +28,10 @@ func TestHandlingVolumeServerHeartbeat(t *testing.T) {
|
||||||
|
|
||||||
dc := topo.GetOrCreateDataCenter("dc1")
|
dc := topo.GetOrCreateDataCenter("dc1")
|
||||||
rack := dc.GetOrCreateRack("rack1")
|
rack := dc.GetOrCreateRack("rack1")
|
||||||
dn := rack.GetOrCreateDataNode("127.0.0.1", 34534, "127.0.0.1", 25, 12)
|
maxVolumeCounts := make(map[string]uint32)
|
||||||
|
maxVolumeCounts[""] = 25
|
||||||
|
maxVolumeCounts["ssd"] = 12
|
||||||
|
dn := rack.GetOrCreateDataNode("127.0.0.1", 34534, "127.0.0.1", maxVolumeCounts)
|
||||||
|
|
||||||
{
|
{
|
||||||
volumeCount := 7
|
volumeCount := 7
|
||||||
|
@ -68,9 +71,11 @@ func TestHandlingVolumeServerHeartbeat(t *testing.T) {
|
||||||
|
|
||||||
topo.SyncDataNodeRegistration(volumeMessages, dn)
|
topo.SyncDataNodeRegistration(volumeMessages, dn)
|
||||||
|
|
||||||
assert(t, "activeVolumeCount1", int(topo.activeVolumeCount), volumeCount*2)
|
usageCounts := topo.diskUsages.usages[types.HardDriveType]
|
||||||
assert(t, "volumeCount", int(topo.volumeCount), volumeCount)
|
|
||||||
assert(t, "ssdVolumeCount", int(topo.ssdVolumeCount), volumeCount)
|
assert(t, "activeVolumeCount1", int(usageCounts.activeVolumeCount), volumeCount)
|
||||||
|
assert(t, "volumeCount", int(usageCounts.volumeCount), volumeCount)
|
||||||
|
assert(t, "ssdVolumeCount", int(topo.diskUsages.usages[types.SsdType].volumeCount), volumeCount)
|
||||||
}
|
}
|
||||||
|
|
||||||
{
|
{
|
||||||
|
@ -97,8 +102,10 @@ func TestHandlingVolumeServerHeartbeat(t *testing.T) {
|
||||||
//layout := topo.GetVolumeLayout("", rp, needle.EMPTY_TTL)
|
//layout := topo.GetVolumeLayout("", rp, needle.EMPTY_TTL)
|
||||||
//assert(t, "writables", len(layout.writables), volumeCount)
|
//assert(t, "writables", len(layout.writables), volumeCount)
|
||||||
|
|
||||||
assert(t, "activeVolumeCount1", int(topo.activeVolumeCount), volumeCount)
|
usageCounts := topo.diskUsages.usages[types.HardDriveType]
|
||||||
assert(t, "volumeCount", int(topo.volumeCount), volumeCount)
|
|
||||||
|
assert(t, "activeVolumeCount1", int(usageCounts.activeVolumeCount), volumeCount)
|
||||||
|
assert(t, "volumeCount", int(usageCounts.volumeCount), volumeCount)
|
||||||
}
|
}
|
||||||
|
|
||||||
{
|
{
|
||||||
|
@ -118,16 +125,18 @@ func TestHandlingVolumeServerHeartbeat(t *testing.T) {
|
||||||
layout := topo.GetVolumeLayout("", rp, needle.EMPTY_TTL, types.HardDriveType)
|
layout := topo.GetVolumeLayout("", rp, needle.EMPTY_TTL, types.HardDriveType)
|
||||||
assert(t, "writables after repeated add", len(layout.writables), volumeCount)
|
assert(t, "writables after repeated add", len(layout.writables), volumeCount)
|
||||||
|
|
||||||
assert(t, "activeVolumeCount1", int(topo.activeVolumeCount), volumeCount)
|
usageCounts := topo.diskUsages.usages[types.HardDriveType]
|
||||||
assert(t, "volumeCount", int(topo.volumeCount), volumeCount)
|
|
||||||
|
assert(t, "activeVolumeCount1", int(usageCounts.activeVolumeCount), volumeCount)
|
||||||
|
assert(t, "volumeCount", int(usageCounts.volumeCount), volumeCount)
|
||||||
|
|
||||||
topo.IncrementalSyncDataNodeRegistration(
|
topo.IncrementalSyncDataNodeRegistration(
|
||||||
nil,
|
nil,
|
||||||
[]*master_pb.VolumeShortInformationMessage{newVolumeShortMessage},
|
[]*master_pb.VolumeShortInformationMessage{newVolumeShortMessage},
|
||||||
dn)
|
dn)
|
||||||
assert(t, "writables after deletion", len(layout.writables), volumeCount-1)
|
assert(t, "writables after deletion", len(layout.writables), volumeCount-1)
|
||||||
assert(t, "activeVolumeCount1", int(topo.activeVolumeCount), volumeCount-1)
|
assert(t, "activeVolumeCount1", int(usageCounts.activeVolumeCount), volumeCount-1)
|
||||||
assert(t, "volumeCount", int(topo.volumeCount), volumeCount-1)
|
assert(t, "volumeCount", int(usageCounts.volumeCount), volumeCount-1)
|
||||||
|
|
||||||
topo.IncrementalSyncDataNodeRegistration(
|
topo.IncrementalSyncDataNodeRegistration(
|
||||||
[]*master_pb.VolumeShortInformationMessage{newVolumeShortMessage},
|
[]*master_pb.VolumeShortInformationMessage{newVolumeShortMessage},
|
||||||
|
@ -147,7 +156,9 @@ func TestHandlingVolumeServerHeartbeat(t *testing.T) {
|
||||||
|
|
||||||
topo.UnRegisterDataNode(dn)
|
topo.UnRegisterDataNode(dn)
|
||||||
|
|
||||||
assert(t, "activeVolumeCount2", int(topo.activeVolumeCount), 0)
|
usageCounts := topo.diskUsages.usages[types.HardDriveType]
|
||||||
|
|
||||||
|
assert(t, "activeVolumeCount2", int(usageCounts.activeVolumeCount), 0)
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -163,7 +174,10 @@ func TestAddRemoveVolume(t *testing.T) {
|
||||||
|
|
||||||
dc := topo.GetOrCreateDataCenter("dc1")
|
dc := topo.GetOrCreateDataCenter("dc1")
|
||||||
rack := dc.GetOrCreateRack("rack1")
|
rack := dc.GetOrCreateRack("rack1")
|
||||||
dn := rack.GetOrCreateDataNode("127.0.0.1", 34534, "127.0.0.1", 25, 12)
|
maxVolumeCounts := make(map[string]uint32)
|
||||||
|
maxVolumeCounts[""] = 25
|
||||||
|
maxVolumeCounts["ssd"] = 12
|
||||||
|
dn := rack.GetOrCreateDataNode("127.0.0.1", 34534, "127.0.0.1", maxVolumeCounts)
|
||||||
|
|
||||||
v := storage.VolumeInfo{
|
v := storage.VolumeInfo{
|
||||||
Id: needle.VolumeId(1),
|
Id: needle.VolumeId(1),
|
||||||
|
|
|
@ -103,7 +103,13 @@ func setup(topologyLayout string) *Topology {
|
||||||
Version: needle.CurrentVersion}
|
Version: needle.CurrentVersion}
|
||||||
server.AddOrUpdateVolume(vi)
|
server.AddOrUpdateVolume(vi)
|
||||||
}
|
}
|
||||||
server.UpAdjustMaxVolumeCountDelta(int64(serverMap["limit"].(float64)))
|
|
||||||
|
disk := server.getOrCreateDisk("")
|
||||||
|
deltaDiskUsages := newDiskUsages()
|
||||||
|
deltaDiskUsage := deltaDiskUsages.getOrCreateDisk("")
|
||||||
|
deltaDiskUsage.maxVolumeCount = int64(serverMap["limit"].(float64))
|
||||||
|
disk.UpAdjustDiskUsageDelta(deltaDiskUsages)
|
||||||
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
Loading…
Reference in a new issue