2015-03-17 18:15:37 +00:00
|
|
|
package main
|
|
|
|
|
|
|
|
import (
|
|
|
|
"flag"
|
|
|
|
"fmt"
|
|
|
|
"io"
|
|
|
|
"os"
|
|
|
|
"path"
|
|
|
|
"strconv"
|
|
|
|
|
2016-06-03 01:09:14 +00:00
|
|
|
"github.com/chrislusf/seaweedfs/weed/glog"
|
|
|
|
"github.com/chrislusf/seaweedfs/weed/storage"
|
|
|
|
"github.com/chrislusf/seaweedfs/weed/util"
|
2018-07-10 05:48:30 +00:00
|
|
|
"github.com/chrislusf/seaweedfs/weed/storage/types"
|
2015-03-17 18:15:37 +00:00
|
|
|
)
|
|
|
|
|
|
|
|
var (
|
|
|
|
fixVolumePath = flag.String("dir", "/tmp", "data directory to store files")
|
|
|
|
fixVolumeCollection = flag.String("collection", "", "the volume collection name")
|
|
|
|
fixVolumeId = flag.Int("volumeId", -1, "a volume id. The volume should already exist in the dir. The volume index file should not exist.")
|
|
|
|
)
|
|
|
|
|
|
|
|
/*
|
|
|
|
This is to resolve an one-time issue that caused inconsistency with .dat and .idx files.
|
2015-03-17 21:22:56 +00:00
|
|
|
In this case, the .dat file contains all data, but some of deletion caused incorrect offset.
|
|
|
|
The .idx has all correct offsets.
|
|
|
|
|
2015-03-17 18:15:37 +00:00
|
|
|
1. fix the .dat file, a new .dat_fixed file will be generated.
|
|
|
|
go run fix_dat.go -volumeId=9 -dir=/Users/chrislu/Downloads
|
|
|
|
2. move the original .dat and .idx files to some backup folder, and rename .dat_fixed to .dat file
|
|
|
|
mv 9.dat_fixed 9.dat
|
|
|
|
3. fix the .idx file with the "weed fix"
|
|
|
|
weed fix -volumeId=9 -dir=/Users/chrislu/Downloads
|
|
|
|
*/
|
|
|
|
func main() {
|
|
|
|
flag.Parse()
|
|
|
|
fileName := strconv.Itoa(*fixVolumeId)
|
|
|
|
if *fixVolumeCollection != "" {
|
|
|
|
fileName = *fixVolumeCollection + "_" + fileName
|
|
|
|
}
|
|
|
|
indexFile, err := os.OpenFile(path.Join(*fixVolumePath, fileName+".idx"), os.O_RDONLY, 0644)
|
|
|
|
if err != nil {
|
2018-06-24 18:37:08 +00:00
|
|
|
glog.Fatalf("Read Volume Index %v", err)
|
2015-03-17 18:15:37 +00:00
|
|
|
}
|
|
|
|
defer indexFile.Close()
|
|
|
|
datFile, err := os.OpenFile(path.Join(*fixVolumePath, fileName+".dat"), os.O_RDONLY, 0644)
|
|
|
|
if err != nil {
|
2018-06-24 18:37:08 +00:00
|
|
|
glog.Fatalf("Read Volume Data %v", err)
|
2015-03-17 18:15:37 +00:00
|
|
|
}
|
|
|
|
defer datFile.Close()
|
|
|
|
|
|
|
|
newDatFile, err := os.Create(path.Join(*fixVolumePath, fileName+".dat_fixed"))
|
|
|
|
if err != nil {
|
2018-06-24 18:37:08 +00:00
|
|
|
glog.Fatalf("Write New Volume Data %v", err)
|
2015-03-17 18:15:37 +00:00
|
|
|
}
|
|
|
|
defer newDatFile.Close()
|
|
|
|
|
2018-06-24 18:37:08 +00:00
|
|
|
superBlock, err := storage.ReadSuperBlock(datFile)
|
|
|
|
if err != nil {
|
|
|
|
glog.Fatalf("Read Volume Data superblock %v", err)
|
|
|
|
}
|
|
|
|
newDatFile.Write(superBlock.Bytes())
|
2015-03-17 18:15:37 +00:00
|
|
|
|
|
|
|
iterateEntries(datFile, indexFile, func(n *storage.Needle, offset int64) {
|
2018-08-01 06:39:56 +00:00
|
|
|
fmt.Printf("needle id=%v name=%s size=%d dataSize=%d\n", n.Id, string(n.Name), n.Size, n.DataSize)
|
2018-06-24 22:19:36 +00:00
|
|
|
s, _, e := n.Append(newDatFile, superBlock.Version())
|
2015-03-17 18:15:37 +00:00
|
|
|
fmt.Printf("size %d error %v\n", s, e)
|
|
|
|
})
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
func iterateEntries(datFile, idxFile *os.File, visitNeedle func(n *storage.Needle, offset int64)) {
|
|
|
|
// start to read index file
|
|
|
|
var readerOffset int64
|
|
|
|
bytes := make([]byte, 16)
|
|
|
|
count, _ := idxFile.ReadAt(bytes, readerOffset)
|
|
|
|
readerOffset += int64(count)
|
|
|
|
|
|
|
|
// start to read dat file
|
2018-06-24 18:37:08 +00:00
|
|
|
superblock, err := storage.ReadSuperBlock(datFile)
|
|
|
|
if err != nil {
|
|
|
|
fmt.Printf("cannot read dat file super block: %v", err)
|
|
|
|
return
|
|
|
|
}
|
|
|
|
offset := int64(superblock.BlockSize())
|
|
|
|
version := superblock.Version()
|
2015-03-17 18:15:37 +00:00
|
|
|
n, rest, err := storage.ReadNeedleHeader(datFile, version, offset)
|
|
|
|
if err != nil {
|
|
|
|
fmt.Printf("cannot read needle header: %v", err)
|
|
|
|
return
|
|
|
|
}
|
|
|
|
fmt.Printf("Needle %+v, rest %d\n", n, rest)
|
|
|
|
for n != nil && count > 0 {
|
|
|
|
// parse index file entry
|
|
|
|
key := util.BytesToUint64(bytes[0:8])
|
|
|
|
offsetFromIndex := util.BytesToUint32(bytes[8:12])
|
|
|
|
sizeFromIndex := util.BytesToUint32(bytes[12:16])
|
|
|
|
count, _ = idxFile.ReadAt(bytes, readerOffset)
|
|
|
|
readerOffset += int64(count)
|
|
|
|
|
|
|
|
if offsetFromIndex != 0 && offset != int64(offsetFromIndex)*8 {
|
|
|
|
//t := offset
|
|
|
|
offset = int64(offsetFromIndex) * 8
|
|
|
|
//fmt.Printf("Offset change %d => %d\n", t, offset)
|
|
|
|
}
|
|
|
|
|
|
|
|
fmt.Printf("key: %d offsetFromIndex %d n.Size %d sizeFromIndex:%d\n", key, offsetFromIndex, n.Size, sizeFromIndex)
|
|
|
|
|
2018-07-24 08:36:04 +00:00
|
|
|
rest = storage.NeedleBodyLength(sizeFromIndex, version)
|
2015-03-17 18:15:37 +00:00
|
|
|
|
|
|
|
func() {
|
|
|
|
defer func() {
|
|
|
|
if r := recover(); r != nil {
|
|
|
|
fmt.Println("Recovered in f", r)
|
|
|
|
}
|
|
|
|
}()
|
2018-07-10 05:48:30 +00:00
|
|
|
if err = n.ReadNeedleBody(datFile, version, offset+int64(types.NeedleEntrySize), rest); err != nil {
|
2015-03-17 18:15:37 +00:00
|
|
|
fmt.Printf("cannot read needle body: offset %d body %d %v\n", offset, rest, err)
|
|
|
|
}
|
|
|
|
}()
|
|
|
|
|
|
|
|
if n.Size <= n.DataSize {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
visitNeedle(n, offset)
|
|
|
|
|
2018-07-10 05:48:30 +00:00
|
|
|
offset += types.NeedleEntrySize + rest
|
2015-03-17 18:15:37 +00:00
|
|
|
//fmt.Printf("==> new entry offset %d\n", offset)
|
|
|
|
if n, rest, err = storage.ReadNeedleHeader(datFile, version, offset); err != nil {
|
|
|
|
if err == io.EOF {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
fmt.Printf("cannot read needle header: %v\n", err)
|
|
|
|
return
|
|
|
|
}
|
|
|
|
//fmt.Printf("new entry needle size:%d rest:%d\n", n.Size, rest)
|
|
|
|
}
|
|
|
|
|
|
|
|
}
|