seaweedfs/weed/shell/command_fs_meta_save.go

222 lines
5.8 KiB
Go
Raw Normal View History

2019-04-16 08:06:32 +00:00
package shell
import (
"flag"
2019-04-16 08:06:32 +00:00
"fmt"
"io"
"os"
"sync"
"sync/atomic"
"time"
2019-04-16 08:06:32 +00:00
"github.com/golang/protobuf/proto"
2019-04-16 08:06:32 +00:00
"github.com/chrislusf/seaweedfs/weed/pb/filer_pb"
"github.com/chrislusf/seaweedfs/weed/util"
)
func init() {
Commands = append(Commands, &commandFsMetaSave{})
2019-04-16 08:06:32 +00:00
}
type commandFsMetaSave struct {
}
func (c *commandFsMetaSave) Name() string {
return "fs.meta.save"
}
func (c *commandFsMetaSave) Help() string {
return `save all directory and file meta data to a local file for metadata backup.
2019-04-16 08:06:32 +00:00
fs.meta.save / # save from the root
fs.meta.save -v -o t.meta / # save from the root, output to t.meta file.
fs.meta.save /path/to/save # save from the directory /path/to/save
fs.meta.save . # save from current directory
fs.meta.save # save from current directory
2019-04-16 08:06:32 +00:00
The meta data will be saved into a local <filer_host>-<port>-<time>.meta file.
These meta data can be later loaded by fs.meta.load command,
2020-03-24 06:07:11 +00:00
Another usage is to export all data chunk file ids used by the files.
fs.meta.save -chunks <filer.chunks>
The output chunks file will contain lines as:
<file key> <tab> <volumeId, fileKey, cookie> <tab> <file name>
This output chunks can be used to find out missing chunks or files.
2019-04-16 08:06:32 +00:00
`
}
func (c *commandFsMetaSave) Do(args []string, commandEnv *CommandEnv, writer io.Writer) (err error) {
2019-04-16 08:06:32 +00:00
fsMetaSaveCommand := flag.NewFlagSet(c.Name(), flag.ContinueOnError)
verbose := fsMetaSaveCommand.Bool("v", false, "print out each processed files")
outputFileName := fsMetaSaveCommand.String("o", "", "output the meta data to this file")
chunksFileName := fsMetaSaveCommand.String("chunks", "", "output all the chunks to this file")
if err = fsMetaSaveCommand.Parse(args); err != nil {
return nil
}
2020-03-24 04:26:15 +00:00
path, parseErr := commandEnv.parseUrl(findInputDirectory(fsMetaSaveCommand.Args()))
2019-12-13 08:22:37 +00:00
if parseErr != nil {
return parseErr
2019-04-16 08:06:32 +00:00
}
if *outputFileName != "" {
fileName := *outputFileName
if fileName == "" {
t := time.Now()
fileName = fmt.Sprintf("%s-%d-%4d%02d%02d-%02d%02d%02d.meta",
commandEnv.option.FilerHost, commandEnv.option.FilerPort, t.Year(), t.Month(), t.Day(), t.Hour(), t.Minute(), t.Second())
}
return doTraverseBfsAndSaving(fileName, commandEnv, writer, path, *verbose, func(dst io.Writer, outputChan chan []byte) {
sizeBuf := make([]byte, 4)
for b := range outputChan {
util.Uint32toBytes(sizeBuf, uint32(len(b)))
dst.Write(sizeBuf)
dst.Write(b)
}
}, func(entry *filer_pb.FullEntry, outputChan chan []byte) (err error) {
bytes, err := proto.Marshal(entry)
if err != nil {
fmt.Fprintf(writer, "marshall error: %v\n", err)
return
}
outputChan <- bytes
return nil
})
}
if *chunksFileName != "" {
return doTraverseBfsAndSaving(*chunksFileName, commandEnv, writer, path, *verbose, func(dst io.Writer, outputChan chan []byte) {
for b := range outputChan {
dst.Write(b)
}
}, func(entry *filer_pb.FullEntry, outputChan chan []byte) (err error) {
for _, chunk := range entry.Entry.Chunks {
dir := entry.Dir
if dir == "/" {
dir = ""
}
outputLine := fmt.Sprintf("%d\t%s\t%s/%s\n", chunk.Fid.FileKey, chunk.FileId, dir, entry.Entry.Name)
outputChan <- []byte(outputLine)
}
return nil
})
2019-12-13 08:22:37 +00:00
}
2019-04-16 08:06:32 +00:00
return err
}
func doTraverseBfsAndSaving(fileName string, commandEnv *CommandEnv, writer io.Writer, path string, verbose bool,
saveFn func(dst io.Writer, outputChan chan []byte),
genFn func(entry *filer_pb.FullEntry, outputChan chan []byte) error) error {
2019-12-13 08:22:37 +00:00
dst, openErr := os.OpenFile(fileName, os.O_WRONLY|os.O_CREATE|os.O_TRUNC, 0644)
if openErr != nil {
return fmt.Errorf("failed to create file %s: %v", fileName, openErr)
}
defer dst.Close()
2019-04-16 08:06:32 +00:00
var wg sync.WaitGroup
wg.Add(1)
outputChan := make(chan []byte, 1024)
go func() {
saveFn(dst, outputChan)
wg.Done()
}()
2019-12-13 08:22:37 +00:00
var dirCount, fileCount uint64
2019-04-16 08:06:32 +00:00
err := doTraverseBfs(writer, commandEnv, util.FullPath(path), func(parentPath util.FullPath, entry *filer_pb.Entry) {
2019-04-16 08:06:32 +00:00
2019-12-13 08:22:37 +00:00
protoMessage := &filer_pb.FullEntry{
Dir: string(parentPath),
Entry: entry,
}
2019-04-16 08:06:32 +00:00
if err := genFn(protoMessage, outputChan); err != nil {
2019-12-13 08:22:37 +00:00
fmt.Fprintf(writer, "marshall error: %v\n", err)
return
}
2019-04-16 08:06:32 +00:00
2019-12-13 08:22:37 +00:00
if entry.IsDirectory {
atomic.AddUint64(&dirCount, 1)
} else {
atomic.AddUint64(&fileCount, 1)
2019-04-16 08:06:32 +00:00
}
if verbose {
2019-12-13 08:22:37 +00:00
println(parentPath.Child(entry.Name))
}
2019-04-16 08:06:32 +00:00
})
close(outputChan)
wg.Wait()
2019-12-13 08:22:37 +00:00
if err == nil {
fmt.Fprintf(writer, "total %d directories, %d files\n", dirCount, fileCount)
2019-12-13 08:22:37 +00:00
}
return err
2019-04-16 08:06:32 +00:00
}
func doTraverseBfs(writer io.Writer, filerClient filer_pb.FilerClient, parentPath util.FullPath, fn func(parentPath util.FullPath, entry *filer_pb.Entry)) (err error) {
K := 5
var jobQueueWg sync.WaitGroup
queue := util.NewQueue()
jobQueueWg.Add(1)
queue.Enqueue(parentPath)
var isTerminating bool
for i := 0; i < K; i++ {
go func() {
for {
if isTerminating {
break
}
t := queue.Dequeue()
if t == nil {
time.Sleep(329 * time.Millisecond)
continue
}
2020-03-23 07:01:34 +00:00
dir := t.(util.FullPath)
2020-02-26 06:23:59 +00:00
processErr := processOneDirectory(writer, filerClient, dir, queue, &jobQueueWg, fn)
if processErr != nil {
err = processErr
}
jobQueueWg.Done()
}
}()
}
jobQueueWg.Wait()
isTerminating = true
return
}
2020-03-23 07:01:34 +00:00
func processOneDirectory(writer io.Writer, filerClient filer_pb.FilerClient, parentPath util.FullPath, queue *util.Queue, jobQueueWg *sync.WaitGroup, fn func(parentPath util.FullPath, entry *filer_pb.Entry)) (err error) {
2019-04-16 08:06:32 +00:00
2020-03-23 07:01:34 +00:00
return filer_pb.ReadDirAllEntries(filerClient, parentPath, "", func(entry *filer_pb.Entry, isLast bool) {
2019-04-16 08:06:32 +00:00
2019-12-13 08:22:37 +00:00
fn(parentPath, entry)
2019-04-16 08:06:32 +00:00
2019-12-13 08:22:37 +00:00
if entry.IsDirectory {
subDir := fmt.Sprintf("%s/%s", parentPath, entry.Name)
if parentPath == "/" {
subDir = "/" + entry.Name
2019-04-16 08:06:32 +00:00
}
2019-12-13 08:22:37 +00:00
jobQueueWg.Add(1)
2020-03-23 07:01:34 +00:00
queue.Enqueue(util.FullPath(subDir))
2019-04-16 08:06:32 +00:00
}
2019-12-13 08:22:37 +00:00
})
2019-04-16 08:06:32 +00:00
}