2019-11-27 11:09:42 +00:00
|
|
|
package shell
|
|
|
|
|
|
|
|
import (
|
|
|
|
"context"
|
|
|
|
"flag"
|
|
|
|
"fmt"
|
|
|
|
"io"
|
|
|
|
"time"
|
|
|
|
|
2019-12-25 17:53:13 +00:00
|
|
|
"google.golang.org/grpc"
|
|
|
|
|
2019-11-27 11:09:42 +00:00
|
|
|
"github.com/chrislusf/seaweedfs/weed/operation"
|
|
|
|
"github.com/chrislusf/seaweedfs/weed/pb/volume_server_pb"
|
|
|
|
"github.com/chrislusf/seaweedfs/weed/storage/needle"
|
|
|
|
)
|
|
|
|
|
|
|
|
func init() {
|
2019-12-25 17:53:13 +00:00
|
|
|
Commands = append(Commands, &commandVolumeTierUpload{})
|
2019-11-27 11:09:42 +00:00
|
|
|
}
|
|
|
|
|
2019-12-25 17:53:13 +00:00
|
|
|
type commandVolumeTierUpload struct {
|
2019-11-27 11:09:42 +00:00
|
|
|
}
|
|
|
|
|
2019-12-25 17:53:13 +00:00
|
|
|
func (c *commandVolumeTierUpload) Name() string {
|
2019-12-12 07:13:04 +00:00
|
|
|
return "volume.tier.upload"
|
2019-11-27 11:09:42 +00:00
|
|
|
}
|
|
|
|
|
2019-12-25 17:53:13 +00:00
|
|
|
func (c *commandVolumeTierUpload) Help() string {
|
2020-01-01 20:37:38 +00:00
|
|
|
return `upload the dat file of a volume to a remote tier
|
2019-11-27 11:09:42 +00:00
|
|
|
|
2019-12-12 07:13:04 +00:00
|
|
|
volume.tier.upload [-collection=""] [-fullPercent=95] [-quietFor=1h]
|
|
|
|
volume.tier.upload [-collection=""] -volumeId=<volume_id> -dest=<storage_backend> [-keepLocalDatFile]
|
2019-11-27 11:09:42 +00:00
|
|
|
|
2019-12-02 23:08:28 +00:00
|
|
|
e.g.:
|
2019-12-12 07:13:04 +00:00
|
|
|
volume.tier.upload -volumeId=7 -dest=s3
|
|
|
|
volume.tier.upload -volumeId=7 -dest=s3.default
|
2019-12-02 23:08:28 +00:00
|
|
|
|
|
|
|
The <storage_backend> is defined in master.toml.
|
|
|
|
For example, "s3.default" in [storage.backend.s3.default]
|
|
|
|
|
|
|
|
This command will move the dat file of a volume to a remote tier.
|
|
|
|
|
|
|
|
SeaweedFS enables scalable and fast local access to lots of files,
|
|
|
|
and the cloud storage is slower by cost efficient. How to combine them together?
|
|
|
|
|
|
|
|
Usually the data follows 80/20 rule: only 20% of data is frequently accessed.
|
|
|
|
We can offload the old volumes to the cloud.
|
|
|
|
|
|
|
|
With this, SeaweedFS can be both fast and scalable, and infinite storage space.
|
|
|
|
Just add more local SeaweedFS volume servers to increase the throughput.
|
|
|
|
|
|
|
|
The index file is still local, and the same O(1) disk read is applied to the remote file.
|
2019-11-27 11:09:42 +00:00
|
|
|
|
|
|
|
`
|
|
|
|
}
|
|
|
|
|
2019-12-25 17:53:13 +00:00
|
|
|
func (c *commandVolumeTierUpload) Do(args []string, commandEnv *CommandEnv, writer io.Writer) (err error) {
|
2019-11-27 11:09:42 +00:00
|
|
|
|
2020-04-23 20:37:31 +00:00
|
|
|
if err = commandEnv.confirmIsLocked(); err != nil {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
2019-11-27 11:09:42 +00:00
|
|
|
tierCommand := flag.NewFlagSet(c.Name(), flag.ContinueOnError)
|
|
|
|
volumeId := tierCommand.Int("volumeId", 0, "the volume id")
|
|
|
|
collection := tierCommand.String("collection", "", "the collection name")
|
|
|
|
fullPercentage := tierCommand.Float64("fullPercent", 95, "the volume reaches the percentage of max volume size")
|
|
|
|
quietPeriod := tierCommand.Duration("quietFor", 24*time.Hour, "select volumes without no writes for this period")
|
2019-12-02 23:08:28 +00:00
|
|
|
dest := tierCommand.String("dest", "", "the target tier name")
|
|
|
|
keepLocalDatFile := tierCommand.Bool("keepLocalDatFile", false, "whether keep local dat file")
|
2019-11-27 11:09:42 +00:00
|
|
|
if err = tierCommand.Parse(args); err != nil {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
vid := needle.VolumeId(*volumeId)
|
|
|
|
|
|
|
|
// volumeId is provided
|
|
|
|
if vid != 0 {
|
2020-02-26 05:50:12 +00:00
|
|
|
return doVolumeTierUpload(commandEnv, writer, *collection, vid, *dest, *keepLocalDatFile)
|
2019-11-27 11:09:42 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// apply to all volumes in the collection
|
|
|
|
// reusing collectVolumeIdsForEcEncode for now
|
2020-02-26 05:50:12 +00:00
|
|
|
volumeIds, err := collectVolumeIdsForEcEncode(commandEnv, *collection, *fullPercentage, *quietPeriod)
|
2019-11-27 11:09:42 +00:00
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
2019-12-25 17:53:13 +00:00
|
|
|
fmt.Printf("tier upload volumes: %v\n", volumeIds)
|
2019-11-27 11:09:42 +00:00
|
|
|
for _, vid := range volumeIds {
|
2020-02-26 05:50:12 +00:00
|
|
|
if err = doVolumeTierUpload(commandEnv, writer, *collection, vid, *dest, *keepLocalDatFile); err != nil {
|
2019-11-27 11:09:42 +00:00
|
|
|
return err
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2020-02-26 05:50:12 +00:00
|
|
|
func doVolumeTierUpload(commandEnv *CommandEnv, writer io.Writer, collection string, vid needle.VolumeId, dest string, keepLocalDatFile bool) (err error) {
|
2019-11-27 11:09:42 +00:00
|
|
|
// find volume location
|
|
|
|
locations, found := commandEnv.MasterClient.GetLocations(uint32(vid))
|
|
|
|
if !found {
|
|
|
|
return fmt.Errorf("volume %d not found", vid)
|
|
|
|
}
|
|
|
|
|
2020-02-26 05:50:12 +00:00
|
|
|
err = markVolumeReadonly(commandEnv.option.GrpcDialOption, needle.VolumeId(vid), locations)
|
2019-12-25 17:53:13 +00:00
|
|
|
if err != nil {
|
|
|
|
return fmt.Errorf("mark volume %d as readonly on %s: %v", vid, locations[0].Url, err)
|
|
|
|
}
|
2019-11-27 11:09:42 +00:00
|
|
|
|
|
|
|
// copy the .dat file to remote tier
|
2020-02-26 05:50:12 +00:00
|
|
|
err = uploadDatToRemoteTier(commandEnv.option.GrpcDialOption, writer, needle.VolumeId(vid), collection, locations[0].Url, dest, keepLocalDatFile)
|
2019-11-27 11:09:42 +00:00
|
|
|
if err != nil {
|
|
|
|
return fmt.Errorf("copy dat file for volume %d on %s to %s: %v", vid, locations[0].Url, dest, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2020-02-26 05:50:12 +00:00
|
|
|
func uploadDatToRemoteTier(grpcDialOption grpc.DialOption, writer io.Writer, volumeId needle.VolumeId, collection string, sourceVolumeServer string, dest string, keepLocalDatFile bool) error {
|
2019-11-27 11:09:42 +00:00
|
|
|
|
2020-02-26 05:50:12 +00:00
|
|
|
err := operation.WithVolumeServerClient(sourceVolumeServer, grpcDialOption, func(volumeServerClient volume_server_pb.VolumeServerClient) error {
|
|
|
|
stream, copyErr := volumeServerClient.VolumeTierMoveDatToRemote(context.Background(), &volume_server_pb.VolumeTierMoveDatToRemoteRequest{
|
2019-12-02 23:08:28 +00:00
|
|
|
VolumeId: uint32(volumeId),
|
|
|
|
Collection: collection,
|
|
|
|
DestinationBackendName: dest,
|
|
|
|
KeepLocalDatFile: keepLocalDatFile,
|
2019-11-27 11:09:42 +00:00
|
|
|
})
|
2019-12-02 23:08:28 +00:00
|
|
|
|
|
|
|
var lastProcessed int64
|
|
|
|
for {
|
|
|
|
resp, recvErr := stream.Recv()
|
|
|
|
if recvErr != nil {
|
|
|
|
if recvErr == io.EOF {
|
|
|
|
break
|
|
|
|
} else {
|
|
|
|
return recvErr
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2019-12-03 04:49:58 +00:00
|
|
|
processingSpeed := float64(resp.Processed-lastProcessed) / 1024.0 / 1024.0
|
2019-12-02 23:08:28 +00:00
|
|
|
|
|
|
|
fmt.Fprintf(writer, "copied %.2f%%, %d bytes, %.2fMB/s\n", resp.ProcessedPercentage, resp.Processed, processingSpeed)
|
|
|
|
|
|
|
|
lastProcessed = resp.Processed
|
|
|
|
}
|
|
|
|
|
2019-11-27 11:09:42 +00:00
|
|
|
return copyErr
|
|
|
|
})
|
|
|
|
|
|
|
|
return err
|
|
|
|
|
|
|
|
}
|