2019-05-18 00:33:49 +00:00
|
|
|
package leveldb
|
|
|
|
|
|
|
|
import (
|
|
|
|
"bytes"
|
|
|
|
"context"
|
|
|
|
"crypto/md5"
|
|
|
|
"fmt"
|
|
|
|
"io"
|
2019-07-03 04:25:53 +00:00
|
|
|
"os"
|
2019-05-18 00:33:49 +00:00
|
|
|
|
|
|
|
"github.com/syndtr/goleveldb/leveldb"
|
|
|
|
"github.com/syndtr/goleveldb/leveldb/opt"
|
|
|
|
leveldb_util "github.com/syndtr/goleveldb/leveldb/util"
|
2019-12-13 08:23:05 +00:00
|
|
|
|
|
|
|
"github.com/chrislusf/seaweedfs/weed/filer2"
|
|
|
|
"github.com/chrislusf/seaweedfs/weed/glog"
|
2020-03-08 01:01:39 +00:00
|
|
|
"github.com/chrislusf/seaweedfs/weed/pb/filer_pb"
|
2019-12-13 08:23:05 +00:00
|
|
|
weed_util "github.com/chrislusf/seaweedfs/weed/util"
|
2019-05-18 00:33:49 +00:00
|
|
|
)
|
|
|
|
|
|
|
|
func init() {
|
|
|
|
filer2.Stores = append(filer2.Stores, &LevelDB2Store{})
|
|
|
|
}
|
|
|
|
|
|
|
|
type LevelDB2Store struct {
|
2019-07-22 04:51:38 +00:00
|
|
|
dbs []*leveldb.DB
|
2019-07-03 04:25:53 +00:00
|
|
|
dbCount int
|
2019-05-18 00:33:49 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
func (store *LevelDB2Store) GetName() string {
|
|
|
|
return "leveldb2"
|
|
|
|
}
|
|
|
|
|
2020-01-29 17:09:55 +00:00
|
|
|
func (store *LevelDB2Store) Initialize(configuration weed_util.Configuration, prefix string) (err error) {
|
|
|
|
dir := configuration.GetString(prefix + "dir")
|
2019-07-03 04:25:53 +00:00
|
|
|
return store.initialize(dir, 8)
|
2019-05-18 00:33:49 +00:00
|
|
|
}
|
|
|
|
|
2019-07-03 04:25:53 +00:00
|
|
|
func (store *LevelDB2Store) initialize(dir string, dbCount int) (err error) {
|
2019-05-18 00:33:49 +00:00
|
|
|
glog.Infof("filer store leveldb2 dir: %s", dir)
|
|
|
|
if err := weed_util.TestFolderWritable(dir); err != nil {
|
|
|
|
return fmt.Errorf("Check Level Folder %s Writable: %s", dir, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
opts := &opt.Options{
|
|
|
|
BlockCacheCapacity: 32 * 1024 * 1024, // default value is 8MiB
|
|
|
|
WriteBuffer: 16 * 1024 * 1024, // default value is 4MiB
|
2019-07-03 04:25:53 +00:00
|
|
|
CompactionTableSizeMultiplier: 4,
|
2019-05-18 00:33:49 +00:00
|
|
|
}
|
|
|
|
|
2019-07-22 04:51:38 +00:00
|
|
|
for d := 0; d < dbCount; d++ {
|
2019-07-03 04:25:53 +00:00
|
|
|
dbFolder := fmt.Sprintf("%s/%02d", dir, d)
|
|
|
|
os.MkdirAll(dbFolder, 0755)
|
2019-07-03 04:28:51 +00:00
|
|
|
db, dbErr := leveldb.OpenFile(dbFolder, opts)
|
|
|
|
if dbErr != nil {
|
|
|
|
glog.Errorf("filer store open dir %s: %v", dbFolder, dbErr)
|
2019-07-03 04:25:53 +00:00
|
|
|
return
|
|
|
|
}
|
|
|
|
store.dbs = append(store.dbs, db)
|
2019-05-18 00:33:49 +00:00
|
|
|
}
|
2019-07-03 04:25:53 +00:00
|
|
|
store.dbCount = dbCount
|
|
|
|
|
2019-05-18 00:33:49 +00:00
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
func (store *LevelDB2Store) BeginTransaction(ctx context.Context) (context.Context, error) {
|
|
|
|
return ctx, nil
|
|
|
|
}
|
|
|
|
func (store *LevelDB2Store) CommitTransaction(ctx context.Context) error {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
func (store *LevelDB2Store) RollbackTransaction(ctx context.Context) error {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func (store *LevelDB2Store) InsertEntry(ctx context.Context, entry *filer2.Entry) (err error) {
|
2019-07-03 04:25:53 +00:00
|
|
|
dir, name := entry.DirAndName()
|
|
|
|
key, partitionId := genKey(dir, name, store.dbCount)
|
2019-05-18 00:33:49 +00:00
|
|
|
|
|
|
|
value, err := entry.EncodeAttributesAndChunks()
|
|
|
|
if err != nil {
|
|
|
|
return fmt.Errorf("encoding %s %+v: %v", entry.FullPath, entry.Attr, err)
|
|
|
|
}
|
|
|
|
|
2019-07-03 04:25:53 +00:00
|
|
|
err = store.dbs[partitionId].Put(key, value, nil)
|
2019-05-18 00:33:49 +00:00
|
|
|
|
|
|
|
if err != nil {
|
|
|
|
return fmt.Errorf("persisting %s : %v", entry.FullPath, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// println("saved", entry.FullPath, "chunks", len(entry.Chunks))
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func (store *LevelDB2Store) UpdateEntry(ctx context.Context, entry *filer2.Entry) (err error) {
|
|
|
|
|
|
|
|
return store.InsertEntry(ctx, entry)
|
|
|
|
}
|
|
|
|
|
|
|
|
func (store *LevelDB2Store) FindEntry(ctx context.Context, fullpath filer2.FullPath) (entry *filer2.Entry, err error) {
|
2019-07-03 04:25:53 +00:00
|
|
|
dir, name := fullpath.DirAndName()
|
|
|
|
key, partitionId := genKey(dir, name, store.dbCount)
|
2019-05-18 00:33:49 +00:00
|
|
|
|
2019-07-03 04:25:53 +00:00
|
|
|
data, err := store.dbs[partitionId].Get(key, nil)
|
2019-05-18 00:33:49 +00:00
|
|
|
|
|
|
|
if err == leveldb.ErrNotFound {
|
2020-03-08 01:01:39 +00:00
|
|
|
return nil, filer_pb.ErrNotFound
|
2019-05-18 00:33:49 +00:00
|
|
|
}
|
|
|
|
if err != nil {
|
|
|
|
return nil, fmt.Errorf("get %s : %v", entry.FullPath, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
entry = &filer2.Entry{
|
|
|
|
FullPath: fullpath,
|
|
|
|
}
|
|
|
|
err = entry.DecodeAttributesAndChunks(data)
|
|
|
|
if err != nil {
|
|
|
|
return entry, fmt.Errorf("decode %s : %v", entry.FullPath, err)
|
|
|
|
}
|
|
|
|
|
2019-07-03 04:33:23 +00:00
|
|
|
// println("read", entry.FullPath, "chunks", len(entry.Chunks), "data", len(data), string(data))
|
2019-05-18 00:33:49 +00:00
|
|
|
|
|
|
|
return entry, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func (store *LevelDB2Store) DeleteEntry(ctx context.Context, fullpath filer2.FullPath) (err error) {
|
2019-07-03 04:25:53 +00:00
|
|
|
dir, name := fullpath.DirAndName()
|
|
|
|
key, partitionId := genKey(dir, name, store.dbCount)
|
2019-05-18 00:33:49 +00:00
|
|
|
|
2019-07-03 04:25:53 +00:00
|
|
|
err = store.dbs[partitionId].Delete(key, nil)
|
2019-05-18 00:33:49 +00:00
|
|
|
if err != nil {
|
|
|
|
return fmt.Errorf("delete %s : %v", fullpath, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2019-12-13 08:23:05 +00:00
|
|
|
func (store *LevelDB2Store) DeleteFolderChildren(ctx context.Context, fullpath filer2.FullPath) (err error) {
|
|
|
|
directoryPrefix, partitionId := genDirectoryKeyPrefix(fullpath, "", store.dbCount)
|
|
|
|
|
|
|
|
batch := new(leveldb.Batch)
|
|
|
|
|
|
|
|
iter := store.dbs[partitionId].NewIterator(&leveldb_util.Range{Start: directoryPrefix}, nil)
|
|
|
|
for iter.Next() {
|
|
|
|
key := iter.Key()
|
|
|
|
if !bytes.HasPrefix(key, directoryPrefix) {
|
|
|
|
break
|
|
|
|
}
|
|
|
|
fileName := getNameFromKey(key)
|
|
|
|
if fileName == "" {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
batch.Delete(append(directoryPrefix, []byte(fileName)...))
|
|
|
|
}
|
|
|
|
iter.Release()
|
|
|
|
|
|
|
|
err = store.dbs[partitionId].Write(batch, nil)
|
|
|
|
|
|
|
|
if err != nil {
|
|
|
|
return fmt.Errorf("delete %s : %v", fullpath, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2019-05-18 00:33:49 +00:00
|
|
|
func (store *LevelDB2Store) ListDirectoryEntries(ctx context.Context, fullpath filer2.FullPath, startFileName string, inclusive bool,
|
|
|
|
limit int) (entries []*filer2.Entry, err error) {
|
|
|
|
|
2019-07-03 04:25:53 +00:00
|
|
|
directoryPrefix, partitionId := genDirectoryKeyPrefix(fullpath, "", store.dbCount)
|
2019-07-08 03:37:48 +00:00
|
|
|
lastFileStart, _ := genDirectoryKeyPrefix(fullpath, startFileName, store.dbCount)
|
2019-05-18 00:33:49 +00:00
|
|
|
|
2019-07-08 03:37:48 +00:00
|
|
|
iter := store.dbs[partitionId].NewIterator(&leveldb_util.Range{Start: lastFileStart}, nil)
|
2019-05-18 00:33:49 +00:00
|
|
|
for iter.Next() {
|
|
|
|
key := iter.Key()
|
|
|
|
if !bytes.HasPrefix(key, directoryPrefix) {
|
|
|
|
break
|
|
|
|
}
|
|
|
|
fileName := getNameFromKey(key)
|
|
|
|
if fileName == "" {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
if fileName == startFileName && !inclusive {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
limit--
|
|
|
|
if limit < 0 {
|
|
|
|
break
|
|
|
|
}
|
|
|
|
entry := &filer2.Entry{
|
|
|
|
FullPath: filer2.NewFullPath(string(fullpath), fileName),
|
|
|
|
}
|
2019-07-03 04:25:53 +00:00
|
|
|
|
2019-07-03 04:33:23 +00:00
|
|
|
// println("list", entry.FullPath, "chunks", len(entry.Chunks))
|
2019-07-03 04:25:53 +00:00
|
|
|
|
2019-05-18 00:33:49 +00:00
|
|
|
if decodeErr := entry.DecodeAttributesAndChunks(iter.Value()); decodeErr != nil {
|
|
|
|
err = decodeErr
|
|
|
|
glog.V(0).Infof("list %s : %v", entry.FullPath, err)
|
|
|
|
break
|
|
|
|
}
|
|
|
|
entries = append(entries, entry)
|
|
|
|
}
|
|
|
|
iter.Release()
|
|
|
|
|
|
|
|
return entries, err
|
|
|
|
}
|
|
|
|
|
2019-07-03 04:25:53 +00:00
|
|
|
func genKey(dirPath, fileName string, dbCount int) (key []byte, partitionId int) {
|
|
|
|
key, partitionId = hashToBytes(dirPath, dbCount)
|
2019-05-18 00:33:49 +00:00
|
|
|
key = append(key, []byte(fileName)...)
|
2019-07-03 04:25:53 +00:00
|
|
|
return key, partitionId
|
2019-05-18 00:33:49 +00:00
|
|
|
}
|
|
|
|
|
2019-07-03 04:25:53 +00:00
|
|
|
func genDirectoryKeyPrefix(fullpath filer2.FullPath, startFileName string, dbCount int) (keyPrefix []byte, partitionId int) {
|
|
|
|
keyPrefix, partitionId = hashToBytes(string(fullpath), dbCount)
|
2019-05-18 00:33:49 +00:00
|
|
|
if len(startFileName) > 0 {
|
|
|
|
keyPrefix = append(keyPrefix, []byte(startFileName)...)
|
|
|
|
}
|
2019-07-03 04:25:53 +00:00
|
|
|
return keyPrefix, partitionId
|
2019-05-18 00:33:49 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
func getNameFromKey(key []byte) string {
|
|
|
|
|
2019-07-03 04:25:53 +00:00
|
|
|
return string(key[md5.Size:])
|
2019-05-18 00:33:49 +00:00
|
|
|
|
|
|
|
}
|
|
|
|
|
2019-07-03 04:25:53 +00:00
|
|
|
// hash directory, and use last byte for partitioning
|
|
|
|
func hashToBytes(dir string, dbCount int) ([]byte, int) {
|
2019-05-18 00:33:49 +00:00
|
|
|
h := md5.New()
|
|
|
|
io.WriteString(h, dir)
|
|
|
|
|
|
|
|
b := h.Sum(nil)
|
|
|
|
|
2019-07-03 04:25:53 +00:00
|
|
|
x := b[len(b)-1]
|
|
|
|
|
2019-07-22 04:51:38 +00:00
|
|
|
return b, int(x) % dbCount
|
2019-05-18 00:33:49 +00:00
|
|
|
}
|