2020-03-30 04:07:55 +00:00
|
|
|
package filer2
|
2020-03-27 11:50:51 +00:00
|
|
|
|
|
|
|
import (
|
|
|
|
"bytes"
|
|
|
|
"context"
|
|
|
|
"fmt"
|
|
|
|
"io"
|
|
|
|
"sync"
|
|
|
|
|
|
|
|
"github.com/chrislusf/seaweedfs/weed/glog"
|
|
|
|
"github.com/chrislusf/seaweedfs/weed/pb/filer_pb"
|
2020-03-28 20:43:31 +00:00
|
|
|
"github.com/chrislusf/seaweedfs/weed/pb/pb_cache"
|
2020-03-27 11:50:51 +00:00
|
|
|
"github.com/chrislusf/seaweedfs/weed/util"
|
|
|
|
"github.com/chrislusf/seaweedfs/weed/wdclient"
|
|
|
|
)
|
|
|
|
|
|
|
|
type ChunkReadAt struct {
|
|
|
|
masterClient *wdclient.MasterClient
|
2020-03-30 04:07:55 +00:00
|
|
|
chunkViews []*ChunkView
|
2020-03-27 11:50:51 +00:00
|
|
|
buffer []byte
|
|
|
|
bufferOffset int64
|
|
|
|
lookupFileId func(fileId string) (targetUrl string, err error)
|
|
|
|
readerLock sync.Mutex
|
2020-03-28 20:43:31 +00:00
|
|
|
|
|
|
|
chunkCache *pb_cache.ChunkCache
|
2020-03-27 11:50:51 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// var _ = io.ReaderAt(&ChunkReadAt{})
|
|
|
|
|
2020-03-30 04:07:55 +00:00
|
|
|
func NewChunkReaderAtFromClient(filerClient filer_pb.FilerClient, chunkViews []*ChunkView, chunkCache *pb_cache.ChunkCache) *ChunkReadAt {
|
2020-03-27 11:50:51 +00:00
|
|
|
|
|
|
|
return &ChunkReadAt{
|
|
|
|
chunkViews: chunkViews,
|
|
|
|
lookupFileId: func(fileId string) (targetUrl string, err error) {
|
|
|
|
err = filerClient.WithFilerClient(func(client filer_pb.SeaweedFilerClient) error {
|
2020-03-30 04:07:55 +00:00
|
|
|
vid := VolumeId(fileId)
|
2020-03-27 11:50:51 +00:00
|
|
|
resp, err := client.LookupVolume(context.Background(), &filer_pb.LookupVolumeRequest{
|
|
|
|
VolumeIds: []string{vid},
|
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
locations := resp.LocationsMap[vid]
|
|
|
|
if locations == nil || len(locations.Locations) == 0 {
|
|
|
|
glog.V(0).Infof("failed to locate %s", fileId)
|
|
|
|
return fmt.Errorf("failed to locate %s", fileId)
|
|
|
|
}
|
|
|
|
|
|
|
|
volumeServerAddress := filerClient.AdjustedUrl(locations.Locations[0].Url)
|
|
|
|
|
|
|
|
targetUrl = fmt.Sprintf("http://%s/%s", volumeServerAddress, fileId)
|
|
|
|
|
|
|
|
return nil
|
|
|
|
})
|
|
|
|
return
|
|
|
|
},
|
|
|
|
bufferOffset: -1,
|
2020-03-28 21:07:16 +00:00
|
|
|
chunkCache: chunkCache,
|
2020-03-27 11:50:51 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func (c *ChunkReadAt) ReadAt(p []byte, offset int64) (n int, err error) {
|
|
|
|
|
|
|
|
c.readerLock.Lock()
|
|
|
|
defer c.readerLock.Unlock()
|
|
|
|
|
|
|
|
for n < len(p) && err == nil {
|
|
|
|
readCount, readErr := c.doReadAt(p[n:], offset+int64(n))
|
|
|
|
n += readCount
|
|
|
|
err = readErr
|
|
|
|
if readCount == 0 {
|
|
|
|
return n, nil
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
func (c *ChunkReadAt) doReadAt(p []byte, offset int64) (n int, err error) {
|
|
|
|
|
|
|
|
var found bool
|
|
|
|
for _, chunk := range c.chunkViews {
|
|
|
|
if chunk.LogicOffset <= offset && offset < chunk.LogicOffset+int64(chunk.Size) {
|
|
|
|
found = true
|
|
|
|
if c.bufferOffset != chunk.LogicOffset {
|
2020-03-28 20:43:31 +00:00
|
|
|
c.buffer, err = c.fetchChunkData(chunk)
|
|
|
|
c.bufferOffset = chunk.LogicOffset
|
2020-03-27 11:50:51 +00:00
|
|
|
}
|
|
|
|
break
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if !found {
|
|
|
|
return 0, io.EOF
|
|
|
|
}
|
|
|
|
|
|
|
|
n = copy(p, c.buffer[offset-c.bufferOffset:])
|
|
|
|
|
|
|
|
// fmt.Printf("> doReadAt [%d,%d), buffer:[%d,%d)\n", offset, offset+int64(n), c.bufferOffset, c.bufferOffset+int64(len(c.buffer)))
|
|
|
|
|
|
|
|
return
|
|
|
|
|
|
|
|
}
|
|
|
|
|
2020-03-30 04:07:55 +00:00
|
|
|
func (c *ChunkReadAt) fetchChunkData(chunkView *ChunkView) (data []byte, err error) {
|
2020-03-27 11:50:51 +00:00
|
|
|
|
2020-03-28 21:07:25 +00:00
|
|
|
// fmt.Printf("fetching %s [%d,%d)\n", chunkView.FileId, chunkView.LogicOffset, chunkView.LogicOffset+int64(chunkView.Size))
|
2020-03-28 20:43:31 +00:00
|
|
|
|
|
|
|
chunkData := c.chunkCache.GetChunk(chunkView.FileId)
|
|
|
|
if chunkData != nil {
|
|
|
|
glog.V(3).Infof("cache hit %s [%d,%d)", chunkView.FileId, chunkView.LogicOffset, chunkView.LogicOffset+int64(chunkView.Size))
|
2020-03-29 07:54:39 +00:00
|
|
|
} else {
|
|
|
|
chunkData, err = c.doFetchFullChunkData(chunkView.FileId, chunkView.CipherKey, chunkView.IsGzipped)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if int64(len(chunkData)) < chunkView.Offset+int64(chunkView.Size) {
|
|
|
|
return nil, fmt.Errorf("unexpected larger chunkView [%d,%d) than chunk %d", chunkView.Offset, chunkView.Offset+int64(chunkView.Size), len(chunkData))
|
2020-03-28 20:43:31 +00:00
|
|
|
}
|
2020-03-27 11:50:51 +00:00
|
|
|
|
2020-03-29 07:54:39 +00:00
|
|
|
data = chunkData[chunkView.Offset : chunkView.Offset+int64(chunkView.Size)]
|
|
|
|
|
|
|
|
c.chunkCache.SetChunk(chunkView.FileId, chunkData)
|
|
|
|
|
|
|
|
return data, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func (c *ChunkReadAt) doFetchFullChunkData(fileId string, cipherKey []byte, isGzipped bool) ([]byte, error) {
|
|
|
|
|
|
|
|
urlString, err := c.lookupFileId(fileId)
|
2020-03-27 11:50:51 +00:00
|
|
|
if err != nil {
|
2020-03-29 07:54:39 +00:00
|
|
|
glog.V(1).Infof("operation LookupFileId %s failed, err: %v", fileId, err)
|
2020-03-28 20:43:31 +00:00
|
|
|
return nil, err
|
2020-03-27 11:50:51 +00:00
|
|
|
}
|
|
|
|
var buffer bytes.Buffer
|
2020-03-29 07:54:39 +00:00
|
|
|
err = util.ReadUrlAsStream(urlString, cipherKey, isGzipped, true, 0, 0, func(data []byte) {
|
2020-03-27 11:50:51 +00:00
|
|
|
buffer.Write(data)
|
|
|
|
})
|
|
|
|
if err != nil {
|
2020-03-29 07:54:39 +00:00
|
|
|
glog.V(1).Infof("read %s failed, err: %v", fileId, err)
|
2020-03-28 20:43:31 +00:00
|
|
|
return nil, err
|
2020-03-27 11:50:51 +00:00
|
|
|
}
|
|
|
|
|
2020-03-29 07:54:39 +00:00
|
|
|
return buffer.Bytes(), nil
|
2020-03-27 11:50:51 +00:00
|
|
|
}
|