mirror of
https://github.com/seaweedfs/seaweedfs.git
synced 2024-01-19 02:48:24 +00:00
refactoring
This commit is contained in:
parent
3aa3991f0f
commit
d6ba97219b
|
@ -5,8 +5,6 @@ import (
|
||||||
"fmt"
|
"fmt"
|
||||||
"io"
|
"io"
|
||||||
"math"
|
"math"
|
||||||
"net/url"
|
|
||||||
"strings"
|
|
||||||
"sync"
|
"sync"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
|
@ -122,44 +120,7 @@ func fetchChunkRange(buffer []byte, lookupFileIdFn wdclient.LookupFileIdFunction
|
||||||
glog.Errorf("operation LookupFileId %s failed, err: %v", fileId, err)
|
glog.Errorf("operation LookupFileId %s failed, err: %v", fileId, err)
|
||||||
return 0, err
|
return 0, err
|
||||||
}
|
}
|
||||||
return retriedFetchChunkData(buffer, urlStrings, cipherKey, isGzipped, false, offset)
|
return util.RetriedFetchChunkData(buffer, urlStrings, cipherKey, isGzipped, false, offset)
|
||||||
}
|
|
||||||
|
|
||||||
func retriedFetchChunkData(buffer []byte, urlStrings []string, cipherKey []byte, isGzipped bool, isFullChunk bool, offset int64) (n int, err error) {
|
|
||||||
|
|
||||||
var shouldRetry bool
|
|
||||||
|
|
||||||
for waitTime := time.Second; waitTime < util.RetryWaitTime; waitTime += waitTime / 2 {
|
|
||||||
for _, urlString := range urlStrings {
|
|
||||||
n = 0
|
|
||||||
if strings.Contains(urlString, "%") {
|
|
||||||
urlString = url.PathEscape(urlString)
|
|
||||||
}
|
|
||||||
shouldRetry, err = util.ReadUrlAsStream(urlString+"?readDeleted=true", cipherKey, isGzipped, isFullChunk, offset, len(buffer), func(data []byte) {
|
|
||||||
if n < len(buffer) {
|
|
||||||
x := copy(buffer[n:], data)
|
|
||||||
n += x
|
|
||||||
}
|
|
||||||
})
|
|
||||||
if !shouldRetry {
|
|
||||||
break
|
|
||||||
}
|
|
||||||
if err != nil {
|
|
||||||
glog.V(0).Infof("read %s failed, err: %v", urlString, err)
|
|
||||||
} else {
|
|
||||||
break
|
|
||||||
}
|
|
||||||
}
|
|
||||||
if err != nil && shouldRetry {
|
|
||||||
glog.V(0).Infof("retry reading in %v", waitTime)
|
|
||||||
time.Sleep(waitTime)
|
|
||||||
} else {
|
|
||||||
break
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
return n, err
|
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func retriedStreamFetchChunkData(writer io.Writer, urlStrings []string, cipherKey []byte, isGzipped bool, isFullChunk bool, offset int64, size int) (err error) {
|
func retriedStreamFetchChunkData(writer io.Writer, urlStrings []string, cipherKey []byte, isGzipped bool, isFullChunk bool, offset int64, size int) (err error) {
|
||||||
|
|
|
@ -2,6 +2,7 @@ package filer
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"fmt"
|
"fmt"
|
||||||
|
"github.com/seaweedfs/seaweedfs/weed/util"
|
||||||
"sync"
|
"sync"
|
||||||
"sync/atomic"
|
"sync/atomic"
|
||||||
"time"
|
"time"
|
||||||
|
@ -170,7 +171,7 @@ func (s *SingleChunkCacher) startCaching() {
|
||||||
|
|
||||||
s.data = mem.Allocate(s.chunkSize)
|
s.data = mem.Allocate(s.chunkSize)
|
||||||
|
|
||||||
_, s.err = retriedFetchChunkData(s.data, urlStrings, s.cipherKey, s.isGzipped, true, 0)
|
_, s.err = util.RetriedFetchChunkData(s.data, urlStrings, s.cipherKey, s.isGzipped, true, 0)
|
||||||
if s.err != nil {
|
if s.err != nil {
|
||||||
mem.Free(s.data)
|
mem.Free(s.data)
|
||||||
s.data = nil
|
s.data = nil
|
||||||
|
|
|
@ -176,7 +176,7 @@ func ReadAll(buffer []byte, masterClient *wdclient.MasterClient, chunks []*filer
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
n, err := retriedFetchChunkData(buffer[idx:idx+int(chunkView.ViewSize)], urlStrings, chunkView.CipherKey, chunkView.IsGzipped, chunkView.IsFullChunk(), chunkView.OffsetInChunk)
|
n, err := util.RetriedFetchChunkData(buffer[idx:idx+int(chunkView.ViewSize)], urlStrings, chunkView.CipherKey, chunkView.IsGzipped, chunkView.IsFullChunk(), chunkView.OffsetInChunk)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
|
@ -10,6 +10,7 @@ import (
|
||||||
"net/http"
|
"net/http"
|
||||||
"net/url"
|
"net/url"
|
||||||
"strings"
|
"strings"
|
||||||
|
"time"
|
||||||
|
|
||||||
"github.com/seaweedfs/seaweedfs/weed/glog"
|
"github.com/seaweedfs/seaweedfs/weed/glog"
|
||||||
)
|
)
|
||||||
|
@ -450,3 +451,40 @@ func (r *CountingReader) Read(p []byte) (n int, err error) {
|
||||||
r.BytesRead += n
|
r.BytesRead += n
|
||||||
return n, err
|
return n, err
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func RetriedFetchChunkData(buffer []byte, urlStrings []string, cipherKey []byte, isGzipped bool, isFullChunk bool, offset int64) (n int, err error) {
|
||||||
|
|
||||||
|
var shouldRetry bool
|
||||||
|
|
||||||
|
for waitTime := time.Second; waitTime < RetryWaitTime; waitTime += waitTime / 2 {
|
||||||
|
for _, urlString := range urlStrings {
|
||||||
|
n = 0
|
||||||
|
if strings.Contains(urlString, "%") {
|
||||||
|
urlString = url.PathEscape(urlString)
|
||||||
|
}
|
||||||
|
shouldRetry, err = ReadUrlAsStream(urlString+"?readDeleted=true", cipherKey, isGzipped, isFullChunk, offset, len(buffer), func(data []byte) {
|
||||||
|
if n < len(buffer) {
|
||||||
|
x := copy(buffer[n:], data)
|
||||||
|
n += x
|
||||||
|
}
|
||||||
|
})
|
||||||
|
if !shouldRetry {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
if err != nil {
|
||||||
|
glog.V(0).Infof("read %s failed, err: %v", urlString, err)
|
||||||
|
} else {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if err != nil && shouldRetry {
|
||||||
|
glog.V(0).Infof("retry reading in %v", waitTime)
|
||||||
|
time.Sleep(waitTime)
|
||||||
|
} else {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return n, err
|
||||||
|
|
||||||
|
}
|
||||||
|
|
Loading…
Reference in a new issue