mirror of
https://github.com/seaweedfs/seaweedfs.git
synced 2024-01-19 02:48:24 +00:00
go fmt
This commit is contained in:
parent
326a4bbaf5
commit
da36abf033
|
@ -90,7 +90,7 @@ func runMaster(cmd *Command, args []string) bool {
|
||||||
grace.SetupProfiling(*masterCpuProfile, *masterMemProfile)
|
grace.SetupProfiling(*masterCpuProfile, *masterMemProfile)
|
||||||
|
|
||||||
parent, _ := util.FullPath(*m.metaFolder).DirAndName()
|
parent, _ := util.FullPath(*m.metaFolder).DirAndName()
|
||||||
if util.FileExists(string(parent)) && !util.FileExists(*m.metaFolder){
|
if util.FileExists(string(parent)) && !util.FileExists(*m.metaFolder) {
|
||||||
os.MkdirAll(*m.metaFolder, 0755)
|
os.MkdirAll(*m.metaFolder, 0755)
|
||||||
}
|
}
|
||||||
if err := util.TestFolderWritable(util.ResolvePath(*m.metaFolder)); err != nil {
|
if err := util.TestFolderWritable(util.ResolvePath(*m.metaFolder)); err != nil {
|
||||||
|
|
|
@ -14,7 +14,7 @@ import (
|
||||||
|
|
||||||
var (
|
var (
|
||||||
concurrentWriterLimit = runtime.NumCPU()
|
concurrentWriterLimit = runtime.NumCPU()
|
||||||
concurrentWriters = util.NewLimitedConcurrentExecutor(4 * concurrentWriterLimit)
|
concurrentWriters = util.NewLimitedConcurrentExecutor(4 * concurrentWriterLimit)
|
||||||
)
|
)
|
||||||
|
|
||||||
type ContinuousDirtyPages struct {
|
type ContinuousDirtyPages struct {
|
||||||
|
|
|
@ -95,7 +95,7 @@ func (c *ContinuousIntervals) AddInterval(data []byte, offset int64) {
|
||||||
// append to the tail and return
|
// append to the tail and return
|
||||||
if len(c.lists) == 1 {
|
if len(c.lists) == 1 {
|
||||||
lastSpan := c.lists[0]
|
lastSpan := c.lists[0]
|
||||||
if lastSpan.Tail.Offset + lastSpan.Tail.Size == offset {
|
if lastSpan.Tail.Offset+lastSpan.Tail.Size == offset {
|
||||||
lastSpan.addNodeToTail(interval)
|
lastSpan.addNodeToTail(interval)
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
|
@ -126,7 +126,7 @@ func (fh *FileHandle) readFromChunks(buff []byte, offset int64) (int64, error) {
|
||||||
|
|
||||||
totalRead, err := fh.f.reader.ReadAt(buff, offset)
|
totalRead, err := fh.f.reader.ReadAt(buff, offset)
|
||||||
|
|
||||||
if err != nil && err != io.EOF{
|
if err != nil && err != io.EOF {
|
||||||
glog.Errorf("file handle read %s: %v", fh.f.fullpath(), err)
|
glog.Errorf("file handle read %s: %v", fh.f.fullpath(), err)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -67,7 +67,7 @@ func (s3a *S3ApiServer) PutBucketHandler(w http.ResponseWriter, r *http.Request)
|
||||||
}); err != nil {
|
}); err != nil {
|
||||||
glog.Errorf("list collection: %v", err)
|
glog.Errorf("list collection: %v", err)
|
||||||
return fmt.Errorf("list collections: %v", err)
|
return fmt.Errorf("list collections: %v", err)
|
||||||
}else {
|
} else {
|
||||||
for _, c := range resp.Collections {
|
for _, c := range resp.Collections {
|
||||||
if bucket == c.Name {
|
if bucket == c.Name {
|
||||||
errCode = s3err.ErrBucketAlreadyExists
|
errCode = s3err.ErrBucketAlreadyExists
|
||||||
|
|
Loading…
Reference in a new issue