mirror of
https://github.com/seaweedfs/seaweedfs.git
synced 2024-01-19 02:48:24 +00:00
renaming
This commit is contained in:
parent
16fe132a20
commit
e912fd15e3
|
@ -48,8 +48,8 @@ func main() {
|
||||||
if *showTextFile {
|
if *showTextFile {
|
||||||
|
|
||||||
data := n.Data
|
data := n.Data
|
||||||
if n.IsGzipped() {
|
if n.IsCompressed() {
|
||||||
if data, err = util2.UnGzipData(data); err != nil {
|
if data, err = util2.UnCompressData(data); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -57,7 +57,7 @@ func main() {
|
||||||
println(string(data))
|
println(string(data))
|
||||||
}
|
}
|
||||||
|
|
||||||
println("-", n.String(), "compressed", n.IsGzipped(), "original size", len(data))
|
println("-", n.String(), "compressed", n.IsCompressed(), "original size", len(data))
|
||||||
}
|
}
|
||||||
return nil
|
return nil
|
||||||
})
|
})
|
||||||
|
|
|
@ -19,6 +19,7 @@ import (
|
||||||
"github.com/chrislusf/seaweedfs/weed/storage/needle_map"
|
"github.com/chrislusf/seaweedfs/weed/storage/needle_map"
|
||||||
"github.com/chrislusf/seaweedfs/weed/storage/super_block"
|
"github.com/chrislusf/seaweedfs/weed/storage/super_block"
|
||||||
"github.com/chrislusf/seaweedfs/weed/storage/types"
|
"github.com/chrislusf/seaweedfs/weed/storage/types"
|
||||||
|
"github.com/chrislusf/seaweedfs/weed/util"
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
|
@ -79,7 +80,7 @@ func printNeedle(vid needle.VolumeId, n *needle.Needle, version needle.Version,
|
||||||
key,
|
key,
|
||||||
n.Name,
|
n.Name,
|
||||||
size,
|
size,
|
||||||
n.IsGzipped(),
|
n.IsCompressed(),
|
||||||
n.Mime,
|
n.Mime,
|
||||||
n.LastModifiedString(),
|
n.LastModifiedString(),
|
||||||
n.Ttl.String(),
|
n.Ttl.String(),
|
||||||
|
@ -108,8 +109,8 @@ func (scanner *VolumeFileScanner4Export) VisitNeedle(n *needle.Needle, offset in
|
||||||
vid := scanner.vid
|
vid := scanner.vid
|
||||||
|
|
||||||
nv, ok := needleMap.Get(n.Id)
|
nv, ok := needleMap.Get(n.Id)
|
||||||
glog.V(3).Infof("key %d offset %d size %d disk_size %d gzip %v ok %v nv %+v",
|
glog.V(3).Infof("key %d offset %d size %d disk_size %d compressed %v ok %v nv %+v",
|
||||||
n.Id, offset, n.Size, n.DiskSize(scanner.version), n.IsGzipped(), ok, nv)
|
n.Id, offset, n.Size, n.DiskSize(scanner.version), n.IsCompressed(), ok, nv)
|
||||||
if ok && nv.Size > 0 && nv.Size != types.TombstoneFileSize && nv.Offset.ToAcutalOffset() == offset {
|
if ok && nv.Size > 0 && nv.Size != types.TombstoneFileSize && nv.Offset.ToAcutalOffset() == offset {
|
||||||
if newerThanUnix >= 0 && n.HasLastModifiedDate() && n.LastModified < uint64(newerThanUnix) {
|
if newerThanUnix >= 0 && n.HasLastModifiedDate() && n.LastModified < uint64(newerThanUnix) {
|
||||||
glog.V(3).Infof("Skipping this file, as it's old enough: LastModified %d vs %d",
|
glog.V(3).Infof("Skipping this file, as it's old enough: LastModified %d vs %d",
|
||||||
|
@ -242,9 +243,12 @@ func writeFile(vid needle.VolumeId, n *needle.Needle) (err error) {
|
||||||
|
|
||||||
fileName := fileNameTemplateBuffer.String()
|
fileName := fileNameTemplateBuffer.String()
|
||||||
|
|
||||||
if n.IsGzipped() && path.Ext(fileName) != ".gz" {
|
if n.IsCompressed() {
|
||||||
|
if util.IsGzippedContent(n.Data) && path.Ext(fileName) != ".gz" {
|
||||||
fileName = fileName + ".gz"
|
fileName = fileName + ".gz"
|
||||||
}
|
}
|
||||||
|
// TODO other compression method
|
||||||
|
}
|
||||||
|
|
||||||
tarHeader.Name, tarHeader.Size = fileName, int64(len(n.Data))
|
tarHeader.Name, tarHeader.Size = fileName, int64(len(n.Data))
|
||||||
if n.HasLastModifiedDate() {
|
if n.HasLastModifiedDate() {
|
||||||
|
|
|
@ -46,7 +46,7 @@ func (scanner *VolumeFileScanner4Fix) ReadNeedleBody() bool {
|
||||||
}
|
}
|
||||||
|
|
||||||
func (scanner *VolumeFileScanner4Fix) VisitNeedle(n *needle.Needle, offset int64, needleHeader, needleBody []byte) error {
|
func (scanner *VolumeFileScanner4Fix) VisitNeedle(n *needle.Needle, offset int64, needleHeader, needleBody []byte) error {
|
||||||
glog.V(2).Infof("key %d offset %d size %d disk_size %d gzip %v", n.Id, offset, n.Size, n.DiskSize(scanner.version), n.IsGzipped())
|
glog.V(2).Infof("key %d offset %d size %d disk_size %d compressed %v", n.Id, offset, n.Size, n.DiskSize(scanner.version), n.IsCompressed())
|
||||||
if n.Size > 0 && n.Size != types.TombstoneFileSize {
|
if n.Size > 0 && n.Size != types.TombstoneFileSize {
|
||||||
pe := scanner.nm.Set(n.Id, types.ToOffset(offset), n.Size)
|
pe := scanner.nm.Set(n.Id, types.ToOffset(offset), n.Size)
|
||||||
glog.V(2).Infof("saved %d with error %v", n.Size, pe)
|
glog.V(2).Infof("saved %d with error %v", n.Size, pe)
|
||||||
|
|
|
@ -53,10 +53,10 @@ func (s ChunkList) Len() int { return len(s) }
|
||||||
func (s ChunkList) Less(i, j int) bool { return s[i].Offset < s[j].Offset }
|
func (s ChunkList) Less(i, j int) bool { return s[i].Offset < s[j].Offset }
|
||||||
func (s ChunkList) Swap(i, j int) { s[i], s[j] = s[j], s[i] }
|
func (s ChunkList) Swap(i, j int) { s[i], s[j] = s[j], s[i] }
|
||||||
|
|
||||||
func LoadChunkManifest(buffer []byte, isGzipped bool) (*ChunkManifest, error) {
|
func LoadChunkManifest(buffer []byte, isCompressed bool) (*ChunkManifest, error) {
|
||||||
if isGzipped {
|
if isCompressed {
|
||||||
var err error
|
var err error
|
||||||
if buffer, err = util.UnGzipData(buffer); err != nil {
|
if buffer, err = util.UnCompressData(buffer); err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -121,7 +121,7 @@ func doUploadData(uploadUrl string, filename string, cipher bool, data []byte, i
|
||||||
}
|
}
|
||||||
} else if isInputGzipped {
|
} else if isInputGzipped {
|
||||||
// just to get the clear data length
|
// just to get the clear data length
|
||||||
clearData, err := util.UnGzipData(data)
|
clearData, err := util.UnCompressData(data)
|
||||||
if err == nil {
|
if err == nil {
|
||||||
clearDataLen = len(clearData)
|
clearDataLen = len(clearData)
|
||||||
}
|
}
|
||||||
|
|
|
@ -11,6 +11,7 @@ import (
|
||||||
"github.com/chrislusf/seaweedfs/weed/util"
|
"github.com/chrislusf/seaweedfs/weed/util"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
// Deprecated
|
||||||
func (vs *VolumeServer) FileGet(req *volume_server_pb.FileGetRequest, stream volume_server_pb.VolumeServer_FileGetServer) error {
|
func (vs *VolumeServer) FileGet(req *volume_server_pb.FileGetRequest, stream volume_server_pb.VolumeServer_FileGetServer) error {
|
||||||
|
|
||||||
headResponse := &volume_server_pb.FileGetResponse{}
|
headResponse := &volume_server_pb.FileGetResponse{}
|
||||||
|
@ -90,10 +91,10 @@ func (vs *VolumeServer) FileGet(req *volume_server_pb.FileGetRequest, stream vol
|
||||||
}
|
}
|
||||||
headResponse.ContentType = mtype
|
headResponse.ContentType = mtype
|
||||||
|
|
||||||
headResponse.IsGzipped = n.IsGzipped()
|
headResponse.IsGzipped = n.IsCompressed()
|
||||||
|
|
||||||
if n.IsGzipped() && req.AcceptGzip {
|
if n.IsCompressed() && req.AcceptGzip {
|
||||||
if n.Data, err = util.UnGzipData(n.Data); err != nil {
|
if n.Data, err = util.UnCompressData(n.Data); err != nil {
|
||||||
glog.V(0).Infof("ungzip %s error: %v", req.FileId, err)
|
glog.V(0).Infof("ungzip %s error: %v", req.FileId, err)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -143,17 +143,19 @@ func (vs *VolumeServer) GetOrHeadHandler(w http.ResponseWriter, r *http.Request)
|
||||||
}
|
}
|
||||||
|
|
||||||
if ext != ".gz" {
|
if ext != ".gz" {
|
||||||
if n.IsGzipped() {
|
if n.IsCompressed() {
|
||||||
if strings.Contains(r.Header.Get("Accept-Encoding"), "gzip") {
|
if strings.Contains(r.Header.Get("Accept-Encoding"), "gzip") {
|
||||||
if _, _, _, shouldResize := shouldResizeImages(ext, r); shouldResize {
|
if _, _, _, shouldResize := shouldResizeImages(ext, r); shouldResize {
|
||||||
if n.Data, err = util.UnGzipData(n.Data); err != nil {
|
if n.Data, err = util.UnCompressData(n.Data); err != nil {
|
||||||
glog.V(0).Infoln("ungzip error:", err, r.URL.Path)
|
glog.V(0).Infoln("ungzip error:", err, r.URL.Path)
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
|
if util.IsGzippedContent(n.Data) {
|
||||||
w.Header().Set("Content-Encoding", "gzip")
|
w.Header().Set("Content-Encoding", "gzip")
|
||||||
}
|
}
|
||||||
|
}
|
||||||
} else {
|
} else {
|
||||||
if n.Data, err = util.UnGzipData(n.Data); err != nil {
|
if n.Data, err = util.UnCompressData(n.Data); err != nil {
|
||||||
glog.V(0).Infoln("ungzip error:", err, r.URL.Path)
|
glog.V(0).Infoln("ungzip error:", err, r.URL.Path)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -172,7 +174,7 @@ func (vs *VolumeServer) tryHandleChunkedFile(n *needle.Needle, fileName string,
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
|
|
||||||
chunkManifest, e := operation.LoadChunkManifest(n.Data, n.IsGzipped())
|
chunkManifest, e := operation.LoadChunkManifest(n.Data, n.IsCompressed())
|
||||||
if e != nil {
|
if e != nil {
|
||||||
glog.V(0).Infof("load chunked manifest (%s) error: %v", r.URL.Path, e)
|
glog.V(0).Infof("load chunked manifest (%s) error: %v", r.URL.Path, e)
|
||||||
return false
|
return false
|
||||||
|
|
|
@ -120,7 +120,7 @@ func (vs *VolumeServer) DeleteHandler(w http.ResponseWriter, r *http.Request) {
|
||||||
count := int64(n.Size)
|
count := int64(n.Size)
|
||||||
|
|
||||||
if n.IsChunkedManifest() {
|
if n.IsChunkedManifest() {
|
||||||
chunkManifest, e := operation.LoadChunkManifest(n.Data, n.IsGzipped())
|
chunkManifest, e := operation.LoadChunkManifest(n.Data, n.IsCompressed())
|
||||||
if e != nil {
|
if e != nil {
|
||||||
writeJsonError(w, r, http.StatusInternalServerError, fmt.Errorf("Load chunks manifest error: %v", e))
|
writeJsonError(w, r, http.StatusInternalServerError, fmt.Errorf("Load chunks manifest error: %v", e))
|
||||||
return
|
return
|
||||||
|
|
|
@ -81,7 +81,7 @@ func CreateNeedleFromRequest(r *http.Request, fixJpgOrientation bool, sizeLimit
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if pu.IsGzipped {
|
if pu.IsGzipped {
|
||||||
n.SetGzipped()
|
n.SetIsCompressed()
|
||||||
}
|
}
|
||||||
if n.LastModified == 0 {
|
if n.LastModified == 0 {
|
||||||
n.LastModified = uint64(time.Now().Unix())
|
n.LastModified = uint64(time.Now().Unix())
|
||||||
|
|
|
@ -51,7 +51,7 @@ func ParseUpload(r *http.Request, sizeLimit int64) (pu *ParsedUpload, e error) {
|
||||||
|
|
||||||
pu.OriginalDataSize = len(pu.Data)
|
pu.OriginalDataSize = len(pu.Data)
|
||||||
pu.UncompressedData = pu.Data
|
pu.UncompressedData = pu.Data
|
||||||
// println("received data", len(pu.Data), "isGzipped", pu.IsGzipped, "mime", pu.MimeType, "name", pu.FileName)
|
// println("received data", len(pu.Data), "isGzipped", pu.IsCompressed, "mime", pu.MimeType, "name", pu.FileName)
|
||||||
if pu.MimeType == "" {
|
if pu.MimeType == "" {
|
||||||
pu.MimeType = http.DetectContentType(pu.Data)
|
pu.MimeType = http.DetectContentType(pu.Data)
|
||||||
// println("detected mimetype to", pu.MimeType)
|
// println("detected mimetype to", pu.MimeType)
|
||||||
|
@ -60,7 +60,7 @@ func ParseUpload(r *http.Request, sizeLimit int64) (pu *ParsedUpload, e error) {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if pu.IsGzipped {
|
if pu.IsGzipped {
|
||||||
if unzipped, e := util.UnGzipData(pu.Data); e == nil {
|
if unzipped, e := util.UnCompressData(pu.Data); e == nil {
|
||||||
pu.OriginalDataSize = len(unzipped)
|
pu.OriginalDataSize = len(unzipped)
|
||||||
pu.UncompressedData = unzipped
|
pu.UncompressedData = unzipped
|
||||||
// println("ungzipped data size", len(unzipped))
|
// println("ungzipped data size", len(unzipped))
|
||||||
|
|
|
@ -13,7 +13,7 @@ import (
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
FlagGzip = 0x01
|
FlagIsCompressed = 0x01
|
||||||
FlagHasName = 0x02
|
FlagHasName = 0x02
|
||||||
FlagHasMime = 0x04
|
FlagHasMime = 0x04
|
||||||
FlagHasLastModifiedDate = 0x08
|
FlagHasLastModifiedDate = 0x08
|
||||||
|
@ -343,11 +343,11 @@ func (n *Needle) ReadNeedleBodyBytes(needleBody []byte, version Version) (err er
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
func (n *Needle) IsGzipped() bool {
|
func (n *Needle) IsCompressed() bool {
|
||||||
return n.Flags&FlagGzip > 0
|
return n.Flags&FlagIsCompressed > 0
|
||||||
}
|
}
|
||||||
func (n *Needle) SetGzipped() {
|
func (n *Needle) SetIsCompressed() {
|
||||||
n.Flags = n.Flags | FlagGzip
|
n.Flags = n.Flags | FlagIsCompressed
|
||||||
}
|
}
|
||||||
func (n *Needle) HasName() bool {
|
func (n *Needle) HasName() bool {
|
||||||
return n.Flags&FlagHasName > 0
|
return n.Flags&FlagHasName > 0
|
||||||
|
|
|
@ -80,7 +80,7 @@ func ReplicatedWrite(masterNode string, s *storage.Store, volumeId needle.Volume
|
||||||
}
|
}
|
||||||
|
|
||||||
// volume server do not know about encryption
|
// volume server do not know about encryption
|
||||||
_, err := operation.UploadData(u.String(), string(n.Name), false, n.Data, n.IsGzipped(), string(n.Mime), pairMap, jwt)
|
_, err := operation.UploadData(u.String(), string(n.Name), false, n.Data, n.IsCompressed(), string(n.Mime), pairMap, jwt)
|
||||||
return err
|
return err
|
||||||
}); err != nil {
|
}); err != nil {
|
||||||
err = fmt.Errorf("failed to write to replicas for volume %d: %v", volumeId, err)
|
err = fmt.Errorf("failed to write to replicas for volume %d: %v", volumeId, err)
|
||||||
|
|
|
@ -25,7 +25,25 @@ func GzipData(input []byte) ([]byte, error) {
|
||||||
}
|
}
|
||||||
return buf.Bytes(), nil
|
return buf.Bytes(), nil
|
||||||
}
|
}
|
||||||
func UnGzipData(input []byte) ([]byte, error) {
|
func UnCompressData(input []byte) ([]byte, error) {
|
||||||
|
if IsGzippedContent(input) {
|
||||||
|
return ungzipData(input)
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
func ungzipData(input []byte) ([]byte, error) {
|
||||||
|
buf := bytes.NewBuffer(input)
|
||||||
|
r, _ := gzip.NewReader(buf)
|
||||||
|
defer r.Close()
|
||||||
|
output, err := ioutil.ReadAll(r)
|
||||||
|
if err != nil {
|
||||||
|
glog.V(2).Infoln("error uncompressing data:", err)
|
||||||
|
}
|
||||||
|
return output, err
|
||||||
|
}
|
||||||
|
|
||||||
|
func ungzipData(input []byte) ([]byte, error) {
|
||||||
buf := bytes.NewBuffer(input)
|
buf := bytes.NewBuffer(input)
|
||||||
r, _ := gzip.NewReader(buf)
|
r, _ := gzip.NewReader(buf)
|
||||||
defer r.Close()
|
defer r.Close()
|
||||||
|
@ -51,6 +69,13 @@ func IsGzippable(ext, mtype string, data []byte) bool {
|
||||||
return isMostlyText
|
return isMostlyText
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func IsGzippedContent(data []byte) bool {
|
||||||
|
if len(data) < 2 {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
return data[0] == 31 && data[1] == 139
|
||||||
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Default more not to gzip since gzip can be done on client side.
|
* Default more not to gzip since gzip can be done on client side.
|
||||||
*/func IsGzippableFileType(ext, mtype string) (shouldBeZipped, iAmSure bool) {
|
*/func IsGzippableFileType(ext, mtype string) (shouldBeZipped, iAmSure bool) {
|
||||||
|
|
|
@ -189,11 +189,11 @@ func NormalizeUrl(url string) string {
|
||||||
return "http://" + url
|
return "http://" + url
|
||||||
}
|
}
|
||||||
|
|
||||||
func ReadUrl(fileUrl string, cipherKey []byte, isGzipped bool, isFullChunk bool, offset int64, size int, buf []byte) (int64, error) {
|
func ReadUrl(fileUrl string, cipherKey []byte, isContentCompressed bool, isFullChunk bool, offset int64, size int, buf []byte) (int64, error) {
|
||||||
|
|
||||||
if cipherKey != nil {
|
if cipherKey != nil {
|
||||||
var n int
|
var n int
|
||||||
err := readEncryptedUrl(fileUrl, cipherKey, isGzipped, isFullChunk, offset, size, func(data []byte) {
|
err := readEncryptedUrl(fileUrl, cipherKey, isContentCompressed, isFullChunk, offset, size, func(data []byte) {
|
||||||
n = copy(buf, data)
|
n = copy(buf, data)
|
||||||
})
|
})
|
||||||
return int64(n), err
|
return int64(n), err
|
||||||
|
@ -300,7 +300,7 @@ func ReadUrlAsStream(fileUrl string, cipherKey []byte, isContentGzipped bool, is
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func readEncryptedUrl(fileUrl string, cipherKey []byte, isContentGzipped bool, isFullChunk bool, offset int64, size int, fn func(data []byte)) error {
|
func readEncryptedUrl(fileUrl string, cipherKey []byte, isContentCompressed bool, isFullChunk bool, offset int64, size int, fn func(data []byte)) error {
|
||||||
encryptedData, err := Get(fileUrl)
|
encryptedData, err := Get(fileUrl)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return fmt.Errorf("fetch %s: %v", fileUrl, err)
|
return fmt.Errorf("fetch %s: %v", fileUrl, err)
|
||||||
|
@ -309,8 +309,8 @@ func readEncryptedUrl(fileUrl string, cipherKey []byte, isContentGzipped bool, i
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return fmt.Errorf("decrypt %s: %v", fileUrl, err)
|
return fmt.Errorf("decrypt %s: %v", fileUrl, err)
|
||||||
}
|
}
|
||||||
if isContentGzipped {
|
if isContentCompressed {
|
||||||
decryptedData, err = UnGzipData(decryptedData)
|
decryptedData, err = UnCompressData(decryptedData)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return fmt.Errorf("unzip decrypt %s: %v", fileUrl, err)
|
return fmt.Errorf("unzip decrypt %s: %v", fileUrl, err)
|
||||||
}
|
}
|
||||||
|
|
Loading…
Reference in a new issue