2015-11-29 23:49:41 +08:00
|
|
|
package operation
|
|
|
|
|
|
|
|
import (
|
|
|
|
"encoding/json"
|
|
|
|
"errors"
|
|
|
|
"fmt"
|
|
|
|
"io"
|
2019-04-15 14:28:24 +08:00
|
|
|
"io/ioutil"
|
2015-11-29 23:49:41 +08:00
|
|
|
"net/http"
|
|
|
|
"sort"
|
2020-03-21 06:38:04 +08:00
|
|
|
"sync"
|
2015-11-29 23:49:41 +08:00
|
|
|
|
2019-04-19 12:43:36 +08:00
|
|
|
"google.golang.org/grpc"
|
|
|
|
|
2018-10-14 15:12:28 +08:00
|
|
|
"github.com/chrislusf/seaweedfs/weed/glog"
|
2018-10-14 15:30:20 +08:00
|
|
|
"github.com/chrislusf/seaweedfs/weed/util"
|
2015-11-29 23:49:41 +08:00
|
|
|
)
|
|
|
|
|
2015-12-01 20:23:50 +08:00
|
|
|
var (
|
|
|
|
// when the remote server does not allow range requests (Accept-Ranges was not set)
|
|
|
|
ErrRangeRequestsNotSupported = errors.New("Range requests are not supported by the remote server")
|
|
|
|
// ErrInvalidRange is returned by Read when trying to read past the end of the file
|
|
|
|
ErrInvalidRange = errors.New("Invalid range")
|
|
|
|
)
|
2015-11-29 23:49:41 +08:00
|
|
|
|
|
|
|
type ChunkInfo struct {
|
2015-12-02 16:35:16 +08:00
|
|
|
Fid string `json:"fid"`
|
|
|
|
Offset int64 `json:"offset"`
|
|
|
|
Size int64 `json:"size"`
|
2015-11-29 23:49:41 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
type ChunkList []*ChunkInfo
|
|
|
|
|
2015-12-01 20:23:50 +08:00
|
|
|
type ChunkManifest struct {
|
2015-12-02 21:27:29 +08:00
|
|
|
Name string `json:"name,omitempty"`
|
|
|
|
Mime string `json:"mime,omitempty"`
|
|
|
|
Size int64 `json:"size,omitempty"`
|
|
|
|
Chunks ChunkList `json:"chunks,omitempty"`
|
2015-12-01 20:23:50 +08:00
|
|
|
}
|
2015-11-29 23:49:41 +08:00
|
|
|
|
2015-12-01 20:23:50 +08:00
|
|
|
// seekable chunked file reader
|
|
|
|
type ChunkedFileReader struct {
|
2020-03-21 06:46:16 +08:00
|
|
|
totalSize int64
|
|
|
|
chunkList []*ChunkInfo
|
|
|
|
master string
|
|
|
|
pos int64
|
|
|
|
pr *io.PipeReader
|
|
|
|
pw *io.PipeWriter
|
|
|
|
mutex sync.Mutex
|
2015-11-29 23:49:41 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
func (s ChunkList) Len() int { return len(s) }
|
|
|
|
func (s ChunkList) Less(i, j int) bool { return s[i].Offset < s[j].Offset }
|
|
|
|
func (s ChunkList) Swap(i, j int) { s[i], s[j] = s[j], s[i] }
|
|
|
|
|
2020-06-20 13:45:27 +08:00
|
|
|
func LoadChunkManifest(buffer []byte, isCompressed bool) (*ChunkManifest, error) {
|
|
|
|
if isCompressed {
|
2015-12-02 21:27:29 +08:00
|
|
|
var err error
|
2020-06-20 23:16:16 +08:00
|
|
|
if buffer, err = util.DecompressData(buffer); err != nil {
|
2020-08-02 04:46:52 +08:00
|
|
|
glog.V(0).Infof("fail to decompress chunk manifest: %v", err)
|
2015-12-02 21:27:29 +08:00
|
|
|
}
|
|
|
|
}
|
2015-12-01 20:23:50 +08:00
|
|
|
cm := ChunkManifest{}
|
2015-12-02 16:35:16 +08:00
|
|
|
if e := json.Unmarshal(buffer, &cm); e != nil {
|
2015-11-29 23:49:41 +08:00
|
|
|
return nil, e
|
|
|
|
}
|
2015-12-01 20:23:50 +08:00
|
|
|
sort.Sort(cm.Chunks)
|
|
|
|
return &cm, nil
|
|
|
|
}
|
|
|
|
|
2015-12-14 22:01:30 +08:00
|
|
|
func (cm *ChunkManifest) Marshal() ([]byte, error) {
|
2015-12-01 20:23:50 +08:00
|
|
|
return json.Marshal(cm)
|
|
|
|
}
|
|
|
|
|
2020-03-19 02:16:45 +08:00
|
|
|
func (cm *ChunkManifest) DeleteChunks(master string, usePublicUrl bool, grpcDialOption grpc.DialOption) error {
|
2018-10-14 15:12:28 +08:00
|
|
|
var fileIds []string
|
2015-12-03 16:27:02 +08:00
|
|
|
for _, ci := range cm.Chunks {
|
2018-10-14 15:12:28 +08:00
|
|
|
fileIds = append(fileIds, ci.Fid)
|
2015-12-01 20:23:50 +08:00
|
|
|
}
|
2020-03-19 02:16:45 +08:00
|
|
|
results, err := DeleteFiles(master, usePublicUrl, grpcDialOption, fileIds)
|
2018-10-14 15:12:28 +08:00
|
|
|
if err != nil {
|
|
|
|
glog.V(0).Infof("delete %+v: %v", fileIds, err)
|
|
|
|
return fmt.Errorf("chunk delete: %v", err)
|
2015-12-01 20:23:50 +08:00
|
|
|
}
|
2018-10-14 15:12:28 +08:00
|
|
|
for _, result := range results {
|
|
|
|
if result.Error != "" {
|
|
|
|
glog.V(0).Infof("delete file %+v: %v", result.FileId, result.Error)
|
|
|
|
return fmt.Errorf("chunk delete %v: %v", result.FileId, result.Error)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2015-12-01 20:23:50 +08:00
|
|
|
return nil
|
2015-11-29 23:49:41 +08:00
|
|
|
}
|
|
|
|
|
2015-12-03 21:35:33 +08:00
|
|
|
func readChunkNeedle(fileUrl string, w io.Writer, offset int64) (written int64, e error) {
|
2015-11-29 23:49:41 +08:00
|
|
|
req, err := http.NewRequest("GET", fileUrl, nil)
|
|
|
|
if err != nil {
|
|
|
|
return written, err
|
|
|
|
}
|
2015-12-01 20:23:50 +08:00
|
|
|
if offset > 0 {
|
|
|
|
req.Header.Set("Range", fmt.Sprintf("bytes=%d-", offset))
|
2015-11-29 23:49:41 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
resp, err := util.Do(req)
|
|
|
|
if err != nil {
|
|
|
|
return written, err
|
|
|
|
}
|
2019-04-15 14:28:24 +08:00
|
|
|
defer func() {
|
|
|
|
io.Copy(ioutil.Discard, resp.Body)
|
|
|
|
resp.Body.Close()
|
|
|
|
}()
|
2015-11-29 23:49:41 +08:00
|
|
|
|
2015-12-01 20:23:50 +08:00
|
|
|
switch resp.StatusCode {
|
|
|
|
case http.StatusRequestedRangeNotSatisfiable:
|
|
|
|
return written, ErrInvalidRange
|
|
|
|
case http.StatusOK:
|
|
|
|
if offset > 0 {
|
|
|
|
return written, ErrRangeRequestsNotSupported
|
|
|
|
}
|
|
|
|
case http.StatusPartialContent:
|
|
|
|
break
|
|
|
|
default:
|
2015-12-03 21:35:33 +08:00
|
|
|
return written, fmt.Errorf("Read chunk needle error: [%d] %s", resp.StatusCode, fileUrl)
|
2015-12-01 20:23:50 +08:00
|
|
|
|
2015-11-29 23:49:41 +08:00
|
|
|
}
|
2015-12-01 20:23:50 +08:00
|
|
|
return io.Copy(w, resp.Body)
|
2015-11-29 23:49:41 +08:00
|
|
|
}
|
|
|
|
|
2020-03-21 06:46:16 +08:00
|
|
|
func NewChunkedFileReader(chunkList []*ChunkInfo, master string) *ChunkedFileReader {
|
|
|
|
var totalSize int64
|
|
|
|
for _, chunk := range chunkList {
|
|
|
|
totalSize += chunk.Size
|
|
|
|
}
|
2020-03-21 11:31:11 +08:00
|
|
|
sort.Sort(ChunkList(chunkList))
|
2020-03-21 06:38:04 +08:00
|
|
|
return &ChunkedFileReader{
|
2020-03-21 06:46:16 +08:00
|
|
|
totalSize: totalSize,
|
|
|
|
chunkList: chunkList,
|
|
|
|
master: master,
|
2020-03-21 06:38:04 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2015-12-01 20:23:50 +08:00
|
|
|
func (cf *ChunkedFileReader) Seek(offset int64, whence int) (int64, error) {
|
|
|
|
var err error
|
|
|
|
switch whence {
|
2020-03-22 13:16:00 +08:00
|
|
|
case io.SeekStart:
|
|
|
|
case io.SeekCurrent:
|
2015-12-01 20:23:50 +08:00
|
|
|
offset += cf.pos
|
2020-03-22 13:16:00 +08:00
|
|
|
case io.SeekEnd:
|
|
|
|
offset = cf.totalSize + offset
|
2015-12-01 20:23:50 +08:00
|
|
|
}
|
2020-03-21 06:46:16 +08:00
|
|
|
if offset > cf.totalSize {
|
2015-12-01 20:23:50 +08:00
|
|
|
err = ErrInvalidRange
|
2015-11-29 23:49:41 +08:00
|
|
|
}
|
2015-12-01 20:23:50 +08:00
|
|
|
if cf.pos != offset {
|
|
|
|
cf.Close()
|
|
|
|
}
|
|
|
|
cf.pos = offset
|
|
|
|
return cf.pos, err
|
|
|
|
}
|
|
|
|
|
|
|
|
func (cf *ChunkedFileReader) WriteTo(w io.Writer) (n int64, err error) {
|
2015-11-29 23:49:41 +08:00
|
|
|
chunkIndex := -1
|
2015-11-30 00:21:42 +08:00
|
|
|
chunkStartOffset := int64(0)
|
2020-03-21 06:46:16 +08:00
|
|
|
for i, ci := range cf.chunkList {
|
2015-12-01 20:23:50 +08:00
|
|
|
if cf.pos >= ci.Offset && cf.pos < ci.Offset+ci.Size {
|
2015-11-29 23:49:41 +08:00
|
|
|
chunkIndex = i
|
2015-12-01 20:23:50 +08:00
|
|
|
chunkStartOffset = cf.pos - ci.Offset
|
2015-11-29 23:49:41 +08:00
|
|
|
break
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if chunkIndex < 0 {
|
2015-12-01 20:23:50 +08:00
|
|
|
return n, ErrInvalidRange
|
2015-11-29 23:49:41 +08:00
|
|
|
}
|
2020-03-21 06:46:16 +08:00
|
|
|
for ; chunkIndex < len(cf.chunkList); chunkIndex++ {
|
|
|
|
ci := cf.chunkList[chunkIndex]
|
2015-12-01 20:23:50 +08:00
|
|
|
// if we need read date from local volume server first?
|
2020-03-21 06:46:16 +08:00
|
|
|
fileUrl, lookupError := LookupFileId(cf.master, ci.Fid)
|
2015-11-29 23:49:41 +08:00
|
|
|
if lookupError != nil {
|
2015-12-01 20:23:50 +08:00
|
|
|
return n, lookupError
|
2015-11-29 23:49:41 +08:00
|
|
|
}
|
2015-12-03 21:35:33 +08:00
|
|
|
if wn, e := readChunkNeedle(fileUrl, w, chunkStartOffset); e != nil {
|
2015-12-01 20:23:50 +08:00
|
|
|
return n, e
|
2015-11-29 23:49:41 +08:00
|
|
|
} else {
|
2015-12-01 20:23:50 +08:00
|
|
|
n += wn
|
|
|
|
cf.pos += wn
|
2015-11-29 23:49:41 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
chunkStartOffset = 0
|
|
|
|
}
|
2015-12-01 20:23:50 +08:00
|
|
|
return n, nil
|
|
|
|
}
|
2015-11-29 23:49:41 +08:00
|
|
|
|
2015-12-01 20:23:50 +08:00
|
|
|
func (cf *ChunkedFileReader) ReadAt(p []byte, off int64) (n int, err error) {
|
|
|
|
cf.Seek(off, 0)
|
|
|
|
return cf.Read(p)
|
2015-11-29 23:49:41 +08:00
|
|
|
}
|
|
|
|
|
2015-12-01 20:23:50 +08:00
|
|
|
func (cf *ChunkedFileReader) Read(p []byte) (int, error) {
|
|
|
|
return cf.getPipeReader().Read(p)
|
2015-11-29 23:49:41 +08:00
|
|
|
}
|
|
|
|
|
2015-12-01 20:23:50 +08:00
|
|
|
func (cf *ChunkedFileReader) Close() (e error) {
|
|
|
|
cf.mutex.Lock()
|
|
|
|
defer cf.mutex.Unlock()
|
|
|
|
return cf.closePipe()
|
|
|
|
}
|
|
|
|
|
|
|
|
func (cf *ChunkedFileReader) closePipe() (e error) {
|
|
|
|
if cf.pr != nil {
|
|
|
|
if err := cf.pr.Close(); err != nil {
|
|
|
|
e = err
|
|
|
|
}
|
|
|
|
}
|
|
|
|
cf.pr = nil
|
|
|
|
if cf.pw != nil {
|
|
|
|
if err := cf.pw.Close(); err != nil {
|
|
|
|
e = err
|
|
|
|
}
|
|
|
|
}
|
|
|
|
cf.pw = nil
|
|
|
|
return e
|
|
|
|
}
|
|
|
|
|
|
|
|
func (cf *ChunkedFileReader) getPipeReader() io.Reader {
|
|
|
|
cf.mutex.Lock()
|
|
|
|
defer cf.mutex.Unlock()
|
|
|
|
if cf.pr != nil && cf.pw != nil {
|
|
|
|
return cf.pr
|
|
|
|
}
|
|
|
|
cf.closePipe()
|
|
|
|
cf.pr, cf.pw = io.Pipe()
|
|
|
|
go func(pw *io.PipeWriter) {
|
|
|
|
_, e := cf.WriteTo(pw)
|
|
|
|
pw.CloseWithError(e)
|
|
|
|
}(cf.pw)
|
|
|
|
return cf.pr
|
2015-11-29 23:49:41 +08:00
|
|
|
}
|