2016-06-03 11:05:34 +08:00
|
|
|
package weed_server
|
|
|
|
|
|
|
|
import (
|
2019-03-16 06:26:09 +08:00
|
|
|
"context"
|
2016-06-03 11:05:34 +08:00
|
|
|
"net/http"
|
2019-02-15 16:09:19 +08:00
|
|
|
"os"
|
2017-01-08 22:34:26 +08:00
|
|
|
"strings"
|
2018-05-28 14:59:49 +08:00
|
|
|
"time"
|
2017-01-08 09:16:29 +08:00
|
|
|
|
2016-06-03 11:05:34 +08:00
|
|
|
"github.com/chrislusf/seaweedfs/weed/glog"
|
|
|
|
"github.com/chrislusf/seaweedfs/weed/operation"
|
2018-05-28 02:52:26 +08:00
|
|
|
"github.com/chrislusf/seaweedfs/weed/pb/filer_pb"
|
2019-02-15 16:09:19 +08:00
|
|
|
"github.com/chrislusf/seaweedfs/weed/security"
|
2019-06-16 03:21:44 +08:00
|
|
|
"github.com/chrislusf/seaweedfs/weed/stats"
|
2020-03-09 15:16:10 +08:00
|
|
|
"github.com/chrislusf/seaweedfs/weed/storage/needle"
|
2016-06-03 11:05:34 +08:00
|
|
|
"github.com/chrislusf/seaweedfs/weed/util"
|
2018-07-22 08:47:59 +08:00
|
|
|
)
|
|
|
|
|
|
|
|
var (
|
|
|
|
OS_UID = uint32(os.Getuid())
|
|
|
|
OS_GID = uint32(os.Getgid())
|
2016-06-03 11:05:34 +08:00
|
|
|
)
|
|
|
|
|
2016-06-03 11:44:50 +08:00
|
|
|
type FilerPostResult struct {
|
|
|
|
Name string `json:"name,omitempty"`
|
2020-02-04 09:04:06 +08:00
|
|
|
Size int64 `json:"size,omitempty"`
|
2016-06-03 11:44:50 +08:00
|
|
|
Error string `json:"error,omitempty"`
|
|
|
|
Fid string `json:"fid,omitempty"`
|
|
|
|
Url string `json:"url,omitempty"`
|
|
|
|
}
|
|
|
|
|
2020-07-20 08:59:43 +08:00
|
|
|
func (fs *FilerServer) assignNewFileInfo(replication, collection, dataCenter, ttlString string, fsync bool) (fileId, urlLocation string, auth security.EncodedJwt, err error) {
|
2019-06-23 13:53:52 +08:00
|
|
|
|
|
|
|
stats.FilerRequestCounter.WithLabelValues("assign").Inc()
|
|
|
|
start := time.Now()
|
|
|
|
defer func() { stats.FilerRequestHistogram.WithLabelValues("assign").Observe(time.Since(start).Seconds()) }()
|
|
|
|
|
2016-06-26 10:50:18 +08:00
|
|
|
ar := &operation.VolumeAssignRequest{
|
|
|
|
Count: 1,
|
|
|
|
Replication: replication,
|
|
|
|
Collection: collection,
|
2020-03-09 15:16:10 +08:00
|
|
|
Ttl: ttlString,
|
2018-07-15 04:36:28 +08:00
|
|
|
DataCenter: dataCenter,
|
2016-06-26 10:50:18 +08:00
|
|
|
}
|
2018-07-10 15:20:50 +08:00
|
|
|
var altRequest *operation.VolumeAssignRequest
|
2018-07-15 04:36:28 +08:00
|
|
|
if dataCenter != "" {
|
2018-07-10 15:20:50 +08:00
|
|
|
altRequest = &operation.VolumeAssignRequest{
|
|
|
|
Count: 1,
|
|
|
|
Replication: replication,
|
|
|
|
Collection: collection,
|
2020-03-09 15:16:10 +08:00
|
|
|
Ttl: ttlString,
|
2018-07-10 15:20:50 +08:00
|
|
|
DataCenter: "",
|
|
|
|
}
|
|
|
|
}
|
2018-07-22 01:39:02 +08:00
|
|
|
|
2019-02-19 04:11:52 +08:00
|
|
|
assignResult, ae := operation.Assign(fs.filer.GetMaster(), fs.grpcDialOption, ar, altRequest)
|
2016-06-08 15:46:14 +08:00
|
|
|
if ae != nil {
|
2018-07-22 01:39:02 +08:00
|
|
|
glog.Errorf("failing to assign a file id: %v", ae)
|
2016-06-08 15:46:14 +08:00
|
|
|
err = ae
|
|
|
|
return
|
|
|
|
}
|
|
|
|
fileId = assignResult.Fid
|
|
|
|
urlLocation = "http://" + assignResult.Url + "/" + assignResult.Fid
|
2020-04-12 14:37:10 +08:00
|
|
|
if fsync {
|
|
|
|
urlLocation += "?fsync=true"
|
|
|
|
}
|
2019-02-15 16:09:19 +08:00
|
|
|
auth = assignResult.Auth
|
2016-06-08 15:46:14 +08:00
|
|
|
return
|
|
|
|
}
|
|
|
|
|
2016-06-03 11:05:34 +08:00
|
|
|
func (fs *FilerServer) PostHandler(w http.ResponseWriter, r *http.Request) {
|
2016-08-06 06:01:30 +08:00
|
|
|
|
2019-03-16 06:55:34 +08:00
|
|
|
ctx := context.Background()
|
|
|
|
|
2016-06-03 11:05:34 +08:00
|
|
|
query := r.URL.Query()
|
2020-04-12 14:37:10 +08:00
|
|
|
collection, replication, fsync := fs.detectCollection(r.RequestURI, query.Get("collection"), query.Get("replication"))
|
2018-07-15 04:36:28 +08:00
|
|
|
dataCenter := query.Get("dataCenter")
|
|
|
|
if dataCenter == "" {
|
|
|
|
dataCenter = fs.option.DataCenter
|
|
|
|
}
|
2020-03-09 15:16:10 +08:00
|
|
|
ttlString := r.URL.Query().Get("ttl")
|
2016-06-03 11:05:34 +08:00
|
|
|
|
2020-03-09 15:16:10 +08:00
|
|
|
// read ttl in seconds
|
|
|
|
ttl, err := needle.ReadTTL(ttlString)
|
|
|
|
ttlSeconds := int32(0)
|
|
|
|
if err == nil {
|
|
|
|
ttlSeconds = int32(ttl.Minutes()) * 60
|
|
|
|
}
|
|
|
|
|
2020-08-09 03:02:06 +08:00
|
|
|
fs.autoChunk(ctx, w, r, replication, collection, dataCenter, ttlSeconds, ttlString, fsync)
|
2019-06-23 13:53:52 +08:00
|
|
|
|
2016-06-03 11:05:34 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
// curl -X DELETE http://localhost:8888/path/to
|
2018-07-25 13:33:26 +08:00
|
|
|
// curl -X DELETE http://localhost:8888/path/to?recursive=true
|
2019-09-12 11:26:20 +08:00
|
|
|
// curl -X DELETE http://localhost:8888/path/to?recursive=true&ignoreRecursiveError=true
|
2019-12-12 06:58:22 +08:00
|
|
|
// curl -X DELETE http://localhost:8888/path/to?recursive=true&skipChunkDeletion=true
|
2016-06-03 11:05:34 +08:00
|
|
|
func (fs *FilerServer) DeleteHandler(w http.ResponseWriter, r *http.Request) {
|
2018-05-14 14:56:16 +08:00
|
|
|
|
2018-07-25 13:33:26 +08:00
|
|
|
isRecursive := r.FormValue("recursive") == "true"
|
2020-01-01 03:52:54 +08:00
|
|
|
if !isRecursive && fs.option.recursiveDelete {
|
|
|
|
if r.FormValue("recursive") != "false" {
|
|
|
|
isRecursive = true
|
|
|
|
}
|
|
|
|
}
|
2019-09-12 11:26:20 +08:00
|
|
|
ignoreRecursiveError := r.FormValue("ignoreRecursiveError") == "true"
|
2019-12-12 06:58:22 +08:00
|
|
|
skipChunkDeletion := r.FormValue("skipChunkDeletion") == "true"
|
2018-07-25 13:33:26 +08:00
|
|
|
|
2020-06-23 01:01:00 +08:00
|
|
|
objectPath := r.URL.Path
|
|
|
|
if len(r.URL.Path) > 1 && strings.HasSuffix(objectPath, "/") {
|
|
|
|
objectPath = objectPath[0 : len(objectPath)-1]
|
|
|
|
}
|
|
|
|
|
2020-08-29 14:48:48 +08:00
|
|
|
err := fs.filer.DeleteEntryMetaAndData(context.Background(), util.FullPath(objectPath), isRecursive, ignoreRecursiveError, !skipChunkDeletion, false, nil)
|
2018-05-14 14:56:16 +08:00
|
|
|
if err != nil {
|
2020-06-23 01:01:00 +08:00
|
|
|
glog.V(1).Infoln("deleting", objectPath, ":", err.Error())
|
2019-12-19 13:04:40 +08:00
|
|
|
httpStatus := http.StatusInternalServerError
|
2020-03-08 09:01:39 +08:00
|
|
|
if err == filer_pb.ErrNotFound {
|
2019-12-19 13:04:40 +08:00
|
|
|
httpStatus = http.StatusNotFound
|
|
|
|
}
|
|
|
|
writeJsonError(w, r, httpStatus, err)
|
2018-05-14 14:56:16 +08:00
|
|
|
return
|
|
|
|
}
|
|
|
|
|
2018-07-22 09:49:47 +08:00
|
|
|
w.WriteHeader(http.StatusNoContent)
|
2016-06-03 11:05:34 +08:00
|
|
|
}
|
2020-02-25 14:28:45 +08:00
|
|
|
|
2020-04-12 14:37:10 +08:00
|
|
|
func (fs *FilerServer) detectCollection(requestURI, qCollection, qReplication string) (collection, replication string, fsync bool) {
|
2020-02-25 14:28:45 +08:00
|
|
|
// default
|
|
|
|
collection = fs.option.Collection
|
|
|
|
replication = fs.option.DefaultReplication
|
|
|
|
|
|
|
|
// get default collection settings
|
|
|
|
if qCollection != "" {
|
|
|
|
collection = qCollection
|
|
|
|
}
|
|
|
|
if qReplication != "" {
|
|
|
|
replication = qReplication
|
|
|
|
}
|
|
|
|
|
|
|
|
// required by buckets folder
|
|
|
|
if strings.HasPrefix(requestURI, fs.filer.DirBucketsPath+"/") {
|
|
|
|
bucketAndObjectKey := requestURI[len(fs.filer.DirBucketsPath)+1:]
|
|
|
|
t := strings.Index(bucketAndObjectKey, "/")
|
|
|
|
if t < 0 {
|
|
|
|
collection = bucketAndObjectKey
|
|
|
|
}
|
|
|
|
if t > 0 {
|
|
|
|
collection = bucketAndObjectKey[:t]
|
|
|
|
}
|
2020-04-12 14:37:10 +08:00
|
|
|
replication, fsync = fs.filer.ReadBucketOption(collection)
|
2020-02-25 14:28:45 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
return
|
|
|
|
}
|