seaweedfs/weed/s3api/s3api_object_handlers.go

441 lines
12 KiB
Go
Raw Normal View History

package s3api
import (
2018-09-12 16:00:57 +08:00
"crypto/md5"
2018-07-22 08:39:10 +08:00
"encoding/json"
2020-02-26 06:38:36 +08:00
"encoding/xml"
"fmt"
2021-03-10 22:52:41 +08:00
"github.com/chrislusf/seaweedfs/weed/filer"
"github.com/pquerna/cachecontrol/cacheobject"
2018-07-22 16:15:11 +08:00
"io"
"io/ioutil"
2018-07-22 08:39:10 +08:00
"net/http"
"net/url"
"sort"
2018-07-23 16:15:59 +08:00
"strings"
"time"
2018-09-12 15:46:12 +08:00
"github.com/chrislusf/seaweedfs/weed/s3api/s3err"
"github.com/gorilla/mux"
2018-09-12 15:46:12 +08:00
"github.com/chrislusf/seaweedfs/weed/glog"
"github.com/chrislusf/seaweedfs/weed/pb/filer_pb"
weed_server "github.com/chrislusf/seaweedfs/weed/server"
2020-02-15 01:46:36 +08:00
"github.com/chrislusf/seaweedfs/weed/util"
)
var (
client *http.Client
)
func init() {
client = &http.Client{Transport: &http.Transport{
MaxIdleConns: 1024,
MaxIdleConnsPerHost: 1024,
}}
}
func (s3a *S3ApiServer) PutObjectHandler(w http.ResponseWriter, r *http.Request) {
// http://docs.aws.amazon.com/AmazonS3/latest/dev/UploadingObjects.html
2020-07-27 03:58:58 +08:00
bucket, object := getBucketAndObject(r)
_, err := validateContentMd5(r.Header)
if err != nil {
2020-09-20 05:09:58 +08:00
writeErrorResponse(w, s3err.ErrInvalidDigest, r.URL)
return
}
if r.Header.Get("Cache-Control") != "" {
if _, err = cacheobject.ParseRequestCacheControl(r.Header.Get("Cache-Control")); err != nil {
writeErrorResponse(w, s3err.ErrInvalidDigest, r.URL)
return
}
}
if r.Header.Get("Expires") != "" {
if _, err = time.Parse(http.TimeFormat, r.Header.Get("Expires")); err != nil {
writeErrorResponse(w, s3err.ErrInvalidDigest, r.URL)
return
}
}
dataReader := r.Body
if s3a.iam.isEnabled() {
rAuthType := getRequestAuthType(r)
2020-09-20 05:09:58 +08:00
var s3ErrCode s3err.ErrorCode
switch rAuthType {
case authTypeStreamingSigned:
dataReader, s3ErrCode = s3a.iam.newSignV4ChunkedReader(r)
case authTypeSignedV2, authTypePresignedV2:
_, s3ErrCode = s3a.iam.isReqAuthenticatedV2(r)
case authTypePresigned, authTypeSigned:
_, s3ErrCode = s3a.iam.reqSignatureV4Verify(r)
}
2020-09-20 05:09:58 +08:00
if s3ErrCode != s3err.ErrNone {
writeErrorResponse(w, s3ErrCode, r.URL)
return
}
} else {
rAuthType := getRequestAuthType(r)
if authTypeAnonymous != rAuthType {
writeErrorResponse(w, s3err.ErrAuthNotSetup, r.URL)
return
}
}
2020-02-15 01:09:15 +08:00
defer dataReader.Close()
2020-07-28 00:58:42 +08:00
if strings.HasSuffix(object, "/") {
if err := s3a.mkdir(s3a.option.BucketsPath, bucket+object, nil); err != nil {
2020-09-20 05:09:58 +08:00
writeErrorResponse(w, s3err.ErrInternalError, r.URL)
2020-07-28 00:58:42 +08:00
return
}
} else {
uploadUrl := fmt.Sprintf("http://%s%s/%s%s", s3a.option.Filer, s3a.option.BucketsPath, bucket, urlPathEscape(object))
2020-07-28 00:58:42 +08:00
etag, errCode := s3a.putToFiler(r, uploadUrl, dataReader)
2020-09-20 05:09:58 +08:00
if errCode != s3err.ErrNone {
2020-07-28 00:58:42 +08:00
writeErrorResponse(w, errCode, r.URL)
return
}
setEtag(w, etag)
}
writeSuccessResponseEmpty(w)
}
2018-07-22 09:49:47 +08:00
func urlPathEscape(object string) string {
var escapedParts []string
for _, part := range strings.Split(object, "/") {
escapedParts = append(escapedParts, url.PathEscape(part))
}
return strings.Join(escapedParts, "/")
}
2018-07-22 09:49:47 +08:00
func (s3a *S3ApiServer) GetObjectHandler(w http.ResponseWriter, r *http.Request) {
2020-07-27 03:58:58 +08:00
bucket, object := getBucketAndObject(r)
2018-09-20 13:01:41 +08:00
2018-07-23 16:15:59 +08:00
if strings.HasSuffix(r.URL.Path, "/") {
2020-09-20 05:09:58 +08:00
writeErrorResponse(w, s3err.ErrNotImplemented, r.URL)
2018-07-23 16:15:59 +08:00
return
}
2018-09-20 13:01:41 +08:00
destUrl := fmt.Sprintf("http://%s%s/%s%s",
s3a.option.Filer, s3a.option.BucketsPath, bucket, urlPathEscape(object))
2018-07-22 09:49:47 +08:00
2019-01-02 18:38:32 +08:00
s3a.proxyToFiler(w, r, destUrl, passThroughResponse)
2018-07-22 09:49:47 +08:00
}
func (s3a *S3ApiServer) HeadObjectHandler(w http.ResponseWriter, r *http.Request) {
2020-07-27 03:58:58 +08:00
bucket, object := getBucketAndObject(r)
2018-09-20 13:01:41 +08:00
destUrl := fmt.Sprintf("http://%s%s/%s%s",
s3a.option.Filer, s3a.option.BucketsPath, bucket, urlPathEscape(object))
2018-07-22 09:49:47 +08:00
2019-01-02 18:38:32 +08:00
s3a.proxyToFiler(w, r, destUrl, passThroughResponse)
2018-07-22 09:49:47 +08:00
}
func (s3a *S3ApiServer) DeleteObjectHandler(w http.ResponseWriter, r *http.Request) {
2020-07-27 03:58:58 +08:00
bucket, object := getBucketAndObject(r)
2018-09-20 13:01:41 +08:00
destUrl := fmt.Sprintf("http://%s%s/%s%s?recursive=true",
s3a.option.Filer, s3a.option.BucketsPath, bucket, urlPathEscape(object))
2018-07-22 09:49:47 +08:00
s3a.proxyToFiler(w, r, destUrl, func(proxyResponse *http.Response, w http.ResponseWriter) {
for k, v := range proxyResponse.Header {
2018-07-22 10:12:44 +08:00
w.Header()[k] = v
}
2020-06-12 00:00:34 +08:00
w.WriteHeader(http.StatusNoContent)
2018-07-22 10:12:44 +08:00
})
2018-07-22 09:49:47 +08:00
}
2020-07-27 03:58:58 +08:00
// / ObjectIdentifier carries key name for the object to delete.
2020-02-26 06:38:36 +08:00
type ObjectIdentifier struct {
ObjectName string `xml:"Key"`
}
// DeleteObjectsRequest - xml carrying the object key names which needs to be deleted.
type DeleteObjectsRequest struct {
// Element to enable quiet mode for the request
Quiet bool
// List of objects to be deleted
Objects []ObjectIdentifier `xml:"Object"`
}
// DeleteError structure.
type DeleteError struct {
Code string
Message string
Key string
}
// DeleteObjectsResponse container for multiple object deletes.
type DeleteObjectsResponse struct {
XMLName xml.Name `xml:"http://s3.amazonaws.com/doc/2006-03-01/ DeleteResult" json:"-"`
// Collection of all deleted objects
DeletedObjects []ObjectIdentifier `xml:"Deleted,omitempty"`
// Collection of errors deleting certain objects.
Errors []DeleteError `xml:"Error,omitempty"`
}
// DeleteMultipleObjectsHandler - Delete multiple objects
func (s3a *S3ApiServer) DeleteMultipleObjectsHandler(w http.ResponseWriter, r *http.Request) {
2020-02-26 06:38:36 +08:00
2020-07-27 03:58:58 +08:00
bucket, _ := getBucketAndObject(r)
2020-02-26 06:38:36 +08:00
deleteXMLBytes, err := ioutil.ReadAll(r.Body)
if err != nil {
2020-09-20 05:09:58 +08:00
writeErrorResponse(w, s3err.ErrInternalError, r.URL)
2020-02-26 06:38:36 +08:00
return
}
deleteObjects := &DeleteObjectsRequest{}
if err := xml.Unmarshal(deleteXMLBytes, deleteObjects); err != nil {
2020-09-20 05:09:58 +08:00
writeErrorResponse(w, s3err.ErrMalformedXML, r.URL)
2020-02-26 06:38:36 +08:00
return
}
var deletedObjects []ObjectIdentifier
var deleteErrors []DeleteError
2021-01-29 05:20:06 +08:00
directoriesWithDeletion := make(map[string]int)
2020-03-23 14:52:55 +08:00
s3a.WithFilerClient(func(client filer_pb.SeaweedFilerClient) error {
2021-01-29 05:20:06 +08:00
// delete file entries
for _, object := range deleteObjects.Objects {
lastSeparator := strings.LastIndex(object.ObjectName, "/")
parentDirectoryPath, entryName, isDeleteData, isRecursive := "", object.ObjectName, true, false
if lastSeparator > 0 && lastSeparator+1 < len(object.ObjectName) {
entryName = object.ObjectName[lastSeparator+1:]
parentDirectoryPath = "/" + object.ObjectName[:lastSeparator]
}
parentDirectoryPath = fmt.Sprintf("%s/%s%s", s3a.option.BucketsPath, bucket, parentDirectoryPath)
err := doDeleteEntry(client, parentDirectoryPath, entryName, isDeleteData, isRecursive)
if err == nil {
2021-01-29 05:20:06 +08:00
directoriesWithDeletion[parentDirectoryPath]++
deletedObjects = append(deletedObjects, object)
2021-03-10 22:52:41 +08:00
} else if strings.Contains(err.Error(), filer.MsgFailDelNonEmptyFolder) {
deletedObjects = append(deletedObjects, object)
} else {
2021-01-29 05:20:06 +08:00
delete(directoriesWithDeletion, parentDirectoryPath)
deleteErrors = append(deleteErrors, DeleteError{
Code: "",
Message: err.Error(),
Key: object.ObjectName,
})
}
2020-02-26 06:38:36 +08:00
}
2021-01-29 05:20:06 +08:00
// purge empty folders, only checking folders with deletions
for len(directoriesWithDeletion) > 0 {
directoriesWithDeletion = s3a.doDeleteEmptyDirectories(client, directoriesWithDeletion)
2021-01-29 05:20:06 +08:00
}
return nil
2020-02-26 06:38:36 +08:00
})
deleteResp := DeleteObjectsResponse{}
if !deleteObjects.Quiet {
deleteResp.DeletedObjects = deletedObjects
}
deleteResp.Errors = deleteErrors
writeSuccessResponseXML(w, encodeResponse(deleteResp))
}
func (s3a *S3ApiServer) doDeleteEmptyDirectories(client filer_pb.SeaweedFilerClient, directoriesWithDeletion map[string]int) (newDirectoriesWithDeletion map[string]int) {
var allDirs []string
for dir, _ := range directoriesWithDeletion {
allDirs = append(allDirs, dir)
}
sort.Slice(allDirs, func(i, j int) bool {
return len(allDirs[i]) > len(allDirs[j])
})
newDirectoriesWithDeletion = make(map[string]int)
for _, dir := range allDirs {
parentDir, dirName := util.FullPath(dir).DirAndName()
if parentDir == s3a.option.BucketsPath {
continue
}
if err := doDeleteEntry(client, parentDir, dirName, false, false); err != nil {
glog.V(4).Infof("directory %s has %d deletion but still not empty: %v", dir, directoriesWithDeletion[dir], err)
} else {
newDirectoriesWithDeletion[parentDir]++
}
}
return
}
var passThroughHeaders = []string{
"response-cache-control",
"response-content-disposition",
"response-content-encoding",
"response-content-language",
"response-content-type",
"response-expires",
}
2020-06-12 01:53:25 +08:00
func (s3a *S3ApiServer) proxyToFiler(w http.ResponseWriter, r *http.Request, destUrl string, responseFn func(proxyResponse *http.Response, w http.ResponseWriter)) {
2018-07-22 09:49:47 +08:00
glog.V(2).Infof("s3 proxying %s to %s", r.Method, destUrl)
proxyReq, err := http.NewRequest(r.Method, destUrl, r.Body)
if err != nil {
glog.Errorf("NewRequest %s: %v", destUrl, err)
2020-09-20 05:09:58 +08:00
writeErrorResponse(w, s3err.ErrInternalError, r.URL)
2018-07-22 09:49:47 +08:00
return
}
proxyReq.Header.Set("Host", s3a.option.Filer)
proxyReq.Header.Set("X-Forwarded-For", r.RemoteAddr)
for header, values := range r.Header {
// handle s3 related headers
passed := false
for _, h := range passThroughHeaders {
if strings.ToLower(header) == h && len(values) > 0 {
proxyReq.Header.Add(header[len("response-"):], values[0])
passed = true
break
}
}
if passed {
continue
}
// handle other headers
2018-07-22 09:49:47 +08:00
for _, value := range values {
proxyReq.Header.Add(header, value)
}
}
resp, postErr := client.Do(proxyReq)
if postErr != nil {
glog.Errorf("post to filer: %v", postErr)
2020-09-20 05:09:58 +08:00
writeErrorResponse(w, s3err.ErrInternalError, r.URL)
2018-07-22 09:49:47 +08:00
return
}
2020-02-15 01:46:36 +08:00
defer util.CloseResponse(resp)
2021-03-18 02:41:34 +08:00
2021-05-24 19:59:44 +08:00
if resp.StatusCode == http.StatusPreconditionFailed {
writeErrorResponse(w, s3err.ErrPreconditionFailed, r.URL)
return
}
if (resp.ContentLength == -1 || resp.StatusCode == 404) && resp.StatusCode != 304 {
if r.Method != "DELETE" {
writeErrorResponse(w, s3err.ErrNoSuchKey, r.URL)
return
}
}
2018-07-22 10:12:44 +08:00
responseFn(resp, w)
2020-02-15 01:46:36 +08:00
2018-07-22 10:12:44 +08:00
}
2020-06-12 01:53:25 +08:00
func passThroughResponse(proxyResponse *http.Response, w http.ResponseWriter) {
for k, v := range proxyResponse.Header {
2018-07-22 09:49:47 +08:00
w.Header()[k] = v
}
if proxyResponse.Header.Get("Content-Range") != "" && proxyResponse.StatusCode == 200 {
w.WriteHeader(http.StatusPartialContent)
} else {
w.WriteHeader(proxyResponse.StatusCode)
}
2020-06-12 01:53:25 +08:00
io.Copy(w, proxyResponse.Body)
2018-07-22 09:49:47 +08:00
}
2020-09-20 05:09:58 +08:00
func (s3a *S3ApiServer) putToFiler(r *http.Request, uploadUrl string, dataReader io.Reader) (etag string, code s3err.ErrorCode) {
2018-09-12 15:46:12 +08:00
hash := md5.New()
var body = io.TeeReader(dataReader, hash)
2018-09-12 15:46:12 +08:00
proxyReq, err := http.NewRequest("PUT", uploadUrl, body)
if err != nil {
glog.Errorf("NewRequest %s: %v", uploadUrl, err)
2020-09-20 05:09:58 +08:00
return "", s3err.ErrInternalError
}
proxyReq.Header.Set("Host", s3a.option.Filer)
proxyReq.Header.Set("X-Forwarded-For", r.RemoteAddr)
for header, values := range r.Header {
for _, value := range values {
proxyReq.Header.Add(header, value)
}
}
resp, postErr := client.Do(proxyReq)
if postErr != nil {
glog.Errorf("post to filer: %v", postErr)
2020-09-20 05:09:58 +08:00
return "", s3err.ErrInternalError
}
defer resp.Body.Close()
2018-09-12 15:46:12 +08:00
etag = fmt.Sprintf("%x", hash.Sum(nil))
resp_body, ra_err := ioutil.ReadAll(resp.Body)
if ra_err != nil {
2020-11-11 18:01:24 +08:00
glog.Errorf("upload to filer response read %d: %v", resp.StatusCode, ra_err)
2020-09-20 05:09:58 +08:00
return etag, s3err.ErrInternalError
}
2018-09-12 15:46:12 +08:00
var ret weed_server.FilerPostResult
unmarshal_err := json.Unmarshal(resp_body, &ret)
if unmarshal_err != nil {
glog.Errorf("failing to read upload to %s : %v", uploadUrl, string(resp_body))
2020-09-20 05:09:58 +08:00
return "", s3err.ErrInternalError
}
if ret.Error != "" {
glog.Errorf("upload to filer error: %v", ret.Error)
return "", filerErrorToS3Error(ret.Error)
}
2020-09-20 05:09:58 +08:00
return etag, s3err.ErrNone
}
func setEtag(w http.ResponseWriter, etag string) {
if etag != "" {
2018-09-12 15:46:12 +08:00
if strings.HasPrefix(etag, "\"") {
w.Header().Set("ETag", etag)
} else {
w.Header().Set("ETag", "\""+etag+"\"")
}
}
}
2020-07-27 03:58:58 +08:00
func getBucketAndObject(r *http.Request) (bucket, object string) {
vars := mux.Vars(r)
bucket = vars["bucket"]
object = vars["object"]
2018-09-12 15:46:12 +08:00
if !strings.HasPrefix(object, "/") {
object = "/" + object
}
2020-07-27 03:58:58 +08:00
return
2018-09-12 15:46:12 +08:00
}
func filerErrorToS3Error(errString string) s3err.ErrorCode {
if strings.HasPrefix(errString, "existing ") && strings.HasSuffix(errString, "is a directory") {
return s3err.ErrExistingObjectIsDirectory
}
return s3err.ErrInternalError
}