seaweedfs/weed/server/filer_server_handlers_read.go

144 lines
3.4 KiB
Go
Raw Normal View History

2016-06-03 11:05:34 +08:00
package weed_server
import (
2016-07-21 12:20:22 +08:00
"io"
2016-06-03 11:05:34 +08:00
"net/http"
2016-07-21 12:20:22 +08:00
"net/url"
2016-06-03 11:05:34 +08:00
"strconv"
"strings"
"github.com/chrislusf/seaweedfs/weed/glog"
2016-07-21 12:20:22 +08:00
"github.com/chrislusf/seaweedfs/weed/operation"
2016-07-18 16:28:24 +08:00
ui "github.com/chrislusf/seaweedfs/weed/server/filer_ui"
2016-07-21 12:20:22 +08:00
"github.com/chrislusf/seaweedfs/weed/util"
2018-05-14 14:56:16 +08:00
"github.com/chrislusf/seaweedfs/weed/filer2"
2016-06-03 11:05:34 +08:00
)
2016-07-21 12:20:22 +08:00
// listDirectoryHandler lists directories and folers under a directory
// files are sorted by name and paginated via "lastFileName" and "limit".
// sub directories are listed on the first page, when "lastFileName"
// is empty.
func (fs *FilerServer) listDirectoryHandler(w http.ResponseWriter, r *http.Request) {
2018-05-14 14:56:16 +08:00
path := r.URL.Path
if strings.HasSuffix(path, "/") && len(path) > 1 {
path = path[:len(path)-1]
2016-06-03 11:05:34 +08:00
}
2018-05-14 14:56:16 +08:00
2016-06-03 11:05:34 +08:00
limit, limit_err := strconv.Atoi(r.FormValue("limit"))
if limit_err != nil {
limit = 100
}
lastFileName := r.FormValue("lastFileName")
2016-06-03 11:05:34 +08:00
2018-05-14 14:56:16 +08:00
entries, err := fs.filer.ListDirectoryEntries(filer2.FullPath(path), lastFileName, false, limit)
2016-06-03 11:05:34 +08:00
2018-05-14 14:56:16 +08:00
if err != nil {
glog.V(0).Infof("listDirectory %s %s $d: %s", path, lastFileName, limit, err)
w.WriteHeader(http.StatusNotFound)
2016-06-03 11:05:34 +08:00
return
}
2016-07-18 16:28:24 +08:00
2018-05-14 14:56:16 +08:00
shouldDisplayLoadMore := len(entries) == limit
if path == "/" {
path = ""
}
2016-07-18 16:28:24 +08:00
args := struct {
Path string
2018-05-14 14:56:16 +08:00
Entries interface{}
Limit int
LastFileName string
ShouldDisplayLoadMore bool
2016-07-18 16:28:24 +08:00
}{
2018-05-14 14:56:16 +08:00
path,
entries,
limit,
lastFileName,
shouldDisplayLoadMore,
2016-06-03 11:05:34 +08:00
}
if r.Header.Get("Accept") == "application/json" {
writeJsonQuiet(w, r, http.StatusOK, args)
} else {
ui.StatusTpl.Execute(w, args)
}
2016-06-03 11:05:34 +08:00
}
2016-07-21 12:20:22 +08:00
func (fs *FilerServer) GetOrHeadHandler(w http.ResponseWriter, r *http.Request, isGetMethod bool) {
2018-05-14 14:56:16 +08:00
path := r.URL.Path
if strings.HasSuffix(path, "/") && len(path) > 1 {
path = path[:len(path)-1]
}
found, entry, err := fs.filer.FindEntry(filer2.FullPath(path))
if !found || err != nil {
glog.V(3).Infof("Not found %s: %v", path, err)
w.WriteHeader(http.StatusNotFound)
return
}
if entry.IsDirectory() {
2016-07-21 12:20:22 +08:00
if fs.disableDirListing {
w.WriteHeader(http.StatusMethodNotAllowed)
return
}
fs.listDirectoryHandler(w, r)
return
}
2018-05-14 14:56:16 +08:00
if len(entry.Chunks) == 0 {
glog.V(3).Infof("Empty %s: %v", path)
w.WriteHeader(http.StatusNoContent)
2016-07-21 12:20:22 +08:00
return
}
2018-05-14 14:56:16 +08:00
// FIXME pick the right fid
fileId := string(entry.Chunks[0].Fid)
2016-07-21 12:20:22 +08:00
urlLocation, err := operation.LookupFileId(fs.getMasterNode(), fileId)
if err != nil {
glog.V(1).Infoln("operation LookupFileId %s failed, err is %s", fileId, err.Error())
w.WriteHeader(http.StatusNotFound)
return
}
urlString := urlLocation
if fs.redirectOnRead {
http.Redirect(w, r, urlString, http.StatusFound)
return
}
u, _ := url.Parse(urlString)
q := u.Query()
for key, values := range r.URL.Query() {
for _, value := range values {
q.Add(key, value)
}
}
u.RawQuery = q.Encode()
request := &http.Request{
Method: r.Method,
URL: u,
Proto: r.Proto,
ProtoMajor: r.ProtoMajor,
ProtoMinor: r.ProtoMinor,
Header: r.Header,
Body: r.Body,
Host: r.Host,
ContentLength: r.ContentLength,
}
glog.V(3).Infoln("retrieving from", u)
resp, do_err := util.Do(request)
if do_err != nil {
glog.V(0).Infoln("failing to connect to volume server", do_err.Error())
writeJsonError(w, r, http.StatusInternalServerError, do_err)
return
}
defer resp.Body.Close()
for k, v := range resp.Header {
w.Header()[k] = v
}
w.WriteHeader(resp.StatusCode)
io.Copy(w, resp.Body)
}