Chris Lu
10 years ago
4 changed files with 369 additions and 312 deletions
-
7go/weed/weed_server/volume_server.go
-
348go/weed/weed_server/volume_server_handlers.go
-
211go/weed/weed_server/volume_server_handlers_read.go
-
115go/weed/weed_server/volume_server_handlers_write.go
@ -0,0 +1,211 @@ |
|||
package weed_server |
|||
|
|||
import ( |
|||
"io" |
|||
"mime" |
|||
"mime/multipart" |
|||
"net/http" |
|||
"strconv" |
|||
"strings" |
|||
"time" |
|||
|
|||
"github.com/chrislusf/weed-fs/go/glog" |
|||
"github.com/chrislusf/weed-fs/go/images" |
|||
"github.com/chrislusf/weed-fs/go/operation" |
|||
"github.com/chrislusf/weed-fs/go/storage" |
|||
) |
|||
|
|||
var fileNameEscaper = strings.NewReplacer("\\", "\\\\", "\"", "\\\"") |
|||
|
|||
func (vs *VolumeServer) GetOrHeadHandler(w http.ResponseWriter, r *http.Request) { |
|||
n := new(storage.Needle) |
|||
vid, fid, filename, ext, _ := parseURLPath(r.URL.Path) |
|||
volumeId, err := storage.NewVolumeId(vid) |
|||
if err != nil { |
|||
glog.V(2).Infoln("parsing error:", err, r.URL.Path) |
|||
w.WriteHeader(http.StatusBadRequest) |
|||
return |
|||
} |
|||
err = n.ParsePath(fid) |
|||
if err != nil { |
|||
glog.V(2).Infoln("parsing fid error:", err, r.URL.Path) |
|||
w.WriteHeader(http.StatusBadRequest) |
|||
return |
|||
} |
|||
|
|||
glog.V(4).Infoln("volume", volumeId, "reading", n) |
|||
if !vs.store.HasVolume(volumeId) { |
|||
lookupResult, err := operation.Lookup(vs.GetMasterNode(), volumeId.String()) |
|||
glog.V(2).Infoln("volume", volumeId, "found on", lookupResult, "error", err) |
|||
if err == nil && len(lookupResult.Locations) > 0 { |
|||
http.Redirect(w, r, "http://"+lookupResult.Locations[0].Url+r.URL.Path, http.StatusMovedPermanently) |
|||
} else { |
|||
glog.V(2).Infoln("lookup error:", err, r.URL.Path) |
|||
w.WriteHeader(http.StatusNotFound) |
|||
} |
|||
return |
|||
} |
|||
cookie := n.Cookie |
|||
count, e := vs.store.Read(volumeId, n) |
|||
glog.V(4).Infoln("read bytes", count, "error", e) |
|||
if e != nil || count <= 0 { |
|||
glog.V(0).Infoln("read error:", e, r.URL.Path) |
|||
w.WriteHeader(http.StatusNotFound) |
|||
return |
|||
} |
|||
if n.Cookie != cookie { |
|||
glog.V(0).Infoln("request", r.URL.Path, "with unmaching cookie seen:", cookie, "expected:", n.Cookie, "from", r.RemoteAddr, "agent", r.UserAgent()) |
|||
w.WriteHeader(http.StatusNotFound) |
|||
return |
|||
} |
|||
if n.LastModified != 0 { |
|||
w.Header().Set("Last-Modified", time.Unix(int64(n.LastModified), 0).UTC().Format(http.TimeFormat)) |
|||
if r.Header.Get("If-Modified-Since") != "" { |
|||
if t, parseError := time.Parse(http.TimeFormat, r.Header.Get("If-Modified-Since")); parseError == nil { |
|||
if t.Unix() >= int64(n.LastModified) { |
|||
w.WriteHeader(http.StatusNotModified) |
|||
return |
|||
} |
|||
} |
|||
} |
|||
} |
|||
etag := n.Etag() |
|||
if inm := r.Header.Get("If-None-Match"); inm == etag { |
|||
w.WriteHeader(http.StatusNotModified) |
|||
return |
|||
} |
|||
w.Header().Set("Etag", etag) |
|||
if n.NameSize > 0 && filename == "" { |
|||
filename = string(n.Name) |
|||
dotIndex := strings.LastIndex(filename, ".") |
|||
if dotIndex > 0 { |
|||
ext = filename[dotIndex:] |
|||
} |
|||
} |
|||
mtype := "" |
|||
if ext != "" { |
|||
mtype = mime.TypeByExtension(ext) |
|||
} |
|||
if n.MimeSize > 0 { |
|||
mt := string(n.Mime) |
|||
if mt != "application/octet-stream" { |
|||
mtype = mt |
|||
} |
|||
} |
|||
if mtype != "" { |
|||
w.Header().Set("Content-Type", mtype) |
|||
} |
|||
if filename != "" { |
|||
w.Header().Set("Content-Disposition", "filename=\""+fileNameEscaper.Replace(filename)+"\"") |
|||
} |
|||
if ext != ".gz" { |
|||
if n.IsGzipped() { |
|||
if strings.Contains(r.Header.Get("Accept-Encoding"), "gzip") { |
|||
w.Header().Set("Content-Encoding", "gzip") |
|||
} else { |
|||
if n.Data, err = storage.UnGzipData(n.Data); err != nil { |
|||
glog.V(0).Infoln("lookup error:", err, r.URL.Path) |
|||
} |
|||
} |
|||
} |
|||
} |
|||
if ext == ".png" || ext == ".jpg" || ext == ".gif" { |
|||
width, height := 0, 0 |
|||
if r.FormValue("width") != "" { |
|||
width, _ = strconv.Atoi(r.FormValue("width")) |
|||
} |
|||
if r.FormValue("height") != "" { |
|||
height, _ = strconv.Atoi(r.FormValue("height")) |
|||
} |
|||
n.Data, _, _ = images.Resized(ext, n.Data, width, height) |
|||
} |
|||
|
|||
w.Header().Set("Accept-Ranges", "bytes") |
|||
if r.Method == "HEAD" { |
|||
w.Header().Set("Content-Length", strconv.Itoa(len(n.Data))) |
|||
return |
|||
} |
|||
rangeReq := r.Header.Get("Range") |
|||
if rangeReq == "" { |
|||
w.Header().Set("Content-Length", strconv.Itoa(len(n.Data))) |
|||
if _, e = w.Write(n.Data); e != nil { |
|||
glog.V(0).Infoln("response write error:", e) |
|||
} |
|||
return |
|||
} |
|||
|
|||
//the rest is dealing with partial content request
|
|||
//mostly copy from src/pkg/net/http/fs.go
|
|||
size := int64(len(n.Data)) |
|||
ranges, err := parseRange(rangeReq, size) |
|||
if err != nil { |
|||
http.Error(w, err.Error(), http.StatusRequestedRangeNotSatisfiable) |
|||
return |
|||
} |
|||
if sumRangesSize(ranges) > size { |
|||
// The total number of bytes in all the ranges
|
|||
// is larger than the size of the file by
|
|||
// itself, so this is probably an attack, or a
|
|||
// dumb client. Ignore the range request.
|
|||
ranges = nil |
|||
return |
|||
} |
|||
if len(ranges) == 0 { |
|||
return |
|||
} |
|||
if len(ranges) == 1 { |
|||
// RFC 2616, Section 14.16:
|
|||
// "When an HTTP message includes the content of a single
|
|||
// range (for example, a response to a request for a
|
|||
// single range, or to a request for a set of ranges
|
|||
// that overlap without any holes), this content is
|
|||
// transmitted with a Content-Range header, and a
|
|||
// Content-Length header showing the number of bytes
|
|||
// actually transferred.
|
|||
// ...
|
|||
// A response to a request for a single range MUST NOT
|
|||
// be sent using the multipart/byteranges media type."
|
|||
ra := ranges[0] |
|||
w.Header().Set("Content-Length", strconv.FormatInt(ra.length, 10)) |
|||
w.Header().Set("Content-Range", ra.contentRange(size)) |
|||
w.WriteHeader(http.StatusPartialContent) |
|||
if _, e = w.Write(n.Data[ra.start : ra.start+ra.length]); e != nil { |
|||
glog.V(0).Infoln("response write error:", e) |
|||
} |
|||
return |
|||
} |
|||
// process mulitple ranges
|
|||
for _, ra := range ranges { |
|||
if ra.start > size { |
|||
http.Error(w, "Out of Range", http.StatusRequestedRangeNotSatisfiable) |
|||
return |
|||
} |
|||
} |
|||
sendSize := rangesMIMESize(ranges, mtype, size) |
|||
pr, pw := io.Pipe() |
|||
mw := multipart.NewWriter(pw) |
|||
w.Header().Set("Content-Type", "multipart/byteranges; boundary="+mw.Boundary()) |
|||
sendContent := pr |
|||
defer pr.Close() // cause writing goroutine to fail and exit if CopyN doesn't finish.
|
|||
go func() { |
|||
for _, ra := range ranges { |
|||
part, err := mw.CreatePart(ra.mimeHeader(mtype, size)) |
|||
if err != nil { |
|||
pw.CloseWithError(err) |
|||
return |
|||
} |
|||
if _, err = part.Write(n.Data[ra.start : ra.start+ra.length]); err != nil { |
|||
pw.CloseWithError(err) |
|||
return |
|||
} |
|||
} |
|||
mw.Close() |
|||
pw.Close() |
|||
}() |
|||
if w.Header().Get("Content-Encoding") == "" { |
|||
w.Header().Set("Content-Length", strconv.FormatInt(sendSize, 10)) |
|||
} |
|||
w.WriteHeader(http.StatusPartialContent) |
|||
io.CopyN(w, sendContent, sendSize) |
|||
|
|||
} |
@ -0,0 +1,115 @@ |
|||
package weed_server |
|||
|
|||
import ( |
|||
"errors" |
|||
"net/http" |
|||
|
|||
"github.com/chrislusf/weed-fs/go/glog" |
|||
"github.com/chrislusf/weed-fs/go/operation" |
|||
"github.com/chrislusf/weed-fs/go/storage" |
|||
"github.com/chrislusf/weed-fs/go/topology" |
|||
) |
|||
|
|||
func (vs *VolumeServer) PostHandler(w http.ResponseWriter, r *http.Request) { |
|||
if e := r.ParseForm(); e != nil { |
|||
glog.V(0).Infoln("form parse error:", e) |
|||
writeJsonError(w, r, http.StatusBadRequest, e) |
|||
return |
|||
} |
|||
vid, _, _, _, _ := parseURLPath(r.URL.Path) |
|||
volumeId, ve := storage.NewVolumeId(vid) |
|||
if ve != nil { |
|||
glog.V(0).Infoln("NewVolumeId error:", ve) |
|||
writeJsonError(w, r, http.StatusBadRequest, ve) |
|||
return |
|||
} |
|||
needle, ne := storage.NewNeedle(r, vs.FixJpgOrientation) |
|||
if ne != nil { |
|||
writeJsonError(w, r, http.StatusBadRequest, ne) |
|||
return |
|||
} |
|||
|
|||
ret := operation.UploadResult{} |
|||
size, errorStatus := topology.ReplicatedWrite(vs.GetMasterNode(), |
|||
vs.store, volumeId, needle, r) |
|||
httpStatus := http.StatusCreated |
|||
if errorStatus != "" { |
|||
httpStatus = http.StatusInternalServerError |
|||
ret.Error = errorStatus |
|||
} |
|||
if needle.HasName() { |
|||
ret.Name = string(needle.Name) |
|||
} |
|||
ret.Size = size |
|||
writeJsonQuiet(w, r, httpStatus, ret) |
|||
} |
|||
|
|||
func (vs *VolumeServer) DeleteHandler(w http.ResponseWriter, r *http.Request) { |
|||
n := new(storage.Needle) |
|||
vid, fid, _, _, _ := parseURLPath(r.URL.Path) |
|||
volumeId, _ := storage.NewVolumeId(vid) |
|||
n.ParsePath(fid) |
|||
|
|||
glog.V(2).Infoln("deleting", n) |
|||
|
|||
cookie := n.Cookie |
|||
count, ok := vs.store.Read(volumeId, n) |
|||
|
|||
if ok != nil { |
|||
m := make(map[string]uint32) |
|||
m["size"] = 0 |
|||
writeJsonQuiet(w, r, http.StatusNotFound, m) |
|||
return |
|||
} |
|||
|
|||
if n.Cookie != cookie { |
|||
glog.V(0).Infoln("delete", r.URL.Path, "with unmaching cookie from ", r.RemoteAddr, "agent", r.UserAgent()) |
|||
return |
|||
} |
|||
|
|||
n.Size = 0 |
|||
ret := topology.ReplicatedDelete(vs.GetMasterNode(), vs.store, volumeId, n, r) |
|||
|
|||
if ret != 0 { |
|||
m := make(map[string]uint32) |
|||
m["size"] = uint32(count) |
|||
writeJsonQuiet(w, r, http.StatusAccepted, m) |
|||
} else { |
|||
writeJsonError(w, r, http.StatusInternalServerError, errors.New("Deletion Failed.")) |
|||
} |
|||
|
|||
} |
|||
|
|||
//Experts only: takes multiple fid parameters. This function does not propagate deletes to replicas.
|
|||
func (vs *VolumeServer) batchDeleteHandler(w http.ResponseWriter, r *http.Request) { |
|||
r.ParseForm() |
|||
var ret []operation.DeleteResult |
|||
for _, fid := range r.Form["fid"] { |
|||
vid, id_cookie, err := operation.ParseFileId(fid) |
|||
if err != nil { |
|||
ret = append(ret, operation.DeleteResult{Fid: fid, Error: err.Error()}) |
|||
continue |
|||
} |
|||
n := new(storage.Needle) |
|||
volumeId, _ := storage.NewVolumeId(vid) |
|||
n.ParsePath(id_cookie) |
|||
glog.V(4).Infoln("batch deleting", n) |
|||
cookie := n.Cookie |
|||
if _, err := vs.store.Read(volumeId, n); err != nil { |
|||
ret = append(ret, operation.DeleteResult{Fid: fid, Error: err.Error()}) |
|||
continue |
|||
} |
|||
if n.Cookie != cookie { |
|||
ret = append(ret, operation.DeleteResult{Fid: fid, Error: "File Random Cookie does not match."}) |
|||
glog.V(0).Infoln("deleting", fid, "with unmaching cookie from ", r.RemoteAddr, "agent", r.UserAgent()) |
|||
return |
|||
} |
|||
if size, err := vs.store.Delete(volumeId, n); err != nil { |
|||
ret = append(ret, operation.DeleteResult{Fid: fid, Error: err.Error()}) |
|||
} else { |
|||
ret = append(ret, operation.DeleteResult{Fid: fid, Size: int(size)}) |
|||
} |
|||
} |
|||
|
|||
writeJsonQuiet(w, r, http.StatusAccepted, ret) |
|||
} |
Write
Preview
Loading…
Cancel
Save
Reference in new issue