Browse Source

skip error when draining reader

fix https://github.com/chrislusf/seaweedfs/issues/1179
pull/1189/head
Chris Lu 5 years ago
parent
commit
89e16bd2e8
  1. 7
      weed/util/http_util.go

7
weed/util/http_util.go

@ -11,6 +11,8 @@ import (
"net/http" "net/http"
"net/url" "net/url"
"strings" "strings"
"github.com/chrislusf/seaweedfs/weed/glog"
) )
var ( var (
@ -210,7 +212,8 @@ func ReadUrl(fileUrl string, offset int64, size int, buf []byte, isReadRange boo
} }
var reader io.ReadCloser var reader io.ReadCloser
switch r.Header.Get("Content-Encoding") {
contentEncoding := r.Header.Get("Content-Encoding")
switch contentEncoding {
case "gzip": case "gzip":
reader, err = gzip.NewReader(r.Body) reader, err = gzip.NewReader(r.Body)
defer reader.Close() defer reader.Close()
@ -242,7 +245,7 @@ func ReadUrl(fileUrl string, offset int64, size int, buf []byte, isReadRange boo
// drains the response body to avoid memory leak // drains the response body to avoid memory leak
data, _ := ioutil.ReadAll(reader) data, _ := ioutil.ReadAll(reader)
if len(data) != 0 { if len(data) != 0 {
err = fmt.Errorf("buffer size is too small. remains %d", len(data))
glog.V(1).Infof("%s reader has remaining %d bytes", contentEncoding, len(data))
} }
return n, err return n, err
} }

Loading…
Cancel
Save