You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

636 lines
16 KiB

5 years ago
5 years ago
3 years ago
3 years ago
4 years ago
2 years ago
2 years ago
more solid weed mount (#4089) * compare chunks by timestamp * fix slab clearing error * fix test compilation * move oldest chunk to sealed, instead of by fullness * lock on fh.entryViewCache * remove verbose logs * revert slat clearing * less logs * less logs * track write and read by timestamp * remove useless logic * add entry lock on file handle release * use mem chunk only, swap file chunk has problems * comment out code that maybe used later * add debug mode to compare data read and write * more efficient readResolvedChunks with linked list * small optimization * fix test compilation * minor fix on writer * add SeparateGarbageChunks * group chunks into sections * turn off debug mode * fix tests * fix tests * tmp enable swap file chunk * Revert "tmp enable swap file chunk" This reverts commit 985137ec472924e4815f258189f6ca9f2168a0a7. * simple refactoring * simple refactoring * do not re-use swap file chunk. Sealed chunks should not be re-used. * comment out debugging facilities * either mem chunk or swap file chunk is fine now * remove orderedMutex as *semaphore.Weighted not found impactful * optimize size calculation for changing large files * optimize performance to avoid going through the long list of chunks * still problems with swap file chunk * rename * tiny optimization * swap file chunk save only successfully read data * fix * enable both mem and swap file chunk * resolve chunks with range * rename * fix chunk interval list * also change file handle chunk group when adding chunks * pick in-active chunk with time-decayed counter * fix compilation * avoid nil with empty fh.entry * refactoring * rename * rename * refactor visible intervals to *list.List * refactor chunkViews to *list.List * add IntervalList for generic interval list * change visible interval to use IntervalList in generics * cahnge chunkViews to *IntervalList[*ChunkView] * use NewFileChunkSection to create * rename variables * refactor * fix renaming leftover * renaming * renaming * add insert interval * interval list adds lock * incrementally add chunks to readers Fixes: 1. set start and stop offset for the value object 2. clone the value object 3. use pointer instead of copy-by-value when passing to interval.Value 4. use insert interval since adding chunk could be out of order * fix tests compilation * fix tests compilation
2 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
more solid weed mount (#4089) * compare chunks by timestamp * fix slab clearing error * fix test compilation * move oldest chunk to sealed, instead of by fullness * lock on fh.entryViewCache * remove verbose logs * revert slat clearing * less logs * less logs * track write and read by timestamp * remove useless logic * add entry lock on file handle release * use mem chunk only, swap file chunk has problems * comment out code that maybe used later * add debug mode to compare data read and write * more efficient readResolvedChunks with linked list * small optimization * fix test compilation * minor fix on writer * add SeparateGarbageChunks * group chunks into sections * turn off debug mode * fix tests * fix tests * tmp enable swap file chunk * Revert "tmp enable swap file chunk" This reverts commit 985137ec472924e4815f258189f6ca9f2168a0a7. * simple refactoring * simple refactoring * do not re-use swap file chunk. Sealed chunks should not be re-used. * comment out debugging facilities * either mem chunk or swap file chunk is fine now * remove orderedMutex as *semaphore.Weighted not found impactful * optimize size calculation for changing large files * optimize performance to avoid going through the long list of chunks * still problems with swap file chunk * rename * tiny optimization * swap file chunk save only successfully read data * fix * enable both mem and swap file chunk * resolve chunks with range * rename * fix chunk interval list * also change file handle chunk group when adding chunks * pick in-active chunk with time-decayed counter * fix compilation * avoid nil with empty fh.entry * refactoring * rename * rename * refactor visible intervals to *list.List * refactor chunkViews to *list.List * add IntervalList for generic interval list * change visible interval to use IntervalList in generics * cahnge chunkViews to *IntervalList[*ChunkView] * use NewFileChunkSection to create * rename variables * refactor * fix renaming leftover * renaming * renaming * add insert interval * interval list adds lock * incrementally add chunks to readers Fixes: 1. set start and stop offset for the value object 2. clone the value object 3. use pointer instead of copy-by-value when passing to interval.Value 4. use insert interval since adding chunk could be out of order * fix tests compilation * fix tests compilation
2 years ago
more solid weed mount (#4089) * compare chunks by timestamp * fix slab clearing error * fix test compilation * move oldest chunk to sealed, instead of by fullness * lock on fh.entryViewCache * remove verbose logs * revert slat clearing * less logs * less logs * track write and read by timestamp * remove useless logic * add entry lock on file handle release * use mem chunk only, swap file chunk has problems * comment out code that maybe used later * add debug mode to compare data read and write * more efficient readResolvedChunks with linked list * small optimization * fix test compilation * minor fix on writer * add SeparateGarbageChunks * group chunks into sections * turn off debug mode * fix tests * fix tests * tmp enable swap file chunk * Revert "tmp enable swap file chunk" This reverts commit 985137ec472924e4815f258189f6ca9f2168a0a7. * simple refactoring * simple refactoring * do not re-use swap file chunk. Sealed chunks should not be re-used. * comment out debugging facilities * either mem chunk or swap file chunk is fine now * remove orderedMutex as *semaphore.Weighted not found impactful * optimize size calculation for changing large files * optimize performance to avoid going through the long list of chunks * still problems with swap file chunk * rename * tiny optimization * swap file chunk save only successfully read data * fix * enable both mem and swap file chunk * resolve chunks with range * rename * fix chunk interval list * also change file handle chunk group when adding chunks * pick in-active chunk with time-decayed counter * fix compilation * avoid nil with empty fh.entry * refactoring * rename * rename * refactor visible intervals to *list.List * refactor chunkViews to *list.List * add IntervalList for generic interval list * change visible interval to use IntervalList in generics * cahnge chunkViews to *IntervalList[*ChunkView] * use NewFileChunkSection to create * rename variables * refactor * fix renaming leftover * renaming * renaming * add insert interval * interval list adds lock * incrementally add chunks to readers Fixes: 1. set start and stop offset for the value object 2. clone the value object 3. use pointer instead of copy-by-value when passing to interval.Value 4. use insert interval since adding chunk could be out of order * fix tests compilation * fix tests compilation
2 years ago
more solid weed mount (#4089) * compare chunks by timestamp * fix slab clearing error * fix test compilation * move oldest chunk to sealed, instead of by fullness * lock on fh.entryViewCache * remove verbose logs * revert slat clearing * less logs * less logs * track write and read by timestamp * remove useless logic * add entry lock on file handle release * use mem chunk only, swap file chunk has problems * comment out code that maybe used later * add debug mode to compare data read and write * more efficient readResolvedChunks with linked list * small optimization * fix test compilation * minor fix on writer * add SeparateGarbageChunks * group chunks into sections * turn off debug mode * fix tests * fix tests * tmp enable swap file chunk * Revert "tmp enable swap file chunk" This reverts commit 985137ec472924e4815f258189f6ca9f2168a0a7. * simple refactoring * simple refactoring * do not re-use swap file chunk. Sealed chunks should not be re-used. * comment out debugging facilities * either mem chunk or swap file chunk is fine now * remove orderedMutex as *semaphore.Weighted not found impactful * optimize size calculation for changing large files * optimize performance to avoid going through the long list of chunks * still problems with swap file chunk * rename * tiny optimization * swap file chunk save only successfully read data * fix * enable both mem and swap file chunk * resolve chunks with range * rename * fix chunk interval list * also change file handle chunk group when adding chunks * pick in-active chunk with time-decayed counter * fix compilation * avoid nil with empty fh.entry * refactoring * rename * rename * refactor visible intervals to *list.List * refactor chunkViews to *list.List * add IntervalList for generic interval list * change visible interval to use IntervalList in generics * cahnge chunkViews to *IntervalList[*ChunkView] * use NewFileChunkSection to create * rename variables * refactor * fix renaming leftover * renaming * renaming * add insert interval * interval list adds lock * incrementally add chunks to readers Fixes: 1. set start and stop offset for the value object 2. clone the value object 3. use pointer instead of copy-by-value when passing to interval.Value 4. use insert interval since adding chunk could be out of order * fix tests compilation * fix tests compilation
2 years ago
4 years ago
5 years ago
5 years ago
more solid weed mount (#4089) * compare chunks by timestamp * fix slab clearing error * fix test compilation * move oldest chunk to sealed, instead of by fullness * lock on fh.entryViewCache * remove verbose logs * revert slat clearing * less logs * less logs * track write and read by timestamp * remove useless logic * add entry lock on file handle release * use mem chunk only, swap file chunk has problems * comment out code that maybe used later * add debug mode to compare data read and write * more efficient readResolvedChunks with linked list * small optimization * fix test compilation * minor fix on writer * add SeparateGarbageChunks * group chunks into sections * turn off debug mode * fix tests * fix tests * tmp enable swap file chunk * Revert "tmp enable swap file chunk" This reverts commit 985137ec472924e4815f258189f6ca9f2168a0a7. * simple refactoring * simple refactoring * do not re-use swap file chunk. Sealed chunks should not be re-used. * comment out debugging facilities * either mem chunk or swap file chunk is fine now * remove orderedMutex as *semaphore.Weighted not found impactful * optimize size calculation for changing large files * optimize performance to avoid going through the long list of chunks * still problems with swap file chunk * rename * tiny optimization * swap file chunk save only successfully read data * fix * enable both mem and swap file chunk * resolve chunks with range * rename * fix chunk interval list * also change file handle chunk group when adding chunks * pick in-active chunk with time-decayed counter * fix compilation * avoid nil with empty fh.entry * refactoring * rename * rename * refactor visible intervals to *list.List * refactor chunkViews to *list.List * add IntervalList for generic interval list * change visible interval to use IntervalList in generics * cahnge chunkViews to *IntervalList[*ChunkView] * use NewFileChunkSection to create * rename variables * refactor * fix renaming leftover * renaming * renaming * add insert interval * interval list adds lock * incrementally add chunks to readers Fixes: 1. set start and stop offset for the value object 2. clone the value object 3. use pointer instead of copy-by-value when passing to interval.Value 4. use insert interval since adding chunk could be out of order * fix tests compilation * fix tests compilation
2 years ago
5 years ago
4 years ago
more solid weed mount (#4089) * compare chunks by timestamp * fix slab clearing error * fix test compilation * move oldest chunk to sealed, instead of by fullness * lock on fh.entryViewCache * remove verbose logs * revert slat clearing * less logs * less logs * track write and read by timestamp * remove useless logic * add entry lock on file handle release * use mem chunk only, swap file chunk has problems * comment out code that maybe used later * add debug mode to compare data read and write * more efficient readResolvedChunks with linked list * small optimization * fix test compilation * minor fix on writer * add SeparateGarbageChunks * group chunks into sections * turn off debug mode * fix tests * fix tests * tmp enable swap file chunk * Revert "tmp enable swap file chunk" This reverts commit 985137ec472924e4815f258189f6ca9f2168a0a7. * simple refactoring * simple refactoring * do not re-use swap file chunk. Sealed chunks should not be re-used. * comment out debugging facilities * either mem chunk or swap file chunk is fine now * remove orderedMutex as *semaphore.Weighted not found impactful * optimize size calculation for changing large files * optimize performance to avoid going through the long list of chunks * still problems with swap file chunk * rename * tiny optimization * swap file chunk save only successfully read data * fix * enable both mem and swap file chunk * resolve chunks with range * rename * fix chunk interval list * also change file handle chunk group when adding chunks * pick in-active chunk with time-decayed counter * fix compilation * avoid nil with empty fh.entry * refactoring * rename * rename * refactor visible intervals to *list.List * refactor chunkViews to *list.List * add IntervalList for generic interval list * change visible interval to use IntervalList in generics * cahnge chunkViews to *IntervalList[*ChunkView] * use NewFileChunkSection to create * rename variables * refactor * fix renaming leftover * renaming * renaming * add insert interval * interval list adds lock * incrementally add chunks to readers Fixes: 1. set start and stop offset for the value object 2. clone the value object 3. use pointer instead of copy-by-value when passing to interval.Value 4. use insert interval since adding chunk could be out of order * fix tests compilation * fix tests compilation
2 years ago
more solid weed mount (#4089) * compare chunks by timestamp * fix slab clearing error * fix test compilation * move oldest chunk to sealed, instead of by fullness * lock on fh.entryViewCache * remove verbose logs * revert slat clearing * less logs * less logs * track write and read by timestamp * remove useless logic * add entry lock on file handle release * use mem chunk only, swap file chunk has problems * comment out code that maybe used later * add debug mode to compare data read and write * more efficient readResolvedChunks with linked list * small optimization * fix test compilation * minor fix on writer * add SeparateGarbageChunks * group chunks into sections * turn off debug mode * fix tests * fix tests * tmp enable swap file chunk * Revert "tmp enable swap file chunk" This reverts commit 985137ec472924e4815f258189f6ca9f2168a0a7. * simple refactoring * simple refactoring * do not re-use swap file chunk. Sealed chunks should not be re-used. * comment out debugging facilities * either mem chunk or swap file chunk is fine now * remove orderedMutex as *semaphore.Weighted not found impactful * optimize size calculation for changing large files * optimize performance to avoid going through the long list of chunks * still problems with swap file chunk * rename * tiny optimization * swap file chunk save only successfully read data * fix * enable both mem and swap file chunk * resolve chunks with range * rename * fix chunk interval list * also change file handle chunk group when adding chunks * pick in-active chunk with time-decayed counter * fix compilation * avoid nil with empty fh.entry * refactoring * rename * rename * refactor visible intervals to *list.List * refactor chunkViews to *list.List * add IntervalList for generic interval list * change visible interval to use IntervalList in generics * cahnge chunkViews to *IntervalList[*ChunkView] * use NewFileChunkSection to create * rename variables * refactor * fix renaming leftover * renaming * renaming * add insert interval * interval list adds lock * incrementally add chunks to readers Fixes: 1. set start and stop offset for the value object 2. clone the value object 3. use pointer instead of copy-by-value when passing to interval.Value 4. use insert interval since adding chunk could be out of order * fix tests compilation * fix tests compilation
2 years ago
5 years ago
5 years ago
5 years ago
2 years ago
2 years ago
4 years ago
4 years ago
  1. package weed_server
  2. import (
  3. "context"
  4. "fmt"
  5. "io"
  6. "os"
  7. "path"
  8. "strings"
  9. "time"
  10. "github.com/seaweedfs/seaweedfs/weed/util/buffered_writer"
  11. "golang.org/x/net/webdav"
  12. "google.golang.org/grpc"
  13. "github.com/seaweedfs/seaweedfs/weed/operation"
  14. "github.com/seaweedfs/seaweedfs/weed/pb"
  15. "github.com/seaweedfs/seaweedfs/weed/pb/filer_pb"
  16. "github.com/seaweedfs/seaweedfs/weed/util"
  17. "github.com/seaweedfs/seaweedfs/weed/util/chunk_cache"
  18. "github.com/seaweedfs/seaweedfs/weed/filer"
  19. "github.com/seaweedfs/seaweedfs/weed/glog"
  20. "github.com/seaweedfs/seaweedfs/weed/security"
  21. )
  22. type WebDavOption struct {
  23. Filer pb.ServerAddress
  24. FilerRootPath string
  25. DomainName string
  26. BucketsPath string
  27. GrpcDialOption grpc.DialOption
  28. Collection string
  29. Replication string
  30. DiskType string
  31. Uid uint32
  32. Gid uint32
  33. Cipher bool
  34. CacheDir string
  35. CacheSizeMB int64
  36. MaxMB int
  37. }
  38. type WebDavServer struct {
  39. option *WebDavOption
  40. secret security.SigningKey
  41. filer *filer.Filer
  42. grpcDialOption grpc.DialOption
  43. Handler *webdav.Handler
  44. }
  45. func max(x, y int64) int64 {
  46. if x <= y {
  47. return y
  48. }
  49. return x
  50. }
  51. func NewWebDavServer(option *WebDavOption) (ws *WebDavServer, err error) {
  52. fs, _ := NewWebDavFileSystem(option)
  53. // Fix no set filer.path , accessing "/" returns "//"
  54. if option.FilerRootPath == "/" {
  55. option.FilerRootPath = ""
  56. }
  57. // filer.path non "/" option means we are accessing filer's sub-folders
  58. if option.FilerRootPath != "" {
  59. fs = NewWrappedFs(fs, path.Clean(option.FilerRootPath))
  60. }
  61. ws = &WebDavServer{
  62. option: option,
  63. grpcDialOption: security.LoadClientTLS(util.GetViper(), "grpc.filer"),
  64. Handler: &webdav.Handler{
  65. FileSystem: fs,
  66. LockSystem: webdav.NewMemLS(),
  67. },
  68. }
  69. return ws, nil
  70. }
  71. // adapted from https://github.com/mattn/davfs/blob/master/plugin/mysql/mysql.go
  72. type WebDavFileSystem struct {
  73. option *WebDavOption
  74. secret security.SigningKey
  75. grpcDialOption grpc.DialOption
  76. chunkCache *chunk_cache.TieredChunkCache
  77. readerCache *filer.ReaderCache
  78. signature int32
  79. }
  80. type FileInfo struct {
  81. name string
  82. size int64
  83. mode os.FileMode
  84. modifiedTime time.Time
  85. etag string
  86. isDirectory bool
  87. }
  88. func (fi *FileInfo) Name() string { return fi.name }
  89. func (fi *FileInfo) Size() int64 { return fi.size }
  90. func (fi *FileInfo) Mode() os.FileMode { return fi.mode }
  91. func (fi *FileInfo) ModTime() time.Time { return fi.modifiedTime }
  92. func (fi *FileInfo) IsDir() bool { return fi.isDirectory }
  93. func (fi *FileInfo) Sys() interface{} { return nil }
  94. func (fi *FileInfo) ETag(ctx context.Context) (string, error) {
  95. return fi.etag, nil
  96. }
  97. type WebDavFile struct {
  98. fs *WebDavFileSystem
  99. name string
  100. isDirectory bool
  101. off int64
  102. entry *filer_pb.Entry
  103. visibleIntervals *filer.IntervalList[*filer.VisibleInterval]
  104. reader io.ReaderAt
  105. bufWriter *buffered_writer.BufferedWriteCloser
  106. }
  107. func NewWebDavFileSystem(option *WebDavOption) (webdav.FileSystem, error) {
  108. cacheUniqueId := util.Md5String([]byte("webdav" + string(option.Filer) + util.Version()))[0:8]
  109. cacheDir := path.Join(option.CacheDir, cacheUniqueId)
  110. os.MkdirAll(cacheDir, os.FileMode(0755))
  111. chunkCache := chunk_cache.NewTieredChunkCache(256, cacheDir, option.CacheSizeMB, 1024*1024)
  112. t := &WebDavFileSystem{
  113. option: option,
  114. chunkCache: chunkCache,
  115. signature: util.RandomInt32(),
  116. }
  117. t.readerCache = filer.NewReaderCache(32, chunkCache, filer.LookupFn(t))
  118. return t, nil
  119. }
  120. var _ = filer_pb.FilerClient(&WebDavFileSystem{})
  121. func (fs *WebDavFileSystem) WithFilerClient(streamingMode bool, fn func(filer_pb.SeaweedFilerClient) error) error {
  122. return pb.WithGrpcClient(streamingMode, fs.signature, func(grpcConnection *grpc.ClientConn) error {
  123. client := filer_pb.NewSeaweedFilerClient(grpcConnection)
  124. return fn(client)
  125. }, fs.option.Filer.ToGrpcAddress(), false, fs.option.GrpcDialOption)
  126. }
  127. func (fs *WebDavFileSystem) AdjustedUrl(location *filer_pb.Location) string {
  128. return location.Url
  129. }
  130. func (fs *WebDavFileSystem) GetDataCenter() string {
  131. return ""
  132. }
  133. func clearName(name string) (string, error) {
  134. slashed := strings.HasSuffix(name, "/")
  135. name = path.Clean(name)
  136. if !strings.HasSuffix(name, "/") && slashed {
  137. name += "/"
  138. }
  139. if !strings.HasPrefix(name, "/") {
  140. return "", os.ErrInvalid
  141. }
  142. return name, nil
  143. }
  144. func (fs *WebDavFileSystem) Mkdir(ctx context.Context, fullDirPath string, perm os.FileMode) error {
  145. glog.V(2).Infof("WebDavFileSystem.Mkdir %v", fullDirPath)
  146. if !strings.HasSuffix(fullDirPath, "/") {
  147. fullDirPath += "/"
  148. }
  149. var err error
  150. if fullDirPath, err = clearName(fullDirPath); err != nil {
  151. return err
  152. }
  153. _, err = fs.stat(ctx, fullDirPath)
  154. if err == nil {
  155. return os.ErrExist
  156. }
  157. return fs.WithFilerClient(false, func(client filer_pb.SeaweedFilerClient) error {
  158. dir, name := util.FullPath(fullDirPath).DirAndName()
  159. request := &filer_pb.CreateEntryRequest{
  160. Directory: dir,
  161. Entry: &filer_pb.Entry{
  162. Name: name,
  163. IsDirectory: true,
  164. Attributes: &filer_pb.FuseAttributes{
  165. Mtime: time.Now().Unix(),
  166. Crtime: time.Now().Unix(),
  167. FileMode: uint32(perm | os.ModeDir),
  168. Uid: fs.option.Uid,
  169. Gid: fs.option.Gid,
  170. },
  171. },
  172. Signatures: []int32{fs.signature},
  173. }
  174. glog.V(1).Infof("mkdir: %v", request)
  175. if err := filer_pb.CreateEntry(client, request); err != nil {
  176. return fmt.Errorf("mkdir %s/%s: %v", dir, name, err)
  177. }
  178. return nil
  179. })
  180. }
  181. func (fs *WebDavFileSystem) OpenFile(ctx context.Context, fullFilePath string, flag int, perm os.FileMode) (webdav.File, error) {
  182. glog.V(2).Infof("WebDavFileSystem.OpenFile %v %x", fullFilePath, flag)
  183. var err error
  184. if fullFilePath, err = clearName(fullFilePath); err != nil {
  185. return nil, err
  186. }
  187. if flag&os.O_CREATE != 0 {
  188. // file should not have / suffix.
  189. if strings.HasSuffix(fullFilePath, "/") {
  190. return nil, os.ErrInvalid
  191. }
  192. _, err = fs.stat(ctx, fullFilePath)
  193. if err == nil {
  194. if flag&os.O_EXCL != 0 {
  195. return nil, os.ErrExist
  196. }
  197. fs.removeAll(ctx, fullFilePath)
  198. }
  199. dir, name := util.FullPath(fullFilePath).DirAndName()
  200. err = fs.WithFilerClient(false, func(client filer_pb.SeaweedFilerClient) error {
  201. if err := filer_pb.CreateEntry(client, &filer_pb.CreateEntryRequest{
  202. Directory: dir,
  203. Entry: &filer_pb.Entry{
  204. Name: name,
  205. IsDirectory: perm&os.ModeDir > 0,
  206. Attributes: &filer_pb.FuseAttributes{
  207. Mtime: 0,
  208. Crtime: time.Now().Unix(),
  209. FileMode: uint32(perm),
  210. Uid: fs.option.Uid,
  211. Gid: fs.option.Gid,
  212. TtlSec: 0,
  213. },
  214. },
  215. Signatures: []int32{fs.signature},
  216. }); err != nil {
  217. return fmt.Errorf("create %s: %v", fullFilePath, err)
  218. }
  219. return nil
  220. })
  221. if err != nil {
  222. return nil, err
  223. }
  224. return &WebDavFile{
  225. fs: fs,
  226. name: fullFilePath,
  227. isDirectory: false,
  228. bufWriter: buffered_writer.NewBufferedWriteCloser(fs.option.MaxMB * 1024 * 1024),
  229. }, nil
  230. }
  231. fi, err := fs.stat(ctx, fullFilePath)
  232. if err != nil {
  233. return nil, os.ErrNotExist
  234. }
  235. if !strings.HasSuffix(fullFilePath, "/") && fi.IsDir() {
  236. fullFilePath += "/"
  237. }
  238. return &WebDavFile{
  239. fs: fs,
  240. name: fullFilePath,
  241. isDirectory: false,
  242. bufWriter: buffered_writer.NewBufferedWriteCloser(fs.option.MaxMB * 1024 * 1024),
  243. }, nil
  244. }
  245. func (fs *WebDavFileSystem) removeAll(ctx context.Context, fullFilePath string) error {
  246. var err error
  247. if fullFilePath, err = clearName(fullFilePath); err != nil {
  248. return err
  249. }
  250. dir, name := util.FullPath(fullFilePath).DirAndName()
  251. return filer_pb.Remove(fs, dir, name, true, false, false, false, []int32{fs.signature})
  252. }
  253. func (fs *WebDavFileSystem) RemoveAll(ctx context.Context, name string) error {
  254. glog.V(2).Infof("WebDavFileSystem.RemoveAll %v", name)
  255. return fs.removeAll(ctx, name)
  256. }
  257. func (fs *WebDavFileSystem) Rename(ctx context.Context, oldName, newName string) error {
  258. glog.V(2).Infof("WebDavFileSystem.Rename %v to %v", oldName, newName)
  259. var err error
  260. if oldName, err = clearName(oldName); err != nil {
  261. return err
  262. }
  263. if newName, err = clearName(newName); err != nil {
  264. return err
  265. }
  266. of, err := fs.stat(ctx, oldName)
  267. if err != nil {
  268. return os.ErrExist
  269. }
  270. if of.IsDir() {
  271. if strings.HasSuffix(oldName, "/") {
  272. oldName = strings.TrimRight(oldName, "/")
  273. }
  274. if strings.HasSuffix(newName, "/") {
  275. newName = strings.TrimRight(newName, "/")
  276. }
  277. }
  278. _, err = fs.stat(ctx, newName)
  279. if err == nil {
  280. return os.ErrExist
  281. }
  282. oldDir, oldBaseName := util.FullPath(oldName).DirAndName()
  283. newDir, newBaseName := util.FullPath(newName).DirAndName()
  284. return fs.WithFilerClient(false, func(client filer_pb.SeaweedFilerClient) error {
  285. request := &filer_pb.AtomicRenameEntryRequest{
  286. OldDirectory: oldDir,
  287. OldName: oldBaseName,
  288. NewDirectory: newDir,
  289. NewName: newBaseName,
  290. }
  291. _, err := client.AtomicRenameEntry(ctx, request)
  292. if err != nil {
  293. return fmt.Errorf("renaming %s/%s => %s/%s: %v", oldDir, oldBaseName, newDir, newBaseName, err)
  294. }
  295. return nil
  296. })
  297. }
  298. func (fs *WebDavFileSystem) stat(ctx context.Context, fullFilePath string) (os.FileInfo, error) {
  299. var err error
  300. if fullFilePath, err = clearName(fullFilePath); err != nil {
  301. return nil, err
  302. }
  303. fullpath := util.FullPath(fullFilePath)
  304. var fi FileInfo
  305. entry, err := filer_pb.GetEntry(fs, fullpath)
  306. if entry == nil {
  307. return nil, os.ErrNotExist
  308. }
  309. if err != nil {
  310. return nil, err
  311. }
  312. fi.size = int64(filer.FileSize(entry))
  313. fi.name = string(fullpath)
  314. fi.mode = os.FileMode(entry.Attributes.FileMode)
  315. fi.modifiedTime = time.Unix(entry.Attributes.Mtime, 0)
  316. fi.etag = filer.ETag(entry)
  317. fi.isDirectory = entry.IsDirectory
  318. if fi.name == "/" {
  319. fi.modifiedTime = time.Now()
  320. fi.isDirectory = true
  321. }
  322. return &fi, nil
  323. }
  324. func (fs *WebDavFileSystem) Stat(ctx context.Context, name string) (os.FileInfo, error) {
  325. glog.V(2).Infof("WebDavFileSystem.Stat %v", name)
  326. return fs.stat(ctx, name)
  327. }
  328. func (f *WebDavFile) saveDataAsChunk(reader io.Reader, name string, offset int64, tsNs int64) (chunk *filer_pb.FileChunk, err error) {
  329. fileId, uploadResult, flushErr, _ := operation.UploadWithRetry(
  330. f.fs,
  331. &filer_pb.AssignVolumeRequest{
  332. Count: 1,
  333. Replication: f.fs.option.Replication,
  334. Collection: f.fs.option.Collection,
  335. DiskType: f.fs.option.DiskType,
  336. Path: name,
  337. },
  338. &operation.UploadOption{
  339. Filename: f.name,
  340. Cipher: f.fs.option.Cipher,
  341. IsInputCompressed: false,
  342. MimeType: "",
  343. PairMap: nil,
  344. },
  345. func(host, fileId string) string {
  346. return fmt.Sprintf("http://%s/%s", host, fileId)
  347. },
  348. reader,
  349. )
  350. if flushErr != nil {
  351. glog.V(0).Infof("upload data %v: %v", f.name, flushErr)
  352. return nil, fmt.Errorf("upload data: %v", flushErr)
  353. }
  354. if uploadResult.Error != "" {
  355. glog.V(0).Infof("upload failure %v: %v", f.name, flushErr)
  356. return nil, fmt.Errorf("upload result: %v", uploadResult.Error)
  357. }
  358. return uploadResult.ToPbFileChunk(fileId, offset, tsNs), nil
  359. }
  360. func (f *WebDavFile) Write(buf []byte) (int, error) {
  361. glog.V(2).Infof("WebDavFileSystem.Write %v", f.name)
  362. fullPath := util.FullPath(f.name)
  363. dir, _ := fullPath.DirAndName()
  364. var getErr error
  365. ctx := context.Background()
  366. if f.entry == nil {
  367. f.entry, getErr = filer_pb.GetEntry(f.fs, fullPath)
  368. }
  369. if f.entry == nil {
  370. return 0, getErr
  371. }
  372. if getErr != nil {
  373. return 0, getErr
  374. }
  375. if f.bufWriter.FlushFunc == nil {
  376. f.bufWriter.FlushFunc = func(data []byte, offset int64) (flushErr error) {
  377. var chunk *filer_pb.FileChunk
  378. chunk, flushErr = f.saveDataAsChunk(util.NewBytesReader(data), f.name, offset, time.Now().UnixNano())
  379. if flushErr != nil {
  380. if f.entry.Attributes.Mtime == 0 {
  381. if err := f.fs.removeAll(ctx, f.name); err != nil {
  382. glog.Errorf("bufWriter.Flush remove file error: %+v", f.name)
  383. }
  384. }
  385. return fmt.Errorf("%s upload result: %v", f.name, flushErr)
  386. }
  387. f.entry.Content = nil
  388. f.entry.Chunks = append(f.entry.GetChunks(), chunk)
  389. return flushErr
  390. }
  391. f.bufWriter.CloseFunc = func() error {
  392. manifestedChunks, manifestErr := filer.MaybeManifestize(f.saveDataAsChunk, f.entry.GetChunks())
  393. if manifestErr != nil {
  394. // not good, but should be ok
  395. glog.V(0).Infof("file %s close MaybeManifestize: %v", f.name, manifestErr)
  396. } else {
  397. f.entry.Chunks = manifestedChunks
  398. }
  399. flushErr := f.fs.WithFilerClient(false, func(client filer_pb.SeaweedFilerClient) error {
  400. f.entry.Attributes.Mtime = time.Now().Unix()
  401. request := &filer_pb.UpdateEntryRequest{
  402. Directory: dir,
  403. Entry: f.entry,
  404. Signatures: []int32{f.fs.signature},
  405. }
  406. if _, err := client.UpdateEntry(ctx, request); err != nil {
  407. return fmt.Errorf("update %s: %v", f.name, err)
  408. }
  409. return nil
  410. })
  411. return flushErr
  412. }
  413. }
  414. written, err := f.bufWriter.Write(buf)
  415. if err == nil {
  416. f.entry.Attributes.FileSize = uint64(max(f.off+int64(written), int64(f.entry.Attributes.FileSize)))
  417. glog.V(3).Infof("WebDavFileSystem.Write %v: written [%d,%d)", f.name, f.off, f.off+int64(len(buf)))
  418. f.off += int64(written)
  419. }
  420. return written, err
  421. }
  422. func (f *WebDavFile) Close() error {
  423. glog.V(2).Infof("WebDavFileSystem.Close %v", f.name)
  424. err := f.bufWriter.Close()
  425. if f.entry != nil {
  426. f.entry = nil
  427. f.visibleIntervals = nil
  428. }
  429. return err
  430. }
  431. func (f *WebDavFile) Read(p []byte) (readSize int, err error) {
  432. glog.V(2).Infof("WebDavFileSystem.Read %v", f.name)
  433. if f.entry == nil {
  434. f.entry, err = filer_pb.GetEntry(f.fs, util.FullPath(f.name))
  435. }
  436. if f.entry == nil {
  437. return 0, err
  438. }
  439. if err != nil {
  440. return 0, err
  441. }
  442. fileSize := int64(filer.FileSize(f.entry))
  443. if fileSize == 0 {
  444. return 0, io.EOF
  445. }
  446. if f.visibleIntervals == nil {
  447. f.visibleIntervals, _ = filer.NonOverlappingVisibleIntervals(filer.LookupFn(f.fs), f.entry.GetChunks(), 0, fileSize)
  448. f.reader = nil
  449. }
  450. if f.reader == nil {
  451. chunkViews := filer.ViewFromVisibleIntervals(f.visibleIntervals, 0, fileSize)
  452. f.reader = filer.NewChunkReaderAtFromClient(f.fs.readerCache, chunkViews, fileSize)
  453. }
  454. readSize, err = f.reader.ReadAt(p, f.off)
  455. glog.V(3).Infof("WebDavFileSystem.Read %v: [%d,%d)", f.name, f.off, f.off+int64(readSize))
  456. f.off += int64(readSize)
  457. if err != nil && err != io.EOF {
  458. glog.Errorf("file read %s: %v", f.name, err)
  459. }
  460. return
  461. }
  462. func (f *WebDavFile) Readdir(count int) (ret []os.FileInfo, err error) {
  463. glog.V(2).Infof("WebDavFileSystem.Readdir %v count %d", f.name, count)
  464. dir, _ := util.FullPath(f.name).DirAndName()
  465. err = filer_pb.ReadDirAllEntries(f.fs, util.FullPath(dir), "", func(entry *filer_pb.Entry, isLast bool) error {
  466. fi := FileInfo{
  467. size: int64(filer.FileSize(entry)),
  468. name: entry.Name,
  469. mode: os.FileMode(entry.Attributes.FileMode),
  470. modifiedTime: time.Unix(entry.Attributes.Mtime, 0),
  471. isDirectory: entry.IsDirectory,
  472. }
  473. if !strings.HasSuffix(fi.name, "/") && fi.IsDir() {
  474. fi.name += "/"
  475. }
  476. glog.V(4).Infof("entry: %v", fi.name)
  477. ret = append(ret, &fi)
  478. return nil
  479. })
  480. if err != nil {
  481. return nil, err
  482. }
  483. old := f.off
  484. if old >= int64(len(ret)) {
  485. if count > 0 {
  486. return nil, io.EOF
  487. }
  488. return nil, nil
  489. }
  490. if count > 0 {
  491. f.off += int64(count)
  492. if f.off > int64(len(ret)) {
  493. f.off = int64(len(ret))
  494. }
  495. } else {
  496. f.off = int64(len(ret))
  497. old = 0
  498. }
  499. return ret[old:f.off], nil
  500. }
  501. func (f *WebDavFile) Seek(offset int64, whence int) (int64, error) {
  502. glog.V(2).Infof("WebDavFile.Seek %v %v %v", f.name, offset, whence)
  503. ctx := context.Background()
  504. var err error
  505. switch whence {
  506. case io.SeekStart:
  507. f.off = 0
  508. case io.SeekEnd:
  509. if fi, err := f.fs.stat(ctx, f.name); err != nil {
  510. return 0, err
  511. } else {
  512. f.off = fi.Size()
  513. }
  514. }
  515. f.off += offset
  516. return f.off, err
  517. }
  518. func (f *WebDavFile) Stat() (os.FileInfo, error) {
  519. glog.V(2).Infof("WebDavFile.Stat %v", f.name)
  520. ctx := context.Background()
  521. return f.fs.stat(ctx, f.name)
  522. }