|
|
package chunk_cache
import ( "fmt" "github.com/chrislusf/seaweedfs/weed/glog" "github.com/chrislusf/seaweedfs/weed/storage" "github.com/chrislusf/seaweedfs/weed/storage/types" "golang.org/x/exp/slices" "path" )
type OnDiskCacheLayer struct { diskCaches []*ChunkCacheVolume }
func NewOnDiskCacheLayer(dir, namePrefix string, diskSize int64, segmentCount int) *OnDiskCacheLayer {
volumeCount, volumeSize := int(diskSize/(30000*1024*1024)), int64(30000*1024*1024) if volumeCount < segmentCount { volumeCount, volumeSize = segmentCount, diskSize/int64(segmentCount) }
c := &OnDiskCacheLayer{} for i := 0; i < volumeCount; i++ { fileName := path.Join(dir, fmt.Sprintf("%s_%d", namePrefix, i)) diskCache, err := LoadOrCreateChunkCacheVolume(fileName, volumeSize) if err != nil { glog.Errorf("failed to add cache %s : %v", fileName, err) } else { c.diskCaches = append(c.diskCaches, diskCache) } }
// keep newest cache to the front
slices.SortFunc(c.diskCaches, func(a, b *ChunkCacheVolume) bool { return a.lastModTime.After(b.lastModTime) }) return c }
func (c *OnDiskCacheLayer) setChunk(needleId types.NeedleId, data []byte) {
if len(c.diskCaches) == 0 { return }
if c.diskCaches[0].fileSize+int64(len(data)) > c.diskCaches[0].sizeLimit { t, resetErr := c.diskCaches[len(c.diskCaches)-1].Reset() if resetErr != nil { glog.Errorf("failed to reset cache file %s", c.diskCaches[len(c.diskCaches)-1].fileName) return } for i := len(c.diskCaches) - 1; i > 0; i-- { c.diskCaches[i] = c.diskCaches[i-1] } c.diskCaches[0] = t }
if err := c.diskCaches[0].WriteNeedle(needleId, data); err != nil { glog.V(0).Infof("cache write %v size %d: %v", needleId, len(data), err) }
}
func (c *OnDiskCacheLayer) getChunk(needleId types.NeedleId) (data []byte) {
var err error
for _, diskCache := range c.diskCaches { data, err = diskCache.GetNeedle(needleId) if err == storage.ErrorNotFound { continue } if err != nil { glog.Errorf("failed to read cache file %s id %d", diskCache.fileName, needleId) continue } if len(data) != 0 { return } }
return nil
}
func (c *OnDiskCacheLayer) getChunkSlice(needleId types.NeedleId, offset, length uint64) (data []byte) {
var err error
for _, diskCache := range c.diskCaches { data, err = diskCache.getNeedleSlice(needleId, offset, length) if err == storage.ErrorNotFound { continue } if err != nil { glog.Warningf("failed to read cache file %s id %d: %v", diskCache.fileName, needleId, err) continue } if len(data) != 0 { return } }
return nil
}
func (c *OnDiskCacheLayer) readChunkAt(buffer []byte, needleId types.NeedleId, offset uint64) (n int, err error) {
for _, diskCache := range c.diskCaches { n, err = diskCache.readNeedleSliceAt(buffer, needleId, offset) if err == storage.ErrorNotFound { continue } if err != nil { glog.Warningf("failed to read cache file %s id %d: %v", diskCache.fileName, needleId, err) continue } if n > 0 { return } }
return
}
func (c *OnDiskCacheLayer) shutdown() {
for _, diskCache := range c.diskCaches { diskCache.Shutdown() }
}
|