You can not select more than 25 topics
			Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
		
		
		
		
		
			
		
			
				
					
					
						
							171 lines
						
					
					
						
							5.4 KiB
						
					
					
				
			
		
		
		
			
			
			
		
		
	
	
							171 lines
						
					
					
						
							5.4 KiB
						
					
					
				
								package mount
							 | 
						|
								
							 | 
						|
								import (
							 | 
						|
									"context"
							 | 
						|
									"fmt"
							 | 
						|
									"io"
							 | 
						|
									"sort"
							 | 
						|
								
							 | 
						|
									"github.com/seaweedfs/seaweedfs/weed/filer"
							 | 
						|
									"github.com/seaweedfs/seaweedfs/weed/glog"
							 | 
						|
									"github.com/seaweedfs/seaweedfs/weed/pb/filer_pb"
							 | 
						|
								)
							 | 
						|
								
							 | 
						|
								func (fh *FileHandle) lockForRead(startOffset int64, size int) {
							 | 
						|
									fh.dirtyPages.LockForRead(startOffset, startOffset+int64(size))
							 | 
						|
								}
							 | 
						|
								func (fh *FileHandle) unlockForRead(startOffset int64, size int) {
							 | 
						|
									fh.dirtyPages.UnlockForRead(startOffset, startOffset+int64(size))
							 | 
						|
								}
							 | 
						|
								
							 | 
						|
								func (fh *FileHandle) readFromDirtyPages(buff []byte, startOffset int64, tsNs int64) (maxStop int64) {
							 | 
						|
									maxStop = fh.dirtyPages.ReadDirtyDataAt(buff, startOffset, tsNs)
							 | 
						|
									return
							 | 
						|
								}
							 | 
						|
								
							 | 
						|
								func (fh *FileHandle) readFromChunks(buff []byte, offset int64) (int64, int64, error) {
							 | 
						|
									return fh.readFromChunksWithContext(context.Background(), buff, offset)
							 | 
						|
								}
							 | 
						|
								
							 | 
						|
								func (fh *FileHandle) readFromChunksWithContext(ctx context.Context, buff []byte, offset int64) (int64, int64, error) {
							 | 
						|
									fh.entryLock.RLock()
							 | 
						|
									defer fh.entryLock.RUnlock()
							 | 
						|
								
							 | 
						|
									fileFullPath := fh.FullPath()
							 | 
						|
								
							 | 
						|
									entry := fh.GetEntry()
							 | 
						|
								
							 | 
						|
									if entry.IsInRemoteOnly() {
							 | 
						|
										glog.V(4).Infof("download remote entry %s", fileFullPath)
							 | 
						|
										err := fh.downloadRemoteEntry(entry)
							 | 
						|
										if err != nil {
							 | 
						|
											glog.V(1).Infof("download remote entry %s: %v", fileFullPath, err)
							 | 
						|
											return 0, 0, err
							 | 
						|
										}
							 | 
						|
									}
							 | 
						|
								
							 | 
						|
									fileSize := int64(entry.Attributes.FileSize)
							 | 
						|
									if fileSize == 0 {
							 | 
						|
										fileSize = int64(filer.FileSize(entry.GetEntry()))
							 | 
						|
									}
							 | 
						|
								
							 | 
						|
									if fileSize == 0 {
							 | 
						|
										glog.V(1).Infof("empty fh %v", fileFullPath)
							 | 
						|
										return 0, 0, io.EOF
							 | 
						|
									} else if offset == fileSize {
							 | 
						|
										return 0, 0, io.EOF
							 | 
						|
									} else if offset >= fileSize {
							 | 
						|
										glog.V(1).Infof("invalid read, fileSize %d, offset %d for %s", fileSize, offset, fileFullPath)
							 | 
						|
										return 0, 0, io.EOF
							 | 
						|
									}
							 | 
						|
								
							 | 
						|
									if offset < int64(len(entry.Content)) {
							 | 
						|
										totalRead := copy(buff, entry.Content[offset:])
							 | 
						|
										glog.V(4).Infof("file handle read cached %s [%d,%d] %d", fileFullPath, offset, offset+int64(totalRead), totalRead)
							 | 
						|
										return int64(totalRead), 0, nil
							 | 
						|
									}
							 | 
						|
								
							 | 
						|
									// Try RDMA acceleration first if available
							 | 
						|
									if fh.wfs.rdmaClient != nil && fh.wfs.option.RdmaEnabled {
							 | 
						|
										totalRead, ts, err := fh.tryRDMARead(ctx, fileSize, buff, offset, entry)
							 | 
						|
										if err == nil {
							 | 
						|
											glog.V(4).Infof("RDMA read successful for %s [%d,%d] %d", fileFullPath, offset, offset+int64(totalRead), totalRead)
							 | 
						|
											return int64(totalRead), ts, nil
							 | 
						|
										}
							 | 
						|
										glog.V(4).Infof("RDMA read failed for %s, falling back to HTTP: %v", fileFullPath, err)
							 | 
						|
									}
							 | 
						|
								
							 | 
						|
									// Fall back to normal chunk reading
							 | 
						|
									totalRead, ts, err := fh.entryChunkGroup.ReadDataAt(ctx, fileSize, buff, offset)
							 | 
						|
								
							 | 
						|
									if err != nil && err != io.EOF {
							 | 
						|
										glog.Errorf("file handle read %s: %v", fileFullPath, err)
							 | 
						|
									}
							 | 
						|
								
							 | 
						|
									// glog.V(4).Infof("file handle read %s [%d,%d] %d : %v", fileFullPath, offset, offset+int64(totalRead), totalRead, err)
							 | 
						|
								
							 | 
						|
									return int64(totalRead), ts, err
							 | 
						|
								}
							 | 
						|
								
							 | 
						|
								// tryRDMARead attempts to read file data using RDMA acceleration
							 | 
						|
								func (fh *FileHandle) tryRDMARead(ctx context.Context, fileSize int64, buff []byte, offset int64, entry *LockedEntry) (int64, int64, error) {
							 | 
						|
									// For now, we'll try to read the chunks directly using RDMA
							 | 
						|
									// This is a simplified approach - in a full implementation, we'd need to
							 | 
						|
									// handle chunk boundaries, multiple chunks, etc.
							 | 
						|
								
							 | 
						|
									chunks := entry.GetEntry().Chunks
							 | 
						|
									if len(chunks) == 0 {
							 | 
						|
										return 0, 0, fmt.Errorf("no chunks available for RDMA read")
							 | 
						|
									}
							 | 
						|
								
							 | 
						|
									// Find the chunk that contains our offset using binary search
							 | 
						|
									var targetChunk *filer_pb.FileChunk
							 | 
						|
									var chunkOffset int64
							 | 
						|
								
							 | 
						|
									// Get cached cumulative offsets for efficient binary search
							 | 
						|
									cumulativeOffsets := fh.getCumulativeOffsets(chunks)
							 | 
						|
								
							 | 
						|
									// Use binary search to find the chunk containing the offset
							 | 
						|
									chunkIndex := sort.Search(len(chunks), func(i int) bool {
							 | 
						|
										return offset < cumulativeOffsets[i+1]
							 | 
						|
									})
							 | 
						|
								
							 | 
						|
									// Verify the chunk actually contains our offset
							 | 
						|
									if chunkIndex < len(chunks) && offset >= cumulativeOffsets[chunkIndex] {
							 | 
						|
										targetChunk = chunks[chunkIndex]
							 | 
						|
										chunkOffset = offset - cumulativeOffsets[chunkIndex]
							 | 
						|
									}
							 | 
						|
								
							 | 
						|
									if targetChunk == nil {
							 | 
						|
										return 0, 0, fmt.Errorf("no chunk found for offset %d", offset)
							 | 
						|
									}
							 | 
						|
								
							 | 
						|
									// Calculate how much to read from this chunk
							 | 
						|
									remainingInChunk := int64(targetChunk.Size) - chunkOffset
							 | 
						|
									readSize := min(int64(len(buff)), remainingInChunk)
							 | 
						|
								
							 | 
						|
									glog.V(4).Infof("RDMA read attempt: chunk=%s (fileId=%s), chunkOffset=%d, readSize=%d",
							 | 
						|
										targetChunk.FileId, targetChunk.FileId, chunkOffset, readSize)
							 | 
						|
								
							 | 
						|
									// Try RDMA read using file ID directly (more efficient)
							 | 
						|
									data, isRDMA, err := fh.wfs.rdmaClient.ReadNeedle(ctx, targetChunk.FileId, uint64(chunkOffset), uint64(readSize))
							 | 
						|
									if err != nil {
							 | 
						|
										return 0, 0, fmt.Errorf("RDMA read failed: %w", err)
							 | 
						|
									}
							 | 
						|
								
							 | 
						|
									if !isRDMA {
							 | 
						|
										return 0, 0, fmt.Errorf("RDMA not available for chunk")
							 | 
						|
									}
							 | 
						|
								
							 | 
						|
									// Copy data to buffer
							 | 
						|
									copied := copy(buff, data)
							 | 
						|
									return int64(copied), targetChunk.ModifiedTsNs, nil
							 | 
						|
								}
							 | 
						|
								
							 | 
						|
								func (fh *FileHandle) downloadRemoteEntry(entry *LockedEntry) error {
							 | 
						|
								
							 | 
						|
									fileFullPath := fh.FullPath()
							 | 
						|
									dir, _ := fileFullPath.DirAndName()
							 | 
						|
								
							 | 
						|
									err := fh.wfs.WithFilerClient(false, func(client filer_pb.SeaweedFilerClient) error {
							 | 
						|
								
							 | 
						|
										request := &filer_pb.CacheRemoteObjectToLocalClusterRequest{
							 | 
						|
											Directory: string(dir),
							 | 
						|
											Name:      entry.Name,
							 | 
						|
										}
							 | 
						|
								
							 | 
						|
										glog.V(4).Infof("download entry: %v", request)
							 | 
						|
										resp, err := client.CacheRemoteObjectToLocalCluster(context.Background(), request)
							 | 
						|
										if err != nil {
							 | 
						|
											return fmt.Errorf("CacheRemoteObjectToLocalCluster file %s: %v", fileFullPath, err)
							 | 
						|
										}
							 | 
						|
								
							 | 
						|
										fh.SetEntry(resp.Entry)
							 | 
						|
								
							 | 
						|
										fh.wfs.metaCache.InsertEntry(context.Background(), filer.FromPbEntry(request.Directory, resp.Entry))
							 | 
						|
								
							 | 
						|
										return nil
							 | 
						|
									})
							 | 
						|
								
							 | 
						|
									return err
							 | 
						|
								}
							 |