You can not select more than 25 topics
			Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
		
		
		
		
		
			
		
			
				
					
					
						
							427 lines
						
					
					
						
							16 KiB
						
					
					
				
			
		
		
		
			
			
			
		
		
	
	
							427 lines
						
					
					
						
							16 KiB
						
					
					
				
								package shell
							 | 
						|
								
							 | 
						|
								import (
							 | 
						|
									"bytes"
							 | 
						|
									"context"
							 | 
						|
									"flag"
							 | 
						|
									"fmt"
							 | 
						|
									"io"
							 | 
						|
									"math"
							 | 
						|
									"net/http"
							 | 
						|
									"sync"
							 | 
						|
									"time"
							 | 
						|
								
							 | 
						|
									"slices"
							 | 
						|
								
							 | 
						|
									"github.com/seaweedfs/seaweedfs/weed/operation"
							 | 
						|
									"github.com/seaweedfs/seaweedfs/weed/pb"
							 | 
						|
									"github.com/seaweedfs/seaweedfs/weed/pb/master_pb"
							 | 
						|
									"github.com/seaweedfs/seaweedfs/weed/pb/volume_server_pb"
							 | 
						|
									"github.com/seaweedfs/seaweedfs/weed/server/constants"
							 | 
						|
									"github.com/seaweedfs/seaweedfs/weed/storage/needle_map"
							 | 
						|
									"google.golang.org/grpc"
							 | 
						|
								)
							 | 
						|
								
							 | 
						|
								func init() {
							 | 
						|
									Commands = append(Commands, &commandVolumeCheckDisk{})
							 | 
						|
								}
							 | 
						|
								
							 | 
						|
								type commandVolumeCheckDisk struct {
							 | 
						|
									env    *CommandEnv
							 | 
						|
									writer io.Writer
							 | 
						|
								}
							 | 
						|
								
							 | 
						|
								func (c *commandVolumeCheckDisk) Name() string {
							 | 
						|
									return "volume.check.disk"
							 | 
						|
								}
							 | 
						|
								
							 | 
						|
								func (c *commandVolumeCheckDisk) Help() string {
							 | 
						|
									return `check all replicated volumes to find and fix inconsistencies. It is optional and resource intensive.
							 | 
						|
								
							 | 
						|
									How it works:
							 | 
						|
									
							 | 
						|
									find all volumes that are replicated
							 | 
						|
									  for each volume id, if there are more than 2 replicas, find one pair with the largest 2 in file count.
							 | 
						|
								      for the pair volume A and B
							 | 
						|
								        append entries in A and not in B to B
							 | 
						|
								        append entries in B and not in A to A
							 | 
						|
								
							 | 
						|
								`
							 | 
						|
								}
							 | 
						|
								
							 | 
						|
								func (c *commandVolumeCheckDisk) HasTag(tag CommandTag) bool {
							 | 
						|
									return tag == ResourceHeavy
							 | 
						|
								}
							 | 
						|
								
							 | 
						|
								func (c *commandVolumeCheckDisk) getVolumeStatusFileCount(vid uint32, dn *master_pb.DataNodeInfo) (totalFileCount, deletedFileCount uint64) {
							 | 
						|
									err := operation.WithVolumeServerClient(false, pb.NewServerAddressWithGrpcPort(dn.Id, int(dn.GrpcPort)), c.env.option.GrpcDialOption, func(volumeServerClient volume_server_pb.VolumeServerClient) error {
							 | 
						|
										resp, reqErr := volumeServerClient.VolumeStatus(context.Background(), &volume_server_pb.VolumeStatusRequest{
							 | 
						|
											VolumeId: uint32(vid),
							 | 
						|
										})
							 | 
						|
										if resp != nil {
							 | 
						|
											totalFileCount = resp.FileCount
							 | 
						|
											deletedFileCount = resp.FileDeletedCount
							 | 
						|
										}
							 | 
						|
										return reqErr
							 | 
						|
									})
							 | 
						|
									if err != nil {
							 | 
						|
										fmt.Fprintf(c.writer, "getting number of files for volume id %d from volumes status: %+v\n", vid, err)
							 | 
						|
									}
							 | 
						|
									return totalFileCount, deletedFileCount
							 | 
						|
								}
							 | 
						|
								
							 | 
						|
								func (c *commandVolumeCheckDisk) eqVolumeFileCount(a, b *VolumeReplica) (bool, bool) {
							 | 
						|
									var waitGroup sync.WaitGroup
							 | 
						|
									var fileCountA, fileCountB, fileDeletedCountA, fileDeletedCountB uint64
							 | 
						|
									waitGroup.Add(1)
							 | 
						|
									go func() {
							 | 
						|
										defer waitGroup.Done()
							 | 
						|
										fileCountA, fileDeletedCountA = c.getVolumeStatusFileCount(a.info.Id, a.location.dataNode)
							 | 
						|
									}()
							 | 
						|
									waitGroup.Add(1)
							 | 
						|
									go func() {
							 | 
						|
										defer waitGroup.Done()
							 | 
						|
										fileCountB, fileDeletedCountB = c.getVolumeStatusFileCount(b.info.Id, b.location.dataNode)
							 | 
						|
									}()
							 | 
						|
									// Trying to synchronize a remote call to two nodes
							 | 
						|
									waitGroup.Wait()
							 | 
						|
									return fileCountA == fileCountB, fileDeletedCountA == fileDeletedCountB
							 | 
						|
								}
							 | 
						|
								
							 | 
						|
								func (c *commandVolumeCheckDisk) shouldSkipVolume(a, b *VolumeReplica, pulseTimeAtSecond int64, syncDeletions, verbose bool) bool {
							 | 
						|
									doSyncDeletedCount := false
							 | 
						|
									if syncDeletions && a.info.DeleteCount != b.info.DeleteCount {
							 | 
						|
										doSyncDeletedCount = true
							 | 
						|
									}
							 | 
						|
									if (a.info.FileCount != b.info.FileCount) || doSyncDeletedCount {
							 | 
						|
										// Do synchronization of volumes, if the modification time was before the last pulsation time
							 | 
						|
										if a.info.ModifiedAtSecond < pulseTimeAtSecond || b.info.ModifiedAtSecond < pulseTimeAtSecond {
							 | 
						|
											return false
							 | 
						|
										}
							 | 
						|
										if eqFileCount, eqDeletedFileCount := c.eqVolumeFileCount(a, b); eqFileCount {
							 | 
						|
											if doSyncDeletedCount && !eqDeletedFileCount {
							 | 
						|
												return false
							 | 
						|
											}
							 | 
						|
											if verbose {
							 | 
						|
												fmt.Fprintf(c.writer, "skipping active volumes %d with the same file counts on %s and %s\n",
							 | 
						|
													a.info.Id, a.location.dataNode.Id, b.location.dataNode.Id)
							 | 
						|
											}
							 | 
						|
										} else {
							 | 
						|
											return false
							 | 
						|
										}
							 | 
						|
									}
							 | 
						|
									return true
							 | 
						|
								}
							 | 
						|
								
							 | 
						|
								func (c *commandVolumeCheckDisk) Do(args []string, commandEnv *CommandEnv, writer io.Writer) (err error) {
							 | 
						|
								
							 | 
						|
									fsckCommand := flag.NewFlagSet(c.Name(), flag.ContinueOnError)
							 | 
						|
									slowMode := fsckCommand.Bool("slow", false, "slow mode checks all replicas even file counts are the same")
							 | 
						|
									verbose := fsckCommand.Bool("v", false, "verbose mode")
							 | 
						|
									volumeId := fsckCommand.Uint("volumeId", 0, "the volume id")
							 | 
						|
									applyChanges := fsckCommand.Bool("force", false, "apply the fix")
							 | 
						|
									syncDeletions := fsckCommand.Bool("syncDeleted", false, "sync of deletions the fix")
							 | 
						|
									nonRepairThreshold := fsckCommand.Float64("nonRepairThreshold", 0.3, "repair when missing keys is not more than this limit")
							 | 
						|
									if err = fsckCommand.Parse(args); err != nil {
							 | 
						|
										return nil
							 | 
						|
									}
							 | 
						|
									infoAboutSimulationMode(writer, *applyChanges, "-force")
							 | 
						|
								
							 | 
						|
									if err = commandEnv.confirmIsLocked(args); err != nil {
							 | 
						|
										return
							 | 
						|
									}
							 | 
						|
								
							 | 
						|
									c.env = commandEnv
							 | 
						|
									c.writer = writer
							 | 
						|
								
							 | 
						|
									// collect topology information
							 | 
						|
									pulseTimeAtSecond := time.Now().Unix() - constants.VolumePulseSeconds*2
							 | 
						|
									topologyInfo, _, err := collectTopologyInfo(commandEnv, 0)
							 | 
						|
									if err != nil {
							 | 
						|
										return err
							 | 
						|
									}
							 | 
						|
									volumeReplicas, _ := collectVolumeReplicaLocations(topologyInfo)
							 | 
						|
								
							 | 
						|
									// pick 1 pairs of volume replica
							 | 
						|
									for _, replicas := range volumeReplicas {
							 | 
						|
										if *volumeId > 0 && replicas[0].info.Id != uint32(*volumeId) {
							 | 
						|
											continue
							 | 
						|
										}
							 | 
						|
										// filter readonly replica
							 | 
						|
										var writableReplicas []*VolumeReplica
							 | 
						|
										for _, replica := range replicas {
							 | 
						|
											if replica.info.ReadOnly {
							 | 
						|
												fmt.Fprintf(writer, "skipping readonly volume %d on %s\n", replica.info.Id, replica.location.dataNode.Id)
							 | 
						|
											} else {
							 | 
						|
												writableReplicas = append(writableReplicas, replica)
							 | 
						|
											}
							 | 
						|
										}
							 | 
						|
								
							 | 
						|
										slices.SortFunc(writableReplicas, func(a, b *VolumeReplica) int {
							 | 
						|
											return int(b.info.FileCount - a.info.FileCount)
							 | 
						|
										})
							 | 
						|
										for len(writableReplicas) >= 2 {
							 | 
						|
											a, b := writableReplicas[0], writableReplicas[1]
							 | 
						|
											if !*slowMode && c.shouldSkipVolume(a, b, pulseTimeAtSecond, *syncDeletions, *verbose) {
							 | 
						|
												// always choose the larger volume to be the source
							 | 
						|
												writableReplicas = append(replicas[:1], writableReplicas[2:]...)
							 | 
						|
												continue
							 | 
						|
											}
							 | 
						|
											if err := c.syncTwoReplicas(a, b, *applyChanges, *syncDeletions, *nonRepairThreshold, *verbose); err != nil {
							 | 
						|
												fmt.Fprintf(writer, "sync volume %d on %s and %s: %v\n", a.info.Id, a.location.dataNode.Id, b.location.dataNode.Id, err)
							 | 
						|
											}
							 | 
						|
											// always choose the larger volume to be the source
							 | 
						|
											if a.info.FileCount > b.info.FileCount {
							 | 
						|
												writableReplicas = append(writableReplicas[:1], writableReplicas[2:]...)
							 | 
						|
											} else {
							 | 
						|
												writableReplicas = writableReplicas[1:]
							 | 
						|
											}
							 | 
						|
										}
							 | 
						|
									}
							 | 
						|
								
							 | 
						|
									return nil
							 | 
						|
								}
							 | 
						|
								
							 | 
						|
								func (c *commandVolumeCheckDisk) syncTwoReplicas(a *VolumeReplica, b *VolumeReplica, applyChanges bool, doSyncDeletions bool, nonRepairThreshold float64, verbose bool) (err error) {
							 | 
						|
									aHasChanges, bHasChanges := true, true
							 | 
						|
									const maxIterations = 5
							 | 
						|
									iteration := 0
							 | 
						|
								
							 | 
						|
									for (aHasChanges || bHasChanges) && iteration < maxIterations {
							 | 
						|
										iteration++
							 | 
						|
										if verbose {
							 | 
						|
											fmt.Fprintf(c.writer, "sync iteration %d for volume %d\n", iteration, a.info.Id)
							 | 
						|
										}
							 | 
						|
								
							 | 
						|
										prevAHasChanges, prevBHasChanges := aHasChanges, bHasChanges
							 | 
						|
										if aHasChanges, bHasChanges, err = c.checkBoth(a, b, applyChanges, doSyncDeletions, nonRepairThreshold, verbose); err != nil {
							 | 
						|
											return err
							 | 
						|
										}
							 | 
						|
								
							 | 
						|
										// Detect if we're stuck in a loop with no progress
							 | 
						|
										if iteration > 1 && prevAHasChanges == aHasChanges && prevBHasChanges == bHasChanges && (aHasChanges || bHasChanges) {
							 | 
						|
											fmt.Fprintf(c.writer, "volume %d sync is not making progress between %s and %s after iteration %d, stopping to prevent infinite loop\n",
							 | 
						|
												a.info.Id, a.location.dataNode.Id, b.location.dataNode.Id, iteration)
							 | 
						|
											return fmt.Errorf("sync not making progress after %d iterations", iteration)
							 | 
						|
										}
							 | 
						|
									}
							 | 
						|
								
							 | 
						|
									if iteration >= maxIterations && (aHasChanges || bHasChanges) {
							 | 
						|
										fmt.Fprintf(c.writer, "volume %d sync reached maximum iterations (%d) between %s and %s, may need manual intervention\n",
							 | 
						|
											a.info.Id, maxIterations, a.location.dataNode.Id, b.location.dataNode.Id)
							 | 
						|
										return fmt.Errorf("reached maximum sync iterations (%d)", maxIterations)
							 | 
						|
									}
							 | 
						|
								
							 | 
						|
									return nil
							 | 
						|
								}
							 | 
						|
								
							 | 
						|
								func (c *commandVolumeCheckDisk) checkBoth(a *VolumeReplica, b *VolumeReplica, applyChanges bool, doSyncDeletions bool, nonRepairThreshold float64, verbose bool) (aHasChanges bool, bHasChanges bool, err error) {
							 | 
						|
									aDB, bDB := needle_map.NewMemDb(), needle_map.NewMemDb()
							 | 
						|
									defer func() {
							 | 
						|
										aDB.Close()
							 | 
						|
										bDB.Close()
							 | 
						|
									}()
							 | 
						|
								
							 | 
						|
									// read index db
							 | 
						|
									readIndexDbCutoffFrom := uint64(time.Now().UnixNano())
							 | 
						|
									if err = readIndexDatabase(aDB, a.info.Collection, a.info.Id, pb.NewServerAddressFromDataNode(a.location.dataNode), verbose, c.writer, c.env.option.GrpcDialOption); err != nil {
							 | 
						|
										return true, true, fmt.Errorf("readIndexDatabase %s volume %d: %v", a.location.dataNode, a.info.Id, err)
							 | 
						|
									}
							 | 
						|
									if err := readIndexDatabase(bDB, b.info.Collection, b.info.Id, pb.NewServerAddressFromDataNode(b.location.dataNode), verbose, c.writer, c.env.option.GrpcDialOption); err != nil {
							 | 
						|
										return true, true, fmt.Errorf("readIndexDatabase %s volume %d: %v", b.location.dataNode, b.info.Id, err)
							 | 
						|
									}
							 | 
						|
								
							 | 
						|
									// find and make up the differences
							 | 
						|
									aHasChanges, err1 := doVolumeCheckDisk(bDB, aDB, b, a, verbose, c.writer, applyChanges, doSyncDeletions, nonRepairThreshold, readIndexDbCutoffFrom, c.env.option.GrpcDialOption)
							 | 
						|
									bHasChanges, err2 := doVolumeCheckDisk(aDB, bDB, a, b, verbose, c.writer, applyChanges, doSyncDeletions, nonRepairThreshold, readIndexDbCutoffFrom, c.env.option.GrpcDialOption)
							 | 
						|
									if err1 != nil {
							 | 
						|
										return aHasChanges, bHasChanges, fmt.Errorf("doVolumeCheckDisk source:%s target:%s volume %d: %v", b.location.dataNode.Id, a.location.dataNode.Id, b.info.Id, err1)
							 | 
						|
									}
							 | 
						|
									if err2 != nil {
							 | 
						|
										return aHasChanges, bHasChanges, fmt.Errorf("doVolumeCheckDisk source:%s target:%s volume %d: %v", a.location.dataNode.Id, b.location.dataNode.Id, a.info.Id, err2)
							 | 
						|
									}
							 | 
						|
									return aHasChanges, bHasChanges, nil
							 | 
						|
								}
							 | 
						|
								
							 | 
						|
								func doVolumeCheckDisk(minuend, subtrahend *needle_map.MemDb, source, target *VolumeReplica, verbose bool, writer io.Writer, applyChanges bool, doSyncDeletions bool, nonRepairThreshold float64, cutoffFromAtNs uint64, grpcDialOption grpc.DialOption) (hasChanges bool, err error) {
							 | 
						|
								
							 | 
						|
									// find missing keys
							 | 
						|
									// hash join, can be more efficient
							 | 
						|
									var missingNeedles []needle_map.NeedleValue
							 | 
						|
									var partiallyDeletedNeedles []needle_map.NeedleValue
							 | 
						|
									var counter int
							 | 
						|
									doCutoffOfLastNeedle := true
							 | 
						|
									minuend.DescendingVisit(func(minuendValue needle_map.NeedleValue) error {
							 | 
						|
										counter++
							 | 
						|
										if subtrahendValue, found := subtrahend.Get(minuendValue.Key); !found {
							 | 
						|
											if minuendValue.Size.IsDeleted() {
							 | 
						|
												return nil
							 | 
						|
											}
							 | 
						|
											if doCutoffOfLastNeedle {
							 | 
						|
												if needleMeta, err := readNeedleMeta(grpcDialOption, pb.NewServerAddressFromDataNode(source.location.dataNode), source.info.Id, minuendValue); err == nil {
							 | 
						|
													// needles older than the cutoff time are not missing yet
							 | 
						|
													if needleMeta.AppendAtNs > cutoffFromAtNs {
							 | 
						|
														return nil
							 | 
						|
													}
							 | 
						|
													doCutoffOfLastNeedle = false
							 | 
						|
												}
							 | 
						|
											}
							 | 
						|
											missingNeedles = append(missingNeedles, minuendValue)
							 | 
						|
										} else {
							 | 
						|
											if minuendValue.Size.IsDeleted() && !subtrahendValue.Size.IsDeleted() {
							 | 
						|
												partiallyDeletedNeedles = append(partiallyDeletedNeedles, minuendValue)
							 | 
						|
											}
							 | 
						|
											if doCutoffOfLastNeedle {
							 | 
						|
												doCutoffOfLastNeedle = false
							 | 
						|
											}
							 | 
						|
										}
							 | 
						|
										return nil
							 | 
						|
									})
							 | 
						|
								
							 | 
						|
									fmt.Fprintf(writer, "volume %d %s has %d entries, %s missed %d and partially deleted %d entries\n",
							 | 
						|
										source.info.Id, source.location.dataNode.Id, counter, target.location.dataNode.Id, len(missingNeedles), len(partiallyDeletedNeedles))
							 | 
						|
								
							 | 
						|
									if counter == 0 || (len(missingNeedles) == 0 && len(partiallyDeletedNeedles) == 0) {
							 | 
						|
										return false, nil
							 | 
						|
									}
							 | 
						|
								
							 | 
						|
									missingNeedlesFraction := float64(len(missingNeedles)) / float64(counter)
							 | 
						|
									if missingNeedlesFraction > nonRepairThreshold {
							 | 
						|
										return false, fmt.Errorf(
							 | 
						|
											"failed to start repair volume %d, percentage of missing keys is greater than the threshold: %.2f > %.2f",
							 | 
						|
											source.info.Id, missingNeedlesFraction, nonRepairThreshold)
							 | 
						|
									}
							 | 
						|
								
							 | 
						|
									for _, needleValue := range missingNeedles {
							 | 
						|
										needleBlob, err := readSourceNeedleBlob(grpcDialOption, pb.NewServerAddressFromDataNode(source.location.dataNode), source.info.Id, needleValue)
							 | 
						|
										if err != nil {
							 | 
						|
											return hasChanges, err
							 | 
						|
										}
							 | 
						|
								
							 | 
						|
										if !applyChanges {
							 | 
						|
											continue
							 | 
						|
										}
							 | 
						|
								
							 | 
						|
										if verbose {
							 | 
						|
											fmt.Fprintf(writer, "read %s %s => %s\n", needleValue.Key.FileId(source.info.Id), source.location.dataNode.Id, target.location.dataNode.Id)
							 | 
						|
										}
							 | 
						|
								
							 | 
						|
										hasChanges = true
							 | 
						|
								
							 | 
						|
										if err = writeNeedleBlobToTarget(grpcDialOption, pb.NewServerAddressFromDataNode(target.location.dataNode), source.info.Id, needleValue, needleBlob); err != nil {
							 | 
						|
											return hasChanges, err
							 | 
						|
										}
							 | 
						|
								
							 | 
						|
									}
							 | 
						|
								
							 | 
						|
									if doSyncDeletions && applyChanges && len(partiallyDeletedNeedles) > 0 {
							 | 
						|
										var fidList []string
							 | 
						|
										for _, needleValue := range partiallyDeletedNeedles {
							 | 
						|
											fidList = append(fidList, needleValue.Key.FileId(source.info.Id))
							 | 
						|
											if verbose {
							 | 
						|
												fmt.Fprintf(writer, "delete %s %s => %s\n", needleValue.Key.FileId(source.info.Id), source.location.dataNode.Id, target.location.dataNode.Id)
							 | 
						|
											}
							 | 
						|
										}
							 | 
						|
										deleteResults := operation.DeleteFileIdsAtOneVolumeServer(
							 | 
						|
											pb.NewServerAddressFromDataNode(target.location.dataNode),
							 | 
						|
											grpcDialOption, fidList, false)
							 | 
						|
								
							 | 
						|
										// Check for errors in results
							 | 
						|
										for _, deleteResult := range deleteResults {
							 | 
						|
											if deleteResult.Error != "" && deleteResult.Error != "not found" {
							 | 
						|
												return hasChanges, fmt.Errorf("delete file %s: %v", deleteResult.FileId, deleteResult.Error)
							 | 
						|
											}
							 | 
						|
											if deleteResult.Status == http.StatusAccepted && deleteResult.Size > 0 {
							 | 
						|
												hasChanges = true
							 | 
						|
											}
							 | 
						|
										}
							 | 
						|
									}
							 | 
						|
									return hasChanges, nil
							 | 
						|
								}
							 | 
						|
								
							 | 
						|
								func readSourceNeedleBlob(grpcDialOption grpc.DialOption, sourceVolumeServer pb.ServerAddress, volumeId uint32, needleValue needle_map.NeedleValue) (needleBlob []byte, err error) {
							 | 
						|
								
							 | 
						|
									err = operation.WithVolumeServerClient(false, sourceVolumeServer, grpcDialOption, func(client volume_server_pb.VolumeServerClient) error {
							 | 
						|
										resp, err := client.ReadNeedleBlob(context.Background(), &volume_server_pb.ReadNeedleBlobRequest{
							 | 
						|
											VolumeId: volumeId,
							 | 
						|
											Offset:   needleValue.Offset.ToActualOffset(),
							 | 
						|
											Size:     int32(needleValue.Size),
							 | 
						|
										})
							 | 
						|
										if err != nil {
							 | 
						|
											return err
							 | 
						|
										}
							 | 
						|
										needleBlob = resp.NeedleBlob
							 | 
						|
										return nil
							 | 
						|
									})
							 | 
						|
									return
							 | 
						|
								}
							 | 
						|
								
							 | 
						|
								func writeNeedleBlobToTarget(grpcDialOption grpc.DialOption, targetVolumeServer pb.ServerAddress, volumeId uint32, needleValue needle_map.NeedleValue, needleBlob []byte) error {
							 | 
						|
								
							 | 
						|
									return operation.WithVolumeServerClient(false, targetVolumeServer, grpcDialOption, func(client volume_server_pb.VolumeServerClient) error {
							 | 
						|
										_, err := client.WriteNeedleBlob(context.Background(), &volume_server_pb.WriteNeedleBlobRequest{
							 | 
						|
											VolumeId:   volumeId,
							 | 
						|
											NeedleId:   uint64(needleValue.Key),
							 | 
						|
											Size:       int32(needleValue.Size),
							 | 
						|
											NeedleBlob: needleBlob,
							 | 
						|
										})
							 | 
						|
										return err
							 | 
						|
									})
							 | 
						|
								
							 | 
						|
								}
							 | 
						|
								
							 | 
						|
								func readIndexDatabase(db *needle_map.MemDb, collection string, volumeId uint32, volumeServer pb.ServerAddress, verbose bool, writer io.Writer, grpcDialOption grpc.DialOption) error {
							 | 
						|
								
							 | 
						|
									var buf bytes.Buffer
							 | 
						|
									if err := copyVolumeIndexFile(collection, volumeId, volumeServer, &buf, verbose, writer, grpcDialOption); err != nil {
							 | 
						|
										return err
							 | 
						|
									}
							 | 
						|
								
							 | 
						|
									if verbose {
							 | 
						|
										fmt.Fprintf(writer, "load collection %s volume %d index size %d from %s ...\n", collection, volumeId, buf.Len(), volumeServer)
							 | 
						|
									}
							 | 
						|
									return db.LoadFilterFromReaderAt(bytes.NewReader(buf.Bytes()), true, false)
							 | 
						|
								}
							 | 
						|
								
							 | 
						|
								func copyVolumeIndexFile(collection string, volumeId uint32, volumeServer pb.ServerAddress, buf *bytes.Buffer, verbose bool, writer io.Writer, grpcDialOption grpc.DialOption) error {
							 | 
						|
								
							 | 
						|
									return operation.WithVolumeServerClient(true, volumeServer, grpcDialOption, func(volumeServerClient volume_server_pb.VolumeServerClient) error {
							 | 
						|
								
							 | 
						|
										ext := ".idx"
							 | 
						|
								
							 | 
						|
										copyFileClient, err := volumeServerClient.CopyFile(context.Background(), &volume_server_pb.CopyFileRequest{
							 | 
						|
											VolumeId:                 volumeId,
							 | 
						|
											Ext:                      ".idx",
							 | 
						|
											CompactionRevision:       math.MaxUint32,
							 | 
						|
											StopOffset:               math.MaxInt64,
							 | 
						|
											Collection:               collection,
							 | 
						|
											IsEcVolume:               false,
							 | 
						|
											IgnoreSourceFileNotFound: false,
							 | 
						|
										})
							 | 
						|
										if err != nil {
							 | 
						|
											return fmt.Errorf("failed to start copying volume %d%s: %v", volumeId, ext, err)
							 | 
						|
										}
							 | 
						|
								
							 | 
						|
										err = writeToBuffer(copyFileClient, buf)
							 | 
						|
										if err != nil {
							 | 
						|
											return fmt.Errorf("failed to copy %d%s from %s: %v", volumeId, ext, volumeServer, err)
							 | 
						|
										}
							 | 
						|
								
							 | 
						|
										return nil
							 | 
						|
								
							 | 
						|
									})
							 | 
						|
								}
							 | 
						|
								
							 | 
						|
								func writeToBuffer(client volume_server_pb.VolumeServer_CopyFileClient, buf *bytes.Buffer) error {
							 | 
						|
									for {
							 | 
						|
										resp, receiveErr := client.Recv()
							 | 
						|
										if receiveErr == io.EOF {
							 | 
						|
											break
							 | 
						|
										}
							 | 
						|
										if receiveErr != nil {
							 | 
						|
											return fmt.Errorf("receiving: %w", receiveErr)
							 | 
						|
										}
							 | 
						|
										buf.Write(resp.FileContent)
							 | 
						|
									}
							 | 
						|
									return nil
							 | 
						|
								}
							 |