You can not select more than 25 topics
			Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
		
		
		
		
		
			
		
			
				
					
					
						
							751 lines
						
					
					
						
							24 KiB
						
					
					
				
			
		
		
		
			
			
			
		
		
	
	
							751 lines
						
					
					
						
							24 KiB
						
					
					
				
								package shell
							 | 
						|
								
							 | 
						|
								import (
							 | 
						|
									"bufio"
							 | 
						|
									"bytes"
							 | 
						|
									"context"
							 | 
						|
									"errors"
							 | 
						|
									"flag"
							 | 
						|
									"fmt"
							 | 
						|
									"io"
							 | 
						|
									"math"
							 | 
						|
									"net/http"
							 | 
						|
									"net/url"
							 | 
						|
									"os"
							 | 
						|
									"path"
							 | 
						|
									"path/filepath"
							 | 
						|
									"strconv"
							 | 
						|
									"strings"
							 | 
						|
									"sync"
							 | 
						|
									"time"
							 | 
						|
								
							 | 
						|
									"github.com/seaweedfs/seaweedfs/weed/filer"
							 | 
						|
									"github.com/seaweedfs/seaweedfs/weed/operation"
							 | 
						|
									"github.com/seaweedfs/seaweedfs/weed/pb"
							 | 
						|
									"github.com/seaweedfs/seaweedfs/weed/pb/filer_pb"
							 | 
						|
									"github.com/seaweedfs/seaweedfs/weed/pb/master_pb"
							 | 
						|
									"github.com/seaweedfs/seaweedfs/weed/pb/volume_server_pb"
							 | 
						|
									"github.com/seaweedfs/seaweedfs/weed/storage"
							 | 
						|
									"github.com/seaweedfs/seaweedfs/weed/storage/needle"
							 | 
						|
									"github.com/seaweedfs/seaweedfs/weed/storage/needle_map"
							 | 
						|
									"github.com/seaweedfs/seaweedfs/weed/storage/types"
							 | 
						|
									"github.com/seaweedfs/seaweedfs/weed/util"
							 | 
						|
									util_http "github.com/seaweedfs/seaweedfs/weed/util/http"
							 | 
						|
									"golang.org/x/sync/errgroup"
							 | 
						|
								)
							 | 
						|
								
							 | 
						|
								func init() {
							 | 
						|
									Commands = append(Commands, &commandVolumeFsck{})
							 | 
						|
								}
							 | 
						|
								
							 | 
						|
								const (
							 | 
						|
									readbufferSize = 16
							 | 
						|
								)
							 | 
						|
								
							 | 
						|
								type commandVolumeFsck struct {
							 | 
						|
									env                      *CommandEnv
							 | 
						|
									writer                   io.Writer
							 | 
						|
									bucketsPath              string
							 | 
						|
									collection               *string
							 | 
						|
									volumeIds                map[uint32]bool
							 | 
						|
									tempFolder               string
							 | 
						|
									verbose                  *bool
							 | 
						|
									forcePurging             *bool
							 | 
						|
									findMissingChunksInFiler *bool
							 | 
						|
									verifyNeedle             *bool
							 | 
						|
								}
							 | 
						|
								
							 | 
						|
								func (c *commandVolumeFsck) Name() string {
							 | 
						|
									return "volume.fsck"
							 | 
						|
								}
							 | 
						|
								
							 | 
						|
								func (c *commandVolumeFsck) Help() string {
							 | 
						|
									return `check all volumes to find entries not used by the filer. It is optional and resource intensive.
							 | 
						|
								
							 | 
						|
									Important assumption!!!
							 | 
						|
										the system is all used by one filer.
							 | 
						|
								
							 | 
						|
									This command works this way:
							 | 
						|
									1. collect all file ids from all volumes, as set A
							 | 
						|
									2. collect all file ids from the filer, as set B
							 | 
						|
									3. find out the set A subtract B
							 | 
						|
								
							 | 
						|
									If -findMissingChunksInFiler is enabled, this works
							 | 
						|
									in a reverse way:
							 | 
						|
									1. collect all file ids from all volumes, as set A
							 | 
						|
									2. collect all file ids from the filer, as set B
							 | 
						|
									3. find out the set B subtract A
							 | 
						|
								
							 | 
						|
								`
							 | 
						|
								}
							 | 
						|
								
							 | 
						|
								func (c *commandVolumeFsck) HasTag(tag CommandTag) bool {
							 | 
						|
									return tag == ResourceHeavy
							 | 
						|
								}
							 | 
						|
								
							 | 
						|
								func (c *commandVolumeFsck) Do(args []string, commandEnv *CommandEnv, writer io.Writer) (err error) {
							 | 
						|
								
							 | 
						|
									fsckCommand := flag.NewFlagSet(c.Name(), flag.ContinueOnError)
							 | 
						|
									c.verbose = fsckCommand.Bool("v", false, "verbose mode")
							 | 
						|
									c.findMissingChunksInFiler = fsckCommand.Bool("findMissingChunksInFiler", false, "see \"help volume.fsck\"")
							 | 
						|
									c.collection = fsckCommand.String("collection", "", "the collection name")
							 | 
						|
									volumeIds := fsckCommand.String("volumeId", "", "comma separated the volume id")
							 | 
						|
									applyPurging := fsckCommand.Bool("reallyDeleteFromVolume", false, "<expert only!> after detection, delete missing data from volumes / delete missing file entries from filer. Currently this only works with default filerGroup.")
							 | 
						|
									c.forcePurging = fsckCommand.Bool("forcePurging", false, "delete missing data from volumes in one replica used together with applyPurging")
							 | 
						|
									purgeAbsent := fsckCommand.Bool("reallyDeleteFilerEntries", false, "<expert only!> delete missing file entries from filer if the corresponding volume is missing for any reason, please ensure all still existing/expected volumes are connected! used together with findMissingChunksInFiler")
							 | 
						|
									tempPath := fsckCommand.String("tempPath", path.Join(os.TempDir()), "path for temporary idx files")
							 | 
						|
									cutoffTimeAgo := fsckCommand.Duration("cutoffTimeAgo", 5*time.Minute, "only include entries  on volume servers before this cutoff time to check orphan chunks")
							 | 
						|
									modifyTimeAgo := fsckCommand.Duration("modifyTimeAgo", 0, "only include entries after this modify time to check orphan chunks")
							 | 
						|
									c.verifyNeedle = fsckCommand.Bool("verifyNeedles", false, "check needles status from volume server")
							 | 
						|
								
							 | 
						|
									if err = fsckCommand.Parse(args); err != nil {
							 | 
						|
										return nil
							 | 
						|
									}
							 | 
						|
								
							 | 
						|
									if err = commandEnv.confirmIsLocked(args); err != nil {
							 | 
						|
										return
							 | 
						|
									}
							 | 
						|
									c.volumeIds = make(map[uint32]bool)
							 | 
						|
									if *volumeIds != "" {
							 | 
						|
										for _, volumeIdStr := range strings.Split(*volumeIds, ",") {
							 | 
						|
											if volumeIdInt, err := strconv.ParseUint(volumeIdStr, 10, 32); err == nil {
							 | 
						|
												c.volumeIds[uint32(volumeIdInt)] = true
							 | 
						|
											} else {
							 | 
						|
												return fmt.Errorf("parse volumeId string %s to int: %v", volumeIdStr, err)
							 | 
						|
											}
							 | 
						|
										}
							 | 
						|
									}
							 | 
						|
									c.env = commandEnv
							 | 
						|
									c.writer = writer
							 | 
						|
								
							 | 
						|
									c.bucketsPath, err = readFilerBucketsPath(commandEnv)
							 | 
						|
									if err != nil {
							 | 
						|
										return fmt.Errorf("read filer buckets path: %w", err)
							 | 
						|
									}
							 | 
						|
								
							 | 
						|
									// create a temp folder
							 | 
						|
									c.tempFolder, err = os.MkdirTemp(*tempPath, "sw_fsck")
							 | 
						|
									if err != nil {
							 | 
						|
										return fmt.Errorf("failed to create temp folder: %w", err)
							 | 
						|
									}
							 | 
						|
									if *c.verbose {
							 | 
						|
										fmt.Fprintf(c.writer, "working directory: %s\n", c.tempFolder)
							 | 
						|
									}
							 | 
						|
									defer os.RemoveAll(c.tempFolder)
							 | 
						|
								
							 | 
						|
									// collect all volume id locations
							 | 
						|
									dataNodeVolumeIdToVInfo, err := c.collectVolumeIds()
							 | 
						|
									if err != nil {
							 | 
						|
										return fmt.Errorf("failed to collect all volume locations: %w", err)
							 | 
						|
									}
							 | 
						|
								
							 | 
						|
									if err != nil {
							 | 
						|
										return fmt.Errorf("read filer buckets path: %w", err)
							 | 
						|
									}
							 | 
						|
								
							 | 
						|
									var collectCutoffFromAtNs int64 = 0
							 | 
						|
									if cutoffTimeAgo.Seconds() != 0 {
							 | 
						|
										collectCutoffFromAtNs = time.Now().Add(-*cutoffTimeAgo).UnixNano()
							 | 
						|
									}
							 | 
						|
									var collectModifyFromAtNs int64 = 0
							 | 
						|
									if modifyTimeAgo.Seconds() != 0 {
							 | 
						|
										collectModifyFromAtNs = time.Now().Add(-*modifyTimeAgo).UnixNano()
							 | 
						|
									}
							 | 
						|
									// collect each volume file ids
							 | 
						|
									eg, _ := errgroup.WithContext(context.Background())
							 | 
						|
									for _dataNodeId, _volumeIdToVInfo := range dataNodeVolumeIdToVInfo {
							 | 
						|
										dataNodeId, volumeIdToVInfo := _dataNodeId, _volumeIdToVInfo
							 | 
						|
										eg.Go(func() error {
							 | 
						|
											for volumeId, vinfo := range volumeIdToVInfo {
							 | 
						|
												if len(c.volumeIds) > 0 {
							 | 
						|
													if _, ok := c.volumeIds[volumeId]; !ok {
							 | 
						|
														delete(volumeIdToVInfo, volumeId)
							 | 
						|
														continue
							 | 
						|
													}
							 | 
						|
												}
							 | 
						|
												if *c.collection != "" && vinfo.collection != *c.collection {
							 | 
						|
													delete(volumeIdToVInfo, volumeId)
							 | 
						|
													continue
							 | 
						|
												}
							 | 
						|
												err = c.collectOneVolumeFileIds(dataNodeId, volumeId, vinfo)
							 | 
						|
												if err != nil {
							 | 
						|
													return fmt.Errorf("failed to collect file ids from volume %d on %s: %v", volumeId, vinfo.server, err)
							 | 
						|
												}
							 | 
						|
											}
							 | 
						|
											if *c.verbose {
							 | 
						|
												fmt.Fprintf(c.writer, "dn %+v filtred %d volumes and locations.\n", dataNodeId, len(dataNodeVolumeIdToVInfo[dataNodeId]))
							 | 
						|
											}
							 | 
						|
											return nil
							 | 
						|
										})
							 | 
						|
									}
							 | 
						|
									err = eg.Wait()
							 | 
						|
									if err != nil {
							 | 
						|
										fmt.Fprintf(c.writer, "got error: %v", err)
							 | 
						|
										return err
							 | 
						|
									}
							 | 
						|
								
							 | 
						|
									if *c.findMissingChunksInFiler {
							 | 
						|
										// collect all filer file ids and paths
							 | 
						|
								
							 | 
						|
										if err = c.collectFilerFileIdAndPaths(dataNodeVolumeIdToVInfo, *purgeAbsent, collectModifyFromAtNs, collectCutoffFromAtNs); err != nil {
							 | 
						|
											return fmt.Errorf("collectFilerFileIdAndPaths: %w", err)
							 | 
						|
										}
							 | 
						|
										for dataNodeId, volumeIdToVInfo := range dataNodeVolumeIdToVInfo {
							 | 
						|
											// for each volume, check filer file ids
							 | 
						|
											if err = c.findFilerChunksMissingInVolumeServers(volumeIdToVInfo, dataNodeId, *applyPurging); err != nil {
							 | 
						|
												return fmt.Errorf("findFilerChunksMissingInVolumeServers: %w", err)
							 | 
						|
											}
							 | 
						|
										}
							 | 
						|
									} else {
							 | 
						|
										// collect all filer file ids
							 | 
						|
										if err = c.collectFilerFileIdAndPaths(dataNodeVolumeIdToVInfo, false, 0, 0); err != nil {
							 | 
						|
											return fmt.Errorf("failed to collect file ids from filer: %w", err)
							 | 
						|
										}
							 | 
						|
										// volume file ids subtract filer file ids
							 | 
						|
										if err = c.findExtraChunksInVolumeServers(dataNodeVolumeIdToVInfo, *applyPurging, uint64(collectModifyFromAtNs), uint64(collectCutoffFromAtNs)); err != nil {
							 | 
						|
											return fmt.Errorf("findExtraChunksInVolumeServers: %w", err)
							 | 
						|
										}
							 | 
						|
									}
							 | 
						|
								
							 | 
						|
									return nil
							 | 
						|
								}
							 | 
						|
								
							 | 
						|
								func (c *commandVolumeFsck) collectFilerFileIdAndPaths(dataNodeVolumeIdToVInfo map[string]map[uint32]VInfo, purgeAbsent bool, collectModifyFromAtNs int64, cutoffFromAtNs int64) error {
							 | 
						|
									if *c.verbose {
							 | 
						|
										fmt.Fprintf(c.writer, "checking each file from filer path %s...\n", c.getCollectFilerFilePath())
							 | 
						|
									}
							 | 
						|
								
							 | 
						|
									files := make(map[uint32]*os.File)
							 | 
						|
									for _, volumeIdToServer := range dataNodeVolumeIdToVInfo {
							 | 
						|
										for vid := range volumeIdToServer {
							 | 
						|
											if _, ok := files[vid]; ok {
							 | 
						|
												continue
							 | 
						|
											}
							 | 
						|
											dst, openErr := os.OpenFile(getFilerFileIdFile(c.tempFolder, vid), os.O_WRONLY|os.O_CREATE|os.O_TRUNC, 0644)
							 | 
						|
											if openErr != nil {
							 | 
						|
												return fmt.Errorf("failed to create file %s: %v", getFilerFileIdFile(c.tempFolder, vid), openErr)
							 | 
						|
											}
							 | 
						|
											files[vid] = dst
							 | 
						|
										}
							 | 
						|
									}
							 | 
						|
									defer func() {
							 | 
						|
										for _, f := range files {
							 | 
						|
											f.Close()
							 | 
						|
										}
							 | 
						|
									}()
							 | 
						|
								
							 | 
						|
									return doTraverseBfsAndSaving(c.env, c.writer, c.getCollectFilerFilePath(), false,
							 | 
						|
										func(entry *filer_pb.FullEntry, outputChan chan interface{}) (err error) {
							 | 
						|
											if *c.verbose && entry.Entry.IsDirectory {
							 | 
						|
												fmt.Fprintf(c.writer, "checking directory %s\n", util.NewFullPath(entry.Dir, entry.Entry.Name))
							 | 
						|
											}
							 | 
						|
											dataChunks, manifestChunks, resolveErr := filer.ResolveChunkManifest(context.Background(), filer.LookupFn(c.env), entry.Entry.GetChunks(), 0, math.MaxInt64)
							 | 
						|
											if resolveErr != nil {
							 | 
						|
												return fmt.Errorf("failed to ResolveChunkManifest: %+v", resolveErr)
							 | 
						|
											}
							 | 
						|
											dataChunks = append(dataChunks, manifestChunks...)
							 | 
						|
											for _, chunk := range dataChunks {
							 | 
						|
												if cutoffFromAtNs != 0 && chunk.ModifiedTsNs > cutoffFromAtNs {
							 | 
						|
													continue
							 | 
						|
												}
							 | 
						|
												if collectModifyFromAtNs != 0 && chunk.ModifiedTsNs < collectModifyFromAtNs {
							 | 
						|
													continue
							 | 
						|
												}
							 | 
						|
												outputChan <- &Item{
							 | 
						|
													vid:     chunk.Fid.VolumeId,
							 | 
						|
													fileKey: chunk.Fid.FileKey,
							 | 
						|
													cookie:  chunk.Fid.Cookie,
							 | 
						|
													path:    util.NewFullPath(entry.Dir, entry.Entry.Name),
							 | 
						|
												}
							 | 
						|
											}
							 | 
						|
											return nil
							 | 
						|
										},
							 | 
						|
										func(outputChan chan interface{}) {
							 | 
						|
											buffer := make([]byte, readbufferSize)
							 | 
						|
											for item := range outputChan {
							 | 
						|
												i := item.(*Item)
							 | 
						|
												if f, ok := files[i.vid]; ok {
							 | 
						|
													util.Uint64toBytes(buffer, i.fileKey)
							 | 
						|
													util.Uint32toBytes(buffer[8:], i.cookie)
							 | 
						|
													util.Uint32toBytes(buffer[12:], uint32(len(i.path)))
							 | 
						|
													f.Write(buffer)
							 | 
						|
													f.Write([]byte(i.path))
							 | 
						|
												} else if *c.findMissingChunksInFiler && len(c.volumeIds) == 0 {
							 | 
						|
													fmt.Fprintf(c.writer, "%d,%x%08x %s volume not found\n", i.vid, i.fileKey, i.cookie, i.path)
							 | 
						|
													if purgeAbsent {
							 | 
						|
														fmt.Printf("deleting path %s after volume not found", i.path)
							 | 
						|
														c.httpDelete(i.path)
							 | 
						|
													}
							 | 
						|
												}
							 | 
						|
											}
							 | 
						|
										})
							 | 
						|
								}
							 | 
						|
								
							 | 
						|
								func (c *commandVolumeFsck) findFilerChunksMissingInVolumeServers(volumeIdToVInfo map[uint32]VInfo, dataNodeId string, applyPurging bool) error {
							 | 
						|
								
							 | 
						|
									for volumeId, vinfo := range volumeIdToVInfo {
							 | 
						|
										checkErr := c.oneVolumeFileIdsCheckOneVolume(dataNodeId, volumeId, applyPurging)
							 | 
						|
										if checkErr != nil {
							 | 
						|
											return fmt.Errorf("failed to collect file ids from volume %d on %s: %v", volumeId, vinfo.server, checkErr)
							 | 
						|
										}
							 | 
						|
									}
							 | 
						|
									return nil
							 | 
						|
								}
							 | 
						|
								
							 | 
						|
								func (c *commandVolumeFsck) findExtraChunksInVolumeServers(dataNodeVolumeIdToVInfo map[string]map[uint32]VInfo, applyPurging bool, modifyFrom, cutoffFrom uint64) error {
							 | 
						|
								
							 | 
						|
									var totalInUseCount, totalOrphanChunkCount, totalOrphanDataSize uint64
							 | 
						|
									volumeIdOrphanFileIds := make(map[uint32]map[string]bool)
							 | 
						|
									isSeveralReplicas := make(map[uint32]bool)
							 | 
						|
									isEcVolumeReplicas := make(map[uint32]bool)
							 | 
						|
									isReadOnlyReplicas := make(map[uint32]bool)
							 | 
						|
									serverReplicas := make(map[uint32][]pb.ServerAddress)
							 | 
						|
									for dataNodeId, volumeIdToVInfo := range dataNodeVolumeIdToVInfo {
							 | 
						|
										for volumeId, vinfo := range volumeIdToVInfo {
							 | 
						|
											inUseCount, orphanFileIds, orphanDataSize, checkErr := c.oneVolumeFileIdsSubtractFilerFileIds(dataNodeId, volumeId, &vinfo, modifyFrom, cutoffFrom)
							 | 
						|
											if checkErr != nil {
							 | 
						|
												return fmt.Errorf("failed to collect file ids from volume %d on %s: %v", volumeId, vinfo.server, checkErr)
							 | 
						|
											}
							 | 
						|
											isSeveralReplicas[volumeId] = false
							 | 
						|
											if _, found := volumeIdOrphanFileIds[volumeId]; !found {
							 | 
						|
												volumeIdOrphanFileIds[volumeId] = make(map[string]bool)
							 | 
						|
											} else {
							 | 
						|
												isSeveralReplicas[volumeId] = true
							 | 
						|
											}
							 | 
						|
											for _, fid := range orphanFileIds {
							 | 
						|
												if isSeveralReplicas[volumeId] {
							 | 
						|
													if _, found := volumeIdOrphanFileIds[volumeId][fid]; !found {
							 | 
						|
														continue
							 | 
						|
													}
							 | 
						|
												}
							 | 
						|
												volumeIdOrphanFileIds[volumeId][fid] = isSeveralReplicas[volumeId]
							 | 
						|
											}
							 | 
						|
								
							 | 
						|
											totalInUseCount += inUseCount
							 | 
						|
											totalOrphanChunkCount += uint64(len(orphanFileIds))
							 | 
						|
											totalOrphanDataSize += orphanDataSize
							 | 
						|
								
							 | 
						|
											if *c.verbose {
							 | 
						|
												for _, fid := range orphanFileIds {
							 | 
						|
													fmt.Fprintf(c.writer, "%s:%s\n", vinfo.collection, fid)
							 | 
						|
												}
							 | 
						|
											}
							 | 
						|
											isEcVolumeReplicas[volumeId] = vinfo.isEcVolume
							 | 
						|
											if isReadOnly, found := isReadOnlyReplicas[volumeId]; !(found && isReadOnly) {
							 | 
						|
												isReadOnlyReplicas[volumeId] = vinfo.isReadOnly
							 | 
						|
											}
							 | 
						|
											serverReplicas[volumeId] = append(serverReplicas[volumeId], vinfo.server)
							 | 
						|
										}
							 | 
						|
								
							 | 
						|
										for volumeId, orphanReplicaFileIds := range volumeIdOrphanFileIds {
							 | 
						|
											if !(applyPurging && len(orphanReplicaFileIds) > 0) {
							 | 
						|
												continue
							 | 
						|
											}
							 | 
						|
											orphanFileIds := []string{}
							 | 
						|
											for fid, foundInAllReplicas := range orphanReplicaFileIds {
							 | 
						|
												if !isSeveralReplicas[volumeId] || *c.forcePurging || (isSeveralReplicas[volumeId] && foundInAllReplicas) {
							 | 
						|
													orphanFileIds = append(orphanFileIds, fid)
							 | 
						|
												}
							 | 
						|
											}
							 | 
						|
											if !(len(orphanFileIds) > 0) {
							 | 
						|
												continue
							 | 
						|
											}
							 | 
						|
											if *c.verbose {
							 | 
						|
												fmt.Fprintf(c.writer, "purging process for volume %d.\n", volumeId)
							 | 
						|
											}
							 | 
						|
								
							 | 
						|
											if isEcVolumeReplicas[volumeId] {
							 | 
						|
												fmt.Fprintf(c.writer, "skip purging for Erasure Coded volume %d.\n", volumeId)
							 | 
						|
												continue
							 | 
						|
											}
							 | 
						|
											for _, server := range serverReplicas[volumeId] {
							 | 
						|
												needleVID := needle.VolumeId(volumeId)
							 | 
						|
								
							 | 
						|
												if isReadOnlyReplicas[volumeId] {
							 | 
						|
													err := markVolumeWritable(c.env.option.GrpcDialOption, needleVID, server, true, false)
							 | 
						|
													if err != nil {
							 | 
						|
														return fmt.Errorf("mark volume %d read/write: %v", volumeId, err)
							 | 
						|
													}
							 | 
						|
													fmt.Fprintf(c.writer, "temporarily marked %d on server %v writable for forced purge\n", volumeId, server)
							 | 
						|
													defer markVolumeWritable(c.env.option.GrpcDialOption, needleVID, server, false, false)
							 | 
						|
								
							 | 
						|
													fmt.Fprintf(c.writer, "marked %d on server %v writable for forced purge\n", volumeId, server)
							 | 
						|
												}
							 | 
						|
								
							 | 
						|
												if *c.verbose {
							 | 
						|
													fmt.Fprintf(c.writer, "purging files from volume %d\n", volumeId)
							 | 
						|
												}
							 | 
						|
								
							 | 
						|
												if err := c.purgeFileIdsForOneVolume(volumeId, orphanFileIds); err != nil {
							 | 
						|
													return fmt.Errorf("purging volume %d: %v", volumeId, err)
							 | 
						|
												}
							 | 
						|
											}
							 | 
						|
										}
							 | 
						|
									}
							 | 
						|
								
							 | 
						|
									if !applyPurging {
							 | 
						|
										var pct float64
							 | 
						|
								
							 | 
						|
										if totalCount := totalOrphanChunkCount + totalInUseCount; totalCount > 0 {
							 | 
						|
											pct = float64(totalOrphanChunkCount) * 100 / (float64(totalCount))
							 | 
						|
										}
							 | 
						|
								
							 | 
						|
										fmt.Fprintf(c.writer, "\nTotal\t\tentries:%d\torphan:%d\t%.2f%%\t%dB\n",
							 | 
						|
											totalOrphanChunkCount+totalInUseCount, totalOrphanChunkCount, pct, totalOrphanDataSize)
							 | 
						|
								
							 | 
						|
										fmt.Fprintf(c.writer, "This could be normal if multiple filers or no filers are used.\n")
							 | 
						|
									}
							 | 
						|
								
							 | 
						|
									if totalOrphanChunkCount == 0 {
							 | 
						|
										fmt.Fprintf(c.writer, "no orphan data\n")
							 | 
						|
									}
							 | 
						|
								
							 | 
						|
									return nil
							 | 
						|
								}
							 | 
						|
								
							 | 
						|
								func (c *commandVolumeFsck) collectOneVolumeFileIds(dataNodeId string, volumeId uint32, vinfo VInfo) error {
							 | 
						|
								
							 | 
						|
									if *c.verbose {
							 | 
						|
										fmt.Fprintf(c.writer, "collecting volume %d file ids from %s ...\n", volumeId, vinfo.server)
							 | 
						|
									}
							 | 
						|
								
							 | 
						|
									return operation.WithVolumeServerClient(false, vinfo.server, c.env.option.GrpcDialOption,
							 | 
						|
										func(volumeServerClient volume_server_pb.VolumeServerClient) error {
							 | 
						|
											ext := ".idx"
							 | 
						|
											if vinfo.isEcVolume {
							 | 
						|
												ext = ".ecx"
							 | 
						|
											}
							 | 
						|
								
							 | 
						|
											copyFileClient, err := volumeServerClient.CopyFile(context.Background(), &volume_server_pb.CopyFileRequest{
							 | 
						|
												VolumeId:                 volumeId,
							 | 
						|
												Ext:                      ext,
							 | 
						|
												CompactionRevision:       math.MaxUint32,
							 | 
						|
												StopOffset:               math.MaxInt64,
							 | 
						|
												Collection:               vinfo.collection,
							 | 
						|
												IsEcVolume:               vinfo.isEcVolume,
							 | 
						|
												IgnoreSourceFileNotFound: false,
							 | 
						|
											})
							 | 
						|
											if err != nil {
							 | 
						|
												return fmt.Errorf("failed to start copying volume %d%s: %v", volumeId, ext, err)
							 | 
						|
											}
							 | 
						|
								
							 | 
						|
											var buf bytes.Buffer
							 | 
						|
											for {
							 | 
						|
												resp, err := copyFileClient.Recv()
							 | 
						|
												if errors.Is(err, io.EOF) {
							 | 
						|
													break
							 | 
						|
												}
							 | 
						|
												if err != nil {
							 | 
						|
													return err
							 | 
						|
												}
							 | 
						|
												buf.Write(resp.FileContent)
							 | 
						|
											}
							 | 
						|
											idxFilename := getVolumeFileIdFile(c.tempFolder, dataNodeId, volumeId)
							 | 
						|
											err = writeToFile(buf.Bytes(), idxFilename)
							 | 
						|
											if err != nil {
							 | 
						|
												return fmt.Errorf("failed to copy %d%s from %s: %v", volumeId, ext, vinfo.server, err)
							 | 
						|
											}
							 | 
						|
								
							 | 
						|
											return nil
							 | 
						|
										})
							 | 
						|
								
							 | 
						|
								}
							 | 
						|
								
							 | 
						|
								type Item struct {
							 | 
						|
									vid     uint32
							 | 
						|
									fileKey uint64
							 | 
						|
									cookie  uint32
							 | 
						|
									path    util.FullPath
							 | 
						|
								}
							 | 
						|
								
							 | 
						|
								func (c *commandVolumeFsck) readFilerFileIdFile(volumeId uint32, fn func(needleId types.NeedleId, itemPath util.FullPath)) error {
							 | 
						|
									fp, err := os.Open(getFilerFileIdFile(c.tempFolder, volumeId))
							 | 
						|
									if err != nil {
							 | 
						|
										return err
							 | 
						|
									}
							 | 
						|
									defer fp.Close()
							 | 
						|
								
							 | 
						|
									br := bufio.NewReader(fp)
							 | 
						|
									buffer := make([]byte, readbufferSize)
							 | 
						|
									var readSize int
							 | 
						|
									var readErr error
							 | 
						|
									item := &Item{vid: volumeId}
							 | 
						|
									for {
							 | 
						|
										readSize, readErr = io.ReadFull(br, buffer)
							 | 
						|
										if errors.Is(readErr, io.EOF) {
							 | 
						|
											break
							 | 
						|
										}
							 | 
						|
										if readErr != nil {
							 | 
						|
											return readErr
							 | 
						|
										}
							 | 
						|
										if readSize != readbufferSize {
							 | 
						|
											return fmt.Errorf("readSize mismatch")
							 | 
						|
										}
							 | 
						|
										item.fileKey = util.BytesToUint64(buffer[:8])
							 | 
						|
										item.cookie = util.BytesToUint32(buffer[8:12])
							 | 
						|
										pathSize := util.BytesToUint32(buffer[12:16])
							 | 
						|
										pathBytes := make([]byte, int(pathSize))
							 | 
						|
										n, err := io.ReadFull(br, pathBytes)
							 | 
						|
										if err != nil {
							 | 
						|
											fmt.Fprintf(c.writer, "%d,%x%08x in unexpected error: %v\n", volumeId, item.fileKey, item.cookie, err)
							 | 
						|
										}
							 | 
						|
										if n != int(pathSize) {
							 | 
						|
											fmt.Fprintf(c.writer, "%d,%x%08x %d unexpected file name size %d\n", volumeId, item.fileKey, item.cookie, pathSize, n)
							 | 
						|
										}
							 | 
						|
										item.path = util.FullPath(pathBytes)
							 | 
						|
										needleId := types.NeedleId(item.fileKey)
							 | 
						|
										fn(needleId, item.path)
							 | 
						|
									}
							 | 
						|
									return nil
							 | 
						|
								}
							 | 
						|
								
							 | 
						|
								func (c *commandVolumeFsck) oneVolumeFileIdsCheckOneVolume(dataNodeId string, volumeId uint32, applyPurging bool) (err error) {
							 | 
						|
									if *c.verbose {
							 | 
						|
										fmt.Fprintf(c.writer, "find missing file chunks in dataNodeId %s volume %d ...\n", dataNodeId, volumeId)
							 | 
						|
									}
							 | 
						|
								
							 | 
						|
									db := needle_map.NewMemDb()
							 | 
						|
									defer db.Close()
							 | 
						|
								
							 | 
						|
									if err = db.LoadFromIdx(getVolumeFileIdFile(c.tempFolder, dataNodeId, volumeId)); err != nil {
							 | 
						|
										return
							 | 
						|
									}
							 | 
						|
									if err = c.readFilerFileIdFile(volumeId, func(needleId types.NeedleId, itemPath util.FullPath) {
							 | 
						|
										if _, found := db.Get(needleId); !found {
							 | 
						|
											fmt.Fprintf(c.writer, "%s\n", itemPath)
							 | 
						|
											if applyPurging {
							 | 
						|
												c.httpDelete(itemPath)
							 | 
						|
											}
							 | 
						|
										}
							 | 
						|
									}); err != nil {
							 | 
						|
										return
							 | 
						|
									}
							 | 
						|
									return nil
							 | 
						|
								}
							 | 
						|
								
							 | 
						|
								func (c *commandVolumeFsck) httpDelete(path util.FullPath) {
							 | 
						|
									req, err := http.NewRequest(http.MethodDelete, "", nil)
							 | 
						|
								
							 | 
						|
									req.URL = &url.URL{
							 | 
						|
										Scheme: "http",
							 | 
						|
										Host:   c.env.option.FilerAddress.ToHttpAddress(),
							 | 
						|
										Path:   string(path),
							 | 
						|
									}
							 | 
						|
									if *c.verbose {
							 | 
						|
										fmt.Fprintf(c.writer, "full HTTP delete request to be sent: %v\n", req)
							 | 
						|
									}
							 | 
						|
									if err != nil {
							 | 
						|
										fmt.Fprintf(c.writer, "HTTP delete request error: %v\n", err)
							 | 
						|
									}
							 | 
						|
								
							 | 
						|
									resp, err := util_http.GetGlobalHttpClient().Do(req)
							 | 
						|
									if err != nil {
							 | 
						|
										fmt.Fprintf(c.writer, "DELETE fetch error: %v\n", err)
							 | 
						|
									}
							 | 
						|
									defer resp.Body.Close()
							 | 
						|
								
							 | 
						|
									_, err = io.ReadAll(resp.Body)
							 | 
						|
									if err != nil {
							 | 
						|
										fmt.Fprintf(c.writer, "DELETE response error: %v\n", err)
							 | 
						|
									}
							 | 
						|
								
							 | 
						|
									if *c.verbose {
							 | 
						|
										fmt.Fprintln(c.writer, "delete response Status : ", resp.Status)
							 | 
						|
										fmt.Fprintln(c.writer, "delete response Headers : ", resp.Header)
							 | 
						|
									}
							 | 
						|
								}
							 | 
						|
								
							 | 
						|
								func (c *commandVolumeFsck) oneVolumeFileIdsSubtractFilerFileIds(dataNodeId string, volumeId uint32, vinfo *VInfo, modifyFrom, cutoffFrom uint64) (inUseCount uint64, orphanFileIds []string, orphanDataSize uint64, err error) {
							 | 
						|
								
							 | 
						|
									volumeFileIdDb := needle_map.NewMemDb()
							 | 
						|
									defer volumeFileIdDb.Close()
							 | 
						|
								
							 | 
						|
									if err = volumeFileIdDb.LoadFromIdx(getVolumeFileIdFile(c.tempFolder, dataNodeId, volumeId)); err != nil {
							 | 
						|
										err = fmt.Errorf("failed to LoadFromIdx %+v", err)
							 | 
						|
										return
							 | 
						|
									}
							 | 
						|
								
							 | 
						|
									if err = c.readFilerFileIdFile(volumeId, func(filerNeedleId types.NeedleId, itemPath util.FullPath) {
							 | 
						|
										inUseCount++
							 | 
						|
										if *c.verifyNeedle {
							 | 
						|
											if needleValue, ok := volumeFileIdDb.Get(filerNeedleId); ok && !needleValue.Size.IsDeleted() {
							 | 
						|
												if _, err := readNeedleStatus(c.env.option.GrpcDialOption, vinfo.server, volumeId, *needleValue); err != nil {
							 | 
						|
													// files may be deleted during copying filesIds
							 | 
						|
													if !strings.Contains(err.Error(), storage.ErrorDeleted.Error()) {
							 | 
						|
														fmt.Fprintf(c.writer, "failed to read %d:%s needle status of file %s: %+v\n",
							 | 
						|
															volumeId, filerNeedleId.String(), itemPath, err)
							 | 
						|
														if *c.forcePurging {
							 | 
						|
															return
							 | 
						|
														}
							 | 
						|
													}
							 | 
						|
												}
							 | 
						|
											}
							 | 
						|
										}
							 | 
						|
								
							 | 
						|
										if err = volumeFileIdDb.Delete(filerNeedleId); err != nil && *c.verbose {
							 | 
						|
											fmt.Fprintf(c.writer, "failed to nm.delete %s(%+v): %+v", itemPath, filerNeedleId, err)
							 | 
						|
										}
							 | 
						|
									}); err != nil {
							 | 
						|
										err = fmt.Errorf("failed to readFilerFileIdFile %+v", err)
							 | 
						|
										return
							 | 
						|
									}
							 | 
						|
								
							 | 
						|
									var orphanFileCount uint64
							 | 
						|
									if err = volumeFileIdDb.AscendingVisit(func(n needle_map.NeedleValue) error {
							 | 
						|
										if n.Size.IsDeleted() {
							 | 
						|
											return nil
							 | 
						|
										}
							 | 
						|
										if cutoffFrom > 0 || modifyFrom > 0 {
							 | 
						|
											return operation.WithVolumeServerClient(false, vinfo.server, c.env.option.GrpcDialOption,
							 | 
						|
												func(volumeServerClient volume_server_pb.VolumeServerClient) error {
							 | 
						|
													resp, err := volumeServerClient.ReadNeedleMeta(context.Background(), &volume_server_pb.ReadNeedleMetaRequest{
							 | 
						|
														VolumeId: volumeId,
							 | 
						|
														NeedleId: types.NeedleIdToUint64(n.Key),
							 | 
						|
														Offset:   n.Offset.ToActualOffset(),
							 | 
						|
														Size:     int32(n.Size),
							 | 
						|
													})
							 | 
						|
													if err != nil {
							 | 
						|
														return fmt.Errorf("read needle meta with id %d from volume %d: %v", n.Key, volumeId, err)
							 | 
						|
													}
							 | 
						|
													if (modifyFrom == 0 || modifyFrom <= resp.AppendAtNs) && (cutoffFrom == 0 || resp.AppendAtNs <= cutoffFrom) {
							 | 
						|
														orphanFileIds = append(orphanFileIds, n.Key.FileId(volumeId))
							 | 
						|
														orphanFileCount++
							 | 
						|
														orphanDataSize += uint64(n.Size)
							 | 
						|
													}
							 | 
						|
													return nil
							 | 
						|
												})
							 | 
						|
										} else {
							 | 
						|
											orphanFileIds = append(orphanFileIds, n.Key.FileId(volumeId))
							 | 
						|
											orphanFileCount++
							 | 
						|
											orphanDataSize += uint64(n.Size)
							 | 
						|
										}
							 | 
						|
										return nil
							 | 
						|
									}); err != nil {
							 | 
						|
										err = fmt.Errorf("failed to AscendingVisit %+v", err)
							 | 
						|
										return
							 | 
						|
									}
							 | 
						|
								
							 | 
						|
									if orphanFileCount > 0 {
							 | 
						|
										pct := float64(orphanFileCount*100) / (float64(orphanFileCount + inUseCount))
							 | 
						|
										fmt.Fprintf(c.writer, "dataNode:%s\tvolume:%d\tentries:%d\torphan:%d\t%.2f%%\t%dB\n",
							 | 
						|
											dataNodeId, volumeId, orphanFileCount+inUseCount, orphanFileCount, pct, orphanDataSize)
							 | 
						|
									}
							 | 
						|
								
							 | 
						|
									return
							 | 
						|
								
							 | 
						|
								}
							 | 
						|
								
							 | 
						|
								type VInfo struct {
							 | 
						|
									server     pb.ServerAddress
							 | 
						|
									collection string
							 | 
						|
									isEcVolume bool
							 | 
						|
									isReadOnly bool
							 | 
						|
								}
							 | 
						|
								
							 | 
						|
								func (c *commandVolumeFsck) collectVolumeIds() (volumeIdToServer map[string]map[uint32]VInfo, err error) {
							 | 
						|
								
							 | 
						|
									if *c.verbose {
							 | 
						|
										fmt.Fprintf(c.writer, "collecting volume id and locations from master ...\n")
							 | 
						|
									}
							 | 
						|
								
							 | 
						|
									volumeIdToServer = make(map[string]map[uint32]VInfo)
							 | 
						|
									// collect topology information
							 | 
						|
									topologyInfo, _, err := collectTopologyInfo(c.env, 0)
							 | 
						|
									if err != nil {
							 | 
						|
										return
							 | 
						|
									}
							 | 
						|
								
							 | 
						|
									eachDataNode(topologyInfo, func(dc DataCenterId, rack RackId, t *master_pb.DataNodeInfo) {
							 | 
						|
										var volumeCount, ecShardCount int
							 | 
						|
										dataNodeId := t.GetId()
							 | 
						|
										for _, diskInfo := range t.DiskInfos {
							 | 
						|
											if _, ok := volumeIdToServer[dataNodeId]; !ok {
							 | 
						|
												volumeIdToServer[dataNodeId] = make(map[uint32]VInfo)
							 | 
						|
											}
							 | 
						|
											for _, vi := range diskInfo.VolumeInfos {
							 | 
						|
												volumeIdToServer[dataNodeId][vi.Id] = VInfo{
							 | 
						|
													server:     pb.NewServerAddressFromDataNode(t),
							 | 
						|
													collection: vi.Collection,
							 | 
						|
													isEcVolume: false,
							 | 
						|
													isReadOnly: vi.ReadOnly,
							 | 
						|
												}
							 | 
						|
												volumeCount += 1
							 | 
						|
											}
							 | 
						|
											for _, ecShardInfo := range diskInfo.EcShardInfos {
							 | 
						|
												volumeIdToServer[dataNodeId][ecShardInfo.Id] = VInfo{
							 | 
						|
													server:     pb.NewServerAddressFromDataNode(t),
							 | 
						|
													collection: ecShardInfo.Collection,
							 | 
						|
													isEcVolume: true,
							 | 
						|
													isReadOnly: true,
							 | 
						|
												}
							 | 
						|
												ecShardCount += 1
							 | 
						|
											}
							 | 
						|
										}
							 | 
						|
										if *c.verbose {
							 | 
						|
											fmt.Fprintf(c.writer, "dn %+v collected %d volumes and %d ec shards.\n", dataNodeId, volumeCount, ecShardCount)
							 | 
						|
										}
							 | 
						|
									})
							 | 
						|
									return
							 | 
						|
								}
							 | 
						|
								
							 | 
						|
								func (c *commandVolumeFsck) purgeFileIdsForOneVolume(volumeId uint32, fileIds []string) (err error) {
							 | 
						|
									fmt.Fprintf(c.writer, "purging orphan data for volume %d...\n", volumeId)
							 | 
						|
									locations, found := c.env.MasterClient.GetLocations(volumeId)
							 | 
						|
									if !found {
							 | 
						|
										return fmt.Errorf("failed to find volume %d locations", volumeId)
							 | 
						|
									}
							 | 
						|
								
							 | 
						|
									resultChan := make(chan []*volume_server_pb.DeleteResult, len(locations))
							 | 
						|
									var wg sync.WaitGroup
							 | 
						|
									for _, location := range locations {
							 | 
						|
										wg.Add(1)
							 | 
						|
										go func(server pb.ServerAddress, fidList []string) {
							 | 
						|
											defer wg.Done()
							 | 
						|
								
							 | 
						|
											deleteResults := operation.DeleteFileIdsAtOneVolumeServer(server, c.env.option.GrpcDialOption, fidList, false)
							 | 
						|
											if deleteResults != nil {
							 | 
						|
												resultChan <- deleteResults
							 | 
						|
											}
							 | 
						|
								
							 | 
						|
										}(location.ServerAddress(), fileIds)
							 | 
						|
									}
							 | 
						|
									wg.Wait()
							 | 
						|
									close(resultChan)
							 | 
						|
								
							 | 
						|
									for results := range resultChan {
							 | 
						|
										for _, result := range results {
							 | 
						|
											if result.Error != "" {
							 | 
						|
												fmt.Fprintf(c.writer, "purge error: %s\n", result.Error)
							 | 
						|
											}
							 | 
						|
										}
							 | 
						|
									}
							 | 
						|
								
							 | 
						|
									return
							 | 
						|
								}
							 | 
						|
								
							 | 
						|
								func (c *commandVolumeFsck) getCollectFilerFilePath() string {
							 | 
						|
									if *c.collection != "" {
							 | 
						|
										return fmt.Sprintf("%s/%s", c.bucketsPath, *c.collection)
							 | 
						|
									}
							 | 
						|
									return "/"
							 | 
						|
								}
							 | 
						|
								
							 | 
						|
								func getVolumeFileIdFile(tempFolder string, dataNodeid string, vid uint32) string {
							 | 
						|
									return filepath.Join(tempFolder, fmt.Sprintf("%s_%d.idx", dataNodeid, vid))
							 | 
						|
								}
							 | 
						|
								
							 | 
						|
								func getFilerFileIdFile(tempFolder string, vid uint32) string {
							 | 
						|
									return filepath.Join(tempFolder, fmt.Sprintf("%d.fid", vid))
							 | 
						|
								}
							 | 
						|
								
							 | 
						|
								func writeToFile(bytes []byte, fileName string) error {
							 | 
						|
									flags := os.O_WRONLY | os.O_CREATE | os.O_TRUNC
							 | 
						|
									dst, err := os.OpenFile(fileName, flags, 0644)
							 | 
						|
									if err != nil {
							 | 
						|
										return nil
							 | 
						|
									}
							 | 
						|
									defer dst.Close()
							 | 
						|
								
							 | 
						|
									dst.Write(bytes)
							 | 
						|
									return nil
							 | 
						|
								}
							 |