|
|
package shell
import ( "bytes" "context" "flag" "fmt" "github.com/seaweedfs/seaweedfs/weed/filer" "github.com/seaweedfs/seaweedfs/weed/operation" "github.com/seaweedfs/seaweedfs/weed/pb" "github.com/seaweedfs/seaweedfs/weed/pb/filer_pb" "github.com/seaweedfs/seaweedfs/weed/pb/master_pb" "github.com/seaweedfs/seaweedfs/weed/pb/volume_server_pb" "github.com/seaweedfs/seaweedfs/weed/storage" "github.com/seaweedfs/seaweedfs/weed/util" "go.uber.org/atomic" "golang.org/x/exp/slices" "io" "math" "strings" "sync" "time" )
func init() { Commands = append(Commands, &commandFsVerify{}) }
type commandFsVerify struct { env *CommandEnv volumeServers []pb.ServerAddress volumeIds map[uint32][]pb.ServerAddress verbose *bool metadataFromLog *bool concurrency *int modifyTimeAgoAtSec int64 writer io.Writer waitChan map[string]chan struct{} waitChanLock sync.RWMutex }
func (c *commandFsVerify) Name() string { return "fs.verify" }
func (c *commandFsVerify) Help() string { return `recursively verify all files under a directory
fs.verify [-v] [-modifyTimeAgo 1h] /buckets/dir
` }
func (c *commandFsVerify) IsResourceHeavy() bool { return false }
func (c *commandFsVerify) Do(args []string, commandEnv *CommandEnv, writer io.Writer) (err error) { c.env = commandEnv c.writer = writer fsVerifyCommand := flag.NewFlagSet(c.Name(), flag.ContinueOnError) c.verbose = fsVerifyCommand.Bool("v", false, "print out each processed files") modifyTimeAgo := fsVerifyCommand.Duration("modifyTimeAgo", 0, "only include files after this modify time to verify") c.concurrency = fsVerifyCommand.Int("concurrency", 0, "number of parallel verification per volume server") c.metadataFromLog = fsVerifyCommand.Bool("metadataFromLog", false, "Using filer log to get metadata") if err = fsVerifyCommand.Parse(args); err != nil { return err }
path, parseErr := commandEnv.parseUrl(findInputDirectory(fsVerifyCommand.Args())) if parseErr != nil { return parseErr }
c.modifyTimeAgoAtSec = int64(modifyTimeAgo.Seconds()) c.volumeIds = make(map[uint32][]pb.ServerAddress) c.waitChan = make(map[string]chan struct{}) c.volumeServers = []pb.ServerAddress{} defer func() { c.modifyTimeAgoAtSec = 0 c.volumeIds = nil c.waitChan = nil c.volumeServers = nil }()
if err := c.collectVolumeIds(); err != nil { return parseErr }
if *c.concurrency > 0 { for _, volumeServer := range c.volumeServers { volumeServerStr := string(volumeServer) c.waitChan[volumeServerStr] = make(chan struct{}, *c.concurrency) defer close(c.waitChan[volumeServerStr]) } } var fCount, eCount uint64 if *c.metadataFromLog { var wg sync.WaitGroup fCount, eCount, err = c.verifyProcessMetadata(path, &wg) wg.Wait() if err != nil { return err } } else { fCount, eCount, err = c.verifyTraverseBfs(path) } fmt.Fprintf(writer, "verified %d files, error %d files \n", fCount, eCount) return err }
func (c *commandFsVerify) collectVolumeIds() error { topologyInfo, _, err := collectTopologyInfo(c.env, 0) if err != nil { return err } eachDataNode(topologyInfo, func(dc string, rack RackId, nodeInfo *master_pb.DataNodeInfo) { for _, diskInfo := range nodeInfo.DiskInfos { for _, vi := range diskInfo.VolumeInfos { volumeServer := pb.NewServerAddressFromDataNode(nodeInfo) c.volumeIds[vi.Id] = append(c.volumeIds[vi.Id], volumeServer) if !slices.Contains(c.volumeServers, volumeServer) { c.volumeServers = append(c.volumeServers, volumeServer) } } for _, vi := range diskInfo.EcShardInfos { volumeServer := pb.NewServerAddressFromDataNode(nodeInfo) c.volumeIds[vi.Id] = append(c.volumeIds[vi.Id], volumeServer) if !slices.Contains(c.volumeServers, volumeServer) { c.volumeServers = append(c.volumeServers, volumeServer) } } } }) return nil }
func (c *commandFsVerify) verifyChunk(volumeServer pb.ServerAddress, fileId *filer_pb.FileId) error { err := operation.WithVolumeServerClient(false, volumeServer, c.env.option.GrpcDialOption, func(client volume_server_pb.VolumeServerClient) error { _, err := client.VolumeNeedleStatus(context.Background(), &volume_server_pb.VolumeNeedleStatusRequest{ VolumeId: fileId.VolumeId, NeedleId: fileId.FileKey}) return err }, ) if err != nil && !strings.Contains(err.Error(), storage.ErrorDeleted.Error()) { return err } return nil }
type ItemEntry struct { chunks []*filer_pb.FileChunk path util.FullPath }
func (c *commandFsVerify) verifyProcessMetadata(path string, wg *sync.WaitGroup) (fileCount uint64, errCount uint64, err error) { processEventFn := func(resp *filer_pb.SubscribeMetadataResponse) error { message := resp.EventNotification if resp.EventNotification.NewEntry == nil { return nil } chunkCount := len(message.NewEntry.Chunks) if chunkCount == 0 { return nil } entryPath := fmt.Sprintf("%s/%s", message.NewParentPath, message.NewEntry.Name) errorChunksCount := atomic.NewUint64(0) if !c.verifyEntry(entryPath, message.NewEntry.Chunks, errorChunksCount, wg) { if err = c.env.WithFilerClient(false, func(client filer_pb.SeaweedFilerClient) error { entryResp, errReq := client.LookupDirectoryEntry(context.Background(), &filer_pb.LookupDirectoryEntryRequest{ Directory: message.NewParentPath, Name: message.NewEntry.Name, }) if errReq != nil { if strings.HasSuffix(errReq.Error(), "no entry is found in filer store") { return nil } return errReq } if entryResp.Entry.Attributes.Mtime == message.NewEntry.Attributes.Mtime && bytes.Equal(entryResp.Entry.Attributes.Md5, message.NewEntry.Attributes.Md5) { fmt.Fprintf(c.writer, "file: %s needles:%d failed:%d\n", entryPath, chunkCount, errorChunksCount.Load()) errCount++ } return nil }); err != nil { return err } return nil } if *c.verbose { fmt.Fprintf(c.writer, "file: %s needles:%d verifed\n", entryPath, chunkCount) } fileCount++ return nil } metadataFollowOption := &pb.MetadataFollowOption{ ClientName: "shell_verify", ClientId: util.RandomInt32(), ClientEpoch: 0, SelfSignature: 0, PathPrefix: path, AdditionalPathPrefixes: nil, DirectoriesToWatch: nil, StartTsNs: time.Now().Add(-1 * time.Second * time.Duration(c.modifyTimeAgoAtSec)).UnixNano(), StopTsNs: time.Now().UnixNano(), EventErrorType: pb.DontLogError, } return fileCount, errCount, pb.FollowMetadata(c.env.option.FilerAddress, c.env.option.GrpcDialOption, metadataFollowOption, processEventFn) }
func (c *commandFsVerify) verifyEntry(path string, chunks []*filer_pb.FileChunk, errorCount *atomic.Uint64, wg *sync.WaitGroup) bool { fileMsg := fmt.Sprintf("file:%s", path) itemIsVerifed := atomic.NewBool(true) for _, chunk := range chunks { if volumeIds, ok := c.volumeIds[chunk.Fid.VolumeId]; ok { for _, volumeServer := range volumeIds { if *c.concurrency == 0 { if err := c.verifyChunk(volumeServer, chunk.Fid); err != nil { if !(*c.metadataFromLog && strings.HasSuffix(err.Error(), "not found")) { fmt.Fprintf(c.writer, "%s failed verify fileId %s: %+v, at volume server %v\n", fileMsg, chunk.GetFileIdString(), err, volumeServer) } if itemIsVerifed.Load() { itemIsVerifed.Store(false) errorCount.Add(1) } } continue } c.waitChanLock.RLock() waitChan, ok := c.waitChan[string(volumeServer)] c.waitChanLock.RUnlock() if !ok { fmt.Fprintf(c.writer, "%s failed to get channel for %s fileId: %s\n", string(volumeServer), fileMsg, chunk.GetFileIdString()) if itemIsVerifed.Load() { itemIsVerifed.Store(false) errorCount.Add(1) } continue } wg.Add(1) waitChan <- struct{}{} go func(fChunk *filer_pb.FileChunk, path string, volumeServer pb.ServerAddress, msg string) { defer wg.Done() if err := c.verifyChunk(volumeServer, fChunk.Fid); err != nil { if !(*c.metadataFromLog && strings.HasSuffix(err.Error(), "not found")) { fmt.Fprintf(c.writer, "%s failed verify fileId %s: %+v, at volume server %v\n", msg, fChunk.GetFileIdString(), err, volumeServer) } if itemIsVerifed.Load() { itemIsVerifed.Store(false) errorCount.Add(1) } } <-waitChan }(chunk, path, volumeServer, fileMsg) } } else { if !*c.metadataFromLog { err := fmt.Errorf("volumeId %d not found", chunk.Fid.VolumeId) fmt.Fprintf(c.writer, "%s failed verify fileId %s: %+v\n", fileMsg, chunk.GetFileIdString(), err) } if itemIsVerifed.Load() { itemIsVerifed.Store(false) errorCount.Add(1) } break } } return itemIsVerifed.Load() }
func (c *commandFsVerify) verifyTraverseBfs(path string) (fileCount uint64, errCount uint64, err error) { timeNowAtSec := time.Now().Unix() return fileCount, errCount, doTraverseBfsAndSaving(c.env, c.writer, path, false, func(entry *filer_pb.FullEntry, outputChan chan interface{}) (err error) { if c.modifyTimeAgoAtSec > 0 { if entry.Entry.Attributes != nil && c.modifyTimeAgoAtSec < timeNowAtSec-entry.Entry.Attributes.Mtime { return nil } } dataChunks, manifestChunks, resolveErr := filer.ResolveChunkManifest(filer.LookupFn(c.env), entry.Entry.GetChunks(), 0, math.MaxInt64) if resolveErr != nil { return fmt.Errorf("failed to ResolveChunkManifest: %+v", resolveErr) } dataChunks = append(dataChunks, manifestChunks...) if len(dataChunks) > 0 { outputChan <- &ItemEntry{ chunks: dataChunks, path: util.NewFullPath(entry.Dir, entry.Entry.Name), } } return nil }, func(outputChan chan interface{}) { var wg sync.WaitGroup itemErrCount := atomic.NewUint64(0) for itemEntry := range outputChan { i := itemEntry.(*ItemEntry) itemPath := string(i.path) if c.verifyEntry(itemPath, i.chunks, itemErrCount, &wg) { if *c.verbose { fmt.Fprintf(c.writer, "file: %s needles:%d verifed\n", itemPath, len(i.chunks)) } fileCount++ } } wg.Wait() errCount = itemErrCount.Load() }) }
|