Browse Source

"golang.org/x/exp/slices" => "slices" and go fmt

pull/6207/merge
chrislu 2 days ago
parent
commit
ec155022e7
  1. 2
      weed/command/master.go
  2. 2
      weed/filer/filechunks2_test.go
  3. 2
      weed/filer/filechunks_read.go
  4. 2
      weed/filer/filer_notify.go
  5. 3
      weed/filer/meta_replay.go
  6. 2
      weed/filer/redis/universal_redis_store.go
  7. 2
      weed/filer/redis3/redis_cluster_store.go
  8. 2
      weed/filer/redis3/redis_sentinel_store.go
  9. 2
      weed/filer/redis3/redis_store.go
  10. 4
      weed/filer/stream.go
  11. 2
      weed/mq/client/cmd/weed_pub_kv/publisher_kv.go
  12. 2
      weed/mq/client/cmd/weed_sub_kv/subscriber_kv.go
  13. 2
      weed/mq/client/pub_client/publish.go
  14. 24
      weed/mq/pub_balancer/allocate_test.go
  15. 4
      weed/mq/sub_coordinator/consumer_group_instance.go
  16. 8
      weed/mq/sub_coordinator/inflight_message_tracker.go
  17. 2
      weed/mq/topic/local_topic.go
  18. 12
      weed/operation/upload_content.go
  19. 4
      weed/pb/filer_pb/filer_client_bfs.go
  20. 2
      weed/replication/sink/azuresink/azure_sink.go
  21. 2
      weed/s3api/filer_multipart.go
  22. 2
      weed/s3api/s3api_acl_helper.go
  23. 2
      weed/s3api/s3api_object_handlers_delete.go
  24. 2
      weed/server/filer_server_handlers_write_cipher.go
  25. 2
      weed/server/filer_server_handlers_write_upload.go
  26. 2
      weed/shell/command_ec_common.go
  27. 2
      weed/shell/command_fs_merge_volumes.go
  28. 2
      weed/shell/command_fs_verify.go
  29. 2
      weed/shell/command_volume_balance.go
  30. 2
      weed/shell/command_volume_check_disk.go
  31. 6
      weed/shell/command_volume_fix_replication.go
  32. 2
      weed/shell/command_volume_list.go
  33. 2
      weed/shell/command_volume_server_evacuate.go
  34. 2
      weed/shell/shell_liner.go
  35. 2
      weed/storage/disk_location_ec.go
  36. 4
      weed/storage/erasure_coding/ec_locate.go
  37. 2
      weed/storage/erasure_coding/ec_volume.go
  38. 2
      weed/topology/data_center.go
  39. 2
      weed/topology/rack.go
  40. 2
      weed/topology/topology_info.go
  41. 2
      weed/util/chunk_cache/on_disk_cache_layer.go
  42. 4
      weed/util/config.go
  43. 2
      weed/util/queue.go
  44. 2
      weed/util/skiplist/name_batch.go

2
weed/command/master.go

@ -11,7 +11,7 @@ import (
hashicorpRaft "github.com/hashicorp/raft"
"golang.org/x/exp/slices"
"slices"
"github.com/gorilla/mux"
"github.com/seaweedfs/raft/protobuf"

2
weed/filer/filechunks2_test.go

@ -2,8 +2,8 @@ package filer
import (
"github.com/stretchr/testify/assert"
"golang.org/x/exp/slices"
"log"
"slices"
"testing"
"github.com/seaweedfs/seaweedfs/weed/glog"

2
weed/filer/filechunks_read.go

@ -3,7 +3,7 @@ package filer
import (
"container/list"
"github.com/seaweedfs/seaweedfs/weed/pb/filer_pb"
"golang.org/x/exp/slices"
"slices"
)
func readResolvedChunks(chunks []*filer_pb.FileChunk, startOffset int64, stopOffset int64) (visibles *IntervalList[*VisibleInterval]) {

2
weed/filer/filer_notify.go

@ -123,7 +123,7 @@ func (f *Filer) ReadPersistedLogBuffer(startPosition log_buffer.MessagePosition,
err = fmt.Errorf("reading from persisted logs: %v", visitErr)
return
}
var logEntry *filer_pb.LogEntry
var logEntry *filer_pb.LogEntry
for {
logEntry, visitErr = visitor.GetNext()
if visitErr != nil {

3
weed/filer/meta_replay.go

@ -37,10 +37,9 @@ func Replay(filerStore FilerStore, resp *filer_pb.SubscribeMetadataResponse) err
return nil
}
// ParallelProcessDirectoryStructure processes each entry in parallel, and also ensure parent directories are processed first.
// This also assumes the parent directories are in the entryChan already.
func ParallelProcessDirectoryStructure(entryChan chan *Entry, concurrency int, eachEntryFn func(entry *Entry)(error)) (firstErr error) {
func ParallelProcessDirectoryStructure(entryChan chan *Entry, concurrency int, eachEntryFn func(entry *Entry) error) (firstErr error) {
executors := util.NewLimitedConcurrentExecutor(concurrency)

2
weed/filer/redis/universal_redis_store.go

@ -3,7 +3,7 @@ package redis
import (
"context"
"fmt"
"golang.org/x/exp/slices"
"slices"
"strings"
"time"

2
weed/filer/redis3/redis_cluster_store.go

@ -1,9 +1,9 @@
package redis3
import (
"github.com/redis/go-redis/v9"
"github.com/go-redsync/redsync/v4"
"github.com/go-redsync/redsync/v4/redis/goredis/v9"
"github.com/redis/go-redis/v9"
"github.com/seaweedfs/seaweedfs/weed/filer"
"github.com/seaweedfs/seaweedfs/weed/util"
)

2
weed/filer/redis3/redis_sentinel_store.go

@ -3,9 +3,9 @@ package redis3
import (
"time"
"github.com/redis/go-redis/v9"
"github.com/go-redsync/redsync/v4"
"github.com/go-redsync/redsync/v4/redis/goredis/v9"
"github.com/redis/go-redis/v9"
"github.com/seaweedfs/seaweedfs/weed/filer"
"github.com/seaweedfs/seaweedfs/weed/util"
)

2
weed/filer/redis3/redis_store.go

@ -1,9 +1,9 @@
package redis3
import (
"github.com/redis/go-redis/v9"
"github.com/go-redsync/redsync/v4"
"github.com/go-redsync/redsync/v4/redis/goredis/v9"
"github.com/redis/go-redis/v9"
"github.com/seaweedfs/seaweedfs/weed/filer"
"github.com/seaweedfs/seaweedfs/weed/util"
)

4
weed/filer/stream.go

@ -9,14 +9,14 @@ import (
"sync"
"time"
"golang.org/x/exp/slices"
"slices"
"github.com/seaweedfs/seaweedfs/weed/glog"
"github.com/seaweedfs/seaweedfs/weed/pb/filer_pb"
"github.com/seaweedfs/seaweedfs/weed/stats"
"github.com/seaweedfs/seaweedfs/weed/util"
"github.com/seaweedfs/seaweedfs/weed/wdclient"
util_http "github.com/seaweedfs/seaweedfs/weed/util/http"
"github.com/seaweedfs/seaweedfs/weed/wdclient"
)
var getLookupFileIdBackoffSchedule = []time.Duration{

2
weed/mq/client/cmd/weed_pub_kv/publisher_kv.go

@ -5,11 +5,11 @@ import (
"fmt"
"github.com/seaweedfs/seaweedfs/weed/mq/client/pub_client"
"github.com/seaweedfs/seaweedfs/weed/mq/topic"
util_http "github.com/seaweedfs/seaweedfs/weed/util/http"
"log"
"strings"
"sync"
"time"
util_http "github.com/seaweedfs/seaweedfs/weed/util/http"
)
var (

2
weed/mq/client/cmd/weed_sub_kv/subscriber_kv.go

@ -7,11 +7,11 @@ import (
"github.com/seaweedfs/seaweedfs/weed/mq/client/sub_client"
"github.com/seaweedfs/seaweedfs/weed/mq/topic"
"github.com/seaweedfs/seaweedfs/weed/util"
util_http "github.com/seaweedfs/seaweedfs/weed/util/http"
"google.golang.org/grpc"
"google.golang.org/grpc/credentials/insecure"
"strings"
"time"
util_http "github.com/seaweedfs/seaweedfs/weed/util/http"
)
var (

2
weed/mq/client/pub_client/publish.go

@ -50,7 +50,7 @@ func (p *TopicPublisher) FinishPublish() error {
inputBuffer.Enqueue(&mq_pb.DataMessage{
TsNs: time.Now().UnixNano(),
Ctrl: &mq_pb.ControlMessage{
IsClose: true,
IsClose: true,
PublisherName: p.config.PublisherName,
},
})

24
weed/mq/pub_balancer/allocate_test.go

@ -95,8 +95,8 @@ func TestEnsureAssignmentsToActiveBrokersX(t *testing.T) {
followerCount: 1,
assignments: []*mq_pb.BrokerPartitionAssignment{
{
LeaderBroker: "",
Partition: &mq_pb.Partition{},
LeaderBroker: "",
Partition: &mq_pb.Partition{},
FollowerBroker: "localhost:2",
},
},
@ -110,8 +110,8 @@ func TestEnsureAssignmentsToActiveBrokersX(t *testing.T) {
followerCount: 1,
assignments: []*mq_pb.BrokerPartitionAssignment{
{
LeaderBroker: "localhost:1",
Partition: &mq_pb.Partition{},
LeaderBroker: "localhost:1",
Partition: &mq_pb.Partition{},
FollowerBroker: "",
},
},
@ -125,8 +125,8 @@ func TestEnsureAssignmentsToActiveBrokersX(t *testing.T) {
followerCount: 1,
assignments: []*mq_pb.BrokerPartitionAssignment{
{
LeaderBroker: "localhost:1",
Partition: &mq_pb.Partition{},
LeaderBroker: "localhost:1",
Partition: &mq_pb.Partition{},
FollowerBroker: "localhost:200",
},
},
@ -140,8 +140,8 @@ func TestEnsureAssignmentsToActiveBrokersX(t *testing.T) {
followerCount: 1,
assignments: []*mq_pb.BrokerPartitionAssignment{
{
LeaderBroker: "localhost:100",
Partition: &mq_pb.Partition{},
LeaderBroker: "localhost:100",
Partition: &mq_pb.Partition{},
FollowerBroker: "localhost:200",
},
},
@ -155,8 +155,8 @@ func TestEnsureAssignmentsToActiveBrokersX(t *testing.T) {
followerCount: 3,
assignments: []*mq_pb.BrokerPartitionAssignment{
{
LeaderBroker: "localhost:1",
Partition: &mq_pb.Partition{},
LeaderBroker: "localhost:1",
Partition: &mq_pb.Partition{},
FollowerBroker: "localhost:2",
},
},
@ -184,8 +184,8 @@ func TestEnsureAssignmentsToActiveBrokersX(t *testing.T) {
followerCount: 3,
assignments: []*mq_pb.BrokerPartitionAssignment{
{
LeaderBroker: "localhost:1",
Partition: &mq_pb.Partition{},
LeaderBroker: "localhost:1",
Partition: &mq_pb.Partition{},
FollowerBroker: "localhost:2",
},
},

4
weed/mq/sub_coordinator/consumer_group_instance.go

@ -16,8 +16,8 @@ type ConsumerGroupInstance struct {
func NewConsumerGroupInstance(instanceId string, maxPartitionCount int32) *ConsumerGroupInstance {
return &ConsumerGroupInstance{
InstanceId: ConsumerGroupInstanceId(instanceId),
ResponseChan: make(chan *mq_pb.SubscriberToSubCoordinatorResponse, 1),
InstanceId: ConsumerGroupInstanceId(instanceId),
ResponseChan: make(chan *mq_pb.SubscriberToSubCoordinatorResponse, 1),
MaxPartitionCount: maxPartitionCount,
}
}

8
weed/mq/sub_coordinator/inflight_message_tracker.go

@ -84,9 +84,9 @@ type TimestampStatus struct {
// RingBuffer represents a circular buffer to hold timestamps.
type RingBuffer struct {
buffer []*TimestampStatus
head int
size int
buffer []*TimestampStatus
head int
size int
maxTimestamp int64
maxAllAckedTs int64
}
@ -111,7 +111,7 @@ func (rb *RingBuffer) EnflightTimestamp(timestamp int64) {
if rb.size < len(rb.buffer) {
rb.size++
} else {
newBuf := newBuffer(2*len(rb.buffer))
newBuf := newBuffer(2 * len(rb.buffer))
for i := 0; i < rb.size; i++ {
newBuf[i] = rb.buffer[(rb.head+len(rb.buffer)-rb.size+i)%len(rb.buffer)]
}

2
weed/mq/topic/local_topic.go

@ -4,7 +4,7 @@ import "sync"
type LocalTopic struct {
Topic
Partitions []*LocalPartition
Partitions []*LocalPartition
partitionLock sync.RWMutex
}

12
weed/operation/upload_content.go

@ -9,11 +9,11 @@ import (
"io"
"mime"
"mime/multipart"
"sync"
"net/http"
"net/textproto"
"path/filepath"
"strings"
"sync"
"time"
"github.com/seaweedfs/seaweedfs/weed/glog"
@ -66,9 +66,9 @@ func (uploadResult *UploadResult) ToPbFileChunk(fileId string, offset int64, tsN
var (
fileNameEscaper = strings.NewReplacer(`\`, `\\`, `"`, `\"`, "\n", "")
uploader *Uploader
uploaderErr error
once sync.Once
uploader *Uploader
uploaderErr error
once sync.Once
)
// HTTPClient interface for testing
@ -82,7 +82,7 @@ type Uploader struct {
}
func NewUploader() (*Uploader, error) {
once.Do(func () {
once.Do(func() {
// With Dial context
var httpClient *util_http_client.HTTPClient
httpClient, uploaderErr = util_http.NewGlobalHttpClient(util_http_client.AddDialContext)
@ -96,7 +96,7 @@ func NewUploader() (*Uploader, error) {
return uploader, uploaderErr
}
func newUploader(httpClient HTTPClient) (*Uploader) {
func newUploader(httpClient HTTPClient) *Uploader {
return &Uploader{
httpClient: httpClient,
}

4
weed/pb/filer_pb/filer_client_bfs.go

@ -69,7 +69,7 @@ func processOneDirectory(filerClient FilerClient, parentPath util.FullPath, queu
}
func StreamBfs(client SeaweedFilerClient, dir util.FullPath, olderThanTsNs int64, fn func(parentPath util.FullPath, entry *Entry)error) (err error) {
func StreamBfs(client SeaweedFilerClient, dir util.FullPath, olderThanTsNs int64, fn func(parentPath util.FullPath, entry *Entry) error) (err error) {
glog.V(0).Infof("TraverseBfsMetadata %v if before %v", dir, time.Unix(0, olderThanTsNs))
ctx, cancel := context.WithCancel(context.Background())
defer cancel()
@ -87,7 +87,7 @@ func StreamBfs(client SeaweedFilerClient, dir util.FullPath, olderThanTsNs int64
}
return fmt.Errorf("traverse bfs metadata: %v", err)
}
if err := fn(util.FullPath(resp.Directory), resp.Entry); err != nil {
if err := fn(util.FullPath(resp.Directory), resp.Entry); err != nil {
return err
}
}

2
weed/replication/sink/azuresink/azure_sink.go

@ -112,7 +112,7 @@ func (g *AzureSink) CreateEntry(key string, entry *filer_pb.Entry, signatures []
appendBlobURL := g.containerURL.NewAppendBlobURL(key)
accessCondition := azblob.BlobAccessConditions{}
if entry.Attributes!=nil && entry.Attributes.Mtime>0 {
if entry.Attributes != nil && entry.Attributes.Mtime > 0 {
accessCondition.ModifiedAccessConditions.IfUnmodifiedSince = time.Unix(entry.Attributes.Mtime, 0)
}

2
weed/s3api/filer_multipart.go

@ -7,9 +7,9 @@ import (
"fmt"
"github.com/seaweedfs/seaweedfs/weed/s3api/s3_constants"
"github.com/seaweedfs/seaweedfs/weed/stats"
"golang.org/x/exp/slices"
"math"
"path/filepath"
"slices"
"sort"
"strconv"
"strings"

2
weed/s3api/s3api_acl_helper.go

@ -9,9 +9,9 @@ import (
"github.com/seaweedfs/seaweedfs/weed/pb/filer_pb"
"github.com/seaweedfs/seaweedfs/weed/s3api/s3_constants"
"github.com/seaweedfs/seaweedfs/weed/s3api/s3err"
util_http "github.com/seaweedfs/seaweedfs/weed/util/http"
"net/http"
"strings"
util_http "github.com/seaweedfs/seaweedfs/weed/util/http"
)
type AccountManager interface {

2
weed/s3api/s3api_object_handlers_delete.go

@ -3,8 +3,10 @@ package s3api
import (
"encoding/xml"
"fmt"
"github.com/seaweedfs/seaweedfs/weed/s3api/s3_constants"
"io"
"net/http"
"slices"
"strings"
"github.com/seaweedfs/seaweedfs/weed/s3api/s3_constants"

2
weed/server/filer_server_handlers_write_cipher.go

@ -53,7 +53,7 @@ func (fs *FilerServer) encrypt(ctx context.Context, w http.ResponseWriter, r *ht
PairMap: pu.PairMap,
Jwt: auth,
}
uploader, uploaderErr := operation.NewUploader()
if uploaderErr != nil {
return nil, fmt.Errorf("uploader initialization error: %v", uploaderErr)

2
weed/server/filer_server_handlers_write_upload.go

@ -11,7 +11,7 @@ import (
"sync"
"time"
"golang.org/x/exp/slices"
"slices"
"github.com/seaweedfs/seaweedfs/weed/glog"
"github.com/seaweedfs/seaweedfs/weed/operation"

2
weed/shell/command_ec_common.go

@ -18,8 +18,8 @@ import (
"github.com/seaweedfs/seaweedfs/weed/storage/needle"
"github.com/seaweedfs/seaweedfs/weed/storage/super_block"
"github.com/seaweedfs/seaweedfs/weed/storage/types"
"golang.org/x/exp/slices"
"google.golang.org/grpc"
"slices"
)
type DataCenterId string

2
weed/shell/command_fs_merge_volumes.go

@ -13,7 +13,7 @@ import (
"github.com/seaweedfs/seaweedfs/weed/storage/needle"
"github.com/seaweedfs/seaweedfs/weed/wdclient"
"golang.org/x/exp/maps"
"golang.org/x/exp/slices"
"slices"
"github.com/seaweedfs/seaweedfs/weed/operation"
"github.com/seaweedfs/seaweedfs/weed/pb/filer_pb"

2
weed/shell/command_fs_verify.go

@ -20,7 +20,7 @@ import (
"github.com/seaweedfs/seaweedfs/weed/storage"
"github.com/seaweedfs/seaweedfs/weed/util"
"go.uber.org/atomic"
"golang.org/x/exp/slices"
"slices"
)
func init() {

2
weed/shell/command_volume_balance.go

@ -13,7 +13,7 @@ import (
"github.com/seaweedfs/seaweedfs/weed/storage/erasure_coding"
"github.com/seaweedfs/seaweedfs/weed/storage/super_block"
"github.com/seaweedfs/seaweedfs/weed/storage/types"
"golang.org/x/exp/slices"
"slices"
"github.com/seaweedfs/seaweedfs/weed/pb/master_pb"
"github.com/seaweedfs/seaweedfs/weed/storage/needle"

2
weed/shell/command_volume_check_disk.go

@ -17,8 +17,8 @@ import (
"github.com/seaweedfs/seaweedfs/weed/pb/volume_server_pb"
"github.com/seaweedfs/seaweedfs/weed/server/constants"
"github.com/seaweedfs/seaweedfs/weed/storage/needle_map"
"golang.org/x/exp/slices"
"google.golang.org/grpc"
"slices"
)
func init() {

6
weed/shell/command_volume_fix_replication.go

@ -13,10 +13,8 @@ import (
"github.com/seaweedfs/seaweedfs/weed/storage/needle"
"github.com/seaweedfs/seaweedfs/weed/storage/needle_map"
"github.com/seaweedfs/seaweedfs/weed/storage/types"
"github.com/seaweedfs/seaweedfs/weed/util"
"golang.org/x/exp/slices"
"google.golang.org/grpc"
"slices"
"github.com/seaweedfs/seaweedfs/weed/operation"
"github.com/seaweedfs/seaweedfs/weed/pb/master_pb"
@ -346,7 +344,7 @@ func (c *commandVolumeFixReplication) fixOneUnderReplicatedVolume(commandEnv *Co
}
}
if resp.ProcessedBytes > 0 {
fmt.Fprintf(writer, "volume %d processed %s bytes\n", replica.info.Id, util.BytesToHumanReadable(uint64(resp.ProcessedBytes)))
fmt.Fprintf(writer, "volume %d processed %d bytes\n", replica.info.Id, resp.ProcessedBytes)
}
}

2
weed/shell/command_volume_list.go

@ -7,8 +7,8 @@ import (
"github.com/seaweedfs/seaweedfs/weed/pb/master_pb"
"github.com/seaweedfs/seaweedfs/weed/storage/erasure_coding"
"github.com/seaweedfs/seaweedfs/weed/storage/types"
"golang.org/x/exp/slices"
"path/filepath"
"slices"
"strings"
"time"

2
weed/shell/command_volume_server_evacuate.go

@ -11,7 +11,7 @@ import (
"github.com/seaweedfs/seaweedfs/weed/storage/needle"
"github.com/seaweedfs/seaweedfs/weed/storage/super_block"
"github.com/seaweedfs/seaweedfs/weed/storage/types"
"golang.org/x/exp/slices"
"slices"
)
func init() {

2
weed/shell/shell_liner.go

@ -8,12 +8,12 @@ import (
"github.com/seaweedfs/seaweedfs/weed/pb/master_pb"
"github.com/seaweedfs/seaweedfs/weed/util"
"github.com/seaweedfs/seaweedfs/weed/util/grace"
"golang.org/x/exp/slices"
"io"
"math/rand"
"os"
"path"
"regexp"
"slices"
"strings"
"github.com/peterh/liner"

2
weed/storage/disk_location_ec.go

@ -8,7 +8,7 @@ import (
"strconv"
"strings"
"golang.org/x/exp/slices"
"slices"
"github.com/seaweedfs/seaweedfs/weed/storage/erasure_coding"
"github.com/seaweedfs/seaweedfs/weed/storage/needle"

4
weed/storage/erasure_coding/ec_locate.go

@ -5,7 +5,7 @@ import (
)
type Interval struct {
BlockIndex int // the index of the block in either the large blocks or the small blocks
BlockIndex int // the index of the block in either the large blocks or the small blocks
InnerBlockOffset int64
Size types.Size
IsLargeBlock bool // whether the block is a large block or a small block
@ -50,7 +50,7 @@ func LocateData(largeBlockLength, smallBlockLength int64, shardDatSize int64, of
func locateOffset(largeBlockLength, smallBlockLength int64, shardDatSize int64, offset int64) (blockIndex int, isLargeBlock bool, nLargeBlockRows int64, innerBlockOffset int64) {
largeRowSize := largeBlockLength * DataShardsCount
nLargeBlockRows = (shardDatSize-1)/ largeBlockLength
nLargeBlockRows = (shardDatSize - 1) / largeBlockLength
// if offset is within the large block area
if offset < nLargeBlockRows*largeRowSize {

2
weed/storage/erasure_coding/ec_volume.go

@ -9,7 +9,7 @@ import (
"sync"
"time"
"golang.org/x/exp/slices"
"slices"
"github.com/seaweedfs/seaweedfs/weed/pb"
"github.com/seaweedfs/seaweedfs/weed/pb/master_pb"

2
weed/topology/data_center.go

@ -2,7 +2,7 @@ package topology
import (
"github.com/seaweedfs/seaweedfs/weed/pb/master_pb"
"golang.org/x/exp/slices"
"slices"
"strings"
)

2
weed/topology/rack.go

@ -4,7 +4,7 @@ import (
"github.com/seaweedfs/seaweedfs/weed/pb/master_pb"
"github.com/seaweedfs/seaweedfs/weed/storage/types"
"github.com/seaweedfs/seaweedfs/weed/util"
"golang.org/x/exp/slices"
"slices"
"strings"
"time"
)

2
weed/topology/topology_info.go

@ -2,7 +2,7 @@ package topology
import (
"github.com/seaweedfs/seaweedfs/weed/pb/master_pb"
"golang.org/x/exp/slices"
"slices"
"strings"
)

2
weed/util/chunk_cache/on_disk_cache_layer.go

@ -5,8 +5,8 @@ import (
"github.com/seaweedfs/seaweedfs/weed/glog"
"github.com/seaweedfs/seaweedfs/weed/storage"
"github.com/seaweedfs/seaweedfs/weed/storage/types"
"golang.org/x/exp/slices"
"path"
"slices"
)
type OnDiskCacheLayer struct {

4
weed/util/config.go

@ -11,7 +11,7 @@ import (
var (
ConfigurationFileDirectory DirectoryValueType
loadSecurityConfigOnce sync.Once
loadSecurityConfigOnce sync.Once
)
type DirectoryValueType string
@ -32,7 +32,7 @@ type Configuration interface {
SetDefault(key string, value interface{})
}
func LoadSecurityConfiguration(){
func LoadSecurityConfiguration() {
loadSecurityConfigOnce.Do(func() {
LoadConfiguration("security", false)
})

2
weed/util/queue.go

@ -4,7 +4,7 @@ import (
"sync"
)
type node[T any]struct {
type node[T any] struct {
data T
next *node[T]
}

2
weed/util/skiplist/name_batch.go

@ -2,8 +2,8 @@ package skiplist
import (
"github.com/seaweedfs/seaweedfs/weed/glog"
"golang.org/x/exp/slices"
"google.golang.org/protobuf/proto"
"slices"
"strings"
)

Loading…
Cancel
Save