You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

169 lines
5.1 KiB

  1. package weed_server
  2. // https://yusufs.medium.com/creating-distributed-kv-database-by-implementing-raft-consensus-using-golang-d0884eef2e28
  3. // https://github.com/Jille/raft-grpc-example/blob/cd5bcab0218f008e044fbeee4facdd01b06018ad/application.go#L18
  4. import (
  5. "fmt"
  6. transport "github.com/Jille/raft-grpc-transport"
  7. "github.com/chrislusf/seaweedfs/weed/glog"
  8. "github.com/chrislusf/seaweedfs/weed/pb"
  9. "github.com/hashicorp/raft"
  10. boltdb "github.com/hashicorp/raft-boltdb"
  11. "google.golang.org/grpc"
  12. "math/rand"
  13. "os"
  14. "path/filepath"
  15. "sort"
  16. "strings"
  17. "time"
  18. )
  19. func getPeerIdx(self pb.ServerAddress, mapPeers map[string]pb.ServerAddress) int {
  20. peers := make([]pb.ServerAddress, 0, len(mapPeers))
  21. for _, peer := range mapPeers {
  22. peers = append(peers, peer)
  23. }
  24. sort.Slice(peers, func(i, j int) bool {
  25. return strings.Compare(string(peers[i]), string(peers[j])) < 0
  26. })
  27. for i, peer := range peers {
  28. if string(peer) == string(self) {
  29. return i
  30. }
  31. }
  32. return -1
  33. }
  34. func (s *RaftServer) AddPeersConfiguration() (cfg raft.Configuration) {
  35. for _, peer := range s.peers {
  36. cfg.Servers = append(cfg.Servers, raft.Server{
  37. Suffrage: raft.Voter,
  38. ID: raft.ServerID(peer.String()),
  39. Address: raft.ServerAddress(peer.ToGrpcAddress()),
  40. })
  41. }
  42. return cfg
  43. }
  44. func (s *RaftServer) UpdatePeers() {
  45. for {
  46. select {
  47. case isLeader := <-s.RaftHashicorp.LeaderCh():
  48. if isLeader {
  49. peerLeader := string(s.serverAddr)
  50. existsPeerName := make(map[string]bool)
  51. for _, server := range s.RaftHashicorp.GetConfiguration().Configuration().Servers {
  52. if string(server.ID) == peerLeader {
  53. continue
  54. }
  55. existsPeerName[string(server.ID)] = true
  56. }
  57. for _, peer := range s.peers {
  58. if peer.String() == peerLeader || existsPeerName[peer.String()] {
  59. continue
  60. }
  61. glog.V(0).Infof("adding new peer: %s", peer.String())
  62. s.RaftHashicorp.AddVoter(
  63. raft.ServerID(peer.String()), raft.ServerAddress(peer.ToGrpcAddress()), 0, 0)
  64. }
  65. for peer, _ := range existsPeerName {
  66. if _, found := s.peers[peer]; !found {
  67. glog.V(0).Infof("removing old peer: %s", peer)
  68. s.RaftHashicorp.RemoveServer(raft.ServerID(peer), 0, 0)
  69. }
  70. }
  71. if _, found := s.peers[peerLeader]; !found {
  72. glog.V(0).Infof("removing old leader peer: %s", peerLeader)
  73. s.RaftHashicorp.RemoveServer(raft.ServerID(peerLeader), 0, 0)
  74. }
  75. }
  76. break
  77. }
  78. }
  79. }
  80. func NewHashicorpRaftServer(option *RaftServerOption) (*RaftServer, error) {
  81. s := &RaftServer{
  82. peers: option.Peers,
  83. serverAddr: option.ServerAddr,
  84. dataDir: option.DataDir,
  85. topo: option.Topo,
  86. }
  87. c := raft.DefaultConfig()
  88. c.LocalID = raft.ServerID(s.serverAddr) // TODO maybee the IP:port address will change
  89. c.NoSnapshotRestoreOnStart = option.RaftResumeState
  90. c.HeartbeatTimeout = time.Duration(float64(option.HeartbeatInterval) * (rand.Float64()*0.25 + 1))
  91. c.ElectionTimeout = option.ElectionTimeout
  92. if c.LeaderLeaseTimeout > c.HeartbeatTimeout {
  93. c.LeaderLeaseTimeout = c.HeartbeatTimeout
  94. }
  95. if glog.V(4) {
  96. c.LogLevel = "Debug"
  97. } else if glog.V(2) {
  98. c.LogLevel = "Info"
  99. } else if glog.V(1) {
  100. c.LogLevel = "Warn"
  101. } else if glog.V(0) {
  102. c.LogLevel = "Error"
  103. }
  104. baseDir := s.dataDir
  105. ldb, err := boltdb.NewBoltStore(filepath.Join(baseDir, "logs.dat"))
  106. if err != nil {
  107. return nil, fmt.Errorf(`boltdb.NewBoltStore(%q): %v`, filepath.Join(baseDir, "logs.dat"), err)
  108. }
  109. sdb, err := boltdb.NewBoltStore(filepath.Join(baseDir, "stable.dat"))
  110. if err != nil {
  111. return nil, fmt.Errorf(`boltdb.NewBoltStore(%q): %v`, filepath.Join(baseDir, "stable.dat"), err)
  112. }
  113. fss, err := raft.NewFileSnapshotStore(baseDir, 3, os.Stderr)
  114. if err != nil {
  115. return nil, fmt.Errorf(`raft.NewFileSnapshotStore(%q, ...): %v`, baseDir, err)
  116. }
  117. s.TransportManager = transport.New(raft.ServerAddress(s.serverAddr), []grpc.DialOption{option.GrpcDialOption})
  118. stateMachine := StateMachine{topo: option.Topo}
  119. s.RaftHashicorp, err = raft.NewRaft(c, &stateMachine, ldb, sdb, fss, s.TransportManager.Transport())
  120. if err != nil {
  121. return nil, fmt.Errorf("raft.NewRaft: %v", err)
  122. }
  123. if option.RaftBootstrap || len(s.RaftHashicorp.GetConfiguration().Configuration().Servers) == 0 {
  124. cfg := s.AddPeersConfiguration()
  125. // Need to get lock, in case all servers do this at the same time.
  126. peerIdx := getPeerIdx(s.serverAddr, s.peers)
  127. timeSpeep := time.Duration(float64(c.LeaderLeaseTimeout) * (rand.Float64()*0.25 + 1) * float64(peerIdx))
  128. glog.V(0).Infof("Bootstrapping idx: %d sleep: %v new cluster: %+v", peerIdx, timeSpeep, cfg)
  129. time.Sleep(timeSpeep)
  130. f := s.RaftHashicorp.BootstrapCluster(cfg)
  131. if err := f.Error(); err != nil {
  132. return nil, fmt.Errorf("raft.Raft.BootstrapCluster: %v", err)
  133. }
  134. } else {
  135. go s.UpdatePeers()
  136. }
  137. ticker := time.NewTicker(c.HeartbeatTimeout * 10)
  138. if glog.V(4) {
  139. go func() {
  140. for {
  141. select {
  142. case <-ticker.C:
  143. cfuture := s.RaftHashicorp.GetConfiguration()
  144. if err = cfuture.Error(); err != nil {
  145. glog.Fatalf("error getting config: %s", err)
  146. }
  147. configuration := cfuture.Configuration()
  148. glog.V(4).Infof("Showing peers known by %s:\n%+v", s.RaftHashicorp.String(), configuration.Servers)
  149. }
  150. }
  151. }()
  152. }
  153. return s, nil
  154. }