node.go 21 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869
  1. package node
  2. import (
  3. "crypto/tls"
  4. "encoding/json"
  5. "io/ioutil"
  6. "net"
  7. "os"
  8. "path/filepath"
  9. "reflect"
  10. "sort"
  11. "sync"
  12. "time"
  13. "github.com/Sirupsen/logrus"
  14. "github.com/boltdb/bolt"
  15. "github.com/docker/swarmkit/agent"
  16. "github.com/docker/swarmkit/agent/exec"
  17. "github.com/docker/swarmkit/api"
  18. "github.com/docker/swarmkit/ca"
  19. "github.com/docker/swarmkit/connectionbroker"
  20. "github.com/docker/swarmkit/ioutils"
  21. "github.com/docker/swarmkit/log"
  22. "github.com/docker/swarmkit/manager"
  23. "github.com/docker/swarmkit/manager/encryption"
  24. "github.com/docker/swarmkit/remotes"
  25. "github.com/docker/swarmkit/xnet"
  26. "github.com/pkg/errors"
  27. "golang.org/x/net/context"
  28. "google.golang.org/grpc"
  29. "google.golang.org/grpc/credentials"
  30. )
  31. const stateFilename = "state.json"
  32. var (
  33. errNodeStarted = errors.New("node: already started")
  34. errNodeNotStarted = errors.New("node: not started")
  35. certDirectory = "certificates"
  36. // ErrInvalidUnlockKey is returned when we can't decrypt the TLS certificate
  37. ErrInvalidUnlockKey = errors.New("node is locked, and needs a valid unlock key")
  38. )
  39. // Config provides values for a Node.
  40. type Config struct {
  41. // Hostname is the name of host for agent instance.
  42. Hostname string
  43. // JoinAddr specifies node that should be used for the initial connection to
  44. // other manager in cluster. This should be only one address and optional,
  45. // the actual remotes come from the stored state.
  46. JoinAddr string
  47. // StateDir specifies the directory the node uses to keep the state of the
  48. // remote managers and certificates.
  49. StateDir string
  50. // JoinToken is the token to be used on the first certificate request.
  51. JoinToken string
  52. // ExternalCAs is a list of CAs to which a manager node
  53. // will make certificate signing requests for node certificates.
  54. ExternalCAs []*api.ExternalCA
  55. // ForceNewCluster creates a new cluster from current raft state.
  56. ForceNewCluster bool
  57. // ListenControlAPI specifies address the control API should listen on.
  58. ListenControlAPI string
  59. // ListenRemoteAPI specifies the address for the remote API that agents
  60. // and raft members connect to.
  61. ListenRemoteAPI string
  62. // AdvertiseRemoteAPI specifies the address that should be advertised
  63. // for connections to the remote API (including the raft service).
  64. AdvertiseRemoteAPI string
  65. // Executor specifies the executor to use for the agent.
  66. Executor exec.Executor
  67. // ElectionTick defines the amount of ticks needed without
  68. // leader to trigger a new election
  69. ElectionTick uint32
  70. // HeartbeatTick defines the amount of ticks between each
  71. // heartbeat sent to other members for health-check purposes
  72. HeartbeatTick uint32
  73. // AutoLockManagers determines whether or not an unlock key will be generated
  74. // when bootstrapping a new cluster for the first time
  75. AutoLockManagers bool
  76. // UnlockKey is the key to unlock a node - used for decrypting at rest. This
  77. // only applies to nodes that have already joined a cluster.
  78. UnlockKey []byte
  79. // Availability allows a user to control the current scheduling status of a node
  80. Availability api.NodeSpec_Availability
  81. }
  82. // Node implements the primary node functionality for a member of a swarm
  83. // cluster. Node handles workloads and may also run as a manager.
  84. type Node struct {
  85. sync.RWMutex
  86. config *Config
  87. remotes *persistentRemotes
  88. connBroker *connectionbroker.Broker
  89. role string
  90. roleCond *sync.Cond
  91. conn *grpc.ClientConn
  92. connCond *sync.Cond
  93. nodeID string
  94. started chan struct{}
  95. startOnce sync.Once
  96. stopped chan struct{}
  97. stopOnce sync.Once
  98. ready chan struct{} // closed when agent has completed registration and manager(if enabled) is ready to receive control requests
  99. closed chan struct{}
  100. err error
  101. agent *agent.Agent
  102. manager *manager.Manager
  103. notifyNodeChange chan *api.Node // used to send role updates from the dispatcher api on promotion/demotion
  104. unlockKey []byte
  105. }
  106. // RemoteAPIAddr returns address on which remote manager api listens.
  107. // Returns nil if node is not manager.
  108. func (n *Node) RemoteAPIAddr() (string, error) {
  109. n.RLock()
  110. defer n.RUnlock()
  111. if n.manager == nil {
  112. return "", errors.Errorf("node is not manager")
  113. }
  114. addr := n.manager.Addr()
  115. if addr == "" {
  116. return "", errors.Errorf("manager addr is not set")
  117. }
  118. return addr, nil
  119. }
  120. // New returns new Node instance.
  121. func New(c *Config) (*Node, error) {
  122. if err := os.MkdirAll(c.StateDir, 0700); err != nil {
  123. return nil, err
  124. }
  125. stateFile := filepath.Join(c.StateDir, stateFilename)
  126. dt, err := ioutil.ReadFile(stateFile)
  127. var p []api.Peer
  128. if err != nil && !os.IsNotExist(err) {
  129. return nil, err
  130. }
  131. if err == nil {
  132. if err := json.Unmarshal(dt, &p); err != nil {
  133. return nil, err
  134. }
  135. }
  136. n := &Node{
  137. remotes: newPersistentRemotes(stateFile, p...),
  138. role: ca.WorkerRole,
  139. config: c,
  140. started: make(chan struct{}),
  141. stopped: make(chan struct{}),
  142. closed: make(chan struct{}),
  143. ready: make(chan struct{}),
  144. notifyNodeChange: make(chan *api.Node, 1),
  145. unlockKey: c.UnlockKey,
  146. }
  147. if n.config.JoinAddr != "" || n.config.ForceNewCluster {
  148. n.remotes = newPersistentRemotes(filepath.Join(n.config.StateDir, stateFilename))
  149. if n.config.JoinAddr != "" {
  150. n.remotes.Observe(api.Peer{Addr: n.config.JoinAddr}, remotes.DefaultObservationWeight)
  151. }
  152. }
  153. n.connBroker = connectionbroker.New(n.remotes)
  154. n.roleCond = sync.NewCond(n.RLocker())
  155. n.connCond = sync.NewCond(n.RLocker())
  156. return n, nil
  157. }
  158. // Start starts a node instance.
  159. func (n *Node) Start(ctx context.Context) error {
  160. err := errNodeStarted
  161. n.startOnce.Do(func() {
  162. close(n.started)
  163. go n.run(ctx)
  164. err = nil // clear error above, only once.
  165. })
  166. return err
  167. }
  168. func (n *Node) run(ctx context.Context) (err error) {
  169. defer func() {
  170. n.err = err
  171. close(n.closed)
  172. }()
  173. ctx, cancel := context.WithCancel(ctx)
  174. defer cancel()
  175. ctx = log.WithModule(ctx, "node")
  176. go func() {
  177. select {
  178. case <-ctx.Done():
  179. case <-n.stopped:
  180. cancel()
  181. }
  182. }()
  183. securityConfig, err := n.loadSecurityConfig(ctx)
  184. if err != nil {
  185. return err
  186. }
  187. taskDBPath := filepath.Join(n.config.StateDir, "worker/tasks.db")
  188. if err := os.MkdirAll(filepath.Dir(taskDBPath), 0777); err != nil {
  189. return err
  190. }
  191. db, err := bolt.Open(taskDBPath, 0666, nil)
  192. if err != nil {
  193. return err
  194. }
  195. defer db.Close()
  196. forceCertRenewal := make(chan struct{})
  197. renewCert := func() {
  198. select {
  199. case forceCertRenewal <- struct{}{}:
  200. case <-ctx.Done():
  201. }
  202. }
  203. go func() {
  204. for {
  205. select {
  206. case <-ctx.Done():
  207. return
  208. case node := <-n.notifyNodeChange:
  209. // If the server is sending us a ForceRenewal State, renew
  210. if node.Certificate.Status.State == api.IssuanceStateRotate {
  211. renewCert()
  212. continue
  213. }
  214. n.Lock()
  215. // If we got a role change, renew
  216. lastRole := n.role
  217. role := ca.WorkerRole
  218. if node.Role == api.NodeRoleManager {
  219. role = ca.ManagerRole
  220. }
  221. if lastRole == role {
  222. n.Unlock()
  223. continue
  224. }
  225. n.Unlock()
  226. renewCert()
  227. }
  228. }
  229. }()
  230. updates := ca.RenewTLSConfig(ctx, securityConfig, n.connBroker, forceCertRenewal)
  231. go func() {
  232. for {
  233. select {
  234. case certUpdate := <-updates:
  235. if certUpdate.Err != nil {
  236. logrus.Warnf("error renewing TLS certificate: %v", certUpdate.Err)
  237. continue
  238. }
  239. n.Lock()
  240. n.role = certUpdate.Role
  241. n.roleCond.Broadcast()
  242. n.Unlock()
  243. case <-ctx.Done():
  244. return
  245. }
  246. }
  247. }()
  248. role := n.role
  249. managerReady := make(chan struct{})
  250. agentReady := make(chan struct{})
  251. var managerErr error
  252. var agentErr error
  253. var wg sync.WaitGroup
  254. wg.Add(2)
  255. go func() {
  256. managerErr = n.superviseManager(ctx, securityConfig, managerReady) // store err and loop
  257. wg.Done()
  258. cancel()
  259. }()
  260. go func() {
  261. agentErr = n.runAgent(ctx, db, securityConfig.ClientTLSCreds, agentReady)
  262. wg.Done()
  263. cancel()
  264. }()
  265. go func() {
  266. <-agentReady
  267. if role == ca.ManagerRole {
  268. workerRole := make(chan struct{})
  269. waitRoleCtx, waitRoleCancel := context.WithCancel(ctx)
  270. go func() {
  271. if n.waitRole(waitRoleCtx, ca.WorkerRole) == nil {
  272. close(workerRole)
  273. }
  274. }()
  275. select {
  276. case <-managerReady:
  277. case <-workerRole:
  278. }
  279. waitRoleCancel()
  280. }
  281. close(n.ready)
  282. }()
  283. wg.Wait()
  284. if managerErr != nil && managerErr != context.Canceled {
  285. return managerErr
  286. }
  287. if agentErr != nil && agentErr != context.Canceled {
  288. return agentErr
  289. }
  290. return err
  291. }
  292. // Stop stops node execution
  293. func (n *Node) Stop(ctx context.Context) error {
  294. select {
  295. case <-n.started:
  296. default:
  297. return errNodeNotStarted
  298. }
  299. // ask agent to clean up assignments
  300. n.Lock()
  301. if n.agent != nil {
  302. if err := n.agent.Leave(ctx); err != nil {
  303. log.G(ctx).WithError(err).Error("agent failed to clean up assignments")
  304. }
  305. }
  306. n.Unlock()
  307. n.stopOnce.Do(func() {
  308. close(n.stopped)
  309. })
  310. select {
  311. case <-n.closed:
  312. return nil
  313. case <-ctx.Done():
  314. return ctx.Err()
  315. }
  316. }
  317. // Err returns the error that caused the node to shutdown or nil. Err blocks
  318. // until the node has fully shut down.
  319. func (n *Node) Err(ctx context.Context) error {
  320. select {
  321. case <-n.closed:
  322. return n.err
  323. case <-ctx.Done():
  324. return ctx.Err()
  325. }
  326. }
  327. func (n *Node) runAgent(ctx context.Context, db *bolt.DB, creds credentials.TransportCredentials, ready chan<- struct{}) error {
  328. waitCtx, waitCancel := context.WithCancel(ctx)
  329. remotesCh := n.remotes.WaitSelect(ctx)
  330. controlCh := n.ListenControlSocket(waitCtx)
  331. waitPeer:
  332. for {
  333. select {
  334. case <-ctx.Done():
  335. break waitPeer
  336. case <-remotesCh:
  337. break waitPeer
  338. case conn := <-controlCh:
  339. if conn != nil {
  340. break waitPeer
  341. }
  342. }
  343. }
  344. waitCancel()
  345. select {
  346. case <-ctx.Done():
  347. return ctx.Err()
  348. default:
  349. }
  350. a, err := agent.New(&agent.Config{
  351. Hostname: n.config.Hostname,
  352. ConnBroker: n.connBroker,
  353. Executor: n.config.Executor,
  354. DB: db,
  355. NotifyNodeChange: n.notifyNodeChange,
  356. Credentials: creds,
  357. })
  358. if err != nil {
  359. return err
  360. }
  361. if err := a.Start(ctx); err != nil {
  362. return err
  363. }
  364. n.Lock()
  365. n.agent = a
  366. n.Unlock()
  367. defer func() {
  368. n.Lock()
  369. n.agent = nil
  370. n.Unlock()
  371. }()
  372. go func() {
  373. <-a.Ready()
  374. close(ready)
  375. }()
  376. // todo: manually call stop on context cancellation?
  377. return a.Err(context.Background())
  378. }
  379. // Ready returns a channel that is closed after node's initialization has
  380. // completes for the first time.
  381. func (n *Node) Ready() <-chan struct{} {
  382. return n.ready
  383. }
  384. func (n *Node) setControlSocket(conn *grpc.ClientConn) {
  385. n.Lock()
  386. if n.conn != nil {
  387. n.conn.Close()
  388. }
  389. n.conn = conn
  390. n.connBroker.SetLocalConn(conn)
  391. n.connCond.Broadcast()
  392. n.Unlock()
  393. }
  394. // ListenControlSocket listens changes of a connection for managing the
  395. // cluster control api
  396. func (n *Node) ListenControlSocket(ctx context.Context) <-chan *grpc.ClientConn {
  397. c := make(chan *grpc.ClientConn, 1)
  398. n.RLock()
  399. conn := n.conn
  400. c <- conn
  401. done := make(chan struct{})
  402. go func() {
  403. select {
  404. case <-ctx.Done():
  405. n.connCond.Broadcast()
  406. case <-done:
  407. }
  408. }()
  409. go func() {
  410. defer close(c)
  411. defer close(done)
  412. defer n.RUnlock()
  413. for {
  414. select {
  415. case <-ctx.Done():
  416. return
  417. default:
  418. }
  419. if conn == n.conn {
  420. n.connCond.Wait()
  421. continue
  422. }
  423. conn = n.conn
  424. select {
  425. case c <- conn:
  426. case <-ctx.Done():
  427. return
  428. }
  429. }
  430. }()
  431. return c
  432. }
  433. // NodeID returns current node's ID. May be empty if not set.
  434. func (n *Node) NodeID() string {
  435. n.RLock()
  436. defer n.RUnlock()
  437. return n.nodeID
  438. }
  439. // Manager returns manager instance started by node. May be nil.
  440. func (n *Node) Manager() *manager.Manager {
  441. n.RLock()
  442. defer n.RUnlock()
  443. return n.manager
  444. }
  445. // Agent returns agent instance started by node. May be nil.
  446. func (n *Node) Agent() *agent.Agent {
  447. n.RLock()
  448. defer n.RUnlock()
  449. return n.agent
  450. }
  451. // Remotes returns a list of known peers known to node.
  452. func (n *Node) Remotes() []api.Peer {
  453. weights := n.remotes.Weights()
  454. remotes := make([]api.Peer, 0, len(weights))
  455. for p := range weights {
  456. remotes = append(remotes, p)
  457. }
  458. return remotes
  459. }
  460. func (n *Node) loadSecurityConfig(ctx context.Context) (*ca.SecurityConfig, error) {
  461. paths := ca.NewConfigPaths(filepath.Join(n.config.StateDir, certDirectory))
  462. var securityConfig *ca.SecurityConfig
  463. krw := ca.NewKeyReadWriter(paths.Node, n.unlockKey, &manager.RaftDEKData{})
  464. if err := krw.Migrate(); err != nil {
  465. return nil, err
  466. }
  467. // Check if we already have a valid certificates on disk.
  468. rootCA, err := ca.GetLocalRootCA(paths.RootCA)
  469. if err != nil && err != ca.ErrNoLocalRootCA {
  470. return nil, err
  471. }
  472. if err == nil {
  473. securityConfig, err = ca.LoadSecurityConfig(ctx, rootCA, krw)
  474. if err != nil {
  475. _, isInvalidKEK := errors.Cause(err).(ca.ErrInvalidKEK)
  476. if isInvalidKEK {
  477. return nil, ErrInvalidUnlockKey
  478. } else if !os.IsNotExist(err) {
  479. return nil, errors.Wrapf(err, "error while loading TLS certificate in %s", paths.Node.Cert)
  480. }
  481. }
  482. }
  483. if securityConfig == nil {
  484. if n.config.JoinAddr == "" {
  485. // if we're not joining a cluster, bootstrap a new one - and we have to set the unlock key
  486. n.unlockKey = nil
  487. if n.config.AutoLockManagers {
  488. n.unlockKey = encryption.GenerateSecretKey()
  489. }
  490. krw = ca.NewKeyReadWriter(paths.Node, n.unlockKey, &manager.RaftDEKData{})
  491. rootCA, err = ca.CreateRootCA(ca.DefaultRootCN, paths.RootCA)
  492. if err != nil {
  493. return nil, err
  494. }
  495. log.G(ctx).Debug("generated CA key and certificate")
  496. } else if err == ca.ErrNoLocalRootCA { // from previous error loading the root CA from disk
  497. rootCA, err = ca.DownloadRootCA(ctx, paths.RootCA, n.config.JoinToken, n.connBroker)
  498. if err != nil {
  499. return nil, err
  500. }
  501. log.G(ctx).Debug("downloaded CA certificate")
  502. }
  503. // Obtain new certs and setup TLS certificates renewal for this node:
  504. // - If certificates weren't present on disk, we call CreateSecurityConfig, which blocks
  505. // until a valid certificate has been issued.
  506. // - We wait for CreateSecurityConfig to finish since we need a certificate to operate.
  507. // Attempt to load certificate from disk
  508. securityConfig, err = ca.LoadSecurityConfig(ctx, rootCA, krw)
  509. if err == nil {
  510. log.G(ctx).WithFields(logrus.Fields{
  511. "node.id": securityConfig.ClientTLSCreds.NodeID(),
  512. }).Debugf("loaded TLS certificate")
  513. } else {
  514. if _, ok := errors.Cause(err).(ca.ErrInvalidKEK); ok {
  515. return nil, ErrInvalidUnlockKey
  516. }
  517. log.G(ctx).WithError(err).Debugf("no node credentials found in: %s", krw.Target())
  518. securityConfig, err = rootCA.CreateSecurityConfig(ctx, krw, ca.CertificateRequestConfig{
  519. Token: n.config.JoinToken,
  520. Availability: n.config.Availability,
  521. ConnBroker: n.connBroker,
  522. })
  523. if err != nil {
  524. return nil, err
  525. }
  526. }
  527. }
  528. n.Lock()
  529. n.role = securityConfig.ClientTLSCreds.Role()
  530. n.nodeID = securityConfig.ClientTLSCreds.NodeID()
  531. n.roleCond.Broadcast()
  532. n.Unlock()
  533. return securityConfig, nil
  534. }
  535. func (n *Node) initManagerConnection(ctx context.Context, ready chan<- struct{}) error {
  536. opts := []grpc.DialOption{}
  537. insecureCreds := credentials.NewTLS(&tls.Config{InsecureSkipVerify: true})
  538. opts = append(opts, grpc.WithTransportCredentials(insecureCreds))
  539. addr := n.config.ListenControlAPI
  540. opts = append(opts, grpc.WithDialer(
  541. func(addr string, timeout time.Duration) (net.Conn, error) {
  542. return xnet.DialTimeoutLocal(addr, timeout)
  543. }))
  544. conn, err := grpc.Dial(addr, opts...)
  545. if err != nil {
  546. return err
  547. }
  548. client := api.NewHealthClient(conn)
  549. for {
  550. resp, err := client.Check(ctx, &api.HealthCheckRequest{Service: "ControlAPI"})
  551. if err != nil {
  552. return err
  553. }
  554. if resp.Status == api.HealthCheckResponse_SERVING {
  555. break
  556. }
  557. time.Sleep(500 * time.Millisecond)
  558. }
  559. n.setControlSocket(conn)
  560. if ready != nil {
  561. close(ready)
  562. }
  563. return nil
  564. }
  565. func (n *Node) waitRole(ctx context.Context, role string) error {
  566. n.roleCond.L.Lock()
  567. if role == n.role {
  568. n.roleCond.L.Unlock()
  569. return nil
  570. }
  571. finishCh := make(chan struct{})
  572. defer close(finishCh)
  573. go func() {
  574. select {
  575. case <-finishCh:
  576. case <-ctx.Done():
  577. // call broadcast to shutdown this function
  578. n.roleCond.Broadcast()
  579. }
  580. }()
  581. defer n.roleCond.L.Unlock()
  582. for role != n.role {
  583. n.roleCond.Wait()
  584. select {
  585. case <-ctx.Done():
  586. return ctx.Err()
  587. default:
  588. }
  589. }
  590. return nil
  591. }
  592. func (n *Node) runManager(ctx context.Context, securityConfig *ca.SecurityConfig, ready chan struct{}, workerRole <-chan struct{}) error {
  593. remoteAddr, _ := n.remotes.Select(n.NodeID())
  594. m, err := manager.New(&manager.Config{
  595. ForceNewCluster: n.config.ForceNewCluster,
  596. RemoteAPI: manager.RemoteAddrs{
  597. ListenAddr: n.config.ListenRemoteAPI,
  598. AdvertiseAddr: n.config.AdvertiseRemoteAPI,
  599. },
  600. ControlAPI: n.config.ListenControlAPI,
  601. SecurityConfig: securityConfig,
  602. ExternalCAs: n.config.ExternalCAs,
  603. JoinRaft: remoteAddr.Addr,
  604. StateDir: n.config.StateDir,
  605. HeartbeatTick: n.config.HeartbeatTick,
  606. ElectionTick: n.config.ElectionTick,
  607. AutoLockManagers: n.config.AutoLockManagers,
  608. UnlockKey: n.unlockKey,
  609. Availability: n.config.Availability,
  610. })
  611. if err != nil {
  612. return err
  613. }
  614. done := make(chan struct{})
  615. var runErr error
  616. go func() {
  617. if err := m.Run(context.Background()); err != nil {
  618. runErr = err
  619. }
  620. close(done)
  621. }()
  622. var clearData bool
  623. defer func() {
  624. n.Lock()
  625. n.manager = nil
  626. n.Unlock()
  627. m.Stop(ctx, clearData)
  628. <-done
  629. n.setControlSocket(nil)
  630. }()
  631. n.Lock()
  632. n.manager = m
  633. n.Unlock()
  634. connCtx, connCancel := context.WithCancel(ctx)
  635. defer connCancel()
  636. go n.initManagerConnection(connCtx, ready)
  637. // wait for manager stop or for role change
  638. select {
  639. case <-done:
  640. return runErr
  641. case <-workerRole:
  642. log.G(ctx).Info("role changed to worker, stopping manager")
  643. clearData = true
  644. case <-m.RemovedFromRaft():
  645. log.G(ctx).Info("manager removed from raft cluster, stopping manager")
  646. clearData = true
  647. case <-ctx.Done():
  648. return ctx.Err()
  649. }
  650. return nil
  651. }
  652. func (n *Node) superviseManager(ctx context.Context, securityConfig *ca.SecurityConfig, ready chan struct{}) error {
  653. for {
  654. if err := n.waitRole(ctx, ca.ManagerRole); err != nil {
  655. return err
  656. }
  657. workerRole := make(chan struct{})
  658. waitRoleCtx, waitRoleCancel := context.WithCancel(ctx)
  659. go func() {
  660. if n.waitRole(waitRoleCtx, ca.WorkerRole) == nil {
  661. close(workerRole)
  662. }
  663. }()
  664. if err := n.runManager(ctx, securityConfig, ready, workerRole); err != nil {
  665. waitRoleCancel()
  666. return errors.Wrap(err, "manager stopped")
  667. }
  668. // If the manager stopped running and our role is still
  669. // "manager", it's possible that the manager was demoted and
  670. // the agent hasn't realized this yet. We should wait for the
  671. // role to change instead of restarting the manager immediately.
  672. timer := time.NewTimer(16 * time.Second)
  673. select {
  674. case <-timer.C:
  675. log.G(ctx).Warn("failed to get worker role after manager stop, restarting manager")
  676. case <-workerRole:
  677. case <-ctx.Done():
  678. timer.Stop()
  679. waitRoleCancel()
  680. return ctx.Err()
  681. }
  682. timer.Stop()
  683. waitRoleCancel()
  684. ready = nil
  685. }
  686. }
  687. type persistentRemotes struct {
  688. sync.RWMutex
  689. c *sync.Cond
  690. remotes.Remotes
  691. storePath string
  692. lastSavedState []api.Peer
  693. }
  694. func newPersistentRemotes(f string, peers ...api.Peer) *persistentRemotes {
  695. pr := &persistentRemotes{
  696. storePath: f,
  697. Remotes: remotes.NewRemotes(peers...),
  698. }
  699. pr.c = sync.NewCond(pr.RLocker())
  700. return pr
  701. }
  702. func (s *persistentRemotes) Observe(peer api.Peer, weight int) {
  703. s.Lock()
  704. defer s.Unlock()
  705. s.Remotes.Observe(peer, weight)
  706. s.c.Broadcast()
  707. if err := s.save(); err != nil {
  708. logrus.Errorf("error writing cluster state file: %v", err)
  709. return
  710. }
  711. return
  712. }
  713. func (s *persistentRemotes) Remove(peers ...api.Peer) {
  714. s.Lock()
  715. defer s.Unlock()
  716. s.Remotes.Remove(peers...)
  717. if err := s.save(); err != nil {
  718. logrus.Errorf("error writing cluster state file: %v", err)
  719. return
  720. }
  721. return
  722. }
  723. func (s *persistentRemotes) save() error {
  724. weights := s.Weights()
  725. remotes := make([]api.Peer, 0, len(weights))
  726. for r := range weights {
  727. remotes = append(remotes, r)
  728. }
  729. sort.Sort(sortablePeers(remotes))
  730. if reflect.DeepEqual(remotes, s.lastSavedState) {
  731. return nil
  732. }
  733. dt, err := json.Marshal(remotes)
  734. if err != nil {
  735. return err
  736. }
  737. s.lastSavedState = remotes
  738. return ioutils.AtomicWriteFile(s.storePath, dt, 0600)
  739. }
  740. // WaitSelect waits until at least one remote becomes available and then selects one.
  741. func (s *persistentRemotes) WaitSelect(ctx context.Context) <-chan api.Peer {
  742. c := make(chan api.Peer, 1)
  743. s.RLock()
  744. done := make(chan struct{})
  745. go func() {
  746. select {
  747. case <-ctx.Done():
  748. s.c.Broadcast()
  749. case <-done:
  750. }
  751. }()
  752. go func() {
  753. defer s.RUnlock()
  754. defer close(c)
  755. defer close(done)
  756. for {
  757. if ctx.Err() != nil {
  758. return
  759. }
  760. p, err := s.Select()
  761. if err == nil {
  762. c <- p
  763. return
  764. }
  765. s.c.Wait()
  766. }
  767. }()
  768. return c
  769. }
  770. // sortablePeers is a sort wrapper for []api.Peer
  771. type sortablePeers []api.Peer
  772. func (sp sortablePeers) Less(i, j int) bool { return sp[i].NodeID < sp[j].NodeID }
  773. func (sp sortablePeers) Len() int { return len(sp) }
  774. func (sp sortablePeers) Swap(i, j int) { sp[i], sp[j] = sp[j], sp[i] }