node.go 22 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877878879880881882883884885886887888889890891892893894895896897898899900901902903904905906907908
  1. package node
  2. import (
  3. "crypto/tls"
  4. "encoding/json"
  5. "io/ioutil"
  6. "net"
  7. "os"
  8. "path/filepath"
  9. "reflect"
  10. "sort"
  11. "sync"
  12. "time"
  13. "github.com/Sirupsen/logrus"
  14. "github.com/boltdb/bolt"
  15. "github.com/docker/docker/pkg/plugingetter"
  16. "github.com/docker/swarmkit/agent"
  17. "github.com/docker/swarmkit/agent/exec"
  18. "github.com/docker/swarmkit/api"
  19. "github.com/docker/swarmkit/ca"
  20. "github.com/docker/swarmkit/connectionbroker"
  21. "github.com/docker/swarmkit/ioutils"
  22. "github.com/docker/swarmkit/log"
  23. "github.com/docker/swarmkit/manager"
  24. "github.com/docker/swarmkit/manager/encryption"
  25. "github.com/docker/swarmkit/remotes"
  26. "github.com/docker/swarmkit/xnet"
  27. "github.com/pkg/errors"
  28. "golang.org/x/net/context"
  29. "google.golang.org/grpc"
  30. "google.golang.org/grpc/credentials"
  31. )
  32. const stateFilename = "state.json"
  33. var (
  34. errNodeStarted = errors.New("node: already started")
  35. errNodeNotStarted = errors.New("node: not started")
  36. certDirectory = "certificates"
  37. // ErrInvalidUnlockKey is returned when we can't decrypt the TLS certificate
  38. ErrInvalidUnlockKey = errors.New("node is locked, and needs a valid unlock key")
  39. )
  40. // Config provides values for a Node.
  41. type Config struct {
  42. // Hostname is the name of host for agent instance.
  43. Hostname string
  44. // JoinAddr specifies node that should be used for the initial connection to
  45. // other manager in cluster. This should be only one address and optional,
  46. // the actual remotes come from the stored state.
  47. JoinAddr string
  48. // StateDir specifies the directory the node uses to keep the state of the
  49. // remote managers and certificates.
  50. StateDir string
  51. // JoinToken is the token to be used on the first certificate request.
  52. JoinToken string
  53. // ExternalCAs is a list of CAs to which a manager node
  54. // will make certificate signing requests for node certificates.
  55. ExternalCAs []*api.ExternalCA
  56. // ForceNewCluster creates a new cluster from current raft state.
  57. ForceNewCluster bool
  58. // ListenControlAPI specifies address the control API should listen on.
  59. ListenControlAPI string
  60. // ListenRemoteAPI specifies the address for the remote API that agents
  61. // and raft members connect to.
  62. ListenRemoteAPI string
  63. // AdvertiseRemoteAPI specifies the address that should be advertised
  64. // for connections to the remote API (including the raft service).
  65. AdvertiseRemoteAPI string
  66. // Executor specifies the executor to use for the agent.
  67. Executor exec.Executor
  68. // ElectionTick defines the amount of ticks needed without
  69. // leader to trigger a new election
  70. ElectionTick uint32
  71. // HeartbeatTick defines the amount of ticks between each
  72. // heartbeat sent to other members for health-check purposes
  73. HeartbeatTick uint32
  74. // AutoLockManagers determines whether or not an unlock key will be generated
  75. // when bootstrapping a new cluster for the first time
  76. AutoLockManagers bool
  77. // UnlockKey is the key to unlock a node - used for decrypting at rest. This
  78. // only applies to nodes that have already joined a cluster.
  79. UnlockKey []byte
  80. // Availability allows a user to control the current scheduling status of a node
  81. Availability api.NodeSpec_Availability
  82. // PluginGetter provides access to docker's plugin inventory.
  83. PluginGetter plugingetter.PluginGetter
  84. }
  85. // Node implements the primary node functionality for a member of a swarm
  86. // cluster. Node handles workloads and may also run as a manager.
  87. type Node struct {
  88. sync.RWMutex
  89. config *Config
  90. remotes *persistentRemotes
  91. connBroker *connectionbroker.Broker
  92. role string
  93. roleCond *sync.Cond
  94. conn *grpc.ClientConn
  95. connCond *sync.Cond
  96. nodeID string
  97. started chan struct{}
  98. startOnce sync.Once
  99. stopped chan struct{}
  100. stopOnce sync.Once
  101. ready chan struct{} // closed when agent has completed registration and manager(if enabled) is ready to receive control requests
  102. closed chan struct{}
  103. err error
  104. agent *agent.Agent
  105. manager *manager.Manager
  106. notifyNodeChange chan *api.Node // used to send role updates from the dispatcher api on promotion/demotion
  107. unlockKey []byte
  108. }
  109. // RemoteAPIAddr returns address on which remote manager api listens.
  110. // Returns nil if node is not manager.
  111. func (n *Node) RemoteAPIAddr() (string, error) {
  112. n.RLock()
  113. defer n.RUnlock()
  114. if n.manager == nil {
  115. return "", errors.New("manager is not running")
  116. }
  117. addr := n.manager.Addr()
  118. if addr == "" {
  119. return "", errors.New("manager addr is not set")
  120. }
  121. return addr, nil
  122. }
  123. // New returns new Node instance.
  124. func New(c *Config) (*Node, error) {
  125. if err := os.MkdirAll(c.StateDir, 0700); err != nil {
  126. return nil, err
  127. }
  128. stateFile := filepath.Join(c.StateDir, stateFilename)
  129. dt, err := ioutil.ReadFile(stateFile)
  130. var p []api.Peer
  131. if err != nil && !os.IsNotExist(err) {
  132. return nil, err
  133. }
  134. if err == nil {
  135. if err := json.Unmarshal(dt, &p); err != nil {
  136. return nil, err
  137. }
  138. }
  139. n := &Node{
  140. remotes: newPersistentRemotes(stateFile, p...),
  141. role: ca.WorkerRole,
  142. config: c,
  143. started: make(chan struct{}),
  144. stopped: make(chan struct{}),
  145. closed: make(chan struct{}),
  146. ready: make(chan struct{}),
  147. notifyNodeChange: make(chan *api.Node, 1),
  148. unlockKey: c.UnlockKey,
  149. }
  150. if n.config.JoinAddr != "" || n.config.ForceNewCluster {
  151. n.remotes = newPersistentRemotes(filepath.Join(n.config.StateDir, stateFilename))
  152. if n.config.JoinAddr != "" {
  153. n.remotes.Observe(api.Peer{Addr: n.config.JoinAddr}, remotes.DefaultObservationWeight)
  154. }
  155. }
  156. n.connBroker = connectionbroker.New(n.remotes)
  157. n.roleCond = sync.NewCond(n.RLocker())
  158. n.connCond = sync.NewCond(n.RLocker())
  159. return n, nil
  160. }
  161. // BindRemote starts a listener that exposes the remote API.
  162. func (n *Node) BindRemote(ctx context.Context, listenAddr string, advertiseAddr string) error {
  163. n.RLock()
  164. defer n.RUnlock()
  165. if n.manager == nil {
  166. return errors.New("manager is not running")
  167. }
  168. return n.manager.BindRemote(ctx, manager.RemoteAddrs{
  169. ListenAddr: listenAddr,
  170. AdvertiseAddr: advertiseAddr,
  171. })
  172. }
  173. // Start starts a node instance.
  174. func (n *Node) Start(ctx context.Context) error {
  175. err := errNodeStarted
  176. n.startOnce.Do(func() {
  177. close(n.started)
  178. go n.run(ctx)
  179. err = nil // clear error above, only once.
  180. })
  181. return err
  182. }
  183. func (n *Node) run(ctx context.Context) (err error) {
  184. defer func() {
  185. n.err = err
  186. close(n.closed)
  187. }()
  188. ctx, cancel := context.WithCancel(ctx)
  189. defer cancel()
  190. ctx = log.WithModule(ctx, "node")
  191. go func() {
  192. select {
  193. case <-ctx.Done():
  194. case <-n.stopped:
  195. cancel()
  196. }
  197. }()
  198. securityConfig, err := n.loadSecurityConfig(ctx)
  199. if err != nil {
  200. return err
  201. }
  202. taskDBPath := filepath.Join(n.config.StateDir, "worker/tasks.db")
  203. if err := os.MkdirAll(filepath.Dir(taskDBPath), 0777); err != nil {
  204. return err
  205. }
  206. db, err := bolt.Open(taskDBPath, 0666, nil)
  207. if err != nil {
  208. return err
  209. }
  210. defer db.Close()
  211. forceCertRenewal := make(chan struct{})
  212. renewCert := func() {
  213. select {
  214. case forceCertRenewal <- struct{}{}:
  215. case <-ctx.Done():
  216. }
  217. }
  218. go func() {
  219. for {
  220. select {
  221. case <-ctx.Done():
  222. return
  223. case node := <-n.notifyNodeChange:
  224. // If the server is sending us a ForceRenewal State, renew
  225. if node.Certificate.Status.State == api.IssuanceStateRotate {
  226. renewCert()
  227. continue
  228. }
  229. n.Lock()
  230. // If we got a role change, renew
  231. lastRole := n.role
  232. role := ca.WorkerRole
  233. if node.Role == api.NodeRoleManager {
  234. role = ca.ManagerRole
  235. }
  236. if lastRole == role {
  237. n.Unlock()
  238. continue
  239. }
  240. n.Unlock()
  241. renewCert()
  242. }
  243. }
  244. }()
  245. updates := ca.RenewTLSConfig(ctx, securityConfig, n.connBroker, forceCertRenewal)
  246. go func() {
  247. for {
  248. select {
  249. case certUpdate := <-updates:
  250. if certUpdate.Err != nil {
  251. logrus.Warnf("error renewing TLS certificate: %v", certUpdate.Err)
  252. continue
  253. }
  254. n.Lock()
  255. n.role = certUpdate.Role
  256. n.roleCond.Broadcast()
  257. n.Unlock()
  258. case <-ctx.Done():
  259. return
  260. }
  261. }
  262. }()
  263. role := n.role
  264. managerReady := make(chan struct{})
  265. agentReady := make(chan struct{})
  266. var managerErr error
  267. var agentErr error
  268. var wg sync.WaitGroup
  269. wg.Add(2)
  270. go func() {
  271. managerErr = n.superviseManager(ctx, securityConfig, managerReady) // store err and loop
  272. wg.Done()
  273. cancel()
  274. }()
  275. go func() {
  276. agentErr = n.runAgent(ctx, db, securityConfig.ClientTLSCreds, agentReady)
  277. wg.Done()
  278. cancel()
  279. }()
  280. go func() {
  281. <-agentReady
  282. if role == ca.ManagerRole {
  283. workerRole := make(chan struct{})
  284. waitRoleCtx, waitRoleCancel := context.WithCancel(ctx)
  285. go func() {
  286. if n.waitRole(waitRoleCtx, ca.WorkerRole) == nil {
  287. close(workerRole)
  288. }
  289. }()
  290. select {
  291. case <-managerReady:
  292. case <-workerRole:
  293. }
  294. waitRoleCancel()
  295. }
  296. close(n.ready)
  297. }()
  298. wg.Wait()
  299. if managerErr != nil && managerErr != context.Canceled {
  300. return managerErr
  301. }
  302. if agentErr != nil && agentErr != context.Canceled {
  303. return agentErr
  304. }
  305. return err
  306. }
  307. // Stop stops node execution
  308. func (n *Node) Stop(ctx context.Context) error {
  309. select {
  310. case <-n.started:
  311. default:
  312. return errNodeNotStarted
  313. }
  314. // ask agent to clean up assignments
  315. n.Lock()
  316. if n.agent != nil {
  317. if err := n.agent.Leave(ctx); err != nil {
  318. log.G(ctx).WithError(err).Error("agent failed to clean up assignments")
  319. }
  320. }
  321. n.Unlock()
  322. n.stopOnce.Do(func() {
  323. close(n.stopped)
  324. })
  325. select {
  326. case <-n.closed:
  327. return nil
  328. case <-ctx.Done():
  329. return ctx.Err()
  330. }
  331. }
  332. // Err returns the error that caused the node to shutdown or nil. Err blocks
  333. // until the node has fully shut down.
  334. func (n *Node) Err(ctx context.Context) error {
  335. select {
  336. case <-n.closed:
  337. return n.err
  338. case <-ctx.Done():
  339. return ctx.Err()
  340. }
  341. }
  342. func (n *Node) runAgent(ctx context.Context, db *bolt.DB, creds credentials.TransportCredentials, ready chan<- struct{}) error {
  343. waitCtx, waitCancel := context.WithCancel(ctx)
  344. remotesCh := n.remotes.WaitSelect(ctx)
  345. controlCh := n.ListenControlSocket(waitCtx)
  346. waitPeer:
  347. for {
  348. select {
  349. case <-ctx.Done():
  350. break waitPeer
  351. case <-remotesCh:
  352. break waitPeer
  353. case conn := <-controlCh:
  354. if conn != nil {
  355. break waitPeer
  356. }
  357. }
  358. }
  359. waitCancel()
  360. select {
  361. case <-ctx.Done():
  362. return ctx.Err()
  363. default:
  364. }
  365. a, err := agent.New(&agent.Config{
  366. Hostname: n.config.Hostname,
  367. ConnBroker: n.connBroker,
  368. Executor: n.config.Executor,
  369. DB: db,
  370. NotifyNodeChange: n.notifyNodeChange,
  371. Credentials: creds,
  372. })
  373. if err != nil {
  374. return err
  375. }
  376. if err := a.Start(ctx); err != nil {
  377. return err
  378. }
  379. n.Lock()
  380. n.agent = a
  381. n.Unlock()
  382. defer func() {
  383. n.Lock()
  384. n.agent = nil
  385. n.Unlock()
  386. }()
  387. go func() {
  388. <-a.Ready()
  389. close(ready)
  390. }()
  391. // todo: manually call stop on context cancellation?
  392. return a.Err(context.Background())
  393. }
  394. // Ready returns a channel that is closed after node's initialization has
  395. // completes for the first time.
  396. func (n *Node) Ready() <-chan struct{} {
  397. return n.ready
  398. }
  399. func (n *Node) setControlSocket(conn *grpc.ClientConn) {
  400. n.Lock()
  401. if n.conn != nil {
  402. n.conn.Close()
  403. }
  404. n.conn = conn
  405. n.connBroker.SetLocalConn(conn)
  406. n.connCond.Broadcast()
  407. n.Unlock()
  408. }
  409. // ListenControlSocket listens changes of a connection for managing the
  410. // cluster control api
  411. func (n *Node) ListenControlSocket(ctx context.Context) <-chan *grpc.ClientConn {
  412. c := make(chan *grpc.ClientConn, 1)
  413. n.RLock()
  414. conn := n.conn
  415. c <- conn
  416. done := make(chan struct{})
  417. go func() {
  418. select {
  419. case <-ctx.Done():
  420. n.connCond.Broadcast()
  421. case <-done:
  422. }
  423. }()
  424. go func() {
  425. defer close(c)
  426. defer close(done)
  427. defer n.RUnlock()
  428. for {
  429. select {
  430. case <-ctx.Done():
  431. return
  432. default:
  433. }
  434. if conn == n.conn {
  435. n.connCond.Wait()
  436. continue
  437. }
  438. conn = n.conn
  439. select {
  440. case c <- conn:
  441. case <-ctx.Done():
  442. return
  443. }
  444. }
  445. }()
  446. return c
  447. }
  448. // NodeID returns current node's ID. May be empty if not set.
  449. func (n *Node) NodeID() string {
  450. n.RLock()
  451. defer n.RUnlock()
  452. return n.nodeID
  453. }
  454. // Manager returns manager instance started by node. May be nil.
  455. func (n *Node) Manager() *manager.Manager {
  456. n.RLock()
  457. defer n.RUnlock()
  458. return n.manager
  459. }
  460. // Agent returns agent instance started by node. May be nil.
  461. func (n *Node) Agent() *agent.Agent {
  462. n.RLock()
  463. defer n.RUnlock()
  464. return n.agent
  465. }
  466. // IsStateDirty returns true if any objects have been added to raft which make
  467. // the state "dirty". Currently, the existence of any object other than the
  468. // default cluster or the local node implies a dirty state.
  469. func (n *Node) IsStateDirty() (bool, error) {
  470. n.RLock()
  471. defer n.RUnlock()
  472. if n.manager == nil {
  473. return false, errors.New("node is not a manager")
  474. }
  475. return n.manager.IsStateDirty()
  476. }
  477. // Remotes returns a list of known peers known to node.
  478. func (n *Node) Remotes() []api.Peer {
  479. weights := n.remotes.Weights()
  480. remotes := make([]api.Peer, 0, len(weights))
  481. for p := range weights {
  482. remotes = append(remotes, p)
  483. }
  484. return remotes
  485. }
  486. func (n *Node) loadSecurityConfig(ctx context.Context) (*ca.SecurityConfig, error) {
  487. paths := ca.NewConfigPaths(filepath.Join(n.config.StateDir, certDirectory))
  488. var securityConfig *ca.SecurityConfig
  489. krw := ca.NewKeyReadWriter(paths.Node, n.unlockKey, &manager.RaftDEKData{})
  490. if err := krw.Migrate(); err != nil {
  491. return nil, err
  492. }
  493. // Check if we already have a valid certificates on disk.
  494. rootCA, err := ca.GetLocalRootCA(paths.RootCA)
  495. if err != nil && err != ca.ErrNoLocalRootCA {
  496. return nil, err
  497. }
  498. if err == nil {
  499. securityConfig, err = ca.LoadSecurityConfig(ctx, rootCA, krw)
  500. if err != nil {
  501. _, isInvalidKEK := errors.Cause(err).(ca.ErrInvalidKEK)
  502. if isInvalidKEK {
  503. return nil, ErrInvalidUnlockKey
  504. } else if !os.IsNotExist(err) {
  505. return nil, errors.Wrapf(err, "error while loading TLS certificate in %s", paths.Node.Cert)
  506. }
  507. }
  508. }
  509. if securityConfig == nil {
  510. if n.config.JoinAddr == "" {
  511. // if we're not joining a cluster, bootstrap a new one - and we have to set the unlock key
  512. n.unlockKey = nil
  513. if n.config.AutoLockManagers {
  514. n.unlockKey = encryption.GenerateSecretKey()
  515. }
  516. krw = ca.NewKeyReadWriter(paths.Node, n.unlockKey, &manager.RaftDEKData{})
  517. rootCA, err = ca.CreateRootCA(ca.DefaultRootCN, paths.RootCA)
  518. if err != nil {
  519. return nil, err
  520. }
  521. log.G(ctx).Debug("generated CA key and certificate")
  522. } else if err == ca.ErrNoLocalRootCA { // from previous error loading the root CA from disk
  523. rootCA, err = ca.DownloadRootCA(ctx, paths.RootCA, n.config.JoinToken, n.connBroker)
  524. if err != nil {
  525. return nil, err
  526. }
  527. log.G(ctx).Debug("downloaded CA certificate")
  528. }
  529. // Obtain new certs and setup TLS certificates renewal for this node:
  530. // - If certificates weren't present on disk, we call CreateSecurityConfig, which blocks
  531. // until a valid certificate has been issued.
  532. // - We wait for CreateSecurityConfig to finish since we need a certificate to operate.
  533. // Attempt to load certificate from disk
  534. securityConfig, err = ca.LoadSecurityConfig(ctx, rootCA, krw)
  535. if err == nil {
  536. log.G(ctx).WithFields(logrus.Fields{
  537. "node.id": securityConfig.ClientTLSCreds.NodeID(),
  538. }).Debugf("loaded TLS certificate")
  539. } else {
  540. if _, ok := errors.Cause(err).(ca.ErrInvalidKEK); ok {
  541. return nil, ErrInvalidUnlockKey
  542. }
  543. log.G(ctx).WithError(err).Debugf("no node credentials found in: %s", krw.Target())
  544. securityConfig, err = rootCA.CreateSecurityConfig(ctx, krw, ca.CertificateRequestConfig{
  545. Token: n.config.JoinToken,
  546. Availability: n.config.Availability,
  547. ConnBroker: n.connBroker,
  548. })
  549. if err != nil {
  550. return nil, err
  551. }
  552. }
  553. }
  554. n.Lock()
  555. n.role = securityConfig.ClientTLSCreds.Role()
  556. n.nodeID = securityConfig.ClientTLSCreds.NodeID()
  557. n.roleCond.Broadcast()
  558. n.Unlock()
  559. return securityConfig, nil
  560. }
  561. func (n *Node) initManagerConnection(ctx context.Context, ready chan<- struct{}) error {
  562. opts := []grpc.DialOption{}
  563. insecureCreds := credentials.NewTLS(&tls.Config{InsecureSkipVerify: true})
  564. opts = append(opts, grpc.WithTransportCredentials(insecureCreds))
  565. addr := n.config.ListenControlAPI
  566. opts = append(opts, grpc.WithDialer(
  567. func(addr string, timeout time.Duration) (net.Conn, error) {
  568. return xnet.DialTimeoutLocal(addr, timeout)
  569. }))
  570. conn, err := grpc.Dial(addr, opts...)
  571. if err != nil {
  572. return err
  573. }
  574. client := api.NewHealthClient(conn)
  575. for {
  576. resp, err := client.Check(ctx, &api.HealthCheckRequest{Service: "ControlAPI"})
  577. if err != nil {
  578. return err
  579. }
  580. if resp.Status == api.HealthCheckResponse_SERVING {
  581. break
  582. }
  583. time.Sleep(500 * time.Millisecond)
  584. }
  585. n.setControlSocket(conn)
  586. if ready != nil {
  587. close(ready)
  588. }
  589. return nil
  590. }
  591. func (n *Node) waitRole(ctx context.Context, role string) error {
  592. n.roleCond.L.Lock()
  593. if role == n.role {
  594. n.roleCond.L.Unlock()
  595. return nil
  596. }
  597. finishCh := make(chan struct{})
  598. defer close(finishCh)
  599. go func() {
  600. select {
  601. case <-finishCh:
  602. case <-ctx.Done():
  603. // call broadcast to shutdown this function
  604. n.roleCond.Broadcast()
  605. }
  606. }()
  607. defer n.roleCond.L.Unlock()
  608. for role != n.role {
  609. n.roleCond.Wait()
  610. select {
  611. case <-ctx.Done():
  612. return ctx.Err()
  613. default:
  614. }
  615. }
  616. return nil
  617. }
  618. func (n *Node) runManager(ctx context.Context, securityConfig *ca.SecurityConfig, ready chan struct{}, workerRole <-chan struct{}) error {
  619. var remoteAPI *manager.RemoteAddrs
  620. if n.config.ListenRemoteAPI != "" {
  621. remoteAPI = &manager.RemoteAddrs{
  622. ListenAddr: n.config.ListenRemoteAPI,
  623. AdvertiseAddr: n.config.AdvertiseRemoteAPI,
  624. }
  625. }
  626. remoteAddr, _ := n.remotes.Select(n.NodeID())
  627. m, err := manager.New(&manager.Config{
  628. ForceNewCluster: n.config.ForceNewCluster,
  629. RemoteAPI: remoteAPI,
  630. ControlAPI: n.config.ListenControlAPI,
  631. SecurityConfig: securityConfig,
  632. ExternalCAs: n.config.ExternalCAs,
  633. JoinRaft: remoteAddr.Addr,
  634. StateDir: n.config.StateDir,
  635. HeartbeatTick: n.config.HeartbeatTick,
  636. ElectionTick: n.config.ElectionTick,
  637. AutoLockManagers: n.config.AutoLockManagers,
  638. UnlockKey: n.unlockKey,
  639. Availability: n.config.Availability,
  640. PluginGetter: n.config.PluginGetter,
  641. })
  642. if err != nil {
  643. return err
  644. }
  645. done := make(chan struct{})
  646. var runErr error
  647. go func() {
  648. if err := m.Run(context.Background()); err != nil {
  649. runErr = err
  650. }
  651. close(done)
  652. }()
  653. var clearData bool
  654. defer func() {
  655. n.Lock()
  656. n.manager = nil
  657. n.Unlock()
  658. m.Stop(ctx, clearData)
  659. <-done
  660. n.setControlSocket(nil)
  661. }()
  662. n.Lock()
  663. n.manager = m
  664. n.Unlock()
  665. connCtx, connCancel := context.WithCancel(ctx)
  666. defer connCancel()
  667. go n.initManagerConnection(connCtx, ready)
  668. // wait for manager stop or for role change
  669. select {
  670. case <-done:
  671. return runErr
  672. case <-workerRole:
  673. log.G(ctx).Info("role changed to worker, stopping manager")
  674. clearData = true
  675. case <-m.RemovedFromRaft():
  676. log.G(ctx).Info("manager removed from raft cluster, stopping manager")
  677. clearData = true
  678. case <-ctx.Done():
  679. return ctx.Err()
  680. }
  681. return nil
  682. }
  683. func (n *Node) superviseManager(ctx context.Context, securityConfig *ca.SecurityConfig, ready chan struct{}) error {
  684. for {
  685. if err := n.waitRole(ctx, ca.ManagerRole); err != nil {
  686. return err
  687. }
  688. workerRole := make(chan struct{})
  689. waitRoleCtx, waitRoleCancel := context.WithCancel(ctx)
  690. go func() {
  691. if n.waitRole(waitRoleCtx, ca.WorkerRole) == nil {
  692. close(workerRole)
  693. }
  694. }()
  695. if err := n.runManager(ctx, securityConfig, ready, workerRole); err != nil {
  696. waitRoleCancel()
  697. return errors.Wrap(err, "manager stopped")
  698. }
  699. // If the manager stopped running and our role is still
  700. // "manager", it's possible that the manager was demoted and
  701. // the agent hasn't realized this yet. We should wait for the
  702. // role to change instead of restarting the manager immediately.
  703. timer := time.NewTimer(16 * time.Second)
  704. select {
  705. case <-timer.C:
  706. log.G(ctx).Warn("failed to get worker role after manager stop, restarting manager")
  707. case <-workerRole:
  708. case <-ctx.Done():
  709. timer.Stop()
  710. waitRoleCancel()
  711. return ctx.Err()
  712. }
  713. timer.Stop()
  714. waitRoleCancel()
  715. ready = nil
  716. }
  717. }
  718. type persistentRemotes struct {
  719. sync.RWMutex
  720. c *sync.Cond
  721. remotes.Remotes
  722. storePath string
  723. lastSavedState []api.Peer
  724. }
  725. func newPersistentRemotes(f string, peers ...api.Peer) *persistentRemotes {
  726. pr := &persistentRemotes{
  727. storePath: f,
  728. Remotes: remotes.NewRemotes(peers...),
  729. }
  730. pr.c = sync.NewCond(pr.RLocker())
  731. return pr
  732. }
  733. func (s *persistentRemotes) Observe(peer api.Peer, weight int) {
  734. s.Lock()
  735. defer s.Unlock()
  736. s.Remotes.Observe(peer, weight)
  737. s.c.Broadcast()
  738. if err := s.save(); err != nil {
  739. logrus.Errorf("error writing cluster state file: %v", err)
  740. return
  741. }
  742. return
  743. }
  744. func (s *persistentRemotes) Remove(peers ...api.Peer) {
  745. s.Lock()
  746. defer s.Unlock()
  747. s.Remotes.Remove(peers...)
  748. if err := s.save(); err != nil {
  749. logrus.Errorf("error writing cluster state file: %v", err)
  750. return
  751. }
  752. return
  753. }
  754. func (s *persistentRemotes) save() error {
  755. weights := s.Weights()
  756. remotes := make([]api.Peer, 0, len(weights))
  757. for r := range weights {
  758. remotes = append(remotes, r)
  759. }
  760. sort.Sort(sortablePeers(remotes))
  761. if reflect.DeepEqual(remotes, s.lastSavedState) {
  762. return nil
  763. }
  764. dt, err := json.Marshal(remotes)
  765. if err != nil {
  766. return err
  767. }
  768. s.lastSavedState = remotes
  769. return ioutils.AtomicWriteFile(s.storePath, dt, 0600)
  770. }
  771. // WaitSelect waits until at least one remote becomes available and then selects one.
  772. func (s *persistentRemotes) WaitSelect(ctx context.Context) <-chan api.Peer {
  773. c := make(chan api.Peer, 1)
  774. s.RLock()
  775. done := make(chan struct{})
  776. go func() {
  777. select {
  778. case <-ctx.Done():
  779. s.c.Broadcast()
  780. case <-done:
  781. }
  782. }()
  783. go func() {
  784. defer s.RUnlock()
  785. defer close(c)
  786. defer close(done)
  787. for {
  788. if ctx.Err() != nil {
  789. return
  790. }
  791. p, err := s.Select()
  792. if err == nil {
  793. c <- p
  794. return
  795. }
  796. s.c.Wait()
  797. }
  798. }()
  799. return c
  800. }
  801. // sortablePeers is a sort wrapper for []api.Peer
  802. type sortablePeers []api.Peer
  803. func (sp sortablePeers) Less(i, j int) bool { return sp[i].NodeID < sp[j].NodeID }
  804. func (sp sortablePeers) Len() int { return len(sp) }
  805. func (sp sortablePeers) Swap(i, j int) { sp[i], sp[j] = sp[j], sp[i] }