swarm.go 17 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620
  1. package cluster // import "github.com/docker/docker/daemon/cluster"
  2. import (
  3. "context"
  4. "fmt"
  5. "net"
  6. "strings"
  7. "time"
  8. "github.com/containerd/log"
  9. "github.com/docker/docker/api/types/backend"
  10. "github.com/docker/docker/api/types/container"
  11. "github.com/docker/docker/api/types/filters"
  12. types "github.com/docker/docker/api/types/swarm"
  13. "github.com/docker/docker/daemon/cluster/convert"
  14. "github.com/docker/docker/errdefs"
  15. "github.com/docker/docker/opts"
  16. "github.com/docker/docker/pkg/stack"
  17. swarmapi "github.com/moby/swarmkit/v2/api"
  18. "github.com/moby/swarmkit/v2/manager/encryption"
  19. swarmnode "github.com/moby/swarmkit/v2/node"
  20. "github.com/pkg/errors"
  21. "google.golang.org/grpc"
  22. )
  23. // Init initializes new cluster from user provided request.
  24. func (c *Cluster) Init(req types.InitRequest) (string, error) {
  25. c.controlMutex.Lock()
  26. defer c.controlMutex.Unlock()
  27. if c.nr != nil {
  28. if req.ForceNewCluster {
  29. // Take c.mu temporarily to wait for presently running
  30. // API handlers to finish before shutting down the node.
  31. c.mu.Lock()
  32. if !c.nr.nodeState.IsManager() {
  33. c.mu.Unlock()
  34. return "", errSwarmNotManager
  35. }
  36. c.mu.Unlock()
  37. if err := c.nr.Stop(); err != nil {
  38. return "", err
  39. }
  40. } else {
  41. return "", errSwarmExists
  42. }
  43. }
  44. if err := validateAndSanitizeInitRequest(&req); err != nil {
  45. return "", errdefs.InvalidParameter(err)
  46. }
  47. listenHost, listenPort, err := resolveListenAddr(req.ListenAddr)
  48. if err != nil {
  49. return "", errdefs.InvalidParameter(err)
  50. }
  51. advertiseHost, advertisePort, err := c.resolveAdvertiseAddr(req.AdvertiseAddr, listenPort)
  52. if err != nil {
  53. return "", err
  54. }
  55. dataPathAddr, err := resolveDataPathAddr(req.DataPathAddr)
  56. if err != nil {
  57. return "", err
  58. }
  59. localAddr := listenHost
  60. // If the local address is undetermined, the advertise address
  61. // will be used as local address, if it belongs to this system.
  62. // If the advertise address is not local, then we try to find
  63. // a system address to use as local address. If this fails,
  64. // we give up and ask the user to pass the listen address.
  65. if net.ParseIP(localAddr).IsUnspecified() {
  66. advertiseIP := net.ParseIP(advertiseHost)
  67. found := false
  68. for _, systemIP := range listSystemIPs() {
  69. if systemIP.Equal(advertiseIP) {
  70. localAddr = advertiseIP.String()
  71. found = true
  72. break
  73. }
  74. }
  75. if !found {
  76. ip, err := c.resolveSystemAddr()
  77. if err != nil {
  78. log.G(context.TODO()).Warnf("Could not find a local address: %v", err)
  79. return "", errMustSpecifyListenAddr
  80. }
  81. localAddr = ip.String()
  82. }
  83. }
  84. if err := validateDefaultAddrPool(req.DefaultAddrPool, req.SubnetSize); err != nil {
  85. return "", err
  86. }
  87. port, err := getDataPathPort(req.DataPathPort)
  88. if err != nil {
  89. return "", err
  90. }
  91. nr, err := c.newNodeRunner(nodeStartConfig{
  92. forceNewCluster: req.ForceNewCluster,
  93. autolock: req.AutoLockManagers,
  94. LocalAddr: localAddr,
  95. ListenAddr: net.JoinHostPort(listenHost, listenPort),
  96. AdvertiseAddr: net.JoinHostPort(advertiseHost, advertisePort),
  97. DataPathAddr: dataPathAddr,
  98. DefaultAddressPool: req.DefaultAddrPool,
  99. SubnetSize: req.SubnetSize,
  100. availability: req.Availability,
  101. DataPathPort: port,
  102. })
  103. if err != nil {
  104. return "", err
  105. }
  106. c.mu.Lock()
  107. c.nr = nr
  108. c.mu.Unlock()
  109. if err := <-nr.Ready(); err != nil {
  110. c.mu.Lock()
  111. c.nr = nil
  112. c.mu.Unlock()
  113. if !req.ForceNewCluster { // if failure on first attempt don't keep state
  114. if err := clearPersistentState(c.root); err != nil {
  115. return "", err
  116. }
  117. }
  118. return "", err
  119. }
  120. state := nr.State()
  121. if state.swarmNode == nil { // should never happen but protect from panic
  122. return "", errors.New("invalid cluster state for spec initialization")
  123. }
  124. if err := initClusterSpec(state.swarmNode, req.Spec); err != nil {
  125. return "", err
  126. }
  127. return state.NodeID(), nil
  128. }
  129. // Join makes current Cluster part of an existing swarm cluster.
  130. func (c *Cluster) Join(req types.JoinRequest) error {
  131. c.controlMutex.Lock()
  132. defer c.controlMutex.Unlock()
  133. c.mu.Lock()
  134. if c.nr != nil {
  135. c.mu.Unlock()
  136. return errors.WithStack(errSwarmExists)
  137. }
  138. c.mu.Unlock()
  139. if err := validateAndSanitizeJoinRequest(&req); err != nil {
  140. return errdefs.InvalidParameter(err)
  141. }
  142. listenHost, listenPort, err := resolveListenAddr(req.ListenAddr)
  143. if err != nil {
  144. return err
  145. }
  146. var advertiseAddr string
  147. if req.AdvertiseAddr != "" {
  148. advertiseHost, advertisePort, err := c.resolveAdvertiseAddr(req.AdvertiseAddr, listenPort)
  149. // For joining, we don't need to provide an advertise address,
  150. // since the remote side can detect it.
  151. if err == nil {
  152. advertiseAddr = net.JoinHostPort(advertiseHost, advertisePort)
  153. }
  154. }
  155. dataPathAddr, err := resolveDataPathAddr(req.DataPathAddr)
  156. if err != nil {
  157. return err
  158. }
  159. nr, err := c.newNodeRunner(nodeStartConfig{
  160. RemoteAddr: req.RemoteAddrs[0],
  161. ListenAddr: net.JoinHostPort(listenHost, listenPort),
  162. AdvertiseAddr: advertiseAddr,
  163. DataPathAddr: dataPathAddr,
  164. joinAddr: req.RemoteAddrs[0],
  165. joinToken: req.JoinToken,
  166. availability: req.Availability,
  167. })
  168. if err != nil {
  169. return err
  170. }
  171. c.mu.Lock()
  172. c.nr = nr
  173. c.mu.Unlock()
  174. timeout := time.NewTimer(swarmConnectTimeout)
  175. defer timeout.Stop()
  176. select {
  177. case <-timeout.C:
  178. return errSwarmJoinTimeoutReached
  179. case err := <-nr.Ready():
  180. if err != nil {
  181. c.mu.Lock()
  182. c.nr = nil
  183. c.mu.Unlock()
  184. if err := clearPersistentState(c.root); err != nil {
  185. return err
  186. }
  187. }
  188. return err
  189. }
  190. }
  191. // Inspect retrieves the configuration properties of a managed swarm cluster.
  192. func (c *Cluster) Inspect() (types.Swarm, error) {
  193. var swarm types.Swarm
  194. if err := c.lockedManagerAction(func(ctx context.Context, state nodeState) error {
  195. s, err := c.inspect(ctx, state)
  196. if err != nil {
  197. return err
  198. }
  199. swarm = s
  200. return nil
  201. }); err != nil {
  202. return types.Swarm{}, err
  203. }
  204. return swarm, nil
  205. }
  206. func (c *Cluster) inspect(ctx context.Context, state nodeState) (types.Swarm, error) {
  207. s, err := getSwarm(ctx, state.controlClient)
  208. if err != nil {
  209. return types.Swarm{}, err
  210. }
  211. return convert.SwarmFromGRPC(*s), nil
  212. }
  213. // Update updates configuration of a managed swarm cluster.
  214. func (c *Cluster) Update(version uint64, spec types.Spec, flags types.UpdateFlags) error {
  215. return c.lockedManagerAction(func(ctx context.Context, state nodeState) error {
  216. swarm, err := getSwarm(ctx, state.controlClient)
  217. if err != nil {
  218. return err
  219. }
  220. // Validate spec name.
  221. if spec.Annotations.Name == "" {
  222. spec.Annotations.Name = "default"
  223. } else if spec.Annotations.Name != "default" {
  224. return errdefs.InvalidParameter(errors.New(`swarm spec must be named "default"`))
  225. }
  226. // In update, client should provide the complete spec of the swarm, including
  227. // Name and Labels. If a field is specified with 0 or nil, then the default value
  228. // will be used to swarmkit.
  229. clusterSpec, err := convert.SwarmSpecToGRPC(spec)
  230. if err != nil {
  231. return errdefs.InvalidParameter(err)
  232. }
  233. _, err = state.controlClient.UpdateCluster(
  234. ctx,
  235. &swarmapi.UpdateClusterRequest{
  236. ClusterID: swarm.ID,
  237. Spec: &clusterSpec,
  238. ClusterVersion: &swarmapi.Version{
  239. Index: version,
  240. },
  241. Rotation: swarmapi.KeyRotation{
  242. WorkerJoinToken: flags.RotateWorkerToken,
  243. ManagerJoinToken: flags.RotateManagerToken,
  244. ManagerUnlockKey: flags.RotateManagerUnlockKey,
  245. },
  246. },
  247. )
  248. return err
  249. })
  250. }
  251. // GetUnlockKey returns the unlock key for the swarm.
  252. func (c *Cluster) GetUnlockKey() (string, error) {
  253. var resp *swarmapi.GetUnlockKeyResponse
  254. if err := c.lockedManagerAction(func(ctx context.Context, state nodeState) error {
  255. client := swarmapi.NewCAClient(state.grpcConn)
  256. r, err := client.GetUnlockKey(ctx, &swarmapi.GetUnlockKeyRequest{})
  257. if err != nil {
  258. return err
  259. }
  260. resp = r
  261. return nil
  262. }); err != nil {
  263. return "", err
  264. }
  265. if len(resp.UnlockKey) == 0 {
  266. // no key
  267. return "", nil
  268. }
  269. return encryption.HumanReadableKey(resp.UnlockKey), nil
  270. }
  271. // UnlockSwarm provides a key to decrypt data that is encrypted at rest.
  272. func (c *Cluster) UnlockSwarm(req types.UnlockRequest) error {
  273. c.controlMutex.Lock()
  274. defer c.controlMutex.Unlock()
  275. c.mu.RLock()
  276. state := c.currentNodeState()
  277. if !state.IsActiveManager() {
  278. // when manager is not active,
  279. // unless it is locked, otherwise return error.
  280. if err := c.errNoManager(state); err != errSwarmLocked {
  281. c.mu.RUnlock()
  282. return err
  283. }
  284. } else {
  285. // when manager is active, return an error of "not locked"
  286. c.mu.RUnlock()
  287. return notLockedError{}
  288. }
  289. // only when swarm is locked, code running reaches here
  290. nr := c.nr
  291. c.mu.RUnlock()
  292. key, err := encryption.ParseHumanReadableKey(req.UnlockKey)
  293. if err != nil {
  294. return errdefs.InvalidParameter(err)
  295. }
  296. config := nr.config
  297. config.lockKey = key
  298. if err := nr.Stop(); err != nil {
  299. return err
  300. }
  301. nr, err = c.newNodeRunner(config)
  302. if err != nil {
  303. return err
  304. }
  305. c.mu.Lock()
  306. c.nr = nr
  307. c.mu.Unlock()
  308. if err := <-nr.Ready(); err != nil {
  309. if errors.Is(err, errSwarmLocked) {
  310. return invalidUnlockKey{}
  311. }
  312. return errors.Errorf("swarm component could not be started: %v", err)
  313. }
  314. return nil
  315. }
  316. // Leave shuts down Cluster and removes current state.
  317. func (c *Cluster) Leave(ctx context.Context, force bool) error {
  318. c.controlMutex.Lock()
  319. defer c.controlMutex.Unlock()
  320. c.mu.Lock()
  321. nr := c.nr
  322. if nr == nil {
  323. c.mu.Unlock()
  324. return errors.WithStack(errNoSwarm)
  325. }
  326. state := c.currentNodeState()
  327. c.mu.Unlock()
  328. if errors.Is(state.err, errSwarmLocked) && !force {
  329. // leave a locked swarm without --force is not allowed
  330. return errors.WithStack(notAvailableError("Swarm is encrypted and locked. Please unlock it first or use `--force` to ignore this message."))
  331. }
  332. if state.IsManager() && !force {
  333. msg := "You are attempting to leave the swarm on a node that is participating as a manager. "
  334. if state.IsActiveManager() {
  335. active, reachable, unreachable, err := managerStats(state.controlClient, state.NodeID())
  336. if err == nil {
  337. if active && removingManagerCausesLossOfQuorum(reachable, unreachable) {
  338. if isLastManager(reachable, unreachable) {
  339. msg += "Removing the last manager erases all current state of the swarm. Use `--force` to ignore this message. "
  340. return errors.WithStack(notAvailableError(msg))
  341. }
  342. msg += fmt.Sprintf("Removing this node leaves %v managers out of %v. Without a Raft quorum your swarm will be inaccessible. ", reachable-1, reachable+unreachable)
  343. }
  344. }
  345. } else {
  346. msg += "Doing so may lose the consensus of your cluster. "
  347. }
  348. msg += "The only way to restore a swarm that has lost consensus is to reinitialize it with `--force-new-cluster`. Use `--force` to suppress this message."
  349. return errors.WithStack(notAvailableError(msg))
  350. }
  351. // release readers in here
  352. if err := nr.Stop(); err != nil {
  353. log.G(ctx).Errorf("failed to shut down cluster node: %v", err)
  354. stack.Dump()
  355. return err
  356. }
  357. c.mu.Lock()
  358. c.nr = nil
  359. c.mu.Unlock()
  360. if nodeID := state.NodeID(); nodeID != "" {
  361. nodeContainers, err := c.listContainerForNode(ctx, nodeID)
  362. if err != nil {
  363. return err
  364. }
  365. for _, id := range nodeContainers {
  366. if err := c.config.Backend.ContainerRm(id, &backend.ContainerRmConfig{ForceRemove: true}); err != nil {
  367. log.G(ctx).Errorf("error removing %v: %v", id, err)
  368. }
  369. }
  370. }
  371. // todo: cleanup optional?
  372. if err := clearPersistentState(c.root); err != nil {
  373. return err
  374. }
  375. c.config.Backend.DaemonLeavesCluster()
  376. return nil
  377. }
  378. // Info returns information about the current cluster state.
  379. func (c *Cluster) Info(ctx context.Context) types.Info {
  380. info := types.Info{
  381. NodeAddr: c.GetAdvertiseAddress(),
  382. }
  383. c.mu.RLock()
  384. defer c.mu.RUnlock()
  385. state := c.currentNodeState()
  386. info.LocalNodeState = state.status
  387. if state.err != nil {
  388. info.Error = state.err.Error()
  389. }
  390. ctx, cancel := c.getRequestContext(ctx)
  391. defer cancel()
  392. if state.IsActiveManager() {
  393. info.ControlAvailable = true
  394. swarm, err := c.inspect(ctx, state)
  395. if err != nil {
  396. info.Error = err.Error()
  397. }
  398. info.Cluster = &swarm.ClusterInfo
  399. if r, err := state.controlClient.ListNodes(
  400. ctx, &swarmapi.ListNodesRequest{},
  401. grpc.MaxCallRecvMsgSize(defaultRecvSizeForListResponse),
  402. ); err != nil {
  403. info.Error = err.Error()
  404. } else {
  405. info.Nodes = len(r.Nodes)
  406. for _, n := range r.Nodes {
  407. if n.ManagerStatus != nil {
  408. info.Managers = info.Managers + 1
  409. }
  410. }
  411. }
  412. switch info.LocalNodeState {
  413. case types.LocalNodeStateInactive, types.LocalNodeStateLocked, types.LocalNodeStateError:
  414. // nothing to do
  415. default:
  416. if info.Managers == 2 {
  417. const warn string = `WARNING: Running Swarm in a two-manager configuration. This configuration provides
  418. no fault tolerance, and poses a high risk to lose control over the cluster.
  419. Refer to https://docs.docker.com/engine/swarm/admin_guide/ to configure the
  420. Swarm for fault-tolerance.`
  421. info.Warnings = append(info.Warnings, warn)
  422. }
  423. }
  424. }
  425. if state.swarmNode != nil {
  426. for _, r := range state.swarmNode.Remotes() {
  427. info.RemoteManagers = append(info.RemoteManagers, types.Peer{NodeID: r.NodeID, Addr: r.Addr})
  428. }
  429. info.NodeID = state.swarmNode.NodeID()
  430. }
  431. return info
  432. }
  433. // Status returns a textual representation of the node's swarm status and role (manager/worker)
  434. func (c *Cluster) Status() string {
  435. c.mu.RLock()
  436. s := c.currentNodeState()
  437. c.mu.RUnlock()
  438. state := string(s.status)
  439. if s.status == types.LocalNodeStateActive {
  440. if s.IsActiveManager() || s.IsManager() {
  441. state += "/manager"
  442. } else {
  443. state += "/worker"
  444. }
  445. }
  446. return state
  447. }
  448. func validateAndSanitizeInitRequest(req *types.InitRequest) error {
  449. var err error
  450. req.ListenAddr, err = validateAddr(req.ListenAddr)
  451. if err != nil {
  452. return fmt.Errorf("invalid ListenAddr %q: %v", req.ListenAddr, err)
  453. }
  454. if req.Spec.Annotations.Name == "" {
  455. req.Spec.Annotations.Name = "default"
  456. } else if req.Spec.Annotations.Name != "default" {
  457. return errors.New(`swarm spec must be named "default"`)
  458. }
  459. return nil
  460. }
  461. func validateAndSanitizeJoinRequest(req *types.JoinRequest) error {
  462. var err error
  463. req.ListenAddr, err = validateAddr(req.ListenAddr)
  464. if err != nil {
  465. return fmt.Errorf("invalid ListenAddr %q: %v", req.ListenAddr, err)
  466. }
  467. if len(req.RemoteAddrs) == 0 {
  468. return errors.New("at least 1 RemoteAddr is required to join")
  469. }
  470. for i := range req.RemoteAddrs {
  471. req.RemoteAddrs[i], err = validateAddr(req.RemoteAddrs[i])
  472. if err != nil {
  473. return fmt.Errorf("invalid remoteAddr %q: %v", req.RemoteAddrs[i], err)
  474. }
  475. }
  476. return nil
  477. }
  478. func validateAddr(addr string) (string, error) {
  479. if addr == "" {
  480. return addr, errors.New("invalid empty address")
  481. }
  482. newaddr, err := opts.ParseTCPAddr(addr, defaultAddr)
  483. if err != nil {
  484. // TODO(thaJeztah) why are we ignoring the error here? Is this to allow "non-tcp" addresses?
  485. return addr, nil
  486. }
  487. return strings.TrimPrefix(newaddr, "tcp://"), nil
  488. }
  489. func initClusterSpec(node *swarmnode.Node, spec types.Spec) error {
  490. ctx, cancel := context.WithTimeout(context.Background(), 5*time.Second)
  491. defer cancel()
  492. for conn := range node.ListenControlSocket(ctx) {
  493. if ctx.Err() != nil {
  494. return ctx.Err()
  495. }
  496. if conn != nil {
  497. client := swarmapi.NewControlClient(conn)
  498. var cluster *swarmapi.Cluster
  499. for i := 0; ; i++ {
  500. lcr, err := client.ListClusters(ctx, &swarmapi.ListClustersRequest{})
  501. if err != nil {
  502. return fmt.Errorf("error on listing clusters: %v", err)
  503. }
  504. if len(lcr.Clusters) == 0 {
  505. if i < 10 {
  506. time.Sleep(200 * time.Millisecond)
  507. continue
  508. }
  509. return errors.New("empty list of clusters was returned")
  510. }
  511. cluster = lcr.Clusters[0]
  512. break
  513. }
  514. // In init, we take the initial default values from swarmkit, and merge
  515. // any non nil or 0 value from spec to GRPC spec. This will leave the
  516. // default value alone.
  517. // Note that this is different from Update(), as in Update() we expect
  518. // user to specify the complete spec of the cluster (as they already know
  519. // the existing one and knows which field to update)
  520. clusterSpec, err := convert.MergeSwarmSpecToGRPC(spec, cluster.Spec)
  521. if err != nil {
  522. return fmt.Errorf("error updating cluster settings: %v", err)
  523. }
  524. _, err = client.UpdateCluster(ctx, &swarmapi.UpdateClusterRequest{
  525. ClusterID: cluster.ID,
  526. ClusterVersion: &cluster.Meta.Version,
  527. Spec: &clusterSpec,
  528. })
  529. if err != nil {
  530. return fmt.Errorf("error updating cluster settings: %v", err)
  531. }
  532. return nil
  533. }
  534. }
  535. return ctx.Err()
  536. }
  537. func (c *Cluster) listContainerForNode(ctx context.Context, nodeID string) ([]string, error) {
  538. var ids []string
  539. containers, err := c.config.Backend.Containers(ctx, &container.ListOptions{
  540. Filters: filters.NewArgs(filters.Arg("label", "com.docker.swarm.node.id="+nodeID)),
  541. })
  542. if err != nil {
  543. return []string{}, err
  544. }
  545. for _, c := range containers {
  546. ids = append(ids, c.ID)
  547. }
  548. return ids, nil
  549. }