swarm.go 17 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619
  1. package cluster // import "github.com/docker/docker/daemon/cluster"
  2. import (
  3. "context"
  4. "fmt"
  5. "net"
  6. "strings"
  7. "time"
  8. "github.com/containerd/containerd/log"
  9. apitypes "github.com/docker/docker/api/types"
  10. "github.com/docker/docker/api/types/filters"
  11. types "github.com/docker/docker/api/types/swarm"
  12. "github.com/docker/docker/daemon/cluster/convert"
  13. "github.com/docker/docker/errdefs"
  14. "github.com/docker/docker/opts"
  15. "github.com/docker/docker/pkg/stack"
  16. swarmapi "github.com/moby/swarmkit/v2/api"
  17. "github.com/moby/swarmkit/v2/manager/encryption"
  18. swarmnode "github.com/moby/swarmkit/v2/node"
  19. "github.com/pkg/errors"
  20. "google.golang.org/grpc"
  21. )
  22. // Init initializes new cluster from user provided request.
  23. func (c *Cluster) Init(req types.InitRequest) (string, error) {
  24. c.controlMutex.Lock()
  25. defer c.controlMutex.Unlock()
  26. if c.nr != nil {
  27. if req.ForceNewCluster {
  28. // Take c.mu temporarily to wait for presently running
  29. // API handlers to finish before shutting down the node.
  30. c.mu.Lock()
  31. if !c.nr.nodeState.IsManager() {
  32. c.mu.Unlock()
  33. return "", errSwarmNotManager
  34. }
  35. c.mu.Unlock()
  36. if err := c.nr.Stop(); err != nil {
  37. return "", err
  38. }
  39. } else {
  40. return "", errSwarmExists
  41. }
  42. }
  43. if err := validateAndSanitizeInitRequest(&req); err != nil {
  44. return "", errdefs.InvalidParameter(err)
  45. }
  46. listenHost, listenPort, err := resolveListenAddr(req.ListenAddr)
  47. if err != nil {
  48. return "", errdefs.InvalidParameter(err)
  49. }
  50. advertiseHost, advertisePort, err := c.resolveAdvertiseAddr(req.AdvertiseAddr, listenPort)
  51. if err != nil {
  52. return "", err
  53. }
  54. dataPathAddr, err := resolveDataPathAddr(req.DataPathAddr)
  55. if err != nil {
  56. return "", err
  57. }
  58. localAddr := listenHost
  59. // If the local address is undetermined, the advertise address
  60. // will be used as local address, if it belongs to this system.
  61. // If the advertise address is not local, then we try to find
  62. // a system address to use as local address. If this fails,
  63. // we give up and ask the user to pass the listen address.
  64. if net.ParseIP(localAddr).IsUnspecified() {
  65. advertiseIP := net.ParseIP(advertiseHost)
  66. found := false
  67. for _, systemIP := range listSystemIPs() {
  68. if systemIP.Equal(advertiseIP) {
  69. localAddr = advertiseIP.String()
  70. found = true
  71. break
  72. }
  73. }
  74. if !found {
  75. ip, err := c.resolveSystemAddr()
  76. if err != nil {
  77. log.G(context.TODO()).Warnf("Could not find a local address: %v", err)
  78. return "", errMustSpecifyListenAddr
  79. }
  80. localAddr = ip.String()
  81. }
  82. }
  83. if err := validateDefaultAddrPool(req.DefaultAddrPool, req.SubnetSize); err != nil {
  84. return "", err
  85. }
  86. port, err := getDataPathPort(req.DataPathPort)
  87. if err != nil {
  88. return "", err
  89. }
  90. nr, err := c.newNodeRunner(nodeStartConfig{
  91. forceNewCluster: req.ForceNewCluster,
  92. autolock: req.AutoLockManagers,
  93. LocalAddr: localAddr,
  94. ListenAddr: net.JoinHostPort(listenHost, listenPort),
  95. AdvertiseAddr: net.JoinHostPort(advertiseHost, advertisePort),
  96. DataPathAddr: dataPathAddr,
  97. DefaultAddressPool: req.DefaultAddrPool,
  98. SubnetSize: req.SubnetSize,
  99. availability: req.Availability,
  100. DataPathPort: port,
  101. })
  102. if err != nil {
  103. return "", err
  104. }
  105. c.mu.Lock()
  106. c.nr = nr
  107. c.mu.Unlock()
  108. if err := <-nr.Ready(); err != nil {
  109. c.mu.Lock()
  110. c.nr = nil
  111. c.mu.Unlock()
  112. if !req.ForceNewCluster { // if failure on first attempt don't keep state
  113. if err := clearPersistentState(c.root); err != nil {
  114. return "", err
  115. }
  116. }
  117. return "", err
  118. }
  119. state := nr.State()
  120. if state.swarmNode == nil { // should never happen but protect from panic
  121. return "", errors.New("invalid cluster state for spec initialization")
  122. }
  123. if err := initClusterSpec(state.swarmNode, req.Spec); err != nil {
  124. return "", err
  125. }
  126. return state.NodeID(), nil
  127. }
  128. // Join makes current Cluster part of an existing swarm cluster.
  129. func (c *Cluster) Join(req types.JoinRequest) error {
  130. c.controlMutex.Lock()
  131. defer c.controlMutex.Unlock()
  132. c.mu.Lock()
  133. if c.nr != nil {
  134. c.mu.Unlock()
  135. return errors.WithStack(errSwarmExists)
  136. }
  137. c.mu.Unlock()
  138. if err := validateAndSanitizeJoinRequest(&req); err != nil {
  139. return errdefs.InvalidParameter(err)
  140. }
  141. listenHost, listenPort, err := resolveListenAddr(req.ListenAddr)
  142. if err != nil {
  143. return err
  144. }
  145. var advertiseAddr string
  146. if req.AdvertiseAddr != "" {
  147. advertiseHost, advertisePort, err := c.resolveAdvertiseAddr(req.AdvertiseAddr, listenPort)
  148. // For joining, we don't need to provide an advertise address,
  149. // since the remote side can detect it.
  150. if err == nil {
  151. advertiseAddr = net.JoinHostPort(advertiseHost, advertisePort)
  152. }
  153. }
  154. dataPathAddr, err := resolveDataPathAddr(req.DataPathAddr)
  155. if err != nil {
  156. return err
  157. }
  158. nr, err := c.newNodeRunner(nodeStartConfig{
  159. RemoteAddr: req.RemoteAddrs[0],
  160. ListenAddr: net.JoinHostPort(listenHost, listenPort),
  161. AdvertiseAddr: advertiseAddr,
  162. DataPathAddr: dataPathAddr,
  163. joinAddr: req.RemoteAddrs[0],
  164. joinToken: req.JoinToken,
  165. availability: req.Availability,
  166. })
  167. if err != nil {
  168. return err
  169. }
  170. c.mu.Lock()
  171. c.nr = nr
  172. c.mu.Unlock()
  173. timeout := time.NewTimer(swarmConnectTimeout)
  174. defer timeout.Stop()
  175. select {
  176. case <-timeout.C:
  177. return errSwarmJoinTimeoutReached
  178. case err := <-nr.Ready():
  179. if err != nil {
  180. c.mu.Lock()
  181. c.nr = nil
  182. c.mu.Unlock()
  183. if err := clearPersistentState(c.root); err != nil {
  184. return err
  185. }
  186. }
  187. return err
  188. }
  189. }
  190. // Inspect retrieves the configuration properties of a managed swarm cluster.
  191. func (c *Cluster) Inspect() (types.Swarm, error) {
  192. var swarm types.Swarm
  193. if err := c.lockedManagerAction(func(ctx context.Context, state nodeState) error {
  194. s, err := c.inspect(ctx, state)
  195. if err != nil {
  196. return err
  197. }
  198. swarm = s
  199. return nil
  200. }); err != nil {
  201. return types.Swarm{}, err
  202. }
  203. return swarm, nil
  204. }
  205. func (c *Cluster) inspect(ctx context.Context, state nodeState) (types.Swarm, error) {
  206. s, err := getSwarm(ctx, state.controlClient)
  207. if err != nil {
  208. return types.Swarm{}, err
  209. }
  210. return convert.SwarmFromGRPC(*s), nil
  211. }
  212. // Update updates configuration of a managed swarm cluster.
  213. func (c *Cluster) Update(version uint64, spec types.Spec, flags types.UpdateFlags) error {
  214. return c.lockedManagerAction(func(ctx context.Context, state nodeState) error {
  215. swarm, err := getSwarm(ctx, state.controlClient)
  216. if err != nil {
  217. return err
  218. }
  219. // Validate spec name.
  220. if spec.Annotations.Name == "" {
  221. spec.Annotations.Name = "default"
  222. } else if spec.Annotations.Name != "default" {
  223. return errdefs.InvalidParameter(errors.New(`swarm spec must be named "default"`))
  224. }
  225. // In update, client should provide the complete spec of the swarm, including
  226. // Name and Labels. If a field is specified with 0 or nil, then the default value
  227. // will be used to swarmkit.
  228. clusterSpec, err := convert.SwarmSpecToGRPC(spec)
  229. if err != nil {
  230. return errdefs.InvalidParameter(err)
  231. }
  232. _, err = state.controlClient.UpdateCluster(
  233. ctx,
  234. &swarmapi.UpdateClusterRequest{
  235. ClusterID: swarm.ID,
  236. Spec: &clusterSpec,
  237. ClusterVersion: &swarmapi.Version{
  238. Index: version,
  239. },
  240. Rotation: swarmapi.KeyRotation{
  241. WorkerJoinToken: flags.RotateWorkerToken,
  242. ManagerJoinToken: flags.RotateManagerToken,
  243. ManagerUnlockKey: flags.RotateManagerUnlockKey,
  244. },
  245. },
  246. )
  247. return err
  248. })
  249. }
  250. // GetUnlockKey returns the unlock key for the swarm.
  251. func (c *Cluster) GetUnlockKey() (string, error) {
  252. var resp *swarmapi.GetUnlockKeyResponse
  253. if err := c.lockedManagerAction(func(ctx context.Context, state nodeState) error {
  254. client := swarmapi.NewCAClient(state.grpcConn)
  255. r, err := client.GetUnlockKey(ctx, &swarmapi.GetUnlockKeyRequest{})
  256. if err != nil {
  257. return err
  258. }
  259. resp = r
  260. return nil
  261. }); err != nil {
  262. return "", err
  263. }
  264. if len(resp.UnlockKey) == 0 {
  265. // no key
  266. return "", nil
  267. }
  268. return encryption.HumanReadableKey(resp.UnlockKey), nil
  269. }
  270. // UnlockSwarm provides a key to decrypt data that is encrypted at rest.
  271. func (c *Cluster) UnlockSwarm(req types.UnlockRequest) error {
  272. c.controlMutex.Lock()
  273. defer c.controlMutex.Unlock()
  274. c.mu.RLock()
  275. state := c.currentNodeState()
  276. if !state.IsActiveManager() {
  277. // when manager is not active,
  278. // unless it is locked, otherwise return error.
  279. if err := c.errNoManager(state); err != errSwarmLocked {
  280. c.mu.RUnlock()
  281. return err
  282. }
  283. } else {
  284. // when manager is active, return an error of "not locked"
  285. c.mu.RUnlock()
  286. return notLockedError{}
  287. }
  288. // only when swarm is locked, code running reaches here
  289. nr := c.nr
  290. c.mu.RUnlock()
  291. key, err := encryption.ParseHumanReadableKey(req.UnlockKey)
  292. if err != nil {
  293. return errdefs.InvalidParameter(err)
  294. }
  295. config := nr.config
  296. config.lockKey = key
  297. if err := nr.Stop(); err != nil {
  298. return err
  299. }
  300. nr, err = c.newNodeRunner(config)
  301. if err != nil {
  302. return err
  303. }
  304. c.mu.Lock()
  305. c.nr = nr
  306. c.mu.Unlock()
  307. if err := <-nr.Ready(); err != nil {
  308. if errors.Is(err, errSwarmLocked) {
  309. return invalidUnlockKey{}
  310. }
  311. return errors.Errorf("swarm component could not be started: %v", err)
  312. }
  313. return nil
  314. }
  315. // Leave shuts down Cluster and removes current state.
  316. func (c *Cluster) Leave(ctx context.Context, force bool) error {
  317. c.controlMutex.Lock()
  318. defer c.controlMutex.Unlock()
  319. c.mu.Lock()
  320. nr := c.nr
  321. if nr == nil {
  322. c.mu.Unlock()
  323. return errors.WithStack(errNoSwarm)
  324. }
  325. state := c.currentNodeState()
  326. c.mu.Unlock()
  327. if errors.Is(state.err, errSwarmLocked) && !force {
  328. // leave a locked swarm without --force is not allowed
  329. return errors.WithStack(notAvailableError("Swarm is encrypted and locked. Please unlock it first or use `--force` to ignore this message."))
  330. }
  331. if state.IsManager() && !force {
  332. msg := "You are attempting to leave the swarm on a node that is participating as a manager. "
  333. if state.IsActiveManager() {
  334. active, reachable, unreachable, err := managerStats(state.controlClient, state.NodeID())
  335. if err == nil {
  336. if active && removingManagerCausesLossOfQuorum(reachable, unreachable) {
  337. if isLastManager(reachable, unreachable) {
  338. msg += "Removing the last manager erases all current state of the swarm. Use `--force` to ignore this message. "
  339. return errors.WithStack(notAvailableError(msg))
  340. }
  341. msg += fmt.Sprintf("Removing this node leaves %v managers out of %v. Without a Raft quorum your swarm will be inaccessible. ", reachable-1, reachable+unreachable)
  342. }
  343. }
  344. } else {
  345. msg += "Doing so may lose the consensus of your cluster. "
  346. }
  347. msg += "The only way to restore a swarm that has lost consensus is to reinitialize it with `--force-new-cluster`. Use `--force` to suppress this message."
  348. return errors.WithStack(notAvailableError(msg))
  349. }
  350. // release readers in here
  351. if err := nr.Stop(); err != nil {
  352. log.G(ctx).Errorf("failed to shut down cluster node: %v", err)
  353. stack.Dump()
  354. return err
  355. }
  356. c.mu.Lock()
  357. c.nr = nil
  358. c.mu.Unlock()
  359. if nodeID := state.NodeID(); nodeID != "" {
  360. nodeContainers, err := c.listContainerForNode(ctx, nodeID)
  361. if err != nil {
  362. return err
  363. }
  364. for _, id := range nodeContainers {
  365. if err := c.config.Backend.ContainerRm(id, &apitypes.ContainerRmConfig{ForceRemove: true}); err != nil {
  366. log.G(ctx).Errorf("error removing %v: %v", id, err)
  367. }
  368. }
  369. }
  370. // todo: cleanup optional?
  371. if err := clearPersistentState(c.root); err != nil {
  372. return err
  373. }
  374. c.config.Backend.DaemonLeavesCluster()
  375. return nil
  376. }
  377. // Info returns information about the current cluster state.
  378. func (c *Cluster) Info() types.Info {
  379. info := types.Info{
  380. NodeAddr: c.GetAdvertiseAddress(),
  381. }
  382. c.mu.RLock()
  383. defer c.mu.RUnlock()
  384. state := c.currentNodeState()
  385. info.LocalNodeState = state.status
  386. if state.err != nil {
  387. info.Error = state.err.Error()
  388. }
  389. ctx, cancel := c.getRequestContext()
  390. defer cancel()
  391. if state.IsActiveManager() {
  392. info.ControlAvailable = true
  393. swarm, err := c.inspect(ctx, state)
  394. if err != nil {
  395. info.Error = err.Error()
  396. }
  397. info.Cluster = &swarm.ClusterInfo
  398. if r, err := state.controlClient.ListNodes(
  399. ctx, &swarmapi.ListNodesRequest{},
  400. grpc.MaxCallRecvMsgSize(defaultRecvSizeForListResponse),
  401. ); err != nil {
  402. info.Error = err.Error()
  403. } else {
  404. info.Nodes = len(r.Nodes)
  405. for _, n := range r.Nodes {
  406. if n.ManagerStatus != nil {
  407. info.Managers = info.Managers + 1
  408. }
  409. }
  410. }
  411. switch info.LocalNodeState {
  412. case types.LocalNodeStateInactive, types.LocalNodeStateLocked, types.LocalNodeStateError:
  413. // nothing to do
  414. default:
  415. if info.Managers == 2 {
  416. const warn string = `WARNING: Running Swarm in a two-manager configuration. This configuration provides
  417. no fault tolerance, and poses a high risk to lose control over the cluster.
  418. Refer to https://docs.docker.com/engine/swarm/admin_guide/ to configure the
  419. Swarm for fault-tolerance.`
  420. info.Warnings = append(info.Warnings, warn)
  421. }
  422. }
  423. }
  424. if state.swarmNode != nil {
  425. for _, r := range state.swarmNode.Remotes() {
  426. info.RemoteManagers = append(info.RemoteManagers, types.Peer{NodeID: r.NodeID, Addr: r.Addr})
  427. }
  428. info.NodeID = state.swarmNode.NodeID()
  429. }
  430. return info
  431. }
  432. // Status returns a textual representation of the node's swarm status and role (manager/worker)
  433. func (c *Cluster) Status() string {
  434. c.mu.RLock()
  435. s := c.currentNodeState()
  436. c.mu.RUnlock()
  437. state := string(s.status)
  438. if s.status == types.LocalNodeStateActive {
  439. if s.IsActiveManager() || s.IsManager() {
  440. state += "/manager"
  441. } else {
  442. state += "/worker"
  443. }
  444. }
  445. return state
  446. }
  447. func validateAndSanitizeInitRequest(req *types.InitRequest) error {
  448. var err error
  449. req.ListenAddr, err = validateAddr(req.ListenAddr)
  450. if err != nil {
  451. return fmt.Errorf("invalid ListenAddr %q: %v", req.ListenAddr, err)
  452. }
  453. if req.Spec.Annotations.Name == "" {
  454. req.Spec.Annotations.Name = "default"
  455. } else if req.Spec.Annotations.Name != "default" {
  456. return errors.New(`swarm spec must be named "default"`)
  457. }
  458. return nil
  459. }
  460. func validateAndSanitizeJoinRequest(req *types.JoinRequest) error {
  461. var err error
  462. req.ListenAddr, err = validateAddr(req.ListenAddr)
  463. if err != nil {
  464. return fmt.Errorf("invalid ListenAddr %q: %v", req.ListenAddr, err)
  465. }
  466. if len(req.RemoteAddrs) == 0 {
  467. return errors.New("at least 1 RemoteAddr is required to join")
  468. }
  469. for i := range req.RemoteAddrs {
  470. req.RemoteAddrs[i], err = validateAddr(req.RemoteAddrs[i])
  471. if err != nil {
  472. return fmt.Errorf("invalid remoteAddr %q: %v", req.RemoteAddrs[i], err)
  473. }
  474. }
  475. return nil
  476. }
  477. func validateAddr(addr string) (string, error) {
  478. if addr == "" {
  479. return addr, errors.New("invalid empty address")
  480. }
  481. newaddr, err := opts.ParseTCPAddr(addr, defaultAddr)
  482. if err != nil {
  483. // TODO(thaJeztah) why are we ignoring the error here? Is this to allow "non-tcp" addresses?
  484. return addr, nil
  485. }
  486. return strings.TrimPrefix(newaddr, "tcp://"), nil
  487. }
  488. func initClusterSpec(node *swarmnode.Node, spec types.Spec) error {
  489. ctx, cancel := context.WithTimeout(context.Background(), 5*time.Second)
  490. defer cancel()
  491. for conn := range node.ListenControlSocket(ctx) {
  492. if ctx.Err() != nil {
  493. return ctx.Err()
  494. }
  495. if conn != nil {
  496. client := swarmapi.NewControlClient(conn)
  497. var cluster *swarmapi.Cluster
  498. for i := 0; ; i++ {
  499. lcr, err := client.ListClusters(ctx, &swarmapi.ListClustersRequest{})
  500. if err != nil {
  501. return fmt.Errorf("error on listing clusters: %v", err)
  502. }
  503. if len(lcr.Clusters) == 0 {
  504. if i < 10 {
  505. time.Sleep(200 * time.Millisecond)
  506. continue
  507. }
  508. return errors.New("empty list of clusters was returned")
  509. }
  510. cluster = lcr.Clusters[0]
  511. break
  512. }
  513. // In init, we take the initial default values from swarmkit, and merge
  514. // any non nil or 0 value from spec to GRPC spec. This will leave the
  515. // default value alone.
  516. // Note that this is different from Update(), as in Update() we expect
  517. // user to specify the complete spec of the cluster (as they already know
  518. // the existing one and knows which field to update)
  519. clusterSpec, err := convert.MergeSwarmSpecToGRPC(spec, cluster.Spec)
  520. if err != nil {
  521. return fmt.Errorf("error updating cluster settings: %v", err)
  522. }
  523. _, err = client.UpdateCluster(ctx, &swarmapi.UpdateClusterRequest{
  524. ClusterID: cluster.ID,
  525. ClusterVersion: &cluster.Meta.Version,
  526. Spec: &clusterSpec,
  527. })
  528. if err != nil {
  529. return fmt.Errorf("error updating cluster settings: %v", err)
  530. }
  531. return nil
  532. }
  533. }
  534. return ctx.Err()
  535. }
  536. func (c *Cluster) listContainerForNode(ctx context.Context, nodeID string) ([]string, error) {
  537. var ids []string
  538. containers, err := c.config.Backend.Containers(ctx, &apitypes.ContainerListOptions{
  539. Filters: filters.NewArgs(filters.Arg("label", "com.docker.swarm.node.id="+nodeID)),
  540. })
  541. if err != nil {
  542. return []string{}, err
  543. }
  544. for _, c := range containers {
  545. ids = append(ids, c.ID)
  546. }
  547. return ids, nil
  548. }