diff --git a/api/client/node/accept.go b/api/client/node/accept.go deleted file mode 100644 index da9e8f14d5..0000000000 --- a/api/client/node/accept.go +++ /dev/null @@ -1,32 +0,0 @@ -package node - -import ( - "fmt" - - "github.com/docker/docker/api/client" - "github.com/docker/docker/cli" - "github.com/docker/engine-api/types/swarm" - "github.com/spf13/cobra" -) - -func newAcceptCommand(dockerCli *client.DockerCli) *cobra.Command { - return &cobra.Command{ - Use: "accept NODE [NODE...]", - Short: "Accept a node in the swarm", - Args: cli.RequiresMinArgs(1), - RunE: func(cmd *cobra.Command, args []string) error { - return runAccept(dockerCli, args) - }, - } -} - -func runAccept(dockerCli *client.DockerCli, nodes []string) error { - accept := func(node *swarm.Node) error { - node.Spec.Membership = swarm.NodeMembershipAccepted - return nil - } - success := func(nodeID string) { - fmt.Fprintf(dockerCli.Out(), "Node %s accepted in the swarm.\n", nodeID) - } - return updateNodes(dockerCli, nodes, accept, success) -} diff --git a/api/client/node/cmd.go b/api/client/node/cmd.go index 96c8e5bc09..f79ea0231c 100644 --- a/api/client/node/cmd.go +++ b/api/client/node/cmd.go @@ -23,7 +23,6 @@ func NewNodeCommand(dockerCli *client.DockerCli) *cobra.Command { }, } cmd.AddCommand( - newAcceptCommand(dockerCli), newDemoteCommand(dockerCli), newInspectCommand(dockerCli), newListCommand(dockerCli), diff --git a/api/client/node/list.go b/api/client/node/list.go index 1109e0ee28..55929adabe 100644 --- a/api/client/node/list.go +++ b/api/client/node/list.go @@ -16,7 +16,7 @@ import ( ) const ( - listItemFmt = "%s\t%s\t%s\t%s\t%s\t%s\n" + listItemFmt = "%s\t%s\t%s\t%s\t%s\n" ) type listOptions struct { @@ -74,11 +74,10 @@ func printTable(out io.Writer, nodes []swarm.Node, info types.Info) { // Ignore flushing errors defer writer.Flush() - fmt.Fprintf(writer, listItemFmt, "ID", "HOSTNAME", "MEMBERSHIP", "STATUS", "AVAILABILITY", "MANAGER STATUS") + fmt.Fprintf(writer, listItemFmt, "ID", "HOSTNAME", "STATUS", "AVAILABILITY", "MANAGER STATUS") for _, node := range nodes { name := node.Description.Hostname availability := string(node.Spec.Availability) - membership := string(node.Spec.Membership) reachability := "" if node.ManagerStatus != nil { @@ -99,7 +98,6 @@ func printTable(out io.Writer, nodes []swarm.Node, info types.Info) { listItemFmt, ID, name, - client.PrettyPrint(membership), client.PrettyPrint(string(node.Status.State)), client.PrettyPrint(availability), client.PrettyPrint(reachability)) diff --git a/api/client/node/opts.go b/api/client/node/opts.go index 381e1dfcea..f387bafc47 100644 --- a/api/client/node/opts.go +++ b/api/client/node/opts.go @@ -12,7 +12,6 @@ import ( type nodeOptions struct { annotations role string - membership string availability string } @@ -45,14 +44,6 @@ func (opts *nodeOptions) ToNodeSpec() (swarm.NodeSpec, error) { return swarm.NodeSpec{}, fmt.Errorf("invalid role %q, only worker and manager are supported", opts.role) } - switch swarm.NodeMembership(strings.ToLower(opts.membership)) { - case swarm.NodeMembershipAccepted: - spec.Membership = swarm.NodeMembershipAccepted - case "": - default: - return swarm.NodeSpec{}, fmt.Errorf("invalid membership %q, only accepted is supported", opts.membership) - } - switch swarm.NodeAvailability(strings.ToLower(opts.availability)) { case swarm.NodeAvailabilityActive: spec.Availability = swarm.NodeAvailabilityActive diff --git a/api/client/node/update.go b/api/client/node/update.go index 5cbdd908d8..1d070dd536 100644 --- a/api/client/node/update.go +++ b/api/client/node/update.go @@ -27,7 +27,6 @@ func newUpdateCommand(dockerCli *client.DockerCli) *cobra.Command { flags := cmd.Flags() flags.StringVar(&nodeOpts.role, flagRole, "", "Role of the node (worker/manager)") - flags.StringVar(&nodeOpts.membership, flagMembership, "", "Membership of the node (accepted/rejected)") flags.StringVar(&nodeOpts.availability, flagAvailability, "", "Availability of the node (active/pause/drain)") flags.Var(&nodeOpts.annotations.labels, flagLabelAdd, "Add or update a node label (key=value)") labelKeys := opts.NewListOpts(nil) @@ -76,13 +75,6 @@ func mergeNodeUpdate(flags *pflag.FlagSet) func(*swarm.Node) error { } spec.Role = swarm.NodeRole(str) } - if flags.Changed(flagMembership) { - str, err := flags.GetString(flagMembership) - if err != nil { - return err - } - spec.Membership = swarm.NodeMembership(str) - } if flags.Changed(flagAvailability) { str, err := flags.GetString(flagAvailability) if err != nil { @@ -115,7 +107,6 @@ func mergeNodeUpdate(flags *pflag.FlagSet) func(*swarm.Node) error { const ( flagRole = "role" - flagMembership = "membership" flagAvailability = "availability" flagLabelAdd = "label-add" flagLabelRemove = "label-rm" diff --git a/api/client/service/opts.go b/api/client/service/opts.go index 52cf2bb5d6..c516522cfb 100644 --- a/api/client/service/opts.go +++ b/api/client/service/opts.go @@ -506,7 +506,7 @@ func addServiceFlags(cmd *cobra.Command, opts *serviceOptions) { flags.StringVar(&opts.endpoint.mode, flagEndpointMode, "", "Endpoint mode (vip or dnsrr)") - flags.BoolVar(&opts.registryAuth, flagRegistryAuth, false, "Send registry authentication details to Swarm agents") + flags.BoolVar(&opts.registryAuth, flagRegistryAuth, false, "Send registry authentication details to swarm agents") flags.StringVar(&opts.logDriver.name, flagLogDriver, "", "Logging driver for service") flags.Var(&opts.logDriver.opts, flagLogOpt, "Logging driver options") diff --git a/api/client/swarm/cmd.go b/api/client/swarm/cmd.go index 03a4c0a1cb..8fcacf63f1 100644 --- a/api/client/swarm/cmd.go +++ b/api/client/swarm/cmd.go @@ -25,6 +25,7 @@ func NewSwarmCommand(dockerCli *client.DockerCli) *cobra.Command { newUpdateCommand(dockerCli), newLeaveCommand(dockerCli), newInspectCommand(dockerCli), + newJoinTokenCommand(dockerCli), ) return cmd } diff --git a/api/client/swarm/init.go b/api/client/swarm/init.go index c13154c296..cbd5325181 100644 --- a/api/client/swarm/init.go +++ b/api/client/swarm/init.go @@ -28,14 +28,11 @@ type initOptions struct { func newInitCommand(dockerCli *client.DockerCli) *cobra.Command { opts := initOptions{ listenAddr: NewListenAddrOption(), - swarmOptions: swarmOptions{ - autoAccept: NewAutoAcceptOption(), - }, } cmd := &cobra.Command{ Use: "init [OPTIONS]", - Short: "Initialize a Swarm", + Short: "Initialize a swarm", Args: cli.NoArgs, RunE: func(cmd *cobra.Command, args []string) error { return runInit(dockerCli, cmd.Flags(), opts) @@ -53,12 +50,6 @@ func runInit(dockerCli *client.DockerCli, flags *pflag.FlagSet, opts initOptions client := dockerCli.Client() ctx := context.Background() - // If no secret was specified, we create a random one - if !flags.Changed("secret") { - opts.secret = generateRandomSecret() - fmt.Fprintf(dockerCli.Out(), "No --secret provided. Generated random secret:\n %s\n\n", opts.secret) - } - req := swarm.InitRequest{ ListenAddr: opts.listenAddr.String(), ForceNewCluster: opts.forceNewCluster, @@ -72,24 +63,5 @@ func runInit(dockerCli *client.DockerCli, flags *pflag.FlagSet, opts initOptions fmt.Fprintf(dockerCli.Out(), "Swarm initialized: current node (%s) is now a manager.\n\n", nodeID) - // Fetch CAHash and Address from the API - info, err := client.Info(ctx) - if err != nil { - return err - } - - node, _, err := client.NodeInspectWithRaw(ctx, nodeID) - if err != nil { - return err - } - - if node.ManagerStatus != nil && info.Swarm.CACertHash != "" { - var secretArgs string - if opts.secret != "" { - secretArgs = "--secret " + opts.secret - } - fmt.Fprintf(dockerCli.Out(), "To add a worker to this swarm, run the following command:\n docker swarm join %s \\\n --ca-hash %s \\\n %s\n", secretArgs, info.Swarm.CACertHash, node.ManagerStatus.Addr) - } - - return nil + return printJoinCommand(ctx, dockerCli, nodeID, true, true) } diff --git a/api/client/swarm/inspect.go b/api/client/swarm/inspect.go index 6419ca74b7..03c3c12c12 100644 --- a/api/client/swarm/inspect.go +++ b/api/client/swarm/inspect.go @@ -18,7 +18,7 @@ func newInspectCommand(dockerCli *client.DockerCli) *cobra.Command { cmd := &cobra.Command{ Use: "inspect [OPTIONS]", - Short: "Inspect the Swarm", + Short: "Inspect the swarm", Args: cli.NoArgs, RunE: func(cmd *cobra.Command, args []string) error { return runInspect(dockerCli, opts) diff --git a/api/client/swarm/join.go b/api/client/swarm/join.go index 0d3cb286cb..63cf9d20c0 100644 --- a/api/client/swarm/join.go +++ b/api/client/swarm/join.go @@ -2,6 +2,7 @@ package swarm import ( "fmt" + "strings" "github.com/docker/docker/api/client" "github.com/docker/docker/cli" @@ -13,9 +14,7 @@ import ( type joinOptions struct { remote string listenAddr NodeAddrOption - manager bool - secret string - CACertHash string + token string } func newJoinCommand(dockerCli *client.DockerCli) *cobra.Command { @@ -25,7 +24,7 @@ func newJoinCommand(dockerCli *client.DockerCli) *cobra.Command { cmd := &cobra.Command{ Use: "join [OPTIONS] HOST:PORT", - Short: "Join a Swarm as a node and/or manager", + Short: "Join a swarm as a node and/or manager", Args: cli.ExactArgs(1), RunE: func(cmd *cobra.Command, args []string) error { opts.remote = args[0] @@ -35,9 +34,7 @@ func newJoinCommand(dockerCli *client.DockerCli) *cobra.Command { flags := cmd.Flags() flags.Var(&opts.listenAddr, flagListenAddr, "Listen address") - flags.BoolVar(&opts.manager, "manager", false, "Try joining as a manager.") - flags.StringVar(&opts.secret, flagSecret, "", "Secret for node acceptance") - flags.StringVar(&opts.CACertHash, "ca-hash", "", "Hash of the Root Certificate Authority certificate used for trusted join") + flags.StringVar(&opts.token, flagToken, "", "Token for entry into the swarm") return cmd } @@ -46,20 +43,29 @@ func runJoin(dockerCli *client.DockerCli, opts joinOptions) error { ctx := context.Background() req := swarm.JoinRequest{ - Manager: opts.manager, - Secret: opts.secret, + JoinToken: opts.token, ListenAddr: opts.listenAddr.String(), RemoteAddrs: []string{opts.remote}, - CACertHash: opts.CACertHash, } err := client.SwarmJoin(ctx, req) if err != nil { return err } - if opts.manager { - fmt.Fprintln(dockerCli.Out(), "This node joined a Swarm as a manager.") - } else { - fmt.Fprintln(dockerCli.Out(), "This node joined a Swarm as a worker.") + + info, err := client.Info(ctx) + if err != nil { + return err } + + _, _, err = client.NodeInspectWithRaw(ctx, info.Swarm.NodeID) + if err != nil { + // TODO(aaronl): is there a better way to do this? + if strings.Contains(err.Error(), "This node is not a swarm manager.") { + fmt.Fprintln(dockerCli.Out(), "This node joined a swarm as a worker.") + } + } else { + fmt.Fprintln(dockerCli.Out(), "This node joined a swarm as a manager.") + } + return nil } diff --git a/api/client/swarm/join_token.go b/api/client/swarm/join_token.go new file mode 100644 index 0000000000..8e6f3fa2de --- /dev/null +++ b/api/client/swarm/join_token.go @@ -0,0 +1,110 @@ +package swarm + +import ( + "errors" + "fmt" + + "github.com/spf13/cobra" + + "github.com/docker/docker/api/client" + "github.com/docker/docker/cli" + "github.com/docker/engine-api/types/swarm" + "golang.org/x/net/context" +) + +const ( + flagRotate = "rotate" + flagQuiet = "quiet" +) + +func newJoinTokenCommand(dockerCli *client.DockerCli) *cobra.Command { + var rotate, quiet bool + + cmd := &cobra.Command{ + Use: "join-token [-q] [--rotate] (worker|manager)", + Short: "Manage join tokens", + Args: cli.ExactArgs(1), + RunE: func(cmd *cobra.Command, args []string) error { + if args[0] != "worker" && args[0] != "manager" { + return errors.New("unknown role " + args[0]) + } + + client := dockerCli.Client() + ctx := context.Background() + + if rotate { + var flags swarm.UpdateFlags + + swarm, err := client.SwarmInspect(ctx) + if err != nil { + return err + } + + if args[0] == "worker" { + flags.RotateWorkerToken = true + } else if args[0] == "manager" { + flags.RotateManagerToken = true + } + + err = client.SwarmUpdate(ctx, swarm.Version, swarm.Spec, flags) + if err != nil { + return err + } + } + + swarm, err := client.SwarmInspect(ctx) + if err != nil { + return err + } + + if quiet { + if args[0] == "worker" { + fmt.Fprintln(dockerCli.Out(), swarm.JoinTokens.Worker) + } else if args[0] == "manager" { + fmt.Fprintln(dockerCli.Out(), swarm.JoinTokens.Manager) + } + } else { + info, err := client.Info(ctx) + if err != nil { + return err + } + return printJoinCommand(ctx, dockerCli, info.Swarm.NodeID, args[0] == "worker", args[0] == "manager") + } + return nil + }, + } + + flags := cmd.Flags() + flags.BoolVar(&rotate, flagRotate, false, "Rotate join token") + flags.BoolVarP(&quiet, flagQuiet, "q", false, "Only display token") + + return cmd +} + +func printJoinCommand(ctx context.Context, dockerCli *client.DockerCli, nodeID string, worker bool, manager bool) error { + client := dockerCli.Client() + + swarm, err := client.SwarmInspect(ctx) + if err != nil { + return err + } + + node, _, err := client.NodeInspectWithRaw(ctx, nodeID) + if err != nil { + return err + } + + if node.ManagerStatus != nil { + if worker { + fmt.Fprintf(dockerCli.Out(), "To add a worker to this swarm, run the following command:\n docker swarm join \\\n --token %s \\\n %s\n", swarm.JoinTokens.Worker, node.ManagerStatus.Addr) + } + if manager { + if worker { + fmt.Fprintln(dockerCli.Out()) + } + fmt.Fprintf(dockerCli.Out(), "To add a manager to this swarm, run the following command:\n docker swarm join \\\n --token %s \\\n %s\n", swarm.JoinTokens.Manager, node.ManagerStatus.Addr) + } + } + + return nil +} diff --git a/api/client/swarm/leave.go b/api/client/swarm/leave.go index b2c13599d4..acaa7c82c1 100644 --- a/api/client/swarm/leave.go +++ b/api/client/swarm/leave.go @@ -19,7 +19,7 @@ func newLeaveCommand(dockerCli *client.DockerCli) *cobra.Command { cmd := &cobra.Command{ Use: "leave [OPTIONS]", - Short: "Leave a Swarm", + Short: "Leave a swarm", Args: cli.NoArgs, RunE: func(cmd *cobra.Command, args []string) error { return runLeave(dockerCli, opts) diff --git a/api/client/swarm/opts.go b/api/client/swarm/opts.go index 92011fea23..3e78e4327c 100644 --- a/api/client/swarm/opts.go +++ b/api/client/swarm/opts.go @@ -15,29 +15,15 @@ import ( const ( defaultListenAddr = "0.0.0.0:2377" - worker = "WORKER" - manager = "MANAGER" - none = "NONE" - - flagAutoAccept = "auto-accept" flagCertExpiry = "cert-expiry" flagDispatcherHeartbeat = "dispatcher-heartbeat" flagListenAddr = "listen-addr" - flagSecret = "secret" + flagToken = "token" flagTaskHistoryLimit = "task-history-limit" flagExternalCA = "external-ca" ) -var ( - defaultPolicies = []swarm.Policy{ - {Role: worker, Autoaccept: true}, - {Role: manager, Autoaccept: false}, - } -) - type swarmOptions struct { - autoAccept AutoAcceptOption - secret string taskHistoryLimit int64 dispatcherHeartbeat time.Duration nodeCertExpiry time.Duration @@ -84,71 +70,6 @@ func NewListenAddrOption() NodeAddrOption { return NewNodeAddrOption(defaultListenAddr) } -// AutoAcceptOption is a value type for auto-accept policy -type AutoAcceptOption struct { - values map[string]struct{} -} - -// String prints a string representation of this option -func (o *AutoAcceptOption) String() string { - keys := []string{} - for key := range o.values { - keys = append(keys, fmt.Sprintf("%s=true", strings.ToLower(key))) - } - return strings.Join(keys, ", ") -} - -// Set sets a new value on this option -func (o *AutoAcceptOption) Set(acceptValues string) error { - for _, value := range strings.Split(acceptValues, ",") { - value = strings.ToUpper(value) - switch value { - case none, worker, manager: - o.values[value] = struct{}{} - default: - return fmt.Errorf("must be one / combination of %s, %s; or NONE", worker, manager) - } - } - // NONE must stand alone, so if any non-NONE setting exist with it, error with conflict - if o.isPresent(none) && len(o.values) > 1 { - return fmt.Errorf("value NONE cannot be specified alongside other node types") - } - return nil -} - -// Type returns the type of this option -func (o *AutoAcceptOption) Type() string { - return "auto-accept" -} - -// Policies returns a representation of this option for the api -func (o *AutoAcceptOption) Policies(secret *string) []swarm.Policy { - policies := []swarm.Policy{} - for _, p := range defaultPolicies { - if len(o.values) != 0 { - if _, ok := o.values[string(p.Role)]; ok { - p.Autoaccept = true - } else { - p.Autoaccept = false - } - } - p.Secret = secret - policies = append(policies, p) - } - return policies -} - -// isPresent returns whether the key exists in the set or not -func (o *AutoAcceptOption) isPresent(key string) bool { - _, c := o.values[key] - return c -} - -// NewAutoAcceptOption returns a new auto-accept option -func NewAutoAcceptOption() AutoAcceptOption { - return AutoAcceptOption{values: make(map[string]struct{})} -} - // ExternalCAOption is a Value type for parsing external CA specifications. type ExternalCAOption struct { values []*swarm.ExternalCA @@ -239,8 +160,6 @@ func parseExternalCA(caSpec string) (*swarm.ExternalCA, error) { } func addSwarmFlags(flags *pflag.FlagSet, opts *swarmOptions) { - flags.Var(&opts.autoAccept, flagAutoAccept, "Auto acceptance policy (worker, manager or none)") - flags.StringVar(&opts.secret, flagSecret, "", "Set secret value needed to join a cluster") flags.Int64Var(&opts.taskHistoryLimit, flagTaskHistoryLimit, 10, "Task history retention limit") flags.DurationVar(&opts.dispatcherHeartbeat, flagDispatcherHeartbeat, time.Duration(5*time.Second), "Dispatcher heartbeat period") flags.DurationVar(&opts.nodeCertExpiry, flagCertExpiry, time.Duration(90*24*time.Hour), "Validity period for node certificates") @@ -249,11 +168,6 @@ func addSwarmFlags(flags *pflag.FlagSet, opts *swarmOptions) { func (opts *swarmOptions) ToSpec() swarm.Spec { spec := swarm.Spec{} - if opts.secret != "" { - spec.AcceptancePolicy.Policies = opts.autoAccept.Policies(&opts.secret) - } else { - spec.AcceptancePolicy.Policies = opts.autoAccept.Policies(nil) - } spec.Orchestration.TaskHistoryRetentionLimit = opts.taskHistoryLimit spec.Dispatcher.HeartbeatPeriod = uint64(opts.dispatcherHeartbeat.Nanoseconds()) spec.CAConfig.NodeCertExpiry = opts.nodeCertExpiry diff --git a/api/client/swarm/opts_test.go b/api/client/swarm/opts_test.go index a6dcdb5254..568dc87302 100644 --- a/api/client/swarm/opts_test.go +++ b/api/client/swarm/opts_test.go @@ -4,7 +4,6 @@ import ( "testing" "github.com/docker/docker/pkg/testutil/assert" - "github.com/docker/engine-api/types/swarm" ) func TestNodeAddrOptionSetHostAndPort(t *testing.T) { @@ -36,101 +35,3 @@ func TestNodeAddrOptionSetInvalidFormat(t *testing.T) { opt := NewListenAddrOption() assert.Error(t, opt.Set("http://localhost:4545"), "Invalid") } - -func TestAutoAcceptOptionSetWorker(t *testing.T) { - opt := NewAutoAcceptOption() - assert.NilError(t, opt.Set("worker")) - assert.Equal(t, opt.isPresent(worker), true) -} - -func TestAutoAcceptOptionSetManager(t *testing.T) { - opt := NewAutoAcceptOption() - assert.NilError(t, opt.Set("manager")) - assert.Equal(t, opt.isPresent(manager), true) -} - -func TestAutoAcceptOptionSetInvalid(t *testing.T) { - opt := NewAutoAcceptOption() - assert.Error(t, opt.Set("bogus"), "must be one / combination") -} - -func TestAutoAcceptOptionSetEmpty(t *testing.T) { - opt := NewAutoAcceptOption() - assert.Error(t, opt.Set(""), "must be one / combination") -} - -func TestAutoAcceptOptionSetNone(t *testing.T) { - opt := NewAutoAcceptOption() - assert.NilError(t, opt.Set("none")) - assert.Equal(t, opt.isPresent(manager), false) - assert.Equal(t, opt.isPresent(worker), false) -} - -func TestAutoAcceptOptionSetTwo(t *testing.T) { - opt := NewAutoAcceptOption() - assert.NilError(t, opt.Set("worker,manager")) - assert.Equal(t, opt.isPresent(manager), true) - assert.Equal(t, opt.isPresent(worker), true) -} - -func TestAutoAcceptOptionSetConflict(t *testing.T) { - opt := NewAutoAcceptOption() - assert.Error(t, opt.Set("none,manager"), "value NONE cannot be specified alongside other node types") - - opt = NewAutoAcceptOption() - assert.Error(t, opt.Set("none,worker"), "value NONE cannot be specified alongside other node types") - - opt = NewAutoAcceptOption() - assert.Error(t, opt.Set("worker,none,manager"), "value NONE cannot be specified alongside other node types") - - opt = NewAutoAcceptOption() - assert.Error(t, opt.Set("worker,manager,none"), "value NONE cannot be specified alongside other node types") -} - -func TestAutoAcceptOptionPoliciesDefault(t *testing.T) { - opt := NewAutoAcceptOption() - secret := "thesecret" - - policies := opt.Policies(&secret) - assert.Equal(t, len(policies), 2) - assert.Equal(t, policies[0], swarm.Policy{ - Role: worker, - Autoaccept: true, - Secret: &secret, - }) - assert.Equal(t, policies[1], swarm.Policy{ - Role: manager, - Autoaccept: false, - Secret: &secret, - }) -} - -func TestAutoAcceptOptionPoliciesWithManager(t *testing.T) { - opt := NewAutoAcceptOption() - secret := "thesecret" - - assert.NilError(t, opt.Set("manager")) - - policies := opt.Policies(&secret) - assert.Equal(t, len(policies), 2) - assert.Equal(t, policies[0], swarm.Policy{ - Role: worker, - Autoaccept: false, - Secret: &secret, - }) - assert.Equal(t, policies[1], swarm.Policy{ - Role: manager, - Autoaccept: true, - Secret: &secret, - }) -} - -func TestAutoAcceptOptionString(t *testing.T) { - opt := NewAutoAcceptOption() - assert.NilError(t, opt.Set("manager")) - assert.NilError(t, opt.Set("worker")) - - repr := opt.String() - assert.Contains(t, repr, "worker=true") - assert.Contains(t, repr, "manager=true") -} diff --git a/api/client/swarm/secret.go b/api/client/swarm/secret.go deleted file mode 100644 index e98f1d6626..0000000000 --- a/api/client/swarm/secret.go +++ /dev/null @@ -1,19 +0,0 @@ -package swarm - -import ( - cryptorand "crypto/rand" - "fmt" - "math/big" -) - -func generateRandomSecret() string { - var secretBytes [generatedSecretEntropyBytes]byte - - if _, err := cryptorand.Read(secretBytes[:]); err != nil { - panic(fmt.Errorf("failed to read random bytes: %v", err)) - } - - var nn big.Int - nn.SetBytes(secretBytes[:]) - return fmt.Sprintf("%0[1]*s", maxGeneratedSecretLength, nn.Text(generatedSecretBase)) -} diff --git a/api/client/swarm/update.go b/api/client/swarm/update.go index 84963b92eb..05d4e91ee0 100644 --- a/api/client/swarm/update.go +++ b/api/client/swarm/update.go @@ -13,11 +13,11 @@ import ( ) func newUpdateCommand(dockerCli *client.DockerCli) *cobra.Command { - opts := swarmOptions{autoAccept: NewAutoAcceptOption()} + opts := swarmOptions{} cmd := &cobra.Command{ Use: "update [OPTIONS]", - Short: "Update the Swarm", + Short: "Update the swarm", Args: cli.NoArgs, RunE: func(cmd *cobra.Command, args []string) error { return runUpdate(dockerCli, cmd.Flags(), opts) @@ -32,6 +32,8 @@ func runUpdate(dockerCli *client.DockerCli, flags *pflag.FlagSet, opts swarmOpti client := dockerCli.Client() ctx := context.Background() + var updateFlags swarm.UpdateFlags + swarm, err := client.SwarmInspect(ctx) if err != nil { return err @@ -42,7 +44,7 @@ func runUpdate(dockerCli *client.DockerCli, flags *pflag.FlagSet, opts swarmOpti return err } - err = client.SwarmUpdate(ctx, swarm.Version, swarm.Spec) + err = client.SwarmUpdate(ctx, swarm.Version, swarm.Spec, updateFlags) if err != nil { return err } @@ -55,21 +57,6 @@ func runUpdate(dockerCli *client.DockerCli, flags *pflag.FlagSet, opts swarmOpti func mergeSwarm(swarm *swarm.Swarm, flags *pflag.FlagSet) error { spec := &swarm.Spec - if flags.Changed(flagAutoAccept) { - value := flags.Lookup(flagAutoAccept).Value.(*AutoAcceptOption) - spec.AcceptancePolicy.Policies = value.Policies(nil) - } - - var psecret *string - if flags.Changed(flagSecret) { - secret, _ := flags.GetString(flagSecret) - psecret = &secret - } - - for i := range spec.AcceptancePolicy.Policies { - spec.AcceptancePolicy.Policies[i].Secret = psecret - } - if flags.Changed(flagTaskHistoryLimit) { spec.Orchestration.TaskHistoryRetentionLimit, _ = flags.GetInt64(flagTaskHistoryLimit) } diff --git a/api/client/system/info.go b/api/client/system/info.go index 20a7ba9067..b9d730d9f4 100644 --- a/api/client/system/info.go +++ b/api/client/system/info.go @@ -85,7 +85,6 @@ func runInfo(dockerCli *client.DockerCli) error { if info.Swarm.ControlAvailable { fmt.Fprintf(dockerCli.Out(), " Managers: %d\n", info.Swarm.Managers) fmt.Fprintf(dockerCli.Out(), " Nodes: %d\n", info.Swarm.Nodes) - ioutils.FprintfIfNotEmpty(dockerCli.Out(), " CA Certificate Hash: %s\n", info.Swarm.CACertHash) } } diff --git a/api/server/router/swarm/backend.go b/api/server/router/swarm/backend.go index f9ee0a1bbe..36ede04b84 100644 --- a/api/server/router/swarm/backend.go +++ b/api/server/router/swarm/backend.go @@ -11,7 +11,7 @@ type Backend interface { Join(req types.JoinRequest) error Leave(force bool) error Inspect() (types.Swarm, error) - Update(uint64, types.Spec) error + Update(uint64, types.Spec, types.UpdateFlags) error GetServices(basictypes.ServiceListOptions) ([]types.Service, error) GetService(string) (types.Service, error) CreateService(types.ServiceSpec, string) (string, error) diff --git a/api/server/router/swarm/cluster_routes.go b/api/server/router/swarm/cluster_routes.go index d91b4a972e..944616bf4b 100644 --- a/api/server/router/swarm/cluster_routes.go +++ b/api/server/router/swarm/cluster_routes.go @@ -66,7 +66,15 @@ func (sr *swarmRouter) updateCluster(ctx context.Context, w http.ResponseWriter, return fmt.Errorf("Invalid swarm version '%s': %s", rawVersion, err.Error()) } - if err := sr.backend.Update(version, swarm); err != nil { + var flags types.UpdateFlags + if r.URL.Query().Get("rotate_worker_token") == "true" { + flags.RotateWorkerToken = true + } + if r.URL.Query().Get("rotate_manager_token") == "true" { + flags.RotateManagerToken = true + } + + if err := sr.backend.Update(version, swarm, flags); err != nil { logrus.Errorf("Error configuring swarm: %v", err) return err } diff --git a/contrib/completion/zsh/_docker b/contrib/completion/zsh/_docker index 1d71374a32..724d869837 100644 --- a/contrib/completion/zsh/_docker +++ b/contrib/completion/zsh/_docker @@ -1087,7 +1087,7 @@ __docker_service_subcommand() { "($help)--name=[Service name]:name: " "($help)*--network=[Network attachments]:network: " "($help)*"{-p=,--publish=}"[Publish a port as a node port]:port: " - "($help)--registry-auth[Send registry authentication details to Swarm agents]" + "($help)--registry-auth[Send registry authentication details to swarm agents]" "($help)--replicas=[Number of tasks]:replicas: " "($help)--reserve-cpu=[Reserve CPUs]:value: " "($help)--reserve-memory=[Reserve Memory]:value: " @@ -1185,11 +1185,11 @@ __docker_service_subcommand() { __docker_swarm_commands() { local -a _docker_swarm_subcommands _docker_swarm_subcommands=( - "init:Initialize a Swarm" - "inspect:Inspect the Swarm" - "join:Join a Swarm as a node and/or manager" - "leave:Leave a Swarm" - "update:Update the Swarm" + "init:Initialize a swarm" + "inspect:Inspect the swarm" + "join:Join a swarm as a node and/or manager" + "leave:Leave a swarm" + "update:Update the swarm" ) _describe -t docker-swarm-commands "docker swarm command" _docker_swarm_subcommands } diff --git a/daemon/cluster/cluster.go b/daemon/cluster/cluster.go index 2385c92007..71d782902a 100644 --- a/daemon/cluster/cluster.go +++ b/daemon/cluster/cluster.go @@ -13,7 +13,6 @@ import ( "google.golang.org/grpc" "github.com/Sirupsen/logrus" - "github.com/docker/distribution/digest" "github.com/docker/docker/daemon/cluster/convert" executorpkg "github.com/docker/docker/daemon/cluster/executor" "github.com/docker/docker/daemon/cluster/executor/container" @@ -42,16 +41,16 @@ const ( ) // ErrNoSwarm is returned on leaving a cluster that was never initialized -var ErrNoSwarm = fmt.Errorf("This node is not part of Swarm") +var ErrNoSwarm = fmt.Errorf("This node is not part of swarm") // ErrSwarmExists is returned on initialize or join request for a cluster that has already been activated -var ErrSwarmExists = fmt.Errorf("This node is already part of a Swarm cluster. Use \"docker swarm leave\" to leave this cluster and join another one.") +var ErrSwarmExists = fmt.Errorf("This node is already part of a swarm cluster. Use \"docker swarm leave\" to leave this cluster and join another one.") // ErrPendingSwarmExists is returned on initialize or join request for a cluster that is already processing a similar request but has not succeeded yet. var ErrPendingSwarmExists = fmt.Errorf("This node is processing an existing join request that has not succeeded yet. Use \"docker swarm leave\" to cancel the current request.") // ErrSwarmJoinTimeoutReached is returned when cluster join could not complete before timeout was reached. -var ErrSwarmJoinTimeoutReached = fmt.Errorf("Timeout was reached before node was joined. Attempt to join the cluster will continue in the background. Use \"docker info\" command to see the current Swarm status of your node.") +var ErrSwarmJoinTimeoutReached = fmt.Errorf("Timeout was reached before node was joined. Attempt to join the cluster will continue in the background. Use \"docker info\" command to see the current swarm status of your node.") // defaultSpec contains some sane defaults if cluster options are missing on init var defaultSpec = types.Spec{ @@ -127,7 +126,7 @@ func New(config Config) (*Cluster, error) { return nil, err } - n, err := c.startNewNode(false, st.ListenAddr, "", "", "", false) + n, err := c.startNewNode(false, st.ListenAddr, "", "") if err != nil { return nil, err } @@ -196,7 +195,7 @@ func (c *Cluster) reconnectOnFailure(n *node) { return } var err error - n, err = c.startNewNode(false, c.listenAddr, c.getRemoteAddress(), "", "", false) + n, err = c.startNewNode(false, c.listenAddr, c.getRemoteAddress(), "") if err != nil { c.err = err close(n.done) @@ -205,7 +204,7 @@ func (c *Cluster) reconnectOnFailure(n *node) { } } -func (c *Cluster) startNewNode(forceNewCluster bool, listenAddr, joinAddr, secret, cahash string, ismanager bool) (*node, error) { +func (c *Cluster) startNewNode(forceNewCluster bool, listenAddr, joinAddr, joinToken string) (*node, error) { if err := c.config.Backend.IsSwarmCompatible(); err != nil { return nil, err } @@ -219,12 +218,10 @@ func (c *Cluster) startNewNode(forceNewCluster bool, listenAddr, joinAddr, secre ListenRemoteAPI: listenAddr, JoinAddr: joinAddr, StateDir: c.root, - CAHash: cahash, - Secret: secret, + JoinToken: joinToken, Executor: container.NewExecutor(c.config.Backend), HeartbeatTick: 1, ElectionTick: 3, - IsManager: ismanager, }) if err != nil { return nil, err @@ -291,7 +288,7 @@ func (c *Cluster) Init(req types.InitRequest) (string, error) { if node := c.node; node != nil { if !req.ForceNewCluster { c.Unlock() - return "", errSwarmExists(node) + return "", ErrSwarmExists } if err := c.stopNode(); err != nil { c.Unlock() @@ -305,7 +302,7 @@ func (c *Cluster) Init(req types.InitRequest) (string, error) { } // todo: check current state existing - n, err := c.startNewNode(req.ForceNewCluster, req.ListenAddr, "", "", "", false) + n, err := c.startNewNode(req.ForceNewCluster, req.ListenAddr, "", "") if err != nil { c.Unlock() return "", err @@ -336,40 +333,32 @@ func (c *Cluster) Join(req types.JoinRequest) error { c.Lock() if node := c.node; node != nil { c.Unlock() - return errSwarmExists(node) + return ErrSwarmExists } if err := validateAndSanitizeJoinRequest(&req); err != nil { c.Unlock() return err } // todo: check current state existing - n, err := c.startNewNode(false, req.ListenAddr, req.RemoteAddrs[0], req.Secret, req.CACertHash, req.Manager) + n, err := c.startNewNode(false, req.ListenAddr, req.RemoteAddrs[0], req.JoinToken) if err != nil { c.Unlock() return err } c.Unlock() - certificateRequested := n.CertificateRequested() - for { - select { - case <-certificateRequested: - if n.NodeMembership() == swarmapi.NodeMembershipPending { - return fmt.Errorf("Your node is in the process of joining the cluster but needs to be accepted by existing cluster member.\nTo accept this node into cluster run \"docker node accept %v\" in an existing cluster manager. Use \"docker info\" command to see the current Swarm status of your node.", n.NodeID()) - } - certificateRequested = nil - case <-time.After(swarmConnectTimeout): - // attempt to connect will continue in background, also reconnecting - go c.reconnectOnFailure(n) - return ErrSwarmJoinTimeoutReached - case <-n.Ready(): - go c.reconnectOnFailure(n) - return nil - case <-n.done: - c.RLock() - defer c.RUnlock() - return c.err - } + select { + case <-time.After(swarmConnectTimeout): + // attempt to connect will continue in background, also reconnecting + go c.reconnectOnFailure(n) + return ErrSwarmJoinTimeoutReached + case <-n.Ready(): + go c.reconnectOnFailure(n) + return nil + case <-n.done: + c.RLock() + defer c.RUnlock() + return c.err } } @@ -489,7 +478,7 @@ func (c *Cluster) Inspect() (types.Swarm, error) { } // Update updates configuration of a managed swarm cluster. -func (c *Cluster) Update(version uint64, spec types.Spec) error { +func (c *Cluster) Update(version uint64, spec types.Spec, flags types.UpdateFlags) error { c.RLock() defer c.RUnlock() @@ -505,7 +494,7 @@ func (c *Cluster) Update(version uint64, spec types.Spec) error { return err } - swarmSpec, err := convert.SwarmSpecToGRPCandMerge(spec, &swarm.Spec) + swarmSpec, err := convert.SwarmSpecToGRPC(spec) if err != nil { return err } @@ -518,6 +507,10 @@ func (c *Cluster) Update(version uint64, spec types.Spec) error { ClusterVersion: &swarmapi.Version{ Index: version, }, + Rotation: swarmapi.JoinTokenRotation{ + RotateWorkerToken: flags.RotateWorkerToken, + RotateManagerToken: flags.RotateManagerToken, + }, }, ) return err @@ -611,10 +604,6 @@ func (c *Cluster) Info() types.Info { } } } - - if swarm, err := getSwarm(ctx, c.client); err == nil && swarm != nil { - info.CACertHash = swarm.RootCA.CACertHash - } } if c.node != nil { @@ -636,12 +625,12 @@ func (c *Cluster) isActiveManager() bool { // Call with read lock. func (c *Cluster) errNoManager() error { if c.node == nil { - return fmt.Errorf("This node is not a Swarm manager. Use \"docker swarm init\" or \"docker swarm join --manager\" to connect this node to Swarm and try again.") + return fmt.Errorf("This node is not a swarm manager. Use \"docker swarm init\" or \"docker swarm join --manager\" to connect this node to swarm and try again.") } if c.node.Manager() != nil { - return fmt.Errorf("This node is not a Swarm manager. Manager is being prepared or has trouble connecting to the cluster.") + return fmt.Errorf("This node is not a swarm manager. Manager is being prepared or has trouble connecting to the cluster.") } - return fmt.Errorf("This node is not a Swarm manager. Worker nodes can't be used to view or modify cluster state. Please run this command on a manager node or promote the current node to a manager.") + return fmt.Errorf("This node is not a swarm manager. Worker nodes can't be used to view or modify cluster state. Please run this command on a manager node or promote the current node to a manager.") } // GetServices returns all services of a managed swarm cluster. @@ -1219,11 +1208,6 @@ func validateAndSanitizeJoinRequest(req *types.JoinRequest) error { return fmt.Errorf("invalid remoteAddr %q: %v", req.RemoteAddrs[i], err) } } - if req.CACertHash != "" { - if _, err := digest.ParseDigest(req.CACertHash); err != nil { - return fmt.Errorf("invalid CACertHash %q, %v", req.CACertHash, err) - } - } return nil } @@ -1238,13 +1222,6 @@ func validateAddr(addr string) (string, error) { return strings.TrimPrefix(newaddr, "tcp://"), nil } -func errSwarmExists(node *node) error { - if node.NodeMembership() != swarmapi.NodeMembershipAccepted { - return ErrPendingSwarmExists - } - return ErrSwarmExists -} - func initClusterSpec(node *node, spec types.Spec) error { ctx, _ := context.WithTimeout(context.Background(), 5*time.Second) for conn := range node.ListenControlSocket(ctx) { @@ -1269,7 +1246,7 @@ func initClusterSpec(node *node, spec types.Spec) error { cluster = lcr.Clusters[0] break } - newspec, err := convert.SwarmSpecToGRPCandMerge(spec, &cluster.Spec) + newspec, err := convert.SwarmSpecToGRPC(spec) if err != nil { return fmt.Errorf("error updating cluster settings: %v", err) } diff --git a/daemon/cluster/convert/node.go b/daemon/cluster/convert/node.go index 04c40ff112..53d7efa428 100644 --- a/daemon/cluster/convert/node.go +++ b/daemon/cluster/convert/node.go @@ -15,7 +15,6 @@ func NodeFromGRPC(n swarmapi.Node) types.Node { ID: n.ID, Spec: types.NodeSpec{ Role: types.NodeRole(strings.ToLower(n.Spec.Role.String())), - Membership: types.NodeMembership(strings.ToLower(n.Spec.Membership.String())), Availability: types.NodeAvailability(strings.ToLower(n.Spec.Availability.String())), }, Status: types.NodeStatus{ @@ -79,12 +78,6 @@ func NodeSpecToGRPC(s types.NodeSpec) (swarmapi.NodeSpec, error) { return swarmapi.NodeSpec{}, fmt.Errorf("invalid Role: %q", s.Role) } - if membership, ok := swarmapi.NodeSpec_Membership_value[strings.ToUpper(string(s.Membership))]; ok { - spec.Membership = swarmapi.NodeSpec_Membership(membership) - } else { - return swarmapi.NodeSpec{}, fmt.Errorf("invalid Membership: %q", s.Membership) - } - if availability, ok := swarmapi.NodeSpec_Availability_value[strings.ToUpper(string(s.Availability))]; ok { spec.Availability = swarmapi.NodeSpec_Availability(availability) } else { diff --git a/daemon/cluster/convert/swarm.go b/daemon/cluster/convert/swarm.go index 331c199c84..a22b02acaa 100644 --- a/daemon/cluster/convert/swarm.go +++ b/daemon/cluster/convert/swarm.go @@ -5,8 +5,6 @@ import ( "strings" "time" - "golang.org/x/crypto/bcrypt" - types "github.com/docker/engine-api/types/swarm" swarmapi "github.com/docker/swarmkit/api" "github.com/docker/swarmkit/protobuf/ptypes" @@ -28,6 +26,10 @@ func SwarmFromGRPC(c swarmapi.Cluster) types.Swarm { ElectionTick: c.Spec.Raft.ElectionTick, }, }, + JoinTokens: types.JoinTokens{ + Worker: c.RootCA.JoinTokens.Worker, + Manager: c.RootCA.JoinTokens.Manager, + }, } heartbeatPeriod, _ := ptypes.Duration(c.Spec.Dispatcher.HeartbeatPeriod) @@ -52,23 +54,11 @@ func SwarmFromGRPC(c swarmapi.Cluster) types.Swarm { swarm.Spec.Name = c.Spec.Annotations.Name swarm.Spec.Labels = c.Spec.Annotations.Labels - for _, policy := range c.Spec.AcceptancePolicy.Policies { - p := types.Policy{ - Role: types.NodeRole(strings.ToLower(policy.Role.String())), - Autoaccept: policy.Autoaccept, - } - if policy.Secret != nil { - secret := string(policy.Secret.Data) - p.Secret = &secret - } - swarm.Spec.AcceptancePolicy.Policies = append(swarm.Spec.AcceptancePolicy.Policies, p) - } - return swarm } -// SwarmSpecToGRPCandMerge converts a Spec to a grpc ClusterSpec and merge AcceptancePolicy from an existing grpc ClusterSpec if provided. -func SwarmSpecToGRPCandMerge(s types.Spec, existingSpec *swarmapi.ClusterSpec) (swarmapi.ClusterSpec, error) { +// SwarmSpecToGRPC converts a Spec to a grpc ClusterSpec. +func SwarmSpecToGRPC(s types.Spec) (swarmapi.ClusterSpec, error) { spec := swarmapi.ClusterSpec{ Annotations: swarmapi.Annotations{ Name: s.Name, @@ -104,63 +94,5 @@ func SwarmSpecToGRPCandMerge(s types.Spec, existingSpec *swarmapi.ClusterSpec) ( }) } - if err := SwarmSpecUpdateAcceptancePolicy(&spec, s.AcceptancePolicy, existingSpec); err != nil { - return swarmapi.ClusterSpec{}, err - } - return spec, nil } - -// SwarmSpecUpdateAcceptancePolicy updates a grpc ClusterSpec using AcceptancePolicy. -func SwarmSpecUpdateAcceptancePolicy(spec *swarmapi.ClusterSpec, acceptancePolicy types.AcceptancePolicy, oldSpec *swarmapi.ClusterSpec) error { - spec.AcceptancePolicy.Policies = nil - hashs := make(map[string][]byte) - - for _, p := range acceptancePolicy.Policies { - role, ok := swarmapi.NodeRole_value[strings.ToUpper(string(p.Role))] - if !ok { - return fmt.Errorf("invalid Role: %q", p.Role) - } - - policy := &swarmapi.AcceptancePolicy_RoleAdmissionPolicy{ - Role: swarmapi.NodeRole(role), - Autoaccept: p.Autoaccept, - } - - if p.Secret != nil { - if *p.Secret == "" { // if provided secret is empty, it means erase previous secret. - policy.Secret = nil - } else { // if provided secret is not empty, we generate a new one. - hashPwd, ok := hashs[*p.Secret] - if !ok { - hashPwd, _ = bcrypt.GenerateFromPassword([]byte(*p.Secret), 0) - hashs[*p.Secret] = hashPwd - } - policy.Secret = &swarmapi.AcceptancePolicy_RoleAdmissionPolicy_Secret{ - Data: hashPwd, - Alg: "bcrypt", - } - } - } else if oldSecret := getOldSecret(oldSpec, policy.Role); oldSecret != nil { // else use the old one. - policy.Secret = &swarmapi.AcceptancePolicy_RoleAdmissionPolicy_Secret{ - Data: oldSecret.Data, - Alg: oldSecret.Alg, - } - } - - spec.AcceptancePolicy.Policies = append(spec.AcceptancePolicy.Policies, policy) - } - return nil -} - -func getOldSecret(oldSpec *swarmapi.ClusterSpec, role swarmapi.NodeRole) *swarmapi.AcceptancePolicy_RoleAdmissionPolicy_Secret { - if oldSpec == nil { - return nil - } - for _, p := range oldSpec.AcceptancePolicy.Policies { - if p.Role == role { - return p.Secret - } - } - return nil -} diff --git a/docs/reference/api/docker_remote_api_v1.24.md b/docs/reference/api/docker_remote_api_v1.24.md index 37e7965d1f..ab18471784 100644 --- a/docs/reference/api/docker_remote_api_v1.24.md +++ b/docs/reference/api/docker_remote_api_v1.24.md @@ -3351,7 +3351,6 @@ List nodes "UpdatedAt": "2016-06-07T20:31:11.999868824Z", "Spec": { "Role": "MANAGER", - "Membership": "ACCEPTED", "Availability": "ACTIVE" }, "Description": { @@ -3481,7 +3480,6 @@ Return low-level information on the node `id` "UpdatedAt": "2016-06-07T20:31:11.999868824Z", "Spec": { "Role": "MANAGER", - "Membership": "ACCEPTED", "Availability": "ACTIVE" }, "Description": { @@ -3595,18 +3593,6 @@ Initialize a new Swarm "ListenAddr": "0.0.0.0:4500", "ForceNewCluster": false, "Spec": { - "AcceptancePolicy": { - "Policies": [ - { - "Role": "MANAGER", - "Autoaccept": false - }, - { - "Role": "WORKER", - "Autoaccept": true - } - ] - }, "Orchestration": {}, "Raft": {}, "Dispatcher": {}, @@ -3676,9 +3662,7 @@ Join an existing new Swarm { "ListenAddr": "0.0.0.0:4500", "RemoteAddrs": ["node1:4500"], - "Secret": "", - "CACertHash": "", - "Manager": false + "JoinToken": "SWMTKN-1-3pu6hszjas19xyp7ghgosyx9k8atbfcr8p2is99znpy26u2lkl-7p73s1dx5in4tatdymyhg9hu2" } **Example response**: @@ -3698,9 +3682,7 @@ JSON Parameters: - **ListenAddr** – Listen address used for inter-manager communication if the node gets promoted to manager, as well as determining the networking interface used for the VXLAN Tunnel Endpoint (VTEP). - **RemoteAddr** – Address of any manager node already participating in the Swarm to join. -- **Secret** – Secret token for joining this Swarm. -- **CACertHash** – Optional hash of the root CA to avoid relying on trust on first use. -- **Manager** – Directly join as a manager (only for a Swarm configured to autoaccept managers). +- **JoinToken** – Secret token for joining this Swarm. ### Leave a Swarm @@ -3741,18 +3723,6 @@ Update a Swarm { "Name": "default", - "AcceptancePolicy": { - "Policies": [ - { - "Role": "WORKER", - "Autoaccept": false - }, - { - "Role": "MANAGER", - "Autoaccept": false - } - ] - }, "Orchestration": { "TaskHistoryRetentionLimit": 10 }, @@ -3767,6 +3737,10 @@ Update a Swarm }, "CAConfig": { "NodeCertExpiry": 7776000000000000 + }, + "JoinTokens": { + "Worker": "SWMTKN-1-3pu6hszjas19xyp7ghgosyx9k8atbfcr8p2is99znpy26u2lkl-1awxwuwd3z9j1z3puu7rcgdbx", + "Manager": "SWMTKN-1-3pu6hszjas19xyp7ghgosyx9k8atbfcr8p2is99znpy26u2lkl-7p73s1dx5in4tatdymyhg9hu2" } } @@ -3777,6 +3751,13 @@ Update a Swarm Content-Length: 0 Content-Type: text/plain; charset=utf-8 +**Query parameters**: + +- **version** – The version number of the swarm object being updated. This is + required to avoid conflicting writes. +- **rotate_worker_token** - Set to `true` to rotate the worker join token. +- **rotate_manager_token** - Set to `true` to rotate the manager join token. + **Status codes**: - **200** – no error @@ -3785,11 +3766,6 @@ Update a Swarm JSON Parameters: -- **Policies** – An array of acceptance policies. - - **Role** – The role that policy applies to (`MANAGER` or `WORKER`) - - **Autoaccept** – A boolean indicating whether nodes joining for that role should be - automatically accepted in the Swarm. - - **Secret** – An optional secret to provide for nodes to join the Swarm. - **Orchestration** – Configuration settings for the orchestration aspects of the Swarm. - **TaskHistoryRetentionLimit** – Maximum number of tasks history stored. - **Raft** – Raft related configuration. @@ -3811,6 +3787,9 @@ JSON Parameters: - **URL** - URL where certificate signing requests should be sent. - **Options** - An object with key/value pairs that are interpreted as protocol-specific options for the external CA driver. +- **JoinTokens** - Tokens that can be used by other nodes to join the Swarm. + - **Worker** - Token to use for joining as a worker. + - **Manager** - Token to use for joining as a manager. ## 3.8 Services @@ -4292,6 +4271,10 @@ Update the service `id`. of: `"Ports": { "/: {}" }` - **VirtualIPs** +**Query parameters**: + +- **version** – The version number of the service object being updated. This is + required to avoid conflicting writes. **Status codes**: diff --git a/docs/reference/api/docker_remote_api_v1.25.md b/docs/reference/api/docker_remote_api_v1.25.md index 84e88c6a9e..56b81074f6 100644 --- a/docs/reference/api/docker_remote_api_v1.25.md +++ b/docs/reference/api/docker_remote_api_v1.25.md @@ -3352,7 +3352,6 @@ List nodes "UpdatedAt": "2016-06-07T20:31:11.999868824Z", "Spec": { "Role": "MANAGER", - "Membership": "ACCEPTED", "Availability": "ACTIVE" }, "Description": { @@ -3482,7 +3481,6 @@ Return low-level information on the node `id` "UpdatedAt": "2016-06-07T20:31:11.999868824Z", "Spec": { "Role": "MANAGER", - "Membership": "ACCEPTED", "Availability": "ACTIVE" }, "Description": { @@ -3596,18 +3594,6 @@ Initialize a new Swarm "ListenAddr": "0.0.0.0:4500", "ForceNewCluster": false, "Spec": { - "AcceptancePolicy": { - "Policies": [ - { - "Role": "MANAGER", - "Autoaccept": false - }, - { - "Role": "WORKER", - "Autoaccept": true - } - ] - }, "Orchestration": {}, "Raft": {}, "Dispatcher": {}, @@ -3677,9 +3663,7 @@ Join an existing new Swarm { "ListenAddr": "0.0.0.0:4500", "RemoteAddrs": ["node1:4500"], - "Secret": "", - "CACertHash": "", - "Manager": false + "JoinToken": "SWMTKN-1-3pu6hszjas19xyp7ghgosyx9k8atbfcr8p2is99znpy26u2lkl-7p73s1dx5in4tatdymyhg9hu2" } **Example response**: @@ -3699,9 +3683,7 @@ JSON Parameters: - **ListenAddr** – Listen address used for inter-manager communication if the node gets promoted to manager, as well as determining the networking interface used for the VXLAN Tunnel Endpoint (VTEP). - **RemoteAddr** – Address of any manager node already participating in the Swarm to join. -- **Secret** – Secret token for joining this Swarm. -- **CACertHash** – Optional hash of the root CA to avoid relying on trust on first use. -- **Manager** – Directly join as a manager (only for a Swarm configured to autoaccept managers). +- **JoinToken** – Secret token for joining this Swarm. ### Leave a Swarm @@ -3742,18 +3724,6 @@ Update a Swarm { "Name": "default", - "AcceptancePolicy": { - "Policies": [ - { - "Role": "WORKER", - "Autoaccept": false - }, - { - "Role": "MANAGER", - "Autoaccept": false - } - ] - }, "Orchestration": { "TaskHistoryRetentionLimit": 10 }, @@ -3768,6 +3738,10 @@ Update a Swarm }, "CAConfig": { "NodeCertExpiry": 7776000000000000 + }, + "JoinTokens": { + "Worker": "SWMTKN-1-3pu6hszjas19xyp7ghgosyx9k8atbfcr8p2is99znpy26u2lkl-1awxwuwd3z9j1z3puu7rcgdbx", + "Manager": "SWMTKN-1-3pu6hszjas19xyp7ghgosyx9k8atbfcr8p2is99znpy26u2lkl-7p73s1dx5in4tatdymyhg9hu2" } } @@ -3778,6 +3752,13 @@ Update a Swarm Content-Length: 0 Content-Type: text/plain; charset=utf-8 +**Query parameters**: + +- **version** – The version number of the swarm object being updated. This is + required to avoid conflicting writes. +- **rotate_worker_token** - Set to `true` to rotate the worker join token. +- **rotate_manager_token** - Set to `true` to rotate the manager join token. + **Status codes**: - **200** – no error @@ -3786,11 +3767,6 @@ Update a Swarm JSON Parameters: -- **Policies** – An array of acceptance policies. - - **Role** – The role that policy applies to (`MANAGER` or `WORKER`) - - **Autoaccept** – A boolean indicating whether nodes joining for that role should be - automatically accepted in the Swarm. - - **Secret** – An optional secret to provide for nodes to join the Swarm. - **Orchestration** – Configuration settings for the orchestration aspects of the Swarm. - **TaskHistoryRetentionLimit** – Maximum number of tasks history stored. - **Raft** – Raft related configuration. @@ -3812,6 +3788,9 @@ JSON Parameters: - **URL** - URL where certificate signing requests should be sent. - **Options** - An object with key/value pairs that are interpreted as protocol-specific options for the external CA driver. +- **JoinTokens** - Tokens that can be used by other nodes to join the Swarm. + - **Worker** - Token to use for joining as a worker. + - **Manager** - Token to use for joining as a manager. ## 3.8 Services @@ -4293,6 +4272,10 @@ Update the service `id`. of: `"Ports": { "/: {}" }` - **VirtualIPs** +**Query parameters**: + +- **version** – The version number of the service object being updated. This is + required to avoid conflicting writes. **Status codes**: diff --git a/docs/reference/commandline/deploy.md b/docs/reference/commandline/deploy.md index 9c9a1929e9..c18a8f58bb 100644 --- a/docs/reference/commandline/deploy.md +++ b/docs/reference/commandline/deploy.md @@ -19,7 +19,7 @@ Create and update a stack from a Distributed Application Bundle (DAB) Options: --file string Path to a Distributed Application Bundle file (Default: STACK.dab) --help Print usage - --registry-auth Send registry authentication details to Swarm agents + --registry-auth Send registry authentication details to swarm agents ``` Create and update a stack from a `dab` file. This command has to be diff --git a/docs/reference/commandline/index.md b/docs/reference/commandline/index.md index 75acb1c7fa..583ddf9f95 100644 --- a/docs/reference/commandline/index.md +++ b/docs/reference/commandline/index.md @@ -111,7 +111,6 @@ read the [`dockerd`](dockerd.md) reference page. | Command | Description | |:--------|:-------------------------------------------------------------------| -| [node accept](node_accept.md) | Accept a node into the swarm | | [node promote](node_promote.md) | Promote a node that is pending a promotion to manager | | [node demote](node_demote.md) | Demotes an existing manager so that it is no longer a manager | | [node inspect](node_inspect.md) | Inspect a node in the swarm | @@ -124,10 +123,11 @@ read the [`dockerd`](dockerd.md) reference page. | Command | Description | |:--------|:-------------------------------------------------------------------| -| [swarm init](swarm_init.md) | Initialize a Swarm | -| [swarm join](swarm_join.md) | Join a Swarm as a manager node or worker node | +| [swarm init](swarm_init.md) | Initialize a swarm | +| [swarm join](swarm_join.md) | Join a swarm as a manager node or worker node | | [swarm leave](swarm_leave.md) | Remove the current node from the swarm | | [swarm update](swarm_update.md) | Update attributes of a swarm | +| [swarm join-token](swarm_join_token.md) | Display or rotate join tokens | ### Swarm service commands diff --git a/docs/reference/commandline/info.md b/docs/reference/commandline/info.md index 7d0743554f..04ccc8ed67 100644 --- a/docs/reference/commandline/info.md +++ b/docs/reference/commandline/info.md @@ -38,7 +38,7 @@ available on the volume where `/var/lib/docker` is mounted. ## Display Docker system information Here is a sample output for a daemon running on Ubuntu, using the overlay -storage driver and a node that is part of a 2 node Swarm cluster: +storage driver and a node that is part of a 2 node swarm cluster: $ docker -D info Containers: 14 diff --git a/docs/reference/commandline/node_accept.md b/docs/reference/commandline/node_accept.md deleted file mode 100644 index 73676c086c..0000000000 --- a/docs/reference/commandline/node_accept.md +++ /dev/null @@ -1,32 +0,0 @@ - - -# node accept - -```markdown -Usage: docker node accept NODE [NODE...] - -Accept a node in the swarm - -Options: - --help Print usage -``` - -Accept a node into the swarm. This command targets a docker engine that is a manager in the swarm cluster. - - -```bash -$ docker node accept -``` - -## Related information - -* [node promote](node_promote.md) -* [node demote](node_demote.md) diff --git a/docs/reference/commandline/node_demote.md b/docs/reference/commandline/node_demote.md index 5d765adfaf..a1baff0e50 100644 --- a/docs/reference/commandline/node_demote.md +++ b/docs/reference/commandline/node_demote.md @@ -29,5 +29,4 @@ $ docker node demote ## Related information -* [node accept](node_accept.md) * [node promote](node_promote.md) diff --git a/docs/reference/commandline/node_inspect.md b/docs/reference/commandline/node_inspect.md index a565bb3af0..6cba8ece37 100644 --- a/docs/reference/commandline/node_inspect.md +++ b/docs/reference/commandline/node_inspect.md @@ -41,7 +41,6 @@ Example output: "UpdatedAt": "2016-06-16T22:52:45.230878043Z", "Spec": { "Role": "manager", - "Membership": "accepted", "Availability": "active" }, "Description": { diff --git a/docs/reference/commandline/node_ls.md b/docs/reference/commandline/node_ls.md index fee03cc588..f446568b3c 100644 --- a/docs/reference/commandline/node_ls.md +++ b/docs/reference/commandline/node_ls.md @@ -30,10 +30,10 @@ Lists all the nodes that the Docker Swarm manager knows about. You can filter us Example output: $ docker node ls - ID HOSTNAME MEMBERSHIP STATUS AVAILABILITY MANAGER STATUS LEADER - 1bcef6utixb0l0ca7gxuivsj0 swarm-worker2 Accepted Ready Active - 38ciaotwjuritcdtn9npbnkuz swarm-worker1 Accepted Ready Active - e216jshn25ckzbvmwlnh5jr3g * swarm-manager1 Accepted Ready Active Reachable Yes + ID HOSTNAME STATUS AVAILABILITY MANAGER STATUS + 1bcef6utixb0l0ca7gxuivsj0 swarm-worker2 Ready Active + 38ciaotwjuritcdtn9npbnkuz swarm-worker1 Ready Active + e216jshn25ckzbvmwlnh5jr3g * swarm-manager1 Ready Active Leader ## Filtering @@ -54,16 +54,16 @@ The `name` filter matches on all or part of a node name. The following filter matches the node with a name equal to `swarm-master` string. $ docker node ls -f name=swarm-manager1 - ID HOSTNAME MEMBERSHIP STATUS AVAILABILITY MANAGER STATUS LEADER - e216jshn25ckzbvmwlnh5jr3g * swarm-manager1 Accepted Ready Active Reachable Yes + ID HOSTNAME STATUS AVAILABILITY MANAGER STATUS + e216jshn25ckzbvmwlnh5jr3g * swarm-manager1 Ready Active Leader ### id The `id` filter matches all or part of a node's id. $ docker node ls -f id=1 - ID HOSTNAME MEMBERSHIP STATUS AVAILABILITY MANAGER STATUS LEADER - 1bcef6utixb0l0ca7gxuivsj0 swarm-worker2 Accepted Ready Active + ID HOSTNAME STATUS AVAILABILITY MANAGER STATUS + 1bcef6utixb0l0ca7gxuivsj0 swarm-worker2 Ready Active #### label @@ -75,8 +75,8 @@ The following filter matches nodes with the `usage` label regardless of its valu ```bash $ docker node ls -f "label=foo" -ID HOSTNAME MEMBERSHIP STATUS AVAILABILITY MANAGER STATUS LEADER -1bcef6utixb0l0ca7gxuivsj0 swarm-worker2 Accepted Ready Active +ID HOSTNAME STATUS AVAILABILITY MANAGER STATUS +1bcef6utixb0l0ca7gxuivsj0 swarm-worker2 Ready Active ``` diff --git a/docs/reference/commandline/node_promote.md b/docs/reference/commandline/node_promote.md index 884fee8fc9..b507e7d489 100644 --- a/docs/reference/commandline/node_promote.md +++ b/docs/reference/commandline/node_promote.md @@ -28,5 +28,4 @@ $ docker node promote ## Related information -* [node accept](node_accept.md) * [node demote](node_demote.md) diff --git a/docs/reference/commandline/node_rm.md b/docs/reference/commandline/node_rm.md index f03f8a19ad..1ed6a39336 100644 --- a/docs/reference/commandline/node_rm.md +++ b/docs/reference/commandline/node_rm.md @@ -23,14 +23,13 @@ Options: --help Print usage ``` -Removes specified nodes from a swarm. Rejects nodes with `Pending` -membership from the swarm. +Removes specified nodes from a swarm. Example output: $ docker node rm swarm-node-02 - Node swarm-node-02 removed from Swarm + Node swarm-node-02 removed from swarm ## Related information diff --git a/docs/reference/commandline/node_update.md b/docs/reference/commandline/node_update.md index f90b40cc97..479db326c1 100644 --- a/docs/reference/commandline/node_update.md +++ b/docs/reference/commandline/node_update.md @@ -21,7 +21,6 @@ Options: --help Print usage --label-add value Add or update a node label (key=value) (default []) --label-rm value Remove a node label if exists (default []) - --membership string Membership of the node (accepted/rejected) --role string Role of the node (worker/manager) ``` diff --git a/docs/reference/commandline/service_create.md b/docs/reference/commandline/service_create.md index 3d81286c4d..795ee395e8 100644 --- a/docs/reference/commandline/service_create.md +++ b/docs/reference/commandline/service_create.md @@ -31,7 +31,7 @@ Options: --name string Service name --network value Network attachments (default []) -p, --publish value Publish a port as a node port (default []) - --registry-auth Send registry authentication details to Swarm agents + --registry-auth Send registry authentication details to swarm agents --replicas value Number of tasks (default none) --reserve-cpu value Reserve CPUs (default 0.000) --reserve-memory value Reserve Memory (default 0 B) diff --git a/docs/reference/commandline/service_update.md b/docs/reference/commandline/service_update.md index 7796d8d0b5..f834ac4b31 100644 --- a/docs/reference/commandline/service_update.md +++ b/docs/reference/commandline/service_update.md @@ -38,7 +38,7 @@ Options: --network-rm value Remove a network by name (default []) --publish-add value Add or update a published port (default []) --publish-rm value Remove a published port by its target port (default []) - --registry-auth Send registry authentication details to Swarm agents + --registry-auth Send registry authentication details to swarm agents --replicas value Number of tasks (default none) --reserve-cpu value Reserve CPUs (default 0.000) --reserve-memory value Reserve Memory (default 0 B) diff --git a/docs/reference/commandline/swarm_init.md b/docs/reference/commandline/swarm_init.md index e7f8f3f166..c7575047af 100644 --- a/docs/reference/commandline/swarm_init.md +++ b/docs/reference/commandline/swarm_init.md @@ -14,74 +14,43 @@ parent = "smn_cli" ```markdown Usage: docker swarm init [OPTIONS] -Initialize a Swarm +Initialize a swarm Options: - --auto-accept value Auto acceptance policy (default worker) --cert-expiry duration Validity period for node certificates (default 2160h0m0s) --dispatcher-heartbeat duration Dispatcher heartbeat period (default 5s) --external-ca value Specifications of one or more certificate signing endpoints --force-new-cluster Force create a new cluster from current state. --help Print usage --listen-addr value Listen address (default 0.0.0.0:2377) - --secret string Set secret value needed to accept nodes into cluster --task-history-limit int Task history retention limit (default 10) ``` -Initialize a Swarm cluster. The docker engine targeted by this command becomes a manager -in the newly created one node Swarm cluster. +Initialize a swarm cluster. The docker engine targeted by this command becomes a manager +in the newly created one node swarm cluster. ```bash $ docker swarm init --listen-addr 192.168.99.121:2377 -No --secret provided. Generated random secret: - 4ao565v9jsuogtq5t8s379ulb - -Swarm initialized: current node (1ujecd0j9n3ro9i6628smdmth) is now a manager. +Swarm initialized: current node (bvz81updecsj6wjz393c09vti) is now a manager. To add a worker to this swarm, run the following command: - docker swarm join --secret 4ao565v9jsuogtq5t8s379ulb \ - --ca-hash sha256:07ce22bd1a7619f2adc0d63bd110479a170e7c4e69df05b67a1aa2705c88ef09 \ - 192.168.99.121:2377 -$ docker node ls -ID HOSTNAME MEMBERSHIP STATUS AVAILABILITY MANAGER STATUS LEADER -1ujecd0j9n3ro9i6628smdmth * manager1 Accepted Ready Active Reachable Yes + docker swarm join \ + --token SWMTKN-1-3pu6hszjas19xyp7ghgosyx9k8atbfcr8p2is99znpy26u2lkl-1awxwuwd3z9j1z3puu7rcgdbx \ + 172.17.0.2:2377 + +To add a manager to this swarm, run the following command: + docker swarm join \ + --token SWMTKN-1-3pu6hszjas19xyp7ghgosyx9k8atbfcr8p2is99znpy26u2lkl-7p73s1dx5in4tatdymyhg9hu2 \ + 172.17.0.2:2377 ``` -If a secret for joining new nodes is not provided with `--secret`, `docker swarm init` will -generate a random one and print it to the terminal (as seen in the example above). To initialize -a swarm with no secret, use `--secret ""`. +`docker swarm init` generates two random tokens, a worker token and a manager token. When you join +a new node to the swarm, the node joins as a worker or manager node based upon the token you pass +to [swarm join](swarm_join.md). -### `--auto-accept value` - -This flag controls node acceptance into the cluster. By default, `worker` nodes are -automatically accepted by the cluster. This can be changed by specifying what kinds of nodes -can be auto-accepted into the cluster. If auto-accept is not turned on, then -[node accept](node_accept.md) can be used to explicitly accept a node into the cluster. - -For example, the following initializes a cluster with auto-acceptance of workers, but not managers - - -```bash -$ docker swarm init --listen-addr 192.168.99.121:2377 --auto-accept worker -``` - -It is possible to pass a comma-separated list of node types. The following initializes a cluster -with auto-acceptance of both `worker` and `manager` nodes - -```bash -$ docker swarm init --listen-addr 192.168.99.121:2377 --auto-accept worker,manager -``` - -To disable auto acceptance, use the `none` option. Note that this option cannot -be combined with other values. When disabling auto acceptance, nodes must be -manually accepted or rejected using `docker node accept` or `docker node rm`. - -The following example enables swarm mode with auto acceptance disabled: - -```bash -$ docker swarm init --listen-addr 192.168.99.121:2377 --auto-accept none -``` +After you create the swarm, you can display or rotate the token using +[swarm join-token](swarm_join_token.md). ### `--cert-expiry` @@ -105,11 +74,7 @@ This flag forces an existing node that was part of a quorum that was lost to res ### `--listen-addr value` -The node listens for inbound Swarm manager traffic on this IP:PORT - -### `--secret string` - -Secret value needed to accept nodes into the Swarm +The node listens for inbound swarm manager traffic on this IP:PORT ### `--task-history-limit` @@ -120,5 +85,5 @@ This flag sets up task history retention limit. * [swarm join](swarm_join.md) * [swarm leave](swarm_leave.md) * [swarm update](swarm_update.md) -* [node accept](node_accept.md) +* [swarm join-token](swarm_join_token.md) * [node rm](node_rm.md) diff --git a/docs/reference/commandline/swarm_join.md b/docs/reference/commandline/swarm_join.md index 0499b10894..a8edc348d7 100644 --- a/docs/reference/commandline/swarm_join.md +++ b/docs/reference/commandline/swarm_join.md @@ -14,55 +14,54 @@ parent = "smn_cli" ```markdown Usage: docker swarm join [OPTIONS] HOST:PORT -Join a Swarm as a node and/or manager +Join a swarm as a node and/or manager Options: - --ca-hash string Hash of the Root Certificate Authority certificate used for trusted join --help Print usage --listen-addr value Listen address (default 0.0.0.0:2377) - --manager Try joining as a manager. - --secret string Secret for node acceptance + --token string Token for entry into the swarm ``` -Join a node to a Swarm cluster. If the `--manager` flag is specified, the docker engine -targeted by this command becomes a `manager`. If it is not specified, it becomes a `worker`. +Join a node to a swarm. The node joins as a manager node or worker node based upon the token you +pass with the `--token` flag. If you pass a manager token, the node joins as a manager. If you +pass a worker token, the node joins as a worker. ### Join a node to swarm as a manager +The example below demonstrates joining a manager node using a manager token. + ```bash -$ docker swarm join --secret 4ao565v9jsuogtq5t8s379ulb --manager --listen-addr 192.168.99.122:2377 192.168.99.121:2377 -This node joined a Swarm as a manager. +$ docker swarm join --token SWMTKN-1-3pu6hszjas19xyp7ghgosyx9k8atbfcr8p2is99znpy26u2lkl-7p73s1dx5in4tatdymyhg9hu2 --listen-addr 192.168.99.122:2377 192.168.99.121:2377 +This node joined a swarm as a manager. $ docker node ls -ID HOSTNAME MEMBERSHIP STATUS AVAILABILITY MANAGER STATUS LEADER -dkp8vy1dq1kxleu9g4u78tlag * manager2 Accepted Ready Active Reachable -dvfxp4zseq4s0rih1selh0d20 manager1 Accepted Ready Active Reachable Yes +ID HOSTNAME STATUS AVAILABILITY MANAGER STATUS +dkp8vy1dq1kxleu9g4u78tlag * manager2 Ready Active Reachable +dvfxp4zseq4s0rih1selh0d20 manager1 Ready Active Leader ``` +A cluster should only have 3-7 managers at most, because a majority of managers must be available +for the cluster to function. Nodes that aren't meant to participate in this management quorum +should join as workers instead. Managers should be stable hosts that have static IP addresses. + ### Join a node to swarm as a worker +The example below demonstrates joining a worker node using a worker token. + ```bash -$ docker swarm join --secret 4ao565v9jsuogtq5t8s379ulb --listen-addr 192.168.99.123:2377 192.168.99.121:2377 -This node joined a Swarm as a worker. +$ docker swarm join --token SWMTKN-1-3pu6hszjas19xyp7ghgosyx9k8atbfcr8p2is99znpy26u2lkl-1awxwuwd3z9j1z3puu7rcgdbx --listen-addr 192.168.99.123:2377 192.168.99.121:2377 +This node joined a swarm as a worker. $ docker node ls -ID HOSTNAME MEMBERSHIP STATUS AVAILABILITY MANAGER STATUS LEADER -7ln70fl22uw2dvjn2ft53m3q5 worker2 Accepted Ready Active -dkp8vy1dq1kxleu9g4u78tlag worker1 Accepted Ready Active Reachable -dvfxp4zseq4s0rih1selh0d20 * manager1 Accepted Ready Active Reachable Yes +ID HOSTNAME STATUS AVAILABILITY MANAGER STATUS +7ln70fl22uw2dvjn2ft53m3q5 worker2 Ready Active +dkp8vy1dq1kxleu9g4u78tlag worker1 Ready Active Reachable +dvfxp4zseq4s0rih1selh0d20 * manager1 Ready Active Leader ``` -### `--ca-hash` - -Hash of the Root Certificate Authority certificate used for trusted join. - ### `--listen-addr value` -The node listens for inbound Swarm manager traffic on this IP:PORT +The node listens for inbound swarm manager traffic on this IP:PORT -### `--manager` - -Joins the node as a manager - -### `--secret string` +### `--token string` Secret value required for nodes to join the swarm diff --git a/docs/reference/commandline/swarm_join_token.md b/docs/reference/commandline/swarm_join_token.md new file mode 100644 index 0000000000..1355f70ba1 --- /dev/null +++ b/docs/reference/commandline/swarm_join_token.md @@ -0,0 +1,76 @@ + + +# swarm join-token + +```markdown +Usage: docker swarm join-token [--rotate] (worker|manager) + +Manage join tokens + +Options: + --help Print usage + -q, --quiet Only display token + --rotate Rotate join token +``` + +Join tokens are secrets that determine whether or not a node will join the swarm as a manager node +or a worker node. You pass the token using the `--token flag` when you run +[swarm join](swarm_join.md). You can access the current tokens or rotate the tokens using +`swarm join-token`. + +Run with only a single `worker` or `manager` argument, it will print a command for joining a new +node to the swarm, including the necessary token: + +```bash +$ docker swarm join-token worker +To add a worker to this swarm, run the following command: + docker swarm join \ + --token SWMTKN-1-3pu6hszjas19xyp7ghgosyx9k8atbfcr8p2is99znpy26u2lkl-1awxwuwd3z9j1z3puu7rcgdbx \ + 172.17.0.2:2377 + +$ docker swarm join-token manager +To add a manager to this swarm, run the following command: + docker swarm join \ + --token SWMTKN-1-3pu6hszjas19xyp7ghgosyx9k8atbfcr8p2is99znpy26u2lkl-7p73s1dx5in4tatdymyhg9hu2 \ + 172.17.0.2:2377 +``` + +Use the `--rotate` flag to generate a new join token for the specified role: + +```bash +$ docker swarm join-token --rotate worker +To add a worker to this swarm, run the following command: + docker swarm join \ + --token SWMTKN-1-3pu6hszjas19xyp7ghgosyx9k8atbfcr8p2is99znpy26u2lkl-b30ljddcqhef9b9v4rs7mel7t \ + 172.17.0.2:2377 +``` + +After using `--rotate`, only the new token will be valid for joining with the specified role. + +The `-q` (or `--quiet`) flag only prints the token: + +```bash +$ docker swarm join-token -q worker +SWMTKN-1-3pu6hszjas19xyp7ghgosyx9k8atbfcr8p2is99znpy26u2lkl-b30ljddcqhef9b9v4rs7mel7t +``` + +### `--rotate` + +Update the join token for a specified role with a new token and print the token. + +### `--quiet` + +Only print the token. Do not print a complete command for joining. + +## Related information + +* [swarm join](swarm_join.md) diff --git a/docs/reference/commandline/swarm_leave.md b/docs/reference/commandline/swarm_leave.md index ffae0d6ab1..e838097c80 100644 --- a/docs/reference/commandline/swarm_leave.md +++ b/docs/reference/commandline/swarm_leave.md @@ -14,7 +14,7 @@ parent = "smn_cli" ```markdown Usage: docker swarm leave [OPTIONS] -Leave a Swarm +Leave a swarm Options: --force Force leave ignoring warnings. @@ -26,10 +26,10 @@ This command causes the node to leave the swarm. On a manager node: ```bash $ docker node ls -ID HOSTNAME MEMBERSHIP STATUS AVAILABILITY MANAGER STATUS LEADER -7ln70fl22uw2dvjn2ft53m3q5 worker2 Accepted Ready Active -dkp8vy1dq1kxleu9g4u78tlag worker1 Accepted Ready Active Reachable -dvfxp4zseq4s0rih1selh0d20 * manager1 Accepted Ready Active Reachable Yes +ID HOSTNAME STATUS AVAILABILITY MANAGER STATUS +7ln70fl22uw2dvjn2ft53m3q5 worker2 Ready Active +dkp8vy1dq1kxleu9g4u78tlag worker1 Ready Active Reachable +dvfxp4zseq4s0rih1selh0d20 * manager1 Ready Active Leader ``` On a worker node: @@ -41,10 +41,10 @@ Node left the default swarm. On a manager node: ```bash $ docker node ls -ID HOSTNAME MEMBERSHIP STATUS AVAILABILITY MANAGER STATUS LEADER -7ln70fl22uw2dvjn2ft53m3q5 worker2 Accepted Down Active -dkp8vy1dq1kxleu9g4u78tlag worker1 Accepted Ready Active Reachable -dvfxp4zseq4s0rih1selh0d20 * manager1 Accepted Ready Active Reachable Yes +ID HOSTNAME STATUS AVAILABILITY MANAGER STATUS +7ln70fl22uw2dvjn2ft53m3q5 worker2 Down Active +dkp8vy1dq1kxleu9g4u78tlag worker1 Ready Active Reachable +dvfxp4zseq4s0rih1selh0d20 * manager1 Ready Active Leader ``` ## Related information diff --git a/docs/reference/commandline/swarm_update.md b/docs/reference/commandline/swarm_update.md index 8ab715e278..6a5eb7305f 100644 --- a/docs/reference/commandline/swarm_update.md +++ b/docs/reference/commandline/swarm_update.md @@ -14,23 +14,21 @@ parent = "smn_cli" ```markdown Usage: docker swarm update [OPTIONS] -Update the Swarm +Update the swarm Options: - --auto-accept value Auto acceptance policy (worker, manager or none) --cert-expiry duration Validity period for node certificates (default 2160h0m0s) --dispatcher-heartbeat duration Dispatcher heartbeat period (default 5s) --external-ca value Specifications of one or more certificate signing endpoints --help Print usage - --secret string Set secret value needed to accept nodes into cluster --task-history-limit int Task history retention limit (default 10) ``` -Updates a Swarm cluster with new parameter values. This command must target a manager node. +Updates a swarm cluster with new parameter values. This command must target a manager node. ```bash -$ docker swarm update --auto-accept manager +$ docker swarm update --cert-expirty 4000h0m0s ``` ## Related information diff --git a/integration-cli/check_test.go b/integration-cli/check_test.go index 06f599407d..ce5bf8a94e 100644 --- a/integration-cli/check_test.go +++ b/integration-cli/check_test.go @@ -216,15 +216,17 @@ func (s *DockerSwarmSuite) AddDaemon(c *check.C, joinSwarm, manager bool) *Swarm if joinSwarm == true { if len(s.daemons) > 0 { + tokens := s.daemons[0].joinTokens(c) + token := tokens.Worker + if manager { + token = tokens.Manager + } c.Assert(d.Join(swarm.JoinRequest{ RemoteAddrs: []string{s.daemons[0].listenAddr}, - Manager: manager}), check.IsNil) - } else { - c.Assert(d.Init(swarm.InitRequest{ - Spec: swarm.Spec{ - AcceptancePolicy: autoAcceptPolicy, - }, + JoinToken: token, }), check.IsNil) + } else { + c.Assert(d.Init(swarm.InitRequest{}), check.IsNil) } } diff --git a/integration-cli/daemon_swarm.go b/integration-cli/daemon_swarm.go index 77a1be493b..bb557fc8a1 100644 --- a/integration-cli/daemon_swarm.go +++ b/integration-cli/daemon_swarm.go @@ -22,14 +22,6 @@ type SwarmDaemon struct { listenAddr string } -// default policy in tests is allow-all -var autoAcceptPolicy = swarm.AcceptancePolicy{ - Policies: []swarm.Policy{ - {Role: swarm.NodeRoleWorker, Autoaccept: true}, - {Role: swarm.NodeRoleManager, Autoaccept: true}, - }, -} - // Init initializes a new swarm cluster. func (d *SwarmDaemon) Init(req swarm.InitRequest) error { if req.ListenAddr == "" { @@ -271,6 +263,28 @@ func (d *SwarmDaemon) updateSwarm(c *check.C, f ...specConstructor) { c.Assert(status, checker.Equals, http.StatusOK, check.Commentf("output: %q", string(out))) } +func (d *SwarmDaemon) rotateTokens(c *check.C) { + var sw swarm.Swarm + status, out, err := d.SockRequest("GET", "/swarm", nil) + c.Assert(err, checker.IsNil) + c.Assert(status, checker.Equals, http.StatusOK, check.Commentf("output: %q", string(out))) + c.Assert(json.Unmarshal(out, &sw), checker.IsNil) + + url := fmt.Sprintf("/swarm/update?version=%d&rotate_worker_token=true&rotate_manager_token=true", sw.Version.Index) + status, out, err = d.SockRequest("POST", url, sw.Spec) + c.Assert(err, checker.IsNil) + c.Assert(status, checker.Equals, http.StatusOK, check.Commentf("output: %q", string(out))) +} + +func (d *SwarmDaemon) joinTokens(c *check.C) swarm.JoinTokens { + var sw swarm.Swarm + status, out, err := d.SockRequest("GET", "/swarm", nil) + c.Assert(err, checker.IsNil) + c.Assert(status, checker.Equals, http.StatusOK, check.Commentf("output: %q", string(out))) + c.Assert(json.Unmarshal(out, &sw), checker.IsNil) + return sw.JoinTokens +} + func (d *SwarmDaemon) checkLocalNodeState(c *check.C) (interface{}, check.CommentInterface) { info, err := d.info() c.Assert(err, checker.IsNil) diff --git a/integration-cli/docker_api_swarm_test.go b/integration-cli/docker_api_swarm_test.go index e33a8e98a8..03ed647dee 100644 --- a/integration-cli/docker_api_swarm_test.go +++ b/integration-cli/docker_api_swarm_test.go @@ -43,7 +43,7 @@ func (s *DockerSwarmSuite) TestApiSwarmInit(c *check.C) { c.Assert(info.ControlAvailable, checker.False) c.Assert(info.LocalNodeState, checker.Equals, swarm.LocalNodeStateInactive) - c.Assert(d2.Join(swarm.JoinRequest{RemoteAddrs: []string{d1.listenAddr}}), checker.IsNil) + c.Assert(d2.Join(swarm.JoinRequest{JoinToken: d1.joinTokens(c).Worker, RemoteAddrs: []string{d1.listenAddr}}), checker.IsNil) info, err = d2.info() c.Assert(err, checker.IsNil) @@ -72,89 +72,29 @@ func (s *DockerSwarmSuite) TestApiSwarmInit(c *check.C) { c.Assert(info.LocalNodeState, checker.Equals, swarm.LocalNodeStateActive) } -func (s *DockerSwarmSuite) TestApiSwarmManualAcceptance(c *check.C) { +func (s *DockerSwarmSuite) TestApiSwarmJoinToken(c *check.C) { testRequires(c, Network) - s.testAPISwarmManualAcceptance(c, "") -} -func (s *DockerSwarmSuite) TestApiSwarmManualAcceptanceSecret(c *check.C) { - testRequires(c, Network) - s.testAPISwarmManualAcceptance(c, "foobaz") -} - -func (s *DockerSwarmSuite) testAPISwarmManualAcceptance(c *check.C, secret string) { d1 := s.AddDaemon(c, false, false) - c.Assert(d1.Init(swarm.InitRequest{ - Spec: swarm.Spec{ - AcceptancePolicy: swarm.AcceptancePolicy{ - Policies: []swarm.Policy{ - {Role: swarm.NodeRoleWorker, Secret: &secret}, - {Role: swarm.NodeRoleManager, Secret: &secret}, - }, - }, - }, - }), checker.IsNil) + c.Assert(d1.Init(swarm.InitRequest{}), checker.IsNil) d2 := s.AddDaemon(c, false, false) err := d2.Join(swarm.JoinRequest{RemoteAddrs: []string{d1.listenAddr}}) c.Assert(err, checker.NotNil) - if secret == "" { - c.Assert(err.Error(), checker.Contains, "needs to be accepted") - info, err := d2.info() - c.Assert(err, checker.IsNil) - c.Assert(info.LocalNodeState, checker.Equals, swarm.LocalNodeStatePending) - c.Assert(d2.Leave(false), checker.IsNil) - info, err = d2.info() - c.Assert(err, checker.IsNil) - c.Assert(info.LocalNodeState, checker.Equals, swarm.LocalNodeStateInactive) - } else { - c.Assert(err.Error(), checker.Contains, "valid secret token is necessary") - info, err := d2.info() - c.Assert(err, checker.IsNil) - c.Assert(info.LocalNodeState, checker.Equals, swarm.LocalNodeStateInactive) - } - d3 := s.AddDaemon(c, false, false) - c.Assert(d3.Join(swarm.JoinRequest{Secret: secret, RemoteAddrs: []string{d1.listenAddr}}), checker.NotNil) - info, err := d3.info() - c.Assert(err, checker.IsNil) - c.Assert(info.LocalNodeState, checker.Equals, swarm.LocalNodeStatePending) - c.Assert(len(info.NodeID), checker.GreaterThan, 5) - d1.updateNode(c, info.NodeID, func(n *swarm.Node) { - n.Spec.Membership = swarm.NodeMembershipAccepted - }) - waitAndAssert(c, defaultReconciliationTimeout, d3.checkLocalNodeState, checker.Equals, swarm.LocalNodeStateActive) -} - -func (s *DockerSwarmSuite) TestApiSwarmSecretAcceptance(c *check.C) { - testRequires(c, Network) - d1 := s.AddDaemon(c, false, false) - secret := "foobar" - c.Assert(d1.Init(swarm.InitRequest{ - Spec: swarm.Spec{ - AcceptancePolicy: swarm.AcceptancePolicy{ - Policies: []swarm.Policy{ - {Role: swarm.NodeRoleWorker, Autoaccept: true, Secret: &secret}, - {Role: swarm.NodeRoleManager, Secret: &secret}, - }, - }, - }, - }), checker.IsNil) - - d2 := s.AddDaemon(c, false, false) - err := d2.Join(swarm.JoinRequest{RemoteAddrs: []string{d1.listenAddr}}) - c.Assert(err, checker.NotNil) - c.Assert(err.Error(), checker.Contains, "secret token is necessary") + c.Assert(err.Error(), checker.Contains, "join token is necessary") info, err := d2.info() c.Assert(err, checker.IsNil) c.Assert(info.LocalNodeState, checker.Equals, swarm.LocalNodeStateInactive) - err = d2.Join(swarm.JoinRequest{Secret: "foobaz", RemoteAddrs: []string{d1.listenAddr}}) + err = d2.Join(swarm.JoinRequest{JoinToken: "foobaz", RemoteAddrs: []string{d1.listenAddr}}) c.Assert(err, checker.NotNil) - c.Assert(err.Error(), checker.Contains, "secret token is necessary") + c.Assert(err.Error(), checker.Contains, "join token is necessary") info, err = d2.info() c.Assert(err, checker.IsNil) c.Assert(info.LocalNodeState, checker.Equals, swarm.LocalNodeStateInactive) - c.Assert(d2.Join(swarm.JoinRequest{Secret: "foobar", RemoteAddrs: []string{d1.listenAddr}}), checker.IsNil) + workerToken := d1.joinTokens(c).Worker + + c.Assert(d2.Join(swarm.JoinRequest{JoinToken: workerToken, RemoteAddrs: []string{d1.listenAddr}}), checker.IsNil) info, err = d2.info() c.Assert(err, checker.IsNil) c.Assert(info.LocalNodeState, checker.Equals, swarm.LocalNodeStateActive) @@ -163,22 +103,19 @@ func (s *DockerSwarmSuite) TestApiSwarmSecretAcceptance(c *check.C) { c.Assert(err, checker.IsNil) c.Assert(info.LocalNodeState, checker.Equals, swarm.LocalNodeStateInactive) - // change secret - d1.updateSwarm(c, func(s *swarm.Spec) { - for i := range s.AcceptancePolicy.Policies { - p := "foobaz" - s.AcceptancePolicy.Policies[i].Secret = &p - } - }) + // change tokens + d1.rotateTokens(c) - err = d2.Join(swarm.JoinRequest{Secret: "foobar", RemoteAddrs: []string{d1.listenAddr}}) + err = d2.Join(swarm.JoinRequest{JoinToken: workerToken, RemoteAddrs: []string{d1.listenAddr}}) c.Assert(err, checker.NotNil) - c.Assert(err.Error(), checker.Contains, "secret token is necessary") + c.Assert(err.Error(), checker.Contains, "join token is necessary") info, err = d2.info() c.Assert(err, checker.IsNil) c.Assert(info.LocalNodeState, checker.Equals, swarm.LocalNodeStateInactive) - c.Assert(d2.Join(swarm.JoinRequest{Secret: "foobaz", RemoteAddrs: []string{d1.listenAddr}}), checker.IsNil) + workerToken = d1.joinTokens(c).Worker + + c.Assert(d2.Join(swarm.JoinRequest{JoinToken: workerToken, RemoteAddrs: []string{d1.listenAddr}}), checker.IsNil) info, err = d2.info() c.Assert(err, checker.IsNil) c.Assert(info.LocalNodeState, checker.Equals, swarm.LocalNodeStateActive) @@ -187,24 +124,17 @@ func (s *DockerSwarmSuite) TestApiSwarmSecretAcceptance(c *check.C) { c.Assert(err, checker.IsNil) c.Assert(info.LocalNodeState, checker.Equals, swarm.LocalNodeStateInactive) - // change policy, don't change secret - d1.updateSwarm(c, func(s *swarm.Spec) { - for i, p := range s.AcceptancePolicy.Policies { - if p.Role == swarm.NodeRoleManager { - s.AcceptancePolicy.Policies[i].Autoaccept = false - } - s.AcceptancePolicy.Policies[i].Secret = nil - } - }) + // change spec, don't change tokens + d1.updateSwarm(c, func(s *swarm.Spec) {}) err = d2.Join(swarm.JoinRequest{RemoteAddrs: []string{d1.listenAddr}}) c.Assert(err, checker.NotNil) - c.Assert(err.Error(), checker.Contains, "secret token is necessary") + c.Assert(err.Error(), checker.Contains, "join token is necessary") info, err = d2.info() c.Assert(err, checker.IsNil) c.Assert(info.LocalNodeState, checker.Equals, swarm.LocalNodeStateInactive) - c.Assert(d2.Join(swarm.JoinRequest{Secret: "foobaz", RemoteAddrs: []string{d1.listenAddr}}), checker.IsNil) + c.Assert(d2.Join(swarm.JoinRequest{JoinToken: workerToken, RemoteAddrs: []string{d1.listenAddr}}), checker.IsNil) info, err = d2.info() c.Assert(err, checker.IsNil) c.Assert(info.LocalNodeState, checker.Equals, swarm.LocalNodeStateActive) @@ -212,51 +142,24 @@ func (s *DockerSwarmSuite) TestApiSwarmSecretAcceptance(c *check.C) { info, err = d2.info() c.Assert(err, checker.IsNil) c.Assert(info.LocalNodeState, checker.Equals, swarm.LocalNodeStateInactive) - - // clear secret - d1.updateSwarm(c, func(s *swarm.Spec) { - for i := range s.AcceptancePolicy.Policies { - p := "" - s.AcceptancePolicy.Policies[i].Secret = &p - } - }) - - c.Assert(d2.Join(swarm.JoinRequest{RemoteAddrs: []string{d1.listenAddr}}), checker.IsNil) - info, err = d2.info() - c.Assert(err, checker.IsNil) - c.Assert(info.LocalNodeState, checker.Equals, swarm.LocalNodeStateActive) - c.Assert(d2.Leave(false), checker.IsNil) - info, err = d2.info() - c.Assert(err, checker.IsNil) - c.Assert(info.LocalNodeState, checker.Equals, swarm.LocalNodeStateInactive) - } func (s *DockerSwarmSuite) TestApiSwarmCAHash(c *check.C) { testRequires(c, Network) d1 := s.AddDaemon(c, true, true) d2 := s.AddDaemon(c, false, false) - err := d2.Join(swarm.JoinRequest{CACertHash: "foobar", RemoteAddrs: []string{d1.listenAddr}}) + splitToken := strings.Split(d1.joinTokens(c).Worker, "-") + splitToken[2] = "1kxftv4ofnc6mt30lmgipg6ngf9luhwqopfk1tz6bdmnkubg0e" + replacementToken := strings.Join(splitToken, "-") + err := d2.Join(swarm.JoinRequest{JoinToken: replacementToken, RemoteAddrs: []string{d1.listenAddr}}) c.Assert(err, checker.NotNil) - c.Assert(err.Error(), checker.Contains, "invalid checksum digest format") - - c.Assert(len(d1.CACertHash), checker.GreaterThan, 0) - c.Assert(d2.Join(swarm.JoinRequest{CACertHash: d1.CACertHash, RemoteAddrs: []string{d1.listenAddr}}), checker.IsNil) + c.Assert(err.Error(), checker.Contains, "remote CA does not match fingerprint") } func (s *DockerSwarmSuite) TestApiSwarmPromoteDemote(c *check.C) { testRequires(c, Network) d1 := s.AddDaemon(c, false, false) - c.Assert(d1.Init(swarm.InitRequest{ - Spec: swarm.Spec{ - AcceptancePolicy: swarm.AcceptancePolicy{ - Policies: []swarm.Policy{ - {Role: swarm.NodeRoleWorker, Autoaccept: true}, - {Role: swarm.NodeRoleManager}, - }, - }, - }, - }), checker.IsNil) + c.Assert(d1.Init(swarm.InitRequest{}), checker.IsNil) d2 := s.AddDaemon(c, true, false) info, err := d2.info() @@ -838,9 +741,7 @@ func (s *DockerSwarmSuite) TestApiSwarmForceNewCluster(c *check.C) { c.Assert(d1.Init(swarm.InitRequest{ ForceNewCluster: true, - Spec: swarm.Spec{ - AcceptancePolicy: autoAcceptPolicy, - }, + Spec: swarm.Spec{}, }), checker.IsNil) waitAndAssert(c, defaultReconciliationTimeout, d1.checkActiveContainerCount, checker.Equals, instances) @@ -937,7 +838,6 @@ func checkClusterHealth(c *check.C, cl []*SwarmDaemon, managerCount, workerCount for _, n := range d.listNodes(c) { c.Assert(n.Status.State, checker.Equals, swarm.NodeStateReady, check.Commentf("state of node %s, reported by %s", n.ID, d.Info.NodeID)) c.Assert(n.Spec.Availability, checker.Equals, swarm.NodeAvailabilityActive, check.Commentf("availability of node %s, reported by %s", n.ID, d.Info.NodeID)) - c.Assert(n.Spec.Membership, checker.Equals, swarm.NodeMembershipAccepted, check.Commentf("membership of node %s, reported by %s", n.ID, d.Info.NodeID)) if n.Spec.Role == swarm.NodeRoleManager { c.Assert(n.ManagerStatus, checker.NotNil, check.Commentf("manager status of node %s (manager), reported by %s", n.ID, d.Info.NodeID)) if n.ManagerStatus.Leader { diff --git a/integration-cli/docker_cli_swarm_test.go b/integration-cli/docker_cli_swarm_test.go index 8f6ee9d54d..087efa1262 100644 --- a/integration-cli/docker_cli_swarm_test.go +++ b/integration-cli/docker_cli_swarm_test.go @@ -25,50 +25,13 @@ func (s *DockerSwarmSuite) TestSwarmUpdate(c *check.C) { return sw[0].Spec } - out, err := d.Cmd("swarm", "update", "--cert-expiry", "30h", "--dispatcher-heartbeat", "11s", "--auto-accept", "manager", "--auto-accept", "worker", "--secret", "foo") + out, err := d.Cmd("swarm", "update", "--cert-expiry", "30h", "--dispatcher-heartbeat", "11s") c.Assert(err, checker.IsNil, check.Commentf("out: %v", out)) spec := getSpec() c.Assert(spec.CAConfig.NodeCertExpiry, checker.Equals, 30*time.Hour) c.Assert(spec.Dispatcher.HeartbeatPeriod, checker.Equals, uint64(11*time.Second)) - c.Assert(spec.AcceptancePolicy.Policies, checker.HasLen, 2) - - for _, p := range spec.AcceptancePolicy.Policies { - c.Assert(p.Autoaccept, checker.Equals, true) - c.Assert(p.Secret, checker.NotNil) - c.Assert(*p.Secret, checker.Not(checker.Equals), "") - } - - out, err = d.Cmd("swarm", "update", "--auto-accept", "none") - c.Assert(err, checker.IsNil, check.Commentf("out: %v", out)) - - spec = getSpec() - c.Assert(spec.CAConfig.NodeCertExpiry, checker.Equals, 30*time.Hour) - c.Assert(spec.Dispatcher.HeartbeatPeriod, checker.Equals, uint64(11*time.Second)) - - c.Assert(spec.AcceptancePolicy.Policies, checker.HasLen, 2) - - for _, p := range spec.AcceptancePolicy.Policies { - c.Assert(p.Autoaccept, checker.Equals, false) - // secret is still set - c.Assert(p.Secret, checker.NotNil) - c.Assert(*p.Secret, checker.Not(checker.Equals), "") - } - - out, err = d.Cmd("swarm", "update", "--auto-accept", "manager", "--secret", "") - c.Assert(err, checker.IsNil, check.Commentf("out: %v", out)) - - spec = getSpec() - - c.Assert(spec.AcceptancePolicy.Policies, checker.HasLen, 2) - - for _, p := range spec.AcceptancePolicy.Policies { - c.Assert(p.Autoaccept, checker.Equals, p.Role == swarm.NodeRoleManager) - // secret has been removed - c.Assert(p.Secret, checker.IsNil) - } - // setting anything under 30m for cert-expiry is not allowed out, err = d.Cmd("swarm", "update", "--cert-expiry", "15m") c.Assert(err, checker.NotNil) @@ -89,37 +52,21 @@ func (s *DockerSwarmSuite) TestSwarmInit(c *check.C) { return sw[0].Spec } - out, err := d.Cmd("swarm", "init", "--cert-expiry", "30h", "--dispatcher-heartbeat", "11s", "--auto-accept", "manager", "--auto-accept", "worker", "--secret", "foo") + out, err := d.Cmd("swarm", "init", "--cert-expiry", "30h", "--dispatcher-heartbeat", "11s") c.Assert(err, checker.IsNil, check.Commentf("out: %v", out)) spec := getSpec() c.Assert(spec.CAConfig.NodeCertExpiry, checker.Equals, 30*time.Hour) c.Assert(spec.Dispatcher.HeartbeatPeriod, checker.Equals, uint64(11*time.Second)) - c.Assert(spec.AcceptancePolicy.Policies, checker.HasLen, 2) - - for _, p := range spec.AcceptancePolicy.Policies { - c.Assert(p.Autoaccept, checker.Equals, true) - c.Assert(p.Secret, checker.NotNil) - c.Assert(*p.Secret, checker.Not(checker.Equals), "") - } - c.Assert(d.Leave(true), checker.IsNil) - out, err = d.Cmd("swarm", "init", "--auto-accept", "none", "--secret", "") + out, err = d.Cmd("swarm", "init") c.Assert(err, checker.IsNil, check.Commentf("out: %v", out)) spec = getSpec() c.Assert(spec.CAConfig.NodeCertExpiry, checker.Equals, 90*24*time.Hour) c.Assert(spec.Dispatcher.HeartbeatPeriod, checker.Equals, uint64(5*time.Second)) - - c.Assert(spec.AcceptancePolicy.Policies, checker.HasLen, 2) - - for _, p := range spec.AcceptancePolicy.Policies { - c.Assert(p.Autoaccept, checker.Equals, false) - c.Assert(p.Secret, checker.IsNil) - } - } func (s *DockerSwarmSuite) TestSwarmInitIPv6(c *check.C) {