Replace secrets with join tokens
Implement the proposal from https://github.com/docker/docker/issues/24430#issuecomment-233100121 Removes acceptance policy and secret in favor of an automatically generated join token that combines the secret, CA hash, and manager/worker role into a single opaque string. Adds a docker swarm join-token subcommand to inspect and rotate the tokens. Signed-off-by: Aaron Lehmann <aaron.lehmann@docker.com>
This commit is contained in:
parent
c435045c30
commit
2cc5bd33ee
46 changed files with 451 additions and 893 deletions
|
@ -1,32 +0,0 @@
|
|||
package node
|
||||
|
||||
import (
|
||||
"fmt"
|
||||
|
||||
"github.com/docker/docker/api/client"
|
||||
"github.com/docker/docker/cli"
|
||||
"github.com/docker/engine-api/types/swarm"
|
||||
"github.com/spf13/cobra"
|
||||
)
|
||||
|
||||
func newAcceptCommand(dockerCli *client.DockerCli) *cobra.Command {
|
||||
return &cobra.Command{
|
||||
Use: "accept NODE [NODE...]",
|
||||
Short: "Accept a node in the swarm",
|
||||
Args: cli.RequiresMinArgs(1),
|
||||
RunE: func(cmd *cobra.Command, args []string) error {
|
||||
return runAccept(dockerCli, args)
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
func runAccept(dockerCli *client.DockerCli, nodes []string) error {
|
||||
accept := func(node *swarm.Node) error {
|
||||
node.Spec.Membership = swarm.NodeMembershipAccepted
|
||||
return nil
|
||||
}
|
||||
success := func(nodeID string) {
|
||||
fmt.Fprintf(dockerCli.Out(), "Node %s accepted in the swarm.\n", nodeID)
|
||||
}
|
||||
return updateNodes(dockerCli, nodes, accept, success)
|
||||
}
|
|
@ -23,7 +23,6 @@ func NewNodeCommand(dockerCli *client.DockerCli) *cobra.Command {
|
|||
},
|
||||
}
|
||||
cmd.AddCommand(
|
||||
newAcceptCommand(dockerCli),
|
||||
newDemoteCommand(dockerCli),
|
||||
newInspectCommand(dockerCli),
|
||||
newListCommand(dockerCli),
|
||||
|
|
|
@ -16,7 +16,7 @@ import (
|
|||
)
|
||||
|
||||
const (
|
||||
listItemFmt = "%s\t%s\t%s\t%s\t%s\t%s\n"
|
||||
listItemFmt = "%s\t%s\t%s\t%s\t%s\n"
|
||||
)
|
||||
|
||||
type listOptions struct {
|
||||
|
@ -74,11 +74,10 @@ func printTable(out io.Writer, nodes []swarm.Node, info types.Info) {
|
|||
// Ignore flushing errors
|
||||
defer writer.Flush()
|
||||
|
||||
fmt.Fprintf(writer, listItemFmt, "ID", "HOSTNAME", "MEMBERSHIP", "STATUS", "AVAILABILITY", "MANAGER STATUS")
|
||||
fmt.Fprintf(writer, listItemFmt, "ID", "HOSTNAME", "STATUS", "AVAILABILITY", "MANAGER STATUS")
|
||||
for _, node := range nodes {
|
||||
name := node.Description.Hostname
|
||||
availability := string(node.Spec.Availability)
|
||||
membership := string(node.Spec.Membership)
|
||||
|
||||
reachability := ""
|
||||
if node.ManagerStatus != nil {
|
||||
|
@ -99,7 +98,6 @@ func printTable(out io.Writer, nodes []swarm.Node, info types.Info) {
|
|||
listItemFmt,
|
||||
ID,
|
||||
name,
|
||||
client.PrettyPrint(membership),
|
||||
client.PrettyPrint(string(node.Status.State)),
|
||||
client.PrettyPrint(availability),
|
||||
client.PrettyPrint(reachability))
|
||||
|
|
|
@ -12,7 +12,6 @@ import (
|
|||
type nodeOptions struct {
|
||||
annotations
|
||||
role string
|
||||
membership string
|
||||
availability string
|
||||
}
|
||||
|
||||
|
@ -45,14 +44,6 @@ func (opts *nodeOptions) ToNodeSpec() (swarm.NodeSpec, error) {
|
|||
return swarm.NodeSpec{}, fmt.Errorf("invalid role %q, only worker and manager are supported", opts.role)
|
||||
}
|
||||
|
||||
switch swarm.NodeMembership(strings.ToLower(opts.membership)) {
|
||||
case swarm.NodeMembershipAccepted:
|
||||
spec.Membership = swarm.NodeMembershipAccepted
|
||||
case "":
|
||||
default:
|
||||
return swarm.NodeSpec{}, fmt.Errorf("invalid membership %q, only accepted is supported", opts.membership)
|
||||
}
|
||||
|
||||
switch swarm.NodeAvailability(strings.ToLower(opts.availability)) {
|
||||
case swarm.NodeAvailabilityActive:
|
||||
spec.Availability = swarm.NodeAvailabilityActive
|
||||
|
|
|
@ -27,7 +27,6 @@ func newUpdateCommand(dockerCli *client.DockerCli) *cobra.Command {
|
|||
|
||||
flags := cmd.Flags()
|
||||
flags.StringVar(&nodeOpts.role, flagRole, "", "Role of the node (worker/manager)")
|
||||
flags.StringVar(&nodeOpts.membership, flagMembership, "", "Membership of the node (accepted/rejected)")
|
||||
flags.StringVar(&nodeOpts.availability, flagAvailability, "", "Availability of the node (active/pause/drain)")
|
||||
flags.Var(&nodeOpts.annotations.labels, flagLabelAdd, "Add or update a node label (key=value)")
|
||||
labelKeys := opts.NewListOpts(nil)
|
||||
|
@ -76,13 +75,6 @@ func mergeNodeUpdate(flags *pflag.FlagSet) func(*swarm.Node) error {
|
|||
}
|
||||
spec.Role = swarm.NodeRole(str)
|
||||
}
|
||||
if flags.Changed(flagMembership) {
|
||||
str, err := flags.GetString(flagMembership)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
spec.Membership = swarm.NodeMembership(str)
|
||||
}
|
||||
if flags.Changed(flagAvailability) {
|
||||
str, err := flags.GetString(flagAvailability)
|
||||
if err != nil {
|
||||
|
@ -115,7 +107,6 @@ func mergeNodeUpdate(flags *pflag.FlagSet) func(*swarm.Node) error {
|
|||
|
||||
const (
|
||||
flagRole = "role"
|
||||
flagMembership = "membership"
|
||||
flagAvailability = "availability"
|
||||
flagLabelAdd = "label-add"
|
||||
flagLabelRemove = "label-rm"
|
||||
|
|
|
@ -506,7 +506,7 @@ func addServiceFlags(cmd *cobra.Command, opts *serviceOptions) {
|
|||
|
||||
flags.StringVar(&opts.endpoint.mode, flagEndpointMode, "", "Endpoint mode (vip or dnsrr)")
|
||||
|
||||
flags.BoolVar(&opts.registryAuth, flagRegistryAuth, false, "Send registry authentication details to Swarm agents")
|
||||
flags.BoolVar(&opts.registryAuth, flagRegistryAuth, false, "Send registry authentication details to swarm agents")
|
||||
|
||||
flags.StringVar(&opts.logDriver.name, flagLogDriver, "", "Logging driver for service")
|
||||
flags.Var(&opts.logDriver.opts, flagLogOpt, "Logging driver options")
|
||||
|
|
|
@ -25,6 +25,7 @@ func NewSwarmCommand(dockerCli *client.DockerCli) *cobra.Command {
|
|||
newUpdateCommand(dockerCli),
|
||||
newLeaveCommand(dockerCli),
|
||||
newInspectCommand(dockerCli),
|
||||
newJoinTokenCommand(dockerCli),
|
||||
)
|
||||
return cmd
|
||||
}
|
||||
|
|
|
@ -28,14 +28,11 @@ type initOptions struct {
|
|||
func newInitCommand(dockerCli *client.DockerCli) *cobra.Command {
|
||||
opts := initOptions{
|
||||
listenAddr: NewListenAddrOption(),
|
||||
swarmOptions: swarmOptions{
|
||||
autoAccept: NewAutoAcceptOption(),
|
||||
},
|
||||
}
|
||||
|
||||
cmd := &cobra.Command{
|
||||
Use: "init [OPTIONS]",
|
||||
Short: "Initialize a Swarm",
|
||||
Short: "Initialize a swarm",
|
||||
Args: cli.NoArgs,
|
||||
RunE: func(cmd *cobra.Command, args []string) error {
|
||||
return runInit(dockerCli, cmd.Flags(), opts)
|
||||
|
@ -53,12 +50,6 @@ func runInit(dockerCli *client.DockerCli, flags *pflag.FlagSet, opts initOptions
|
|||
client := dockerCli.Client()
|
||||
ctx := context.Background()
|
||||
|
||||
// If no secret was specified, we create a random one
|
||||
if !flags.Changed("secret") {
|
||||
opts.secret = generateRandomSecret()
|
||||
fmt.Fprintf(dockerCli.Out(), "No --secret provided. Generated random secret:\n %s\n\n", opts.secret)
|
||||
}
|
||||
|
||||
req := swarm.InitRequest{
|
||||
ListenAddr: opts.listenAddr.String(),
|
||||
ForceNewCluster: opts.forceNewCluster,
|
||||
|
@ -72,24 +63,5 @@ func runInit(dockerCli *client.DockerCli, flags *pflag.FlagSet, opts initOptions
|
|||
|
||||
fmt.Fprintf(dockerCli.Out(), "Swarm initialized: current node (%s) is now a manager.\n\n", nodeID)
|
||||
|
||||
// Fetch CAHash and Address from the API
|
||||
info, err := client.Info(ctx)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
node, _, err := client.NodeInspectWithRaw(ctx, nodeID)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
if node.ManagerStatus != nil && info.Swarm.CACertHash != "" {
|
||||
var secretArgs string
|
||||
if opts.secret != "" {
|
||||
secretArgs = "--secret " + opts.secret
|
||||
}
|
||||
fmt.Fprintf(dockerCli.Out(), "To add a worker to this swarm, run the following command:\n docker swarm join %s \\\n --ca-hash %s \\\n %s\n", secretArgs, info.Swarm.CACertHash, node.ManagerStatus.Addr)
|
||||
}
|
||||
|
||||
return nil
|
||||
return printJoinCommand(ctx, dockerCli, nodeID, true, true)
|
||||
}
|
||||
|
|
|
@ -18,7 +18,7 @@ func newInspectCommand(dockerCli *client.DockerCli) *cobra.Command {
|
|||
|
||||
cmd := &cobra.Command{
|
||||
Use: "inspect [OPTIONS]",
|
||||
Short: "Inspect the Swarm",
|
||||
Short: "Inspect the swarm",
|
||||
Args: cli.NoArgs,
|
||||
RunE: func(cmd *cobra.Command, args []string) error {
|
||||
return runInspect(dockerCli, opts)
|
||||
|
|
|
@ -2,6 +2,7 @@ package swarm
|
|||
|
||||
import (
|
||||
"fmt"
|
||||
"strings"
|
||||
|
||||
"github.com/docker/docker/api/client"
|
||||
"github.com/docker/docker/cli"
|
||||
|
@ -13,9 +14,7 @@ import (
|
|||
type joinOptions struct {
|
||||
remote string
|
||||
listenAddr NodeAddrOption
|
||||
manager bool
|
||||
secret string
|
||||
CACertHash string
|
||||
token string
|
||||
}
|
||||
|
||||
func newJoinCommand(dockerCli *client.DockerCli) *cobra.Command {
|
||||
|
@ -25,7 +24,7 @@ func newJoinCommand(dockerCli *client.DockerCli) *cobra.Command {
|
|||
|
||||
cmd := &cobra.Command{
|
||||
Use: "join [OPTIONS] HOST:PORT",
|
||||
Short: "Join a Swarm as a node and/or manager",
|
||||
Short: "Join a swarm as a node and/or manager",
|
||||
Args: cli.ExactArgs(1),
|
||||
RunE: func(cmd *cobra.Command, args []string) error {
|
||||
opts.remote = args[0]
|
||||
|
@ -35,9 +34,7 @@ func newJoinCommand(dockerCli *client.DockerCli) *cobra.Command {
|
|||
|
||||
flags := cmd.Flags()
|
||||
flags.Var(&opts.listenAddr, flagListenAddr, "Listen address")
|
||||
flags.BoolVar(&opts.manager, "manager", false, "Try joining as a manager.")
|
||||
flags.StringVar(&opts.secret, flagSecret, "", "Secret for node acceptance")
|
||||
flags.StringVar(&opts.CACertHash, "ca-hash", "", "Hash of the Root Certificate Authority certificate used for trusted join")
|
||||
flags.StringVar(&opts.token, flagToken, "", "Token for entry into the swarm")
|
||||
return cmd
|
||||
}
|
||||
|
||||
|
@ -46,20 +43,29 @@ func runJoin(dockerCli *client.DockerCli, opts joinOptions) error {
|
|||
ctx := context.Background()
|
||||
|
||||
req := swarm.JoinRequest{
|
||||
Manager: opts.manager,
|
||||
Secret: opts.secret,
|
||||
JoinToken: opts.token,
|
||||
ListenAddr: opts.listenAddr.String(),
|
||||
RemoteAddrs: []string{opts.remote},
|
||||
CACertHash: opts.CACertHash,
|
||||
}
|
||||
err := client.SwarmJoin(ctx, req)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
if opts.manager {
|
||||
fmt.Fprintln(dockerCli.Out(), "This node joined a Swarm as a manager.")
|
||||
} else {
|
||||
fmt.Fprintln(dockerCli.Out(), "This node joined a Swarm as a worker.")
|
||||
|
||||
info, err := client.Info(ctx)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
_, _, err = client.NodeInspectWithRaw(ctx, info.Swarm.NodeID)
|
||||
if err != nil {
|
||||
// TODO(aaronl): is there a better way to do this?
|
||||
if strings.Contains(err.Error(), "This node is not a swarm manager.") {
|
||||
fmt.Fprintln(dockerCli.Out(), "This node joined a swarm as a worker.")
|
||||
}
|
||||
} else {
|
||||
fmt.Fprintln(dockerCli.Out(), "This node joined a swarm as a manager.")
|
||||
}
|
||||
|
||||
return nil
|
||||
}
|
||||
|
|
110
api/client/swarm/join_token.go
Normal file
110
api/client/swarm/join_token.go
Normal file
|
@ -0,0 +1,110 @@
|
|||
package swarm
|
||||
|
||||
import (
|
||||
"errors"
|
||||
"fmt"
|
||||
|
||||
"github.com/spf13/cobra"
|
||||
|
||||
"github.com/docker/docker/api/client"
|
||||
"github.com/docker/docker/cli"
|
||||
"github.com/docker/engine-api/types/swarm"
|
||||
"golang.org/x/net/context"
|
||||
)
|
||||
|
||||
const (
|
||||
flagRotate = "rotate"
|
||||
flagQuiet = "quiet"
|
||||
)
|
||||
|
||||
func newJoinTokenCommand(dockerCli *client.DockerCli) *cobra.Command {
|
||||
var rotate, quiet bool
|
||||
|
||||
cmd := &cobra.Command{
|
||||
Use: "join-token [-q] [--rotate] (worker|manager)",
|
||||
Short: "Manage join tokens",
|
||||
Args: cli.ExactArgs(1),
|
||||
RunE: func(cmd *cobra.Command, args []string) error {
|
||||
if args[0] != "worker" && args[0] != "manager" {
|
||||
return errors.New("unknown role " + args[0])
|
||||
}
|
||||
|
||||
client := dockerCli.Client()
|
||||
ctx := context.Background()
|
||||
|
||||
if rotate {
|
||||
var flags swarm.UpdateFlags
|
||||
|
||||
swarm, err := client.SwarmInspect(ctx)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
if args[0] == "worker" {
|
||||
flags.RotateWorkerToken = true
|
||||
} else if args[0] == "manager" {
|
||||
flags.RotateManagerToken = true
|
||||
}
|
||||
|
||||
err = client.SwarmUpdate(ctx, swarm.Version, swarm.Spec, flags)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
}
|
||||
|
||||
swarm, err := client.SwarmInspect(ctx)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
if quiet {
|
||||
if args[0] == "worker" {
|
||||
fmt.Fprintln(dockerCli.Out(), swarm.JoinTokens.Worker)
|
||||
} else if args[0] == "manager" {
|
||||
fmt.Fprintln(dockerCli.Out(), swarm.JoinTokens.Manager)
|
||||
}
|
||||
} else {
|
||||
info, err := client.Info(ctx)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
return printJoinCommand(ctx, dockerCli, info.Swarm.NodeID, args[0] == "worker", args[0] == "manager")
|
||||
}
|
||||
return nil
|
||||
},
|
||||
}
|
||||
|
||||
flags := cmd.Flags()
|
||||
flags.BoolVar(&rotate, flagRotate, false, "Rotate join token")
|
||||
flags.BoolVarP(&quiet, flagQuiet, "q", false, "Only display token")
|
||||
|
||||
return cmd
|
||||
}
|
||||
|
||||
func printJoinCommand(ctx context.Context, dockerCli *client.DockerCli, nodeID string, worker bool, manager bool) error {
|
||||
client := dockerCli.Client()
|
||||
|
||||
swarm, err := client.SwarmInspect(ctx)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
node, _, err := client.NodeInspectWithRaw(ctx, nodeID)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
if node.ManagerStatus != nil {
|
||||
if worker {
|
||||
fmt.Fprintf(dockerCli.Out(), "To add a worker to this swarm, run the following command:\n docker swarm join \\\n --token %s \\\n %s\n", swarm.JoinTokens.Worker, node.ManagerStatus.Addr)
|
||||
}
|
||||
if manager {
|
||||
if worker {
|
||||
fmt.Fprintln(dockerCli.Out())
|
||||
}
|
||||
fmt.Fprintf(dockerCli.Out(), "To add a manager to this swarm, run the following command:\n docker swarm join \\\n --token %s \\\n %s\n", swarm.JoinTokens.Manager, node.ManagerStatus.Addr)
|
||||
}
|
||||
}
|
||||
|
||||
return nil
|
||||
}
|
|
@ -19,7 +19,7 @@ func newLeaveCommand(dockerCli *client.DockerCli) *cobra.Command {
|
|||
|
||||
cmd := &cobra.Command{
|
||||
Use: "leave [OPTIONS]",
|
||||
Short: "Leave a Swarm",
|
||||
Short: "Leave a swarm",
|
||||
Args: cli.NoArgs,
|
||||
RunE: func(cmd *cobra.Command, args []string) error {
|
||||
return runLeave(dockerCli, opts)
|
||||
|
|
|
@ -15,29 +15,15 @@ import (
|
|||
const (
|
||||
defaultListenAddr = "0.0.0.0:2377"
|
||||
|
||||
worker = "WORKER"
|
||||
manager = "MANAGER"
|
||||
none = "NONE"
|
||||
|
||||
flagAutoAccept = "auto-accept"
|
||||
flagCertExpiry = "cert-expiry"
|
||||
flagDispatcherHeartbeat = "dispatcher-heartbeat"
|
||||
flagListenAddr = "listen-addr"
|
||||
flagSecret = "secret"
|
||||
flagToken = "token"
|
||||
flagTaskHistoryLimit = "task-history-limit"
|
||||
flagExternalCA = "external-ca"
|
||||
)
|
||||
|
||||
var (
|
||||
defaultPolicies = []swarm.Policy{
|
||||
{Role: worker, Autoaccept: true},
|
||||
{Role: manager, Autoaccept: false},
|
||||
}
|
||||
)
|
||||
|
||||
type swarmOptions struct {
|
||||
autoAccept AutoAcceptOption
|
||||
secret string
|
||||
taskHistoryLimit int64
|
||||
dispatcherHeartbeat time.Duration
|
||||
nodeCertExpiry time.Duration
|
||||
|
@ -84,71 +70,6 @@ func NewListenAddrOption() NodeAddrOption {
|
|||
return NewNodeAddrOption(defaultListenAddr)
|
||||
}
|
||||
|
||||
// AutoAcceptOption is a value type for auto-accept policy
|
||||
type AutoAcceptOption struct {
|
||||
values map[string]struct{}
|
||||
}
|
||||
|
||||
// String prints a string representation of this option
|
||||
func (o *AutoAcceptOption) String() string {
|
||||
keys := []string{}
|
||||
for key := range o.values {
|
||||
keys = append(keys, fmt.Sprintf("%s=true", strings.ToLower(key)))
|
||||
}
|
||||
return strings.Join(keys, ", ")
|
||||
}
|
||||
|
||||
// Set sets a new value on this option
|
||||
func (o *AutoAcceptOption) Set(acceptValues string) error {
|
||||
for _, value := range strings.Split(acceptValues, ",") {
|
||||
value = strings.ToUpper(value)
|
||||
switch value {
|
||||
case none, worker, manager:
|
||||
o.values[value] = struct{}{}
|
||||
default:
|
||||
return fmt.Errorf("must be one / combination of %s, %s; or NONE", worker, manager)
|
||||
}
|
||||
}
|
||||
// NONE must stand alone, so if any non-NONE setting exist with it, error with conflict
|
||||
if o.isPresent(none) && len(o.values) > 1 {
|
||||
return fmt.Errorf("value NONE cannot be specified alongside other node types")
|
||||
}
|
||||
return nil
|
||||
}
|
||||
|
||||
// Type returns the type of this option
|
||||
func (o *AutoAcceptOption) Type() string {
|
||||
return "auto-accept"
|
||||
}
|
||||
|
||||
// Policies returns a representation of this option for the api
|
||||
func (o *AutoAcceptOption) Policies(secret *string) []swarm.Policy {
|
||||
policies := []swarm.Policy{}
|
||||
for _, p := range defaultPolicies {
|
||||
if len(o.values) != 0 {
|
||||
if _, ok := o.values[string(p.Role)]; ok {
|
||||
p.Autoaccept = true
|
||||
} else {
|
||||
p.Autoaccept = false
|
||||
}
|
||||
}
|
||||
p.Secret = secret
|
||||
policies = append(policies, p)
|
||||
}
|
||||
return policies
|
||||
}
|
||||
|
||||
// isPresent returns whether the key exists in the set or not
|
||||
func (o *AutoAcceptOption) isPresent(key string) bool {
|
||||
_, c := o.values[key]
|
||||
return c
|
||||
}
|
||||
|
||||
// NewAutoAcceptOption returns a new auto-accept option
|
||||
func NewAutoAcceptOption() AutoAcceptOption {
|
||||
return AutoAcceptOption{values: make(map[string]struct{})}
|
||||
}
|
||||
|
||||
// ExternalCAOption is a Value type for parsing external CA specifications.
|
||||
type ExternalCAOption struct {
|
||||
values []*swarm.ExternalCA
|
||||
|
@ -239,8 +160,6 @@ func parseExternalCA(caSpec string) (*swarm.ExternalCA, error) {
|
|||
}
|
||||
|
||||
func addSwarmFlags(flags *pflag.FlagSet, opts *swarmOptions) {
|
||||
flags.Var(&opts.autoAccept, flagAutoAccept, "Auto acceptance policy (worker, manager or none)")
|
||||
flags.StringVar(&opts.secret, flagSecret, "", "Set secret value needed to join a cluster")
|
||||
flags.Int64Var(&opts.taskHistoryLimit, flagTaskHistoryLimit, 10, "Task history retention limit")
|
||||
flags.DurationVar(&opts.dispatcherHeartbeat, flagDispatcherHeartbeat, time.Duration(5*time.Second), "Dispatcher heartbeat period")
|
||||
flags.DurationVar(&opts.nodeCertExpiry, flagCertExpiry, time.Duration(90*24*time.Hour), "Validity period for node certificates")
|
||||
|
@ -249,11 +168,6 @@ func addSwarmFlags(flags *pflag.FlagSet, opts *swarmOptions) {
|
|||
|
||||
func (opts *swarmOptions) ToSpec() swarm.Spec {
|
||||
spec := swarm.Spec{}
|
||||
if opts.secret != "" {
|
||||
spec.AcceptancePolicy.Policies = opts.autoAccept.Policies(&opts.secret)
|
||||
} else {
|
||||
spec.AcceptancePolicy.Policies = opts.autoAccept.Policies(nil)
|
||||
}
|
||||
spec.Orchestration.TaskHistoryRetentionLimit = opts.taskHistoryLimit
|
||||
spec.Dispatcher.HeartbeatPeriod = uint64(opts.dispatcherHeartbeat.Nanoseconds())
|
||||
spec.CAConfig.NodeCertExpiry = opts.nodeCertExpiry
|
||||
|
|
|
@ -4,7 +4,6 @@ import (
|
|||
"testing"
|
||||
|
||||
"github.com/docker/docker/pkg/testutil/assert"
|
||||
"github.com/docker/engine-api/types/swarm"
|
||||
)
|
||||
|
||||
func TestNodeAddrOptionSetHostAndPort(t *testing.T) {
|
||||
|
@ -36,101 +35,3 @@ func TestNodeAddrOptionSetInvalidFormat(t *testing.T) {
|
|||
opt := NewListenAddrOption()
|
||||
assert.Error(t, opt.Set("http://localhost:4545"), "Invalid")
|
||||
}
|
||||
|
||||
func TestAutoAcceptOptionSetWorker(t *testing.T) {
|
||||
opt := NewAutoAcceptOption()
|
||||
assert.NilError(t, opt.Set("worker"))
|
||||
assert.Equal(t, opt.isPresent(worker), true)
|
||||
}
|
||||
|
||||
func TestAutoAcceptOptionSetManager(t *testing.T) {
|
||||
opt := NewAutoAcceptOption()
|
||||
assert.NilError(t, opt.Set("manager"))
|
||||
assert.Equal(t, opt.isPresent(manager), true)
|
||||
}
|
||||
|
||||
func TestAutoAcceptOptionSetInvalid(t *testing.T) {
|
||||
opt := NewAutoAcceptOption()
|
||||
assert.Error(t, opt.Set("bogus"), "must be one / combination")
|
||||
}
|
||||
|
||||
func TestAutoAcceptOptionSetEmpty(t *testing.T) {
|
||||
opt := NewAutoAcceptOption()
|
||||
assert.Error(t, opt.Set(""), "must be one / combination")
|
||||
}
|
||||
|
||||
func TestAutoAcceptOptionSetNone(t *testing.T) {
|
||||
opt := NewAutoAcceptOption()
|
||||
assert.NilError(t, opt.Set("none"))
|
||||
assert.Equal(t, opt.isPresent(manager), false)
|
||||
assert.Equal(t, opt.isPresent(worker), false)
|
||||
}
|
||||
|
||||
func TestAutoAcceptOptionSetTwo(t *testing.T) {
|
||||
opt := NewAutoAcceptOption()
|
||||
assert.NilError(t, opt.Set("worker,manager"))
|
||||
assert.Equal(t, opt.isPresent(manager), true)
|
||||
assert.Equal(t, opt.isPresent(worker), true)
|
||||
}
|
||||
|
||||
func TestAutoAcceptOptionSetConflict(t *testing.T) {
|
||||
opt := NewAutoAcceptOption()
|
||||
assert.Error(t, opt.Set("none,manager"), "value NONE cannot be specified alongside other node types")
|
||||
|
||||
opt = NewAutoAcceptOption()
|
||||
assert.Error(t, opt.Set("none,worker"), "value NONE cannot be specified alongside other node types")
|
||||
|
||||
opt = NewAutoAcceptOption()
|
||||
assert.Error(t, opt.Set("worker,none,manager"), "value NONE cannot be specified alongside other node types")
|
||||
|
||||
opt = NewAutoAcceptOption()
|
||||
assert.Error(t, opt.Set("worker,manager,none"), "value NONE cannot be specified alongside other node types")
|
||||
}
|
||||
|
||||
func TestAutoAcceptOptionPoliciesDefault(t *testing.T) {
|
||||
opt := NewAutoAcceptOption()
|
||||
secret := "thesecret"
|
||||
|
||||
policies := opt.Policies(&secret)
|
||||
assert.Equal(t, len(policies), 2)
|
||||
assert.Equal(t, policies[0], swarm.Policy{
|
||||
Role: worker,
|
||||
Autoaccept: true,
|
||||
Secret: &secret,
|
||||
})
|
||||
assert.Equal(t, policies[1], swarm.Policy{
|
||||
Role: manager,
|
||||
Autoaccept: false,
|
||||
Secret: &secret,
|
||||
})
|
||||
}
|
||||
|
||||
func TestAutoAcceptOptionPoliciesWithManager(t *testing.T) {
|
||||
opt := NewAutoAcceptOption()
|
||||
secret := "thesecret"
|
||||
|
||||
assert.NilError(t, opt.Set("manager"))
|
||||
|
||||
policies := opt.Policies(&secret)
|
||||
assert.Equal(t, len(policies), 2)
|
||||
assert.Equal(t, policies[0], swarm.Policy{
|
||||
Role: worker,
|
||||
Autoaccept: false,
|
||||
Secret: &secret,
|
||||
})
|
||||
assert.Equal(t, policies[1], swarm.Policy{
|
||||
Role: manager,
|
||||
Autoaccept: true,
|
||||
Secret: &secret,
|
||||
})
|
||||
}
|
||||
|
||||
func TestAutoAcceptOptionString(t *testing.T) {
|
||||
opt := NewAutoAcceptOption()
|
||||
assert.NilError(t, opt.Set("manager"))
|
||||
assert.NilError(t, opt.Set("worker"))
|
||||
|
||||
repr := opt.String()
|
||||
assert.Contains(t, repr, "worker=true")
|
||||
assert.Contains(t, repr, "manager=true")
|
||||
}
|
||||
|
|
|
@ -1,19 +0,0 @@
|
|||
package swarm
|
||||
|
||||
import (
|
||||
cryptorand "crypto/rand"
|
||||
"fmt"
|
||||
"math/big"
|
||||
)
|
||||
|
||||
func generateRandomSecret() string {
|
||||
var secretBytes [generatedSecretEntropyBytes]byte
|
||||
|
||||
if _, err := cryptorand.Read(secretBytes[:]); err != nil {
|
||||
panic(fmt.Errorf("failed to read random bytes: %v", err))
|
||||
}
|
||||
|
||||
var nn big.Int
|
||||
nn.SetBytes(secretBytes[:])
|
||||
return fmt.Sprintf("%0[1]*s", maxGeneratedSecretLength, nn.Text(generatedSecretBase))
|
||||
}
|
|
@ -13,11 +13,11 @@ import (
|
|||
)
|
||||
|
||||
func newUpdateCommand(dockerCli *client.DockerCli) *cobra.Command {
|
||||
opts := swarmOptions{autoAccept: NewAutoAcceptOption()}
|
||||
opts := swarmOptions{}
|
||||
|
||||
cmd := &cobra.Command{
|
||||
Use: "update [OPTIONS]",
|
||||
Short: "Update the Swarm",
|
||||
Short: "Update the swarm",
|
||||
Args: cli.NoArgs,
|
||||
RunE: func(cmd *cobra.Command, args []string) error {
|
||||
return runUpdate(dockerCli, cmd.Flags(), opts)
|
||||
|
@ -32,6 +32,8 @@ func runUpdate(dockerCli *client.DockerCli, flags *pflag.FlagSet, opts swarmOpti
|
|||
client := dockerCli.Client()
|
||||
ctx := context.Background()
|
||||
|
||||
var updateFlags swarm.UpdateFlags
|
||||
|
||||
swarm, err := client.SwarmInspect(ctx)
|
||||
if err != nil {
|
||||
return err
|
||||
|
@ -42,7 +44,7 @@ func runUpdate(dockerCli *client.DockerCli, flags *pflag.FlagSet, opts swarmOpti
|
|||
return err
|
||||
}
|
||||
|
||||
err = client.SwarmUpdate(ctx, swarm.Version, swarm.Spec)
|
||||
err = client.SwarmUpdate(ctx, swarm.Version, swarm.Spec, updateFlags)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
@ -55,21 +57,6 @@ func runUpdate(dockerCli *client.DockerCli, flags *pflag.FlagSet, opts swarmOpti
|
|||
func mergeSwarm(swarm *swarm.Swarm, flags *pflag.FlagSet) error {
|
||||
spec := &swarm.Spec
|
||||
|
||||
if flags.Changed(flagAutoAccept) {
|
||||
value := flags.Lookup(flagAutoAccept).Value.(*AutoAcceptOption)
|
||||
spec.AcceptancePolicy.Policies = value.Policies(nil)
|
||||
}
|
||||
|
||||
var psecret *string
|
||||
if flags.Changed(flagSecret) {
|
||||
secret, _ := flags.GetString(flagSecret)
|
||||
psecret = &secret
|
||||
}
|
||||
|
||||
for i := range spec.AcceptancePolicy.Policies {
|
||||
spec.AcceptancePolicy.Policies[i].Secret = psecret
|
||||
}
|
||||
|
||||
if flags.Changed(flagTaskHistoryLimit) {
|
||||
spec.Orchestration.TaskHistoryRetentionLimit, _ = flags.GetInt64(flagTaskHistoryLimit)
|
||||
}
|
||||
|
|
|
@ -85,7 +85,6 @@ func runInfo(dockerCli *client.DockerCli) error {
|
|||
if info.Swarm.ControlAvailable {
|
||||
fmt.Fprintf(dockerCli.Out(), " Managers: %d\n", info.Swarm.Managers)
|
||||
fmt.Fprintf(dockerCli.Out(), " Nodes: %d\n", info.Swarm.Nodes)
|
||||
ioutils.FprintfIfNotEmpty(dockerCli.Out(), " CA Certificate Hash: %s\n", info.Swarm.CACertHash)
|
||||
}
|
||||
}
|
||||
|
||||
|
|
|
@ -11,7 +11,7 @@ type Backend interface {
|
|||
Join(req types.JoinRequest) error
|
||||
Leave(force bool) error
|
||||
Inspect() (types.Swarm, error)
|
||||
Update(uint64, types.Spec) error
|
||||
Update(uint64, types.Spec, types.UpdateFlags) error
|
||||
GetServices(basictypes.ServiceListOptions) ([]types.Service, error)
|
||||
GetService(string) (types.Service, error)
|
||||
CreateService(types.ServiceSpec, string) (string, error)
|
||||
|
|
|
@ -66,7 +66,15 @@ func (sr *swarmRouter) updateCluster(ctx context.Context, w http.ResponseWriter,
|
|||
return fmt.Errorf("Invalid swarm version '%s': %s", rawVersion, err.Error())
|
||||
}
|
||||
|
||||
if err := sr.backend.Update(version, swarm); err != nil {
|
||||
var flags types.UpdateFlags
|
||||
if r.URL.Query().Get("rotate_worker_token") == "true" {
|
||||
flags.RotateWorkerToken = true
|
||||
}
|
||||
if r.URL.Query().Get("rotate_manager_token") == "true" {
|
||||
flags.RotateManagerToken = true
|
||||
}
|
||||
|
||||
if err := sr.backend.Update(version, swarm, flags); err != nil {
|
||||
logrus.Errorf("Error configuring swarm: %v", err)
|
||||
return err
|
||||
}
|
||||
|
|
|
@ -1087,7 +1087,7 @@ __docker_service_subcommand() {
|
|||
"($help)--name=[Service name]:name: "
|
||||
"($help)*--network=[Network attachments]:network: "
|
||||
"($help)*"{-p=,--publish=}"[Publish a port as a node port]:port: "
|
||||
"($help)--registry-auth[Send registry authentication details to Swarm agents]"
|
||||
"($help)--registry-auth[Send registry authentication details to swarm agents]"
|
||||
"($help)--replicas=[Number of tasks]:replicas: "
|
||||
"($help)--reserve-cpu=[Reserve CPUs]:value: "
|
||||
"($help)--reserve-memory=[Reserve Memory]:value: "
|
||||
|
@ -1185,11 +1185,11 @@ __docker_service_subcommand() {
|
|||
__docker_swarm_commands() {
|
||||
local -a _docker_swarm_subcommands
|
||||
_docker_swarm_subcommands=(
|
||||
"init:Initialize a Swarm"
|
||||
"inspect:Inspect the Swarm"
|
||||
"join:Join a Swarm as a node and/or manager"
|
||||
"leave:Leave a Swarm"
|
||||
"update:Update the Swarm"
|
||||
"init:Initialize a swarm"
|
||||
"inspect:Inspect the swarm"
|
||||
"join:Join a swarm as a node and/or manager"
|
||||
"leave:Leave a swarm"
|
||||
"update:Update the swarm"
|
||||
)
|
||||
_describe -t docker-swarm-commands "docker swarm command" _docker_swarm_subcommands
|
||||
}
|
||||
|
|
|
@ -13,7 +13,6 @@ import (
|
|||
"google.golang.org/grpc"
|
||||
|
||||
"github.com/Sirupsen/logrus"
|
||||
"github.com/docker/distribution/digest"
|
||||
"github.com/docker/docker/daemon/cluster/convert"
|
||||
executorpkg "github.com/docker/docker/daemon/cluster/executor"
|
||||
"github.com/docker/docker/daemon/cluster/executor/container"
|
||||
|
@ -42,16 +41,16 @@ const (
|
|||
)
|
||||
|
||||
// ErrNoSwarm is returned on leaving a cluster that was never initialized
|
||||
var ErrNoSwarm = fmt.Errorf("This node is not part of Swarm")
|
||||
var ErrNoSwarm = fmt.Errorf("This node is not part of swarm")
|
||||
|
||||
// ErrSwarmExists is returned on initialize or join request for a cluster that has already been activated
|
||||
var ErrSwarmExists = fmt.Errorf("This node is already part of a Swarm cluster. Use \"docker swarm leave\" to leave this cluster and join another one.")
|
||||
var ErrSwarmExists = fmt.Errorf("This node is already part of a swarm cluster. Use \"docker swarm leave\" to leave this cluster and join another one.")
|
||||
|
||||
// ErrPendingSwarmExists is returned on initialize or join request for a cluster that is already processing a similar request but has not succeeded yet.
|
||||
var ErrPendingSwarmExists = fmt.Errorf("This node is processing an existing join request that has not succeeded yet. Use \"docker swarm leave\" to cancel the current request.")
|
||||
|
||||
// ErrSwarmJoinTimeoutReached is returned when cluster join could not complete before timeout was reached.
|
||||
var ErrSwarmJoinTimeoutReached = fmt.Errorf("Timeout was reached before node was joined. Attempt to join the cluster will continue in the background. Use \"docker info\" command to see the current Swarm status of your node.")
|
||||
var ErrSwarmJoinTimeoutReached = fmt.Errorf("Timeout was reached before node was joined. Attempt to join the cluster will continue in the background. Use \"docker info\" command to see the current swarm status of your node.")
|
||||
|
||||
// defaultSpec contains some sane defaults if cluster options are missing on init
|
||||
var defaultSpec = types.Spec{
|
||||
|
@ -127,7 +126,7 @@ func New(config Config) (*Cluster, error) {
|
|||
return nil, err
|
||||
}
|
||||
|
||||
n, err := c.startNewNode(false, st.ListenAddr, "", "", "", false)
|
||||
n, err := c.startNewNode(false, st.ListenAddr, "", "")
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
|
@ -196,7 +195,7 @@ func (c *Cluster) reconnectOnFailure(n *node) {
|
|||
return
|
||||
}
|
||||
var err error
|
||||
n, err = c.startNewNode(false, c.listenAddr, c.getRemoteAddress(), "", "", false)
|
||||
n, err = c.startNewNode(false, c.listenAddr, c.getRemoteAddress(), "")
|
||||
if err != nil {
|
||||
c.err = err
|
||||
close(n.done)
|
||||
|
@ -205,7 +204,7 @@ func (c *Cluster) reconnectOnFailure(n *node) {
|
|||
}
|
||||
}
|
||||
|
||||
func (c *Cluster) startNewNode(forceNewCluster bool, listenAddr, joinAddr, secret, cahash string, ismanager bool) (*node, error) {
|
||||
func (c *Cluster) startNewNode(forceNewCluster bool, listenAddr, joinAddr, joinToken string) (*node, error) {
|
||||
if err := c.config.Backend.IsSwarmCompatible(); err != nil {
|
||||
return nil, err
|
||||
}
|
||||
|
@ -219,12 +218,10 @@ func (c *Cluster) startNewNode(forceNewCluster bool, listenAddr, joinAddr, secre
|
|||
ListenRemoteAPI: listenAddr,
|
||||
JoinAddr: joinAddr,
|
||||
StateDir: c.root,
|
||||
CAHash: cahash,
|
||||
Secret: secret,
|
||||
JoinToken: joinToken,
|
||||
Executor: container.NewExecutor(c.config.Backend),
|
||||
HeartbeatTick: 1,
|
||||
ElectionTick: 3,
|
||||
IsManager: ismanager,
|
||||
})
|
||||
if err != nil {
|
||||
return nil, err
|
||||
|
@ -291,7 +288,7 @@ func (c *Cluster) Init(req types.InitRequest) (string, error) {
|
|||
if node := c.node; node != nil {
|
||||
if !req.ForceNewCluster {
|
||||
c.Unlock()
|
||||
return "", errSwarmExists(node)
|
||||
return "", ErrSwarmExists
|
||||
}
|
||||
if err := c.stopNode(); err != nil {
|
||||
c.Unlock()
|
||||
|
@ -305,7 +302,7 @@ func (c *Cluster) Init(req types.InitRequest) (string, error) {
|
|||
}
|
||||
|
||||
// todo: check current state existing
|
||||
n, err := c.startNewNode(req.ForceNewCluster, req.ListenAddr, "", "", "", false)
|
||||
n, err := c.startNewNode(req.ForceNewCluster, req.ListenAddr, "", "")
|
||||
if err != nil {
|
||||
c.Unlock()
|
||||
return "", err
|
||||
|
@ -336,40 +333,32 @@ func (c *Cluster) Join(req types.JoinRequest) error {
|
|||
c.Lock()
|
||||
if node := c.node; node != nil {
|
||||
c.Unlock()
|
||||
return errSwarmExists(node)
|
||||
return ErrSwarmExists
|
||||
}
|
||||
if err := validateAndSanitizeJoinRequest(&req); err != nil {
|
||||
c.Unlock()
|
||||
return err
|
||||
}
|
||||
// todo: check current state existing
|
||||
n, err := c.startNewNode(false, req.ListenAddr, req.RemoteAddrs[0], req.Secret, req.CACertHash, req.Manager)
|
||||
n, err := c.startNewNode(false, req.ListenAddr, req.RemoteAddrs[0], req.JoinToken)
|
||||
if err != nil {
|
||||
c.Unlock()
|
||||
return err
|
||||
}
|
||||
c.Unlock()
|
||||
|
||||
certificateRequested := n.CertificateRequested()
|
||||
for {
|
||||
select {
|
||||
case <-certificateRequested:
|
||||
if n.NodeMembership() == swarmapi.NodeMembershipPending {
|
||||
return fmt.Errorf("Your node is in the process of joining the cluster but needs to be accepted by existing cluster member.\nTo accept this node into cluster run \"docker node accept %v\" in an existing cluster manager. Use \"docker info\" command to see the current Swarm status of your node.", n.NodeID())
|
||||
}
|
||||
certificateRequested = nil
|
||||
case <-time.After(swarmConnectTimeout):
|
||||
// attempt to connect will continue in background, also reconnecting
|
||||
go c.reconnectOnFailure(n)
|
||||
return ErrSwarmJoinTimeoutReached
|
||||
case <-n.Ready():
|
||||
go c.reconnectOnFailure(n)
|
||||
return nil
|
||||
case <-n.done:
|
||||
c.RLock()
|
||||
defer c.RUnlock()
|
||||
return c.err
|
||||
}
|
||||
select {
|
||||
case <-time.After(swarmConnectTimeout):
|
||||
// attempt to connect will continue in background, also reconnecting
|
||||
go c.reconnectOnFailure(n)
|
||||
return ErrSwarmJoinTimeoutReached
|
||||
case <-n.Ready():
|
||||
go c.reconnectOnFailure(n)
|
||||
return nil
|
||||
case <-n.done:
|
||||
c.RLock()
|
||||
defer c.RUnlock()
|
||||
return c.err
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -489,7 +478,7 @@ func (c *Cluster) Inspect() (types.Swarm, error) {
|
|||
}
|
||||
|
||||
// Update updates configuration of a managed swarm cluster.
|
||||
func (c *Cluster) Update(version uint64, spec types.Spec) error {
|
||||
func (c *Cluster) Update(version uint64, spec types.Spec, flags types.UpdateFlags) error {
|
||||
c.RLock()
|
||||
defer c.RUnlock()
|
||||
|
||||
|
@ -505,7 +494,7 @@ func (c *Cluster) Update(version uint64, spec types.Spec) error {
|
|||
return err
|
||||
}
|
||||
|
||||
swarmSpec, err := convert.SwarmSpecToGRPCandMerge(spec, &swarm.Spec)
|
||||
swarmSpec, err := convert.SwarmSpecToGRPC(spec)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
@ -518,6 +507,10 @@ func (c *Cluster) Update(version uint64, spec types.Spec) error {
|
|||
ClusterVersion: &swarmapi.Version{
|
||||
Index: version,
|
||||
},
|
||||
Rotation: swarmapi.JoinTokenRotation{
|
||||
RotateWorkerToken: flags.RotateWorkerToken,
|
||||
RotateManagerToken: flags.RotateManagerToken,
|
||||
},
|
||||
},
|
||||
)
|
||||
return err
|
||||
|
@ -611,10 +604,6 @@ func (c *Cluster) Info() types.Info {
|
|||
}
|
||||
}
|
||||
}
|
||||
|
||||
if swarm, err := getSwarm(ctx, c.client); err == nil && swarm != nil {
|
||||
info.CACertHash = swarm.RootCA.CACertHash
|
||||
}
|
||||
}
|
||||
|
||||
if c.node != nil {
|
||||
|
@ -636,12 +625,12 @@ func (c *Cluster) isActiveManager() bool {
|
|||
// Call with read lock.
|
||||
func (c *Cluster) errNoManager() error {
|
||||
if c.node == nil {
|
||||
return fmt.Errorf("This node is not a Swarm manager. Use \"docker swarm init\" or \"docker swarm join --manager\" to connect this node to Swarm and try again.")
|
||||
return fmt.Errorf("This node is not a swarm manager. Use \"docker swarm init\" or \"docker swarm join --manager\" to connect this node to swarm and try again.")
|
||||
}
|
||||
if c.node.Manager() != nil {
|
||||
return fmt.Errorf("This node is not a Swarm manager. Manager is being prepared or has trouble connecting to the cluster.")
|
||||
return fmt.Errorf("This node is not a swarm manager. Manager is being prepared or has trouble connecting to the cluster.")
|
||||
}
|
||||
return fmt.Errorf("This node is not a Swarm manager. Worker nodes can't be used to view or modify cluster state. Please run this command on a manager node or promote the current node to a manager.")
|
||||
return fmt.Errorf("This node is not a swarm manager. Worker nodes can't be used to view or modify cluster state. Please run this command on a manager node or promote the current node to a manager.")
|
||||
}
|
||||
|
||||
// GetServices returns all services of a managed swarm cluster.
|
||||
|
@ -1219,11 +1208,6 @@ func validateAndSanitizeJoinRequest(req *types.JoinRequest) error {
|
|||
return fmt.Errorf("invalid remoteAddr %q: %v", req.RemoteAddrs[i], err)
|
||||
}
|
||||
}
|
||||
if req.CACertHash != "" {
|
||||
if _, err := digest.ParseDigest(req.CACertHash); err != nil {
|
||||
return fmt.Errorf("invalid CACertHash %q, %v", req.CACertHash, err)
|
||||
}
|
||||
}
|
||||
return nil
|
||||
}
|
||||
|
||||
|
@ -1238,13 +1222,6 @@ func validateAddr(addr string) (string, error) {
|
|||
return strings.TrimPrefix(newaddr, "tcp://"), nil
|
||||
}
|
||||
|
||||
func errSwarmExists(node *node) error {
|
||||
if node.NodeMembership() != swarmapi.NodeMembershipAccepted {
|
||||
return ErrPendingSwarmExists
|
||||
}
|
||||
return ErrSwarmExists
|
||||
}
|
||||
|
||||
func initClusterSpec(node *node, spec types.Spec) error {
|
||||
ctx, _ := context.WithTimeout(context.Background(), 5*time.Second)
|
||||
for conn := range node.ListenControlSocket(ctx) {
|
||||
|
@ -1269,7 +1246,7 @@ func initClusterSpec(node *node, spec types.Spec) error {
|
|||
cluster = lcr.Clusters[0]
|
||||
break
|
||||
}
|
||||
newspec, err := convert.SwarmSpecToGRPCandMerge(spec, &cluster.Spec)
|
||||
newspec, err := convert.SwarmSpecToGRPC(spec)
|
||||
if err != nil {
|
||||
return fmt.Errorf("error updating cluster settings: %v", err)
|
||||
}
|
||||
|
|
|
@ -15,7 +15,6 @@ func NodeFromGRPC(n swarmapi.Node) types.Node {
|
|||
ID: n.ID,
|
||||
Spec: types.NodeSpec{
|
||||
Role: types.NodeRole(strings.ToLower(n.Spec.Role.String())),
|
||||
Membership: types.NodeMembership(strings.ToLower(n.Spec.Membership.String())),
|
||||
Availability: types.NodeAvailability(strings.ToLower(n.Spec.Availability.String())),
|
||||
},
|
||||
Status: types.NodeStatus{
|
||||
|
@ -79,12 +78,6 @@ func NodeSpecToGRPC(s types.NodeSpec) (swarmapi.NodeSpec, error) {
|
|||
return swarmapi.NodeSpec{}, fmt.Errorf("invalid Role: %q", s.Role)
|
||||
}
|
||||
|
||||
if membership, ok := swarmapi.NodeSpec_Membership_value[strings.ToUpper(string(s.Membership))]; ok {
|
||||
spec.Membership = swarmapi.NodeSpec_Membership(membership)
|
||||
} else {
|
||||
return swarmapi.NodeSpec{}, fmt.Errorf("invalid Membership: %q", s.Membership)
|
||||
}
|
||||
|
||||
if availability, ok := swarmapi.NodeSpec_Availability_value[strings.ToUpper(string(s.Availability))]; ok {
|
||||
spec.Availability = swarmapi.NodeSpec_Availability(availability)
|
||||
} else {
|
||||
|
|
|
@ -5,8 +5,6 @@ import (
|
|||
"strings"
|
||||
"time"
|
||||
|
||||
"golang.org/x/crypto/bcrypt"
|
||||
|
||||
types "github.com/docker/engine-api/types/swarm"
|
||||
swarmapi "github.com/docker/swarmkit/api"
|
||||
"github.com/docker/swarmkit/protobuf/ptypes"
|
||||
|
@ -28,6 +26,10 @@ func SwarmFromGRPC(c swarmapi.Cluster) types.Swarm {
|
|||
ElectionTick: c.Spec.Raft.ElectionTick,
|
||||
},
|
||||
},
|
||||
JoinTokens: types.JoinTokens{
|
||||
Worker: c.RootCA.JoinTokens.Worker,
|
||||
Manager: c.RootCA.JoinTokens.Manager,
|
||||
},
|
||||
}
|
||||
|
||||
heartbeatPeriod, _ := ptypes.Duration(c.Spec.Dispatcher.HeartbeatPeriod)
|
||||
|
@ -52,23 +54,11 @@ func SwarmFromGRPC(c swarmapi.Cluster) types.Swarm {
|
|||
swarm.Spec.Name = c.Spec.Annotations.Name
|
||||
swarm.Spec.Labels = c.Spec.Annotations.Labels
|
||||
|
||||
for _, policy := range c.Spec.AcceptancePolicy.Policies {
|
||||
p := types.Policy{
|
||||
Role: types.NodeRole(strings.ToLower(policy.Role.String())),
|
||||
Autoaccept: policy.Autoaccept,
|
||||
}
|
||||
if policy.Secret != nil {
|
||||
secret := string(policy.Secret.Data)
|
||||
p.Secret = &secret
|
||||
}
|
||||
swarm.Spec.AcceptancePolicy.Policies = append(swarm.Spec.AcceptancePolicy.Policies, p)
|
||||
}
|
||||
|
||||
return swarm
|
||||
}
|
||||
|
||||
// SwarmSpecToGRPCandMerge converts a Spec to a grpc ClusterSpec and merge AcceptancePolicy from an existing grpc ClusterSpec if provided.
|
||||
func SwarmSpecToGRPCandMerge(s types.Spec, existingSpec *swarmapi.ClusterSpec) (swarmapi.ClusterSpec, error) {
|
||||
// SwarmSpecToGRPC converts a Spec to a grpc ClusterSpec.
|
||||
func SwarmSpecToGRPC(s types.Spec) (swarmapi.ClusterSpec, error) {
|
||||
spec := swarmapi.ClusterSpec{
|
||||
Annotations: swarmapi.Annotations{
|
||||
Name: s.Name,
|
||||
|
@ -104,63 +94,5 @@ func SwarmSpecToGRPCandMerge(s types.Spec, existingSpec *swarmapi.ClusterSpec) (
|
|||
})
|
||||
}
|
||||
|
||||
if err := SwarmSpecUpdateAcceptancePolicy(&spec, s.AcceptancePolicy, existingSpec); err != nil {
|
||||
return swarmapi.ClusterSpec{}, err
|
||||
}
|
||||
|
||||
return spec, nil
|
||||
}
|
||||
|
||||
// SwarmSpecUpdateAcceptancePolicy updates a grpc ClusterSpec using AcceptancePolicy.
|
||||
func SwarmSpecUpdateAcceptancePolicy(spec *swarmapi.ClusterSpec, acceptancePolicy types.AcceptancePolicy, oldSpec *swarmapi.ClusterSpec) error {
|
||||
spec.AcceptancePolicy.Policies = nil
|
||||
hashs := make(map[string][]byte)
|
||||
|
||||
for _, p := range acceptancePolicy.Policies {
|
||||
role, ok := swarmapi.NodeRole_value[strings.ToUpper(string(p.Role))]
|
||||
if !ok {
|
||||
return fmt.Errorf("invalid Role: %q", p.Role)
|
||||
}
|
||||
|
||||
policy := &swarmapi.AcceptancePolicy_RoleAdmissionPolicy{
|
||||
Role: swarmapi.NodeRole(role),
|
||||
Autoaccept: p.Autoaccept,
|
||||
}
|
||||
|
||||
if p.Secret != nil {
|
||||
if *p.Secret == "" { // if provided secret is empty, it means erase previous secret.
|
||||
policy.Secret = nil
|
||||
} else { // if provided secret is not empty, we generate a new one.
|
||||
hashPwd, ok := hashs[*p.Secret]
|
||||
if !ok {
|
||||
hashPwd, _ = bcrypt.GenerateFromPassword([]byte(*p.Secret), 0)
|
||||
hashs[*p.Secret] = hashPwd
|
||||
}
|
||||
policy.Secret = &swarmapi.AcceptancePolicy_RoleAdmissionPolicy_Secret{
|
||||
Data: hashPwd,
|
||||
Alg: "bcrypt",
|
||||
}
|
||||
}
|
||||
} else if oldSecret := getOldSecret(oldSpec, policy.Role); oldSecret != nil { // else use the old one.
|
||||
policy.Secret = &swarmapi.AcceptancePolicy_RoleAdmissionPolicy_Secret{
|
||||
Data: oldSecret.Data,
|
||||
Alg: oldSecret.Alg,
|
||||
}
|
||||
}
|
||||
|
||||
spec.AcceptancePolicy.Policies = append(spec.AcceptancePolicy.Policies, policy)
|
||||
}
|
||||
return nil
|
||||
}
|
||||
|
||||
func getOldSecret(oldSpec *swarmapi.ClusterSpec, role swarmapi.NodeRole) *swarmapi.AcceptancePolicy_RoleAdmissionPolicy_Secret {
|
||||
if oldSpec == nil {
|
||||
return nil
|
||||
}
|
||||
for _, p := range oldSpec.AcceptancePolicy.Policies {
|
||||
if p.Role == role {
|
||||
return p.Secret
|
||||
}
|
||||
}
|
||||
return nil
|
||||
}
|
||||
|
|
|
@ -3351,7 +3351,6 @@ List nodes
|
|||
"UpdatedAt": "2016-06-07T20:31:11.999868824Z",
|
||||
"Spec": {
|
||||
"Role": "MANAGER",
|
||||
"Membership": "ACCEPTED",
|
||||
"Availability": "ACTIVE"
|
||||
},
|
||||
"Description": {
|
||||
|
@ -3481,7 +3480,6 @@ Return low-level information on the node `id`
|
|||
"UpdatedAt": "2016-06-07T20:31:11.999868824Z",
|
||||
"Spec": {
|
||||
"Role": "MANAGER",
|
||||
"Membership": "ACCEPTED",
|
||||
"Availability": "ACTIVE"
|
||||
},
|
||||
"Description": {
|
||||
|
@ -3595,18 +3593,6 @@ Initialize a new Swarm
|
|||
"ListenAddr": "0.0.0.0:4500",
|
||||
"ForceNewCluster": false,
|
||||
"Spec": {
|
||||
"AcceptancePolicy": {
|
||||
"Policies": [
|
||||
{
|
||||
"Role": "MANAGER",
|
||||
"Autoaccept": false
|
||||
},
|
||||
{
|
||||
"Role": "WORKER",
|
||||
"Autoaccept": true
|
||||
}
|
||||
]
|
||||
},
|
||||
"Orchestration": {},
|
||||
"Raft": {},
|
||||
"Dispatcher": {},
|
||||
|
@ -3676,9 +3662,7 @@ Join an existing new Swarm
|
|||
{
|
||||
"ListenAddr": "0.0.0.0:4500",
|
||||
"RemoteAddrs": ["node1:4500"],
|
||||
"Secret": "",
|
||||
"CACertHash": "",
|
||||
"Manager": false
|
||||
"JoinToken": "SWMTKN-1-3pu6hszjas19xyp7ghgosyx9k8atbfcr8p2is99znpy26u2lkl-7p73s1dx5in4tatdymyhg9hu2"
|
||||
}
|
||||
|
||||
**Example response**:
|
||||
|
@ -3698,9 +3682,7 @@ JSON Parameters:
|
|||
- **ListenAddr** – Listen address used for inter-manager communication if the node gets promoted to
|
||||
manager, as well as determining the networking interface used for the VXLAN Tunnel Endpoint (VTEP).
|
||||
- **RemoteAddr** – Address of any manager node already participating in the Swarm to join.
|
||||
- **Secret** – Secret token for joining this Swarm.
|
||||
- **CACertHash** – Optional hash of the root CA to avoid relying on trust on first use.
|
||||
- **Manager** – Directly join as a manager (only for a Swarm configured to autoaccept managers).
|
||||
- **JoinToken** – Secret token for joining this Swarm.
|
||||
|
||||
### Leave a Swarm
|
||||
|
||||
|
@ -3741,18 +3723,6 @@ Update a Swarm
|
|||
|
||||
{
|
||||
"Name": "default",
|
||||
"AcceptancePolicy": {
|
||||
"Policies": [
|
||||
{
|
||||
"Role": "WORKER",
|
||||
"Autoaccept": false
|
||||
},
|
||||
{
|
||||
"Role": "MANAGER",
|
||||
"Autoaccept": false
|
||||
}
|
||||
]
|
||||
},
|
||||
"Orchestration": {
|
||||
"TaskHistoryRetentionLimit": 10
|
||||
},
|
||||
|
@ -3767,6 +3737,10 @@ Update a Swarm
|
|||
},
|
||||
"CAConfig": {
|
||||
"NodeCertExpiry": 7776000000000000
|
||||
},
|
||||
"JoinTokens": {
|
||||
"Worker": "SWMTKN-1-3pu6hszjas19xyp7ghgosyx9k8atbfcr8p2is99znpy26u2lkl-1awxwuwd3z9j1z3puu7rcgdbx",
|
||||
"Manager": "SWMTKN-1-3pu6hszjas19xyp7ghgosyx9k8atbfcr8p2is99znpy26u2lkl-7p73s1dx5in4tatdymyhg9hu2"
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -3777,6 +3751,13 @@ Update a Swarm
|
|||
Content-Length: 0
|
||||
Content-Type: text/plain; charset=utf-8
|
||||
|
||||
**Query parameters**:
|
||||
|
||||
- **version** – The version number of the swarm object being updated. This is
|
||||
required to avoid conflicting writes.
|
||||
- **rotate_worker_token** - Set to `true` to rotate the worker join token.
|
||||
- **rotate_manager_token** - Set to `true` to rotate the manager join token.
|
||||
|
||||
**Status codes**:
|
||||
|
||||
- **200** – no error
|
||||
|
@ -3785,11 +3766,6 @@ Update a Swarm
|
|||
|
||||
JSON Parameters:
|
||||
|
||||
- **Policies** – An array of acceptance policies.
|
||||
- **Role** – The role that policy applies to (`MANAGER` or `WORKER`)
|
||||
- **Autoaccept** – A boolean indicating whether nodes joining for that role should be
|
||||
automatically accepted in the Swarm.
|
||||
- **Secret** – An optional secret to provide for nodes to join the Swarm.
|
||||
- **Orchestration** – Configuration settings for the orchestration aspects of the Swarm.
|
||||
- **TaskHistoryRetentionLimit** – Maximum number of tasks history stored.
|
||||
- **Raft** – Raft related configuration.
|
||||
|
@ -3811,6 +3787,9 @@ JSON Parameters:
|
|||
- **URL** - URL where certificate signing requests should be sent.
|
||||
- **Options** - An object with key/value pairs that are interpreted
|
||||
as protocol-specific options for the external CA driver.
|
||||
- **JoinTokens** - Tokens that can be used by other nodes to join the Swarm.
|
||||
- **Worker** - Token to use for joining as a worker.
|
||||
- **Manager** - Token to use for joining as a manager.
|
||||
|
||||
## 3.8 Services
|
||||
|
||||
|
@ -4292,6 +4271,10 @@ Update the service `id`.
|
|||
of: `"Ports": { "<port>/<tcp|udp>: {}" }`
|
||||
- **VirtualIPs**
|
||||
|
||||
**Query parameters**:
|
||||
|
||||
- **version** – The version number of the service object being updated. This is
|
||||
required to avoid conflicting writes.
|
||||
|
||||
**Status codes**:
|
||||
|
||||
|
|
|
@ -3352,7 +3352,6 @@ List nodes
|
|||
"UpdatedAt": "2016-06-07T20:31:11.999868824Z",
|
||||
"Spec": {
|
||||
"Role": "MANAGER",
|
||||
"Membership": "ACCEPTED",
|
||||
"Availability": "ACTIVE"
|
||||
},
|
||||
"Description": {
|
||||
|
@ -3482,7 +3481,6 @@ Return low-level information on the node `id`
|
|||
"UpdatedAt": "2016-06-07T20:31:11.999868824Z",
|
||||
"Spec": {
|
||||
"Role": "MANAGER",
|
||||
"Membership": "ACCEPTED",
|
||||
"Availability": "ACTIVE"
|
||||
},
|
||||
"Description": {
|
||||
|
@ -3596,18 +3594,6 @@ Initialize a new Swarm
|
|||
"ListenAddr": "0.0.0.0:4500",
|
||||
"ForceNewCluster": false,
|
||||
"Spec": {
|
||||
"AcceptancePolicy": {
|
||||
"Policies": [
|
||||
{
|
||||
"Role": "MANAGER",
|
||||
"Autoaccept": false
|
||||
},
|
||||
{
|
||||
"Role": "WORKER",
|
||||
"Autoaccept": true
|
||||
}
|
||||
]
|
||||
},
|
||||
"Orchestration": {},
|
||||
"Raft": {},
|
||||
"Dispatcher": {},
|
||||
|
@ -3677,9 +3663,7 @@ Join an existing new Swarm
|
|||
{
|
||||
"ListenAddr": "0.0.0.0:4500",
|
||||
"RemoteAddrs": ["node1:4500"],
|
||||
"Secret": "",
|
||||
"CACertHash": "",
|
||||
"Manager": false
|
||||
"JoinToken": "SWMTKN-1-3pu6hszjas19xyp7ghgosyx9k8atbfcr8p2is99znpy26u2lkl-7p73s1dx5in4tatdymyhg9hu2"
|
||||
}
|
||||
|
||||
**Example response**:
|
||||
|
@ -3699,9 +3683,7 @@ JSON Parameters:
|
|||
- **ListenAddr** – Listen address used for inter-manager communication if the node gets promoted to
|
||||
manager, as well as determining the networking interface used for the VXLAN Tunnel Endpoint (VTEP).
|
||||
- **RemoteAddr** – Address of any manager node already participating in the Swarm to join.
|
||||
- **Secret** – Secret token for joining this Swarm.
|
||||
- **CACertHash** – Optional hash of the root CA to avoid relying on trust on first use.
|
||||
- **Manager** – Directly join as a manager (only for a Swarm configured to autoaccept managers).
|
||||
- **JoinToken** – Secret token for joining this Swarm.
|
||||
|
||||
### Leave a Swarm
|
||||
|
||||
|
@ -3742,18 +3724,6 @@ Update a Swarm
|
|||
|
||||
{
|
||||
"Name": "default",
|
||||
"AcceptancePolicy": {
|
||||
"Policies": [
|
||||
{
|
||||
"Role": "WORKER",
|
||||
"Autoaccept": false
|
||||
},
|
||||
{
|
||||
"Role": "MANAGER",
|
||||
"Autoaccept": false
|
||||
}
|
||||
]
|
||||
},
|
||||
"Orchestration": {
|
||||
"TaskHistoryRetentionLimit": 10
|
||||
},
|
||||
|
@ -3768,6 +3738,10 @@ Update a Swarm
|
|||
},
|
||||
"CAConfig": {
|
||||
"NodeCertExpiry": 7776000000000000
|
||||
},
|
||||
"JoinTokens": {
|
||||
"Worker": "SWMTKN-1-3pu6hszjas19xyp7ghgosyx9k8atbfcr8p2is99znpy26u2lkl-1awxwuwd3z9j1z3puu7rcgdbx",
|
||||
"Manager": "SWMTKN-1-3pu6hszjas19xyp7ghgosyx9k8atbfcr8p2is99znpy26u2lkl-7p73s1dx5in4tatdymyhg9hu2"
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -3778,6 +3752,13 @@ Update a Swarm
|
|||
Content-Length: 0
|
||||
Content-Type: text/plain; charset=utf-8
|
||||
|
||||
**Query parameters**:
|
||||
|
||||
- **version** – The version number of the swarm object being updated. This is
|
||||
required to avoid conflicting writes.
|
||||
- **rotate_worker_token** - Set to `true` to rotate the worker join token.
|
||||
- **rotate_manager_token** - Set to `true` to rotate the manager join token.
|
||||
|
||||
**Status codes**:
|
||||
|
||||
- **200** – no error
|
||||
|
@ -3786,11 +3767,6 @@ Update a Swarm
|
|||
|
||||
JSON Parameters:
|
||||
|
||||
- **Policies** – An array of acceptance policies.
|
||||
- **Role** – The role that policy applies to (`MANAGER` or `WORKER`)
|
||||
- **Autoaccept** – A boolean indicating whether nodes joining for that role should be
|
||||
automatically accepted in the Swarm.
|
||||
- **Secret** – An optional secret to provide for nodes to join the Swarm.
|
||||
- **Orchestration** – Configuration settings for the orchestration aspects of the Swarm.
|
||||
- **TaskHistoryRetentionLimit** – Maximum number of tasks history stored.
|
||||
- **Raft** – Raft related configuration.
|
||||
|
@ -3812,6 +3788,9 @@ JSON Parameters:
|
|||
- **URL** - URL where certificate signing requests should be sent.
|
||||
- **Options** - An object with key/value pairs that are interpreted
|
||||
as protocol-specific options for the external CA driver.
|
||||
- **JoinTokens** - Tokens that can be used by other nodes to join the Swarm.
|
||||
- **Worker** - Token to use for joining as a worker.
|
||||
- **Manager** - Token to use for joining as a manager.
|
||||
|
||||
## 3.8 Services
|
||||
|
||||
|
@ -4293,6 +4272,10 @@ Update the service `id`.
|
|||
of: `"Ports": { "<port>/<tcp|udp>: {}" }`
|
||||
- **VirtualIPs**
|
||||
|
||||
**Query parameters**:
|
||||
|
||||
- **version** – The version number of the service object being updated. This is
|
||||
required to avoid conflicting writes.
|
||||
|
||||
**Status codes**:
|
||||
|
||||
|
|
|
@ -19,7 +19,7 @@ Create and update a stack from a Distributed Application Bundle (DAB)
|
|||
Options:
|
||||
--file string Path to a Distributed Application Bundle file (Default: STACK.dab)
|
||||
--help Print usage
|
||||
--registry-auth Send registry authentication details to Swarm agents
|
||||
--registry-auth Send registry authentication details to swarm agents
|
||||
```
|
||||
|
||||
Create and update a stack from a `dab` file. This command has to be
|
||||
|
|
|
@ -111,7 +111,6 @@ read the [`dockerd`](dockerd.md) reference page.
|
|||
|
||||
| Command | Description |
|
||||
|:--------|:-------------------------------------------------------------------|
|
||||
| [node accept](node_accept.md) | Accept a node into the swarm |
|
||||
| [node promote](node_promote.md) | Promote a node that is pending a promotion to manager |
|
||||
| [node demote](node_demote.md) | Demotes an existing manager so that it is no longer a manager |
|
||||
| [node inspect](node_inspect.md) | Inspect a node in the swarm |
|
||||
|
@ -124,10 +123,11 @@ read the [`dockerd`](dockerd.md) reference page.
|
|||
|
||||
| Command | Description |
|
||||
|:--------|:-------------------------------------------------------------------|
|
||||
| [swarm init](swarm_init.md) | Initialize a Swarm |
|
||||
| [swarm join](swarm_join.md) | Join a Swarm as a manager node or worker node |
|
||||
| [swarm init](swarm_init.md) | Initialize a swarm |
|
||||
| [swarm join](swarm_join.md) | Join a swarm as a manager node or worker node |
|
||||
| [swarm leave](swarm_leave.md) | Remove the current node from the swarm |
|
||||
| [swarm update](swarm_update.md) | Update attributes of a swarm |
|
||||
| [swarm join-token](swarm_join_token.md) | Display or rotate join tokens |
|
||||
|
||||
### Swarm service commands
|
||||
|
||||
|
|
|
@ -38,7 +38,7 @@ available on the volume where `/var/lib/docker` is mounted.
|
|||
## Display Docker system information
|
||||
|
||||
Here is a sample output for a daemon running on Ubuntu, using the overlay
|
||||
storage driver and a node that is part of a 2 node Swarm cluster:
|
||||
storage driver and a node that is part of a 2 node swarm cluster:
|
||||
|
||||
$ docker -D info
|
||||
Containers: 14
|
||||
|
|
|
@ -1,32 +0,0 @@
|
|||
<!--[metadata]>
|
||||
+++
|
||||
title = "node accept"
|
||||
description = "The node accept command description and usage"
|
||||
keywords = ["node, accept"]
|
||||
[menu.main]
|
||||
parent = "smn_cli"
|
||||
+++
|
||||
<![end-metadata]-->
|
||||
|
||||
# node accept
|
||||
|
||||
```markdown
|
||||
Usage: docker node accept NODE [NODE...]
|
||||
|
||||
Accept a node in the swarm
|
||||
|
||||
Options:
|
||||
--help Print usage
|
||||
```
|
||||
|
||||
Accept a node into the swarm. This command targets a docker engine that is a manager in the swarm cluster.
|
||||
|
||||
|
||||
```bash
|
||||
$ docker node accept <node name>
|
||||
```
|
||||
|
||||
## Related information
|
||||
|
||||
* [node promote](node_promote.md)
|
||||
* [node demote](node_demote.md)
|
|
@ -29,5 +29,4 @@ $ docker node demote <node name>
|
|||
|
||||
## Related information
|
||||
|
||||
* [node accept](node_accept.md)
|
||||
* [node promote](node_promote.md)
|
||||
|
|
|
@ -41,7 +41,6 @@ Example output:
|
|||
"UpdatedAt": "2016-06-16T22:52:45.230878043Z",
|
||||
"Spec": {
|
||||
"Role": "manager",
|
||||
"Membership": "accepted",
|
||||
"Availability": "active"
|
||||
},
|
||||
"Description": {
|
||||
|
|
|
@ -30,10 +30,10 @@ Lists all the nodes that the Docker Swarm manager knows about. You can filter us
|
|||
Example output:
|
||||
|
||||
$ docker node ls
|
||||
ID HOSTNAME MEMBERSHIP STATUS AVAILABILITY MANAGER STATUS LEADER
|
||||
1bcef6utixb0l0ca7gxuivsj0 swarm-worker2 Accepted Ready Active
|
||||
38ciaotwjuritcdtn9npbnkuz swarm-worker1 Accepted Ready Active
|
||||
e216jshn25ckzbvmwlnh5jr3g * swarm-manager1 Accepted Ready Active Reachable Yes
|
||||
ID HOSTNAME STATUS AVAILABILITY MANAGER STATUS
|
||||
1bcef6utixb0l0ca7gxuivsj0 swarm-worker2 Ready Active
|
||||
38ciaotwjuritcdtn9npbnkuz swarm-worker1 Ready Active
|
||||
e216jshn25ckzbvmwlnh5jr3g * swarm-manager1 Ready Active Leader
|
||||
|
||||
|
||||
## Filtering
|
||||
|
@ -54,16 +54,16 @@ The `name` filter matches on all or part of a node name.
|
|||
The following filter matches the node with a name equal to `swarm-master` string.
|
||||
|
||||
$ docker node ls -f name=swarm-manager1
|
||||
ID HOSTNAME MEMBERSHIP STATUS AVAILABILITY MANAGER STATUS LEADER
|
||||
e216jshn25ckzbvmwlnh5jr3g * swarm-manager1 Accepted Ready Active Reachable Yes
|
||||
ID HOSTNAME STATUS AVAILABILITY MANAGER STATUS
|
||||
e216jshn25ckzbvmwlnh5jr3g * swarm-manager1 Ready Active Leader
|
||||
|
||||
### id
|
||||
|
||||
The `id` filter matches all or part of a node's id.
|
||||
|
||||
$ docker node ls -f id=1
|
||||
ID HOSTNAME MEMBERSHIP STATUS AVAILABILITY MANAGER STATUS LEADER
|
||||
1bcef6utixb0l0ca7gxuivsj0 swarm-worker2 Accepted Ready Active
|
||||
ID HOSTNAME STATUS AVAILABILITY MANAGER STATUS
|
||||
1bcef6utixb0l0ca7gxuivsj0 swarm-worker2 Ready Active
|
||||
|
||||
|
||||
#### label
|
||||
|
@ -75,8 +75,8 @@ The following filter matches nodes with the `usage` label regardless of its valu
|
|||
|
||||
```bash
|
||||
$ docker node ls -f "label=foo"
|
||||
ID HOSTNAME MEMBERSHIP STATUS AVAILABILITY MANAGER STATUS LEADER
|
||||
1bcef6utixb0l0ca7gxuivsj0 swarm-worker2 Accepted Ready Active
|
||||
ID HOSTNAME STATUS AVAILABILITY MANAGER STATUS
|
||||
1bcef6utixb0l0ca7gxuivsj0 swarm-worker2 Ready Active
|
||||
```
|
||||
|
||||
|
||||
|
|
|
@ -28,5 +28,4 @@ $ docker node promote <node name>
|
|||
|
||||
## Related information
|
||||
|
||||
* [node accept](node_accept.md)
|
||||
* [node demote](node_demote.md)
|
||||
|
|
|
@ -23,14 +23,13 @@ Options:
|
|||
--help Print usage
|
||||
```
|
||||
|
||||
Removes specified nodes from a swarm. Rejects nodes with `Pending`
|
||||
membership from the swarm.
|
||||
Removes specified nodes from a swarm.
|
||||
|
||||
|
||||
Example output:
|
||||
|
||||
$ docker node rm swarm-node-02
|
||||
Node swarm-node-02 removed from Swarm
|
||||
Node swarm-node-02 removed from swarm
|
||||
|
||||
|
||||
## Related information
|
||||
|
|
|
@ -21,7 +21,6 @@ Options:
|
|||
--help Print usage
|
||||
--label-add value Add or update a node label (key=value) (default [])
|
||||
--label-rm value Remove a node label if exists (default [])
|
||||
--membership string Membership of the node (accepted/rejected)
|
||||
--role string Role of the node (worker/manager)
|
||||
```
|
||||
|
||||
|
|
|
@ -31,7 +31,7 @@ Options:
|
|||
--name string Service name
|
||||
--network value Network attachments (default [])
|
||||
-p, --publish value Publish a port as a node port (default [])
|
||||
--registry-auth Send registry authentication details to Swarm agents
|
||||
--registry-auth Send registry authentication details to swarm agents
|
||||
--replicas value Number of tasks (default none)
|
||||
--reserve-cpu value Reserve CPUs (default 0.000)
|
||||
--reserve-memory value Reserve Memory (default 0 B)
|
||||
|
|
|
@ -38,7 +38,7 @@ Options:
|
|||
--network-rm value Remove a network by name (default [])
|
||||
--publish-add value Add or update a published port (default [])
|
||||
--publish-rm value Remove a published port by its target port (default [])
|
||||
--registry-auth Send registry authentication details to Swarm agents
|
||||
--registry-auth Send registry authentication details to swarm agents
|
||||
--replicas value Number of tasks (default none)
|
||||
--reserve-cpu value Reserve CPUs (default 0.000)
|
||||
--reserve-memory value Reserve Memory (default 0 B)
|
||||
|
|
|
@ -14,74 +14,43 @@ parent = "smn_cli"
|
|||
```markdown
|
||||
Usage: docker swarm init [OPTIONS]
|
||||
|
||||
Initialize a Swarm
|
||||
Initialize a swarm
|
||||
|
||||
Options:
|
||||
--auto-accept value Auto acceptance policy (default worker)
|
||||
--cert-expiry duration Validity period for node certificates (default 2160h0m0s)
|
||||
--dispatcher-heartbeat duration Dispatcher heartbeat period (default 5s)
|
||||
--external-ca value Specifications of one or more certificate signing endpoints
|
||||
--force-new-cluster Force create a new cluster from current state.
|
||||
--help Print usage
|
||||
--listen-addr value Listen address (default 0.0.0.0:2377)
|
||||
--secret string Set secret value needed to accept nodes into cluster
|
||||
--task-history-limit int Task history retention limit (default 10)
|
||||
```
|
||||
|
||||
Initialize a Swarm cluster. The docker engine targeted by this command becomes a manager
|
||||
in the newly created one node Swarm cluster.
|
||||
Initialize a swarm cluster. The docker engine targeted by this command becomes a manager
|
||||
in the newly created one node swarm cluster.
|
||||
|
||||
|
||||
```bash
|
||||
$ docker swarm init --listen-addr 192.168.99.121:2377
|
||||
No --secret provided. Generated random secret:
|
||||
4ao565v9jsuogtq5t8s379ulb
|
||||
|
||||
Swarm initialized: current node (1ujecd0j9n3ro9i6628smdmth) is now a manager.
|
||||
Swarm initialized: current node (bvz81updecsj6wjz393c09vti) is now a manager.
|
||||
|
||||
To add a worker to this swarm, run the following command:
|
||||
docker swarm join --secret 4ao565v9jsuogtq5t8s379ulb \
|
||||
--ca-hash sha256:07ce22bd1a7619f2adc0d63bd110479a170e7c4e69df05b67a1aa2705c88ef09 \
|
||||
192.168.99.121:2377
|
||||
$ docker node ls
|
||||
ID HOSTNAME MEMBERSHIP STATUS AVAILABILITY MANAGER STATUS LEADER
|
||||
1ujecd0j9n3ro9i6628smdmth * manager1 Accepted Ready Active Reachable Yes
|
||||
docker swarm join \
|
||||
--token SWMTKN-1-3pu6hszjas19xyp7ghgosyx9k8atbfcr8p2is99znpy26u2lkl-1awxwuwd3z9j1z3puu7rcgdbx \
|
||||
172.17.0.2:2377
|
||||
|
||||
To add a manager to this swarm, run the following command:
|
||||
docker swarm join \
|
||||
--token SWMTKN-1-3pu6hszjas19xyp7ghgosyx9k8atbfcr8p2is99znpy26u2lkl-7p73s1dx5in4tatdymyhg9hu2 \
|
||||
172.17.0.2:2377
|
||||
```
|
||||
|
||||
If a secret for joining new nodes is not provided with `--secret`, `docker swarm init` will
|
||||
generate a random one and print it to the terminal (as seen in the example above). To initialize
|
||||
a swarm with no secret, use `--secret ""`.
|
||||
`docker swarm init` generates two random tokens, a worker token and a manager token. When you join
|
||||
a new node to the swarm, the node joins as a worker or manager node based upon the token you pass
|
||||
to [swarm join](swarm_join.md).
|
||||
|
||||
### `--auto-accept value`
|
||||
|
||||
This flag controls node acceptance into the cluster. By default, `worker` nodes are
|
||||
automatically accepted by the cluster. This can be changed by specifying what kinds of nodes
|
||||
can be auto-accepted into the cluster. If auto-accept is not turned on, then
|
||||
[node accept](node_accept.md) can be used to explicitly accept a node into the cluster.
|
||||
|
||||
For example, the following initializes a cluster with auto-acceptance of workers, but not managers
|
||||
|
||||
|
||||
```bash
|
||||
$ docker swarm init --listen-addr 192.168.99.121:2377 --auto-accept worker
|
||||
```
|
||||
|
||||
It is possible to pass a comma-separated list of node types. The following initializes a cluster
|
||||
with auto-acceptance of both `worker` and `manager` nodes
|
||||
|
||||
```bash
|
||||
$ docker swarm init --listen-addr 192.168.99.121:2377 --auto-accept worker,manager
|
||||
```
|
||||
|
||||
To disable auto acceptance, use the `none` option. Note that this option cannot
|
||||
be combined with other values. When disabling auto acceptance, nodes must be
|
||||
manually accepted or rejected using `docker node accept` or `docker node rm`.
|
||||
|
||||
The following example enables swarm mode with auto acceptance disabled:
|
||||
|
||||
```bash
|
||||
$ docker swarm init --listen-addr 192.168.99.121:2377 --auto-accept none
|
||||
```
|
||||
After you create the swarm, you can display or rotate the token using
|
||||
[swarm join-token](swarm_join_token.md).
|
||||
|
||||
### `--cert-expiry`
|
||||
|
||||
|
@ -105,11 +74,7 @@ This flag forces an existing node that was part of a quorum that was lost to res
|
|||
|
||||
### `--listen-addr value`
|
||||
|
||||
The node listens for inbound Swarm manager traffic on this IP:PORT
|
||||
|
||||
### `--secret string`
|
||||
|
||||
Secret value needed to accept nodes into the Swarm
|
||||
The node listens for inbound swarm manager traffic on this IP:PORT
|
||||
|
||||
### `--task-history-limit`
|
||||
|
||||
|
@ -120,5 +85,5 @@ This flag sets up task history retention limit.
|
|||
* [swarm join](swarm_join.md)
|
||||
* [swarm leave](swarm_leave.md)
|
||||
* [swarm update](swarm_update.md)
|
||||
* [node accept](node_accept.md)
|
||||
* [swarm join-token](swarm_join_token.md)
|
||||
* [node rm](node_rm.md)
|
||||
|
|
|
@ -14,55 +14,54 @@ parent = "smn_cli"
|
|||
```markdown
|
||||
Usage: docker swarm join [OPTIONS] HOST:PORT
|
||||
|
||||
Join a Swarm as a node and/or manager
|
||||
Join a swarm as a node and/or manager
|
||||
|
||||
Options:
|
||||
--ca-hash string Hash of the Root Certificate Authority certificate used for trusted join
|
||||
--help Print usage
|
||||
--listen-addr value Listen address (default 0.0.0.0:2377)
|
||||
--manager Try joining as a manager.
|
||||
--secret string Secret for node acceptance
|
||||
--token string Token for entry into the swarm
|
||||
```
|
||||
|
||||
Join a node to a Swarm cluster. If the `--manager` flag is specified, the docker engine
|
||||
targeted by this command becomes a `manager`. If it is not specified, it becomes a `worker`.
|
||||
Join a node to a swarm. The node joins as a manager node or worker node based upon the token you
|
||||
pass with the `--token` flag. If you pass a manager token, the node joins as a manager. If you
|
||||
pass a worker token, the node joins as a worker.
|
||||
|
||||
### Join a node to swarm as a manager
|
||||
|
||||
The example below demonstrates joining a manager node using a manager token.
|
||||
|
||||
```bash
|
||||
$ docker swarm join --secret 4ao565v9jsuogtq5t8s379ulb --manager --listen-addr 192.168.99.122:2377 192.168.99.121:2377
|
||||
This node joined a Swarm as a manager.
|
||||
$ docker swarm join --token SWMTKN-1-3pu6hszjas19xyp7ghgosyx9k8atbfcr8p2is99znpy26u2lkl-7p73s1dx5in4tatdymyhg9hu2 --listen-addr 192.168.99.122:2377 192.168.99.121:2377
|
||||
This node joined a swarm as a manager.
|
||||
$ docker node ls
|
||||
ID HOSTNAME MEMBERSHIP STATUS AVAILABILITY MANAGER STATUS LEADER
|
||||
dkp8vy1dq1kxleu9g4u78tlag * manager2 Accepted Ready Active Reachable
|
||||
dvfxp4zseq4s0rih1selh0d20 manager1 Accepted Ready Active Reachable Yes
|
||||
ID HOSTNAME STATUS AVAILABILITY MANAGER STATUS
|
||||
dkp8vy1dq1kxleu9g4u78tlag * manager2 Ready Active Reachable
|
||||
dvfxp4zseq4s0rih1selh0d20 manager1 Ready Active Leader
|
||||
```
|
||||
|
||||
A cluster should only have 3-7 managers at most, because a majority of managers must be available
|
||||
for the cluster to function. Nodes that aren't meant to participate in this management quorum
|
||||
should join as workers instead. Managers should be stable hosts that have static IP addresses.
|
||||
|
||||
### Join a node to swarm as a worker
|
||||
|
||||
The example below demonstrates joining a worker node using a worker token.
|
||||
|
||||
```bash
|
||||
$ docker swarm join --secret 4ao565v9jsuogtq5t8s379ulb --listen-addr 192.168.99.123:2377 192.168.99.121:2377
|
||||
This node joined a Swarm as a worker.
|
||||
$ docker swarm join --token SWMTKN-1-3pu6hszjas19xyp7ghgosyx9k8atbfcr8p2is99znpy26u2lkl-1awxwuwd3z9j1z3puu7rcgdbx --listen-addr 192.168.99.123:2377 192.168.99.121:2377
|
||||
This node joined a swarm as a worker.
|
||||
$ docker node ls
|
||||
ID HOSTNAME MEMBERSHIP STATUS AVAILABILITY MANAGER STATUS LEADER
|
||||
7ln70fl22uw2dvjn2ft53m3q5 worker2 Accepted Ready Active
|
||||
dkp8vy1dq1kxleu9g4u78tlag worker1 Accepted Ready Active Reachable
|
||||
dvfxp4zseq4s0rih1selh0d20 * manager1 Accepted Ready Active Reachable Yes
|
||||
ID HOSTNAME STATUS AVAILABILITY MANAGER STATUS
|
||||
7ln70fl22uw2dvjn2ft53m3q5 worker2 Ready Active
|
||||
dkp8vy1dq1kxleu9g4u78tlag worker1 Ready Active Reachable
|
||||
dvfxp4zseq4s0rih1selh0d20 * manager1 Ready Active Leader
|
||||
```
|
||||
|
||||
### `--ca-hash`
|
||||
|
||||
Hash of the Root Certificate Authority certificate used for trusted join.
|
||||
|
||||
### `--listen-addr value`
|
||||
|
||||
The node listens for inbound Swarm manager traffic on this IP:PORT
|
||||
The node listens for inbound swarm manager traffic on this IP:PORT
|
||||
|
||||
### `--manager`
|
||||
|
||||
Joins the node as a manager
|
||||
|
||||
### `--secret string`
|
||||
### `--token string`
|
||||
|
||||
Secret value required for nodes to join the swarm
|
||||
|
||||
|
|
76
docs/reference/commandline/swarm_join_token.md
Normal file
76
docs/reference/commandline/swarm_join_token.md
Normal file
|
@ -0,0 +1,76 @@
|
|||
<!--[metadata]>
|
||||
+++
|
||||
title = "swarm join-token"
|
||||
description = "The swarm join-token command description and usage"
|
||||
keywords = ["swarm, join-token"]
|
||||
advisory = "rc"
|
||||
[menu.main]
|
||||
parent = "smn_cli"
|
||||
+++
|
||||
<![end-metadata]-->
|
||||
|
||||
# swarm join-token
|
||||
|
||||
```markdown
|
||||
Usage: docker swarm join-token [--rotate] (worker|manager)
|
||||
|
||||
Manage join tokens
|
||||
|
||||
Options:
|
||||
--help Print usage
|
||||
-q, --quiet Only display token
|
||||
--rotate Rotate join token
|
||||
```
|
||||
|
||||
Join tokens are secrets that determine whether or not a node will join the swarm as a manager node
|
||||
or a worker node. You pass the token using the `--token flag` when you run
|
||||
[swarm join](swarm_join.md). You can access the current tokens or rotate the tokens using
|
||||
`swarm join-token`.
|
||||
|
||||
Run with only a single `worker` or `manager` argument, it will print a command for joining a new
|
||||
node to the swarm, including the necessary token:
|
||||
|
||||
```bash
|
||||
$ docker swarm join-token worker
|
||||
To add a worker to this swarm, run the following command:
|
||||
docker swarm join \
|
||||
--token SWMTKN-1-3pu6hszjas19xyp7ghgosyx9k8atbfcr8p2is99znpy26u2lkl-1awxwuwd3z9j1z3puu7rcgdbx \
|
||||
172.17.0.2:2377
|
||||
|
||||
$ docker swarm join-token manager
|
||||
To add a manager to this swarm, run the following command:
|
||||
docker swarm join \
|
||||
--token SWMTKN-1-3pu6hszjas19xyp7ghgosyx9k8atbfcr8p2is99znpy26u2lkl-7p73s1dx5in4tatdymyhg9hu2 \
|
||||
172.17.0.2:2377
|
||||
```
|
||||
|
||||
Use the `--rotate` flag to generate a new join token for the specified role:
|
||||
|
||||
```bash
|
||||
$ docker swarm join-token --rotate worker
|
||||
To add a worker to this swarm, run the following command:
|
||||
docker swarm join \
|
||||
--token SWMTKN-1-3pu6hszjas19xyp7ghgosyx9k8atbfcr8p2is99znpy26u2lkl-b30ljddcqhef9b9v4rs7mel7t \
|
||||
172.17.0.2:2377
|
||||
```
|
||||
|
||||
After using `--rotate`, only the new token will be valid for joining with the specified role.
|
||||
|
||||
The `-q` (or `--quiet`) flag only prints the token:
|
||||
|
||||
```bash
|
||||
$ docker swarm join-token -q worker
|
||||
SWMTKN-1-3pu6hszjas19xyp7ghgosyx9k8atbfcr8p2is99znpy26u2lkl-b30ljddcqhef9b9v4rs7mel7t
|
||||
```
|
||||
|
||||
### `--rotate`
|
||||
|
||||
Update the join token for a specified role with a new token and print the token.
|
||||
|
||||
### `--quiet`
|
||||
|
||||
Only print the token. Do not print a complete command for joining.
|
||||
|
||||
## Related information
|
||||
|
||||
* [swarm join](swarm_join.md)
|
|
@ -14,7 +14,7 @@ parent = "smn_cli"
|
|||
```markdown
|
||||
Usage: docker swarm leave [OPTIONS]
|
||||
|
||||
Leave a Swarm
|
||||
Leave a swarm
|
||||
|
||||
Options:
|
||||
--force Force leave ignoring warnings.
|
||||
|
@ -26,10 +26,10 @@ This command causes the node to leave the swarm.
|
|||
On a manager node:
|
||||
```bash
|
||||
$ docker node ls
|
||||
ID HOSTNAME MEMBERSHIP STATUS AVAILABILITY MANAGER STATUS LEADER
|
||||
7ln70fl22uw2dvjn2ft53m3q5 worker2 Accepted Ready Active
|
||||
dkp8vy1dq1kxleu9g4u78tlag worker1 Accepted Ready Active Reachable
|
||||
dvfxp4zseq4s0rih1selh0d20 * manager1 Accepted Ready Active Reachable Yes
|
||||
ID HOSTNAME STATUS AVAILABILITY MANAGER STATUS
|
||||
7ln70fl22uw2dvjn2ft53m3q5 worker2 Ready Active
|
||||
dkp8vy1dq1kxleu9g4u78tlag worker1 Ready Active Reachable
|
||||
dvfxp4zseq4s0rih1selh0d20 * manager1 Ready Active Leader
|
||||
```
|
||||
|
||||
On a worker node:
|
||||
|
@ -41,10 +41,10 @@ Node left the default swarm.
|
|||
On a manager node:
|
||||
```bash
|
||||
$ docker node ls
|
||||
ID HOSTNAME MEMBERSHIP STATUS AVAILABILITY MANAGER STATUS LEADER
|
||||
7ln70fl22uw2dvjn2ft53m3q5 worker2 Accepted Down Active
|
||||
dkp8vy1dq1kxleu9g4u78tlag worker1 Accepted Ready Active Reachable
|
||||
dvfxp4zseq4s0rih1selh0d20 * manager1 Accepted Ready Active Reachable Yes
|
||||
ID HOSTNAME STATUS AVAILABILITY MANAGER STATUS
|
||||
7ln70fl22uw2dvjn2ft53m3q5 worker2 Down Active
|
||||
dkp8vy1dq1kxleu9g4u78tlag worker1 Ready Active Reachable
|
||||
dvfxp4zseq4s0rih1selh0d20 * manager1 Ready Active Leader
|
||||
```
|
||||
|
||||
## Related information
|
||||
|
|
|
@ -14,23 +14,21 @@ parent = "smn_cli"
|
|||
```markdown
|
||||
Usage: docker swarm update [OPTIONS]
|
||||
|
||||
Update the Swarm
|
||||
Update the swarm
|
||||
|
||||
Options:
|
||||
--auto-accept value Auto acceptance policy (worker, manager or none)
|
||||
--cert-expiry duration Validity period for node certificates (default 2160h0m0s)
|
||||
--dispatcher-heartbeat duration Dispatcher heartbeat period (default 5s)
|
||||
--external-ca value Specifications of one or more certificate signing endpoints
|
||||
--help Print usage
|
||||
--secret string Set secret value needed to accept nodes into cluster
|
||||
--task-history-limit int Task history retention limit (default 10)
|
||||
```
|
||||
|
||||
Updates a Swarm cluster with new parameter values. This command must target a manager node.
|
||||
Updates a swarm cluster with new parameter values. This command must target a manager node.
|
||||
|
||||
|
||||
```bash
|
||||
$ docker swarm update --auto-accept manager
|
||||
$ docker swarm update --cert-expirty 4000h0m0s
|
||||
```
|
||||
|
||||
## Related information
|
||||
|
|
|
@ -216,15 +216,17 @@ func (s *DockerSwarmSuite) AddDaemon(c *check.C, joinSwarm, manager bool) *Swarm
|
|||
|
||||
if joinSwarm == true {
|
||||
if len(s.daemons) > 0 {
|
||||
tokens := s.daemons[0].joinTokens(c)
|
||||
token := tokens.Worker
|
||||
if manager {
|
||||
token = tokens.Manager
|
||||
}
|
||||
c.Assert(d.Join(swarm.JoinRequest{
|
||||
RemoteAddrs: []string{s.daemons[0].listenAddr},
|
||||
Manager: manager}), check.IsNil)
|
||||
} else {
|
||||
c.Assert(d.Init(swarm.InitRequest{
|
||||
Spec: swarm.Spec{
|
||||
AcceptancePolicy: autoAcceptPolicy,
|
||||
},
|
||||
JoinToken: token,
|
||||
}), check.IsNil)
|
||||
} else {
|
||||
c.Assert(d.Init(swarm.InitRequest{}), check.IsNil)
|
||||
}
|
||||
}
|
||||
|
||||
|
|
|
@ -22,14 +22,6 @@ type SwarmDaemon struct {
|
|||
listenAddr string
|
||||
}
|
||||
|
||||
// default policy in tests is allow-all
|
||||
var autoAcceptPolicy = swarm.AcceptancePolicy{
|
||||
Policies: []swarm.Policy{
|
||||
{Role: swarm.NodeRoleWorker, Autoaccept: true},
|
||||
{Role: swarm.NodeRoleManager, Autoaccept: true},
|
||||
},
|
||||
}
|
||||
|
||||
// Init initializes a new swarm cluster.
|
||||
func (d *SwarmDaemon) Init(req swarm.InitRequest) error {
|
||||
if req.ListenAddr == "" {
|
||||
|
@ -271,6 +263,28 @@ func (d *SwarmDaemon) updateSwarm(c *check.C, f ...specConstructor) {
|
|||
c.Assert(status, checker.Equals, http.StatusOK, check.Commentf("output: %q", string(out)))
|
||||
}
|
||||
|
||||
func (d *SwarmDaemon) rotateTokens(c *check.C) {
|
||||
var sw swarm.Swarm
|
||||
status, out, err := d.SockRequest("GET", "/swarm", nil)
|
||||
c.Assert(err, checker.IsNil)
|
||||
c.Assert(status, checker.Equals, http.StatusOK, check.Commentf("output: %q", string(out)))
|
||||
c.Assert(json.Unmarshal(out, &sw), checker.IsNil)
|
||||
|
||||
url := fmt.Sprintf("/swarm/update?version=%d&rotate_worker_token=true&rotate_manager_token=true", sw.Version.Index)
|
||||
status, out, err = d.SockRequest("POST", url, sw.Spec)
|
||||
c.Assert(err, checker.IsNil)
|
||||
c.Assert(status, checker.Equals, http.StatusOK, check.Commentf("output: %q", string(out)))
|
||||
}
|
||||
|
||||
func (d *SwarmDaemon) joinTokens(c *check.C) swarm.JoinTokens {
|
||||
var sw swarm.Swarm
|
||||
status, out, err := d.SockRequest("GET", "/swarm", nil)
|
||||
c.Assert(err, checker.IsNil)
|
||||
c.Assert(status, checker.Equals, http.StatusOK, check.Commentf("output: %q", string(out)))
|
||||
c.Assert(json.Unmarshal(out, &sw), checker.IsNil)
|
||||
return sw.JoinTokens
|
||||
}
|
||||
|
||||
func (d *SwarmDaemon) checkLocalNodeState(c *check.C) (interface{}, check.CommentInterface) {
|
||||
info, err := d.info()
|
||||
c.Assert(err, checker.IsNil)
|
||||
|
|
|
@ -43,7 +43,7 @@ func (s *DockerSwarmSuite) TestApiSwarmInit(c *check.C) {
|
|||
c.Assert(info.ControlAvailable, checker.False)
|
||||
c.Assert(info.LocalNodeState, checker.Equals, swarm.LocalNodeStateInactive)
|
||||
|
||||
c.Assert(d2.Join(swarm.JoinRequest{RemoteAddrs: []string{d1.listenAddr}}), checker.IsNil)
|
||||
c.Assert(d2.Join(swarm.JoinRequest{JoinToken: d1.joinTokens(c).Worker, RemoteAddrs: []string{d1.listenAddr}}), checker.IsNil)
|
||||
|
||||
info, err = d2.info()
|
||||
c.Assert(err, checker.IsNil)
|
||||
|
@ -72,89 +72,29 @@ func (s *DockerSwarmSuite) TestApiSwarmInit(c *check.C) {
|
|||
c.Assert(info.LocalNodeState, checker.Equals, swarm.LocalNodeStateActive)
|
||||
}
|
||||
|
||||
func (s *DockerSwarmSuite) TestApiSwarmManualAcceptance(c *check.C) {
|
||||
func (s *DockerSwarmSuite) TestApiSwarmJoinToken(c *check.C) {
|
||||
testRequires(c, Network)
|
||||
s.testAPISwarmManualAcceptance(c, "")
|
||||
}
|
||||
func (s *DockerSwarmSuite) TestApiSwarmManualAcceptanceSecret(c *check.C) {
|
||||
testRequires(c, Network)
|
||||
s.testAPISwarmManualAcceptance(c, "foobaz")
|
||||
}
|
||||
|
||||
func (s *DockerSwarmSuite) testAPISwarmManualAcceptance(c *check.C, secret string) {
|
||||
d1 := s.AddDaemon(c, false, false)
|
||||
c.Assert(d1.Init(swarm.InitRequest{
|
||||
Spec: swarm.Spec{
|
||||
AcceptancePolicy: swarm.AcceptancePolicy{
|
||||
Policies: []swarm.Policy{
|
||||
{Role: swarm.NodeRoleWorker, Secret: &secret},
|
||||
{Role: swarm.NodeRoleManager, Secret: &secret},
|
||||
},
|
||||
},
|
||||
},
|
||||
}), checker.IsNil)
|
||||
c.Assert(d1.Init(swarm.InitRequest{}), checker.IsNil)
|
||||
|
||||
d2 := s.AddDaemon(c, false, false)
|
||||
err := d2.Join(swarm.JoinRequest{RemoteAddrs: []string{d1.listenAddr}})
|
||||
c.Assert(err, checker.NotNil)
|
||||
if secret == "" {
|
||||
c.Assert(err.Error(), checker.Contains, "needs to be accepted")
|
||||
info, err := d2.info()
|
||||
c.Assert(err, checker.IsNil)
|
||||
c.Assert(info.LocalNodeState, checker.Equals, swarm.LocalNodeStatePending)
|
||||
c.Assert(d2.Leave(false), checker.IsNil)
|
||||
info, err = d2.info()
|
||||
c.Assert(err, checker.IsNil)
|
||||
c.Assert(info.LocalNodeState, checker.Equals, swarm.LocalNodeStateInactive)
|
||||
} else {
|
||||
c.Assert(err.Error(), checker.Contains, "valid secret token is necessary")
|
||||
info, err := d2.info()
|
||||
c.Assert(err, checker.IsNil)
|
||||
c.Assert(info.LocalNodeState, checker.Equals, swarm.LocalNodeStateInactive)
|
||||
}
|
||||
d3 := s.AddDaemon(c, false, false)
|
||||
c.Assert(d3.Join(swarm.JoinRequest{Secret: secret, RemoteAddrs: []string{d1.listenAddr}}), checker.NotNil)
|
||||
info, err := d3.info()
|
||||
c.Assert(err, checker.IsNil)
|
||||
c.Assert(info.LocalNodeState, checker.Equals, swarm.LocalNodeStatePending)
|
||||
c.Assert(len(info.NodeID), checker.GreaterThan, 5)
|
||||
d1.updateNode(c, info.NodeID, func(n *swarm.Node) {
|
||||
n.Spec.Membership = swarm.NodeMembershipAccepted
|
||||
})
|
||||
waitAndAssert(c, defaultReconciliationTimeout, d3.checkLocalNodeState, checker.Equals, swarm.LocalNodeStateActive)
|
||||
}
|
||||
|
||||
func (s *DockerSwarmSuite) TestApiSwarmSecretAcceptance(c *check.C) {
|
||||
testRequires(c, Network)
|
||||
d1 := s.AddDaemon(c, false, false)
|
||||
secret := "foobar"
|
||||
c.Assert(d1.Init(swarm.InitRequest{
|
||||
Spec: swarm.Spec{
|
||||
AcceptancePolicy: swarm.AcceptancePolicy{
|
||||
Policies: []swarm.Policy{
|
||||
{Role: swarm.NodeRoleWorker, Autoaccept: true, Secret: &secret},
|
||||
{Role: swarm.NodeRoleManager, Secret: &secret},
|
||||
},
|
||||
},
|
||||
},
|
||||
}), checker.IsNil)
|
||||
|
||||
d2 := s.AddDaemon(c, false, false)
|
||||
err := d2.Join(swarm.JoinRequest{RemoteAddrs: []string{d1.listenAddr}})
|
||||
c.Assert(err, checker.NotNil)
|
||||
c.Assert(err.Error(), checker.Contains, "secret token is necessary")
|
||||
c.Assert(err.Error(), checker.Contains, "join token is necessary")
|
||||
info, err := d2.info()
|
||||
c.Assert(err, checker.IsNil)
|
||||
c.Assert(info.LocalNodeState, checker.Equals, swarm.LocalNodeStateInactive)
|
||||
|
||||
err = d2.Join(swarm.JoinRequest{Secret: "foobaz", RemoteAddrs: []string{d1.listenAddr}})
|
||||
err = d2.Join(swarm.JoinRequest{JoinToken: "foobaz", RemoteAddrs: []string{d1.listenAddr}})
|
||||
c.Assert(err, checker.NotNil)
|
||||
c.Assert(err.Error(), checker.Contains, "secret token is necessary")
|
||||
c.Assert(err.Error(), checker.Contains, "join token is necessary")
|
||||
info, err = d2.info()
|
||||
c.Assert(err, checker.IsNil)
|
||||
c.Assert(info.LocalNodeState, checker.Equals, swarm.LocalNodeStateInactive)
|
||||
|
||||
c.Assert(d2.Join(swarm.JoinRequest{Secret: "foobar", RemoteAddrs: []string{d1.listenAddr}}), checker.IsNil)
|
||||
workerToken := d1.joinTokens(c).Worker
|
||||
|
||||
c.Assert(d2.Join(swarm.JoinRequest{JoinToken: workerToken, RemoteAddrs: []string{d1.listenAddr}}), checker.IsNil)
|
||||
info, err = d2.info()
|
||||
c.Assert(err, checker.IsNil)
|
||||
c.Assert(info.LocalNodeState, checker.Equals, swarm.LocalNodeStateActive)
|
||||
|
@ -163,22 +103,19 @@ func (s *DockerSwarmSuite) TestApiSwarmSecretAcceptance(c *check.C) {
|
|||
c.Assert(err, checker.IsNil)
|
||||
c.Assert(info.LocalNodeState, checker.Equals, swarm.LocalNodeStateInactive)
|
||||
|
||||
// change secret
|
||||
d1.updateSwarm(c, func(s *swarm.Spec) {
|
||||
for i := range s.AcceptancePolicy.Policies {
|
||||
p := "foobaz"
|
||||
s.AcceptancePolicy.Policies[i].Secret = &p
|
||||
}
|
||||
})
|
||||
// change tokens
|
||||
d1.rotateTokens(c)
|
||||
|
||||
err = d2.Join(swarm.JoinRequest{Secret: "foobar", RemoteAddrs: []string{d1.listenAddr}})
|
||||
err = d2.Join(swarm.JoinRequest{JoinToken: workerToken, RemoteAddrs: []string{d1.listenAddr}})
|
||||
c.Assert(err, checker.NotNil)
|
||||
c.Assert(err.Error(), checker.Contains, "secret token is necessary")
|
||||
c.Assert(err.Error(), checker.Contains, "join token is necessary")
|
||||
info, err = d2.info()
|
||||
c.Assert(err, checker.IsNil)
|
||||
c.Assert(info.LocalNodeState, checker.Equals, swarm.LocalNodeStateInactive)
|
||||
|
||||
c.Assert(d2.Join(swarm.JoinRequest{Secret: "foobaz", RemoteAddrs: []string{d1.listenAddr}}), checker.IsNil)
|
||||
workerToken = d1.joinTokens(c).Worker
|
||||
|
||||
c.Assert(d2.Join(swarm.JoinRequest{JoinToken: workerToken, RemoteAddrs: []string{d1.listenAddr}}), checker.IsNil)
|
||||
info, err = d2.info()
|
||||
c.Assert(err, checker.IsNil)
|
||||
c.Assert(info.LocalNodeState, checker.Equals, swarm.LocalNodeStateActive)
|
||||
|
@ -187,24 +124,17 @@ func (s *DockerSwarmSuite) TestApiSwarmSecretAcceptance(c *check.C) {
|
|||
c.Assert(err, checker.IsNil)
|
||||
c.Assert(info.LocalNodeState, checker.Equals, swarm.LocalNodeStateInactive)
|
||||
|
||||
// change policy, don't change secret
|
||||
d1.updateSwarm(c, func(s *swarm.Spec) {
|
||||
for i, p := range s.AcceptancePolicy.Policies {
|
||||
if p.Role == swarm.NodeRoleManager {
|
||||
s.AcceptancePolicy.Policies[i].Autoaccept = false
|
||||
}
|
||||
s.AcceptancePolicy.Policies[i].Secret = nil
|
||||
}
|
||||
})
|
||||
// change spec, don't change tokens
|
||||
d1.updateSwarm(c, func(s *swarm.Spec) {})
|
||||
|
||||
err = d2.Join(swarm.JoinRequest{RemoteAddrs: []string{d1.listenAddr}})
|
||||
c.Assert(err, checker.NotNil)
|
||||
c.Assert(err.Error(), checker.Contains, "secret token is necessary")
|
||||
c.Assert(err.Error(), checker.Contains, "join token is necessary")
|
||||
info, err = d2.info()
|
||||
c.Assert(err, checker.IsNil)
|
||||
c.Assert(info.LocalNodeState, checker.Equals, swarm.LocalNodeStateInactive)
|
||||
|
||||
c.Assert(d2.Join(swarm.JoinRequest{Secret: "foobaz", RemoteAddrs: []string{d1.listenAddr}}), checker.IsNil)
|
||||
c.Assert(d2.Join(swarm.JoinRequest{JoinToken: workerToken, RemoteAddrs: []string{d1.listenAddr}}), checker.IsNil)
|
||||
info, err = d2.info()
|
||||
c.Assert(err, checker.IsNil)
|
||||
c.Assert(info.LocalNodeState, checker.Equals, swarm.LocalNodeStateActive)
|
||||
|
@ -212,51 +142,24 @@ func (s *DockerSwarmSuite) TestApiSwarmSecretAcceptance(c *check.C) {
|
|||
info, err = d2.info()
|
||||
c.Assert(err, checker.IsNil)
|
||||
c.Assert(info.LocalNodeState, checker.Equals, swarm.LocalNodeStateInactive)
|
||||
|
||||
// clear secret
|
||||
d1.updateSwarm(c, func(s *swarm.Spec) {
|
||||
for i := range s.AcceptancePolicy.Policies {
|
||||
p := ""
|
||||
s.AcceptancePolicy.Policies[i].Secret = &p
|
||||
}
|
||||
})
|
||||
|
||||
c.Assert(d2.Join(swarm.JoinRequest{RemoteAddrs: []string{d1.listenAddr}}), checker.IsNil)
|
||||
info, err = d2.info()
|
||||
c.Assert(err, checker.IsNil)
|
||||
c.Assert(info.LocalNodeState, checker.Equals, swarm.LocalNodeStateActive)
|
||||
c.Assert(d2.Leave(false), checker.IsNil)
|
||||
info, err = d2.info()
|
||||
c.Assert(err, checker.IsNil)
|
||||
c.Assert(info.LocalNodeState, checker.Equals, swarm.LocalNodeStateInactive)
|
||||
|
||||
}
|
||||
|
||||
func (s *DockerSwarmSuite) TestApiSwarmCAHash(c *check.C) {
|
||||
testRequires(c, Network)
|
||||
d1 := s.AddDaemon(c, true, true)
|
||||
d2 := s.AddDaemon(c, false, false)
|
||||
err := d2.Join(swarm.JoinRequest{CACertHash: "foobar", RemoteAddrs: []string{d1.listenAddr}})
|
||||
splitToken := strings.Split(d1.joinTokens(c).Worker, "-")
|
||||
splitToken[2] = "1kxftv4ofnc6mt30lmgipg6ngf9luhwqopfk1tz6bdmnkubg0e"
|
||||
replacementToken := strings.Join(splitToken, "-")
|
||||
err := d2.Join(swarm.JoinRequest{JoinToken: replacementToken, RemoteAddrs: []string{d1.listenAddr}})
|
||||
c.Assert(err, checker.NotNil)
|
||||
c.Assert(err.Error(), checker.Contains, "invalid checksum digest format")
|
||||
|
||||
c.Assert(len(d1.CACertHash), checker.GreaterThan, 0)
|
||||
c.Assert(d2.Join(swarm.JoinRequest{CACertHash: d1.CACertHash, RemoteAddrs: []string{d1.listenAddr}}), checker.IsNil)
|
||||
c.Assert(err.Error(), checker.Contains, "remote CA does not match fingerprint")
|
||||
}
|
||||
|
||||
func (s *DockerSwarmSuite) TestApiSwarmPromoteDemote(c *check.C) {
|
||||
testRequires(c, Network)
|
||||
d1 := s.AddDaemon(c, false, false)
|
||||
c.Assert(d1.Init(swarm.InitRequest{
|
||||
Spec: swarm.Spec{
|
||||
AcceptancePolicy: swarm.AcceptancePolicy{
|
||||
Policies: []swarm.Policy{
|
||||
{Role: swarm.NodeRoleWorker, Autoaccept: true},
|
||||
{Role: swarm.NodeRoleManager},
|
||||
},
|
||||
},
|
||||
},
|
||||
}), checker.IsNil)
|
||||
c.Assert(d1.Init(swarm.InitRequest{}), checker.IsNil)
|
||||
d2 := s.AddDaemon(c, true, false)
|
||||
|
||||
info, err := d2.info()
|
||||
|
@ -838,9 +741,7 @@ func (s *DockerSwarmSuite) TestApiSwarmForceNewCluster(c *check.C) {
|
|||
|
||||
c.Assert(d1.Init(swarm.InitRequest{
|
||||
ForceNewCluster: true,
|
||||
Spec: swarm.Spec{
|
||||
AcceptancePolicy: autoAcceptPolicy,
|
||||
},
|
||||
Spec: swarm.Spec{},
|
||||
}), checker.IsNil)
|
||||
|
||||
waitAndAssert(c, defaultReconciliationTimeout, d1.checkActiveContainerCount, checker.Equals, instances)
|
||||
|
@ -937,7 +838,6 @@ func checkClusterHealth(c *check.C, cl []*SwarmDaemon, managerCount, workerCount
|
|||
for _, n := range d.listNodes(c) {
|
||||
c.Assert(n.Status.State, checker.Equals, swarm.NodeStateReady, check.Commentf("state of node %s, reported by %s", n.ID, d.Info.NodeID))
|
||||
c.Assert(n.Spec.Availability, checker.Equals, swarm.NodeAvailabilityActive, check.Commentf("availability of node %s, reported by %s", n.ID, d.Info.NodeID))
|
||||
c.Assert(n.Spec.Membership, checker.Equals, swarm.NodeMembershipAccepted, check.Commentf("membership of node %s, reported by %s", n.ID, d.Info.NodeID))
|
||||
if n.Spec.Role == swarm.NodeRoleManager {
|
||||
c.Assert(n.ManagerStatus, checker.NotNil, check.Commentf("manager status of node %s (manager), reported by %s", n.ID, d.Info.NodeID))
|
||||
if n.ManagerStatus.Leader {
|
||||
|
|
|
@ -25,50 +25,13 @@ func (s *DockerSwarmSuite) TestSwarmUpdate(c *check.C) {
|
|||
return sw[0].Spec
|
||||
}
|
||||
|
||||
out, err := d.Cmd("swarm", "update", "--cert-expiry", "30h", "--dispatcher-heartbeat", "11s", "--auto-accept", "manager", "--auto-accept", "worker", "--secret", "foo")
|
||||
out, err := d.Cmd("swarm", "update", "--cert-expiry", "30h", "--dispatcher-heartbeat", "11s")
|
||||
c.Assert(err, checker.IsNil, check.Commentf("out: %v", out))
|
||||
|
||||
spec := getSpec()
|
||||
c.Assert(spec.CAConfig.NodeCertExpiry, checker.Equals, 30*time.Hour)
|
||||
c.Assert(spec.Dispatcher.HeartbeatPeriod, checker.Equals, uint64(11*time.Second))
|
||||
|
||||
c.Assert(spec.AcceptancePolicy.Policies, checker.HasLen, 2)
|
||||
|
||||
for _, p := range spec.AcceptancePolicy.Policies {
|
||||
c.Assert(p.Autoaccept, checker.Equals, true)
|
||||
c.Assert(p.Secret, checker.NotNil)
|
||||
c.Assert(*p.Secret, checker.Not(checker.Equals), "")
|
||||
}
|
||||
|
||||
out, err = d.Cmd("swarm", "update", "--auto-accept", "none")
|
||||
c.Assert(err, checker.IsNil, check.Commentf("out: %v", out))
|
||||
|
||||
spec = getSpec()
|
||||
c.Assert(spec.CAConfig.NodeCertExpiry, checker.Equals, 30*time.Hour)
|
||||
c.Assert(spec.Dispatcher.HeartbeatPeriod, checker.Equals, uint64(11*time.Second))
|
||||
|
||||
c.Assert(spec.AcceptancePolicy.Policies, checker.HasLen, 2)
|
||||
|
||||
for _, p := range spec.AcceptancePolicy.Policies {
|
||||
c.Assert(p.Autoaccept, checker.Equals, false)
|
||||
// secret is still set
|
||||
c.Assert(p.Secret, checker.NotNil)
|
||||
c.Assert(*p.Secret, checker.Not(checker.Equals), "")
|
||||
}
|
||||
|
||||
out, err = d.Cmd("swarm", "update", "--auto-accept", "manager", "--secret", "")
|
||||
c.Assert(err, checker.IsNil, check.Commentf("out: %v", out))
|
||||
|
||||
spec = getSpec()
|
||||
|
||||
c.Assert(spec.AcceptancePolicy.Policies, checker.HasLen, 2)
|
||||
|
||||
for _, p := range spec.AcceptancePolicy.Policies {
|
||||
c.Assert(p.Autoaccept, checker.Equals, p.Role == swarm.NodeRoleManager)
|
||||
// secret has been removed
|
||||
c.Assert(p.Secret, checker.IsNil)
|
||||
}
|
||||
|
||||
// setting anything under 30m for cert-expiry is not allowed
|
||||
out, err = d.Cmd("swarm", "update", "--cert-expiry", "15m")
|
||||
c.Assert(err, checker.NotNil)
|
||||
|
@ -89,37 +52,21 @@ func (s *DockerSwarmSuite) TestSwarmInit(c *check.C) {
|
|||
return sw[0].Spec
|
||||
}
|
||||
|
||||
out, err := d.Cmd("swarm", "init", "--cert-expiry", "30h", "--dispatcher-heartbeat", "11s", "--auto-accept", "manager", "--auto-accept", "worker", "--secret", "foo")
|
||||
out, err := d.Cmd("swarm", "init", "--cert-expiry", "30h", "--dispatcher-heartbeat", "11s")
|
||||
c.Assert(err, checker.IsNil, check.Commentf("out: %v", out))
|
||||
|
||||
spec := getSpec()
|
||||
c.Assert(spec.CAConfig.NodeCertExpiry, checker.Equals, 30*time.Hour)
|
||||
c.Assert(spec.Dispatcher.HeartbeatPeriod, checker.Equals, uint64(11*time.Second))
|
||||
|
||||
c.Assert(spec.AcceptancePolicy.Policies, checker.HasLen, 2)
|
||||
|
||||
for _, p := range spec.AcceptancePolicy.Policies {
|
||||
c.Assert(p.Autoaccept, checker.Equals, true)
|
||||
c.Assert(p.Secret, checker.NotNil)
|
||||
c.Assert(*p.Secret, checker.Not(checker.Equals), "")
|
||||
}
|
||||
|
||||
c.Assert(d.Leave(true), checker.IsNil)
|
||||
|
||||
out, err = d.Cmd("swarm", "init", "--auto-accept", "none", "--secret", "")
|
||||
out, err = d.Cmd("swarm", "init")
|
||||
c.Assert(err, checker.IsNil, check.Commentf("out: %v", out))
|
||||
|
||||
spec = getSpec()
|
||||
c.Assert(spec.CAConfig.NodeCertExpiry, checker.Equals, 90*24*time.Hour)
|
||||
c.Assert(spec.Dispatcher.HeartbeatPeriod, checker.Equals, uint64(5*time.Second))
|
||||
|
||||
c.Assert(spec.AcceptancePolicy.Policies, checker.HasLen, 2)
|
||||
|
||||
for _, p := range spec.AcceptancePolicy.Policies {
|
||||
c.Assert(p.Autoaccept, checker.Equals, false)
|
||||
c.Assert(p.Secret, checker.IsNil)
|
||||
}
|
||||
|
||||
}
|
||||
|
||||
func (s *DockerSwarmSuite) TestSwarmInitIPv6(c *check.C) {
|
||||
|
|
Loading…
Reference in a new issue