mirror of https://github.com/docker/cli.git
236 lines
6.5 KiB
Go
236 lines
6.5 KiB
Go
package service
|
|
|
|
import (
|
|
"context"
|
|
"fmt"
|
|
"testing"
|
|
"time"
|
|
|
|
"github.com/docker/cli/opts"
|
|
"github.com/docker/docker/api/types"
|
|
"github.com/docker/docker/api/types/container"
|
|
"github.com/docker/docker/api/types/swarm"
|
|
"gotest.tools/assert"
|
|
is "gotest.tools/assert/cmp"
|
|
)
|
|
|
|
func TestMemBytesString(t *testing.T) {
|
|
var mem opts.MemBytes = 1048576
|
|
assert.Check(t, is.Equal("1MiB", mem.String()))
|
|
}
|
|
|
|
func TestMemBytesSetAndValue(t *testing.T) {
|
|
var mem opts.MemBytes
|
|
assert.NilError(t, mem.Set("5kb"))
|
|
assert.Check(t, is.Equal(int64(5120), mem.Value()))
|
|
}
|
|
|
|
func TestNanoCPUsString(t *testing.T) {
|
|
var cpus opts.NanoCPUs = 6100000000
|
|
assert.Check(t, is.Equal("6.100", cpus.String()))
|
|
}
|
|
|
|
func TestNanoCPUsSetAndValue(t *testing.T) {
|
|
var cpus opts.NanoCPUs
|
|
assert.NilError(t, cpus.Set("0.35"))
|
|
assert.Check(t, is.Equal(int64(350000000), cpus.Value()))
|
|
}
|
|
|
|
func TestUint64OptString(t *testing.T) {
|
|
value := uint64(2345678)
|
|
opt := Uint64Opt{value: &value}
|
|
assert.Check(t, is.Equal("2345678", opt.String()))
|
|
|
|
opt = Uint64Opt{}
|
|
assert.Check(t, is.Equal("", opt.String()))
|
|
}
|
|
|
|
func TestUint64OptSetAndValue(t *testing.T) {
|
|
var opt Uint64Opt
|
|
assert.NilError(t, opt.Set("14445"))
|
|
assert.Check(t, is.Equal(uint64(14445), *opt.Value()))
|
|
}
|
|
|
|
func TestHealthCheckOptionsToHealthConfig(t *testing.T) {
|
|
dur := time.Second
|
|
opt := healthCheckOptions{
|
|
cmd: "curl",
|
|
interval: opts.PositiveDurationOpt{DurationOpt: *opts.NewDurationOpt(&dur)},
|
|
timeout: opts.PositiveDurationOpt{DurationOpt: *opts.NewDurationOpt(&dur)},
|
|
startPeriod: opts.PositiveDurationOpt{DurationOpt: *opts.NewDurationOpt(&dur)},
|
|
retries: 10,
|
|
}
|
|
config, err := opt.toHealthConfig()
|
|
assert.NilError(t, err)
|
|
assert.Check(t, is.DeepEqual(&container.HealthConfig{
|
|
Test: []string{"CMD-SHELL", "curl"},
|
|
Interval: time.Second,
|
|
Timeout: time.Second,
|
|
StartPeriod: time.Second,
|
|
Retries: 10,
|
|
}, config))
|
|
}
|
|
|
|
func TestHealthCheckOptionsToHealthConfigNoHealthcheck(t *testing.T) {
|
|
opt := healthCheckOptions{
|
|
noHealthcheck: true,
|
|
}
|
|
config, err := opt.toHealthConfig()
|
|
assert.NilError(t, err)
|
|
assert.Check(t, is.DeepEqual(&container.HealthConfig{
|
|
Test: []string{"NONE"},
|
|
}, config))
|
|
}
|
|
|
|
func TestHealthCheckOptionsToHealthConfigConflict(t *testing.T) {
|
|
opt := healthCheckOptions{
|
|
cmd: "curl",
|
|
noHealthcheck: true,
|
|
}
|
|
_, err := opt.toHealthConfig()
|
|
assert.Error(t, err, "--no-healthcheck conflicts with --health-* options")
|
|
}
|
|
|
|
func TestResourceOptionsToResourceRequirements(t *testing.T) {
|
|
incorrectOptions := []resourceOptions{
|
|
{
|
|
resGenericResources: []string{"foo=bar", "foo=1"},
|
|
},
|
|
{
|
|
resGenericResources: []string{"foo=bar", "foo=baz"},
|
|
},
|
|
{
|
|
resGenericResources: []string{"foo=bar"},
|
|
},
|
|
{
|
|
resGenericResources: []string{"foo=1", "foo=2"},
|
|
},
|
|
}
|
|
|
|
for _, opt := range incorrectOptions {
|
|
_, err := opt.ToResourceRequirements()
|
|
assert.Check(t, is.ErrorContains(err, ""))
|
|
}
|
|
|
|
correctOptions := []resourceOptions{
|
|
{
|
|
resGenericResources: []string{"foo=1"},
|
|
},
|
|
{
|
|
resGenericResources: []string{"foo=1", "bar=2"},
|
|
},
|
|
}
|
|
|
|
for _, opt := range correctOptions {
|
|
r, err := opt.ToResourceRequirements()
|
|
assert.NilError(t, err)
|
|
assert.Check(t, is.Len(r.Reservations.GenericResources, len(opt.resGenericResources)))
|
|
}
|
|
|
|
}
|
|
|
|
func TestToServiceNetwork(t *testing.T) {
|
|
nws := []types.NetworkResource{
|
|
{Name: "aaa-network", ID: "id555"},
|
|
{Name: "mmm-network", ID: "id999"},
|
|
{Name: "zzz-network", ID: "id111"},
|
|
}
|
|
|
|
client := &fakeClient{
|
|
networkInspectFunc: func(ctx context.Context, networkID string, options types.NetworkInspectOptions) (types.NetworkResource, error) {
|
|
for _, network := range nws {
|
|
if network.ID == networkID || network.Name == networkID {
|
|
return network, nil
|
|
}
|
|
}
|
|
return types.NetworkResource{}, fmt.Errorf("network not found: %s", networkID)
|
|
},
|
|
}
|
|
|
|
nwo := opts.NetworkOpt{}
|
|
nwo.Set("zzz-network")
|
|
nwo.Set("mmm-network")
|
|
nwo.Set("aaa-network")
|
|
|
|
o := newServiceOptions()
|
|
o.mode = "replicated"
|
|
o.networks = nwo
|
|
|
|
ctx := context.Background()
|
|
flags := newCreateCommand(nil).Flags()
|
|
service, err := o.ToService(ctx, client, flags)
|
|
assert.NilError(t, err)
|
|
assert.Check(t, is.DeepEqual([]swarm.NetworkAttachmentConfig{{Target: "id111"}, {Target: "id555"}, {Target: "id999"}}, service.TaskTemplate.Networks))
|
|
}
|
|
|
|
func TestToServiceUpdateRollback(t *testing.T) {
|
|
expected := swarm.ServiceSpec{
|
|
UpdateConfig: &swarm.UpdateConfig{
|
|
Parallelism: 23,
|
|
Delay: 34 * time.Second,
|
|
Monitor: 54321 * time.Nanosecond,
|
|
FailureAction: "pause",
|
|
MaxFailureRatio: 0.6,
|
|
Order: "stop-first",
|
|
},
|
|
RollbackConfig: &swarm.UpdateConfig{
|
|
Parallelism: 12,
|
|
Delay: 23 * time.Second,
|
|
Monitor: 12345 * time.Nanosecond,
|
|
FailureAction: "continue",
|
|
MaxFailureRatio: 0.5,
|
|
Order: "start-first",
|
|
},
|
|
}
|
|
|
|
// Note: in test-situation, the flags are only used to detect if an option
|
|
// was set; the actual value itself is read from the serviceOptions below.
|
|
flags := newCreateCommand(nil).Flags()
|
|
flags.Set("update-parallelism", "23")
|
|
flags.Set("update-delay", "34s")
|
|
flags.Set("update-monitor", "54321ns")
|
|
flags.Set("update-failure-action", "pause")
|
|
flags.Set("update-max-failure-ratio", "0.6")
|
|
flags.Set("update-order", "stop-first")
|
|
|
|
flags.Set("rollback-parallelism", "12")
|
|
flags.Set("rollback-delay", "23s")
|
|
flags.Set("rollback-monitor", "12345ns")
|
|
flags.Set("rollback-failure-action", "continue")
|
|
flags.Set("rollback-max-failure-ratio", "0.5")
|
|
flags.Set("rollback-order", "start-first")
|
|
|
|
o := newServiceOptions()
|
|
o.mode = "replicated"
|
|
o.update = updateOptions{
|
|
parallelism: 23,
|
|
delay: 34 * time.Second,
|
|
monitor: 54321 * time.Nanosecond,
|
|
onFailure: "pause",
|
|
maxFailureRatio: 0.6,
|
|
order: "stop-first",
|
|
}
|
|
o.rollback = updateOptions{
|
|
parallelism: 12,
|
|
delay: 23 * time.Second,
|
|
monitor: 12345 * time.Nanosecond,
|
|
onFailure: "continue",
|
|
maxFailureRatio: 0.5,
|
|
order: "start-first",
|
|
}
|
|
|
|
service, err := o.ToService(context.Background(), &fakeClient{}, flags)
|
|
assert.NilError(t, err)
|
|
assert.Check(t, is.DeepEqual(service.UpdateConfig, expected.UpdateConfig))
|
|
assert.Check(t, is.DeepEqual(service.RollbackConfig, expected.RollbackConfig))
|
|
}
|
|
|
|
func TestToServiceMaxReplicasGlobalModeConflict(t *testing.T) {
|
|
opt := serviceOptions{
|
|
mode: "global",
|
|
maxReplicas: 1,
|
|
}
|
|
_, err := opt.ToServiceMode()
|
|
assert.Error(t, err, "replicas-max-per-node can only be used with replicated mode")
|
|
}
|