/* Copyright The containerd Authors. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. */ package client import ( "bytes" "context" "fmt" "io" "os" "os/exec" "path" "path/filepath" "runtime" "strings" "syscall" "testing" "time" apievents "github.com/containerd/containerd/v2/api/events" . "github.com/containerd/containerd/v2/client" "github.com/containerd/containerd/v2/core/containers" "github.com/containerd/containerd/v2/core/images" _ "github.com/containerd/containerd/v2/core/runtime" "github.com/containerd/containerd/v2/core/runtime/v2/runc/options" "github.com/containerd/containerd/v2/pkg/cio" "github.com/containerd/containerd/v2/pkg/namespaces" "github.com/containerd/containerd/v2/pkg/oci" "github.com/containerd/containerd/v2/plugins" gogotypes "github.com/containerd/containerd/v2/protobuf/types" "github.com/containerd/continuity/fs" "github.com/containerd/errdefs" "github.com/containerd/go-runc" "github.com/containerd/log/logtest" "github.com/containerd/platforms" "github.com/containerd/typeurl/v2" specs "github.com/opencontainers/runtime-spec/specs-go" "github.com/stretchr/testify/require" ) func empty() cio.Creator { return cio.NullIO } func TestContainerList(t *testing.T) { client, err := newClient(t, address) if err != nil { t.Fatal(err) } defer client.Close() ctx, cancel := testContext(t) defer cancel() containers, err := client.Containers(ctx) if err != nil { t.Fatalf("container list returned error %v", err) } if len(containers) != 0 { t.Errorf("expected 0 containers but received %d", len(containers)) } } func TestNewContainer(t *testing.T) { t.Parallel() id := t.Name() client, err := newClient(t, address) if err != nil { t.Fatal(err) } defer client.Close() ctx, cancel := testContext(t) defer cancel() container, err := client.NewContainer(ctx, id, WithNewSpec()) if err != nil { t.Fatal(err) } defer container.Delete(ctx) if container.ID() != id { t.Errorf("expected container id %q but received %q", id, container.ID()) } if _, err = container.Spec(ctx); err != nil { t.Fatal(err) } if err := container.Delete(ctx); err != nil { t.Fatal(err) } } func TestContainerStart(t *testing.T) { t.Parallel() client, err := newClient(t, address) if err != nil { t.Fatal(err) } defer client.Close() var ( image Image ctx, cancel = testContext(t) id = t.Name() ) defer cancel() image, err = client.GetImage(ctx, testImage) if err != nil { t.Fatal(err) } container, err := client.NewContainer(ctx, id, WithNewSnapshot(id, image), WithNewSpec(oci.WithImageConfig(image), withExitStatus(7))) if err != nil { t.Fatal(err) } defer container.Delete(ctx, WithSnapshotCleanup) task, err := container.NewTask(ctx, empty()) if err != nil { t.Fatal(err) } defer task.Delete(ctx) statusC, err := task.Wait(ctx) if err != nil { t.Fatal(err) } if runtime.GOOS != "windows" { // task.Pid not implemented on Windows if pid := task.Pid(); pid < 1 { t.Errorf("invalid task pid %d", pid) } } if err := task.Start(ctx); err != nil { t.Error(err) task.Delete(ctx) return } status := <-statusC code, _, err := status.Result() if err != nil { t.Fatal(err) } if code != 7 { t.Errorf("expected status 7 from wait but received %d", code) } deleteStatus, err := task.Delete(ctx) if err != nil { t.Fatal(err) } if ec := deleteStatus.ExitCode(); ec != 7 { t.Errorf("expected status 7 from delete but received %d", ec) } } func readShimPath(taskID string) (string, error) { runtime := plugins.RuntimePluginV2.String() + ".task" shimBinaryNamePath := filepath.Join(defaultState, runtime, testNamespace, taskID, "shim-binary-path") shimPath, err := os.ReadFile(shimBinaryNamePath) if err != nil { return "", err } return string(shimPath), nil } func copyShim(shimPath string) (string, error) { tempPath := filepath.Join(os.TempDir(), filepath.Base(shimPath)) if err := fs.CopyFile(tempPath, shimPath); err != nil { return "", err } fi, err := os.Stat(shimPath) if err != nil { return "", err } if err := os.Chmod(tempPath, fi.Mode().Perm()); err != nil { return "", err } return tempPath, nil } func TestContainerStartWithAbsRuntimePath(t *testing.T) { t.Parallel() client, err := newClient(t, address) if err != nil { t.Fatal(err) } defer client.Close() var ( image Image ctx, cancel = testContext(t) id = t.Name() ) defer cancel() image, err = client.GetImage(ctx, testImage) if err != nil { t.Fatal(err) } container, err := client.NewContainer(ctx, id, WithNewSnapshot(id, image), WithNewSpec(oci.WithImageConfig(image), withExitStatus(7))) if err != nil { t.Fatal(err) } defer container.Delete(ctx, WithSnapshotCleanup) // create a temp task to read the default shim path task, err := container.NewTask(ctx, empty()) if err != nil { t.Fatal(err) } defaultShimPath, err := readShimPath(task.ID()) if err != nil { t.Fatal(err) } // remove the temp task if _, err := task.Delete(ctx, WithProcessKill); err != nil { t.Fatal(err) } tempShimPath, err := copyShim(defaultShimPath) if err != nil { t.Fatal(err) } defer os.Remove(tempShimPath) task, err = container.NewTask(ctx, empty(), WithRuntimePath(tempShimPath)) if err != nil { t.Fatal(err) } shimPath, err := readShimPath(task.ID()) if err != nil { t.Fatal(err) } if shimPath != tempShimPath { t.Fatalf("The task's shim path is %s, does not used the specified runtime path: %s", shimPath, tempShimPath) } statusC, err := task.Wait(ctx) if err != nil { t.Fatal(err) } if runtime.GOOS != "windows" { // task.Pid not implemented on Windows if pid := task.Pid(); pid < 1 { t.Errorf("invalid task pid %d", pid) } } if err := task.Start(ctx); err != nil { t.Error(err) task.Delete(ctx) return } status := <-statusC code, _, err := status.Result() if err != nil { t.Fatal(err) } if code != 7 { t.Errorf("expected status 7 from wait but received %d", code) } deleteStatus, err := task.Delete(ctx) if err != nil { t.Fatal(err) } if ec := deleteStatus.ExitCode(); ec != 7 { t.Errorf("expected status 7 from delete but received %d", ec) } } func TestContainerOutput(t *testing.T) { t.Parallel() client, err := newClient(t, address) if err != nil { t.Fatal(err) } defer client.Close() var ( image Image ctx, cancel = testContext(t) id = t.Name() expected = "kingkoye" ) defer cancel() image, err = client.GetImage(ctx, testImage) if err != nil { t.Fatal(err) } container, err := client.NewContainer(ctx, id, WithNewSnapshot(id, image), WithNewSpec(oci.WithImageConfig(image), withProcessArgs("echo", expected))) if err != nil { t.Fatal(err) } defer container.Delete(ctx, WithSnapshotCleanup) stdout := bytes.NewBuffer(nil) task, err := container.NewTask(ctx, cio.NewCreator(withStdout(stdout))) if err != nil { t.Fatal(err) } defer task.Delete(ctx) statusC, err := task.Wait(ctx) if err != nil { t.Fatal(err) } if err := task.Start(ctx); err != nil { t.Fatal(err) } status := <-statusC code, _, err := status.Result() if code != 0 { t.Errorf("expected status 0 but received %d: %v", code, err) } if _, err := task.Delete(ctx); err != nil { t.Fatal(err) } actual := stdout.String() // echo adds a new line expected = expected + newLine if actual != expected { t.Errorf("expected output %q but received %q", expected, actual) } } func withStdout(stdout io.Writer) cio.Opt { return func(streams *cio.Streams) { streams.Stdin = bytes.NewReader(nil) streams.Stdout = stdout streams.Stderr = io.Discard } } func TestContainerExec(t *testing.T) { t.Parallel() client, err := newClient(t, address) if err != nil { t.Fatal(err) } defer client.Close() var ( image Image ctx, cancel = testContext(t) id = t.Name() ) defer cancel() image, err = client.GetImage(ctx, testImage) if err != nil { t.Fatal(err) } container, err := client.NewContainer(ctx, id, WithNewSnapshot(id, image), WithNewSpec(oci.WithImageConfig(image), longCommand)) if err != nil { t.Fatal(err) } defer container.Delete(ctx, WithSnapshotCleanup) task, err := container.NewTask(ctx, empty()) if err != nil { t.Fatal(err) } defer task.Delete(ctx) finishedC, err := task.Wait(ctx) if err != nil { t.Fatal(err) } if err := task.Start(ctx); err != nil { t.Fatal(err) } spec, err := container.Spec(ctx) if err != nil { t.Fatal(err) } // start an exec process without running the original container process info processSpec := spec.Process withExecExitStatus(processSpec, 6) execID := t.Name() + "_exec" process, err := task.Exec(ctx, execID, processSpec, empty()) if err != nil { t.Fatal(err) } processStatusC, err := process.Wait(ctx) if err != nil { t.Fatal(err) } if err := process.Start(ctx); err != nil { t.Fatal(err) } // wait for the exec to return status := <-processStatusC code, _, err := status.Result() if err != nil { t.Fatal(err) } if code != 6 { t.Errorf("expected exec exit code 6 but received %d", code) } deleteStatus, err := process.Delete(ctx) if err != nil { t.Fatal(err) } if ec := deleteStatus.ExitCode(); ec != 6 { t.Errorf("expected delete exit code 6 but received %d", ec) } if err := task.Kill(ctx, syscall.SIGKILL); err != nil { t.Error(err) } <-finishedC } func TestContainerLargeExecArgs(t *testing.T) { t.Parallel() client, err := newClient(t, address) if err != nil { t.Fatal(err) } defer client.Close() var ( image Image ctx, cancel = testContext(t) id = t.Name() ) defer cancel() image, err = client.GetImage(ctx, testImage) if err != nil { t.Fatal(err) } container, err := client.NewContainer(ctx, id, WithNewSnapshot(id, image), WithNewSpec(oci.WithImageConfig(image), longCommand)) if err != nil { t.Fatal(err) } defer container.Delete(ctx, WithSnapshotCleanup) task, err := container.NewTask(ctx, empty()) if err != nil { t.Fatal(err) } defer task.Delete(ctx) finishedC, err := task.Wait(ctx) if err != nil { t.Fatal(err) } if err := task.Start(ctx); err != nil { t.Fatal(err) } spec, err := container.Spec(ctx) if err != nil { t.Fatal(err) } processSpec := spec.Process withExecArgs(processSpec, "echo", strings.Repeat("a", 20000)) execID := t.Name() + "_exec" process, err := task.Exec(ctx, execID, processSpec, empty()) if err != nil { t.Fatal(err) } processStatusC, err := process.Wait(ctx) if err != nil { t.Fatal(err) } if err := process.Start(ctx); err != nil { t.Fatal(err) } // wait for the exec to return status := <-processStatusC if _, _, err := status.Result(); err != nil { t.Fatal(err) } if _, err := process.Delete(ctx); err != nil { t.Fatal(err) } if err := task.Kill(ctx, syscall.SIGKILL); err != nil { t.Error(err) } <-finishedC } func TestContainerPids(t *testing.T) { t.Parallel() client, err := newClient(t, address) if err != nil { t.Fatal(err) } defer client.Close() var ( image Image ctx, cancel = testContext(t) id = t.Name() ) defer cancel() image, err = client.GetImage(ctx, testImage) if err != nil { t.Fatal(err) } container, err := client.NewContainer(ctx, id, WithNewSnapshot(id, image), WithNewSpec(oci.WithImageConfig(image), longCommand)) if err != nil { t.Fatal(err) } defer container.Delete(ctx, WithSnapshotCleanup) task, err := container.NewTask(ctx, empty()) if err != nil { t.Fatal(err) } defer task.Delete(ctx) statusC, err := task.Wait(ctx) if err != nil { t.Fatal(err) } if err := task.Start(ctx); err != nil { t.Fatal(err) } taskPid := task.Pid() if taskPid < 1 { t.Errorf("invalid task pid %d", taskPid) } processes, err := task.Pids(ctx) if err != nil { t.Fatal(err) } l := len(processes) // The point of this test is to see that we successfully can get all of // the pids running in the container and they match the number expected, // but for Windows this concept is a bit different. Windows containers // essentially go through the usermode boot phase of the operating system, // and have quite a few processes and system services running outside of // the "init" process you specify. Because of this, there's not a great // way to say "there should only be N processes running" like we can ensure // for Linux based off the process we asked to run. // // With all that said, on Windows lets check that we're greater than one // ("init" + system services/procs) if runtime.GOOS == "windows" { if l <= 1 { t.Errorf("expected more than one process but received %d", l) } } else { // 2 processes, 1 for sh and one for sleep if l != 2 { t.Errorf("expected 2 process but received %d", l) } } var found bool for _, p := range processes { if p.Pid == taskPid { found = true break } } if !found { t.Errorf("pid %d must be in %+v", taskPid, processes) } if err := task.Kill(ctx, syscall.SIGKILL); err != nil { select { case s := <-statusC: t.Log(s.Result()) default: } t.Error(err) } <-statusC } func TestContainerCloseIO(t *testing.T) { t.Parallel() client, err := newClient(t, address) if err != nil { t.Fatal(err) } defer client.Close() var ( image Image ctx, cancel = testContext(t) id = t.Name() ) defer cancel() image, err = client.GetImage(ctx, testImage) if err != nil { t.Fatal(err) } container, err := client.NewContainer(ctx, id, WithNewSnapshot(id, image), WithNewSpec(oci.WithImageConfig(image), withCat())) if err != nil { t.Fatal(err) } defer container.Delete(ctx, WithSnapshotCleanup) stdout := bytes.NewBuffer(nil) r, w, err := os.Pipe() if err != nil { t.Fatal(err) } task, err := container.NewTask(ctx, cio.NewCreator(cio.WithStreams(r, stdout, io.Discard))) if err != nil { t.Fatal(err) } defer task.Delete(ctx) statusC, err := task.Wait(ctx) if err != nil { t.Fatal(err) } if err := task.Start(ctx); err != nil { t.Fatal(err) } w.Close() if err := task.CloseIO(ctx, WithStdinCloser); err != nil { t.Error(err) } <-statusC } func TestDeleteRunningContainer(t *testing.T) { t.Parallel() client, err := newClient(t, address) if err != nil { t.Fatal(err) } defer client.Close() var ( image Image ctx, cancel = testContext(t) id = t.Name() ) defer cancel() image, err = client.GetImage(ctx, testImage) if err != nil { t.Fatal(err) } container, err := client.NewContainer(ctx, id, WithNewSnapshot(id, image), WithNewSpec(oci.WithImageConfig(image), longCommand)) if err != nil { t.Fatal(err) } defer container.Delete(ctx, WithSnapshotCleanup) task, err := container.NewTask(ctx, empty()) if err != nil { t.Fatal(err) } defer task.Delete(ctx) statusC, err := task.Wait(ctx) if err != nil { t.Fatal(err) } if err := task.Start(ctx); err != nil { t.Fatal(err) } err = container.Delete(ctx, WithSnapshotCleanup) if err == nil { t.Error("delete did not error with running task") } if !errdefs.IsFailedPrecondition(err) { t.Errorf("expected error %q but received %q", errdefs.ErrFailedPrecondition, err) } if err := task.Kill(ctx, syscall.SIGKILL); err != nil { t.Fatal(err) } <-statusC } func TestContainerKill(t *testing.T) { t.Parallel() client, err := newClient(t, address) if err != nil { t.Fatal(err) } defer client.Close() var ( image Image ctx, cancel = testContext(t) id = t.Name() ) defer cancel() image, err = client.GetImage(ctx, testImage) if err != nil { t.Fatal(err) } container, err := client.NewContainer(ctx, id, WithNewSnapshot(id, image), WithNewSpec(oci.WithImageConfig(image), longCommand)) if err != nil { t.Fatal(err) } defer container.Delete(ctx) task, err := container.NewTask(ctx, empty()) if err != nil { t.Fatal(err) } defer task.Delete(ctx) statusC, err := task.Wait(ctx) if err != nil { t.Fatal(err) } if err := task.Start(ctx); err != nil { t.Fatal(err) } if err := task.Kill(ctx, syscall.SIGKILL); err != nil { t.Fatal(err) } <-statusC err = task.Kill(ctx, syscall.SIGTERM) if err == nil { t.Fatal("second call to kill should return an error") } if !errdefs.IsNotFound(err) { t.Errorf("expected error %q but received %q", errdefs.ErrNotFound, err) } } func TestKillContainerDeletedByRunc(t *testing.T) { t.Parallel() if runtime.GOOS == "windows" { t.Skip("Test relies on runc and is not supported on Windows") } // We skip this case when runtime is crun. // More information in https://github.com/containerd/containerd/pull/4214#discussion_r422769497 if f := os.Getenv("RUNC_FLAVOR"); f != "" && f != "runc" { t.Skip("test requires runc") } client, err := newClient(t, address) if err != nil { t.Fatal(err) } defer client.Close() var ( image Image ctx, cancel = testContext(t) id = t.Name() runcRoot = "/tmp/runc-test" ) defer cancel() image, err = client.GetImage(ctx, testImage) if err != nil { t.Fatal(err) } container, err := client.NewContainer(ctx, id, WithNewSnapshot(id, image), WithNewSpec(oci.WithImageConfig(image), longCommand), WithRuntime(client.Runtime(), &options.Options{Root: runcRoot})) if err != nil { t.Fatal(err) } defer container.Delete(ctx) task, err := container.NewTask(ctx, empty()) if err != nil { t.Fatal(err) } defer task.Delete(ctx) statusC, err := task.Wait(ctx) if err != nil { t.Fatal(err) } if err := task.Start(ctx); err != nil { t.Fatal(err) } rcmd := &runc.Runc{ Root: path.Join(runcRoot, testNamespace), } if err := rcmd.Delete(ctx, id, &runc.DeleteOpts{Force: true}); err != nil { t.Fatal(err) } err = task.Kill(ctx, syscall.SIGKILL) if err == nil { t.Fatal("kill should return NotFound error") } else if !errdefs.IsNotFound(err) { t.Errorf("expected error %q but received %q", errdefs.ErrNotFound, err) } select { case <-statusC: case <-time.After(2 * time.Second): t.Errorf("unexpected timeout when try to get exited container's status") } } func TestContainerNoBinaryExists(t *testing.T) { t.Parallel() client, err := newClient(t, address) if err != nil { t.Fatal(err) } defer client.Close() var ( image Image ctx, cancel = testContext(t) id = t.Name() ) defer cancel() image, err = client.GetImage(ctx, testImage) if err != nil { t.Fatal(err) } container, err := client.NewContainer(ctx, id, WithNewSnapshot(id, image), WithNewSpec(oci.WithImageConfig(image), oci.WithProcessArgs("nothing"))) if err != nil { t.Fatal(err) } defer container.Delete(ctx, WithSnapshotCleanup) task, err := container.NewTask(ctx, empty()) switch runtime.GOOS { case "windows": if err != nil { t.Fatalf("failed to create task %v", err) } defer task.Delete(ctx, WithProcessKill) if err := task.Start(ctx); err == nil { t.Error("task.Start() should return an error when binary does not exist") } default: if err == nil { t.Error("NewTask should return an error when binary does not exist") task.Delete(ctx) } } } func TestContainerExecNoBinaryExists(t *testing.T) { t.Parallel() client, err := newClient(t, address) if err != nil { t.Fatal(err) } defer client.Close() var ( image Image ctx, cancel = testContext(t) id = t.Name() ) defer cancel() image, err = client.GetImage(ctx, testImage) if err != nil { t.Fatal(err) } container, err := client.NewContainer(ctx, id, WithNewSnapshot(id, image), WithNewSpec(oci.WithImageConfig(image), longCommand)) if err != nil { t.Fatal(err) } defer container.Delete(ctx, WithSnapshotCleanup) task, err := container.NewTask(ctx, empty()) if err != nil { t.Fatal(err) } defer task.Delete(ctx) finishedC, err := task.Wait(ctx) if err != nil { t.Error(err) } if err := task.Start(ctx); err != nil { t.Fatal(err) } spec, err := container.Spec(ctx) if err != nil { t.Fatal(err) } // start an exec process without running the original container process processSpec := spec.Process processSpec.Args = []string{ "none", } execID := t.Name() + "_exec" process, err := task.Exec(ctx, execID, processSpec, empty()) if err != nil { t.Fatal(err) } defer process.Delete(ctx) if err := process.Start(ctx); err == nil { t.Error("Process.Start should fail when process does not exist") } if err := task.Kill(ctx, syscall.SIGKILL); err != nil { t.Error(err) } <-finishedC } func TestWaitStoppedTask(t *testing.T) { t.Parallel() client, err := newClient(t, address) if err != nil { t.Fatal(err) } defer client.Close() var ( image Image ctx, cancel = testContext(t) id = t.Name() ) defer cancel() image, err = client.GetImage(ctx, testImage) if err != nil { t.Fatal(err) } container, err := client.NewContainer(ctx, id, WithNewSnapshot(id, image), WithNewSpec(oci.WithImageConfig(image), withExitStatus(7))) if err != nil { t.Fatal(err) } defer container.Delete(ctx, WithSnapshotCleanup) task, err := container.NewTask(ctx, empty()) if err != nil { t.Fatal(err) } defer task.Delete(ctx) statusC, err := task.Wait(ctx) if err != nil { t.Fatal(err) } if runtime.GOOS != "windows" { // Getting the pid is not currently implemented on windows if pid := task.Pid(); pid < 1 { t.Errorf("invalid task pid %d", pid) } } if err := task.Start(ctx); err != nil { t.Error(err) task.Delete(ctx) return } // wait for the task to stop then call wait again <-statusC statusC, err = task.Wait(ctx) if err != nil { t.Fatal(err) } status := <-statusC code, _, err := status.Result() if err != nil { t.Fatal(err) } if code != 7 { t.Errorf("exit status from stopped task should be 7 but received %d", code) } } func TestWaitStoppedProcess(t *testing.T) { t.Parallel() client, err := newClient(t, address) if err != nil { t.Fatal(err) } defer client.Close() var ( image Image ctx, cancel = testContext(t) id = t.Name() ) defer cancel() image, err = client.GetImage(ctx, testImage) if err != nil { t.Fatal(err) } container, err := client.NewContainer(ctx, id, WithNewSnapshot(id, image), WithNewSpec(oci.WithImageConfig(image), longCommand)) if err != nil { t.Fatal(err) } defer container.Delete(ctx, WithSnapshotCleanup) task, err := container.NewTask(ctx, empty()) if err != nil { t.Fatal(err) } defer task.Delete(ctx) finishedC, err := task.Wait(ctx) if err != nil { t.Error(err) } if err := task.Start(ctx); err != nil { t.Fatal(err) } spec, err := container.Spec(ctx) if err != nil { t.Fatal(err) } // start an exec process without running the original container process info processSpec := spec.Process withExecExitStatus(processSpec, 6) execID := t.Name() + "_exec" process, err := task.Exec(ctx, execID, processSpec, empty()) if err != nil { t.Fatal(err) } defer process.Delete(ctx) statusC, err := process.Wait(ctx) if err != nil { t.Fatal(err) } if err := process.Start(ctx); err != nil { t.Fatal(err) } // wait for the exec to return <-statusC // try to wait on the process after it has stopped statusC, err = process.Wait(ctx) if err != nil { t.Fatal(err) } status := <-statusC code, _, err := status.Result() if err != nil { t.Fatal(err) } if code != 6 { t.Errorf("exit status from stopped process should be 6 but received %d", code) } if err := task.Kill(ctx, syscall.SIGKILL); err != nil { t.Error(err) } <-finishedC } func TestTaskForceDelete(t *testing.T) { t.Parallel() client, err := newClient(t, address) if err != nil { t.Fatal(err) } defer client.Close() var ( image Image ctx, cancel = testContext(t) id = t.Name() ) defer cancel() image, err = client.GetImage(ctx, testImage) if err != nil { t.Fatal(err) } container, err := client.NewContainer(ctx, id, WithNewSnapshot(id, image), WithNewSpec(oci.WithImageConfig(image), longCommand)) if err != nil { t.Fatal(err) } defer container.Delete(ctx, WithSnapshotCleanup) task, err := container.NewTask(ctx, empty()) if err != nil { t.Fatal(err) } if err := task.Start(ctx); err != nil { t.Fatal(err) } if _, err := task.Delete(ctx); err == nil { t.Error("task.Delete of a running task should create an error") } if _, err := task.Delete(ctx, WithProcessKill); err != nil { t.Fatal(err) } } func TestProcessForceDelete(t *testing.T) { t.Parallel() client, err := newClient(t, address) if err != nil { t.Fatal(err) } defer client.Close() var ( image Image ctx, cancel = testContext(t) id = t.Name() ) defer cancel() image, err = client.GetImage(ctx, testImage) if err != nil { t.Fatal(err) } container, err := client.NewContainer(ctx, id, WithNewSnapshot(id, image), WithNewSpec(oci.WithImageConfig(image), longCommand)) if err != nil { t.Fatal(err) } defer container.Delete(ctx, WithSnapshotCleanup) task, err := container.NewTask(ctx, empty()) if err != nil { t.Fatal(err) } defer task.Delete(ctx) statusC, err := task.Wait(ctx) if err != nil { t.Fatal(err) } // task must be started on windows if err := task.Start(ctx); err != nil { t.Fatal(err) } spec, err := container.Spec(ctx) if err != nil { t.Fatal(err) } processSpec := spec.Process if runtime.GOOS == "windows" { withExecArgs(processSpec, "cmd", "/c", "ping -t localhost") } else { withExecArgs(processSpec, "/bin/sh", "-c", "while true; do sleep 1; done") } execID := t.Name() + "_exec" process, err := task.Exec(ctx, execID, processSpec, empty()) if err != nil { t.Fatal(err) } if err := process.Start(ctx); err != nil { t.Fatal(err) } if _, err := process.Delete(ctx); err == nil { t.Error("process.Delete should return an error when process is running") } if _, err := process.Delete(ctx, WithProcessKill); err != nil { t.Error(err) } if err := task.Kill(ctx, syscall.SIGKILL); err != nil { t.Fatal(err) } <-statusC } func TestContainerHostname(t *testing.T) { t.Parallel() client, err := newClient(t, address) if err != nil { t.Fatal(err) } defer client.Close() var ( image Image ctx, cancel = testContext(t) id = t.Name() expected = "myhostname" ) defer cancel() image, err = client.GetImage(ctx, testImage) if err != nil { t.Fatal(err) } container, err := client.NewContainer(ctx, id, WithNewSnapshot(id, image), WithNewSpec(oci.WithImageConfig(image), withProcessArgs("hostname"), oci.WithHostname(expected), )) if err != nil { t.Fatal(err) } defer container.Delete(ctx, WithSnapshotCleanup) stdout := bytes.NewBuffer(nil) task, err := container.NewTask(ctx, cio.NewCreator(withStdout(stdout))) if err != nil { t.Fatal(err) } defer task.Delete(ctx) statusC, err := task.Wait(ctx) if err != nil { t.Fatal(err) } if err := task.Start(ctx); err != nil { t.Fatal(err) } status := <-statusC code, _, err := status.Result() if err != nil { t.Fatal(err) } if code != 0 { t.Errorf("expected status 0 but received %d", code) } if _, err := task.Delete(ctx); err != nil { t.Fatal(err) } cutset := "\n" if runtime.GOOS == "windows" { cutset = "\r\n" } actual := strings.TrimSuffix(stdout.String(), cutset) if actual != expected { t.Errorf("expected output %q but received %q", expected, actual) } } func TestContainerExitedAtSet(t *testing.T) { t.Parallel() client, err := newClient(t, address) if err != nil { t.Fatal(err) } defer client.Close() var ( image Image ctx, cancel = testContext(t) id = t.Name() ) defer cancel() image, err = client.GetImage(ctx, testImage) if err != nil { t.Fatal(err) } container, err := client.NewContainer(ctx, id, WithNewSnapshot(id, image), WithNewSpec(oci.WithImageConfig(image), withTrue())) if err != nil { t.Fatal(err) } defer container.Delete(ctx, WithSnapshotCleanup) task, err := container.NewTask(ctx, empty()) if err != nil { t.Fatal(err) } defer task.Delete(ctx) statusC, err := task.Wait(ctx) if err != nil { t.Error(err) } startTime := time.Now() if err := task.Start(ctx); err != nil { t.Fatal(err) } status := <-statusC code, _, err := status.Result() if code != 0 { t.Errorf("expected status 0 but received %d (err: %v)", code, err) } if s, err := task.Status(ctx); err != nil { t.Errorf("failed to retrieve status: %v", err) } else if s.ExitTime.After(startTime) == false { t.Errorf("exit time is not after start time: %v <= %v", startTime, s.ExitTime) } if _, err := task.Delete(ctx); err != nil { t.Fatal(err) } } func TestDeleteContainerExecCreated(t *testing.T) { t.Parallel() client, err := newClient(t, address) if err != nil { t.Fatal(err) } defer client.Close() var ( image Image ctx, cancel = testContext(t) id = t.Name() ) defer cancel() image, err = client.GetImage(ctx, testImage) if err != nil { t.Fatal(err) } container, err := client.NewContainer(ctx, id, WithNewSnapshot(id, image), WithNewSpec(oci.WithImageConfig(image), longCommand)) if err != nil { t.Fatal(err) } defer container.Delete(ctx, WithSnapshotCleanup) task, err := container.NewTask(ctx, empty()) if err != nil { t.Fatal(err) } defer task.Delete(ctx) finished, err := task.Wait(ctx) if err != nil { t.Error(err) } if err := task.Start(ctx); err != nil { t.Fatal(err) } spec, err := container.Spec(ctx) if err != nil { t.Fatal(err) } // start an exec process without running the original container process info processSpec := spec.Process withExecExitStatus(processSpec, 6) execID := t.Name() + "_exec" process, err := task.Exec(ctx, execID, processSpec, empty()) if err != nil { t.Fatal(err) } deleteStatus, err := process.Delete(ctx) if err != nil { t.Fatal(err) } if ec := deleteStatus.ExitCode(); ec != 0 { t.Errorf("expected delete exit code 0 but received %d", ec) } if err := task.Kill(ctx, syscall.SIGKILL); err != nil { t.Error(err) } <-finished } func TestContainerMetrics(t *testing.T) { t.Parallel() client, err := newClient(t, address) if err != nil { t.Fatal(err) } defer client.Close() var ( image Image ctx, cancel = testContext(t) id = t.Name() ) defer cancel() image, err = client.GetImage(ctx, testImage) if err != nil { t.Fatal(err) } container, err := client.NewContainer(ctx, id, WithNewSnapshot(id, image), WithNewSpec(oci.WithImageConfig(image), longCommand)) if err != nil { t.Fatal(err) } defer container.Delete(ctx, WithSnapshotCleanup) task, err := container.NewTask(ctx, empty()) if err != nil { t.Fatal(err) } defer task.Delete(ctx, WithProcessKill) statusC, err := task.Wait(ctx) if err != nil { t.Fatal(err) } metric, err := task.Metrics(ctx) if err != nil { t.Error(err) return } if metric.ID != id { t.Errorf("expected metric id %q but received %q", id, metric.ID) } if err := task.Kill(ctx, syscall.SIGKILL); err != nil { t.Fatal(err) } <-statusC } func TestDeletedContainerMetrics(t *testing.T) { t.Parallel() client, err := newClient(t, address) if err != nil { t.Fatal(err) } defer client.Close() var ( image Image ctx, cancel = testContext(t) id = t.Name() ) defer cancel() image, err = client.GetImage(ctx, testImage) if err != nil { t.Fatal(err) } container, err := client.NewContainer(ctx, id, WithNewSnapshot(id, image), WithNewSpec(oci.WithImageConfig(image), withExitStatus(0))) if err != nil { t.Fatal(err) } defer container.Delete(ctx, WithSnapshotCleanup) task, err := container.NewTask(ctx, empty()) if err != nil { t.Fatal(err) } defer task.Delete(ctx) if err := task.Start(ctx); err != nil { t.Fatal(err) } statusC, err := task.Wait(ctx) if err != nil { t.Fatal(err) } <-statusC if _, err := task.Delete(ctx); err != nil { t.Fatal(err) } if _, err := task.Metrics(ctx); err == nil { t.Errorf("Getting metrics of deleted task should have failed") } } func TestContainerExtensions(t *testing.T) { t.Parallel() ctx, cancel := testContext(t) defer cancel() id := t.Name() client, err := newClient(t, address) if err != nil { t.Fatal(err) } defer client.Close() ext := gogotypes.Any{TypeUrl: "test.ext.url", Value: []byte("hello")} container, err := client.NewContainer(ctx, id, WithNewSpec(), WithContainerExtension("hello", &ext)) if err != nil { t.Fatal(err) } defer container.Delete(ctx) checkExt := func(container Container) { cExts, err := container.Extensions(ctx) if err != nil { t.Fatal(err) } if len(cExts) != 1 { t.Errorf("expected 1 container extension") } if actual := cExts["hello"].GetTypeUrl(); actual != ext.TypeUrl { t.Errorf("got unexpected type url for extension: %s", actual) } if actual := cExts["hello"].GetValue(); !bytes.Equal(actual, ext.Value) { t.Errorf("expected extension value %q, got: %q", ext.Value, actual) } } checkExt(container) container, err = client.LoadContainer(ctx, container.ID()) if err != nil { t.Fatal(err) } checkExt(container) } func TestContainerUpdate(t *testing.T) { t.Parallel() ctx, cancel := testContext(t) defer cancel() id := t.Name() client, err := newClient(t, address) if err != nil { t.Fatal(err) } defer client.Close() container, err := client.NewContainer(ctx, id, WithNewSpec()) if err != nil { t.Fatal(err) } defer container.Delete(ctx) spec, err := container.Spec(ctx) if err != nil { t.Fatal(err) } const hostname = "updated-hostname" spec.Hostname = hostname if err := container.Update(ctx, func(ctx context.Context, client *Client, c *containers.Container) error { a, err := typeurl.MarshalAny(spec) if err != nil { return err } c.Spec = a return nil }); err != nil { t.Fatal(err) } if spec, err = container.Spec(ctx); err != nil { t.Fatal(err) } if spec.Hostname != hostname { t.Errorf("hostname %q != %q", spec.Hostname, hostname) } } func TestContainerInfo(t *testing.T) { t.Parallel() ctx, cancel := testContext(t) defer cancel() id := t.Name() client, err := newClient(t, address) if err != nil { t.Fatal(err) } defer client.Close() container, err := client.NewContainer(ctx, id, WithNewSpec()) if err != nil { t.Fatal(err) } defer container.Delete(ctx) info, err := container.Info(ctx) if err != nil { t.Fatal(err) } if info.ID != container.ID() { t.Fatalf("info.ID=%s != container.ID()=%s", info.ID, container.ID()) } } func TestContainerLabels(t *testing.T) { t.Parallel() ctx, cancel := testContext(t) defer cancel() id := t.Name() client, err := newClient(t, address) if err != nil { t.Fatal(err) } defer client.Close() container, err := client.NewContainer(ctx, id, WithNewSpec(), WithContainerLabels(map[string]string{ "test": "yes", })) if err != nil { t.Fatal(err) } defer container.Delete(ctx) labels, err := container.Labels(ctx) if err != nil { t.Fatal(err) } if labels["test"] != "yes" { t.Fatalf("expected label \"test\" to be \"yes\"") } labels["test"] = "no" if labels, err = container.SetLabels(ctx, labels); err != nil { t.Fatal(err) } if labels["test"] != "no" { t.Fatalf("expected label \"test\" to be \"no\"") } } func TestContainerHook(t *testing.T) { // OCI hooks aren't implemented on Windows. This test will actually run fine on Windows if there's a 'ps' binary in the users PATH, but // there's not any actual hook functionality being tested as any of the OCI fields are plain ignored for Windows containers. if runtime.GOOS == "windows" { t.Skip() } t.Parallel() client, err := newClient(t, address) if err != nil { t.Fatal(err) } defer client.Close() var ( image Image ctx, cancel = testContext(t) id = t.Name() ) defer cancel() image, err = client.GetImage(ctx, testImage) if err != nil { t.Fatal(err) } hook := func(_ context.Context, _ oci.Client, _ *containers.Container, s *specs.Spec) error { if s.Hooks == nil { s.Hooks = &specs.Hooks{} } path, err := exec.LookPath("containerd") if err != nil { return err } psPath, err := exec.LookPath("ps") if err != nil { return err } s.Hooks.Prestart = []specs.Hook{ { Path: path, Args: []string{ "containerd", "oci-hook", "--", psPath, "--pid", "{{pid}}", }, Env: os.Environ(), }, } return nil } container, err := client.NewContainer(ctx, id, WithNewSnapshot(id, image), WithNewSpec(oci.WithImageConfig(image), hook)) if err != nil { t.Fatal(err) } defer container.Delete(ctx, WithSnapshotCleanup) task, err := container.NewTask(ctx, empty()) if err != nil { t.Fatal(err) } defer task.Delete(ctx, WithProcessKill) } func TestShimSockLength(t *testing.T) { t.Parallel() // Max length of namespace should be 76 namespace := strings.Repeat("n", 76) ctx, cancel := context.WithCancel(context.Background()) defer cancel() ctx = namespaces.WithNamespace(ctx, namespace) client, err := newClient(t, address) if err != nil { t.Fatal(err) } defer client.Close() image, err := client.Pull(ctx, testImage, WithPlatformMatcher(platforms.Default()), WithPullUnpack, ) if err != nil { t.Fatal(err) } id := strings.Repeat("c", 64) // We don't have limitation with length of container name, // but 64 bytes of sha256 is the common case container, err := client.NewContainer(ctx, id, WithNewSnapshot(id, image), WithNewSpec(oci.WithImageConfig(image), withExitStatus(0)), ) if err != nil { t.Fatal(err) } defer container.Delete(ctx, WithSnapshotCleanup) task, err := container.NewTask(ctx, empty()) if err != nil { t.Fatal(err) } defer task.Delete(ctx) statusC, err := task.Wait(ctx) if err != nil { t.Fatal(err) } if err := task.Start(ctx); err != nil { t.Fatal(err) } <-statusC } func TestContainerExecLargeOutputWithTTY(t *testing.T) { if runtime.GOOS == "windows" { t.Skip("Test does not run on Windows") } t.Parallel() client, err := newClient(t, address) if err != nil { t.Fatal(err) } defer client.Close() var ( image Image ctx, cancel = testContext(t) id = t.Name() ) defer cancel() image, err = client.GetImage(ctx, testImage) if err != nil { t.Fatal(err) } container, err := client.NewContainer(ctx, id, WithNewSnapshot(id, image), WithNewSpec(oci.WithImageConfig(image), longCommand)) if err != nil { t.Fatal(err) } defer container.Delete(ctx, WithSnapshotCleanup) task, err := container.NewTask(ctx, empty()) if err != nil { t.Fatal(err) } defer task.Delete(ctx) finishedC, err := task.Wait(ctx) if err != nil { t.Fatal(err) } if err := task.Start(ctx); err != nil { t.Fatal(err) } for i := 0; i < 100; i++ { spec, err := container.Spec(ctx) if err != nil { t.Fatal(err) } // start an exec process without running the original container process info processSpec := spec.Process withExecArgs(processSpec, "sh", "-c", `seq -s " " 1000000`) stdout := bytes.NewBuffer(nil) execID := t.Name() + "_exec" process, err := task.Exec(ctx, execID, processSpec, cio.NewCreator(withStdout(stdout), withProcessTTY())) if err != nil { t.Fatal(err) } processStatusC, err := process.Wait(ctx) if err != nil { t.Fatal(err) } if err := process.Start(ctx); err != nil { t.Fatal(err) } // wait for the exec to return status := <-processStatusC code, _, err := status.Result() if err != nil { t.Fatal(err) } if code != 0 { t.Errorf("expected exec exit code 0 but received %d", code) } if _, err := process.Delete(ctx); err != nil { t.Fatal(err) } const expectedSuffix = "999999 1000000" stdoutString := stdout.String() if !strings.Contains(stdoutString, expectedSuffix) { t.Fatalf("process output does not end with %q at iteration %d, here are the last 20 characters of the output:\n\n %q", expectedSuffix, i, stdoutString[len(stdoutString)-20:]) } } if err := task.Kill(ctx, syscall.SIGKILL); err != nil { t.Error(err) } <-finishedC } func TestShortRunningTaskPid(t *testing.T) { t.Parallel() client, err := newClient(t, address) if err != nil { t.Fatal(err) } defer client.Close() var ( image Image ctx, cancel = testContext(t) id = t.Name() ) defer cancel() image, err = client.GetImage(ctx, testImage) if err != nil { t.Fatal(err) } container, err := client.NewContainer(ctx, id, WithNewSnapshot(id, image), WithNewSpec(oci.WithImageConfig(image), shortCommand)) if err != nil { t.Fatal(err) } defer container.Delete(ctx, WithSnapshotCleanup) task, err := container.NewTask(ctx, empty()) if err != nil { t.Fatal(err) } defer task.Delete(ctx) finishedC, err := task.Wait(ctx) if err != nil { t.Fatal(err) } if err := task.Start(ctx); err != nil { t.Fatal(err) } int32PID := int32(task.Pid()) if int32PID <= 0 { t.Errorf("Unexpected task pid %d", int32PID) } <-finishedC } func withProcessTTY() cio.Opt { return func(opt *cio.Streams) { cio.WithTerminal(opt) } } // TestRegressionIssue4769 verifies the number of task exit events. // // Issue: https://github.com/containerd/containerd/issues/4769. func TestRegressionIssue4769(t *testing.T) { t.Parallel() client, err := newClient(t, address) if err != nil { t.Fatal(err) } defer client.Close() // use unique namespace to get unique task events id := t.Name() ns := fmt.Sprintf("%s-%s", testNamespace, id) ctx, cancel := context.WithCancel(context.Background()) defer cancel() ctx = namespaces.WithNamespace(ctx, ns) ctx = logtest.WithT(ctx, t) image, err := client.Pull(ctx, testImage, WithPullUnpack) if err != nil { t.Fatal(err) } defer client.ImageService().Delete(ctx, testImage, images.SynchronousDelete()) container, err := client.NewContainer(ctx, id, WithNewSnapshot(id, image), WithNewSpec(oci.WithImageConfig(image), withTrue()), WithRuntime(client.Runtime(), nil), ) if err != nil { t.Fatal(err) } defer container.Delete(ctx, WithSnapshotCleanup) task, err := container.NewTask(ctx, empty()) if err != nil { t.Fatal(err) } defer task.Delete(ctx) statusC, err := task.Wait(ctx) if err != nil { t.Fatal(err) } eventStream, errC := client.EventService().Subscribe(ctx, "namespace=="+ns+",topic~=|^/tasks/exit|") if err := task.Start(ctx); err != nil { t.Fatal(err) } var timeout = 3 * time.Second select { case et := <-statusC: if got := et.ExitCode(); got != 0 { t.Fatal(fmt.Errorf("expect zero exit status, but got %v", got)) } case <-time.After(timeout): t.Fatal(fmt.Errorf("failed to get exit event in time")) } // start to check events select { case et := <-eventStream: if et.Event == nil { t.Fatal(fmt.Errorf("unexpected empty event: %+v", et)) } v, err := typeurl.UnmarshalAny(et.Event) if err != nil { t.Fatal(fmt.Errorf("failed to unmarshal event: %w", err)) } if e, ok := v.(*apievents.TaskExit); !ok { t.Fatal(fmt.Errorf("unexpected event type: %+v", v)) } else if e.ExitStatus != 0 { t.Fatal(fmt.Errorf("expect zero exit status, but got %v", e.ExitStatus)) } case err := <-errC: t.Fatal(fmt.Errorf("unexpected error from event service: %w", err)) case <-time.After(timeout): t.Fatal(fmt.Errorf("failed to get exit event in time")) } if _, err := task.Delete(ctx); err != nil { t.Fatal(err) } // check duplicate event should not show up select { case event := <-eventStream: t.Fatal(fmt.Errorf("unexpected exit event: %+v", event)) case err := <-errC: t.Fatal(fmt.Errorf("unexpected error from event service: %w", err)) case <-time.After(timeout): } } // TestRegressionIssue6429 should not send exit event out if command is not found. // // Issue: https://github.com/containerd/containerd/issues/6429. func TestRegressionIssue6429(t *testing.T) { t.Parallel() if runtime.GOOS == "windows" { t.Skip("Test relies on runc") } client, err := newClient(t, address) if err != nil { t.Fatal(err) } defer client.Close() // use unique namespace to get unique task events id := t.Name() ns := fmt.Sprintf("%s-%s", testNamespace, id) ctx, cancel := context.WithCancel(context.Background()) defer cancel() ctx = namespaces.WithNamespace(ctx, ns) ctx = logtest.WithT(ctx, t) image, err := client.Pull(ctx, testImage, WithPullUnpack) if err != nil { t.Fatal(err) } defer client.ImageService().Delete(ctx, testImage, images.SynchronousDelete()) container, err := client.NewContainer(ctx, id, WithNewSnapshot(id, image), WithNewSpec(oci.WithImageConfig(image), withProcessArgs("notfound404")), WithRuntime(client.Runtime(), nil), ) if err != nil { t.Fatal(err) } defer container.Delete(ctx, WithSnapshotCleanup) eventStream, errC := client.EventService().Subscribe(ctx, "namespace=="+ns+",topic~=|^/tasks/exit|") if _, err := container.NewTask(ctx, empty()); err == nil { t.Fatalf("expected error but got nil") } var timeout = 10 * time.Second // start to check events select { case et := <-eventStream: t.Fatal(fmt.Errorf("unexpected task exit event: %+v", et)) case err := <-errC: t.Fatal(fmt.Errorf("unexpected error from event service: %w", err)) case <-time.After(timeout): } } func TestDaemonRestart(t *testing.T) { client, err := newClient(t, address) if err != nil { t.Fatal(err) } defer client.Close() var ( image Image ctx, cancel = testContext(t) id = t.Name() ) defer cancel() image, err = client.GetImage(ctx, testImage) if err != nil { t.Fatal(err) } container, err := client.NewContainer(ctx, id, WithNewSnapshot(id, image), WithNewSpec(oci.WithImageConfig(image), longCommand)) if err != nil { t.Fatal(err) } defer container.Delete(ctx, WithSnapshotCleanup) task, err := container.NewTask(ctx, empty()) if err != nil { t.Fatal(err) } defer func() { if _, err := task.Delete(ctx, WithProcessKill); err != nil { t.Logf("failed to delete task: %v", err) } }() statusC, err := task.Wait(ctx) if err != nil { t.Fatal(err) } if err := task.Start(ctx); err != nil { t.Fatal(err) } var exitStatus ExitStatus if err := ctrd.Restart(func() { exitStatus = <-statusC }); err != nil { t.Fatal(err) } if exitStatus.Error() == nil { t.Errorf(`first task.Wait() should have failed with "transport is closing"`) } // NOTE(gabriel-samfira): Windows needs a bit more time to restart. // Increase timeout from 2 seconds to 10 seconds to avoid deadline // exceeded errors. waitCtx, waitCancel := context.WithTimeout(ctx, 10*time.Second) serving, err := client.IsServing(waitCtx) waitCancel() if !serving { t.Fatalf("containerd did not start within 2s: %v", err) } statusC, err = task.Wait(ctx) if err != nil { t.Fatal(err) } if err := task.Kill(ctx, syscall.SIGKILL); err != nil { t.Fatal(err) } <-statusC } type directIO struct { cio.DirectIO } // ioCreate returns IO available for use with task creation func (f *directIO) IOCreate(id string) (cio.IO, error) { return f, nil } // ioAttach returns IO available for use with task attachment func (f *directIO) IOAttach(set *cio.FIFOSet) (cio.IO, error) { return f, nil } func (f *directIO) Cancel() { // nothing to cancel as all operations are handled externally } // Close closes all open fds func (f *directIO) Close() error { err := f.Stdin.Close() if f.Stdout != nil { if err2 := f.Stdout.Close(); err == nil { err = err2 } } if f.Stderr != nil { if err2 := f.Stderr.Close(); err == nil { err = err2 } } return err } // Delete removes the underlying directory containing fifos func (f *directIO) Delete() error { return f.DirectIO.Close() } func initContainerAndCheckChildrenDieOnKill(t *testing.T, opts ...oci.SpecOpts) { client, err := newClient(t, address) if err != nil { t.Fatal(err) } defer client.Close() var ( image Image ctx, cancel = testContext(t) id = t.Name() ) defer cancel() image, err = client.GetImage(ctx, testImage) if err != nil { t.Fatal(err) } opts = append(opts, oci.WithImageConfig(image)) opts = append(opts, longCommand) container, err := client.NewContainer(ctx, id, WithNewSnapshot(id, image), WithNewSpec(opts...), ) if err != nil { t.Fatal(err) } defer container.Delete(ctx, WithSnapshotCleanup) task, err := container.NewTask(ctx, cio.NullIO) if err != nil { t.Fatal(err) } defer task.Delete(ctx) statusC, err := task.Wait(ctx) if err != nil { t.Fatal(err) } if err := task.Start(ctx); err != nil { t.Fatal(err) } if err := task.Kill(ctx, syscall.SIGKILL); err != nil { t.Error(err) } // Give the shim time to reap the init process and kill the orphans select { case <-statusC: case <-time.After(100 * time.Millisecond): } command := []string{"ps", "ax"} if runtime.GOOS == "windows" { command = []string{"tasklist"} } b, err := exec.Command(command[0], command[1:]...).CombinedOutput() if err != nil { t.Fatal(err) } // The container is using longCommand, which contains sleep 1 on Linux, and ping -t localhost on Windows. if strings.Contains(string(b), "sleep 1") || strings.Contains(string(b), "ping -t localhost") { t.Fatalf("killing init didn't kill all its children:\n%v", string(b)) } if _, err := task.Delete(ctx, WithProcessKill); err != nil { t.Error(err) } } func TestContainerKillInitKillsChildWhenNotHostPid(t *testing.T) { initContainerAndCheckChildrenDieOnKill(t) } func TestTaskResize(t *testing.T) { if runtime.GOOS == "windows" { // PowerShell can, but nanoserver images don't have that. t.Skipf("%q doesn't have a way to query its terminal size", testImage) } t.Parallel() client, err := newClient(t, address) if err != nil { t.Fatal(err) } defer client.Close() var ( image Image ctx, cancel = testContext(t) id = t.Name() ) defer cancel() image, err = client.GetImage(ctx, testImage) if err != nil { t.Fatal(err) } container, err := client.NewContainer(ctx, id, WithNewSnapshot(id, image), WithNewSpec(oci.WithImageConfig(image), withProcessArgs("/bin/stty", "size"), oci.WithTTY), ) if err != nil { t.Fatal(err) } defer container.Delete(ctx, WithSnapshotCleanup) stdout := &bytes.Buffer{} task, err := container.NewTask(ctx, cio.NewCreator(cio.WithStreams(nil, stdout, nil), cio.WithTerminal), ) if err != nil { t.Fatal(err) } defer task.Delete(ctx) if err := task.Resize(ctx, 12, 34); err != nil { t.Fatal(err) } err = task.Start(ctx) if err != nil { t.Fatal(err) } statusC, err := task.Wait(ctx) if err != nil { t.Fatal(err) } <-statusC task.Kill(ctx, syscall.SIGKILL) _, err = task.Delete(ctx) if err != nil { t.Fatal(err) } require.Equal(t, "34 12\r\n", stdout.String()) } func TestContainerImage(t *testing.T) { t.Parallel() ctx, cancel := testContext(t) defer cancel() id := t.Name() client, err := newClient(t, address) if err != nil { t.Fatal(err) } defer client.Close() image, err := client.GetImage(ctx, testImage) if err != nil { t.Fatal(err) } container, err := client.NewContainer(ctx, id, WithNewSpec(), WithImage(image)) if err != nil { t.Fatal(err) } defer container.Delete(ctx) i, err := container.Image(ctx) if err != nil { t.Fatal(err) } if i.Name() != image.Name() { t.Fatalf("expected container image name %s but received %s", image.Name(), i.Name()) } } func TestContainerNoImage(t *testing.T) { t.Parallel() ctx, cancel := testContext(t) defer cancel() id := t.Name() client, err := newClient(t, address) if err != nil { t.Fatal(err) } defer client.Close() container, err := client.NewContainer(ctx, id, WithNewSpec()) if err != nil { t.Fatal(err) } defer container.Delete(ctx) _, err = container.Image(ctx) if err == nil { t.Fatal("error should not be nil when container is created without an image") } if !errdefs.IsNotFound(err) { t.Fatalf("expected error to be %s but received %s", errdefs.ErrNotFound, err) } } func TestContainerNoSTDIN(t *testing.T) { t.Parallel() client, err := newClient(t, address) if err != nil { t.Fatal(err) } defer client.Close() var ( image Image ctx, cancel = testContext(t) id = t.Name() ) defer cancel() image, err = client.GetImage(ctx, testImage) if err != nil { t.Fatal(err) } container, err := client.NewContainer(ctx, id, WithNewSnapshot(id, image), WithNewSpec(oci.WithImageConfig(image), withExitStatus(0))) if err != nil { t.Fatal(err) } defer container.Delete(ctx, WithSnapshotCleanup) task, err := container.NewTask(ctx, cio.NewCreator(cio.WithStreams(nil, io.Discard, io.Discard))) if err != nil { t.Fatal(err) } defer task.Delete(ctx) statusC, err := task.Wait(ctx) if err != nil { t.Fatal(err) } if err := task.Start(ctx); err != nil { t.Fatal(err) } status := <-statusC code, _, err := status.Result() if err != nil { t.Fatal(err) } if code != 0 { t.Errorf("expected status 0 from wait but received %d", code) } } func TestTaskSpec(t *testing.T) { t.Parallel() client, err := newClient(t, address) if err != nil { t.Fatal(err) } defer client.Close() var ( image Image ctx, cancel = testContext(t) id = t.Name() ) defer cancel() image, err = client.GetImage(ctx, testImage) if err != nil { t.Fatal(err) } container, err := client.NewContainer(ctx, id, WithNewSnapshot(id, image), WithNewSpec(oci.WithImageConfig(image), longCommand)) if err != nil { t.Fatal(err) } defer container.Delete(ctx, WithSnapshotCleanup) task, err := container.NewTask(ctx, empty()) if err != nil { t.Fatal(err) } defer task.Delete(ctx) statusC, err := task.Wait(ctx) if err != nil { t.Fatal(err) } spec, err := task.Spec(ctx) if err != nil { t.Fatal(err) } if spec == nil { t.Fatal("spec from task is nil") } direct, err := newDirectIO(ctx, false) if err != nil { t.Fatal(err) } defer direct.Delete() lt, err := container.Task(ctx, direct.IOAttach) if err != nil { t.Fatal(err) } spec, err = lt.Spec(ctx) if err != nil { t.Fatal(err) } if spec == nil { t.Fatal("spec from loaded task is nil") } if err := task.Kill(ctx, syscall.SIGKILL); err != nil { t.Fatal(err) } <-statusC } func TestContainerUsername(t *testing.T) { t.Parallel() client, err := newClient(t, address) if err != nil { t.Fatal(err) } defer client.Close() var ( image Image ctx, cancel = testContext(t) id = t.Name() ) defer cancel() image, err = client.GetImage(ctx, testImage) if err != nil { t.Fatal(err) } username := "www-data" command := []string{ "id", "-u", } expectedOutput := "33" if runtime.GOOS == "windows" { username = "ContainerUser" command = []string{ "echo", `%USERNAME%`, } expectedOutput = "ContainerUser" } // the www-data user in the busybox image has a uid of 33 container, err := client.NewContainer(ctx, id, WithNewSnapshot(id, image), WithNewSpec(oci.WithImageConfig(image), oci.WithUsername(username), withProcessArgs(command...)), ) if err != nil { t.Fatal(err) } defer container.Delete(ctx, WithSnapshotCleanup) buf := bytes.NewBuffer(nil) task, err := container.NewTask(ctx, cio.NewCreator(withStdout(buf))) if err != nil { t.Fatal(err) } statusC, err := task.Wait(ctx) if err != nil { t.Fatal(err) } if err := task.Start(ctx); err != nil { t.Fatal(err) } <-statusC if _, err := task.Delete(ctx); err != nil { t.Fatal(err) } output := strings.TrimSuffix(buf.String(), newLine) if output != expectedOutput { t.Errorf("expected %s uid to be %s but received %q", username, expectedOutput, buf.String()) } } func TestContainerPTY(t *testing.T) { t.Parallel() client, err := newClient(t, address) if err != nil { t.Fatal(err) } defer client.Close() var ( image Image ctx, cancel = testContext(t) id = t.Name() ) defer cancel() image, err = client.GetImage(ctx, testImage) if err != nil { t.Fatal(err) } container, err := client.NewContainer(ctx, id, WithNewSnapshot(id, image), WithNewSpec(oci.WithImageConfig(image), oci.WithTTY, withProcessArgs("echo", "hello"))) if err != nil { t.Fatal(err) } defer container.Delete(ctx, WithSnapshotCleanup) buf := bytes.NewBuffer(nil) task, err := container.NewTask(ctx, cio.NewCreator(withStdout(buf), withProcessTTY())) if err != nil { t.Fatal(err) } defer task.Delete(ctx) statusC, err := task.Wait(ctx) if err != nil { t.Error(err) } if err := task.Start(ctx); err != nil { t.Fatal(err) } <-statusC if _, err := task.Delete(ctx); err != nil { t.Fatal(err) } out := buf.String() if !strings.ContainsAny(fmt.Sprintf("%#q", out), `\x00`) { t.Fatal(`expected \x00 in output`) } }