
Motivation: For pod-level user namespaces, it's impossible to force the container runtime to join an existing network namespace after creating a new user namespace. According to the capabilities section in [user_namespaces(7)][1], a network namespace created by containerd is owned by the root user namespace. When the container runtime (like runc or crun) creates a new user namespace, it becomes a child of the root user namespace. Processes within this child user namespace are not permitted to access resources owned by the parent user namespace. If the network namespace is not owned by the new user namespace, the container runtime will fail to mount /sys due to the [sysfs: Restrict mounting sysfs][2] patch. Referencing the [cap_capable][3] function in Linux, a process can access a resource if: * The resource is owned by the process's user namespace, and the process has the required capability. * The resource is owned by a child of the process's user namespace, and the owner's user namespace was created by the process's UID. In the context of pod-level user namespaces, the CRI plugin delegates the creation of the network namespace to the container runtime when running the pause container. After the pause container is initialized, the CRI plugin pins the pause container's network namespace into `/run/netns` and then executes the `CNI_ADD` command over it. However, if the pause container is terminated during the pinning process, the CRI plugin might encounter a PID cycle, leading to the `CNI_ADD` command operating on an incorrect network namespace. Moreover, rolling back the `RunPodSandbox` API is complex due to the delegation of network namespace creation. As highlighted in issue #10363, the CRI plugin can lose IP information after a containerd restart, making it challenging to maintain robustness in the RunPodSandbox API. Solution: Allow containerd to create a new user namespace and then create the network namespace within that user namespace. This way, the CRI plugin can force the container runtime to join both the user namespace and the network namespace. Since the network namespace is owned by the newly created user namespace, the container runtime will have the necessary permissions to mount `/sys` on the container's root filesystem. As a result, delegation of network namespace creation is no longer needed. NOTE: * The CRI plugin does not need to pin the newly created user namespace as it does with the network namespace, because the kernel allows retrieving a user namespace reference via [ioctl_ns(2)][4]. As a result, the podsandbox implementation can obtain the user namespace using the `netnsPath` parameter. [1]: <https://man7.org/linux/man-pages/man7/user_namespaces.7.html> [2]: <7dc5dbc879
> [3]: <2c85ebc57b/security/commoncap.c (L65)
> [4]: <https://man7.org/linux/man-pages/man2/ioctl_ns.2.html> Signed-off-by: Wei Fu <fuweid89@gmail.com>
335 lines
12 KiB
Go
335 lines
12 KiB
Go
/*
|
|
Copyright The containerd Authors.
|
|
|
|
Licensed under the Apache License, Version 2.0 (the "License");
|
|
you may not use this file except in compliance with the License.
|
|
You may obtain a copy of the License at
|
|
|
|
http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
Unless required by applicable law or agreed to in writing, software
|
|
distributed under the License is distributed on an "AS IS" BASIS,
|
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
See the License for the specific language governing permissions and
|
|
limitations under the License.
|
|
*/
|
|
|
|
package podsandbox
|
|
|
|
import (
|
|
"context"
|
|
"errors"
|
|
"fmt"
|
|
|
|
"github.com/containerd/log"
|
|
"github.com/containerd/nri"
|
|
v1 "github.com/containerd/nri/types/v1"
|
|
"github.com/containerd/typeurl/v2"
|
|
"github.com/davecgh/go-spew/spew"
|
|
"github.com/opencontainers/selinux/go-selinux"
|
|
runtime "k8s.io/cri-api/pkg/apis/runtime/v1"
|
|
|
|
containerd "github.com/containerd/containerd/v2/client"
|
|
"github.com/containerd/containerd/v2/core/sandbox"
|
|
"github.com/containerd/containerd/v2/core/snapshots"
|
|
criconfig "github.com/containerd/containerd/v2/internal/cri/config"
|
|
crilabels "github.com/containerd/containerd/v2/internal/cri/labels"
|
|
customopts "github.com/containerd/containerd/v2/internal/cri/opts"
|
|
"github.com/containerd/containerd/v2/internal/cri/server/podsandbox/types"
|
|
imagestore "github.com/containerd/containerd/v2/internal/cri/store/image"
|
|
sandboxstore "github.com/containerd/containerd/v2/internal/cri/store/sandbox"
|
|
ctrdutil "github.com/containerd/containerd/v2/internal/cri/util"
|
|
containerdio "github.com/containerd/containerd/v2/pkg/cio"
|
|
"github.com/containerd/errdefs"
|
|
)
|
|
|
|
func init() {
|
|
typeurl.Register(&sandboxstore.Metadata{},
|
|
"github.com/containerd/cri/pkg/store/sandbox", "Metadata")
|
|
}
|
|
|
|
type CleanupErr struct {
|
|
error
|
|
}
|
|
|
|
// Start creates resources required for the sandbox and starts the sandbox. If an error occurs, Start attempts to tear
|
|
// down the created resources. If an error occurs while tearing down resources, a zero-valued response is returned
|
|
// alongside the error. If the teardown was successful, a nil response is returned with the error.
|
|
// TODO(samuelkarp) Determine whether this error indication is reasonable to retain once controller.Delete is implemented.
|
|
func (c *Controller) Start(ctx context.Context, id string) (cin sandbox.ControllerInstance, retErr error) {
|
|
var cleanupErr error
|
|
defer func() {
|
|
if retErr != nil && cleanupErr != nil {
|
|
log.G(ctx).WithField("id", id).WithError(cleanupErr).Errorf("failed to fully teardown sandbox resources after earlier error: %s", retErr)
|
|
retErr = errors.Join(retErr, CleanupErr{cleanupErr})
|
|
}
|
|
}()
|
|
podSandbox := c.store.Get(id)
|
|
if podSandbox == nil {
|
|
return cin, fmt.Errorf("unable to find pod sandbox with id %q: %w", id, errdefs.ErrNotFound)
|
|
}
|
|
metadata := podSandbox.Metadata
|
|
|
|
var (
|
|
config = metadata.Config
|
|
labels = map[string]string{}
|
|
)
|
|
|
|
sandboxImage := c.getSandboxImageName()
|
|
// Ensure sandbox container image snapshot.
|
|
image, err := c.ensureImageExists(ctx, sandboxImage, config, metadata.RuntimeHandler)
|
|
if err != nil {
|
|
return cin, fmt.Errorf("failed to get sandbox image %q: %w", sandboxImage, err)
|
|
}
|
|
|
|
containerdImage, err := c.toContainerdImage(ctx, *image)
|
|
if err != nil {
|
|
return cin, fmt.Errorf("failed to get image from containerd %q: %w", image.ID, err)
|
|
}
|
|
|
|
ociRuntime, err := c.config.GetSandboxRuntime(config, metadata.RuntimeHandler)
|
|
if err != nil {
|
|
return cin, fmt.Errorf("failed to get sandbox runtime: %w", err)
|
|
}
|
|
log.G(ctx).WithField("podsandboxid", id).Debugf("use OCI runtime %+v", ociRuntime)
|
|
|
|
labels["oci_runtime_type"] = ociRuntime.Type
|
|
|
|
// Create sandbox container root directories.
|
|
sandboxRootDir := c.getSandboxRootDir(id)
|
|
if err := c.os.MkdirAll(sandboxRootDir, 0755); err != nil {
|
|
return cin, fmt.Errorf("failed to create sandbox root directory %q: %w",
|
|
sandboxRootDir, err)
|
|
}
|
|
defer func() {
|
|
if retErr != nil && cleanupErr == nil {
|
|
// Cleanup the sandbox root directory.
|
|
if cleanupErr = c.os.RemoveAll(sandboxRootDir); cleanupErr != nil {
|
|
log.G(ctx).WithError(cleanupErr).Errorf("Failed to remove sandbox root directory %q",
|
|
sandboxRootDir)
|
|
}
|
|
}
|
|
}()
|
|
|
|
volatileSandboxRootDir := c.getVolatileSandboxRootDir(id)
|
|
if err := c.os.MkdirAll(volatileSandboxRootDir, 0755); err != nil {
|
|
return cin, fmt.Errorf("failed to create volatile sandbox root directory %q: %w",
|
|
volatileSandboxRootDir, err)
|
|
}
|
|
defer func() {
|
|
if retErr != nil && cleanupErr == nil {
|
|
deferCtx, deferCancel := ctrdutil.DeferContext()
|
|
defer deferCancel()
|
|
// Cleanup the volatile sandbox root directory.
|
|
if cleanupErr = ensureRemoveAll(deferCtx, volatileSandboxRootDir); cleanupErr != nil {
|
|
log.G(ctx).WithError(cleanupErr).Errorf("Failed to remove volatile sandbox root directory %q",
|
|
volatileSandboxRootDir)
|
|
}
|
|
}
|
|
}()
|
|
|
|
// Create sandbox container.
|
|
// NOTE: sandboxContainerSpec SHOULD NOT have side
|
|
// effect, e.g. accessing/creating files, so that we can test
|
|
// it safely.
|
|
spec, err := c.sandboxContainerSpec(id, config, &image.ImageSpec.Config, metadata.NetNSPath, ociRuntime.PodAnnotations)
|
|
if err != nil {
|
|
return cin, fmt.Errorf("failed to generate sandbox container spec: %w", err)
|
|
}
|
|
log.G(ctx).WithField("podsandboxid", id).Debugf("sandbox container spec: %#+v", spew.NewFormatter(spec))
|
|
|
|
metadata.ProcessLabel = spec.Process.SelinuxLabel
|
|
defer func() {
|
|
if retErr != nil {
|
|
selinux.ReleaseLabel(metadata.ProcessLabel)
|
|
}
|
|
}()
|
|
labels["selinux_label"] = metadata.ProcessLabel
|
|
|
|
// handle any KVM based runtime
|
|
if err := modifyProcessLabel(ociRuntime.Type, spec); err != nil {
|
|
return cin, err
|
|
}
|
|
|
|
if config.GetLinux().GetSecurityContext().GetPrivileged() {
|
|
// If privileged don't set selinux label, but we still record the MCS label so that
|
|
// the unused label can be freed later.
|
|
spec.Process.SelinuxLabel = ""
|
|
}
|
|
|
|
// Generate spec options that will be applied to the spec later.
|
|
specOpts, err := c.sandboxContainerSpecOpts(config, &image.ImageSpec.Config)
|
|
if err != nil {
|
|
return cin, fmt.Errorf("failed to generate sandbox container spec options: %w", err)
|
|
}
|
|
|
|
sandboxLabels := buildLabels(config.Labels, image.ImageSpec.Config.Labels, crilabels.ContainerKindSandbox)
|
|
|
|
snapshotterOpt := []snapshots.Opt{snapshots.WithLabels(snapshots.FilterInheritedLabels(config.Annotations))}
|
|
extraSOpts, err := sandboxSnapshotterOpts(config)
|
|
if err != nil {
|
|
return cin, err
|
|
}
|
|
snapshotterOpt = append(snapshotterOpt, extraSOpts...)
|
|
|
|
opts := []containerd.NewContainerOpts{
|
|
containerd.WithSnapshotter(c.imageService.RuntimeSnapshotter(ctx, ociRuntime)),
|
|
customopts.WithNewSnapshot(id, containerdImage, snapshotterOpt...),
|
|
containerd.WithSpec(spec, specOpts...),
|
|
containerd.WithContainerLabels(sandboxLabels),
|
|
containerd.WithContainerExtension(crilabels.SandboxMetadataExtension, &metadata),
|
|
containerd.WithRuntime(ociRuntime.Type, podSandbox.Runtime.Options),
|
|
}
|
|
|
|
container, err := c.client.NewContainer(ctx, id, opts...)
|
|
if err != nil {
|
|
return cin, fmt.Errorf("failed to create containerd container: %w", err)
|
|
}
|
|
podSandbox.Container = container
|
|
defer func() {
|
|
if retErr != nil && cleanupErr == nil {
|
|
deferCtx, deferCancel := ctrdutil.DeferContext()
|
|
defer deferCancel()
|
|
if cleanupErr = container.Delete(deferCtx, containerd.WithSnapshotCleanup); cleanupErr != nil {
|
|
log.G(ctx).WithError(cleanupErr).Errorf("Failed to delete containerd container %q", id)
|
|
}
|
|
podSandbox.Container = nil
|
|
}
|
|
}()
|
|
|
|
// Setup files required for the sandbox.
|
|
if err = c.setupSandboxFiles(id, config); err != nil {
|
|
return cin, fmt.Errorf("failed to setup sandbox files: %w", err)
|
|
}
|
|
defer func() {
|
|
if retErr != nil && cleanupErr == nil {
|
|
if cleanupErr = c.cleanupSandboxFiles(id, config); cleanupErr != nil {
|
|
log.G(ctx).WithError(cleanupErr).Errorf("Failed to cleanup sandbox files in %q",
|
|
sandboxRootDir)
|
|
}
|
|
}
|
|
}()
|
|
|
|
// Update sandbox created timestamp.
|
|
info, err := container.Info(ctx)
|
|
if err != nil {
|
|
return cin, fmt.Errorf("failed to get sandbox container info: %w", err)
|
|
}
|
|
|
|
// Create sandbox task in containerd.
|
|
log.G(ctx).Tracef("Create sandbox container (id=%q, name=%q).", id, metadata.Name)
|
|
|
|
var taskOpts []containerd.NewTaskOpts
|
|
if ociRuntime.Path != "" {
|
|
taskOpts = append(taskOpts, containerd.WithRuntimePath(ociRuntime.Path))
|
|
}
|
|
|
|
// We don't need stdio for sandbox container.
|
|
task, err := container.NewTask(ctx, containerdio.NullIO, taskOpts...)
|
|
if err != nil {
|
|
return cin, fmt.Errorf("failed to create containerd task: %w", err)
|
|
}
|
|
defer func() {
|
|
if retErr != nil && cleanupErr == nil {
|
|
deferCtx, deferCancel := ctrdutil.DeferContext()
|
|
defer deferCancel()
|
|
// Cleanup the sandbox container if an error is returned.
|
|
if _, err := task.Delete(deferCtx, WithNRISandboxDelete(id), containerd.WithProcessKill); err != nil && !errdefs.IsNotFound(err) {
|
|
log.G(ctx).WithError(err).Errorf("Failed to delete sandbox container %q", id)
|
|
cleanupErr = err
|
|
}
|
|
}
|
|
}()
|
|
|
|
// wait is a long running background request, no timeout needed.
|
|
exitCh, err := task.Wait(ctrdutil.NamespacedContext())
|
|
if err != nil {
|
|
return cin, fmt.Errorf("failed to wait for sandbox container task: %w", err)
|
|
}
|
|
|
|
nric, err := nri.New()
|
|
if err != nil {
|
|
return cin, fmt.Errorf("unable to create nri client: %w", err)
|
|
}
|
|
if nric != nil {
|
|
nriSB := &nri.Sandbox{
|
|
ID: id,
|
|
Labels: config.Labels,
|
|
}
|
|
if _, err := nric.InvokeWithSandbox(ctx, task, v1.Create, nriSB); err != nil {
|
|
return cin, fmt.Errorf("nri invoke: %w", err)
|
|
}
|
|
}
|
|
|
|
if err := task.Start(ctx); err != nil {
|
|
return cin, fmt.Errorf("failed to start sandbox container task %q: %w", id, err)
|
|
}
|
|
pid := task.Pid()
|
|
if err := podSandbox.Status.Update(func(status sandboxstore.Status) (sandboxstore.Status, error) {
|
|
status.Pid = pid
|
|
status.State = sandboxstore.StateReady
|
|
status.CreatedAt = info.CreatedAt
|
|
return status, nil
|
|
}); err != nil {
|
|
return cin, fmt.Errorf("failed to update status of pod sandbox %q: %w", id, err)
|
|
}
|
|
|
|
cin.SandboxID = id
|
|
cin.Pid = task.Pid()
|
|
cin.CreatedAt = info.CreatedAt
|
|
cin.Labels = labels
|
|
|
|
go func() {
|
|
if err := c.waitSandboxExit(ctrdutil.NamespacedContext(), podSandbox, exitCh); err != nil {
|
|
log.G(context.Background()).Warnf("failed to wait pod sandbox exit %v", err)
|
|
}
|
|
}()
|
|
|
|
return
|
|
}
|
|
|
|
func (c *Controller) Create(_ctx context.Context, info sandbox.Sandbox, opts ...sandbox.CreateOpt) error {
|
|
metadata := sandboxstore.Metadata{}
|
|
if err := info.GetExtension(MetadataKey, &metadata); err != nil {
|
|
return fmt.Errorf("failed to get sandbox %q metadata: %w", info.ID, err)
|
|
}
|
|
podSandbox := types.NewPodSandbox(info.ID, sandboxstore.Status{State: sandboxstore.StateUnknown})
|
|
podSandbox.Metadata = metadata
|
|
podSandbox.Runtime = info.Runtime
|
|
return c.store.Save(podSandbox)
|
|
}
|
|
|
|
func (c *Controller) ensureImageExists(ctx context.Context, ref string, config *runtime.PodSandboxConfig, runtimeHandler string) (*imagestore.Image, error) {
|
|
image, err := c.imageService.LocalResolve(ref)
|
|
if err != nil && !errdefs.IsNotFound(err) {
|
|
return nil, fmt.Errorf("failed to get image %q: %w", ref, err)
|
|
}
|
|
if err == nil {
|
|
return &image, nil
|
|
}
|
|
// Pull image to ensure the image exists
|
|
// TODO: Cleaner interface
|
|
imageID, err := c.imageService.PullImage(ctx, ref, nil, config, runtimeHandler)
|
|
if err != nil {
|
|
return nil, fmt.Errorf("failed to pull image %q: %w", ref, err)
|
|
}
|
|
newImage, err := c.imageService.GetImage(imageID)
|
|
if err != nil {
|
|
// It's still possible that someone removed the image right after it is pulled.
|
|
return nil, fmt.Errorf("failed to get image %q after pulling: %w", imageID, err)
|
|
}
|
|
return &newImage, nil
|
|
}
|
|
|
|
func (c *Controller) getSandboxImageName() string {
|
|
// returns the name of the sandbox image used to scope pod shared resources used by the pod's containers,
|
|
// if empty return the default sandbox image.
|
|
if c.imageService != nil {
|
|
sandboxImage := c.imageService.PinnedImage("sandbox")
|
|
if sandboxImage != "" {
|
|
return sandboxImage
|
|
}
|
|
}
|
|
return criconfig.DefaultSandboxImage
|
|
}
|