
Split out the criService-agnostic bits of nri-api* from pkg/cri/server to pkg/cri/nri to allow sharing a single implementation betwen the server and sbserver versions. Rework the interfaces to not require access to package internals. Signed-off-by: Krisztian Litkey <krisztian.litkey@intel.com>
157 lines
5.3 KiB
Go
157 lines
5.3 KiB
Go
//go:build !darwin && !freebsd
|
|
|
|
/*
|
|
Copyright The containerd Authors.
|
|
|
|
Licensed under the Apache License, Version 2.0 (the "License");
|
|
you may not use this file except in compliance with the License.
|
|
You may obtain a copy of the License at
|
|
|
|
http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
Unless required by applicable law or agreed to in writing, software
|
|
distributed under the License is distributed on an "AS IS" BASIS,
|
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
See the License for the specific language governing permissions and
|
|
limitations under the License.
|
|
*/
|
|
|
|
package server
|
|
|
|
import (
|
|
"context"
|
|
gocontext "context"
|
|
"fmt"
|
|
|
|
"github.com/containerd/containerd"
|
|
"github.com/containerd/containerd/containers"
|
|
"github.com/containerd/containerd/errdefs"
|
|
"github.com/containerd/containerd/log"
|
|
"github.com/containerd/typeurl/v2"
|
|
runtimespec "github.com/opencontainers/runtime-spec/specs-go"
|
|
runtime "k8s.io/cri-api/pkg/apis/runtime/v1"
|
|
|
|
containerstore "github.com/containerd/containerd/pkg/cri/store/container"
|
|
ctrdutil "github.com/containerd/containerd/pkg/cri/util"
|
|
)
|
|
|
|
// UpdateContainerResources updates ContainerConfig of the container.
|
|
func (c *criService) UpdateContainerResources(ctx context.Context, r *runtime.UpdateContainerResourcesRequest) (retRes *runtime.UpdateContainerResourcesResponse, retErr error) {
|
|
container, err := c.containerStore.Get(r.GetContainerId())
|
|
if err != nil {
|
|
return nil, fmt.Errorf("failed to find container: %w", err)
|
|
}
|
|
|
|
sandbox, err := c.sandboxStore.Get(container.SandboxID)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
resources := r.GetLinux()
|
|
updated, err := c.nri.UpdateContainerResources(ctx, &sandbox, &container, resources)
|
|
if err != nil {
|
|
return nil, fmt.Errorf("NRI container update failed: %w", err)
|
|
}
|
|
if updated != nil {
|
|
*resources = *updated
|
|
}
|
|
|
|
// Update resources in status update transaction, so that:
|
|
// 1) There won't be race condition with container start.
|
|
// 2) There won't be concurrent resource update to the same container.
|
|
if err := container.Status.UpdateSync(func(status containerstore.Status) (containerstore.Status, error) {
|
|
return c.updateContainerResources(ctx, container, r, status)
|
|
}); err != nil {
|
|
return nil, fmt.Errorf("failed to update resources: %w", err)
|
|
}
|
|
|
|
err = c.nri.PostUpdateContainerResources(ctx, &sandbox, &container)
|
|
if err != nil {
|
|
log.G(ctx).WithError(err).Errorf("NRI post-update notification failed")
|
|
}
|
|
|
|
return &runtime.UpdateContainerResourcesResponse{}, nil
|
|
}
|
|
|
|
func (c *criService) updateContainerResources(ctx context.Context,
|
|
cntr containerstore.Container,
|
|
r *runtime.UpdateContainerResourcesRequest,
|
|
status containerstore.Status) (newStatus containerstore.Status, retErr error) {
|
|
|
|
newStatus = status
|
|
id := cntr.ID
|
|
// Do not update the container when there is a removal in progress.
|
|
if status.Removing {
|
|
return newStatus, fmt.Errorf("container %q is in removing state", id)
|
|
}
|
|
|
|
// Update container spec. If the container is not started yet, updating
|
|
// spec makes sure that the resource limits are correct when start;
|
|
// if the container is already started, updating spec is still required,
|
|
// the spec will become our source of truth for resource limits.
|
|
oldSpec, err := cntr.Container.Spec(ctx)
|
|
if err != nil {
|
|
return newStatus, fmt.Errorf("failed to get container spec: %w", err)
|
|
}
|
|
newSpec, err := updateOCIResource(ctx, oldSpec, r, c.config)
|
|
if err != nil {
|
|
return newStatus, fmt.Errorf("failed to update resource in spec: %w", err)
|
|
}
|
|
|
|
if err := updateContainerSpec(ctx, cntr.Container, newSpec); err != nil {
|
|
return newStatus, err
|
|
}
|
|
defer func() {
|
|
if retErr != nil {
|
|
deferCtx, deferCancel := ctrdutil.DeferContext()
|
|
defer deferCancel()
|
|
// Reset spec on error.
|
|
if err := updateContainerSpec(deferCtx, cntr.Container, oldSpec); err != nil {
|
|
log.G(ctx).WithError(err).Errorf("Failed to update spec %+v for container %q", oldSpec, id)
|
|
}
|
|
} else {
|
|
// Update container status only when the spec is updated
|
|
newStatus = copyResourcesToStatus(newSpec, status)
|
|
}
|
|
}()
|
|
|
|
// If container is not running, only update spec is enough, new resource
|
|
// limit will be applied when container start.
|
|
if status.State() != runtime.ContainerState_CONTAINER_RUNNING {
|
|
return newStatus, nil
|
|
}
|
|
|
|
task, err := cntr.Container.Task(ctx, nil)
|
|
if err != nil {
|
|
if errdefs.IsNotFound(err) {
|
|
// Task exited already.
|
|
return newStatus, nil
|
|
}
|
|
return newStatus, fmt.Errorf("failed to get task: %w", err)
|
|
}
|
|
// newSpec.Linux / newSpec.Windows won't be nil
|
|
if err := task.Update(ctx, containerd.WithResources(getResources(newSpec))); err != nil {
|
|
if errdefs.IsNotFound(err) {
|
|
// Task exited already.
|
|
return newStatus, nil
|
|
}
|
|
return newStatus, fmt.Errorf("failed to update resources: %w", err)
|
|
}
|
|
return newStatus, nil
|
|
}
|
|
|
|
// updateContainerSpec updates container spec.
|
|
func updateContainerSpec(ctx context.Context, cntr containerd.Container, spec *runtimespec.Spec) error {
|
|
any, err := typeurl.MarshalAny(spec)
|
|
if err != nil {
|
|
return fmt.Errorf("failed to marshal spec %+v: %w", spec, err)
|
|
}
|
|
if err := cntr.Update(ctx, func(ctx gocontext.Context, client *containerd.Client, c *containers.Container) error {
|
|
c.Spec = any
|
|
return nil
|
|
}); err != nil {
|
|
return fmt.Errorf("failed to update container spec: %w", err)
|
|
}
|
|
return nil
|
|
}
|