
This allows container runtimes to propagate an image signature verification error through the CRI and display that to the end user during image pull. There is no other behavioral difference compared to a regular image pull failure. Signed-off-by: Sascha Grunert <sgrunert@redhat.com>
208 lines
8.0 KiB
Go
208 lines
8.0 KiB
Go
/*
|
|
Copyright 2016 The Kubernetes Authors.
|
|
|
|
Licensed under the Apache License, Version 2.0 (the "License");
|
|
you may not use this file except in compliance with the License.
|
|
You may obtain a copy of the License at
|
|
|
|
http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
Unless required by applicable law or agreed to in writing, software
|
|
distributed under the License is distributed on an "AS IS" BASIS,
|
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
See the License for the specific language governing permissions and
|
|
limitations under the License.
|
|
*/
|
|
|
|
package images
|
|
|
|
import (
|
|
"context"
|
|
"fmt"
|
|
"time"
|
|
|
|
dockerref "github.com/docker/distribution/reference"
|
|
v1 "k8s.io/api/core/v1"
|
|
"k8s.io/apimachinery/pkg/types"
|
|
"k8s.io/client-go/tools/record"
|
|
"k8s.io/client-go/util/flowcontrol"
|
|
"k8s.io/klog/v2"
|
|
|
|
runtimeapi "k8s.io/cri-api/pkg/apis/runtime/v1"
|
|
crierrors "k8s.io/cri-api/pkg/errors"
|
|
kubecontainer "k8s.io/kubernetes/pkg/kubelet/container"
|
|
"k8s.io/kubernetes/pkg/kubelet/events"
|
|
)
|
|
|
|
type ImagePodPullingTimeRecorder interface {
|
|
RecordImageStartedPulling(podUID types.UID)
|
|
RecordImageFinishedPulling(podUID types.UID)
|
|
}
|
|
|
|
// imageManager provides the functionalities for image pulling.
|
|
type imageManager struct {
|
|
recorder record.EventRecorder
|
|
imageService kubecontainer.ImageService
|
|
backOff *flowcontrol.Backoff
|
|
// It will check the presence of the image, and report the 'image pulling', image pulled' events correspondingly.
|
|
puller imagePuller
|
|
|
|
podPullingTimeRecorder ImagePodPullingTimeRecorder
|
|
}
|
|
|
|
var _ ImageManager = &imageManager{}
|
|
|
|
// NewImageManager instantiates a new ImageManager object.
|
|
func NewImageManager(recorder record.EventRecorder, imageService kubecontainer.ImageService, imageBackOff *flowcontrol.Backoff, serialized bool, maxParallelImagePulls *int32, qps float32, burst int, podPullingTimeRecorder ImagePodPullingTimeRecorder) ImageManager {
|
|
imageService = throttleImagePulling(imageService, qps, burst)
|
|
|
|
var puller imagePuller
|
|
if serialized {
|
|
puller = newSerialImagePuller(imageService)
|
|
} else {
|
|
puller = newParallelImagePuller(imageService, maxParallelImagePulls)
|
|
}
|
|
return &imageManager{
|
|
recorder: recorder,
|
|
imageService: imageService,
|
|
backOff: imageBackOff,
|
|
puller: puller,
|
|
podPullingTimeRecorder: podPullingTimeRecorder,
|
|
}
|
|
}
|
|
|
|
// shouldPullImage returns whether we should pull an image according to
|
|
// the presence and pull policy of the image.
|
|
func shouldPullImage(container *v1.Container, imagePresent bool) bool {
|
|
if container.ImagePullPolicy == v1.PullNever {
|
|
return false
|
|
}
|
|
|
|
if container.ImagePullPolicy == v1.PullAlways ||
|
|
(container.ImagePullPolicy == v1.PullIfNotPresent && (!imagePresent)) {
|
|
return true
|
|
}
|
|
|
|
return false
|
|
}
|
|
|
|
// records an event using ref, event msg. log to glog using prefix, msg, logFn
|
|
func (m *imageManager) logIt(ref *v1.ObjectReference, eventtype, event, prefix, msg string, logFn func(args ...interface{})) {
|
|
if ref != nil {
|
|
m.recorder.Event(ref, eventtype, event, msg)
|
|
} else {
|
|
logFn(fmt.Sprint(prefix, " ", msg))
|
|
}
|
|
}
|
|
|
|
// EnsureImageExists pulls the image for the specified pod and container, and returns
|
|
// (imageRef, error message, error).
|
|
func (m *imageManager) EnsureImageExists(ctx context.Context, pod *v1.Pod, container *v1.Container, pullSecrets []v1.Secret, podSandboxConfig *runtimeapi.PodSandboxConfig) (string, string, error) {
|
|
logPrefix := fmt.Sprintf("%s/%s/%s", pod.Namespace, pod.Name, container.Image)
|
|
ref, err := kubecontainer.GenerateContainerRef(pod, container)
|
|
if err != nil {
|
|
klog.ErrorS(err, "Couldn't make a ref to pod", "pod", klog.KObj(pod), "containerName", container.Name)
|
|
}
|
|
|
|
// If the image contains no tag or digest, a default tag should be applied.
|
|
image, err := applyDefaultImageTag(container.Image)
|
|
if err != nil {
|
|
msg := fmt.Sprintf("Failed to apply default image tag %q: %v", container.Image, err)
|
|
m.logIt(ref, v1.EventTypeWarning, events.FailedToInspectImage, logPrefix, msg, klog.Warning)
|
|
return "", msg, ErrInvalidImageName
|
|
}
|
|
|
|
var podAnnotations []kubecontainer.Annotation
|
|
for k, v := range pod.GetAnnotations() {
|
|
podAnnotations = append(podAnnotations, kubecontainer.Annotation{
|
|
Name: k,
|
|
Value: v,
|
|
})
|
|
}
|
|
|
|
spec := kubecontainer.ImageSpec{
|
|
Image: image,
|
|
Annotations: podAnnotations,
|
|
}
|
|
imageRef, err := m.imageService.GetImageRef(ctx, spec)
|
|
if err != nil {
|
|
msg := fmt.Sprintf("Failed to inspect image %q: %v", container.Image, err)
|
|
m.logIt(ref, v1.EventTypeWarning, events.FailedToInspectImage, logPrefix, msg, klog.Warning)
|
|
return "", msg, ErrImageInspect
|
|
}
|
|
|
|
present := imageRef != ""
|
|
if !shouldPullImage(container, present) {
|
|
if present {
|
|
msg := fmt.Sprintf("Container image %q already present on machine", container.Image)
|
|
m.logIt(ref, v1.EventTypeNormal, events.PulledImage, logPrefix, msg, klog.Info)
|
|
return imageRef, "", nil
|
|
}
|
|
msg := fmt.Sprintf("Container image %q is not present with pull policy of Never", container.Image)
|
|
m.logIt(ref, v1.EventTypeWarning, events.ErrImageNeverPullPolicy, logPrefix, msg, klog.Warning)
|
|
return "", msg, ErrImageNeverPull
|
|
}
|
|
|
|
backOffKey := fmt.Sprintf("%s_%s", pod.UID, container.Image)
|
|
if m.backOff.IsInBackOffSinceUpdate(backOffKey, m.backOff.Clock.Now()) {
|
|
msg := fmt.Sprintf("Back-off pulling image %q", container.Image)
|
|
m.logIt(ref, v1.EventTypeNormal, events.BackOffPullImage, logPrefix, msg, klog.Info)
|
|
return "", msg, ErrImagePullBackOff
|
|
}
|
|
m.podPullingTimeRecorder.RecordImageStartedPulling(pod.UID)
|
|
m.logIt(ref, v1.EventTypeNormal, events.PullingImage, logPrefix, fmt.Sprintf("Pulling image %q", container.Image), klog.Info)
|
|
startTime := time.Now()
|
|
pullChan := make(chan pullResult)
|
|
m.puller.pullImage(ctx, spec, pullSecrets, pullChan, podSandboxConfig)
|
|
imagePullResult := <-pullChan
|
|
if imagePullResult.err != nil {
|
|
m.logIt(ref, v1.EventTypeWarning, events.FailedToPullImage, logPrefix, fmt.Sprintf("Failed to pull image %q: %v", container.Image, imagePullResult.err), klog.Warning)
|
|
m.backOff.Next(backOffKey, m.backOff.Clock.Now())
|
|
|
|
msg, err := evalCRIPullErr(container, imagePullResult.err)
|
|
return "", msg, err
|
|
}
|
|
m.podPullingTimeRecorder.RecordImageFinishedPulling(pod.UID)
|
|
m.logIt(ref, v1.EventTypeNormal, events.PulledImage, logPrefix, fmt.Sprintf("Successfully pulled image %q in %v (%v including waiting)",
|
|
container.Image, imagePullResult.pullDuration.Truncate(time.Millisecond), time.Since(startTime).Truncate(time.Millisecond)), klog.Info)
|
|
m.backOff.GC()
|
|
return imagePullResult.imageRef, "", nil
|
|
}
|
|
|
|
func evalCRIPullErr(container *v1.Container, err error) (errMsg string, errRes error) {
|
|
// Error assertions via errors.Is is not supported by gRPC (remote runtime) errors right now.
|
|
// See https://github.com/grpc/grpc-go/issues/3616
|
|
if err.Error() == crierrors.ErrRegistryUnavailable.Error() {
|
|
errMsg = fmt.Sprintf("image pull failed for %s because the registry is unavailable.", container.Image)
|
|
return errMsg, crierrors.ErrRegistryUnavailable
|
|
}
|
|
|
|
if err.Error() == crierrors.ErrSignatureValidationFailed.Error() {
|
|
errMsg = fmt.Sprintf("image pull failed for %s because the signature validation failed.", container.Image)
|
|
return errMsg, crierrors.ErrSignatureValidationFailed
|
|
}
|
|
|
|
// Fallback for no specific error
|
|
return err.Error(), ErrImagePull
|
|
}
|
|
|
|
// applyDefaultImageTag parses a docker image string, if it doesn't contain any tag or digest,
|
|
// a default tag will be applied.
|
|
func applyDefaultImageTag(image string) (string, error) {
|
|
named, err := dockerref.ParseNormalizedNamed(image)
|
|
if err != nil {
|
|
return "", fmt.Errorf("couldn't parse image reference %q: %v", image, err)
|
|
}
|
|
_, isTagged := named.(dockerref.Tagged)
|
|
_, isDigested := named.(dockerref.Digested)
|
|
if !isTagged && !isDigested {
|
|
// we just concatenate the image name with the default tag here instead
|
|
// of using dockerref.WithTag(named, ...) because that would cause the
|
|
// image to be fully qualified as docker.io/$name if it's a short name
|
|
// (e.g. just busybox). We don't want that to happen to keep the CRI
|
|
// agnostic wrt image names and default hostnames.
|
|
image = image + ":latest"
|
|
}
|
|
return image, nil
|
|
}
|