
This changes the text registration so that tags for which the framework has a dedicated API (features, feature gates, slow, serial, etc.) those APIs are used. Arbitrary, custom tags are still left in place for now.
130 lines
4.9 KiB
Go
130 lines
4.9 KiB
Go
/*
|
|
Copyright 2019 The Kubernetes Authors.
|
|
|
|
Licensed under the Apache License, Version 2.0 (the "License");
|
|
you may not use this file except in compliance with the License.
|
|
You may obtain a copy of the License at
|
|
|
|
http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
Unless required by applicable law or agreed to in writing, software
|
|
distributed under the License is distributed on an "AS IS" BASIS,
|
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
See the License for the specific language governing permissions and
|
|
limitations under the License.
|
|
*/
|
|
|
|
package gcp
|
|
|
|
import (
|
|
"context"
|
|
"fmt"
|
|
"time"
|
|
|
|
"github.com/onsi/ginkgo/v2"
|
|
v1 "k8s.io/api/core/v1"
|
|
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
|
"k8s.io/apimachinery/pkg/fields"
|
|
"k8s.io/apimachinery/pkg/labels"
|
|
clientset "k8s.io/client-go/kubernetes"
|
|
"k8s.io/kubernetes/test/e2e/feature"
|
|
"k8s.io/kubernetes/test/e2e/framework"
|
|
e2enode "k8s.io/kubernetes/test/e2e/framework/node"
|
|
e2epod "k8s.io/kubernetes/test/e2e/framework/pod"
|
|
"k8s.io/kubernetes/test/e2e/framework/providers/gce"
|
|
e2eskipper "k8s.io/kubernetes/test/e2e/framework/skipper"
|
|
testutils "k8s.io/kubernetes/test/utils"
|
|
admissionapi "k8s.io/pod-security-admission/api"
|
|
)
|
|
|
|
const (
|
|
// recreateNodeReadyAgainTimeout is how long a node is allowed to become "Ready" after it is recreated before
|
|
// the test is considered failed.
|
|
recreateNodeReadyAgainTimeout = 10 * time.Minute
|
|
)
|
|
|
|
var _ = SIGDescribe("Recreate", feature.Recreate, func() {
|
|
f := framework.NewDefaultFramework("recreate")
|
|
f.NamespacePodSecurityLevel = admissionapi.LevelPrivileged
|
|
var originalNodes []v1.Node
|
|
var originalPodNames []string
|
|
var ps *testutils.PodStore
|
|
systemNamespace := metav1.NamespaceSystem
|
|
ginkgo.BeforeEach(func(ctx context.Context) {
|
|
e2eskipper.SkipUnlessProviderIs("gce", "gke")
|
|
var err error
|
|
numNodes, err := e2enode.TotalRegistered(ctx, f.ClientSet)
|
|
framework.ExpectNoError(err)
|
|
originalNodes, err = e2enode.CheckReady(ctx, f.ClientSet, numNodes, framework.NodeReadyInitialTimeout)
|
|
framework.ExpectNoError(err)
|
|
|
|
framework.Logf("Got the following nodes before recreate %v", nodeNames(originalNodes))
|
|
|
|
ps, err = testutils.NewPodStore(f.ClientSet, systemNamespace, labels.Everything(), fields.Everything())
|
|
framework.ExpectNoError(err)
|
|
allPods := ps.List()
|
|
originalPods := e2epod.FilterNonRestartablePods(allPods)
|
|
originalPodNames = make([]string, len(originalPods))
|
|
for i, p := range originalPods {
|
|
originalPodNames[i] = p.ObjectMeta.Name
|
|
}
|
|
|
|
if !e2epod.CheckPodsRunningReadyOrSucceeded(ctx, f.ClientSet, systemNamespace, originalPodNames, framework.PodReadyBeforeTimeout) {
|
|
framework.Failf("At least one pod wasn't running and ready or succeeded at test start.")
|
|
}
|
|
|
|
})
|
|
|
|
ginkgo.AfterEach(func(ctx context.Context) {
|
|
if ginkgo.CurrentSpecReport().Failed() {
|
|
// Make sure that addon/system pods are running, so dump
|
|
// events for the kube-system namespace on failures
|
|
ginkgo.By(fmt.Sprintf("Collecting events from namespace %q.", systemNamespace))
|
|
events, err := f.ClientSet.CoreV1().Events(systemNamespace).List(ctx, metav1.ListOptions{})
|
|
framework.ExpectNoError(err)
|
|
|
|
for _, e := range events.Items {
|
|
framework.Logf("event for %v: %v %v: %v", e.InvolvedObject.Name, e.Source, e.Reason, e.Message)
|
|
}
|
|
}
|
|
if ps != nil {
|
|
ps.Stop()
|
|
}
|
|
})
|
|
|
|
ginkgo.It("recreate nodes and ensure they function upon restart", func(ctx context.Context) {
|
|
testRecreate(ctx, f.ClientSet, ps, systemNamespace, originalNodes, originalPodNames)
|
|
})
|
|
})
|
|
|
|
// Recreate all the nodes in the test instance group
|
|
func testRecreate(ctx context.Context, c clientset.Interface, ps *testutils.PodStore, systemNamespace string, nodes []v1.Node, podNames []string) {
|
|
err := gce.RecreateNodes(c, nodes)
|
|
if err != nil {
|
|
framework.Failf("Test failed; failed to start the restart instance group command.")
|
|
}
|
|
|
|
err = gce.WaitForNodeBootIdsToChange(ctx, c, nodes, recreateNodeReadyAgainTimeout)
|
|
if err != nil {
|
|
framework.Failf("Test failed; failed to recreate at least one node in %v.", recreateNodeReadyAgainTimeout)
|
|
}
|
|
|
|
nodesAfter, err := e2enode.CheckReady(ctx, c, len(nodes), framework.RestartNodeReadyAgainTimeout)
|
|
framework.ExpectNoError(err)
|
|
framework.Logf("Got the following nodes after recreate: %v", nodeNames(nodesAfter))
|
|
|
|
if len(nodes) != len(nodesAfter) {
|
|
framework.Failf("Had %d nodes before nodes were recreated, but now only have %d",
|
|
len(nodes), len(nodesAfter))
|
|
}
|
|
|
|
// Make sure the pods from before node recreation are running/completed
|
|
podCheckStart := time.Now()
|
|
podNamesAfter, err := e2epod.WaitForNRestartablePods(ctx, ps, len(podNames), framework.RestartPodReadyAgainTimeout)
|
|
framework.ExpectNoError(err)
|
|
remaining := framework.RestartPodReadyAgainTimeout - time.Since(podCheckStart)
|
|
if !e2epod.CheckPodsRunningReadyOrSucceeded(ctx, c, systemNamespace, podNamesAfter, remaining) {
|
|
framework.Failf("At least one pod wasn't running and ready after the restart.")
|
|
}
|
|
}
|