Merge pull request #123399 from pohly/test-context-use-cancellation
test: use cancellation from ktesting
This commit is contained in:
		| @@ -1538,12 +1538,10 @@ func TestStalePodDisruption(t *testing.T) { | ||||
| 	} | ||||
| 	for name, tc := range cases { | ||||
| 		t.Run(name, func(t *testing.T) { | ||||
| 			_, ctx := ktesting.NewTestContext(t) | ||||
| 			ctx, cancel := context.WithCancel(ctx) | ||||
| 			defer cancel() | ||||
| 			dc, _ := newFakeDisruptionControllerWithTime(ctx, now) | ||||
| 			go dc.Run(ctx) | ||||
| 			if _, err := dc.coreClient.CoreV1().Pods(tc.pod.Namespace).Create(ctx, tc.pod, metav1.CreateOptions{}); err != nil { | ||||
| 			tCtx := ktesting.Init(t) | ||||
| 			dc, _ := newFakeDisruptionControllerWithTime(tCtx, now) | ||||
| 			go dc.Run(tCtx) | ||||
| 			if _, err := dc.coreClient.CoreV1().Pods(tc.pod.Namespace).Create(tCtx, tc.pod, metav1.CreateOptions{}); err != nil { | ||||
| 				t.Fatalf("Failed to create pod: %v", err) | ||||
| 			} | ||||
| 			dc.clock.Sleep(tc.timePassed) | ||||
| @@ -1552,7 +1550,7 @@ func TestStalePodDisruption(t *testing.T) { | ||||
| 			} | ||||
| 			diff := "" | ||||
| 			if err := wait.Poll(100*time.Millisecond, wait.ForeverTestTimeout, func() (bool, error) { | ||||
| 				pod, err := dc.kubeClient.CoreV1().Pods(tc.pod.Namespace).Get(ctx, tc.pod.Name, metav1.GetOptions{}) | ||||
| 				pod, err := dc.kubeClient.CoreV1().Pods(tc.pod.Namespace).Get(tCtx, tc.pod.Name, metav1.GetOptions{}) | ||||
| 				if err != nil { | ||||
| 					t.Fatalf("Failed getting updated pod: %v", err) | ||||
| 				} | ||||
|   | ||||
| @@ -137,11 +137,9 @@ func expectNoPodUpdate(t *testing.T, ch <-chan kubetypes.PodUpdate) { | ||||
| } | ||||
|  | ||||
| func TestNewPodAdded(t *testing.T) { | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
| 	tCtx := ktesting.Init(t) | ||||
|  | ||||
| 	channel, ch, config := createPodConfigTester(ctx, PodConfigNotificationIncremental) | ||||
| 	channel, ch, config := createPodConfigTester(tCtx, PodConfigNotificationIncremental) | ||||
|  | ||||
| 	// see an update | ||||
| 	podUpdate := CreatePodUpdate(kubetypes.ADD, TestSource, CreateValidPod("foo", "new")) | ||||
| @@ -153,11 +151,9 @@ func TestNewPodAdded(t *testing.T) { | ||||
| } | ||||
|  | ||||
| func TestNewPodAddedInvalidNamespace(t *testing.T) { | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
| 	tCtx := ktesting.Init(t) | ||||
|  | ||||
| 	channel, ch, config := createPodConfigTester(ctx, PodConfigNotificationIncremental) | ||||
| 	channel, ch, config := createPodConfigTester(tCtx, PodConfigNotificationIncremental) | ||||
|  | ||||
| 	// see an update | ||||
| 	podUpdate := CreatePodUpdate(kubetypes.ADD, TestSource, CreateValidPod("foo", "")) | ||||
| @@ -169,11 +165,9 @@ func TestNewPodAddedInvalidNamespace(t *testing.T) { | ||||
| } | ||||
|  | ||||
| func TestNewPodAddedDefaultNamespace(t *testing.T) { | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
| 	tCtx := ktesting.Init(t) | ||||
|  | ||||
| 	channel, ch, config := createPodConfigTester(ctx, PodConfigNotificationIncremental) | ||||
| 	channel, ch, config := createPodConfigTester(tCtx, PodConfigNotificationIncremental) | ||||
|  | ||||
| 	// see an update | ||||
| 	podUpdate := CreatePodUpdate(kubetypes.ADD, TestSource, CreateValidPod("foo", "default")) | ||||
| @@ -185,11 +179,9 @@ func TestNewPodAddedDefaultNamespace(t *testing.T) { | ||||
| } | ||||
|  | ||||
| func TestNewPodAddedDifferentNamespaces(t *testing.T) { | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
| 	tCtx := ktesting.Init(t) | ||||
|  | ||||
| 	channel, ch, config := createPodConfigTester(ctx, PodConfigNotificationIncremental) | ||||
| 	channel, ch, config := createPodConfigTester(tCtx, PodConfigNotificationIncremental) | ||||
|  | ||||
| 	// see an update | ||||
| 	podUpdate := CreatePodUpdate(kubetypes.ADD, TestSource, CreateValidPod("foo", "default")) | ||||
| @@ -206,11 +198,9 @@ func TestNewPodAddedDifferentNamespaces(t *testing.T) { | ||||
| } | ||||
|  | ||||
| func TestInvalidPodFiltered(t *testing.T) { | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
| 	tCtx := ktesting.Init(t) | ||||
|  | ||||
| 	channel, ch, _ := createPodConfigTester(ctx, PodConfigNotificationIncremental) | ||||
| 	channel, ch, _ := createPodConfigTester(tCtx, PodConfigNotificationIncremental) | ||||
|  | ||||
| 	// see an update | ||||
| 	podUpdate := CreatePodUpdate(kubetypes.ADD, TestSource, CreateValidPod("foo", "new")) | ||||
| @@ -224,11 +214,9 @@ func TestInvalidPodFiltered(t *testing.T) { | ||||
| } | ||||
|  | ||||
| func TestNewPodAddedSnapshotAndUpdates(t *testing.T) { | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
| 	tCtx := ktesting.Init(t) | ||||
|  | ||||
| 	channel, ch, config := createPodConfigTester(ctx, PodConfigNotificationSnapshotAndUpdates) | ||||
| 	channel, ch, config := createPodConfigTester(tCtx, PodConfigNotificationSnapshotAndUpdates) | ||||
|  | ||||
| 	// see an set | ||||
| 	podUpdate := CreatePodUpdate(kubetypes.ADD, TestSource, CreateValidPod("foo", "new")) | ||||
| @@ -246,11 +234,9 @@ func TestNewPodAddedSnapshotAndUpdates(t *testing.T) { | ||||
| } | ||||
|  | ||||
| func TestNewPodAddedSnapshot(t *testing.T) { | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
| 	tCtx := ktesting.Init(t) | ||||
|  | ||||
| 	channel, ch, config := createPodConfigTester(ctx, PodConfigNotificationSnapshot) | ||||
| 	channel, ch, config := createPodConfigTester(tCtx, PodConfigNotificationSnapshot) | ||||
|  | ||||
| 	// see an set | ||||
| 	podUpdate := CreatePodUpdate(kubetypes.ADD, TestSource, CreateValidPod("foo", "new")) | ||||
| @@ -268,11 +254,9 @@ func TestNewPodAddedSnapshot(t *testing.T) { | ||||
| } | ||||
|  | ||||
| func TestNewPodAddedUpdatedRemoved(t *testing.T) { | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
| 	tCtx := ktesting.Init(t) | ||||
|  | ||||
| 	channel, ch, _ := createPodConfigTester(ctx, PodConfigNotificationIncremental) | ||||
| 	channel, ch, _ := createPodConfigTester(tCtx, PodConfigNotificationIncremental) | ||||
|  | ||||
| 	// should register an add | ||||
| 	podUpdate := CreatePodUpdate(kubetypes.ADD, TestSource, CreateValidPod("foo", "new")) | ||||
| @@ -295,11 +279,9 @@ func TestNewPodAddedUpdatedRemoved(t *testing.T) { | ||||
| } | ||||
|  | ||||
| func TestNewPodAddedDelete(t *testing.T) { | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
| 	tCtx := ktesting.Init(t) | ||||
|  | ||||
| 	channel, ch, _ := createPodConfigTester(ctx, PodConfigNotificationIncremental) | ||||
| 	channel, ch, _ := createPodConfigTester(tCtx, PodConfigNotificationIncremental) | ||||
|  | ||||
| 	// should register an add | ||||
| 	addedPod := CreateValidPod("foo", "new") | ||||
| @@ -318,11 +300,9 @@ func TestNewPodAddedDelete(t *testing.T) { | ||||
| } | ||||
|  | ||||
| func TestNewPodAddedUpdatedSet(t *testing.T) { | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
| 	tCtx := ktesting.Init(t) | ||||
|  | ||||
| 	channel, ch, _ := createPodConfigTester(ctx, PodConfigNotificationIncremental) | ||||
| 	channel, ch, _ := createPodConfigTester(tCtx, PodConfigNotificationIncremental) | ||||
|  | ||||
| 	// should register an add | ||||
| 	podUpdate := CreatePodUpdate(kubetypes.ADD, TestSource, CreateValidPod("foo", "new"), CreateValidPod("foo2", "new"), CreateValidPod("foo3", "new")) | ||||
| @@ -344,9 +324,7 @@ func TestNewPodAddedUpdatedSet(t *testing.T) { | ||||
| } | ||||
|  | ||||
| func TestNewPodAddedSetReconciled(t *testing.T) { | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
| 	tCtx := ktesting.Init(t) | ||||
|  | ||||
| 	// Create and touch new test pods, return the new pods and touched pod. We should create new pod list | ||||
| 	// before touching to avoid data race. | ||||
| @@ -370,7 +348,7 @@ func TestNewPodAddedSetReconciled(t *testing.T) { | ||||
| 	} { | ||||
| 		var podWithStatusChange *v1.Pod | ||||
| 		pods, _ := newTestPods(false, false) | ||||
| 		channel, ch, _ := createPodConfigTester(ctx, PodConfigNotificationIncremental) | ||||
| 		channel, ch, _ := createPodConfigTester(tCtx, PodConfigNotificationIncremental) | ||||
|  | ||||
| 		// Use SET to initialize the config, especially initialize the source set | ||||
| 		channel <- CreatePodUpdate(kubetypes.SET, TestSource, pods...) | ||||
| @@ -393,9 +371,7 @@ func TestNewPodAddedSetReconciled(t *testing.T) { | ||||
| } | ||||
|  | ||||
| func TestInitialEmptySet(t *testing.T) { | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
| 	tCtx := ktesting.Init(t) | ||||
|  | ||||
| 	for _, test := range []struct { | ||||
| 		mode PodConfigNotificationMode | ||||
| @@ -405,7 +381,7 @@ func TestInitialEmptySet(t *testing.T) { | ||||
| 		{PodConfigNotificationSnapshot, kubetypes.SET}, | ||||
| 		{PodConfigNotificationSnapshotAndUpdates, kubetypes.SET}, | ||||
| 	} { | ||||
| 		channel, ch, _ := createPodConfigTester(ctx, test.mode) | ||||
| 		channel, ch, _ := createPodConfigTester(tCtx, test.mode) | ||||
|  | ||||
| 		// should register an empty PodUpdate operation | ||||
| 		podUpdate := CreatePodUpdate(kubetypes.SET, TestSource) | ||||
| @@ -422,11 +398,9 @@ func TestInitialEmptySet(t *testing.T) { | ||||
| } | ||||
|  | ||||
| func TestPodUpdateAnnotations(t *testing.T) { | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
| 	tCtx := ktesting.Init(t) | ||||
|  | ||||
| 	channel, ch, _ := createPodConfigTester(ctx, PodConfigNotificationIncremental) | ||||
| 	channel, ch, _ := createPodConfigTester(tCtx, PodConfigNotificationIncremental) | ||||
|  | ||||
| 	pod := CreateValidPod("foo2", "new") | ||||
| 	pod.Annotations = make(map[string]string) | ||||
| @@ -455,11 +429,9 @@ func TestPodUpdateAnnotations(t *testing.T) { | ||||
| } | ||||
|  | ||||
| func TestPodUpdateLabels(t *testing.T) { | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
| 	tCtx := ktesting.Init(t) | ||||
|  | ||||
| 	channel, ch, _ := createPodConfigTester(ctx, PodConfigNotificationIncremental) | ||||
| 	channel, ch, _ := createPodConfigTester(tCtx, PodConfigNotificationIncremental) | ||||
|  | ||||
| 	pod := CreateValidPod("foo2", "new") | ||||
| 	pod.Labels = make(map[string]string) | ||||
| @@ -479,11 +451,9 @@ func TestPodUpdateLabels(t *testing.T) { | ||||
| } | ||||
|  | ||||
| func TestPodConfigRace(t *testing.T) { | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
| 	tCtx := ktesting.Init(t) | ||||
|  | ||||
| 	eventBroadcaster := record.NewBroadcaster(record.WithContext(ctx)) | ||||
| 	eventBroadcaster := record.NewBroadcaster(record.WithContext(tCtx)) | ||||
| 	config := NewPodConfig(PodConfigNotificationIncremental, eventBroadcaster.NewRecorder(scheme.Scheme, v1.EventSource{Component: "kubelet"}), &mockPodStartupSLIObserver{}) | ||||
| 	seenSources := sets.NewString(TestSource) | ||||
| 	var wg sync.WaitGroup | ||||
| @@ -491,7 +461,7 @@ func TestPodConfigRace(t *testing.T) { | ||||
| 	wg.Add(2) | ||||
|  | ||||
| 	go func() { | ||||
| 		ctx, cancel := context.WithCancel(ctx) | ||||
| 		ctx, cancel := context.WithCancel(tCtx) | ||||
| 		defer cancel() | ||||
| 		defer wg.Done() | ||||
| 		for i := 0; i < iterations; i++ { | ||||
|   | ||||
| @@ -75,11 +75,10 @@ func setup(t *testing.T, groupVersions ...schema.GroupVersion) (context.Context, | ||||
| 	return setupWithResources(t, groupVersions, nil) | ||||
| } | ||||
|  | ||||
| func setupWithResources(t *testing.T, groupVersions []schema.GroupVersion, resources []schema.GroupVersionResource) (context.Context, clientset.Interface, *restclient.Config, framework.TearDownFunc) { | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| func setupWithResources(t *testing.T, groupVersions []schema.GroupVersion, resources []schema.GroupVersionResource) (context.Context, clientset.Interface /* TODO (pohly): return ktesting.TContext */, *restclient.Config, framework.TearDownFunc) { | ||||
| 	tCtx := ktesting.Init(t) | ||||
|  | ||||
| 	client, config, teardown := framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 	client, config, teardown := framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 		ModifyServerConfig: func(config *controlplane.Config) { | ||||
| 			if len(groupVersions) > 0 || len(resources) > 0 { | ||||
| 				resourceConfig := controlplane.DefaultAPIResourceConfigSource() | ||||
| @@ -91,11 +90,11 @@ func setupWithResources(t *testing.T, groupVersions []schema.GroupVersion, resou | ||||
| 	}) | ||||
|  | ||||
| 	newTeardown := func() { | ||||
| 		cancel() | ||||
| 		tCtx.Cancel("tearing down apiserver") | ||||
| 		teardown() | ||||
| 	} | ||||
|  | ||||
| 	return ctx, client, config, newTeardown | ||||
| 	return tCtx, client, config, newTeardown | ||||
| } | ||||
|  | ||||
| func verifyStatusCode(t *testing.T, transport http.RoundTripper, verb, URL, body string, expectedStatusCode int) { | ||||
| @@ -375,12 +374,10 @@ func TestListOptions(t *testing.T) { | ||||
|  | ||||
| 	for _, watchCacheEnabled := range []bool{true, false} { | ||||
| 		t.Run(fmt.Sprintf("watchCacheEnabled=%t", watchCacheEnabled), func(t *testing.T) { | ||||
| 			_, ctx := ktesting.NewTestContext(t) | ||||
| 			ctx, cancel := context.WithCancel(ctx) | ||||
| 			defer cancel() | ||||
| 			tCtx := ktesting.Init(t) | ||||
|  | ||||
| 			var storageTransport *storagebackend.TransportConfig | ||||
| 			clientSet, _, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 			clientSet, _, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 				ModifyServerRunOptions: func(opts *options.ServerRunOptions) { | ||||
| 					opts.Etcd.EnableWatchCache = watchCacheEnabled | ||||
| 					storageTransport = &opts.Etcd.StorageConfig.Transport | ||||
| @@ -397,7 +394,7 @@ func TestListOptions(t *testing.T) { | ||||
| 			for i := 0; i < 15; i++ { | ||||
| 				rs := newRS(ns.Name) | ||||
| 				rs.Name = fmt.Sprintf("test-%d", i) | ||||
| 				created, err := rsClient.Create(context.Background(), rs, metav1.CreateOptions{}) | ||||
| 				created, err := rsClient.Create(tCtx, rs, metav1.CreateOptions{}) | ||||
| 				if err != nil { | ||||
| 					t.Fatal(err) | ||||
| 				} | ||||
| @@ -407,7 +404,7 @@ func TestListOptions(t *testing.T) { | ||||
| 				// delete the first 5, and then compact them | ||||
| 				if i < 5 { | ||||
| 					var zero int64 | ||||
| 					if err := rsClient.Delete(context.Background(), rs.Name, metav1.DeleteOptions{GracePeriodSeconds: &zero}); err != nil { | ||||
| 					if err := rsClient.Delete(tCtx, rs.Name, metav1.DeleteOptions{GracePeriodSeconds: &zero}); err != nil { | ||||
| 						t.Fatal(err) | ||||
| 					} | ||||
| 					oldestUncompactedRv = created.ResourceVersion | ||||
| @@ -427,12 +424,12 @@ func TestListOptions(t *testing.T) { | ||||
| 			if err != nil { | ||||
| 				t.Fatal(err) | ||||
| 			} | ||||
| 			_, err = kvClient.Compact(context.Background(), int64(revision)) | ||||
| 			_, err = kvClient.Compact(tCtx, int64(revision)) | ||||
| 			if err != nil { | ||||
| 				t.Fatal(err) | ||||
| 			} | ||||
|  | ||||
| 			listObj, err := rsClient.List(context.Background(), metav1.ListOptions{ | ||||
| 			listObj, err := rsClient.List(tCtx, metav1.ListOptions{ | ||||
| 				Limit: 6, | ||||
| 			}) | ||||
| 			if err != nil { | ||||
| @@ -618,11 +615,9 @@ func TestListResourceVersion0(t *testing.T) { | ||||
|  | ||||
| 	for _, tc := range testcases { | ||||
| 		t.Run(tc.name, func(t *testing.T) { | ||||
| 			_, ctx := ktesting.NewTestContext(t) | ||||
| 			ctx, cancel := context.WithCancel(ctx) | ||||
| 			defer cancel() | ||||
| 			tCtx := ktesting.Init(t) | ||||
|  | ||||
| 			clientSet, _, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 			clientSet, _, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 				ModifyServerRunOptions: func(opts *options.ServerRunOptions) { | ||||
| 					opts.Etcd.EnableWatchCache = tc.watchCacheEnabled | ||||
| 				}, | ||||
| @@ -637,7 +632,7 @@ func TestListResourceVersion0(t *testing.T) { | ||||
| 			for i := 0; i < 10; i++ { | ||||
| 				rs := newRS(ns.Name) | ||||
| 				rs.Name = fmt.Sprintf("test-%d", i) | ||||
| 				if _, err := rsClient.Create(ctx, rs, metav1.CreateOptions{}); err != nil { | ||||
| 				if _, err := rsClient.Create(tCtx, rs, metav1.CreateOptions{}); err != nil { | ||||
| 					t.Fatal(err) | ||||
| 				} | ||||
| 			} | ||||
| @@ -645,7 +640,7 @@ func TestListResourceVersion0(t *testing.T) { | ||||
| 			if tc.watchCacheEnabled { | ||||
| 				// poll until the watch cache has the full list in memory | ||||
| 				err := wait.PollImmediate(time.Second, wait.ForeverTestTimeout, func() (bool, error) { | ||||
| 					list, err := clientSet.AppsV1().ReplicaSets(ns.Name).List(ctx, metav1.ListOptions{ResourceVersion: "0"}) | ||||
| 					list, err := clientSet.AppsV1().ReplicaSets(ns.Name).List(tCtx, metav1.ListOptions{ResourceVersion: "0"}) | ||||
| 					if err != nil { | ||||
| 						return false, err | ||||
| 					} | ||||
| @@ -657,12 +652,12 @@ func TestListResourceVersion0(t *testing.T) { | ||||
| 			} | ||||
|  | ||||
| 			pagerFn := func(opts metav1.ListOptions) (runtime.Object, error) { | ||||
| 				return rsClient.List(ctx, opts) | ||||
| 				return rsClient.List(tCtx, opts) | ||||
| 			} | ||||
|  | ||||
| 			p := pager.New(pager.SimplePageFunc(pagerFn)) | ||||
| 			p.PageSize = 3 | ||||
| 			listObj, _, err := p.List(ctx, metav1.ListOptions{ResourceVersion: "0"}) | ||||
| 			listObj, _, err := p.List(tCtx, metav1.ListOptions{ResourceVersion: "0"}) | ||||
| 			if err != nil { | ||||
| 				t.Fatalf("Unexpected list error: %v", err) | ||||
| 			} | ||||
|   | ||||
| @@ -136,9 +136,7 @@ func TestClientCARecreate(t *testing.T) { | ||||
| } | ||||
|  | ||||
| func testClientCA(t *testing.T, recreate bool) { | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
| 	tCtx := ktesting.Init(t) | ||||
|  | ||||
| 	frontProxyCA, err := newTestCAWithClient( | ||||
| 		pkix.Name{ | ||||
| @@ -175,7 +173,7 @@ func testClientCA(t *testing.T, recreate bool) { | ||||
| 	clientCAFilename := "" | ||||
| 	frontProxyCAFilename := "" | ||||
|  | ||||
| 	kubeClient, kubeconfig, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 	kubeClient, kubeconfig, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 		ModifyServerRunOptions: func(opts *options.ServerRunOptions) { | ||||
| 			opts.GenericServerRunOptions.MaxRequestBodyBytes = 1024 * 1024 | ||||
| 			clientCAFilename = opts.Authentication.ClientCert.ClientCA | ||||
| @@ -305,7 +303,7 @@ func testClientCA(t *testing.T, recreate bool) { | ||||
| 	} | ||||
|  | ||||
| 	// Call an endpoint to make sure we are authenticated | ||||
| 	_, err = testClient.CoreV1().Nodes().List(ctx, metav1.ListOptions{}) | ||||
| 	_, err = testClient.CoreV1().Nodes().List(tCtx, metav1.ListOptions{}) | ||||
| 	if err != nil { | ||||
| 		t.Error(err) | ||||
| 	} | ||||
| @@ -473,13 +471,11 @@ func TestServingCertRecreate(t *testing.T) { | ||||
| } | ||||
|  | ||||
| func testServingCert(t *testing.T, recreate bool) { | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
| 	tCtx := ktesting.Init(t) | ||||
|  | ||||
| 	var servingCertPath string | ||||
|  | ||||
| 	_, kubeconfig, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 	_, kubeconfig, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 		ModifyServerRunOptions: func(opts *options.ServerRunOptions) { | ||||
| 			opts.GenericServerRunOptions.MaxRequestBodyBytes = 1024 * 1024 | ||||
| 			servingCertPath = opts.SecureServing.ServerCert.CertDirectory | ||||
| @@ -518,11 +514,9 @@ func testServingCert(t *testing.T, recreate bool) { | ||||
| func TestSNICert(t *testing.T) { | ||||
| 	var servingCertPath string | ||||
|  | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
| 	tCtx := ktesting.Init(t) | ||||
|  | ||||
| 	_, kubeconfig, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 	_, kubeconfig, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 		ModifyServerRunOptions: func(opts *options.ServerRunOptions) { | ||||
| 			opts.GenericServerRunOptions.MaxRequestBodyBytes = 1024 * 1024 | ||||
| 			servingCertPath = opts.SecureServing.ServerCert.CertDirectory | ||||
|   | ||||
| @@ -48,10 +48,9 @@ const ( | ||||
| ) | ||||
|  | ||||
| func setup(t testing.TB, maxReadonlyRequestsInFlight, maxMutatingRequestsInFlight int) (context.Context, *rest.Config, framework.TearDownFunc) { | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	tCtx := ktesting.Init(t) | ||||
|  | ||||
| 	_, kubeConfig, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 	_, kubeConfig, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 		ModifyServerRunOptions: func(opts *options.ServerRunOptions) { | ||||
| 			// Ensure all clients are allowed to send requests. | ||||
| 			opts.Authorization.Modes = []string{"AlwaysAllow"} | ||||
| @@ -61,10 +60,10 @@ func setup(t testing.TB, maxReadonlyRequestsInFlight, maxMutatingRequestsInFligh | ||||
| 	}) | ||||
|  | ||||
| 	newTeardown := func() { | ||||
| 		cancel() | ||||
| 		tCtx.Cancel("tearing down apiserver") | ||||
| 		tearDownFn() | ||||
| 	} | ||||
| 	return ctx, kubeConfig, newTeardown | ||||
| 	return tCtx, kubeConfig, newTeardown | ||||
| } | ||||
|  | ||||
| func TestPriorityLevelIsolation(t *testing.T) { | ||||
|   | ||||
| @@ -145,11 +145,8 @@ func (d *noxuDelayingAuthorizer) Authorize(ctx context.Context, a authorizer.Att | ||||
| // Secondarily, this test also checks the observed seat utilizations against values derived from expecting that | ||||
| // the throughput observed by the client equals the execution throughput observed by the server. | ||||
| func TestConcurrencyIsolation(t *testing.T) { | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
|  | ||||
| 	_, kubeConfig, closeFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 	tCtx := ktesting.Init(t) | ||||
| 	_, kubeConfig, closeFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 		ModifyServerRunOptions: func(opts *options.ServerRunOptions) { | ||||
| 			// Ensure all clients are allowed to send requests. | ||||
| 			opts.Authorization.Modes = []string{"AlwaysAllow"} | ||||
| @@ -190,7 +187,7 @@ func TestConcurrencyIsolation(t *testing.T) { | ||||
| 	wg.Add(noxu1NumGoroutines) | ||||
| 	streamRequests(noxu1NumGoroutines, func() { | ||||
| 		start := time.Now() | ||||
| 		_, err := noxu1Client.CoreV1().Namespaces().Get(ctx, "default", metav1.GetOptions{}) | ||||
| 		_, err := noxu1Client.CoreV1().Namespaces().Get(tCtx, "default", metav1.GetOptions{}) | ||||
| 		duration := time.Since(start).Seconds() | ||||
| 		noxu1LatMeasure.update(duration) | ||||
| 		if err != nil { | ||||
| @@ -203,7 +200,7 @@ func TestConcurrencyIsolation(t *testing.T) { | ||||
| 	wg.Add(noxu2NumGoroutines) | ||||
| 	streamRequests(noxu2NumGoroutines, func() { | ||||
| 		start := time.Now() | ||||
| 		_, err := noxu2Client.CoreV1().Namespaces().Get(ctx, "default", metav1.GetOptions{}) | ||||
| 		_, err := noxu2Client.CoreV1().Namespaces().Get(tCtx, "default", metav1.GetOptions{}) | ||||
| 		duration := time.Since(start).Seconds() | ||||
| 		noxu2LatMeasure.update(duration) | ||||
| 		if err != nil { | ||||
|   | ||||
| @@ -17,7 +17,6 @@ limitations under the License. | ||||
| package apiserver | ||||
|  | ||||
| import ( | ||||
| 	"context" | ||||
| 	"fmt" | ||||
| 	"strings" | ||||
| 	"testing" | ||||
| @@ -33,11 +32,9 @@ import ( | ||||
|  | ||||
| // Tests that the apiserver limits the number of operations in a json patch. | ||||
| func TestMaxJSONPatchOperations(t *testing.T) { | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
| 	tCtx := ktesting.Init(t) | ||||
|  | ||||
| 	clientSet, _, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 	clientSet, _, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 		ModifyServerRunOptions: func(opts *options.ServerRunOptions) { | ||||
| 			opts.GenericServerRunOptions.MaxRequestBodyBytes = 1024 * 1024 | ||||
| 		}, | ||||
| @@ -55,13 +52,13 @@ func TestMaxJSONPatchOperations(t *testing.T) { | ||||
| 			Name: "test", | ||||
| 		}, | ||||
| 	} | ||||
| 	_, err := clientSet.CoreV1().Secrets("default").Create(ctx, secret, metav1.CreateOptions{}) | ||||
| 	_, err := clientSet.CoreV1().Secrets("default").Create(tCtx, secret, metav1.CreateOptions{}) | ||||
| 	if err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} | ||||
|  | ||||
| 	err = c.Patch(types.JSONPatchType).AbsPath(fmt.Sprintf("/api/v1/namespaces/default/secrets/test")). | ||||
| 		Body(hugePatch).Do(ctx).Error() | ||||
| 		Body(hugePatch).Do(tCtx).Error() | ||||
| 	if err == nil { | ||||
| 		t.Fatalf("unexpected no error") | ||||
| 	} | ||||
|   | ||||
| @@ -17,7 +17,6 @@ limitations under the License. | ||||
| package apiserver | ||||
|  | ||||
| import ( | ||||
| 	"context" | ||||
| 	"strings" | ||||
| 	"testing" | ||||
|  | ||||
| @@ -31,11 +30,8 @@ import ( | ||||
|  | ||||
| // Tests that the apiserver limits the resource size in write operations. | ||||
| func TestMaxResourceSize(t *testing.T) { | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
|  | ||||
| 	clientSet, _, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{}) | ||||
| 	tCtx := ktesting.Init(t) | ||||
| 	clientSet, _, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{}) | ||||
| 	defer tearDownFn() | ||||
|  | ||||
| 	hugeData := []byte(strings.Repeat("x", 3*1024*1024+1)) | ||||
| @@ -45,7 +41,7 @@ func TestMaxResourceSize(t *testing.T) { | ||||
| 	c := clientSet.CoreV1().RESTClient() | ||||
| 	t.Run("Create should limit the request body size", func(t *testing.T) { | ||||
| 		err := c.Post().AbsPath("/api/v1/namespaces/default/pods"). | ||||
| 			Body(hugeData).Do(ctx).Error() | ||||
| 			Body(hugeData).Do(tCtx).Error() | ||||
| 		if err == nil { | ||||
| 			t.Fatalf("unexpected no error") | ||||
| 		} | ||||
| @@ -61,14 +57,14 @@ func TestMaxResourceSize(t *testing.T) { | ||||
| 			Name: "test", | ||||
| 		}, | ||||
| 	} | ||||
| 	_, err := clientSet.CoreV1().Secrets("default").Create(ctx, secret, metav1.CreateOptions{}) | ||||
| 	_, err := clientSet.CoreV1().Secrets("default").Create(tCtx, secret, metav1.CreateOptions{}) | ||||
| 	if err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} | ||||
|  | ||||
| 	t.Run("Update should limit the request body size", func(t *testing.T) { | ||||
| 		err = c.Put().AbsPath("/api/v1/namespaces/default/secrets/test"). | ||||
| 			Body(hugeData).Do(ctx).Error() | ||||
| 			Body(hugeData).Do(tCtx).Error() | ||||
| 		if err == nil { | ||||
| 			t.Fatalf("unexpected no error") | ||||
| 		} | ||||
| @@ -79,7 +75,7 @@ func TestMaxResourceSize(t *testing.T) { | ||||
| 	}) | ||||
| 	t.Run("Patch should limit the request body size", func(t *testing.T) { | ||||
| 		err = c.Patch(types.JSONPatchType).AbsPath("/api/v1/namespaces/default/secrets/test"). | ||||
| 			Body(hugeData).Do(ctx).Error() | ||||
| 			Body(hugeData).Do(tCtx).Error() | ||||
| 		if err == nil { | ||||
| 			t.Fatalf("unexpected no error") | ||||
| 		} | ||||
| @@ -94,7 +90,7 @@ func TestMaxResourceSize(t *testing.T) { | ||||
| 		} | ||||
| 		patchBody := []byte(`[{"op":"add","path":"/foo","value":` + strings.Repeat("[", 3*1024*1024/2-100) + strings.Repeat("]", 3*1024*1024/2-100) + `}]`) | ||||
| 		err = rest.Patch(types.JSONPatchType).AbsPath("/api/v1/namespaces/default/secrets/test"). | ||||
| 			Body(patchBody).Do(ctx).Error() | ||||
| 			Body(patchBody).Do(tCtx).Error() | ||||
| 		if err != nil && !apierrors.IsBadRequest(err) { | ||||
| 			t.Errorf("expected success or bad request err, got %v", err) | ||||
| 		} | ||||
| @@ -105,7 +101,7 @@ func TestMaxResourceSize(t *testing.T) { | ||||
| 		} | ||||
| 		patchBody := []byte(`[{"op":"add","path":"/foo","value":0` + strings.Repeat(" ", 3*1024*1024-100) + `}]`) | ||||
| 		err = rest.Patch(types.JSONPatchType).AbsPath("/api/v1/namespaces/default/secrets/test"). | ||||
| 			Body(patchBody).Do(ctx).Error() | ||||
| 			Body(patchBody).Do(tCtx).Error() | ||||
| 		if err != nil { | ||||
| 			t.Errorf("unexpected error: %v", err) | ||||
| 		} | ||||
| @@ -116,7 +112,7 @@ func TestMaxResourceSize(t *testing.T) { | ||||
| 		} | ||||
| 		patchBody := []byte(`{"value":` + strings.Repeat("[", 3*1024*1024/2-100) + strings.Repeat("]", 3*1024*1024/2-100) + `}`) | ||||
| 		err = rest.Patch(types.MergePatchType).AbsPath("/api/v1/namespaces/default/secrets/test"). | ||||
| 			Body(patchBody).Do(ctx).Error() | ||||
| 			Body(patchBody).Do(tCtx).Error() | ||||
| 		if err != nil && !apierrors.IsBadRequest(err) { | ||||
| 			t.Errorf("expected success or bad request err, got %v", err) | ||||
| 		} | ||||
| @@ -127,7 +123,7 @@ func TestMaxResourceSize(t *testing.T) { | ||||
| 		} | ||||
| 		patchBody := []byte(`{"value":0` + strings.Repeat(" ", 3*1024*1024-100) + `}`) | ||||
| 		err = rest.Patch(types.MergePatchType).AbsPath("/api/v1/namespaces/default/secrets/test"). | ||||
| 			Body(patchBody).Do(ctx).Error() | ||||
| 			Body(patchBody).Do(tCtx).Error() | ||||
| 		if err != nil { | ||||
| 			t.Errorf("unexpected error: %v", err) | ||||
| 		} | ||||
| @@ -138,7 +134,7 @@ func TestMaxResourceSize(t *testing.T) { | ||||
| 		} | ||||
| 		patchBody := []byte(`{"value":` + strings.Repeat("[", 3*1024*1024/2-100) + strings.Repeat("]", 3*1024*1024/2-100) + `}`) | ||||
| 		err = rest.Patch(types.StrategicMergePatchType).AbsPath("/api/v1/namespaces/default/secrets/test"). | ||||
| 			Body(patchBody).Do(ctx).Error() | ||||
| 			Body(patchBody).Do(tCtx).Error() | ||||
| 		if err != nil && !apierrors.IsBadRequest(err) { | ||||
| 			t.Errorf("expected success or bad request err, got %v", err) | ||||
| 		} | ||||
| @@ -149,7 +145,7 @@ func TestMaxResourceSize(t *testing.T) { | ||||
| 		} | ||||
| 		patchBody := []byte(`{"value":0` + strings.Repeat(" ", 3*1024*1024-100) + `}`) | ||||
| 		err = rest.Patch(types.StrategicMergePatchType).AbsPath("/api/v1/namespaces/default/secrets/test"). | ||||
| 			Body(patchBody).Do(ctx).Error() | ||||
| 			Body(patchBody).Do(tCtx).Error() | ||||
| 		if err != nil { | ||||
| 			t.Errorf("unexpected error: %v", err) | ||||
| 		} | ||||
| @@ -160,7 +156,7 @@ func TestMaxResourceSize(t *testing.T) { | ||||
| 		} | ||||
| 		patchBody := []byte(`{"value":` + strings.Repeat("[", 3*1024*1024/2-100) + strings.Repeat("]", 3*1024*1024/2-100) + `}`) | ||||
| 		err = rest.Patch(types.ApplyPatchType).Param("fieldManager", "test").AbsPath("/api/v1/namespaces/default/secrets/test"). | ||||
| 			Body(patchBody).Do(ctx).Error() | ||||
| 			Body(patchBody).Do(tCtx).Error() | ||||
| 		if err != nil && !apierrors.IsBadRequest(err) { | ||||
| 			t.Errorf("expected success or bad request err, got %#v", err) | ||||
| 		} | ||||
| @@ -171,14 +167,14 @@ func TestMaxResourceSize(t *testing.T) { | ||||
| 		} | ||||
| 		patchBody := []byte(`{"apiVersion":"v1","kind":"Secret"` + strings.Repeat(" ", 3*1024*1024-100) + `}`) | ||||
| 		err = rest.Patch(types.ApplyPatchType).Param("fieldManager", "test").AbsPath("/api/v1/namespaces/default/secrets/test"). | ||||
| 			Body(patchBody).Do(ctx).Error() | ||||
| 			Body(patchBody).Do(tCtx).Error() | ||||
| 		if err != nil { | ||||
| 			t.Errorf("unexpected error: %v", err) | ||||
| 		} | ||||
| 	}) | ||||
| 	t.Run("Delete should limit the request body size", func(t *testing.T) { | ||||
| 		err = c.Delete().AbsPath("/api/v1/namespaces/default/secrets/test"). | ||||
| 			Body(hugeData).Do(ctx).Error() | ||||
| 			Body(hugeData).Do(tCtx).Error() | ||||
| 		if err == nil { | ||||
| 			t.Fatalf("unexpected no error") | ||||
| 		} | ||||
| @@ -202,7 +198,7 @@ values: ` + strings.Repeat("[", 3*1024*1024)) | ||||
| 			SetHeader("Content-Type", "application/yaml"). | ||||
| 			AbsPath("/api/v1/namespaces/default/configmaps"). | ||||
| 			Body(yamlBody). | ||||
| 			DoRaw(ctx) | ||||
| 			DoRaw(tCtx) | ||||
| 		if !apierrors.IsRequestEntityTooLargeError(err) { | ||||
| 			t.Errorf("expected too large error, got %v", err) | ||||
| 		} | ||||
| @@ -225,7 +221,7 @@ values: ` + strings.Repeat("[", 3*1024*1024/2-500) + strings.Repeat("]", 3*1024* | ||||
| 			SetHeader("Content-Type", "application/yaml"). | ||||
| 			AbsPath("/api/v1/namespaces/default/configmaps"). | ||||
| 			Body(yamlBody). | ||||
| 			DoRaw(ctx) | ||||
| 			DoRaw(tCtx) | ||||
| 		if !apierrors.IsBadRequest(err) { | ||||
| 			t.Errorf("expected bad request, got %v", err) | ||||
| 		} | ||||
| @@ -248,7 +244,7 @@ values: ` + strings.Repeat("[", 3*1024*1024-1000)) | ||||
| 			SetHeader("Content-Type", "application/yaml"). | ||||
| 			AbsPath("/api/v1/namespaces/default/configmaps"). | ||||
| 			Body(yamlBody). | ||||
| 			DoRaw(ctx) | ||||
| 			DoRaw(tCtx) | ||||
| 		if !apierrors.IsBadRequest(err) { | ||||
| 			t.Errorf("expected bad request, got %v", err) | ||||
| 		} | ||||
| @@ -269,7 +265,7 @@ values: ` + strings.Repeat("[", 3*1024*1024-1000)) | ||||
| 			SetHeader("Content-Type", "application/json"). | ||||
| 			AbsPath("/api/v1/namespaces/default/configmaps"). | ||||
| 			Body(jsonBody). | ||||
| 			DoRaw(ctx) | ||||
| 			DoRaw(tCtx) | ||||
| 		if !apierrors.IsRequestEntityTooLargeError(err) { | ||||
| 			t.Errorf("expected too large error, got %v", err) | ||||
| 		} | ||||
| @@ -293,7 +289,7 @@ values: ` + strings.Repeat("[", 3*1024*1024-1000)) | ||||
| 			SetHeader("Content-Type", "application/json"). | ||||
| 			AbsPath("/api/v1/namespaces/default/configmaps"). | ||||
| 			Body(jsonBody). | ||||
| 			DoRaw(ctx) | ||||
| 			DoRaw(tCtx) | ||||
| 		// TODO(liggitt): expect bad request on deep nesting, rather than success on dropped unknown field data | ||||
| 		if err != nil && !apierrors.IsBadRequest(err) { | ||||
| 			t.Errorf("expected bad request, got %v", err) | ||||
| @@ -318,7 +314,7 @@ values: ` + strings.Repeat("[", 3*1024*1024-1000)) | ||||
| 			SetHeader("Content-Type", "application/json"). | ||||
| 			AbsPath("/api/v1/namespaces/default/configmaps"). | ||||
| 			Body(jsonBody). | ||||
| 			DoRaw(ctx) | ||||
| 			DoRaw(tCtx) | ||||
| 		if !apierrors.IsBadRequest(err) { | ||||
| 			t.Errorf("expected bad request, got %v", err) | ||||
| 		} | ||||
|   | ||||
| @@ -17,7 +17,6 @@ limitations under the License. | ||||
| package openapi | ||||
|  | ||||
| import ( | ||||
| 	"context" | ||||
| 	"encoding/json" | ||||
| 	"net/http" | ||||
| 	"testing" | ||||
| @@ -56,10 +55,7 @@ func TestEnablingOpenAPIEnumTypes(t *testing.T) { | ||||
| 		}, | ||||
| 	} { | ||||
| 		t.Run(tc.name, func(t *testing.T) { | ||||
| 			_, ctx := ktesting.NewTestContext(t) | ||||
| 			ctx, cancel := context.WithCancel(ctx) | ||||
| 			defer cancel() | ||||
|  | ||||
| 			tCtx := ktesting.Init(t) | ||||
| 			defer featuregatetesting.SetFeatureGateDuringTest(t, utilfeature.DefaultFeatureGate, features.OpenAPIEnums, tc.featureEnabled)() | ||||
|  | ||||
| 			getDefinitionsFn := openapi.GetOpenAPIDefinitionsWithoutDisabledFeatures(func(ref common.ReferenceCallback) map[string]common.OpenAPIDefinition { | ||||
| @@ -79,7 +75,7 @@ func TestEnablingOpenAPIEnumTypes(t *testing.T) { | ||||
| 				return defs | ||||
| 			}) | ||||
|  | ||||
| 			_, kubeConfig, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 			_, kubeConfig, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 				ModifyServerConfig: func(config *controlplane.Config) { | ||||
| 					config.GenericConfig.OpenAPIConfig = framework.DefaultOpenAPIConfig() | ||||
| 					config.GenericConfig.OpenAPIConfig.GetDefinitions = getDefinitionsFn | ||||
|   | ||||
| @@ -43,11 +43,8 @@ import ( | ||||
| ) | ||||
|  | ||||
| func TestOpenAPIV3SpecRoundTrip(t *testing.T) { | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
|  | ||||
| 	_, kubeConfig, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{}) | ||||
| 	tCtx := ktesting.Init(t) | ||||
| 	_, kubeConfig, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{}) | ||||
| 	defer tearDownFn() | ||||
|  | ||||
| 	paths := []string{ | ||||
| @@ -189,11 +186,8 @@ func TestOpenAPIV3ProtoRoundtrip(t *testing.T) { | ||||
| 	// See https://github.com/kubernetes/kubernetes/issues/106387 for more details | ||||
| 	t.Skip("Skipping OpenAPI V3 Proto roundtrip test") | ||||
|  | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
|  | ||||
| 	_, kubeConfig, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{}) | ||||
| 	tCtx := ktesting.Init(t) | ||||
| 	_, kubeConfig, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{}) | ||||
| 	defer tearDownFn() | ||||
|  | ||||
| 	rt, err := restclient.TransportFor(kubeConfig) | ||||
|   | ||||
| @@ -78,11 +78,8 @@ Bgqc+dJN9xS9Ah5gLiGQJ6C4niUA11piCpvMsy+j/LQ1Erx47KMar5fuMXYk7iPq | ||||
| -----END CERTIFICATE----- | ||||
| `)) | ||||
|  | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
|  | ||||
| 	clientSet, _, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 	tCtx := ktesting.Init(t) | ||||
| 	clientSet, _, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 		ModifyServerRunOptions: func(opts *options.ServerRunOptions) { | ||||
| 			opts.GenericServerRunOptions.MaxRequestBodyBytes = 1024 * 1024 | ||||
| 			// I have no idea what this cert is, but it doesn't matter, we just want something that always fails validation | ||||
| @@ -97,7 +94,7 @@ Bgqc+dJN9xS9Ah5gLiGQJ6C4niUA11piCpvMsy+j/LQ1Erx47KMar5fuMXYk7iPq | ||||
| 	})) | ||||
| 	defer fakeKubeletServer.Close() | ||||
|  | ||||
| 	pod := prepareFakeNodeAndPod(ctx, t, clientSet, fakeKubeletServer) | ||||
| 	pod := prepareFakeNodeAndPod(tCtx, t, clientSet, fakeKubeletServer) | ||||
|  | ||||
| 	insecureResult := clientSet.CoreV1().Pods("ns").GetLogs(pod.Name, &corev1.PodLogOptions{InsecureSkipTLSVerifyBackend: true}).Do(context.TODO()) | ||||
| 	if err := insecureResult.Error(); err != nil { | ||||
| @@ -109,7 +106,7 @@ Bgqc+dJN9xS9Ah5gLiGQJ6C4niUA11piCpvMsy+j/LQ1Erx47KMar5fuMXYk7iPq | ||||
| 		t.Fatal(insecureStatusCode) | ||||
| 	} | ||||
|  | ||||
| 	secureResult := clientSet.CoreV1().Pods("ns").GetLogs(pod.Name, &corev1.PodLogOptions{}).Do(ctx) | ||||
| 	secureResult := clientSet.CoreV1().Pods("ns").GetLogs(pod.Name, &corev1.PodLogOptions{}).Do(tCtx) | ||||
| 	if err := secureResult.Error(); err == nil || !strings.Contains(err.Error(), "x509: certificate signed by unknown authority") { | ||||
| 		t.Fatal(err) | ||||
| 	} | ||||
|   | ||||
| @@ -85,11 +85,8 @@ func multiEtcdSetup(ctx context.Context, t *testing.T) (clientset.Interface, fra | ||||
| } | ||||
|  | ||||
| func TestWatchCacheUpdatedByEtcd(t *testing.T) { | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
|  | ||||
| 	c, closeFn := multiEtcdSetup(ctx, t) | ||||
| 	tCtx := ktesting.Init(t) | ||||
| 	c, closeFn := multiEtcdSetup(tCtx, t) | ||||
| 	defer closeFn() | ||||
|  | ||||
| 	makeConfigMap := func(name string) *v1.ConfigMap { | ||||
| @@ -102,11 +99,11 @@ func TestWatchCacheUpdatedByEtcd(t *testing.T) { | ||||
| 		return &v1.Event{ObjectMeta: metav1.ObjectMeta{Name: name}} | ||||
| 	} | ||||
|  | ||||
| 	cm, err := c.CoreV1().ConfigMaps("default").Create(ctx, makeConfigMap("name"), metav1.CreateOptions{}) | ||||
| 	cm, err := c.CoreV1().ConfigMaps("default").Create(tCtx, makeConfigMap("name"), metav1.CreateOptions{}) | ||||
| 	if err != nil { | ||||
| 		t.Errorf("Couldn't create configmap: %v", err) | ||||
| 	} | ||||
| 	ev, err := c.CoreV1().Events("default").Create(ctx, makeEvent("name"), metav1.CreateOptions{}) | ||||
| 	ev, err := c.CoreV1().Events("default").Create(tCtx, makeEvent("name"), metav1.CreateOptions{}) | ||||
| 	if err != nil { | ||||
| 		t.Errorf("Couldn't create event: %v", err) | ||||
| 	} | ||||
| @@ -120,7 +117,7 @@ func TestWatchCacheUpdatedByEtcd(t *testing.T) { | ||||
| 	// resources (being the last updates). | ||||
| 	t.Logf("Waiting for configmaps watchcache synced to %s", cm.ResourceVersion) | ||||
| 	if err := wait.Poll(100*time.Millisecond, wait.ForeverTestTimeout, func() (bool, error) { | ||||
| 		res, err := c.CoreV1().ConfigMaps("default").List(ctx, listOptions) | ||||
| 		res, err := c.CoreV1().ConfigMaps("default").List(tCtx, listOptions) | ||||
| 		if err != nil { | ||||
| 			return false, nil | ||||
| 		} | ||||
| @@ -130,7 +127,7 @@ func TestWatchCacheUpdatedByEtcd(t *testing.T) { | ||||
| 	} | ||||
| 	t.Logf("Waiting for events watchcache synced to %s", ev.ResourceVersion) | ||||
| 	if err := wait.Poll(100*time.Millisecond, wait.ForeverTestTimeout, func() (bool, error) { | ||||
| 		res, err := c.CoreV1().Events("default").List(ctx, listOptions) | ||||
| 		res, err := c.CoreV1().Events("default").List(tCtx, listOptions) | ||||
| 		if err != nil { | ||||
| 			return false, nil | ||||
| 		} | ||||
| @@ -141,14 +138,14 @@ func TestWatchCacheUpdatedByEtcd(t *testing.T) { | ||||
|  | ||||
| 	// Create a secret, that is stored in the same etcd as configmap, but | ||||
| 	// different than events. | ||||
| 	se, err := c.CoreV1().Secrets("default").Create(ctx, makeSecret("name"), metav1.CreateOptions{}) | ||||
| 	se, err := c.CoreV1().Secrets("default").Create(tCtx, makeSecret("name"), metav1.CreateOptions{}) | ||||
| 	if err != nil { | ||||
| 		t.Errorf("Couldn't create secret: %v", err) | ||||
| 	} | ||||
|  | ||||
| 	t.Logf("Waiting for configmaps watchcache synced to %s", se.ResourceVersion) | ||||
| 	if err := wait.Poll(100*time.Millisecond, wait.ForeverTestTimeout, func() (bool, error) { | ||||
| 		res, err := c.CoreV1().ConfigMaps("default").List(ctx, listOptions) | ||||
| 		res, err := c.CoreV1().ConfigMaps("default").List(tCtx, listOptions) | ||||
| 		if err != nil { | ||||
| 			return false, nil | ||||
| 		} | ||||
| @@ -158,7 +155,7 @@ func TestWatchCacheUpdatedByEtcd(t *testing.T) { | ||||
| 	} | ||||
| 	t.Logf("Waiting for events watchcache NOT synced to %s", se.ResourceVersion) | ||||
| 	if err := wait.Poll(100*time.Millisecond, 5*time.Second, func() (bool, error) { | ||||
| 		res, err := c.CoreV1().Events("default").List(ctx, listOptions) | ||||
| 		res, err := c.CoreV1().Events("default").List(tCtx, listOptions) | ||||
| 		if err != nil { | ||||
| 			return false, nil | ||||
| 		} | ||||
| @@ -169,11 +166,8 @@ func TestWatchCacheUpdatedByEtcd(t *testing.T) { | ||||
| } | ||||
|  | ||||
| func BenchmarkListFromWatchCache(b *testing.B) { | ||||
| 	_, ctx := ktesting.NewTestContext(b) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
|  | ||||
| 	c, _, tearDownFn := framework.StartTestServer(ctx, b, framework.TestServerSetup{ | ||||
| 	tCtx := ktesting.Init(b) | ||||
| 	c, _, tearDownFn := framework.StartTestServer(tCtx, b, framework.TestServerSetup{ | ||||
| 		ModifyServerConfig: func(config *controlplane.Config) { | ||||
| 			// Switch off endpoints reconciler to avoid unnecessary operations. | ||||
| 			config.ExtraConfig.EndpointReconcilerType = reconcilers.NoneEndpointReconcilerType | ||||
| @@ -194,7 +188,7 @@ func BenchmarkListFromWatchCache(b *testing.B) { | ||||
| 			ns := &v1.Namespace{ | ||||
| 				ObjectMeta: metav1.ObjectMeta{Name: fmt.Sprintf("namespace-%d", index)}, | ||||
| 			} | ||||
| 			ns, err := c.CoreV1().Namespaces().Create(ctx, ns, metav1.CreateOptions{}) | ||||
| 			ns, err := c.CoreV1().Namespaces().Create(tCtx, ns, metav1.CreateOptions{}) | ||||
| 			if err != nil { | ||||
| 				errCh <- err | ||||
| 				return | ||||
| @@ -206,7 +200,7 @@ func BenchmarkListFromWatchCache(b *testing.B) { | ||||
| 						Name: fmt.Sprintf("secret-%d", j), | ||||
| 					}, | ||||
| 				} | ||||
| 				_, err := c.CoreV1().Secrets(ns.Name).Create(ctx, secret, metav1.CreateOptions{}) | ||||
| 				_, err := c.CoreV1().Secrets(ns.Name).Create(tCtx, secret, metav1.CreateOptions{}) | ||||
| 				if err != nil { | ||||
| 					errCh <- err | ||||
| 					return | ||||
| @@ -227,7 +221,7 @@ func BenchmarkListFromWatchCache(b *testing.B) { | ||||
| 		ResourceVersion: "0", | ||||
| 	} | ||||
| 	for i := 0; i < b.N; i++ { | ||||
| 		secrets, err := c.CoreV1().Secrets("").List(ctx, opts) | ||||
| 		secrets, err := c.CoreV1().Secrets("").List(tCtx, opts) | ||||
| 		if err != nil { | ||||
| 			b.Errorf("failed to list secrets: %v", err) | ||||
| 		} | ||||
|   | ||||
| @@ -58,11 +58,8 @@ func alwaysAlice(req *http.Request) (*authenticator.Response, bool, error) { | ||||
| } | ||||
|  | ||||
| func TestSubjectAccessReview(t *testing.T) { | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
|  | ||||
| 	clientset, _, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 	tCtx := ktesting.Init(t) | ||||
| 	clientset, _, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 		ModifyServerConfig: func(config *controlplane.Config) { | ||||
| 			// Unset BearerToken to disable BearerToken authenticator. | ||||
| 			config.GenericConfig.LoopbackClientConfig.BearerToken = "" | ||||
| @@ -132,7 +129,7 @@ func TestSubjectAccessReview(t *testing.T) { | ||||
| 	} | ||||
|  | ||||
| 	for _, test := range tests { | ||||
| 		response, err := clientset.AuthorizationV1().SubjectAccessReviews().Create(ctx, test.sar, metav1.CreateOptions{}) | ||||
| 		response, err := clientset.AuthorizationV1().SubjectAccessReviews().Create(tCtx, test.sar, metav1.CreateOptions{}) | ||||
| 		switch { | ||||
| 		case err == nil && len(test.expectedError) == 0: | ||||
|  | ||||
| @@ -154,9 +151,7 @@ func TestSubjectAccessReview(t *testing.T) { | ||||
| } | ||||
|  | ||||
| func TestSelfSubjectAccessReview(t *testing.T) { | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
| 	tCtx := ktesting.Init(t) | ||||
|  | ||||
| 	var mutex sync.Mutex | ||||
|  | ||||
| @@ -174,7 +169,7 @@ func TestSelfSubjectAccessReview(t *testing.T) { | ||||
| 		}, true, nil | ||||
| 	} | ||||
|  | ||||
| 	clientset, _, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 	clientset, _, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 		ModifyServerConfig: func(config *controlplane.Config) { | ||||
| 			// Unset BearerToken to disable BearerToken authenticator. | ||||
| 			config.GenericConfig.LoopbackClientConfig.BearerToken = "" | ||||
| @@ -235,7 +230,7 @@ func TestSelfSubjectAccessReview(t *testing.T) { | ||||
| 		username = test.username | ||||
| 		mutex.Unlock() | ||||
|  | ||||
| 		response, err := clientset.AuthorizationV1().SelfSubjectAccessReviews().Create(ctx, test.sar, metav1.CreateOptions{}) | ||||
| 		response, err := clientset.AuthorizationV1().SelfSubjectAccessReviews().Create(tCtx, test.sar, metav1.CreateOptions{}) | ||||
| 		switch { | ||||
| 		case err == nil && len(test.expectedError) == 0: | ||||
|  | ||||
| @@ -257,11 +252,8 @@ func TestSelfSubjectAccessReview(t *testing.T) { | ||||
| } | ||||
|  | ||||
| func TestLocalSubjectAccessReview(t *testing.T) { | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
|  | ||||
| 	clientset, _, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 	tCtx := ktesting.Init(t) | ||||
| 	clientset, _, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 		ModifyServerConfig: func(config *controlplane.Config) { | ||||
| 			// Unset BearerToken to disable BearerToken authenticator. | ||||
| 			config.GenericConfig.LoopbackClientConfig.BearerToken = "" | ||||
| @@ -359,7 +351,7 @@ func TestLocalSubjectAccessReview(t *testing.T) { | ||||
| 	} | ||||
|  | ||||
| 	for _, test := range tests { | ||||
| 		response, err := clientset.AuthorizationV1().LocalSubjectAccessReviews(test.namespace).Create(ctx, test.sar, metav1.CreateOptions{}) | ||||
| 		response, err := clientset.AuthorizationV1().LocalSubjectAccessReviews(test.namespace).Create(tCtx, test.sar, metav1.CreateOptions{}) | ||||
| 		switch { | ||||
| 		case err == nil && len(test.expectedError) == 0: | ||||
|  | ||||
|   | ||||
| @@ -454,11 +454,8 @@ func getTestRequests(namespace string) []testRequest { | ||||
| // | ||||
| // TODO(etune): write a fuzz test of the REST API. | ||||
| func TestAuthModeAlwaysAllow(t *testing.T) { | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
|  | ||||
| 	kubeClient, kubeConfig, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 	tCtx := ktesting.Init(t) | ||||
| 	kubeClient, kubeConfig, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 		ModifyServerRunOptions: func(opts *options.ServerRunOptions) { | ||||
| 			// Disable ServiceAccount admission plugin as we don't have serviceaccount controller running. | ||||
| 			opts.Admission.GenericAdmission.DisablePlugins = []string{"ServiceAccount"} | ||||
| @@ -562,11 +559,8 @@ func getPreviousResourceVersionKey(url, id string) string { | ||||
| } | ||||
|  | ||||
| func TestAuthModeAlwaysDeny(t *testing.T) { | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
|  | ||||
| 	kubeClient, kubeConfig, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 	tCtx := ktesting.Init(t) | ||||
| 	kubeClient, kubeConfig, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 		ModifyServerRunOptions: func(opts *options.ServerRunOptions) { | ||||
| 			// Disable ServiceAccount admission plugin as we don't have serviceaccount controller running. | ||||
| 			opts.Admission.GenericAdmission.DisablePlugins = []string{"ServiceAccount"} | ||||
| @@ -610,11 +604,8 @@ func TestAuthModeAlwaysDeny(t *testing.T) { | ||||
| // TestAliceNotForbiddenOrUnauthorized tests a user who is known to | ||||
| // the authentication system and authorized to do any actions. | ||||
| func TestAliceNotForbiddenOrUnauthorized(t *testing.T) { | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
|  | ||||
| 	kubeClient, kubeConfig, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 	tCtx := ktesting.Init(t) | ||||
| 	kubeClient, kubeConfig, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 		ModifyServerRunOptions: func(opts *options.ServerRunOptions) { | ||||
| 			// Disable ServiceAccount admission plugin as we don't have serviceaccount controller running. | ||||
| 			opts.Admission.GenericAdmission.DisablePlugins = []string{"ServiceAccount"} | ||||
| @@ -690,11 +681,8 @@ func TestAliceNotForbiddenOrUnauthorized(t *testing.T) { | ||||
| // the authentication system but not authorized to do any actions | ||||
| // should receive "Forbidden". | ||||
| func TestBobIsForbidden(t *testing.T) { | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
|  | ||||
| 	kubeClient, kubeConfig, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 	tCtx := ktesting.Init(t) | ||||
| 	kubeClient, kubeConfig, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 		ModifyServerRunOptions: func(opts *options.ServerRunOptions) { | ||||
| 			// Disable ServiceAccount admission plugin as we don't have serviceaccount controller running. | ||||
| 			opts.Admission.GenericAdmission.DisablePlugins = []string{"ServiceAccount"} | ||||
| @@ -743,11 +731,8 @@ func TestBobIsForbidden(t *testing.T) { | ||||
| // An authorization module is installed in this scenario for integration | ||||
| // test purposes, but requests aren't expected to reach it. | ||||
| func TestUnknownUserIsUnauthorized(t *testing.T) { | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
|  | ||||
| 	kubeClient, kubeConfig, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 	tCtx := ktesting.Init(t) | ||||
| 	kubeClient, kubeConfig, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 		ModifyServerRunOptions: func(opts *options.ServerRunOptions) { | ||||
| 			// Disable ServiceAccount admission plugin as we don't have serviceaccount controller running. | ||||
| 			opts.Admission.GenericAdmission.DisablePlugins = []string{"ServiceAccount"} | ||||
| @@ -819,11 +804,8 @@ func (impersonateAuthorizer) Authorize(ctx context.Context, a authorizer.Attribu | ||||
| } | ||||
|  | ||||
| func TestImpersonateIsForbidden(t *testing.T) { | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
|  | ||||
| 	kubeClient, kubeConfig, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 	tCtx := ktesting.Init(t) | ||||
| 	kubeClient, kubeConfig, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 		ModifyServerRunOptions: func(opts *options.ServerRunOptions) { | ||||
| 			// Disable ServiceAccount admission plugin as we don't have serviceaccount controller running. | ||||
| 			opts.Admission.GenericAdmission.DisablePlugins = []string{"ServiceAccount"} | ||||
| @@ -1125,13 +1107,11 @@ func (a *trackingAuthorizer) Authorize(ctx context.Context, attributes authorize | ||||
|  | ||||
| // TestAuthorizationAttributeDetermination tests that authorization attributes are built correctly | ||||
| func TestAuthorizationAttributeDetermination(t *testing.T) { | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
| 	tCtx := ktesting.Init(t) | ||||
|  | ||||
| 	trackingAuthorizer := &trackingAuthorizer{} | ||||
|  | ||||
| 	kubeClient, kubeConfig, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 	kubeClient, kubeConfig, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 		ModifyServerRunOptions: func(opts *options.ServerRunOptions) { | ||||
| 			// Disable ServiceAccount admission plugin as we don't have serviceaccount controller running. | ||||
| 			opts.Admission.GenericAdmission.DisablePlugins = []string{"ServiceAccount"} | ||||
| @@ -1201,11 +1181,9 @@ func TestAuthorizationAttributeDetermination(t *testing.T) { | ||||
| // TestNamespaceAuthorization tests that authorization can be controlled | ||||
| // by namespace. | ||||
| func TestNamespaceAuthorization(t *testing.T) { | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
| 	tCtx := ktesting.Init(t) | ||||
|  | ||||
| 	kubeClient, kubeConfig, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 	kubeClient, kubeConfig, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 		ModifyServerRunOptions: func(opts *options.ServerRunOptions) { | ||||
| 			// Disable ServiceAccount admission plugin as we don't have serviceaccount controller running. | ||||
| 			opts.Admission.GenericAdmission.DisablePlugins = []string{"ServiceAccount"} | ||||
| @@ -1306,11 +1284,9 @@ func TestNamespaceAuthorization(t *testing.T) { | ||||
| // TestKindAuthorization tests that authorization can be controlled | ||||
| // by namespace. | ||||
| func TestKindAuthorization(t *testing.T) { | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
| 	tCtx := ktesting.Init(t) | ||||
|  | ||||
| 	kubeClient, kubeConfig, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 	kubeClient, kubeConfig, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 		ModifyServerRunOptions: func(opts *options.ServerRunOptions) { | ||||
| 			// Disable ServiceAccount admission plugin as we don't have serviceaccount controller running. | ||||
| 			opts.Admission.GenericAdmission.DisablePlugins = []string{"ServiceAccount"} | ||||
| @@ -1393,11 +1369,8 @@ func TestKindAuthorization(t *testing.T) { | ||||
| // TestReadOnlyAuthorization tests that authorization can be controlled | ||||
| // by namespace. | ||||
| func TestReadOnlyAuthorization(t *testing.T) { | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
|  | ||||
| 	kubeClient, kubeConfig, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 	tCtx := ktesting.Init(t) | ||||
| 	kubeClient, kubeConfig, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 		ModifyServerRunOptions: func(opts *options.ServerRunOptions) { | ||||
| 			// Disable ServiceAccount admission plugin as we don't have serviceaccount controller running. | ||||
| 			opts.Admission.GenericAdmission.DisablePlugins = []string{"ServiceAccount"} | ||||
| @@ -1461,10 +1434,7 @@ func TestWebhookTokenAuthenticatorCustomDial(t *testing.T) { | ||||
| } | ||||
|  | ||||
| func testWebhookTokenAuthenticator(customDialer bool, t *testing.T) { | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
|  | ||||
| 	tCtx := ktesting.Init(t) | ||||
| 	authServer := newTestWebhookTokenAuthServer() | ||||
| 	defer authServer.Close() | ||||
| 	var authenticator authenticator.Request | ||||
| @@ -1480,7 +1450,7 @@ func testWebhookTokenAuthenticator(customDialer bool, t *testing.T) { | ||||
| 		t.Fatalf("error starting webhook token authenticator server: %v", err) | ||||
| 	} | ||||
|  | ||||
| 	kubeClient, kubeConfig, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 	kubeClient, kubeConfig, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 		ModifyServerRunOptions: func(opts *options.ServerRunOptions) { | ||||
| 			// Disable ServiceAccount admission plugin as we don't have serviceaccount controller running. | ||||
| 			opts.Admission.GenericAdmission.DisablePlugins = []string{"ServiceAccount"} | ||||
|   | ||||
| @@ -18,7 +18,6 @@ package auth | ||||
|  | ||||
| import ( | ||||
| 	"bytes" | ||||
| 	"context" | ||||
| 	"fmt" | ||||
| 	"io" | ||||
| 	"net/http" | ||||
| @@ -121,13 +120,10 @@ func TestBootstrapTokenAuth(t *testing.T) { | ||||
| 	} | ||||
| 	for _, test := range tests { | ||||
| 		t.Run(test.name, func(t *testing.T) { | ||||
| 			_, ctx := ktesting.NewTestContext(t) | ||||
| 			ctx, cancel := context.WithCancel(ctx) | ||||
| 			defer cancel() | ||||
|  | ||||
| 			tCtx := ktesting.Init(t) | ||||
| 			authenticator := group.NewAuthenticatedGroupAdder(bearertoken.New(bootstrap.NewTokenAuthenticator(bootstrapSecrets{test.secret}))) | ||||
|  | ||||
| 			kubeClient, kubeConfig, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 			kubeClient, kubeConfig, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 				ModifyServerRunOptions: func(opts *options.ServerRunOptions) { | ||||
| 					opts.Authorization.Modes = []string{"AlwaysAllow"} | ||||
| 				}, | ||||
|   | ||||
| @@ -54,11 +54,8 @@ func TestDynamicClientBuilder(t *testing.T) { | ||||
| 		t.Fatalf("parse duration failed: %v", err) | ||||
| 	} | ||||
|  | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
|  | ||||
| 	baseClient, baseConfig, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 	tCtx := ktesting.Init(t) | ||||
| 	baseClient, baseConfig, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 		ModifyServerRunOptions: func(opts *options.ServerRunOptions) { | ||||
| 			opts.ServiceAccountSigningKeyFile = tmpfile.Name() | ||||
| 			opts.ServiceAccountTokenMaxExpiration = maxExpirationDuration | ||||
| @@ -102,7 +99,7 @@ func TestDynamicClientBuilder(t *testing.T) { | ||||
|  | ||||
| 	// We want to trigger token rotation here by deleting service account | ||||
| 	// the dynamic client was using. | ||||
| 	if err = dymClient.CoreV1().ServiceAccounts(ns).Delete(ctx, saName, metav1.DeleteOptions{}); err != nil { | ||||
| 	if err = dymClient.CoreV1().ServiceAccounts(ns).Delete(tCtx, saName, metav1.DeleteOptions{}); err != nil { | ||||
| 		t.Fatalf("delete service account %s failed: %v", saName, err) | ||||
| 	} | ||||
| 	time.Sleep(time.Second * 10) | ||||
|   | ||||
| @@ -26,7 +26,6 @@ import ( | ||||
| 	"reflect" | ||||
| 	"strings" | ||||
| 	"testing" | ||||
| 	"time" | ||||
|  | ||||
| 	rbacapi "k8s.io/api/rbac/v1" | ||||
| 	metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" | ||||
| @@ -538,10 +537,7 @@ func TestRBAC(t *testing.T) { | ||||
| 				"user-with-no-permissions":         {Name: "user-with-no-permissions"}, | ||||
| 			}))) | ||||
|  | ||||
| 			_, ctx := ktesting.NewTestContext(t) | ||||
| 			ctx, cancel := context.WithCancel(ctx) | ||||
| 			defer cancel() | ||||
|  | ||||
| 			tCtx := ktesting.Init(t) | ||||
| 			var tearDownAuthorizerFn func() | ||||
| 			defer func() { | ||||
| 				if tearDownAuthorizerFn != nil { | ||||
| @@ -549,7 +545,7 @@ func TestRBAC(t *testing.T) { | ||||
| 				} | ||||
| 			}() | ||||
|  | ||||
| 			_, kubeConfig, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 			_, kubeConfig, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 				ModifyServerRunOptions: func(opts *options.ServerRunOptions) { | ||||
| 					// Disable ServiceAccount admission plugin as we don't have serviceaccount controller running. | ||||
| 					// Also disable namespace lifecycle to workaroung the test limitation that first creates | ||||
| @@ -675,23 +671,20 @@ func TestRBAC(t *testing.T) { | ||||
| } | ||||
|  | ||||
| func TestBootstrapping(t *testing.T) { | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithTimeout(ctx, 30*time.Second) | ||||
| 	defer cancel() | ||||
|  | ||||
| 	clientset, _, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 	tCtx := ktesting.Init(t) | ||||
| 	clientset, _, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 		ModifyServerRunOptions: func(opts *options.ServerRunOptions) { | ||||
| 			opts.Authorization.Modes = []string{"RBAC"} | ||||
| 		}, | ||||
| 	}) | ||||
| 	defer tearDownFn() | ||||
|  | ||||
| 	watcher, err := clientset.RbacV1().ClusterRoles().Watch(ctx, metav1.ListOptions{ResourceVersion: "0"}) | ||||
| 	watcher, err := clientset.RbacV1().ClusterRoles().Watch(tCtx, metav1.ListOptions{ResourceVersion: "0"}) | ||||
| 	if err != nil { | ||||
| 		t.Fatalf("unexpected error: %v", err) | ||||
| 	} | ||||
|  | ||||
| 	_, err = watchtools.UntilWithoutRetry(ctx, watcher, func(event watch.Event) (bool, error) { | ||||
| 	_, err = watchtools.UntilWithoutRetry(tCtx, watcher, func(event watch.Event) (bool, error) { | ||||
| 		if event.Type != watch.Added { | ||||
| 			return false, nil | ||||
| 		} | ||||
| @@ -701,7 +694,7 @@ func TestBootstrapping(t *testing.T) { | ||||
| 		t.Fatalf("unexpected error: %v", err) | ||||
| 	} | ||||
|  | ||||
| 	clusterRoles, err := clientset.RbacV1().ClusterRoles().List(ctx, metav1.ListOptions{}) | ||||
| 	clusterRoles, err := clientset.RbacV1().ClusterRoles().List(tCtx, metav1.ListOptions{}) | ||||
| 	if err != nil { | ||||
| 		t.Fatalf("unexpected error: %v", err) | ||||
| 	} | ||||
| @@ -717,7 +710,7 @@ func TestBootstrapping(t *testing.T) { | ||||
|  | ||||
| 	t.Errorf("missing cluster-admin: %v", clusterRoles) | ||||
|  | ||||
| 	healthBytes, err := clientset.Discovery().RESTClient().Get().AbsPath("/healthz/poststarthook/rbac/bootstrap-roles").DoRaw(ctx) | ||||
| 	healthBytes, err := clientset.Discovery().RESTClient().Get().AbsPath("/healthz/poststarthook/rbac/bootstrap-roles").DoRaw(tCtx) | ||||
| 	if err != nil { | ||||
| 		t.Error(err) | ||||
| 	} | ||||
| @@ -736,11 +729,8 @@ func TestDiscoveryUpgradeBootstrapping(t *testing.T) { | ||||
|  | ||||
| 	etcdConfig := framework.SharedEtcd() | ||||
|  | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
|  | ||||
| 	client, _, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 	tCtx := ktesting.Init(t) | ||||
| 	client, _, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 		ModifyServerRunOptions: func(opts *options.ServerRunOptions) { | ||||
| 			// Ensure we're using the same etcd across apiserver restarts. | ||||
| 			opts.Etcd.StorageConfig = *etcdConfig | ||||
| @@ -751,7 +741,7 @@ func TestDiscoveryUpgradeBootstrapping(t *testing.T) { | ||||
| 	// Modify the default RBAC discovery ClusterRoleBidnings to look more like the defaults that | ||||
| 	// existed prior to v1.14, but with user modifications. | ||||
| 	t.Logf("Modifying default `system:discovery` ClusterRoleBinding") | ||||
| 	discRoleBinding, err := client.RbacV1().ClusterRoleBindings().Get(ctx, "system:discovery", metav1.GetOptions{}) | ||||
| 	discRoleBinding, err := client.RbacV1().ClusterRoleBindings().Get(tCtx, "system:discovery", metav1.GetOptions{}) | ||||
| 	if err != nil { | ||||
| 		t.Fatalf("Failed to get `system:discovery` ClusterRoleBinding: %v", err) | ||||
| 	} | ||||
| @@ -764,21 +754,21 @@ func TestDiscoveryUpgradeBootstrapping(t *testing.T) { | ||||
| 			APIGroup: "rbac.authorization.k8s.io", | ||||
| 		}, | ||||
| 	} | ||||
| 	if discRoleBinding, err = client.RbacV1().ClusterRoleBindings().Update(ctx, discRoleBinding, metav1.UpdateOptions{}); err != nil { | ||||
| 	if discRoleBinding, err = client.RbacV1().ClusterRoleBindings().Update(tCtx, discRoleBinding, metav1.UpdateOptions{}); err != nil { | ||||
| 		t.Fatalf("Failed to update `system:discovery` ClusterRoleBinding: %v", err) | ||||
| 	} | ||||
| 	t.Logf("Modifying default `system:basic-user` ClusterRoleBinding") | ||||
| 	basicUserRoleBinding, err := client.RbacV1().ClusterRoleBindings().Get(ctx, "system:basic-user", metav1.GetOptions{}) | ||||
| 	basicUserRoleBinding, err := client.RbacV1().ClusterRoleBindings().Get(tCtx, "system:basic-user", metav1.GetOptions{}) | ||||
| 	if err != nil { | ||||
| 		t.Fatalf("Failed to get `system:basic-user` ClusterRoleBinding: %v", err) | ||||
| 	} | ||||
| 	basicUserRoleBinding.Annotations["rbac.authorization.kubernetes.io/autoupdate"] = "false" | ||||
| 	basicUserRoleBinding.Annotations["rbac-discovery-upgrade-test"] = "pass" | ||||
| 	if basicUserRoleBinding, err = client.RbacV1().ClusterRoleBindings().Update(ctx, basicUserRoleBinding, metav1.UpdateOptions{}); err != nil { | ||||
| 	if basicUserRoleBinding, err = client.RbacV1().ClusterRoleBindings().Update(tCtx, basicUserRoleBinding, metav1.UpdateOptions{}); err != nil { | ||||
| 		t.Fatalf("Failed to update `system:basic-user` ClusterRoleBinding: %v", err) | ||||
| 	} | ||||
| 	t.Logf("Deleting default `system:public-info-viewer` ClusterRoleBinding") | ||||
| 	if err = client.RbacV1().ClusterRoleBindings().Delete(ctx, "system:public-info-viewer", metav1.DeleteOptions{}); err != nil { | ||||
| 	if err = client.RbacV1().ClusterRoleBindings().Delete(tCtx, "system:public-info-viewer", metav1.DeleteOptions{}); err != nil { | ||||
| 		t.Fatalf("Failed to delete `system:public-info-viewer` ClusterRoleBinding: %v", err) | ||||
| 	} | ||||
|  | ||||
| @@ -788,7 +778,7 @@ func TestDiscoveryUpgradeBootstrapping(t *testing.T) { | ||||
|  | ||||
| 	// Check that upgraded API servers inherit `system:public-info-viewer` settings from | ||||
| 	// `system:discovery`, and respect auto-reconciliation annotations. | ||||
| 	client, _, tearDownFn = framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 	client, _, tearDownFn = framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 		ModifyServerRunOptions: func(opts *options.ServerRunOptions) { | ||||
| 			// Ensure we're using the same etcd across apiserver restarts. | ||||
| 			opts.Etcd.StorageConfig = *etcdConfig | ||||
| @@ -796,21 +786,21 @@ func TestDiscoveryUpgradeBootstrapping(t *testing.T) { | ||||
| 		}, | ||||
| 	}) | ||||
|  | ||||
| 	newDiscRoleBinding, err := client.RbacV1().ClusterRoleBindings().Get(ctx, "system:discovery", metav1.GetOptions{}) | ||||
| 	newDiscRoleBinding, err := client.RbacV1().ClusterRoleBindings().Get(tCtx, "system:discovery", metav1.GetOptions{}) | ||||
| 	if err != nil { | ||||
| 		t.Fatalf("Failed to get `system:discovery` ClusterRoleBinding: %v", err) | ||||
| 	} | ||||
| 	if !reflect.DeepEqual(newDiscRoleBinding, discRoleBinding) { | ||||
| 		t.Errorf("`system:discovery` should have been unmodified. Wanted: %v, got %v", discRoleBinding, newDiscRoleBinding) | ||||
| 	} | ||||
| 	newBasicUserRoleBinding, err := client.RbacV1().ClusterRoleBindings().Get(ctx, "system:basic-user", metav1.GetOptions{}) | ||||
| 	newBasicUserRoleBinding, err := client.RbacV1().ClusterRoleBindings().Get(tCtx, "system:basic-user", metav1.GetOptions{}) | ||||
| 	if err != nil { | ||||
| 		t.Fatalf("Failed to get `system:basic-user` ClusterRoleBinding: %v", err) | ||||
| 	} | ||||
| 	if !reflect.DeepEqual(newBasicUserRoleBinding, basicUserRoleBinding) { | ||||
| 		t.Errorf("`system:basic-user` should have been unmodified. Wanted: %v, got %v", basicUserRoleBinding, newBasicUserRoleBinding) | ||||
| 	} | ||||
| 	publicInfoViewerRoleBinding, err := client.RbacV1().ClusterRoleBindings().Get(ctx, "system:public-info-viewer", metav1.GetOptions{}) | ||||
| 	publicInfoViewerRoleBinding, err := client.RbacV1().ClusterRoleBindings().Get(tCtx, "system:public-info-viewer", metav1.GetOptions{}) | ||||
| 	if err != nil { | ||||
| 		t.Fatalf("Failed to get `system:public-info-viewer` ClusterRoleBinding: %v", err) | ||||
| 	} | ||||
|   | ||||
| @@ -86,16 +86,13 @@ func TestGetsSelfAttributes(t *testing.T) { | ||||
| 		}, | ||||
| 	} | ||||
|  | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
|  | ||||
| 	tCtx := ktesting.Init(t) | ||||
| 	var respMu sync.RWMutex | ||||
| 	response := &user.DefaultInfo{ | ||||
| 		Name: "stub", | ||||
| 	} | ||||
|  | ||||
| 	kubeClient, _, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 	kubeClient, _, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 		ModifyServerRunOptions: func(opts *options.ServerRunOptions) { | ||||
| 			opts.APIEnablement.RuntimeConfig.Set("authentication.k8s.io/v1alpha1=true") | ||||
| 			opts.APIEnablement.RuntimeConfig.Set("authentication.k8s.io/v1beta1=true") | ||||
| @@ -122,7 +119,7 @@ func TestGetsSelfAttributes(t *testing.T) { | ||||
|  | ||||
| 			res, err := kubeClient.AuthenticationV1alpha1(). | ||||
| 				SelfSubjectReviews(). | ||||
| 				Create(ctx, &authenticationv1alpha1.SelfSubjectReview{}, metav1.CreateOptions{}) | ||||
| 				Create(tCtx, &authenticationv1alpha1.SelfSubjectReview{}, metav1.CreateOptions{}) | ||||
| 			if err != nil { | ||||
| 				t.Fatalf("unexpected error: %v", err) | ||||
| 			} | ||||
| @@ -149,7 +146,7 @@ func TestGetsSelfAttributes(t *testing.T) { | ||||
|  | ||||
| 			res2, err := kubeClient.AuthenticationV1beta1(). | ||||
| 				SelfSubjectReviews(). | ||||
| 				Create(ctx, &authenticationv1beta1.SelfSubjectReview{}, metav1.CreateOptions{}) | ||||
| 				Create(tCtx, &authenticationv1beta1.SelfSubjectReview{}, metav1.CreateOptions{}) | ||||
| 			if err != nil { | ||||
| 				t.Fatalf("unexpected error: %v", err) | ||||
| 			} | ||||
| @@ -208,11 +205,8 @@ func TestGetsSelfAttributesError(t *testing.T) { | ||||
| 	toggle := &atomic.Value{} | ||||
| 	toggle.Store(true) | ||||
|  | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
|  | ||||
| 	kubeClient, _, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 	tCtx := ktesting.Init(t) | ||||
| 	kubeClient, _, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 		ModifyServerRunOptions: func(opts *options.ServerRunOptions) { | ||||
| 			opts.APIEnablement.RuntimeConfig.Set("authentication.k8s.io/v1alpha1=true") | ||||
| 			opts.APIEnablement.RuntimeConfig.Set("authentication.k8s.io/v1beta1=true") | ||||
| @@ -244,7 +238,7 @@ func TestGetsSelfAttributesError(t *testing.T) { | ||||
|  | ||||
| 		_, err := kubeClient.AuthenticationV1alpha1(). | ||||
| 			SelfSubjectReviews(). | ||||
| 			Create(ctx, &authenticationv1alpha1.SelfSubjectReview{}, metav1.CreateOptions{}) | ||||
| 			Create(tCtx, &authenticationv1alpha1.SelfSubjectReview{}, metav1.CreateOptions{}) | ||||
| 		if err == nil { | ||||
| 			t.Fatalf("expected error: %v, got nil", err) | ||||
| 		} | ||||
| @@ -260,7 +254,7 @@ func TestGetsSelfAttributesError(t *testing.T) { | ||||
|  | ||||
| 		_, err := kubeClient.AuthenticationV1beta1(). | ||||
| 			SelfSubjectReviews(). | ||||
| 			Create(ctx, &authenticationv1beta1.SelfSubjectReview{}, metav1.CreateOptions{}) | ||||
| 			Create(tCtx, &authenticationv1beta1.SelfSubjectReview{}, metav1.CreateOptions{}) | ||||
| 		if err == nil { | ||||
| 			t.Fatalf("expected error: %v, got nil", err) | ||||
| 		} | ||||
|   | ||||
| @@ -79,9 +79,7 @@ func TestServiceAccountTokenCreate(t *testing.T) { | ||||
|  | ||||
| 	var tokenGenerator serviceaccount.TokenGenerator | ||||
|  | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
| 	tCtx := ktesting.Init(t) | ||||
|  | ||||
| 	// Enable the node token improvements feature gates prior to starting the apiserver, as the node getter is | ||||
| 	// conditionally passed to the service account token generator based on feature enablement. | ||||
| @@ -91,7 +89,7 @@ func TestServiceAccountTokenCreate(t *testing.T) { | ||||
|  | ||||
| 	// Start the server | ||||
| 	var serverAddress string | ||||
| 	kubeClient, kubeConfig, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 	kubeClient, kubeConfig, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 		ModifyServerRunOptions: func(opts *options.ServerRunOptions) { | ||||
| 			// Disable ServiceAccount admission plugin as we don't have serviceaccount controller running. | ||||
| 			opts.Admission.GenericAdmission.DisablePlugins = []string{"ServiceAccount"} | ||||
| @@ -193,7 +191,7 @@ func TestServiceAccountTokenCreate(t *testing.T) { | ||||
| 		} | ||||
|  | ||||
| 		warningHandler.clear() | ||||
| 		if resp, err := cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(ctx, sa.Name, treq, metav1.CreateOptions{}); err == nil { | ||||
| 		if resp, err := cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(tCtx, sa.Name, treq, metav1.CreateOptions{}); err == nil { | ||||
| 			t.Fatalf("expected err creating token for nonexistant svcacct but got: %#v", resp) | ||||
| 		} | ||||
| 		warningHandler.assertEqual(t, nil) | ||||
| @@ -202,18 +200,18 @@ func TestServiceAccountTokenCreate(t *testing.T) { | ||||
|  | ||||
| 		treqWithBadName := treq.DeepCopy() | ||||
| 		treqWithBadName.Name = "invalid-name" | ||||
| 		if resp, err := cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(ctx, sa.Name, treqWithBadName, metav1.CreateOptions{}); err == nil || !strings.Contains(err.Error(), "must match the service account name") { | ||||
| 		if resp, err := cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(tCtx, sa.Name, treqWithBadName, metav1.CreateOptions{}); err == nil || !strings.Contains(err.Error(), "must match the service account name") { | ||||
| 			t.Fatalf("expected err creating token with mismatched name but got: %#v", resp) | ||||
| 		} | ||||
|  | ||||
| 		treqWithBadNamespace := treq.DeepCopy() | ||||
| 		treqWithBadNamespace.Namespace = "invalid-namespace" | ||||
| 		if resp, err := cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(ctx, sa.Name, treqWithBadNamespace, metav1.CreateOptions{}); err == nil || !strings.Contains(err.Error(), "does not match the namespace") { | ||||
| 		if resp, err := cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(tCtx, sa.Name, treqWithBadNamespace, metav1.CreateOptions{}); err == nil || !strings.Contains(err.Error(), "does not match the namespace") { | ||||
| 			t.Fatalf("expected err creating token with mismatched namespace but got: %#v, %v", resp, err) | ||||
| 		} | ||||
|  | ||||
| 		warningHandler.clear() | ||||
| 		treq, err = cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(ctx, sa.Name, treq, metav1.CreateOptions{}) | ||||
| 		treq, err = cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(tCtx, sa.Name, treq, metav1.CreateOptions{}) | ||||
| 		if err != nil { | ||||
| 			t.Fatalf("err: %v", err) | ||||
| 		} | ||||
| @@ -259,7 +257,7 @@ func TestServiceAccountTokenCreate(t *testing.T) { | ||||
| 		} | ||||
|  | ||||
| 		warningHandler.clear() | ||||
| 		if resp, err := cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(ctx, sa.Name, treq, metav1.CreateOptions{}); err == nil { | ||||
| 		if resp, err := cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(tCtx, sa.Name, treq, metav1.CreateOptions{}); err == nil { | ||||
| 			t.Fatalf("expected err creating token for nonexistant svcacct but got: %#v", resp) | ||||
| 		} | ||||
| 		warningHandler.assertEqual(t, nil) | ||||
| @@ -267,7 +265,7 @@ func TestServiceAccountTokenCreate(t *testing.T) { | ||||
| 		defer del() | ||||
|  | ||||
| 		warningHandler.clear() | ||||
| 		if resp, err := cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(ctx, sa.Name, treq, metav1.CreateOptions{}); err == nil { | ||||
| 		if resp, err := cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(tCtx, sa.Name, treq, metav1.CreateOptions{}); err == nil { | ||||
| 			t.Fatalf("expected err creating token bound to nonexistant pod but got: %#v", resp) | ||||
| 		} | ||||
| 		warningHandler.assertEqual(t, nil) | ||||
| @@ -277,21 +275,21 @@ func TestServiceAccountTokenCreate(t *testing.T) { | ||||
| 		// right uid | ||||
| 		treq.Spec.BoundObjectRef.UID = pod.UID | ||||
| 		warningHandler.clear() | ||||
| 		if _, err := cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(ctx, sa.Name, treq, metav1.CreateOptions{}); err != nil { | ||||
| 		if _, err := cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(tCtx, sa.Name, treq, metav1.CreateOptions{}); err != nil { | ||||
| 			t.Fatalf("err: %v", err) | ||||
| 		} | ||||
| 		warningHandler.assertEqual(t, nil) | ||||
| 		// wrong uid | ||||
| 		treq.Spec.BoundObjectRef.UID = wrongUID | ||||
| 		warningHandler.clear() | ||||
| 		if resp, err := cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(ctx, sa.Name, treq, metav1.CreateOptions{}); err == nil { | ||||
| 		if resp, err := cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(tCtx, sa.Name, treq, metav1.CreateOptions{}); err == nil { | ||||
| 			t.Fatalf("expected err creating token bound to pod with wrong uid but got: %#v", resp) | ||||
| 		} | ||||
| 		warningHandler.assertEqual(t, nil) | ||||
| 		// no uid | ||||
| 		treq.Spec.BoundObjectRef.UID = noUID | ||||
| 		warningHandler.clear() | ||||
| 		treq, err = cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(ctx, sa.Name, treq, metav1.CreateOptions{}) | ||||
| 		treq, err = cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(tCtx, sa.Name, treq, metav1.CreateOptions{}) | ||||
| 		if err != nil { | ||||
| 			t.Fatalf("err: %v", err) | ||||
| 		} | ||||
| @@ -333,7 +331,7 @@ func TestServiceAccountTokenCreate(t *testing.T) { | ||||
| 			} | ||||
|  | ||||
| 			warningHandler.clear() | ||||
| 			if resp, err := cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(ctx, sa.Name, treq, metav1.CreateOptions{}); err == nil { | ||||
| 			if resp, err := cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(tCtx, sa.Name, treq, metav1.CreateOptions{}); err == nil { | ||||
| 				t.Fatalf("expected err creating token for nonexistant svcacct but got: %#v", resp) | ||||
| 			} | ||||
| 			warningHandler.assertEqual(t, nil) | ||||
| @@ -341,7 +339,7 @@ func TestServiceAccountTokenCreate(t *testing.T) { | ||||
| 			defer del() | ||||
|  | ||||
| 			warningHandler.clear() | ||||
| 			if resp, err := cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(ctx, sa.Name, treq, metav1.CreateOptions{}); err == nil { | ||||
| 			if resp, err := cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(tCtx, sa.Name, treq, metav1.CreateOptions{}); err == nil { | ||||
| 				t.Fatalf("expected err creating token bound to nonexistant pod but got: %#v", resp) | ||||
| 			} | ||||
| 			warningHandler.assertEqual(t, nil) | ||||
| @@ -357,21 +355,21 @@ func TestServiceAccountTokenCreate(t *testing.T) { | ||||
| 			// right uid | ||||
| 			treq.Spec.BoundObjectRef.UID = pod.UID | ||||
| 			warningHandler.clear() | ||||
| 			if _, err := cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(ctx, sa.Name, treq, metav1.CreateOptions{}); err != nil { | ||||
| 			if _, err := cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(tCtx, sa.Name, treq, metav1.CreateOptions{}); err != nil { | ||||
| 				t.Fatalf("err: %v", err) | ||||
| 			} | ||||
| 			warningHandler.assertEqual(t, nil) | ||||
| 			// wrong uid | ||||
| 			treq.Spec.BoundObjectRef.UID = wrongUID | ||||
| 			warningHandler.clear() | ||||
| 			if resp, err := cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(ctx, sa.Name, treq, metav1.CreateOptions{}); err == nil { | ||||
| 			if resp, err := cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(tCtx, sa.Name, treq, metav1.CreateOptions{}); err == nil { | ||||
| 				t.Fatalf("expected err creating token bound to pod with wrong uid but got: %#v", resp) | ||||
| 			} | ||||
| 			warningHandler.assertEqual(t, nil) | ||||
| 			// no uid | ||||
| 			treq.Spec.BoundObjectRef.UID = noUID | ||||
| 			warningHandler.clear() | ||||
| 			treq, err = cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(ctx, sa.Name, treq, metav1.CreateOptions{}) | ||||
| 			treq, err = cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(tCtx, sa.Name, treq, metav1.CreateOptions{}) | ||||
| 			if err != nil { | ||||
| 				t.Fatalf("err: %v", err) | ||||
| 			} | ||||
| @@ -437,7 +435,7 @@ func TestServiceAccountTokenCreate(t *testing.T) { | ||||
| 			}, | ||||
| 		} | ||||
| 		warningHandler.clear() | ||||
| 		if resp, err := cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(ctx, sa.Name, treq, metav1.CreateOptions{}); err == nil { | ||||
| 		if resp, err := cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(tCtx, sa.Name, treq, metav1.CreateOptions{}); err == nil { | ||||
| 			t.Fatalf("expected err creating token with featuregate disabled but got: %#v", resp) | ||||
| 		} else if err.Error() != "cannot bind token to a Node object as the \"ServiceAccountTokenNodeBinding\" feature-gate is disabled" { | ||||
| 			t.Fatalf("expected error due to feature gate being disabled, but got: %s", err.Error()) | ||||
| @@ -459,7 +457,7 @@ func TestServiceAccountTokenCreate(t *testing.T) { | ||||
| 		} | ||||
|  | ||||
| 		warningHandler.clear() | ||||
| 		if resp, err := cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(ctx, sa.Name, treq, metav1.CreateOptions{}); err == nil { | ||||
| 		if resp, err := cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(tCtx, sa.Name, treq, metav1.CreateOptions{}); err == nil { | ||||
| 			t.Fatalf("expected err creating token for nonexistant svcacct but got: %#v", resp) | ||||
| 		} | ||||
| 		warningHandler.assertEqual(t, nil) | ||||
| @@ -467,7 +465,7 @@ func TestServiceAccountTokenCreate(t *testing.T) { | ||||
| 		defer del() | ||||
|  | ||||
| 		warningHandler.clear() | ||||
| 		if resp, err := cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(ctx, sa.Name, treq, metav1.CreateOptions{}); err == nil { | ||||
| 		if resp, err := cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(tCtx, sa.Name, treq, metav1.CreateOptions{}); err == nil { | ||||
| 			t.Fatalf("expected err creating token bound to nonexistant node but got: %#v", resp) | ||||
| 		} | ||||
| 		warningHandler.assertEqual(t, nil) | ||||
| @@ -477,21 +475,21 @@ func TestServiceAccountTokenCreate(t *testing.T) { | ||||
| 		// right uid | ||||
| 		treq.Spec.BoundObjectRef.UID = node.UID | ||||
| 		warningHandler.clear() | ||||
| 		if _, err := cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(ctx, sa.Name, treq, metav1.CreateOptions{}); err != nil { | ||||
| 		if _, err := cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(tCtx, sa.Name, treq, metav1.CreateOptions{}); err != nil { | ||||
| 			t.Fatalf("err: %v", err) | ||||
| 		} | ||||
| 		warningHandler.assertEqual(t, nil) | ||||
| 		// wrong uid | ||||
| 		treq.Spec.BoundObjectRef.UID = wrongUID | ||||
| 		warningHandler.clear() | ||||
| 		if resp, err := cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(ctx, sa.Name, treq, metav1.CreateOptions{}); err == nil { | ||||
| 		if resp, err := cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(tCtx, sa.Name, treq, metav1.CreateOptions{}); err == nil { | ||||
| 			t.Fatalf("expected err creating token bound to node with wrong uid but got: %#v", resp) | ||||
| 		} | ||||
| 		warningHandler.assertEqual(t, nil) | ||||
| 		// no uid | ||||
| 		treq.Spec.BoundObjectRef.UID = noUID | ||||
| 		warningHandler.clear() | ||||
| 		treq, err = cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(ctx, sa.Name, treq, metav1.CreateOptions{}) | ||||
| 		treq, err = cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(tCtx, sa.Name, treq, metav1.CreateOptions{}) | ||||
| 		if err != nil { | ||||
| 			t.Fatalf("err: %v", err) | ||||
| 		} | ||||
| @@ -523,7 +521,7 @@ func TestServiceAccountTokenCreate(t *testing.T) { | ||||
| 		} | ||||
|  | ||||
| 		warningHandler.clear() | ||||
| 		if resp, err := cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(ctx, sa.Name, treq, metav1.CreateOptions{}); err == nil { | ||||
| 		if resp, err := cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(tCtx, sa.Name, treq, metav1.CreateOptions{}); err == nil { | ||||
| 			t.Fatalf("expected err creating token for nonexistant svcacct but got: %#v", resp) | ||||
| 		} | ||||
| 		warningHandler.assertEqual(t, nil) | ||||
| @@ -531,7 +529,7 @@ func TestServiceAccountTokenCreate(t *testing.T) { | ||||
| 		defer del() | ||||
|  | ||||
| 		warningHandler.clear() | ||||
| 		if resp, err := cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(ctx, sa.Name, treq, metav1.CreateOptions{}); err == nil { | ||||
| 		if resp, err := cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(tCtx, sa.Name, treq, metav1.CreateOptions{}); err == nil { | ||||
| 			t.Fatalf("expected err creating token bound to nonexistant secret but got: %#v", resp) | ||||
| 		} | ||||
| 		warningHandler.assertEqual(t, nil) | ||||
| @@ -541,21 +539,21 @@ func TestServiceAccountTokenCreate(t *testing.T) { | ||||
| 		// right uid | ||||
| 		treq.Spec.BoundObjectRef.UID = secret.UID | ||||
| 		warningHandler.clear() | ||||
| 		if _, err := cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(ctx, sa.Name, treq, metav1.CreateOptions{}); err != nil { | ||||
| 		if _, err := cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(tCtx, sa.Name, treq, metav1.CreateOptions{}); err != nil { | ||||
| 			t.Fatalf("err: %v", err) | ||||
| 		} | ||||
| 		warningHandler.assertEqual(t, nil) | ||||
| 		// wrong uid | ||||
| 		treq.Spec.BoundObjectRef.UID = wrongUID | ||||
| 		warningHandler.clear() | ||||
| 		if resp, err := cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(ctx, sa.Name, treq, metav1.CreateOptions{}); err == nil { | ||||
| 		if resp, err := cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(tCtx, sa.Name, treq, metav1.CreateOptions{}); err == nil { | ||||
| 			t.Fatalf("expected err creating token bound to secret with wrong uid but got: %#v", resp) | ||||
| 		} | ||||
| 		warningHandler.assertEqual(t, nil) | ||||
| 		// no uid | ||||
| 		treq.Spec.BoundObjectRef.UID = noUID | ||||
| 		warningHandler.clear() | ||||
| 		treq, err = cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(ctx, sa.Name, treq, metav1.CreateOptions{}) | ||||
| 		treq, err = cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(tCtx, sa.Name, treq, metav1.CreateOptions{}) | ||||
| 		if err != nil { | ||||
| 			t.Fatalf("err: %v", err) | ||||
| 		} | ||||
| @@ -591,7 +589,7 @@ func TestServiceAccountTokenCreate(t *testing.T) { | ||||
| 		defer del() | ||||
|  | ||||
| 		warningHandler.clear() | ||||
| 		if resp, err := cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(ctx, sa.Name, treq, metav1.CreateOptions{}); err == nil { | ||||
| 		if resp, err := cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(tCtx, sa.Name, treq, metav1.CreateOptions{}); err == nil { | ||||
| 			t.Fatalf("expected err but got: %#v", resp) | ||||
| 		} | ||||
| 		warningHandler.assertEqual(t, nil) | ||||
| @@ -608,7 +606,7 @@ func TestServiceAccountTokenCreate(t *testing.T) { | ||||
| 		defer del() | ||||
|  | ||||
| 		warningHandler.clear() | ||||
| 		treq, err = cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(ctx, sa.Name, treq, metav1.CreateOptions{}) | ||||
| 		treq, err = cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(tCtx, sa.Name, treq, metav1.CreateOptions{}) | ||||
| 		if err != nil { | ||||
| 			t.Fatalf("err: %v", err) | ||||
| 		} | ||||
| @@ -662,7 +660,7 @@ func TestServiceAccountTokenCreate(t *testing.T) { | ||||
| 		treq.Spec.BoundObjectRef.UID = pod.UID | ||||
|  | ||||
| 		warningHandler.clear() | ||||
| 		treq, err = cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(ctx, sa.Name, treq, metav1.CreateOptions{}) | ||||
| 		treq, err = cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(tCtx, sa.Name, treq, metav1.CreateOptions{}) | ||||
| 		if err != nil { | ||||
| 			t.Fatalf("err: %v", err) | ||||
| 		} | ||||
| @@ -718,7 +716,7 @@ func TestServiceAccountTokenCreate(t *testing.T) { | ||||
| 		treq.Spec.BoundObjectRef.UID = pod.UID | ||||
|  | ||||
| 		warningHandler.clear() | ||||
| 		treq, err = cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(ctx, sa.Name, treq, metav1.CreateOptions{}) | ||||
| 		treq, err = cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(tCtx, sa.Name, treq, metav1.CreateOptions{}) | ||||
| 		if err != nil { | ||||
| 			t.Fatalf("err: %v", err) | ||||
| 		} | ||||
| @@ -760,7 +758,7 @@ func TestServiceAccountTokenCreate(t *testing.T) { | ||||
| 		defer del() | ||||
|  | ||||
| 		warningHandler.clear() | ||||
| 		treq, err = cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(ctx, sa.Name, treq, metav1.CreateOptions{}) | ||||
| 		treq, err = cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(tCtx, sa.Name, treq, metav1.CreateOptions{}) | ||||
| 		if err != nil { | ||||
| 			t.Fatalf("err: %v", err) | ||||
| 		} | ||||
| @@ -778,7 +776,7 @@ func TestServiceAccountTokenCreate(t *testing.T) { | ||||
| 		defer del() | ||||
|  | ||||
| 		warningHandler.clear() | ||||
| 		treq, err = cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(ctx, sa.Name, treq, metav1.CreateOptions{}) | ||||
| 		treq, err = cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(tCtx, sa.Name, treq, metav1.CreateOptions{}) | ||||
| 		if err != nil { | ||||
| 			t.Fatalf("err: %v", err) | ||||
| 		} | ||||
| @@ -808,7 +806,7 @@ func TestServiceAccountTokenCreate(t *testing.T) { | ||||
|  | ||||
| 		treq.Spec.BoundObjectRef.UID = originalPod.UID | ||||
| 		warningHandler.clear() | ||||
| 		if treq, err = cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(ctx, sa.Name, treq, metav1.CreateOptions{}); err != nil { | ||||
| 		if treq, err = cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(tCtx, sa.Name, treq, metav1.CreateOptions{}); err != nil { | ||||
| 			t.Fatalf("err: %v", err) | ||||
| 		} | ||||
| 		warningHandler.assertEqual(t, nil) | ||||
| @@ -851,7 +849,7 @@ func TestServiceAccountTokenCreate(t *testing.T) { | ||||
|  | ||||
| 		treq.Spec.BoundObjectRef.UID = originalSecret.UID | ||||
| 		warningHandler.clear() | ||||
| 		if treq, err = cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(ctx, sa.Name, treq, metav1.CreateOptions{}); err != nil { | ||||
| 		if treq, err = cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(tCtx, sa.Name, treq, metav1.CreateOptions{}); err != nil { | ||||
| 			t.Fatalf("err: %v", err) | ||||
| 		} | ||||
| 		warningHandler.assertEqual(t, nil) | ||||
| @@ -896,7 +894,7 @@ func TestServiceAccountTokenCreate(t *testing.T) { | ||||
|  | ||||
| 		treq.Spec.BoundObjectRef.UID = originalSecret.UID | ||||
| 		warningHandler.clear() | ||||
| 		if treq, err = cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(ctx, sa.Name, treq, metav1.CreateOptions{}); err != nil { | ||||
| 		if treq, err = cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(tCtx, sa.Name, treq, metav1.CreateOptions{}); err != nil { | ||||
| 			t.Fatalf("err: %v", err) | ||||
| 		} | ||||
| 		warningHandler.assertEqual(t, nil) | ||||
| @@ -942,7 +940,7 @@ func TestServiceAccountTokenCreate(t *testing.T) { | ||||
|  | ||||
| 		treq.Spec.BoundObjectRef.UID = originalSecret.UID | ||||
| 		warningHandler.clear() | ||||
| 		if treq, err = cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(ctx, sa.Name, treq, metav1.CreateOptions{}); err != nil { | ||||
| 		if treq, err = cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(tCtx, sa.Name, treq, metav1.CreateOptions{}); err != nil { | ||||
| 			t.Fatalf("err: %v", err) | ||||
| 		} | ||||
| 		warningHandler.assertEqual(t, []string{fmt.Sprintf("requested expiration of %d seconds shortened to %d seconds", tooLongExpirationTime, maxExpirationSeconds)}) | ||||
| @@ -971,7 +969,7 @@ func TestServiceAccountTokenCreate(t *testing.T) { | ||||
|  | ||||
| 		t.Log("get token") | ||||
| 		warningHandler.clear() | ||||
| 		tokenRequest, err := cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(ctx, sa.Name, | ||||
| 		tokenRequest, err := cs.CoreV1().ServiceAccounts(sa.Namespace).CreateToken(tCtx, sa.Name, | ||||
| 			&authenticationv1.TokenRequest{ | ||||
| 				Spec: authenticationv1.TokenRequestSpec{ | ||||
| 					Audiences: []string{"api"}, | ||||
|   | ||||
| @@ -190,11 +190,8 @@ func TestStatus(t *testing.T) { | ||||
|  | ||||
| 	for _, tc := range testCases { | ||||
| 		t.Run(tc.name, func(t *testing.T) { | ||||
| 			_, ctx := ktesting.NewTestContext(t) | ||||
| 			ctx, cancel := context.WithCancel(ctx) | ||||
| 			defer cancel() | ||||
|  | ||||
| 			_, kubeConfig, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 			tCtx := ktesting.Init(t) | ||||
| 			_, kubeConfig, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 				ModifyServerRunOptions: func(options *options.ServerRunOptions) { | ||||
| 					if tc.modifyOptions != nil { | ||||
| 						tc.modifyOptions(options) | ||||
|   | ||||
| @@ -62,9 +62,7 @@ func setup(t *testing.T) (context.Context, kubeapiservertesting.TearDownFunc, *d | ||||
| } | ||||
|  | ||||
| func setupWithServerSetup(t *testing.T, serverSetup framework.TestServerSetup) (context.Context, kubeapiservertesting.TearDownFunc, *daemon.DaemonSetsController, informers.SharedInformerFactory, clientset.Interface) { | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
|  | ||||
| 	tCtx := ktesting.Init(t) | ||||
| 	modifyServerRunOptions := serverSetup.ModifyServerRunOptions | ||||
| 	serverSetup.ModifyServerRunOptions = func(opts *options.ServerRunOptions) { | ||||
| 		if modifyServerRunOptions != nil { | ||||
| @@ -79,12 +77,12 @@ func setupWithServerSetup(t *testing.T, serverSetup framework.TestServerSetup) ( | ||||
| 		) | ||||
| 	} | ||||
|  | ||||
| 	clientSet, config, closeFn := framework.StartTestServer(ctx, t, serverSetup) | ||||
| 	clientSet, config, closeFn := framework.StartTestServer(tCtx, t, serverSetup) | ||||
|  | ||||
| 	resyncPeriod := 12 * time.Hour | ||||
| 	informers := informers.NewSharedInformerFactory(clientset.NewForConfigOrDie(restclient.AddUserAgent(config, "daemonset-informers")), resyncPeriod) | ||||
| 	dc, err := daemon.NewDaemonSetsController( | ||||
| 		ctx, | ||||
| 		tCtx, | ||||
| 		informers.Apps().V1().DaemonSets(), | ||||
| 		informers.Apps().V1().ControllerRevisions(), | ||||
| 		informers.Core().V1().Pods(), | ||||
| @@ -101,7 +99,7 @@ func setupWithServerSetup(t *testing.T, serverSetup framework.TestServerSetup) ( | ||||
| 	}) | ||||
|  | ||||
| 	sched, err := scheduler.New( | ||||
| 		ctx, | ||||
| 		tCtx, | ||||
| 		clientSet, | ||||
| 		informers, | ||||
| 		nil, | ||||
| @@ -111,16 +109,16 @@ func setupWithServerSetup(t *testing.T, serverSetup framework.TestServerSetup) ( | ||||
| 		t.Fatalf("Couldn't create scheduler: %v", err) | ||||
| 	} | ||||
|  | ||||
| 	eventBroadcaster.StartRecordingToSink(ctx.Done()) | ||||
| 	go sched.Run(ctx) | ||||
| 	eventBroadcaster.StartRecordingToSink(tCtx.Done()) | ||||
| 	go sched.Run(tCtx) | ||||
|  | ||||
| 	tearDownFn := func() { | ||||
| 		cancel() | ||||
| 		tCtx.Cancel("tearing down apiserver") | ||||
| 		closeFn() | ||||
| 		eventBroadcaster.Shutdown() | ||||
| 	} | ||||
|  | ||||
| 	return ctx, tearDownFn, dc, informers, clientSet | ||||
| 	return tCtx, tearDownFn, dc, informers, clientSet | ||||
| } | ||||
|  | ||||
| func testLabels() map[string]string { | ||||
|   | ||||
| @@ -17,7 +17,6 @@ limitations under the License. | ||||
| package defaulttolerationseconds | ||||
|  | ||||
| import ( | ||||
| 	"context" | ||||
| 	"testing" | ||||
|  | ||||
| 	v1 "k8s.io/api/core/v1" | ||||
| @@ -30,11 +29,8 @@ import ( | ||||
| ) | ||||
|  | ||||
| func TestAdmission(t *testing.T) { | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
|  | ||||
| 	client, _, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 	tCtx := ktesting.Init(t) | ||||
| 	client, _, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 		ModifyServerConfig: func(cfg *controlplane.Config) { | ||||
| 			cfg.GenericConfig.EnableProfiling = true | ||||
| 			cfg.GenericConfig.AdmissionControl = defaulttolerationseconds.NewDefaultTolerationSeconds() | ||||
| @@ -60,7 +56,7 @@ func TestAdmission(t *testing.T) { | ||||
| 		}, | ||||
| 	} | ||||
|  | ||||
| 	updatedPod, err := client.CoreV1().Pods(pod.Namespace).Create(ctx, &pod, metav1.CreateOptions{}) | ||||
| 	updatedPod, err := client.CoreV1().Pods(pod.Namespace).Create(tCtx, &pod, metav1.CreateOptions{}) | ||||
| 	if err != nil { | ||||
| 		t.Fatalf("error creating pod: %v", err) | ||||
| 	} | ||||
|   | ||||
| @@ -114,18 +114,16 @@ func setup(ctx context.Context, t *testing.T) (*kubeapiservertesting.TestServer, | ||||
| } | ||||
|  | ||||
| func TestPDBWithScaleSubresource(t *testing.T) { | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
|  | ||||
| 	s, pdbc, informers, clientSet, apiExtensionClient, dynamicClient := setup(ctx, t) | ||||
| 	tCtx := ktesting.Init(t) | ||||
| 	s, pdbc, informers, clientSet, apiExtensionClient, dynamicClient := setup(tCtx, t) | ||||
| 	defer s.TearDownFn() | ||||
| 	defer cancel() | ||||
| 	defer tCtx.Cancel("test has completed") | ||||
|  | ||||
| 	nsName := "pdb-scale-subresource" | ||||
| 	createNs(ctx, t, nsName, clientSet) | ||||
| 	createNs(tCtx, t, nsName, clientSet) | ||||
|  | ||||
| 	informers.Start(ctx.Done()) | ||||
| 	go pdbc.Run(ctx) | ||||
| 	informers.Start(tCtx.Done()) | ||||
| 	go pdbc.Run(tCtx) | ||||
|  | ||||
| 	crdDefinition := newCustomResourceDefinition() | ||||
| 	etcd.CreateTestCRDs(t, apiExtensionClient, true, crdDefinition) | ||||
| @@ -148,7 +146,7 @@ func TestPDBWithScaleSubresource(t *testing.T) { | ||||
| 			}, | ||||
| 		}, | ||||
| 	} | ||||
| 	createdResource, err := resourceClient.Create(ctx, resource, metav1.CreateOptions{}) | ||||
| 	createdResource, err := resourceClient.Create(tCtx, resource, metav1.CreateOptions{}) | ||||
| 	if err != nil { | ||||
| 		t.Error(err) | ||||
| 	} | ||||
| @@ -164,7 +162,7 @@ func TestPDBWithScaleSubresource(t *testing.T) { | ||||
| 		}, | ||||
| 	} | ||||
| 	for i := 0; i < replicas; i++ { | ||||
| 		createPod(ctx, t, fmt.Sprintf("pod-%d", i), nsName, map[string]string{"app": "test-crd"}, clientSet, ownerRefs) | ||||
| 		createPod(tCtx, t, fmt.Sprintf("pod-%d", i), nsName, map[string]string{"app": "test-crd"}, clientSet, ownerRefs) | ||||
| 	} | ||||
|  | ||||
| 	waitToObservePods(t, informers.Core().V1().Pods().Informer(), 4, v1.PodRunning) | ||||
| @@ -183,13 +181,13 @@ func TestPDBWithScaleSubresource(t *testing.T) { | ||||
| 			}, | ||||
| 		}, | ||||
| 	} | ||||
| 	if _, err := clientSet.PolicyV1().PodDisruptionBudgets(nsName).Create(ctx, pdb, metav1.CreateOptions{}); err != nil { | ||||
| 	if _, err := clientSet.PolicyV1().PodDisruptionBudgets(nsName).Create(tCtx, pdb, metav1.CreateOptions{}); err != nil { | ||||
| 		t.Errorf("Error creating PodDisruptionBudget: %v", err) | ||||
| 	} | ||||
|  | ||||
| 	waitPDBStable(ctx, t, clientSet, 4, nsName, pdb.Name) | ||||
| 	waitPDBStable(tCtx, t, clientSet, 4, nsName, pdb.Name) | ||||
|  | ||||
| 	newPdb, err := clientSet.PolicyV1().PodDisruptionBudgets(nsName).Get(ctx, pdb.Name, metav1.GetOptions{}) | ||||
| 	newPdb, err := clientSet.PolicyV1().PodDisruptionBudgets(nsName).Get(tCtx, pdb.Name, metav1.GetOptions{}) | ||||
| 	if err != nil { | ||||
| 		t.Errorf("Error getting PodDisruptionBudget: %v", err) | ||||
| 	} | ||||
| @@ -248,37 +246,35 @@ func TestEmptySelector(t *testing.T) { | ||||
|  | ||||
| 	for i, tc := range testcases { | ||||
| 		t.Run(tc.name, func(t *testing.T) { | ||||
| 			_, ctx := ktesting.NewTestContext(t) | ||||
| 			ctx, cancel := context.WithCancel(ctx) | ||||
|  | ||||
| 			s, pdbc, informers, clientSet, _, _ := setup(ctx, t) | ||||
| 			tCtx := ktesting.Init(t) | ||||
| 			s, pdbc, informers, clientSet, _, _ := setup(tCtx, t) | ||||
| 			defer s.TearDownFn() | ||||
| 			defer cancel() | ||||
| 			defer tCtx.Cancel("test has completed") | ||||
|  | ||||
| 			nsName := fmt.Sprintf("pdb-empty-selector-%d", i) | ||||
| 			createNs(ctx, t, nsName, clientSet) | ||||
| 			createNs(tCtx, t, nsName, clientSet) | ||||
|  | ||||
| 			informers.Start(ctx.Done()) | ||||
| 			go pdbc.Run(ctx) | ||||
| 			informers.Start(tCtx.Done()) | ||||
| 			go pdbc.Run(tCtx) | ||||
|  | ||||
| 			replicas := 4 | ||||
| 			minAvailable := intstr.FromInt32(2) | ||||
|  | ||||
| 			for j := 0; j < replicas; j++ { | ||||
| 				createPod(ctx, t, fmt.Sprintf("pod-%d", j), nsName, map[string]string{"app": "test-crd"}, | ||||
| 				createPod(tCtx, t, fmt.Sprintf("pod-%d", j), nsName, map[string]string{"app": "test-crd"}, | ||||
| 					clientSet, []metav1.OwnerReference{}) | ||||
| 			} | ||||
|  | ||||
| 			waitToObservePods(t, informers.Core().V1().Pods().Informer(), 4, v1.PodRunning) | ||||
|  | ||||
| 			pdbName := "test-pdb" | ||||
| 			if err := tc.createPDBFunc(ctx, clientSet, s.EtcdClient, s.EtcdStoragePrefix, pdbName, nsName, minAvailable); err != nil { | ||||
| 			if err := tc.createPDBFunc(tCtx, clientSet, s.EtcdClient, s.EtcdStoragePrefix, pdbName, nsName, minAvailable); err != nil { | ||||
| 				t.Errorf("Error creating PodDisruptionBudget: %v", err) | ||||
| 			} | ||||
|  | ||||
| 			waitPDBStable(ctx, t, clientSet, tc.expectedCurrentHealthy, nsName, pdbName) | ||||
| 			waitPDBStable(tCtx, t, clientSet, tc.expectedCurrentHealthy, nsName, pdbName) | ||||
|  | ||||
| 			newPdb, err := clientSet.PolicyV1().PodDisruptionBudgets(nsName).Get(ctx, pdbName, metav1.GetOptions{}) | ||||
| 			newPdb, err := clientSet.PolicyV1().PodDisruptionBudgets(nsName).Get(tCtx, pdbName, metav1.GetOptions{}) | ||||
| 			if err != nil { | ||||
| 				t.Errorf("Error getting PodDisruptionBudget: %v", err) | ||||
| 			} | ||||
| @@ -364,36 +360,34 @@ func TestSelectorsForPodsWithoutLabels(t *testing.T) { | ||||
|  | ||||
| 	for i, tc := range testcases { | ||||
| 		t.Run(tc.name, func(t *testing.T) { | ||||
| 			_, ctx := ktesting.NewTestContext(t) | ||||
| 			ctx, cancel := context.WithCancel(ctx) | ||||
|  | ||||
| 			s, pdbc, informers, clientSet, _, _ := setup(ctx, t) | ||||
| 			tCtx := ktesting.Init(t) | ||||
| 			s, pdbc, informers, clientSet, _, _ := setup(tCtx, t) | ||||
| 			defer s.TearDownFn() | ||||
| 			defer cancel() | ||||
| 			defer tCtx.Cancel("test has completed") | ||||
|  | ||||
| 			nsName := fmt.Sprintf("pdb-selectors-%d", i) | ||||
| 			createNs(ctx, t, nsName, clientSet) | ||||
| 			createNs(tCtx, t, nsName, clientSet) | ||||
|  | ||||
| 			informers.Start(ctx.Done()) | ||||
| 			go pdbc.Run(ctx) | ||||
| 			informers.Start(tCtx.Done()) | ||||
| 			go pdbc.Run(tCtx) | ||||
|  | ||||
| 			minAvailable := intstr.FromInt32(1) | ||||
|  | ||||
| 			// Create the PDB first and wait for it to settle. | ||||
| 			pdbName := "test-pdb" | ||||
| 			if err := tc.createPDBFunc(ctx, clientSet, s.EtcdClient, s.EtcdStoragePrefix, pdbName, nsName, minAvailable); err != nil { | ||||
| 			if err := tc.createPDBFunc(tCtx, clientSet, s.EtcdClient, s.EtcdStoragePrefix, pdbName, nsName, minAvailable); err != nil { | ||||
| 				t.Errorf("Error creating PodDisruptionBudget: %v", err) | ||||
| 			} | ||||
| 			waitPDBStable(ctx, t, clientSet, 0, nsName, pdbName) | ||||
| 			waitPDBStable(tCtx, t, clientSet, 0, nsName, pdbName) | ||||
|  | ||||
| 			// Create a pod and wait for it be reach the running phase. | ||||
| 			createPod(ctx, t, "pod", nsName, map[string]string{}, clientSet, []metav1.OwnerReference{}) | ||||
| 			createPod(tCtx, t, "pod", nsName, map[string]string{}, clientSet, []metav1.OwnerReference{}) | ||||
| 			waitToObservePods(t, informers.Core().V1().Pods().Informer(), 1, v1.PodRunning) | ||||
|  | ||||
| 			// Then verify that the added pod are picked up by the disruption controller. | ||||
| 			waitPDBStable(ctx, t, clientSet, 1, nsName, pdbName) | ||||
| 			waitPDBStable(tCtx, t, clientSet, 1, nsName, pdbName) | ||||
|  | ||||
| 			newPdb, err := clientSet.PolicyV1().PodDisruptionBudgets(nsName).Get(ctx, pdbName, metav1.GetOptions{}) | ||||
| 			newPdb, err := clientSet.PolicyV1().PodDisruptionBudgets(nsName).Get(tCtx, pdbName, metav1.GetOptions{}) | ||||
| 			if err != nil { | ||||
| 				t.Errorf("Error getting PodDisruptionBudget: %v", err) | ||||
| 			} | ||||
| @@ -642,19 +636,17 @@ func TestPatchCompatibility(t *testing.T) { | ||||
| } | ||||
|  | ||||
| func TestStalePodDisruption(t *testing.T) { | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
|  | ||||
| 	s, pdbc, informers, clientSet, _, _ := setup(ctx, t) | ||||
| 	tCtx := ktesting.Init(t) | ||||
| 	s, pdbc, informers, clientSet, _, _ := setup(tCtx, t) | ||||
| 	defer s.TearDownFn() | ||||
| 	defer cancel() | ||||
| 	defer tCtx.Cancel("test has completed") | ||||
|  | ||||
| 	nsName := "pdb-stale-pod-disruption" | ||||
| 	createNs(ctx, t, nsName, clientSet) | ||||
| 	createNs(tCtx, t, nsName, clientSet) | ||||
|  | ||||
| 	informers.Start(ctx.Done()) | ||||
| 	informers.WaitForCacheSync(ctx.Done()) | ||||
| 	go pdbc.Run(ctx) | ||||
| 	informers.Start(tCtx.Done()) | ||||
| 	informers.WaitForCacheSync(tCtx.Done()) | ||||
| 	go pdbc.Run(tCtx) | ||||
|  | ||||
| 	cases := map[string]struct { | ||||
| 		deletePod      bool | ||||
| @@ -723,20 +715,20 @@ func TestStalePodDisruption(t *testing.T) { | ||||
| 				Reason:             tc.reason, | ||||
| 				LastTransitionTime: metav1.Now(), | ||||
| 			}) | ||||
| 			pod, err = clientSet.CoreV1().Pods(nsName).UpdateStatus(ctx, pod, metav1.UpdateOptions{}) | ||||
| 			pod, err = clientSet.CoreV1().Pods(nsName).UpdateStatus(tCtx, pod, metav1.UpdateOptions{}) | ||||
| 			if err != nil { | ||||
| 				t.Fatalf("Failed updating pod: %v", err) | ||||
| 			} | ||||
|  | ||||
| 			if tc.deletePod { | ||||
| 				if err := clientSet.CoreV1().Pods(nsName).Delete(ctx, name, metav1.DeleteOptions{}); err != nil { | ||||
| 				if err := clientSet.CoreV1().Pods(nsName).Delete(tCtx, name, metav1.DeleteOptions{}); err != nil { | ||||
| 					t.Fatalf("Failed to delete pod: %v", err) | ||||
| 				} | ||||
| 			} | ||||
| 			time.Sleep(stalePodDisruptionTimeout) | ||||
| 			diff := "" | ||||
| 			if err := wait.PollImmediate(100*time.Millisecond, wait.ForeverTestTimeout, func() (done bool, err error) { | ||||
| 				pod, err = clientSet.CoreV1().Pods(nsName).Get(ctx, name, metav1.GetOptions{}) | ||||
| 				pod, err = clientSet.CoreV1().Pods(nsName).Get(tCtx, name, metav1.GetOptions{}) | ||||
| 				if err != nil { | ||||
| 					return false, err | ||||
| 				} | ||||
|   | ||||
| @@ -17,7 +17,6 @@ limitations under the License. | ||||
| package dualstack | ||||
|  | ||||
| import ( | ||||
| 	"context" | ||||
| 	"encoding/json" | ||||
| 	"fmt" | ||||
| 	"reflect" | ||||
| @@ -46,11 +45,8 @@ func TestCreateServiceSingleStackIPv4(t *testing.T) { | ||||
| 	// Create an IPv4 single stack control-plane | ||||
| 	serviceCIDR := "10.0.0.0/16" | ||||
|  | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
|  | ||||
| 	client, _, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 	tCtx := ktesting.Init(t) | ||||
| 	client, _, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 		ModifyServerRunOptions: func(opts *options.ServerRunOptions) { | ||||
| 			opts.ServiceClusterIPRanges = serviceCIDR | ||||
| 		}, | ||||
| @@ -59,7 +55,7 @@ func TestCreateServiceSingleStackIPv4(t *testing.T) { | ||||
|  | ||||
| 	// Wait until the default "kubernetes" service is created. | ||||
| 	if err := wait.Poll(250*time.Millisecond, time.Minute, func() (bool, error) { | ||||
| 		_, err := client.CoreV1().Services(metav1.NamespaceDefault).Get(ctx, "kubernetes", metav1.GetOptions{}) | ||||
| 		_, err := client.CoreV1().Services(metav1.NamespaceDefault).Get(tCtx, "kubernetes", metav1.GetOptions{}) | ||||
| 		if err != nil && !apierrors.IsNotFound(err) { | ||||
| 			return false, err | ||||
| 		} | ||||
| @@ -252,7 +248,7 @@ func TestCreateServiceSingleStackIPv4(t *testing.T) { | ||||
| 			} | ||||
|  | ||||
| 			// create the service | ||||
| 			_, err := client.CoreV1().Services(metav1.NamespaceDefault).Create(ctx, svc, metav1.CreateOptions{}) | ||||
| 			_, err := client.CoreV1().Services(metav1.NamespaceDefault).Create(tCtx, svc, metav1.CreateOptions{}) | ||||
| 			if (err != nil) != tc.expectError { | ||||
| 				t.Errorf("Test failed expected result: %v received %v ", tc.expectError, err) | ||||
| 			} | ||||
| @@ -261,7 +257,7 @@ func TestCreateServiceSingleStackIPv4(t *testing.T) { | ||||
| 				return | ||||
| 			} | ||||
| 			// validate the service was created correctly if it was not expected to fail | ||||
| 			svc, err = client.CoreV1().Services(metav1.NamespaceDefault).Get(ctx, svc.Name, metav1.GetOptions{}) | ||||
| 			svc, err = client.CoreV1().Services(metav1.NamespaceDefault).Get(tCtx, svc.Name, metav1.GetOptions{}) | ||||
| 			if err != nil { | ||||
| 				t.Errorf("Unexpected error to get the service %s %v", svc.Name, err) | ||||
| 			} | ||||
| @@ -277,11 +273,8 @@ func TestCreateServiceDualStackIPv6(t *testing.T) { | ||||
| 	// Create an IPv6 only dual stack control-plane | ||||
| 	serviceCIDR := "2001:db8:1::/112" | ||||
|  | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
|  | ||||
| 	client, _, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 	tCtx := ktesting.Init(t) | ||||
| 	client, _, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 		ModifyServerRunOptions: func(opts *options.ServerRunOptions) { | ||||
| 			opts.ServiceClusterIPRanges = serviceCIDR | ||||
| 			opts.GenericServerRunOptions.AdvertiseAddress = netutils.ParseIPSloppy("2001:db8::10") | ||||
| @@ -291,7 +284,7 @@ func TestCreateServiceDualStackIPv6(t *testing.T) { | ||||
|  | ||||
| 	// Wait until the default "kubernetes" service is created. | ||||
| 	if err := wait.Poll(250*time.Millisecond, time.Minute, func() (bool, error) { | ||||
| 		_, err := client.CoreV1().Services(metav1.NamespaceDefault).Get(ctx, "kubernetes", metav1.GetOptions{}) | ||||
| 		_, err := client.CoreV1().Services(metav1.NamespaceDefault).Get(tCtx, "kubernetes", metav1.GetOptions{}) | ||||
| 		if err != nil && !apierrors.IsNotFound(err) { | ||||
| 			return false, err | ||||
| 		} | ||||
| @@ -471,7 +464,7 @@ func TestCreateServiceDualStackIPv6(t *testing.T) { | ||||
| 			} | ||||
|  | ||||
| 			// create the service | ||||
| 			_, err := client.CoreV1().Services(metav1.NamespaceDefault).Create(ctx, svc, metav1.CreateOptions{}) | ||||
| 			_, err := client.CoreV1().Services(metav1.NamespaceDefault).Create(tCtx, svc, metav1.CreateOptions{}) | ||||
| 			if (err != nil) != tc.expectError { | ||||
| 				t.Errorf("Test failed expected result: %v received %v ", tc.expectError, err) | ||||
| 			} | ||||
| @@ -480,7 +473,7 @@ func TestCreateServiceDualStackIPv6(t *testing.T) { | ||||
| 				return | ||||
| 			} | ||||
| 			// validate the service was created correctly if it was not expected to fail | ||||
| 			svc, err = client.CoreV1().Services(metav1.NamespaceDefault).Get(ctx, svc.Name, metav1.GetOptions{}) | ||||
| 			svc, err = client.CoreV1().Services(metav1.NamespaceDefault).Get(tCtx, svc.Name, metav1.GetOptions{}) | ||||
| 			if err != nil { | ||||
| 				t.Errorf("Unexpected error to get the service %s %v", svc.Name, err) | ||||
| 			} | ||||
| @@ -497,11 +490,8 @@ func TestCreateServiceDualStackIPv4IPv6(t *testing.T) { | ||||
| 	serviceCIDR := "10.0.0.0/16" | ||||
| 	secondaryServiceCIDR := "2001:db8:1::/112" | ||||
|  | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
|  | ||||
| 	client, _, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 	tCtx := ktesting.Init(t) | ||||
| 	client, _, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 		ModifyServerRunOptions: func(opts *options.ServerRunOptions) { | ||||
| 			opts.ServiceClusterIPRanges = fmt.Sprintf("%s,%s", serviceCIDR, secondaryServiceCIDR) | ||||
| 		}, | ||||
| @@ -510,7 +500,7 @@ func TestCreateServiceDualStackIPv4IPv6(t *testing.T) { | ||||
|  | ||||
| 	// Wait until the default "kubernetes" service is created. | ||||
| 	if err := wait.Poll(250*time.Millisecond, time.Minute, func() (bool, error) { | ||||
| 		_, err := client.CoreV1().Services(metav1.NamespaceDefault).Get(ctx, "kubernetes", metav1.GetOptions{}) | ||||
| 		_, err := client.CoreV1().Services(metav1.NamespaceDefault).Get(tCtx, "kubernetes", metav1.GetOptions{}) | ||||
| 		if err != nil && !apierrors.IsNotFound(err) { | ||||
| 			return false, err | ||||
| 		} | ||||
| @@ -737,7 +727,7 @@ func TestCreateServiceDualStackIPv4IPv6(t *testing.T) { | ||||
| 			} | ||||
|  | ||||
| 			// create a service | ||||
| 			_, err := client.CoreV1().Services(metav1.NamespaceDefault).Create(ctx, svc, metav1.CreateOptions{}) | ||||
| 			_, err := client.CoreV1().Services(metav1.NamespaceDefault).Create(tCtx, svc, metav1.CreateOptions{}) | ||||
| 			if (err != nil) != tc.expectError { | ||||
| 				t.Errorf("Test failed expected result: %v received %v ", tc.expectError, err) | ||||
| 			} | ||||
| @@ -746,7 +736,7 @@ func TestCreateServiceDualStackIPv4IPv6(t *testing.T) { | ||||
| 				return | ||||
| 			} | ||||
| 			// validate the service was created correctly if it was not expected to fail | ||||
| 			svc, err = client.CoreV1().Services(metav1.NamespaceDefault).Get(ctx, svc.Name, metav1.GetOptions{}) | ||||
| 			svc, err = client.CoreV1().Services(metav1.NamespaceDefault).Get(tCtx, svc.Name, metav1.GetOptions{}) | ||||
| 			if err != nil { | ||||
| 				t.Errorf("Unexpected error to get the service %s %v", svc.Name, err) | ||||
| 			} | ||||
| @@ -764,11 +754,8 @@ func TestCreateServiceDualStackIPv6IPv4(t *testing.T) { | ||||
| 	serviceCIDR := "2001:db8:1::/112" | ||||
| 	secondaryServiceCIDR := "10.0.0.0/16" | ||||
|  | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
|  | ||||
| 	client, _, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 	tCtx := ktesting.Init(t) | ||||
| 	client, _, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 		ModifyServerRunOptions: func(opts *options.ServerRunOptions) { | ||||
| 			opts.ServiceClusterIPRanges = fmt.Sprintf("%s,%s", serviceCIDR, secondaryServiceCIDR) | ||||
| 			opts.GenericServerRunOptions.AdvertiseAddress = netutils.ParseIPSloppy("2001:db8::10") | ||||
| @@ -778,7 +765,7 @@ func TestCreateServiceDualStackIPv6IPv4(t *testing.T) { | ||||
|  | ||||
| 	// Wait until the default "kubernetes" service is created. | ||||
| 	if err := wait.Poll(250*time.Millisecond, time.Minute, func() (bool, error) { | ||||
| 		_, err := client.CoreV1().Services(metav1.NamespaceDefault).Get(ctx, "kubernetes", metav1.GetOptions{}) | ||||
| 		_, err := client.CoreV1().Services(metav1.NamespaceDefault).Get(tCtx, "kubernetes", metav1.GetOptions{}) | ||||
| 		if err != nil && !apierrors.IsNotFound(err) { | ||||
| 			return false, err | ||||
| 		} | ||||
| @@ -789,7 +776,7 @@ func TestCreateServiceDualStackIPv6IPv4(t *testing.T) { | ||||
|  | ||||
| 	// verify client is working | ||||
| 	if err := wait.PollImmediate(5*time.Second, 2*time.Minute, func() (bool, error) { | ||||
| 		_, err := client.CoreV1().Endpoints("default").Get(ctx, "kubernetes", metav1.GetOptions{}) | ||||
| 		_, err := client.CoreV1().Endpoints("default").Get(tCtx, "kubernetes", metav1.GetOptions{}) | ||||
| 		if err != nil { | ||||
| 			t.Logf("error fetching endpoints: %v", err) | ||||
| 			return false, nil | ||||
| @@ -960,7 +947,7 @@ func TestCreateServiceDualStackIPv6IPv4(t *testing.T) { | ||||
| 			} | ||||
|  | ||||
| 			// create a service | ||||
| 			_, err := client.CoreV1().Services(metav1.NamespaceDefault).Create(ctx, svc, metav1.CreateOptions{}) | ||||
| 			_, err := client.CoreV1().Services(metav1.NamespaceDefault).Create(tCtx, svc, metav1.CreateOptions{}) | ||||
| 			if (err != nil) != tc.expectError { | ||||
| 				t.Errorf("Test failed expected result: %v received %v ", tc.expectError, err) | ||||
| 			} | ||||
| @@ -969,7 +956,7 @@ func TestCreateServiceDualStackIPv6IPv4(t *testing.T) { | ||||
| 				return | ||||
| 			} | ||||
| 			// validate the service was created correctly if it was not expected to fail | ||||
| 			svc, err = client.CoreV1().Services(metav1.NamespaceDefault).Get(ctx, svc.Name, metav1.GetOptions{}) | ||||
| 			svc, err = client.CoreV1().Services(metav1.NamespaceDefault).Get(tCtx, svc.Name, metav1.GetOptions{}) | ||||
| 			if err != nil { | ||||
| 				t.Errorf("Unexpected error to get the service %s %v", svc.Name, err) | ||||
| 			} | ||||
| @@ -987,11 +974,8 @@ func TestUpgradeDowngrade(t *testing.T) { | ||||
| 	serviceCIDR := "10.0.0.0/16" | ||||
| 	secondaryServiceCIDR := "2001:db8:1::/112" | ||||
|  | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
|  | ||||
| 	client, _, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 	tCtx := ktesting.Init(t) | ||||
| 	client, _, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 		ModifyServerRunOptions: func(opts *options.ServerRunOptions) { | ||||
| 			opts.ServiceClusterIPRanges = fmt.Sprintf("%s,%s", serviceCIDR, secondaryServiceCIDR) | ||||
| 		}, | ||||
| @@ -1000,7 +984,7 @@ func TestUpgradeDowngrade(t *testing.T) { | ||||
|  | ||||
| 	// Wait until the default "kubernetes" service is created. | ||||
| 	if err := wait.Poll(250*time.Millisecond, time.Minute, func() (bool, error) { | ||||
| 		_, err := client.CoreV1().Services(metav1.NamespaceDefault).Get(ctx, "kubernetes", metav1.GetOptions{}) | ||||
| 		_, err := client.CoreV1().Services(metav1.NamespaceDefault).Get(tCtx, "kubernetes", metav1.GetOptions{}) | ||||
| 		if err != nil && !apierrors.IsNotFound(err) { | ||||
| 			return false, err | ||||
| 		} | ||||
| @@ -1027,12 +1011,12 @@ func TestUpgradeDowngrade(t *testing.T) { | ||||
| 	} | ||||
|  | ||||
| 	// create a service | ||||
| 	_, err := client.CoreV1().Services(metav1.NamespaceDefault).Create(ctx, svc, metav1.CreateOptions{}) | ||||
| 	_, err := client.CoreV1().Services(metav1.NamespaceDefault).Create(tCtx, svc, metav1.CreateOptions{}) | ||||
| 	if err != nil { | ||||
| 		t.Fatalf("unexpected error while creating service:%v", err) | ||||
| 	} | ||||
| 	// validate the service was created correctly if it was not expected to fail | ||||
| 	svc, err = client.CoreV1().Services(metav1.NamespaceDefault).Get(ctx, svc.Name, metav1.GetOptions{}) | ||||
| 	svc, err = client.CoreV1().Services(metav1.NamespaceDefault).Get(tCtx, svc.Name, metav1.GetOptions{}) | ||||
| 	if err != nil { | ||||
| 		t.Fatalf("Unexpected error to get the service %s %v", svc.Name, err) | ||||
| 	} | ||||
| @@ -1044,7 +1028,7 @@ func TestUpgradeDowngrade(t *testing.T) { | ||||
| 	// upgrade it | ||||
| 	requireDualStack := v1.IPFamilyPolicyRequireDualStack | ||||
| 	svc.Spec.IPFamilyPolicy = &requireDualStack | ||||
| 	upgraded, err := client.CoreV1().Services(metav1.NamespaceDefault).Update(ctx, svc, metav1.UpdateOptions{}) | ||||
| 	upgraded, err := client.CoreV1().Services(metav1.NamespaceDefault).Update(tCtx, svc, metav1.UpdateOptions{}) | ||||
| 	if err != nil { | ||||
| 		t.Fatalf("unexpected error upgrading service to dual stack. %v", err) | ||||
| 	} | ||||
| @@ -1057,7 +1041,7 @@ func TestUpgradeDowngrade(t *testing.T) { | ||||
| 	upgraded.Spec.IPFamilyPolicy = &singleStack | ||||
| 	upgraded.Spec.ClusterIPs = upgraded.Spec.ClusterIPs[0:1] | ||||
| 	upgraded.Spec.IPFamilies = upgraded.Spec.IPFamilies[0:1] | ||||
| 	downgraded, err := client.CoreV1().Services(metav1.NamespaceDefault).Update(ctx, upgraded, metav1.UpdateOptions{}) | ||||
| 	downgraded, err := client.CoreV1().Services(metav1.NamespaceDefault).Update(tCtx, upgraded, metav1.UpdateOptions{}) | ||||
| 	if err != nil { | ||||
| 		t.Fatalf("unexpected error downgrading service to single stack. %v", err) | ||||
| 	} | ||||
| @@ -1067,7 +1051,7 @@ func TestUpgradeDowngrade(t *testing.T) { | ||||
|  | ||||
| 	// run test again this time without removing secondary IPFamily or ClusterIP | ||||
| 	downgraded.Spec.IPFamilyPolicy = &requireDualStack | ||||
| 	upgradedAgain, err := client.CoreV1().Services(metav1.NamespaceDefault).Update(ctx, downgraded, metav1.UpdateOptions{}) | ||||
| 	upgradedAgain, err := client.CoreV1().Services(metav1.NamespaceDefault).Update(tCtx, downgraded, metav1.UpdateOptions{}) | ||||
| 	if err != nil { | ||||
| 		t.Fatalf("unexpected error upgrading service to dual stack. %v", err) | ||||
| 	} | ||||
| @@ -1078,7 +1062,7 @@ func TestUpgradeDowngrade(t *testing.T) { | ||||
| 	upgradedAgain.Spec.IPFamilyPolicy = &singleStack | ||||
| 	// api-server automatically  removes the secondary ClusterIP and IPFamily | ||||
| 	// when a servie is downgraded. | ||||
| 	downgradedAgain, err := client.CoreV1().Services(metav1.NamespaceDefault).Update(ctx, upgradedAgain, metav1.UpdateOptions{}) | ||||
| 	downgradedAgain, err := client.CoreV1().Services(metav1.NamespaceDefault).Update(tCtx, upgradedAgain, metav1.UpdateOptions{}) | ||||
| 	if err != nil { | ||||
| 		t.Fatalf("unexpected error downgrading service to single stack. %v", err) | ||||
| 	} | ||||
| @@ -1094,11 +1078,8 @@ func TestConvertToFromExternalName(t *testing.T) { | ||||
| 	serviceCIDR := "10.0.0.0/16" | ||||
| 	secondaryServiceCIDR := "2001:db8:1::/112" | ||||
|  | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
|  | ||||
| 	client, _, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 	tCtx := ktesting.Init(t) | ||||
| 	client, _, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 		ModifyServerRunOptions: func(opts *options.ServerRunOptions) { | ||||
| 			opts.ServiceClusterIPRanges = fmt.Sprintf("%s,%s", serviceCIDR, secondaryServiceCIDR) | ||||
| 		}, | ||||
| @@ -1107,7 +1088,7 @@ func TestConvertToFromExternalName(t *testing.T) { | ||||
|  | ||||
| 	// Wait until the default "kubernetes" service is created. | ||||
| 	if err := wait.Poll(250*time.Millisecond, time.Minute, func() (bool, error) { | ||||
| 		_, err := client.CoreV1().Services(metav1.NamespaceDefault).Get(ctx, "kubernetes", metav1.GetOptions{}) | ||||
| 		_, err := client.CoreV1().Services(metav1.NamespaceDefault).Get(tCtx, "kubernetes", metav1.GetOptions{}) | ||||
| 		if err != nil && !apierrors.IsNotFound(err) { | ||||
| 			return false, err | ||||
| 		} | ||||
| @@ -1133,12 +1114,12 @@ func TestConvertToFromExternalName(t *testing.T) { | ||||
| 	} | ||||
|  | ||||
| 	// create a service | ||||
| 	_, err := client.CoreV1().Services(metav1.NamespaceDefault).Create(ctx, svc, metav1.CreateOptions{}) | ||||
| 	_, err := client.CoreV1().Services(metav1.NamespaceDefault).Create(tCtx, svc, metav1.CreateOptions{}) | ||||
| 	if err != nil { | ||||
| 		t.Fatalf("unexpected error while creating service:%v", err) | ||||
| 	} | ||||
| 	// validate the service was created correctly if it was not expected to fail | ||||
| 	svc, err = client.CoreV1().Services(metav1.NamespaceDefault).Get(ctx, svc.Name, metav1.GetOptions{}) | ||||
| 	svc, err = client.CoreV1().Services(metav1.NamespaceDefault).Get(tCtx, svc.Name, metav1.GetOptions{}) | ||||
| 	if err != nil { | ||||
| 		t.Fatalf("Unexpected error to get the service %s %v", svc.Name, err) | ||||
| 	} | ||||
| @@ -1152,7 +1133,7 @@ func TestConvertToFromExternalName(t *testing.T) { | ||||
| 	svc.Spec.ClusterIP = "" // not clearing ClusterIPs | ||||
| 	svc.Spec.ExternalName = "something.somewhere" | ||||
|  | ||||
| 	externalNameSvc, err := client.CoreV1().Services(metav1.NamespaceDefault).Update(ctx, svc, metav1.UpdateOptions{}) | ||||
| 	externalNameSvc, err := client.CoreV1().Services(metav1.NamespaceDefault).Update(tCtx, svc, metav1.UpdateOptions{}) | ||||
| 	if err != nil { | ||||
| 		t.Fatalf("unexpected error converting service to external name. %v", err) | ||||
| 	} | ||||
| @@ -1164,7 +1145,7 @@ func TestConvertToFromExternalName(t *testing.T) { | ||||
| 	// convert to a ClusterIP service | ||||
| 	externalNameSvc.Spec.Type = v1.ServiceTypeClusterIP | ||||
| 	externalNameSvc.Spec.ExternalName = "" | ||||
| 	clusterIPSvc, err := client.CoreV1().Services(metav1.NamespaceDefault).Update(ctx, externalNameSvc, metav1.UpdateOptions{}) | ||||
| 	clusterIPSvc, err := client.CoreV1().Services(metav1.NamespaceDefault).Update(tCtx, externalNameSvc, metav1.UpdateOptions{}) | ||||
| 	if err != nil { | ||||
| 		t.Fatalf("unexpected error converting service to ClusterIP. %v", err) | ||||
| 	} | ||||
| @@ -1179,11 +1160,8 @@ func TestPreferDualStack(t *testing.T) { | ||||
| 	serviceCIDR := "10.0.0.0/16" | ||||
| 	secondaryServiceCIDR := "2001:db8:1::/112" | ||||
|  | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
|  | ||||
| 	client, _, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 	tCtx := ktesting.Init(t) | ||||
| 	client, _, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 		ModifyServerRunOptions: func(opts *options.ServerRunOptions) { | ||||
| 			opts.ServiceClusterIPRanges = fmt.Sprintf("%s,%s", serviceCIDR, secondaryServiceCIDR) | ||||
| 		}, | ||||
| @@ -1192,7 +1170,7 @@ func TestPreferDualStack(t *testing.T) { | ||||
|  | ||||
| 	// Wait until the default "kubernetes" service is created. | ||||
| 	if err := wait.Poll(250*time.Millisecond, time.Minute, func() (bool, error) { | ||||
| 		_, err := client.CoreV1().Services(metav1.NamespaceDefault).Get(ctx, "kubernetes", metav1.GetOptions{}) | ||||
| 		_, err := client.CoreV1().Services(metav1.NamespaceDefault).Get(tCtx, "kubernetes", metav1.GetOptions{}) | ||||
| 		if err != nil && !apierrors.IsNotFound(err) { | ||||
| 			return false, err | ||||
| 		} | ||||
| @@ -1222,12 +1200,12 @@ func TestPreferDualStack(t *testing.T) { | ||||
| 	} | ||||
|  | ||||
| 	// create a service | ||||
| 	_, err := client.CoreV1().Services(metav1.NamespaceDefault).Create(ctx, svc, metav1.CreateOptions{}) | ||||
| 	_, err := client.CoreV1().Services(metav1.NamespaceDefault).Create(tCtx, svc, metav1.CreateOptions{}) | ||||
| 	if err != nil { | ||||
| 		t.Fatalf("unexpected error while creating service:%v", err) | ||||
| 	} | ||||
| 	// validate the service was created correctly if it was not expected to fail | ||||
| 	svc, err = client.CoreV1().Services(metav1.NamespaceDefault).Get(ctx, svc.Name, metav1.GetOptions{}) | ||||
| 	svc, err = client.CoreV1().Services(metav1.NamespaceDefault).Get(tCtx, svc.Name, metav1.GetOptions{}) | ||||
| 	if err != nil { | ||||
| 		t.Fatalf("Unexpected error to get the service %s %v", svc.Name, err) | ||||
| 	} | ||||
| @@ -1238,7 +1216,7 @@ func TestPreferDualStack(t *testing.T) { | ||||
|  | ||||
| 	// update it | ||||
| 	svc.Spec.Selector = map[string]string{"foo": "bar"} | ||||
| 	upgraded, err := client.CoreV1().Services(metav1.NamespaceDefault).Update(ctx, svc, metav1.UpdateOptions{}) | ||||
| 	upgraded, err := client.CoreV1().Services(metav1.NamespaceDefault).Update(tCtx, svc, metav1.UpdateOptions{}) | ||||
| 	if err != nil { | ||||
| 		t.Fatalf("unexpected error upgrading service to dual stack. %v", err) | ||||
| 	} | ||||
| @@ -1256,11 +1234,8 @@ func TestServiceUpdate(t *testing.T) { | ||||
| 	// Create an IPv4 single stack control-plane | ||||
| 	serviceCIDR := "10.0.0.0/16" | ||||
|  | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
|  | ||||
| 	client, _, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 	tCtx := ktesting.Init(t) | ||||
| 	client, _, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 		ModifyServerRunOptions: func(opts *options.ServerRunOptions) { | ||||
| 			opts.ServiceClusterIPRanges = serviceCIDR | ||||
| 		}, | ||||
| @@ -1269,7 +1244,7 @@ func TestServiceUpdate(t *testing.T) { | ||||
|  | ||||
| 	// Wait until the default "kubernetes" service is created. | ||||
| 	if err := wait.Poll(250*time.Millisecond, time.Minute, func() (bool, error) { | ||||
| 		_, err := client.CoreV1().Services(metav1.NamespaceDefault).Get(ctx, "kubernetes", metav1.GetOptions{}) | ||||
| 		_, err := client.CoreV1().Services(metav1.NamespaceDefault).Get(tCtx, "kubernetes", metav1.GetOptions{}) | ||||
| 		if err != nil && !apierrors.IsNotFound(err) { | ||||
| 			return false, err | ||||
| 		} | ||||
| @@ -1295,26 +1270,26 @@ func TestServiceUpdate(t *testing.T) { | ||||
| 	} | ||||
|  | ||||
| 	// create the service | ||||
| 	_, err := client.CoreV1().Services(metav1.NamespaceDefault).Create(ctx, svc, metav1.CreateOptions{}) | ||||
| 	_, err := client.CoreV1().Services(metav1.NamespaceDefault).Create(tCtx, svc, metav1.CreateOptions{}) | ||||
| 	// if no error was expected validate the service otherwise return | ||||
| 	if err != nil { | ||||
| 		t.Errorf("unexpected error creating service:%v", err) | ||||
| 		return | ||||
| 	} | ||||
|  | ||||
| 	svc, err = client.CoreV1().Services(metav1.NamespaceDefault).Get(ctx, svc.Name, metav1.GetOptions{}) | ||||
| 	svc, err = client.CoreV1().Services(metav1.NamespaceDefault).Get(tCtx, svc.Name, metav1.GetOptions{}) | ||||
| 	if err != nil { | ||||
| 		t.Errorf("Unexpected error to get the service %s %v", svc.Name, err) | ||||
| 	} | ||||
|  | ||||
| 	// update using put | ||||
| 	svc.Labels = map[string]string{"x": "y"} | ||||
| 	_, err = client.CoreV1().Services(metav1.NamespaceDefault).Update(ctx, svc, metav1.UpdateOptions{}) | ||||
| 	_, err = client.CoreV1().Services(metav1.NamespaceDefault).Update(tCtx, svc, metav1.UpdateOptions{}) | ||||
| 	if err != nil { | ||||
| 		t.Errorf("Unexpected error updating the service %s %v", svc.Name, err) | ||||
| 	} | ||||
|  | ||||
| 	_, err = client.CoreV1().Services(metav1.NamespaceDefault).Get(ctx, svc.Name, metav1.GetOptions{}) | ||||
| 	_, err = client.CoreV1().Services(metav1.NamespaceDefault).Get(tCtx, svc.Name, metav1.GetOptions{}) | ||||
| 	if err != nil { | ||||
| 		t.Fatalf("Unexpected error to get the service %s %v", svc.Name, err) | ||||
| 	} | ||||
| @@ -1329,12 +1304,12 @@ func TestServiceUpdate(t *testing.T) { | ||||
| 		t.Fatalf("failed to json.Marshal labels: %v", err) | ||||
| 	} | ||||
|  | ||||
| 	_, err = client.CoreV1().Services(metav1.NamespaceDefault).Patch(ctx, svc.Name, types.StrategicMergePatchType, patchBytes, metav1.PatchOptions{}) | ||||
| 	_, err = client.CoreV1().Services(metav1.NamespaceDefault).Patch(tCtx, svc.Name, types.StrategicMergePatchType, patchBytes, metav1.PatchOptions{}) | ||||
| 	if err != nil { | ||||
| 		t.Fatalf("unexpected error patching service using strategic merge patch. %v", err) | ||||
| 	} | ||||
|  | ||||
| 	current, err := client.CoreV1().Services(metav1.NamespaceDefault).Get(ctx, svc.Name, metav1.GetOptions{}) | ||||
| 	current, err := client.CoreV1().Services(metav1.NamespaceDefault).Get(tCtx, svc.Name, metav1.GetOptions{}) | ||||
| 	if err != nil { | ||||
| 		t.Fatalf("Unexpected error to get the service %s %v", svc.Name, err) | ||||
| 	} | ||||
| @@ -1356,13 +1331,13 @@ func TestServiceUpdate(t *testing.T) { | ||||
| 		t.Fatalf("unexpected error creating json patch. %v", err) | ||||
| 	} | ||||
|  | ||||
| 	_, err = client.CoreV1().Services(metav1.NamespaceDefault).Patch(ctx, svc.Name, types.MergePatchType, patchBytes, metav1.PatchOptions{}) | ||||
| 	_, err = client.CoreV1().Services(metav1.NamespaceDefault).Patch(tCtx, svc.Name, types.MergePatchType, patchBytes, metav1.PatchOptions{}) | ||||
| 	if err != nil { | ||||
| 		t.Fatalf("unexpected error patching service using merge patch. %v", err) | ||||
| 	} | ||||
|  | ||||
| 	// validate the service was created correctly if it was not expected to fail | ||||
| 	_, err = client.CoreV1().Services(metav1.NamespaceDefault).Get(ctx, svc.Name, metav1.GetOptions{}) | ||||
| 	_, err = client.CoreV1().Services(metav1.NamespaceDefault).Get(tCtx, svc.Name, metav1.GetOptions{}) | ||||
| 	if err != nil { | ||||
| 		t.Fatalf("Unexpected error to get the service %s %v", svc.Name, err) | ||||
| 	} | ||||
| @@ -1418,14 +1393,12 @@ func validateServiceAndClusterIPFamily(svc *v1.Service, expectedIPFamilies []v1. | ||||
| func TestUpgradeServicePreferToDualStack(t *testing.T) { | ||||
| 	sharedEtcd := framework.SharedEtcd() | ||||
|  | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
| 	tCtx := ktesting.Init(t) | ||||
|  | ||||
| 	// Create an IPv4 only dual stack control-plane | ||||
| 	serviceCIDR := "192.168.0.0/24" | ||||
|  | ||||
| 	client, _, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 	client, _, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 		ModifyServerRunOptions: func(opts *options.ServerRunOptions) { | ||||
| 			opts.Etcd.StorageConfig = *sharedEtcd | ||||
| 			opts.ServiceClusterIPRanges = serviceCIDR | ||||
| @@ -1434,7 +1407,7 @@ func TestUpgradeServicePreferToDualStack(t *testing.T) { | ||||
|  | ||||
| 	// Wait until the default "kubernetes" service is created. | ||||
| 	if err := wait.Poll(250*time.Millisecond, time.Minute, func() (bool, error) { | ||||
| 		_, err := client.CoreV1().Services(metav1.NamespaceDefault).Get(ctx, "kubernetes", metav1.GetOptions{}) | ||||
| 		_, err := client.CoreV1().Services(metav1.NamespaceDefault).Get(tCtx, "kubernetes", metav1.GetOptions{}) | ||||
| 		if err != nil && !apierrors.IsNotFound(err) { | ||||
| 			return false, err | ||||
| 		} | ||||
| @@ -1465,12 +1438,12 @@ func TestUpgradeServicePreferToDualStack(t *testing.T) { | ||||
| 	} | ||||
|  | ||||
| 	// create the service | ||||
| 	_, err := client.CoreV1().Services(metav1.NamespaceDefault).Create(ctx, svc, metav1.CreateOptions{}) | ||||
| 	_, err := client.CoreV1().Services(metav1.NamespaceDefault).Create(tCtx, svc, metav1.CreateOptions{}) | ||||
| 	if err != nil { | ||||
| 		t.Fatalf("Unexpected error: %v", err) | ||||
| 	} | ||||
| 	// validate the service was created correctly if it was not expected to fail | ||||
| 	svc, err = client.CoreV1().Services(metav1.NamespaceDefault).Get(ctx, svc.Name, metav1.GetOptions{}) | ||||
| 	svc, err = client.CoreV1().Services(metav1.NamespaceDefault).Get(tCtx, svc.Name, metav1.GetOptions{}) | ||||
| 	if err != nil { | ||||
| 		t.Fatalf("Unexpected error to get the service %s %v", svc.Name, err) | ||||
| 	} | ||||
| @@ -1483,7 +1456,7 @@ func TestUpgradeServicePreferToDualStack(t *testing.T) { | ||||
|  | ||||
| 	secondaryServiceCIDR := "2001:db8:1::/112" | ||||
|  | ||||
| 	client, _, tearDownFn = framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 	client, _, tearDownFn = framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 		ModifyServerRunOptions: func(opts *options.ServerRunOptions) { | ||||
| 			opts.Etcd.StorageConfig = *sharedEtcd | ||||
| 			opts.ServiceClusterIPRanges = fmt.Sprintf("%s,%s", serviceCIDR, secondaryServiceCIDR) | ||||
| @@ -1493,7 +1466,7 @@ func TestUpgradeServicePreferToDualStack(t *testing.T) { | ||||
|  | ||||
| 	// Wait until the default "kubernetes" service is created. | ||||
| 	if err = wait.Poll(250*time.Millisecond, time.Minute, func() (bool, error) { | ||||
| 		_, err := client.CoreV1().Services(metav1.NamespaceDefault).Get(ctx, "kubernetes", metav1.GetOptions{}) | ||||
| 		_, err := client.CoreV1().Services(metav1.NamespaceDefault).Get(tCtx, "kubernetes", metav1.GetOptions{}) | ||||
| 		if err != nil && !apierrors.IsNotFound(err) { | ||||
| 			return false, err | ||||
| 		} | ||||
| @@ -1502,7 +1475,7 @@ func TestUpgradeServicePreferToDualStack(t *testing.T) { | ||||
| 		t.Fatalf("creating kubernetes service timed out") | ||||
| 	} | ||||
| 	// validate the service was created correctly if it was not expected to fail | ||||
| 	svc, err = client.CoreV1().Services(metav1.NamespaceDefault).Get(ctx, svc.Name, metav1.GetOptions{}) | ||||
| 	svc, err = client.CoreV1().Services(metav1.NamespaceDefault).Get(tCtx, svc.Name, metav1.GetOptions{}) | ||||
| 	if err != nil { | ||||
| 		t.Fatalf("Unexpected error to get the service %s %v", svc.Name, err) | ||||
| 	} | ||||
| @@ -1515,15 +1488,13 @@ func TestUpgradeServicePreferToDualStack(t *testing.T) { | ||||
| func TestDowngradeServicePreferToDualStack(t *testing.T) { | ||||
| 	sharedEtcd := framework.SharedEtcd() | ||||
|  | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
| 	tCtx := ktesting.Init(t) | ||||
|  | ||||
| 	// Create a dual stack control-plane | ||||
| 	serviceCIDR := "192.168.0.0/24" | ||||
| 	secondaryServiceCIDR := "2001:db8:1::/112" | ||||
|  | ||||
| 	client, _, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 	client, _, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 		ModifyServerRunOptions: func(opts *options.ServerRunOptions) { | ||||
| 			opts.Etcd.StorageConfig = *sharedEtcd | ||||
| 			opts.ServiceClusterIPRanges = fmt.Sprintf("%s,%s", serviceCIDR, secondaryServiceCIDR) | ||||
| @@ -1532,7 +1503,7 @@ func TestDowngradeServicePreferToDualStack(t *testing.T) { | ||||
|  | ||||
| 	// Wait until the default "kubernetes" service is created. | ||||
| 	if err := wait.Poll(250*time.Millisecond, time.Minute, func() (bool, error) { | ||||
| 		_, err := client.CoreV1().Services(metav1.NamespaceDefault).Get(ctx, "kubernetes", metav1.GetOptions{}) | ||||
| 		_, err := client.CoreV1().Services(metav1.NamespaceDefault).Get(tCtx, "kubernetes", metav1.GetOptions{}) | ||||
| 		if err != nil && !apierrors.IsNotFound(err) { | ||||
| 			return false, err | ||||
| 		} | ||||
| @@ -1561,12 +1532,12 @@ func TestDowngradeServicePreferToDualStack(t *testing.T) { | ||||
| 		}, | ||||
| 	} | ||||
| 	// create the service | ||||
| 	_, err := client.CoreV1().Services(metav1.NamespaceDefault).Create(ctx, svc, metav1.CreateOptions{}) | ||||
| 	_, err := client.CoreV1().Services(metav1.NamespaceDefault).Create(tCtx, svc, metav1.CreateOptions{}) | ||||
| 	if err != nil { | ||||
| 		t.Fatalf("Unexpected error: %v", err) | ||||
| 	} | ||||
| 	// validate the service was created correctly if it was not expected to fail | ||||
| 	svc, err = client.CoreV1().Services(metav1.NamespaceDefault).Get(ctx, svc.Name, metav1.GetOptions{}) | ||||
| 	svc, err = client.CoreV1().Services(metav1.NamespaceDefault).Get(tCtx, svc.Name, metav1.GetOptions{}) | ||||
| 	if err != nil { | ||||
| 		t.Fatalf("Unexpected error to get the service %s %v", svc.Name, err) | ||||
| 	} | ||||
| @@ -1577,7 +1548,7 @@ func TestDowngradeServicePreferToDualStack(t *testing.T) { | ||||
| 	tearDownFn() | ||||
|  | ||||
| 	// reset secondary | ||||
| 	client, _, tearDownFn = framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 	client, _, tearDownFn = framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 		ModifyServerRunOptions: func(opts *options.ServerRunOptions) { | ||||
| 			opts.Etcd.StorageConfig = *sharedEtcd | ||||
| 			opts.ServiceClusterIPRanges = serviceCIDR | ||||
| @@ -1587,7 +1558,7 @@ func TestDowngradeServicePreferToDualStack(t *testing.T) { | ||||
|  | ||||
| 	// Wait until the default "kubernetes" service is created. | ||||
| 	if err = wait.Poll(250*time.Millisecond, time.Minute, func() (bool, error) { | ||||
| 		_, err := client.CoreV1().Services(metav1.NamespaceDefault).Get(ctx, "kubernetes", metav1.GetOptions{}) | ||||
| 		_, err := client.CoreV1().Services(metav1.NamespaceDefault).Get(tCtx, "kubernetes", metav1.GetOptions{}) | ||||
| 		if err != nil && !apierrors.IsNotFound(err) { | ||||
| 			return false, err | ||||
| 		} | ||||
| @@ -1596,7 +1567,7 @@ func TestDowngradeServicePreferToDualStack(t *testing.T) { | ||||
| 		t.Fatalf("creating kubernetes service timed out") | ||||
| 	} | ||||
| 	// validate the service is still there. | ||||
| 	svc, err = client.CoreV1().Services(metav1.NamespaceDefault).Get(ctx, svc.Name, metav1.GetOptions{}) | ||||
| 	svc, err = client.CoreV1().Services(metav1.NamespaceDefault).Get(tCtx, svc.Name, metav1.GetOptions{}) | ||||
| 	if err != nil { | ||||
| 		t.Fatalf("Unexpected error to get the service %s %v", svc.Name, err) | ||||
| 	} | ||||
| @@ -1616,11 +1587,8 @@ type specMergePatch struct { | ||||
|  | ||||
| // tests success when converting ClusterIP:Headless service to ExternalName | ||||
| func Test_ServiceChangeTypeHeadlessToExternalNameWithPatch(t *testing.T) { | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
|  | ||||
| 	client, _, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{}) | ||||
| 	tCtx := ktesting.Init(t) | ||||
| 	client, _, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{}) | ||||
| 	defer tearDownFn() | ||||
|  | ||||
| 	ns := framework.CreateNamespaceOrDie(client, "test-service-allocate-node-ports", t) | ||||
| @@ -1638,7 +1606,7 @@ func Test_ServiceChangeTypeHeadlessToExternalNameWithPatch(t *testing.T) { | ||||
| 	} | ||||
|  | ||||
| 	var err error | ||||
| 	service, err = client.CoreV1().Services(ns.Name).Create(ctx, service, metav1.CreateOptions{}) | ||||
| 	service, err = client.CoreV1().Services(ns.Name).Create(tCtx, service, metav1.CreateOptions{}) | ||||
| 	if err != nil { | ||||
| 		t.Fatalf("Error creating test service: %v", err) | ||||
| 	} | ||||
| @@ -1654,7 +1622,7 @@ func Test_ServiceChangeTypeHeadlessToExternalNameWithPatch(t *testing.T) { | ||||
| 		t.Fatalf("failed to json.Marshal ports: %v", err) | ||||
| 	} | ||||
|  | ||||
| 	_, err = client.CoreV1().Services(ns.Name).Patch(ctx, service.Name, types.StrategicMergePatchType, patchBytes, metav1.PatchOptions{}) | ||||
| 	_, err = client.CoreV1().Services(ns.Name).Patch(tCtx, service.Name, types.StrategicMergePatchType, patchBytes, metav1.PatchOptions{}) | ||||
| 	if err != nil { | ||||
| 		t.Fatalf("unexpected error patching service using strategic merge patch. %v", err) | ||||
| 	} | ||||
|   | ||||
| @@ -69,18 +69,16 @@ const ( | ||||
| func TestConcurrentEvictionRequests(t *testing.T) { | ||||
| 	podNameFormat := "test-pod-%d" | ||||
|  | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
|  | ||||
| 	closeFn, rm, informers, _, clientSet := rmSetup(ctx, t) | ||||
| 	tCtx := ktesting.Init(t) | ||||
| 	closeFn, rm, informers, _, clientSet := rmSetup(tCtx, t) | ||||
| 	defer closeFn() | ||||
|  | ||||
| 	ns := framework.CreateNamespaceOrDie(clientSet, "concurrent-eviction-requests", t) | ||||
| 	defer framework.DeleteNamespaceOrDie(clientSet, ns, t) | ||||
| 	defer cancel() | ||||
| 	defer tCtx.Cancel("test has completed") | ||||
|  | ||||
| 	informers.Start(ctx.Done()) | ||||
| 	go rm.Run(ctx) | ||||
| 	informers.Start(tCtx.Done()) | ||||
| 	go rm.Run(tCtx) | ||||
|  | ||||
| 	var gracePeriodSeconds int64 = 30 | ||||
| 	deleteOption := metav1.DeleteOptions{ | ||||
| @@ -184,18 +182,16 @@ func TestConcurrentEvictionRequests(t *testing.T) { | ||||
|  | ||||
| // TestTerminalPodEviction ensures that PDB is not checked for terminal pods. | ||||
| func TestTerminalPodEviction(t *testing.T) { | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
|  | ||||
| 	closeFn, rm, informers, _, clientSet := rmSetup(ctx, t) | ||||
| 	tCtx := ktesting.Init(t) | ||||
| 	closeFn, rm, informers, _, clientSet := rmSetup(tCtx, t) | ||||
| 	defer closeFn() | ||||
|  | ||||
| 	ns := framework.CreateNamespaceOrDie(clientSet, "terminalpod-eviction", t) | ||||
| 	defer framework.DeleteNamespaceOrDie(clientSet, ns, t) | ||||
| 	defer cancel() | ||||
| 	defer tCtx.Cancel("test has completed") | ||||
|  | ||||
| 	informers.Start(ctx.Done()) | ||||
| 	go rm.Run(ctx) | ||||
| 	informers.Start(tCtx.Done()) | ||||
| 	go rm.Run(tCtx) | ||||
|  | ||||
| 	var gracePeriodSeconds int64 = 30 | ||||
| 	deleteOption := metav1.DeleteOptions{ | ||||
| @@ -260,15 +256,13 @@ func TestTerminalPodEviction(t *testing.T) { | ||||
|  | ||||
| // TestEvictionVersions ensures the eviction endpoint accepts and returns the correct API versions | ||||
| func TestEvictionVersions(t *testing.T) { | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
|  | ||||
| 	closeFn, rm, informers, config, clientSet := rmSetup(ctx, t) | ||||
| 	tCtx := ktesting.Init(t) | ||||
| 	closeFn, rm, informers, config, clientSet := rmSetup(tCtx, t) | ||||
| 	defer closeFn() | ||||
| 	defer cancel() | ||||
| 	defer tCtx.Cancel("test has completed") | ||||
|  | ||||
| 	informers.Start(ctx.Done()) | ||||
| 	go rm.Run(ctx) | ||||
| 	informers.Start(tCtx.Done()) | ||||
| 	go rm.Run(tCtx) | ||||
|  | ||||
| 	ns := "default" | ||||
| 	subresource := "eviction" | ||||
| @@ -386,29 +380,27 @@ func TestEvictionWithFinalizers(t *testing.T) { | ||||
| 	} | ||||
| 	for name, tc := range cases { | ||||
| 		t.Run(name, func(t *testing.T) { | ||||
| 			_, ctx := ktesting.NewTestContext(t) | ||||
| 			ctx, cancel := context.WithCancel(ctx) | ||||
|  | ||||
| 			closeFn, rm, informers, _, clientSet := rmSetup(ctx, t) | ||||
| 			tCtx := ktesting.Init(t) | ||||
| 			closeFn, rm, informers, _, clientSet := rmSetup(tCtx, t) | ||||
| 			defer closeFn() | ||||
|  | ||||
| 			ns := framework.CreateNamespaceOrDie(clientSet, "eviction-with-finalizers", t) | ||||
| 			defer framework.DeleteNamespaceOrDie(clientSet, ns, t) | ||||
| 			defer featuregatetesting.SetFeatureGateDuringTest(t, feature.DefaultFeatureGate, features.PodDisruptionConditions, tc.enablePodDisruptionConditions)() | ||||
| 			defer cancel() | ||||
| 			defer tCtx.Cancel("test has completed") | ||||
|  | ||||
| 			informers.Start(ctx.Done()) | ||||
| 			go rm.Run(ctx) | ||||
| 			informers.Start(tCtx.Done()) | ||||
| 			go rm.Run(tCtx) | ||||
|  | ||||
| 			pod := newPod("pod") | ||||
| 			pod.ObjectMeta.Finalizers = []string{"test.k8s.io/finalizer"} | ||||
| 			if _, err := clientSet.CoreV1().Pods(ns.Name).Create(ctx, pod, metav1.CreateOptions{}); err != nil { | ||||
| 			if _, err := clientSet.CoreV1().Pods(ns.Name).Create(tCtx, pod, metav1.CreateOptions{}); err != nil { | ||||
| 				t.Errorf("Failed to create pod: %v", err) | ||||
| 			} | ||||
|  | ||||
| 			pod.Status.Phase = tc.phase | ||||
| 			addPodConditionReady(pod) | ||||
| 			if _, err := clientSet.CoreV1().Pods(ns.Name).UpdateStatus(ctx, pod, metav1.UpdateOptions{}); err != nil { | ||||
| 			if _, err := clientSet.CoreV1().Pods(ns.Name).UpdateStatus(tCtx, pod, metav1.UpdateOptions{}); err != nil { | ||||
| 				t.Fatal(err) | ||||
| 			} | ||||
|  | ||||
| @@ -420,12 +412,12 @@ func TestEvictionWithFinalizers(t *testing.T) { | ||||
|  | ||||
| 			eviction := newV1Eviction(ns.Name, pod.Name, deleteOption) | ||||
|  | ||||
| 			err := clientSet.PolicyV1().Evictions(ns.Name).Evict(ctx, eviction) | ||||
| 			err := clientSet.PolicyV1().Evictions(ns.Name).Evict(tCtx, eviction) | ||||
| 			if err != nil { | ||||
| 				t.Fatalf("Eviction of pod failed %v", err) | ||||
| 			} | ||||
|  | ||||
| 			updatedPod, e := clientSet.CoreV1().Pods(ns.Name).Get(ctx, pod.Name, metav1.GetOptions{}) | ||||
| 			updatedPod, e := clientSet.CoreV1().Pods(ns.Name).Get(tCtx, pod.Name, metav1.GetOptions{}) | ||||
| 			if e != nil { | ||||
| 				t.Fatalf("Failed to get the pod %q with error: %q", klog.KObj(pod), e) | ||||
| 			} | ||||
| @@ -474,19 +466,17 @@ func TestEvictionWithUnhealthyPodEvictionPolicy(t *testing.T) { | ||||
| 	} | ||||
| 	for name, tc := range cases { | ||||
| 		t.Run(name, func(t *testing.T) { | ||||
| 			_, ctx := ktesting.NewTestContext(t) | ||||
| 			ctx, cancel := context.WithCancel(ctx) | ||||
|  | ||||
| 			tCtx := ktesting.Init(t) | ||||
| 			defer featuregatetesting.SetFeatureGateDuringTest(t, feature.DefaultFeatureGate, features.PDBUnhealthyPodEvictionPolicy, tc.enableUnhealthyPodEvictionPolicy)() | ||||
| 			closeFn, rm, informers, _, clientSet := rmSetup(ctx, t) | ||||
| 			closeFn, rm, informers, _, clientSet := rmSetup(tCtx, t) | ||||
| 			defer closeFn() | ||||
|  | ||||
| 			ns := framework.CreateNamespaceOrDie(clientSet, "eviction-with-pdb-pod-healthy-policy", t) | ||||
| 			defer framework.DeleteNamespaceOrDie(clientSet, ns, t) | ||||
| 			defer cancel() | ||||
| 			defer tCtx.Cancel("test has completed") | ||||
|  | ||||
| 			informers.Start(ctx.Done()) | ||||
| 			go rm.Run(ctx) | ||||
| 			informers.Start(tCtx.Done()) | ||||
| 			go rm.Run(tCtx) | ||||
|  | ||||
| 			pod := newPod("pod") | ||||
| 			if _, err := clientSet.CoreV1().Pods(ns.Name).Create(context.TODO(), pod, metav1.CreateOptions{}); err != nil { | ||||
| @@ -498,7 +488,7 @@ func TestEvictionWithUnhealthyPodEvictionPolicy(t *testing.T) { | ||||
| 				addPodConditionReady(pod) | ||||
| 			} | ||||
|  | ||||
| 			if _, err := clientSet.CoreV1().Pods(ns.Name).UpdateStatus(ctx, pod, metav1.UpdateOptions{}); err != nil { | ||||
| 			if _, err := clientSet.CoreV1().Pods(ns.Name).UpdateStatus(tCtx, pod, metav1.UpdateOptions{}); err != nil { | ||||
| 				t.Fatal(err) | ||||
| 			} | ||||
|  | ||||
| @@ -524,7 +514,7 @@ func TestEvictionWithUnhealthyPodEvictionPolicy(t *testing.T) { | ||||
|  | ||||
| 			deleteOption := metav1.DeleteOptions{} | ||||
| 			eviction := newV1Eviction(ns.Name, pod.Name, deleteOption) | ||||
| 			err := policyV1NoRetriesClient.Evictions(ns.Name).Evict(ctx, eviction) | ||||
| 			err := policyV1NoRetriesClient.Evictions(ns.Name).Evict(tCtx, eviction) | ||||
| 			if err != nil { | ||||
| 				t.Fatalf("Eviction of pod failed %v", err) | ||||
| 			} | ||||
| @@ -570,18 +560,16 @@ func TestEvictionWithPrecondition(t *testing.T) { | ||||
| 	} | ||||
| 	for name, tc := range cases { | ||||
| 		t.Run(name, func(t *testing.T) { | ||||
| 			_, ctx := ktesting.NewTestContext(t) | ||||
| 			ctx, cancel := context.WithCancel(ctx) | ||||
|  | ||||
| 			closeFn, rm, informers, _, clientSet := rmSetup(ctx, t) | ||||
| 			tCtx := ktesting.Init(t) | ||||
| 			closeFn, rm, informers, _, clientSet := rmSetup(tCtx, t) | ||||
| 			defer closeFn() | ||||
|  | ||||
| 			ns := framework.CreateNamespaceOrDie(clientSet, "eviction-with-preconditions", t) | ||||
| 			defer framework.DeleteNamespaceOrDie(clientSet, ns, t) | ||||
|  | ||||
| 			defer cancel() | ||||
| 			informers.Start(ctx.Done()) | ||||
| 			go rm.Run(ctx) | ||||
| 			defer tCtx.Cancel("test has completed") | ||||
| 			informers.Start(tCtx.Done()) | ||||
| 			go rm.Run(tCtx) | ||||
|  | ||||
| 			pod := newPod("pod") | ||||
| 			pod, err := clientSet.CoreV1().Pods(ns.Name).Create(context.TODO(), pod, metav1.CreateOptions{}) | ||||
| @@ -593,7 +581,7 @@ func TestEvictionWithPrecondition(t *testing.T) { | ||||
| 			addPodConditionReady(pod) | ||||
|  | ||||
| 			// generate a new resource version | ||||
| 			updatedPod, err := clientSet.CoreV1().Pods(ns.Name).UpdateStatus(ctx, pod, metav1.UpdateOptions{}) | ||||
| 			updatedPod, err := clientSet.CoreV1().Pods(ns.Name).UpdateStatus(tCtx, pod, metav1.UpdateOptions{}) | ||||
| 			if err != nil { | ||||
| 				t.Fatal(err) | ||||
| 			} | ||||
| @@ -629,7 +617,7 @@ func TestEvictionWithPrecondition(t *testing.T) { | ||||
| 			policyV1NoRetriesClient := policyv1client.New(policyV1NoRetriesRESTClient) | ||||
|  | ||||
| 			eviction := newV1Eviction(ns.Name, updatedPod.Name, deleteOption) | ||||
| 			err = policyV1NoRetriesClient.Evictions(ns.Name).Evict(ctx, eviction) | ||||
| 			err = policyV1NoRetriesClient.Evictions(ns.Name).Evict(tCtx, eviction) | ||||
| 			if err != nil && !tc.shouldErr { | ||||
| 				t.Fatalf("Eviction of pod failed %q", err) | ||||
| 			} | ||||
|   | ||||
| @@ -17,7 +17,6 @@ limitations under the License. | ||||
| package apiserver | ||||
|  | ||||
| import ( | ||||
| 	"context" | ||||
| 	"sync/atomic" | ||||
| 	"testing" | ||||
| 	"time" | ||||
| @@ -42,11 +41,8 @@ func TestWebhookLoopback(t *testing.T) { | ||||
|  | ||||
| 	called := int32(0) | ||||
|  | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
|  | ||||
| 	client, _, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 	tCtx := ktesting.Init(t) | ||||
| 	client, _, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 		ModifyServerRunOptions: func(opts *options.ServerRunOptions) { | ||||
| 		}, | ||||
| 		ModifyServerConfig: func(config *controlplane.Config) { | ||||
| @@ -72,7 +68,7 @@ func TestWebhookLoopback(t *testing.T) { | ||||
|  | ||||
| 	fail := admissionregistrationv1.Fail | ||||
| 	noSideEffects := admissionregistrationv1.SideEffectClassNone | ||||
| 	_, err := client.AdmissionregistrationV1().MutatingWebhookConfigurations().Create(ctx, &admissionregistrationv1.MutatingWebhookConfiguration{ | ||||
| 	_, err := client.AdmissionregistrationV1().MutatingWebhookConfigurations().Create(tCtx, &admissionregistrationv1.MutatingWebhookConfiguration{ | ||||
| 		ObjectMeta: metav1.ObjectMeta{Name: "webhooktest.example.com"}, | ||||
| 		Webhooks: []admissionregistrationv1.MutatingWebhook{{ | ||||
| 			Name: "webhooktest.example.com", | ||||
| @@ -93,7 +89,7 @@ func TestWebhookLoopback(t *testing.T) { | ||||
| 	} | ||||
|  | ||||
| 	err = wait.PollImmediate(100*time.Millisecond, 30*time.Second, func() (done bool, err error) { | ||||
| 		_, err = client.CoreV1().ConfigMaps("default").Create(ctx, &v1.ConfigMap{ | ||||
| 		_, err = client.CoreV1().ConfigMaps("default").Create(tCtx, &v1.ConfigMap{ | ||||
| 			ObjectMeta: metav1.ObjectMeta{Name: "webhook-test"}, | ||||
| 			Data:       map[string]string{"invalid key": "value"}, | ||||
| 		}, metav1.CreateOptions{}) | ||||
|   | ||||
| @@ -59,6 +59,7 @@ type TestServerSetup struct { | ||||
| type TearDownFunc func() | ||||
|  | ||||
| // StartTestServer runs a kube-apiserver, optionally calling out to the setup.ModifyServerRunOptions and setup.ModifyServerConfig functions | ||||
| // TODO (pohly): convert to ktesting contexts | ||||
| func StartTestServer(ctx context.Context, t testing.TB, setup TestServerSetup) (client.Interface, *rest.Config, TearDownFunc) { | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
|  | ||||
|   | ||||
| @@ -248,12 +248,12 @@ func setupWithServer(t *testing.T, result *kubeapiservertesting.TestServer, work | ||||
| 	sharedInformers := informers.NewSharedInformerFactory(clientSet, 0) | ||||
| 	metadataInformers := metadatainformer.NewSharedInformerFactory(metadataClient, 0) | ||||
|  | ||||
| 	logger, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	tCtx := ktesting.Init(t) | ||||
| 	logger := tCtx.Logger() | ||||
| 	alwaysStarted := make(chan struct{}) | ||||
| 	close(alwaysStarted) | ||||
| 	gc, err := garbagecollector.NewGarbageCollector( | ||||
| 		ctx, | ||||
| 		tCtx, | ||||
| 		clientSet, | ||||
| 		metadataClient, | ||||
| 		restMapper, | ||||
| @@ -266,7 +266,7 @@ func setupWithServer(t *testing.T, result *kubeapiservertesting.TestServer, work | ||||
| 	} | ||||
|  | ||||
| 	tearDown := func() { | ||||
| 		cancel() | ||||
| 		tCtx.Cancel("tearing down") | ||||
| 		result.TearDownFn() | ||||
| 	} | ||||
| 	syncPeriod := 5 * time.Second | ||||
| @@ -276,9 +276,9 @@ func setupWithServer(t *testing.T, result *kubeapiservertesting.TestServer, work | ||||
| 			// client. This is a leaky abstraction and assumes behavior about the REST | ||||
| 			// mapper, but we'll deal with it for now. | ||||
| 			restMapper.Reset() | ||||
| 		}, syncPeriod, ctx.Done()) | ||||
| 		go gc.Run(ctx, workers) | ||||
| 		go gc.Sync(ctx, clientSet.Discovery(), syncPeriod) | ||||
| 		}, syncPeriod, tCtx.Done()) | ||||
| 		go gc.Run(tCtx, workers) | ||||
| 		go gc.Sync(tCtx, clientSet.Discovery(), syncPeriod) | ||||
| 	} | ||||
|  | ||||
| 	if workerCount > 0 { | ||||
|   | ||||
| @@ -17,7 +17,6 @@ limitations under the License. | ||||
| package network | ||||
|  | ||||
| import ( | ||||
| 	"context" | ||||
| 	"encoding/json" | ||||
| 	"fmt" | ||||
| 	"testing" | ||||
| @@ -45,11 +44,8 @@ func TestServicesFinalizersRepairLoop(t *testing.T) { | ||||
| 	clusterIP := "10.0.0.20" | ||||
| 	interval := 5 * time.Second | ||||
|  | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
|  | ||||
| 	client, _, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 	tCtx := ktesting.Init(t) | ||||
| 	client, _, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 		ModifyServerRunOptions: func(opts *options.ServerRunOptions) { | ||||
| 			opts.ServiceClusterIPRanges = serviceCIDR | ||||
| 		}, | ||||
| @@ -61,7 +57,7 @@ func TestServicesFinalizersRepairLoop(t *testing.T) { | ||||
|  | ||||
| 	// verify client is working | ||||
| 	if err := wait.PollImmediate(5*time.Second, 2*time.Minute, func() (bool, error) { | ||||
| 		_, err := client.CoreV1().Endpoints(metav1.NamespaceDefault).Get(ctx, "kubernetes", metav1.GetOptions{}) | ||||
| 		_, err := client.CoreV1().Endpoints(metav1.NamespaceDefault).Get(tCtx, "kubernetes", metav1.GetOptions{}) | ||||
| 		if err != nil { | ||||
| 			t.Logf("error fetching endpoints: %v", err) | ||||
| 			return false, nil | ||||
| @@ -90,20 +86,20 @@ func TestServicesFinalizersRepairLoop(t *testing.T) { | ||||
| 	} | ||||
|  | ||||
| 	// Create service | ||||
| 	if _, err := client.CoreV1().Services(metav1.NamespaceDefault).Create(ctx, &svcNodePort, metav1.CreateOptions{}); err != nil { | ||||
| 	if _, err := client.CoreV1().Services(metav1.NamespaceDefault).Create(tCtx, &svcNodePort, metav1.CreateOptions{}); err != nil { | ||||
| 		t.Errorf("unexpected error creating service: %v", err) | ||||
| 	} | ||||
| 	t.Logf("Created service: %s", svcNodePort.Name) | ||||
|  | ||||
| 	// Check the service has been created correctly | ||||
| 	svc, err := client.CoreV1().Services(metav1.NamespaceDefault).Get(ctx, svcNodePort.Name, metav1.GetOptions{}) | ||||
| 	svc, err := client.CoreV1().Services(metav1.NamespaceDefault).Get(tCtx, svcNodePort.Name, metav1.GetOptions{}) | ||||
| 	if err != nil || svc.Spec.ClusterIP != clusterIP { | ||||
| 		t.Errorf("created service is not correct: %v", err) | ||||
| 	} | ||||
| 	t.Logf("Service created successfully: %v", svc) | ||||
|  | ||||
| 	// Delete service | ||||
| 	if err := client.CoreV1().Services(metav1.NamespaceDefault).Delete(ctx, svcNodePort.Name, metav1.DeleteOptions{}); err != nil { | ||||
| 	if err := client.CoreV1().Services(metav1.NamespaceDefault).Delete(tCtx, svcNodePort.Name, metav1.DeleteOptions{}); err != nil { | ||||
| 		t.Errorf("unexpected error deleting service: %v", err) | ||||
| 	} | ||||
| 	t.Logf("Deleted service: %s", svcNodePort.Name) | ||||
| @@ -112,26 +108,26 @@ func TestServicesFinalizersRepairLoop(t *testing.T) { | ||||
| 	time.Sleep(interval + 1) | ||||
|  | ||||
| 	// Check that the service was not deleted and the IP is already allocated | ||||
| 	svc, err = client.CoreV1().Services(metav1.NamespaceDefault).Get(ctx, svcNodePort.Name, metav1.GetOptions{}) | ||||
| 	svc, err = client.CoreV1().Services(metav1.NamespaceDefault).Get(tCtx, svcNodePort.Name, metav1.GetOptions{}) | ||||
| 	if err != nil || svc.Spec.ClusterIP != clusterIP { | ||||
| 		t.Errorf("created service is not correct: %v", err) | ||||
| 	} | ||||
| 	t.Logf("Service after Delete: %v", svc) | ||||
|  | ||||
| 	// Remove the finalizer | ||||
| 	if _, err = client.CoreV1().Services(metav1.NamespaceDefault).Patch(ctx, svcNodePort.Name, types.JSONPatchType, []byte(`[{"op":"remove","path":"/metadata/finalizers"}]`), metav1.PatchOptions{}); err != nil { | ||||
| 	if _, err = client.CoreV1().Services(metav1.NamespaceDefault).Patch(tCtx, svcNodePort.Name, types.JSONPatchType, []byte(`[{"op":"remove","path":"/metadata/finalizers"}]`), metav1.PatchOptions{}); err != nil { | ||||
| 		t.Errorf("unexpected error removing finalizer: %v", err) | ||||
| 	} | ||||
| 	t.Logf("Removed service finalizer: %s", svcNodePort.Name) | ||||
|  | ||||
| 	// Check that the service was deleted | ||||
| 	_, err = client.CoreV1().Services(metav1.NamespaceDefault).Get(ctx, svcNodePort.Name, metav1.GetOptions{}) | ||||
| 	_, err = client.CoreV1().Services(metav1.NamespaceDefault).Get(tCtx, svcNodePort.Name, metav1.GetOptions{}) | ||||
| 	if err == nil { | ||||
| 		t.Errorf("service was not delete: %v", err) | ||||
| 	} | ||||
|  | ||||
| 	// Try to create service again | ||||
| 	if _, err := client.CoreV1().Services(metav1.NamespaceDefault).Create(ctx, &svcNodePort, metav1.CreateOptions{}); err != nil { | ||||
| 	if _, err := client.CoreV1().Services(metav1.NamespaceDefault).Create(tCtx, &svcNodePort, metav1.CreateOptions{}); err != nil { | ||||
| 		t.Errorf("unexpected error creating service: %v", err) | ||||
| 	} | ||||
| 	t.Logf("Created service: %s", svcNodePort.Name) | ||||
| @@ -141,11 +137,8 @@ func TestServicesFinalizersPatchStatus(t *testing.T) { | ||||
| 	serviceCIDR := "10.0.0.0/16" | ||||
| 	clusterIP := "10.0.0.21" | ||||
| 	nodePort := 30443 | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
|  | ||||
| 	client, _, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 	tCtx := ktesting.Init(t) | ||||
| 	client, _, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 		ModifyServerRunOptions: func(opts *options.ServerRunOptions) { | ||||
| 			opts.ServiceClusterIPRanges = serviceCIDR | ||||
| 		}, | ||||
| @@ -176,26 +169,26 @@ func TestServicesFinalizersPatchStatus(t *testing.T) { | ||||
| 			defer framework.DeleteNamespaceOrDie(client, ns, t) | ||||
|  | ||||
| 			// Create service | ||||
| 			if _, err := client.CoreV1().Services(ns.Name).Create(ctx, &svcNodePort, metav1.CreateOptions{}); err != nil { | ||||
| 			if _, err := client.CoreV1().Services(ns.Name).Create(tCtx, &svcNodePort, metav1.CreateOptions{}); err != nil { | ||||
| 				t.Fatalf("unexpected error creating service: %v", err) | ||||
| 			} | ||||
| 			t.Logf("Created service: %s", svcNodePort.Name) | ||||
|  | ||||
| 			// Check the service has been created correctly | ||||
| 			svc, err := client.CoreV1().Services(ns.Name).Get(ctx, svcNodePort.Name, metav1.GetOptions{}) | ||||
| 			svc, err := client.CoreV1().Services(ns.Name).Get(tCtx, svcNodePort.Name, metav1.GetOptions{}) | ||||
| 			if err != nil || svc.Spec.ClusterIP != clusterIP { | ||||
| 				t.Fatalf("created service is not correct: %v", err) | ||||
| 			} | ||||
| 			t.Logf("Service created successfully: %+v", svc) | ||||
|  | ||||
| 			// Delete service | ||||
| 			if err := client.CoreV1().Services(ns.Name).Delete(ctx, svcNodePort.Name, metav1.DeleteOptions{}); err != nil { | ||||
| 			if err := client.CoreV1().Services(ns.Name).Delete(tCtx, svcNodePort.Name, metav1.DeleteOptions{}); err != nil { | ||||
| 				t.Fatalf("unexpected error deleting service: %v", err) | ||||
| 			} | ||||
| 			t.Logf("Deleted service: %s", svcNodePort.Name) | ||||
|  | ||||
| 			// Check that the service was not deleted and the IP is already allocated | ||||
| 			svc, err = client.CoreV1().Services(ns.Name).Get(ctx, svcNodePort.Name, metav1.GetOptions{}) | ||||
| 			svc, err = client.CoreV1().Services(ns.Name).Get(tCtx, svcNodePort.Name, metav1.GetOptions{}) | ||||
| 			if err != nil || | ||||
| 				svc.Spec.ClusterIP != clusterIP || | ||||
| 				int(svc.Spec.Ports[0].NodePort) != nodePort || | ||||
| @@ -214,18 +207,18 @@ func TestServicesFinalizersPatchStatus(t *testing.T) { | ||||
| 			} | ||||
|  | ||||
| 			if testcase == "spec" { | ||||
| 				if _, err = client.CoreV1().Services(ns.Name).Patch(ctx, svcNodePort.Name, types.StrategicMergePatchType, patchBytes, metav1.PatchOptions{}); err != nil { | ||||
| 				if _, err = client.CoreV1().Services(ns.Name).Patch(tCtx, svcNodePort.Name, types.StrategicMergePatchType, patchBytes, metav1.PatchOptions{}); err != nil { | ||||
| 					t.Fatalf("unexpected error removing finalizer: %v", err) | ||||
| 				} | ||||
| 			} else { | ||||
| 				if _, err = client.CoreV1().Services(ns.Name).Patch(ctx, svcNodePort.Name, types.StrategicMergePatchType, patchBytes, metav1.PatchOptions{}, "status"); err != nil { | ||||
| 				if _, err = client.CoreV1().Services(ns.Name).Patch(tCtx, svcNodePort.Name, types.StrategicMergePatchType, patchBytes, metav1.PatchOptions{}, "status"); err != nil { | ||||
| 					t.Fatalf("unexpected error removing finalizer: %v", err) | ||||
| 				} | ||||
| 			} | ||||
| 			t.Logf("Removed service finalizer: %s", svcNodePort.Name) | ||||
|  | ||||
| 			// Check that the service was deleted | ||||
| 			_, err = client.CoreV1().Services(ns.Name).Get(ctx, svcNodePort.Name, metav1.GetOptions{}) | ||||
| 			_, err = client.CoreV1().Services(ns.Name).Get(tCtx, svcNodePort.Name, metav1.GetOptions{}) | ||||
| 			if err == nil { | ||||
| 				t.Fatalf("service was not delete: %v", err) | ||||
| 			} | ||||
| @@ -233,11 +226,11 @@ func TestServicesFinalizersPatchStatus(t *testing.T) { | ||||
| 			// Try to create service again without the finalizer to check the ClusterIP and NodePort are deallocated | ||||
| 			svc = svcNodePort.DeepCopy() | ||||
| 			svc.Finalizers = []string{} | ||||
| 			if _, err := client.CoreV1().Services(ns.Name).Create(ctx, svc, metav1.CreateOptions{}); err != nil { | ||||
| 			if _, err := client.CoreV1().Services(ns.Name).Create(tCtx, svc, metav1.CreateOptions{}); err != nil { | ||||
| 				t.Fatalf("unexpected error creating service: %v", err) | ||||
| 			} | ||||
| 			// Delete service | ||||
| 			if err := client.CoreV1().Services(ns.Name).Delete(ctx, svc.Name, metav1.DeleteOptions{}); err != nil { | ||||
| 			if err := client.CoreV1().Services(ns.Name).Delete(tCtx, svc.Name, metav1.DeleteOptions{}); err != nil { | ||||
| 				t.Fatalf("unexpected error deleting service: %v", err) | ||||
| 			} | ||||
| 		}) | ||||
| @@ -248,11 +241,8 @@ func TestServicesFinalizersPatchStatus(t *testing.T) { | ||||
| func TestServiceCIDR28bits(t *testing.T) { | ||||
| 	serviceCIDR := "10.0.0.0/28" | ||||
|  | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
|  | ||||
| 	client, _, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 	tCtx := ktesting.Init(t) | ||||
| 	client, _, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 		ModifyServerRunOptions: func(opts *options.ServerRunOptions) { | ||||
| 			opts.ServiceClusterIPRanges = serviceCIDR | ||||
| 		}, | ||||
| @@ -261,7 +251,7 @@ func TestServiceCIDR28bits(t *testing.T) { | ||||
|  | ||||
| 	// Wait until the default "kubernetes" service is created. | ||||
| 	if err := wait.Poll(250*time.Millisecond, time.Minute, func() (bool, error) { | ||||
| 		_, err := client.CoreV1().Services(metav1.NamespaceDefault).Get(ctx, "kubernetes", metav1.GetOptions{}) | ||||
| 		_, err := client.CoreV1().Services(metav1.NamespaceDefault).Get(tCtx, "kubernetes", metav1.GetOptions{}) | ||||
| 		if err != nil { | ||||
| 			return false, err | ||||
| 		} | ||||
| @@ -288,7 +278,7 @@ func TestServiceCIDR28bits(t *testing.T) { | ||||
| 		}, | ||||
| 	} | ||||
|  | ||||
| 	_, err := client.CoreV1().Services(ns.Name).Create(ctx, service, metav1.CreateOptions{}) | ||||
| 	_, err := client.CoreV1().Services(ns.Name).Create(tCtx, service, metav1.CreateOptions{}) | ||||
| 	if err != nil { | ||||
| 		t.Fatalf("Error creating test service: %v", err) | ||||
| 	} | ||||
|   | ||||
| @@ -60,9 +60,7 @@ const ( | ||||
| //	quota_test.go:100: Took 4.196205966s to scale up without quota | ||||
| //	quota_test.go:115: Took 12.021640372s to scale up with quota | ||||
| func TestQuota(t *testing.T) { | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
| 	ctx := ktesting.Init(t) | ||||
|  | ||||
| 	// Set up a API server | ||||
| 	_, kubeConfig, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
|   | ||||
| @@ -55,11 +55,8 @@ const ( | ||||
| ) | ||||
|  | ||||
| func TestServiceAccountAutoCreate(t *testing.T) { | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
|  | ||||
| 	c, _, stopFunc, _, err := startServiceAccountTestServerAndWaitForCaches(ctx, t) | ||||
| 	tCtx := ktesting.Init(t) | ||||
| 	c, _, stopFunc, _, err := startServiceAccountTestServerAndWaitForCaches(tCtx, t) | ||||
| 	defer stopFunc() | ||||
| 	if err != nil { | ||||
| 		t.Fatalf("failed to setup ServiceAccounts server: %v", err) | ||||
| @@ -68,7 +65,7 @@ func TestServiceAccountAutoCreate(t *testing.T) { | ||||
| 	ns := "test-service-account-creation" | ||||
|  | ||||
| 	// Create namespace | ||||
| 	_, err = c.CoreV1().Namespaces().Create(ctx, &v1.Namespace{ObjectMeta: metav1.ObjectMeta{Name: ns}}, metav1.CreateOptions{}) | ||||
| 	_, err = c.CoreV1().Namespaces().Create(tCtx, &v1.Namespace{ObjectMeta: metav1.ObjectMeta{Name: ns}}, metav1.CreateOptions{}) | ||||
| 	if err != nil { | ||||
| 		t.Fatalf("could not create namespace: %v", err) | ||||
| 	} | ||||
| @@ -80,7 +77,7 @@ func TestServiceAccountAutoCreate(t *testing.T) { | ||||
| 	} | ||||
|  | ||||
| 	// Delete service account | ||||
| 	err = c.CoreV1().ServiceAccounts(ns).Delete(ctx, defaultUser.Name, metav1.DeleteOptions{}) | ||||
| 	err = c.CoreV1().ServiceAccounts(ns).Delete(tCtx, defaultUser.Name, metav1.DeleteOptions{}) | ||||
| 	if err != nil { | ||||
| 		t.Fatalf("Could not delete default serviceaccount: %v", err) | ||||
| 	} | ||||
| @@ -96,11 +93,8 @@ func TestServiceAccountAutoCreate(t *testing.T) { | ||||
| } | ||||
|  | ||||
| func TestServiceAccountTokenAutoMount(t *testing.T) { | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
|  | ||||
| 	c, _, stopFunc, _, err := startServiceAccountTestServerAndWaitForCaches(ctx, t) | ||||
| 	tCtx := ktesting.Init(t) | ||||
| 	c, _, stopFunc, _, err := startServiceAccountTestServerAndWaitForCaches(tCtx, t) | ||||
| 	defer stopFunc() | ||||
| 	if err != nil { | ||||
| 		t.Fatalf("failed to setup ServiceAccounts server: %v", err) | ||||
| @@ -109,7 +103,7 @@ func TestServiceAccountTokenAutoMount(t *testing.T) { | ||||
| 	ns := "auto-mount-ns" | ||||
|  | ||||
| 	// Create "my" namespace | ||||
| 	_, err = c.CoreV1().Namespaces().Create(ctx, &v1.Namespace{ObjectMeta: metav1.ObjectMeta{Name: ns}}, metav1.CreateOptions{}) | ||||
| 	_, err = c.CoreV1().Namespaces().Create(tCtx, &v1.Namespace{ObjectMeta: metav1.ObjectMeta{Name: ns}}, metav1.CreateOptions{}) | ||||
| 	if err != nil && !apierrors.IsAlreadyExists(err) { | ||||
| 		t.Fatalf("could not create namespace: %v", err) | ||||
| 	} | ||||
| @@ -127,7 +121,7 @@ func TestServiceAccountTokenAutoMount(t *testing.T) { | ||||
| 		}, | ||||
| 	} | ||||
|  | ||||
| 	createdPod, err := c.CoreV1().Pods(ns).Create(ctx, &protoPod, metav1.CreateOptions{}) | ||||
| 	createdPod, err := c.CoreV1().Pods(ns).Create(tCtx, &protoPod, metav1.CreateOptions{}) | ||||
| 	if err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} | ||||
| @@ -142,11 +136,8 @@ func TestServiceAccountTokenAutoMount(t *testing.T) { | ||||
| } | ||||
|  | ||||
| func TestServiceAccountTokenAuthentication(t *testing.T) { | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
|  | ||||
| 	c, config, stopFunc, _, err := startServiceAccountTestServerAndWaitForCaches(ctx, t) | ||||
| 	tCtx := ktesting.Init(t) | ||||
| 	c, config, stopFunc, _, err := startServiceAccountTestServerAndWaitForCaches(tCtx, t) | ||||
| 	defer stopFunc() | ||||
| 	if err != nil { | ||||
| 		t.Fatalf("failed to setup ServiceAccounts server: %v", err) | ||||
| @@ -156,19 +147,19 @@ func TestServiceAccountTokenAuthentication(t *testing.T) { | ||||
| 	otherns := "other-ns" | ||||
|  | ||||
| 	// Create "my" namespace | ||||
| 	_, err = c.CoreV1().Namespaces().Create(ctx, &v1.Namespace{ObjectMeta: metav1.ObjectMeta{Name: myns}}, metav1.CreateOptions{}) | ||||
| 	_, err = c.CoreV1().Namespaces().Create(tCtx, &v1.Namespace{ObjectMeta: metav1.ObjectMeta{Name: myns}}, metav1.CreateOptions{}) | ||||
| 	if err != nil && !apierrors.IsAlreadyExists(err) { | ||||
| 		t.Fatalf("could not create namespace: %v", err) | ||||
| 	} | ||||
|  | ||||
| 	// Create "other" namespace | ||||
| 	_, err = c.CoreV1().Namespaces().Create(ctx, &v1.Namespace{ObjectMeta: metav1.ObjectMeta{Name: otherns}}, metav1.CreateOptions{}) | ||||
| 	_, err = c.CoreV1().Namespaces().Create(tCtx, &v1.Namespace{ObjectMeta: metav1.ObjectMeta{Name: otherns}}, metav1.CreateOptions{}) | ||||
| 	if err != nil && !apierrors.IsAlreadyExists(err) { | ||||
| 		t.Fatalf("could not create namespace: %v", err) | ||||
| 	} | ||||
|  | ||||
| 	// Create "ro" user in myns | ||||
| 	roSA, err := c.CoreV1().ServiceAccounts(myns).Create(ctx, &v1.ServiceAccount{ObjectMeta: metav1.ObjectMeta{Name: readOnlyServiceAccountName}}, metav1.CreateOptions{}) | ||||
| 	roSA, err := c.CoreV1().ServiceAccounts(myns).Create(tCtx, &v1.ServiceAccount{ObjectMeta: metav1.ObjectMeta{Name: readOnlyServiceAccountName}}, metav1.CreateOptions{}) | ||||
| 	if err != nil { | ||||
| 		t.Fatalf("Service Account not created: %v", err) | ||||
| 	} | ||||
| @@ -183,13 +174,13 @@ func TestServiceAccountTokenAuthentication(t *testing.T) { | ||||
| 	roClient := clientset.NewForConfigOrDie(&roClientConfig) | ||||
| 	doServiceAccountAPIRequests(t, roClient, myns, true, true, false) | ||||
| 	doServiceAccountAPIRequests(t, roClient, otherns, true, false, false) | ||||
| 	err = c.CoreV1().Secrets(myns).Delete(ctx, roTokenName, metav1.DeleteOptions{}) | ||||
| 	err = c.CoreV1().Secrets(myns).Delete(tCtx, roTokenName, metav1.DeleteOptions{}) | ||||
| 	if err != nil { | ||||
| 		t.Fatalf("could not delete token: %v", err) | ||||
| 	} | ||||
| 	// wait for delete to be observed and reacted to via watch | ||||
| 	err = wait.PollImmediate(100*time.Millisecond, 30*time.Second, func() (bool, error) { | ||||
| 		_, err := roClient.CoreV1().Secrets(myns).List(ctx, metav1.ListOptions{}) | ||||
| 		_, err := roClient.CoreV1().Secrets(myns).List(tCtx, metav1.ListOptions{}) | ||||
| 		if err == nil { | ||||
| 			t.Logf("token is still valid, waiting") | ||||
| 			return false, nil | ||||
| @@ -206,7 +197,7 @@ func TestServiceAccountTokenAuthentication(t *testing.T) { | ||||
| 	doServiceAccountAPIRequests(t, roClient, myns, false, false, false) | ||||
|  | ||||
| 	// Create "rw" user in myns | ||||
| 	rwSA, err := c.CoreV1().ServiceAccounts(myns).Create(ctx, &v1.ServiceAccount{ObjectMeta: metav1.ObjectMeta{Name: readWriteServiceAccountName}}, metav1.CreateOptions{}) | ||||
| 	rwSA, err := c.CoreV1().ServiceAccounts(myns).Create(tCtx, &v1.ServiceAccount{ObjectMeta: metav1.ObjectMeta{Name: readWriteServiceAccountName}}, metav1.CreateOptions{}) | ||||
| 	if err != nil { | ||||
| 		t.Fatalf("Service Account not created: %v", err) | ||||
| 	} | ||||
| @@ -223,11 +214,8 @@ func TestServiceAccountTokenAuthentication(t *testing.T) { | ||||
| } | ||||
|  | ||||
| func TestLegacyServiceAccountTokenTracking(t *testing.T) { | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
|  | ||||
| 	c, config, stopFunc, _, err := startServiceAccountTestServerAndWaitForCaches(ctx, t) | ||||
| 	tCtx := ktesting.Init(t) | ||||
| 	c, config, stopFunc, _, err := startServiceAccountTestServerAndWaitForCaches(tCtx, t) | ||||
| 	defer stopFunc() | ||||
| 	if err != nil { | ||||
| 		t.Fatalf("failed to setup ServiceAccounts server: %v", err) | ||||
| @@ -235,11 +223,11 @@ func TestLegacyServiceAccountTokenTracking(t *testing.T) { | ||||
|  | ||||
| 	// create service account | ||||
| 	myns := "auth-ns" | ||||
| 	_, err = c.CoreV1().Namespaces().Create(ctx, &v1.Namespace{ObjectMeta: metav1.ObjectMeta{Name: myns}}, metav1.CreateOptions{}) | ||||
| 	_, err = c.CoreV1().Namespaces().Create(tCtx, &v1.Namespace{ObjectMeta: metav1.ObjectMeta{Name: myns}}, metav1.CreateOptions{}) | ||||
| 	if err != nil && !apierrors.IsAlreadyExists(err) { | ||||
| 		t.Fatalf("could not create namespace: %v", err) | ||||
| 	} | ||||
| 	mysa, err := c.CoreV1().ServiceAccounts(myns).Create(ctx, &v1.ServiceAccount{ObjectMeta: metav1.ObjectMeta{Name: readOnlyServiceAccountName}}, metav1.CreateOptions{}) | ||||
| 	mysa, err := c.CoreV1().ServiceAccounts(myns).Create(tCtx, &v1.ServiceAccount{ObjectMeta: metav1.ObjectMeta{Name: readOnlyServiceAccountName}}, metav1.CreateOptions{}) | ||||
| 	if err != nil { | ||||
| 		t.Fatalf("Service Account not created: %v", err) | ||||
| 	} | ||||
| @@ -298,7 +286,7 @@ func TestLegacyServiceAccountTokenTracking(t *testing.T) { | ||||
| 			} | ||||
| 			wg.Wait() | ||||
| 			dateAfter := time.Now().UTC().Format(dateFormat) | ||||
| 			liveSecret, err := c.CoreV1().Secrets(myns).Get(ctx, test.secretName, metav1.GetOptions{}) | ||||
| 			liveSecret, err := c.CoreV1().Secrets(myns).Get(tCtx, test.secretName, metav1.GetOptions{}) | ||||
| 			if err != nil { | ||||
| 				t.Fatalf("Could not get secret: %v", err) | ||||
| 			} | ||||
| @@ -325,7 +313,7 @@ func TestLegacyServiceAccountTokenTracking(t *testing.T) { | ||||
| 	// configmap should exist with 'since' timestamp. | ||||
| 	if err = wait.PollImmediate(time.Millisecond*10, wait.ForeverTestTimeout, func() (bool, error) { | ||||
| 		dateBefore := time.Now().UTC().Format("2006-01-02") | ||||
| 		configMap, err := c.CoreV1().ConfigMaps(metav1.NamespaceSystem).Get(ctx, legacytokentracking.ConfigMapName, metav1.GetOptions{}) | ||||
| 		configMap, err := c.CoreV1().ConfigMaps(metav1.NamespaceSystem).Get(tCtx, legacytokentracking.ConfigMapName, metav1.GetOptions{}) | ||||
| 		if err != nil { | ||||
| 			return false, fmt.Errorf("failed to get %q configmap, err %w", legacytokentracking.ConfigMapDataKey, err) | ||||
| 		} | ||||
|   | ||||
| @@ -44,11 +44,8 @@ func TestServiceAlloc(t *testing.T) { | ||||
| 	// Create an IPv4 single stack control-plane | ||||
| 	serviceCIDR := "192.168.0.0/29" | ||||
|  | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
|  | ||||
| 	client, _, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 	tCtx := ktesting.Init(t) | ||||
| 	client, _, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 		ModifyServerRunOptions: func(opts *options.ServerRunOptions) { | ||||
| 			opts.ServiceClusterIPRanges = serviceCIDR | ||||
| 		}, | ||||
| @@ -120,11 +117,8 @@ func TestServiceAllocIPAddress(t *testing.T) { | ||||
| 	serviceCIDR := "2001:db8::/64" | ||||
| 	defer featuregatetesting.SetFeatureGateDuringTest(t, utilfeature.DefaultFeatureGate, features.MultiCIDRServiceAllocator, true)() | ||||
|  | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	defer cancel() | ||||
|  | ||||
| 	client, _, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 	tCtx := ktesting.Init(t) | ||||
| 	client, _, tearDownFn := framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 		ModifyServerRunOptions: func(opts *options.ServerRunOptions) { | ||||
| 			opts.ServiceClusterIPRanges = serviceCIDR | ||||
| 			opts.GenericServerRunOptions.AdvertiseAddress = netutils.ParseIPSloppy("2001:db8::10") | ||||
| @@ -149,7 +143,7 @@ func TestServiceAllocIPAddress(t *testing.T) { | ||||
|  | ||||
| 	// Wait until the default "kubernetes" service is created. | ||||
| 	if err := wait.Poll(250*time.Millisecond, time.Minute, func() (bool, error) { | ||||
| 		_, err := client.CoreV1().Services(metav1.NamespaceDefault).Get(ctx, "kubernetes", metav1.GetOptions{}) | ||||
| 		_, err := client.CoreV1().Services(metav1.NamespaceDefault).Get(tCtx, "kubernetes", metav1.GetOptions{}) | ||||
| 		if err != nil && !apierrors.IsNotFound(err) { | ||||
| 			return false, err | ||||
| 		} | ||||
| @@ -160,11 +154,11 @@ func TestServiceAllocIPAddress(t *testing.T) { | ||||
|  | ||||
| 	// create 5 random services and check that the Services have an IP associated | ||||
| 	for i := 0; i < 5; i++ { | ||||
| 		svc, err := client.CoreV1().Services(metav1.NamespaceDefault).Create(ctx, svc(i), metav1.CreateOptions{}) | ||||
| 		svc, err := client.CoreV1().Services(metav1.NamespaceDefault).Create(tCtx, svc(i), metav1.CreateOptions{}) | ||||
| 		if err != nil { | ||||
| 			t.Error(err) | ||||
| 		} | ||||
| 		_, err = client.NetworkingV1alpha1().IPAddresses().Get(ctx, svc.Spec.ClusterIP, metav1.GetOptions{}) | ||||
| 		_, err = client.NetworkingV1alpha1().IPAddresses().Get(tCtx, svc.Spec.ClusterIP, metav1.GetOptions{}) | ||||
| 		if err != nil { | ||||
| 			t.Error(err) | ||||
| 		} | ||||
|   | ||||
| @@ -504,11 +504,10 @@ func UpdateNodeStatus(cs clientset.Interface, node *v1.Node) error { | ||||
| // It registers cleanup functions to t.Cleanup(), they will be called when the test completes, | ||||
| // no need to do this again. | ||||
| func InitTestAPIServer(t *testing.T, nsPrefix string, admission admission.Interface) *TestContext { | ||||
| 	_, ctx := ktesting.NewTestContext(t) | ||||
| 	ctx, cancel := context.WithCancel(ctx) | ||||
| 	testCtx := &TestContext{Ctx: ctx} | ||||
| 	tCtx := ktesting.Init(t) | ||||
| 	testCtx := &TestContext{Ctx: tCtx} | ||||
|  | ||||
| 	testCtx.ClientSet, testCtx.KubeConfig, testCtx.CloseFn = framework.StartTestServer(ctx, t, framework.TestServerSetup{ | ||||
| 	testCtx.ClientSet, testCtx.KubeConfig, testCtx.CloseFn = framework.StartTestServer(tCtx, t, framework.TestServerSetup{ | ||||
| 		ModifyServerRunOptions: func(options *options.ServerRunOptions) { | ||||
| 			options.Admission.GenericAdmission.DisablePlugins = []string{"ServiceAccount", "TaintNodesByCondition", "Priority", "StorageObjectInUseProtection"} | ||||
| 			if utilfeature.DefaultFeatureGate.Enabled(features.DynamicResourceAllocation) { | ||||
| @@ -536,7 +535,7 @@ func InitTestAPIServer(t *testing.T, nsPrefix string, admission admission.Interf | ||||
|  | ||||
| 	oldCloseFn := testCtx.CloseFn | ||||
| 	testCtx.CloseFn = func() { | ||||
| 		cancel() | ||||
| 		tCtx.Cancel("tearing down apiserver") | ||||
| 		oldCloseFn() | ||||
| 	} | ||||
|  | ||||
|   | ||||
| @@ -17,7 +17,6 @@ limitations under the License. | ||||
| package ktesting | ||||
|  | ||||
| import ( | ||||
| 	"context" | ||||
| 	"flag" | ||||
| 	"fmt" | ||||
| 	"testing" | ||||
| @@ -41,14 +40,9 @@ func SetDefaultVerbosity(v int) { | ||||
| 	_ = f.Value.Set(fmt.Sprintf("%d", v)) | ||||
| } | ||||
|  | ||||
| // NewTestContext is a drop-in replacement for ktesting.NewTestContext | ||||
| // NewTestContext is a replacement for ktesting.NewTestContext | ||||
| // which returns a more versatile context. | ||||
| // | ||||
| // The type of that context is still context.Context because replacing | ||||
| // it with TContext breaks tests which use `WithCancel`. | ||||
| // | ||||
| // TODO(pohly): change all of that code together with changing the return type. | ||||
| func NewTestContext(tb testing.TB) (klog.Logger, context.Context) { | ||||
| func NewTestContext(tb testing.TB) (klog.Logger, TContext) { | ||||
| 	tCtx := Init(tb) | ||||
| 	return tCtx.Logger(), tCtx | ||||
| } | ||||
|   | ||||
		Reference in New Issue
	
	Block a user
	 Kubernetes Prow Robot
					Kubernetes Prow Robot