Fix some assertions for integ tests
Signed-off-by: Henry Wang <henwang@amazon.com>
This commit is contained in:
		| @@ -252,7 +252,7 @@ func TestContainerListStatsWithIdFilter(t *testing.T) { | ||||
|  | ||||
| 		t.Logf("Verify container stats for %s", id) | ||||
| 		for _, s := range stats { | ||||
| 			require.Equal(t, s.GetAttributes().GetId(), id) | ||||
| 			require.Equal(t, id, s.GetAttributes().GetId()) | ||||
| 			testStats(t, s, containerConfigMap[id]) | ||||
| 		} | ||||
| 	} | ||||
| @@ -406,9 +406,9 @@ func testStats(t *testing.T, | ||||
| 	require.NotEmpty(t, s.GetAttributes().GetId()) | ||||
| 	require.NotEmpty(t, s.GetAttributes().GetMetadata()) | ||||
| 	require.NotEmpty(t, s.GetAttributes().GetAnnotations()) | ||||
| 	require.Equal(t, s.GetAttributes().GetLabels(), config.Labels) | ||||
| 	require.Equal(t, s.GetAttributes().GetAnnotations(), config.Annotations) | ||||
| 	require.Equal(t, s.GetAttributes().GetMetadata().Name, config.Metadata.Name) | ||||
| 	require.Equal(t, config.Labels, s.GetAttributes().GetLabels()) | ||||
| 	require.Equal(t, config.Annotations, s.GetAttributes().GetAnnotations()) | ||||
| 	require.Equal(t, config.Metadata.Name, s.GetAttributes().GetMetadata().Name) | ||||
| 	require.NotEmpty(t, s.GetAttributes().GetLabels()) | ||||
| 	require.NotEmpty(t, s.GetCpu().GetTimestamp()) | ||||
| 	require.NotEmpty(t, s.GetCpu().GetUsageCoreNanoSeconds().GetValue()) | ||||
|   | ||||
| @@ -67,7 +67,7 @@ func TestSharedPidMultiProcessContainerStop(t *testing.T) { | ||||
| 			t.Log("The container state should be exited") | ||||
| 			s, err := runtimeService.ContainerStatus(cn) | ||||
| 			require.NoError(t, err) | ||||
| 			assert.Equal(t, s.GetState(), runtime.ContainerState_CONTAINER_EXITED) | ||||
| 			assert.Equal(t, runtime.ContainerState_CONTAINER_EXITED, s.GetState()) | ||||
| 		}) | ||||
| 	} | ||||
| } | ||||
| @@ -126,5 +126,5 @@ func TestContainerStopCancellation(t *testing.T) { | ||||
| 	t.Log("The container state should be exited") | ||||
| 	s, err := runtimeService.ContainerStatus(cn) | ||||
| 	require.NoError(t, err) | ||||
| 	assert.Equal(t, s.GetState(), runtime.ContainerState_CONTAINER_EXITED) | ||||
| 	assert.Equal(t, runtime.ContainerState_CONTAINER_EXITED, s.GetState()) | ||||
| } | ||||
|   | ||||
| @@ -53,7 +53,7 @@ func TestContainerLifecycleWithoutImageRef(t *testing.T) { | ||||
| 	t.Log("Container status should be running") | ||||
| 	status, err := runtimeService.ContainerStatus(cn) | ||||
| 	require.NoError(t, err) | ||||
| 	assert.Equal(t, status.GetState(), runtime.ContainerState_CONTAINER_RUNNING) | ||||
| 	assert.Equal(t, runtime.ContainerState_CONTAINER_RUNNING, status.GetState()) | ||||
|  | ||||
| 	t.Logf("Stop container") | ||||
| 	err = runtimeService.StopContainer(cn, 1) | ||||
| @@ -62,5 +62,5 @@ func TestContainerLifecycleWithoutImageRef(t *testing.T) { | ||||
| 	t.Log("Container status should be exited") | ||||
| 	status, err = runtimeService.ContainerStatus(cn) | ||||
| 	require.NoError(t, err) | ||||
| 	assert.Equal(t, status.GetState(), runtime.ContainerState_CONTAINER_EXITED) | ||||
| 	assert.Equal(t, runtime.ContainerState_CONTAINER_EXITED, status.GetState()) | ||||
| } | ||||
|   | ||||
| @@ -117,18 +117,18 @@ func TestContainerdImage(t *testing.T) { | ||||
| 	t.Logf("the image should be marked as managed") | ||||
| 	imgByRef, err := containerdClient.GetImage(ctx, testImage) | ||||
| 	assert.NoError(t, err) | ||||
| 	assert.Equal(t, imgByRef.Labels()["io.cri-containerd.image"], "managed") | ||||
| 	assert.Equal(t, "managed", imgByRef.Labels()["io.cri-containerd.image"]) | ||||
|  | ||||
| 	t.Logf("the image id should be created and managed") | ||||
| 	imgByID, err := containerdClient.GetImage(ctx, id) | ||||
| 	assert.NoError(t, err) | ||||
| 	assert.Equal(t, imgByID.Labels()["io.cri-containerd.image"], "managed") | ||||
| 	assert.Equal(t, "managed", imgByID.Labels()["io.cri-containerd.image"]) | ||||
|  | ||||
| 	t.Logf("the image should be labeled") | ||||
| 	img, err := containerdClient.GetImage(ctx, testImage) | ||||
| 	assert.NoError(t, err) | ||||
| 	assert.Equal(t, img.Labels()["foo"], "bar") | ||||
| 	assert.Equal(t, img.Labels()[labels.ImageLabelKey], labels.ImageLabelValue) | ||||
| 	assert.Equal(t, "bar", img.Labels()["foo"]) | ||||
| 	assert.Equal(t, labels.ImageLabelValue, img.Labels()[labels.ImageLabelKey]) | ||||
|  | ||||
| 	t.Logf("the image should be pinned") | ||||
| 	i, err = imageService.ImageStatus(&runtime.ImageSpec{Image: testImage}) | ||||
| @@ -225,7 +225,7 @@ func TestContainerdSandboxImage(t *testing.T) { | ||||
| 	pauseImg, err := containerdClient.GetImage(ctx, pauseImage) | ||||
| 	require.NoError(t, err) | ||||
| 	t.Log("ensure correct labels are set on pause image") | ||||
| 	assert.Equal(t, pauseImg.Labels()["io.cri-containerd.pinned"], "pinned") | ||||
| 	assert.Equal(t, "pinned", pauseImg.Labels()["io.cri-containerd.pinned"]) | ||||
|  | ||||
| 	t.Log("pause image should be seen by cri plugin") | ||||
| 	pimg, err := imageService.ImageStatus(&runtime.ImageSpec{Image: pauseImage}) | ||||
|   | ||||
| @@ -271,12 +271,12 @@ func testCRIImagePullTimeoutByNoDataTransferred(t *testing.T) { | ||||
| 			}, | ||||
| 		}) | ||||
|  | ||||
| 		assert.Equal(t, errors.Unwrap(err), context.Canceled, "[%v] expected canceled error, but got (%v)", idx, err) | ||||
| 		assert.Equal(t, mirrorSrv.limiter.clearHitCircuitBreaker(), true, "[%v] expected to hit circuit breaker", idx) | ||||
| 		assert.Equal(t, context.Canceled, errors.Unwrap(err), "[%v] expected canceled error, but got (%v)", idx, err) | ||||
| 		assert.True(t, mirrorSrv.limiter.clearHitCircuitBreaker(), "[%v] expected to hit circuit breaker", idx) | ||||
|  | ||||
| 		// cleanup the temp data by sync delete | ||||
| 		lid, ok := leases.FromContext(dctx) | ||||
| 		assert.Equal(t, ok, true) | ||||
| 		assert.True(t, ok) | ||||
| 		err = cli.LeasesService().Delete(ctx, leases.Lease{ID: lid}, leases.SynchronousDelete) | ||||
| 		assert.NoError(t, err) | ||||
| 	} | ||||
|   | ||||
| @@ -67,7 +67,7 @@ func TestRunPodSandboxWithSetupCNIFailure(t *testing.T) { | ||||
| 	t.Logf("Create a sandbox") | ||||
| 	_, err := runtimeService.RunPodSandbox(sbConfig, failpointRuntimeHandler) | ||||
| 	require.Error(t, err) | ||||
| 	require.Equal(t, true, strings.Contains(err.Error(), "you-shall-not-pass!")) | ||||
| 	require.ErrorContains(t, err, "you-shall-not-pass!") | ||||
|  | ||||
| 	t.Logf("Retry to create sandbox with same config") | ||||
| 	sb, err := runtimeService.RunPodSandbox(sbConfig, failpointRuntimeHandler) | ||||
| @@ -95,7 +95,7 @@ func TestRunPodSandboxWithShimStartFailure(t *testing.T) { | ||||
| 	t.Logf("Create a sandbox") | ||||
| 	_, err := runtimeService.RunPodSandbox(sbConfig, failpointRuntimeHandler) | ||||
| 	require.Error(t, err) | ||||
| 	require.Equal(t, true, strings.Contains(err.Error(), "no hard feelings")) | ||||
| 	require.ErrorContains(t, err, "no hard feelings") | ||||
| } | ||||
|  | ||||
| // TestRunPodSandboxWithShimDeleteFailure should keep the sandbox record if | ||||
| @@ -130,16 +130,16 @@ func TestRunPodSandboxWithShimDeleteFailure(t *testing.T) { | ||||
| 			require.Len(t, l, 1) | ||||
|  | ||||
| 			sb := l[0] | ||||
| 			require.Equal(t, sb.State, criapiv1.PodSandboxState_SANDBOX_NOTREADY) | ||||
| 			require.Equal(t, sb.Metadata.Name, sbConfig.Metadata.Name) | ||||
| 			require.Equal(t, sb.Metadata.Namespace, sbConfig.Metadata.Namespace) | ||||
| 			require.Equal(t, sb.Metadata.Uid, sbConfig.Metadata.Uid) | ||||
| 			require.Equal(t, sb.Metadata.Attempt, sbConfig.Metadata.Attempt) | ||||
| 			require.Equal(t, criapiv1.PodSandboxState_SANDBOX_NOTREADY, sb.State) | ||||
| 			require.Equal(t, sbConfig.Metadata.Name, sb.Metadata.Name) | ||||
| 			require.Equal(t, sbConfig.Metadata.Namespace, sb.Metadata.Namespace) | ||||
| 			require.Equal(t, sbConfig.Metadata.Uid, sb.Metadata.Uid) | ||||
| 			require.Equal(t, sbConfig.Metadata.Attempt, sb.Metadata.Attempt) | ||||
|  | ||||
| 			t.Log("Check PodSandboxStatus") | ||||
| 			sbStatus, err := runtimeService.PodSandboxStatus(sb.Id) | ||||
| 			require.NoError(t, err) | ||||
| 			require.Equal(t, sbStatus.State, criapiv1.PodSandboxState_SANDBOX_NOTREADY) | ||||
| 			require.Equal(t, criapiv1.PodSandboxState_SANDBOX_NOTREADY, sbStatus.State) | ||||
| 			require.Greater(t, len(sbStatus.Network.Ip), 0) | ||||
|  | ||||
| 			if restart { | ||||
| @@ -150,13 +150,13 @@ func TestRunPodSandboxWithShimDeleteFailure(t *testing.T) { | ||||
| 				l, err = runtimeService.ListPodSandbox(&criapiv1.PodSandboxFilter{Id: sb.Id}) | ||||
| 				require.NoError(t, err) | ||||
| 				require.Len(t, l, 1) | ||||
| 				require.Equal(t, l[0].State, criapiv1.PodSandboxState_SANDBOX_NOTREADY) | ||||
| 				require.Equal(t, criapiv1.PodSandboxState_SANDBOX_NOTREADY, l[0].State) | ||||
|  | ||||
| 				t.Log("Check PodSandboxStatus") | ||||
| 				sbStatus, err := runtimeService.PodSandboxStatus(sb.Id) | ||||
| 				require.NoError(t, err) | ||||
| 				t.Log(sbStatus.Network) | ||||
| 				require.Equal(t, sbStatus.State, criapiv1.PodSandboxState_SANDBOX_NOTREADY) | ||||
| 				require.Equal(t, criapiv1.PodSandboxState_SANDBOX_NOTREADY, sbStatus.State) | ||||
| 			} | ||||
|  | ||||
| 			t.Log("Cleanup leaky sandbox") | ||||
| @@ -206,11 +206,11 @@ func TestRunPodSandboxWithShimStartAndTeardownCNIFailure(t *testing.T) { | ||||
| 			require.Len(t, l, 1) | ||||
|  | ||||
| 			sb := l[0] | ||||
| 			require.Equal(t, sb.State, criapiv1.PodSandboxState_SANDBOX_NOTREADY) | ||||
| 			require.Equal(t, sb.Metadata.Name, sbConfig.Metadata.Name) | ||||
| 			require.Equal(t, sb.Metadata.Namespace, sbConfig.Metadata.Namespace) | ||||
| 			require.Equal(t, sb.Metadata.Uid, sbConfig.Metadata.Uid) | ||||
| 			require.Equal(t, sb.Metadata.Attempt, sbConfig.Metadata.Attempt) | ||||
| 			require.Equal(t, criapiv1.PodSandboxState_SANDBOX_NOTREADY, sb.State) | ||||
| 			require.Equal(t, sbConfig.Metadata.Name, sb.Metadata.Name) | ||||
| 			require.Equal(t, sbConfig.Metadata.Namespace, sb.Metadata.Namespace) | ||||
| 			require.Equal(t, sbConfig.Metadata.Uid, sb.Metadata.Uid) | ||||
| 			require.Equal(t, sbConfig.Metadata.Attempt, sb.Metadata.Attempt) | ||||
|  | ||||
| 			if restart { | ||||
| 				t.Log("Restart containerd") | ||||
| @@ -220,7 +220,7 @@ func TestRunPodSandboxWithShimStartAndTeardownCNIFailure(t *testing.T) { | ||||
| 				l, err = runtimeService.ListPodSandbox(&criapiv1.PodSandboxFilter{Id: sb.Id}) | ||||
| 				require.NoError(t, err) | ||||
| 				require.Len(t, l, 1) | ||||
| 				require.Equal(t, l[0].State, criapiv1.PodSandboxState_SANDBOX_NOTREADY) | ||||
| 				require.Equal(t, criapiv1.PodSandboxState_SANDBOX_NOTREADY, l[0].State) | ||||
| 			} | ||||
|  | ||||
| 			t.Log("Cleanup leaky sandbox") | ||||
| @@ -287,11 +287,11 @@ func TestRunPodSandboxAndTeardownCNISlow(t *testing.T) { | ||||
| 		require.NoError(t, err) | ||||
| 	}() | ||||
|  | ||||
| 	assert.Equal(t, sb.State, criapiv1.PodSandboxState_SANDBOX_NOTREADY) | ||||
| 	assert.Equal(t, sb.Metadata.Name, sbConfig.Metadata.Name) | ||||
| 	assert.Equal(t, sb.Metadata.Namespace, sbConfig.Metadata.Namespace) | ||||
| 	assert.Equal(t, sb.Metadata.Uid, sbConfig.Metadata.Uid) | ||||
| 	assert.Equal(t, sb.Metadata.Attempt, sbConfig.Metadata.Attempt) | ||||
| 	assert.Equal(t, criapiv1.PodSandboxState_SANDBOX_NOTREADY, sb.State) | ||||
| 	assert.Equal(t, sbConfig.Metadata.Name, sb.Metadata.Name) | ||||
| 	assert.Equal(t, sbConfig.Metadata.Namespace, sb.Metadata.Namespace) | ||||
| 	assert.Equal(t, sbConfig.Metadata.Uid, sb.Metadata.Uid) | ||||
| 	assert.Equal(t, sbConfig.Metadata.Attempt, sb.Metadata.Attempt) | ||||
|  | ||||
| 	if os.Getenv("DISABLE_CRI_SANDBOXES") != "" { | ||||
| 		// non-sbserver | ||||
|   | ||||
| @@ -42,7 +42,7 @@ func TestTruncIndex(t *testing.T) { | ||||
| 	t.Logf("Get image status by truncindex, truncID: %s", imgTruncID) | ||||
| 	res, err := imageService.ImageStatus(&runtimeapi.ImageSpec{Image: imgTruncID}) | ||||
| 	require.NoError(t, err) | ||||
| 	require.NotEqual(t, nil, res) | ||||
| 	require.NotNil(t, res) | ||||
| 	assert.Equal(t, imgID, res.Id) | ||||
|  | ||||
| 	// TODO(yanxuean): for failure test case where there are two images with the same truncindex. | ||||
|   | ||||
		Reference in New Issue
	
	Block a user
	 Henry Wang
					Henry Wang