Fix some assertions for integ tests
Signed-off-by: Henry Wang <henwang@amazon.com>
This commit is contained in:
		| @@ -252,7 +252,7 @@ func TestContainerListStatsWithIdFilter(t *testing.T) { | |||||||
|  |  | ||||||
| 		t.Logf("Verify container stats for %s", id) | 		t.Logf("Verify container stats for %s", id) | ||||||
| 		for _, s := range stats { | 		for _, s := range stats { | ||||||
| 			require.Equal(t, s.GetAttributes().GetId(), id) | 			require.Equal(t, id, s.GetAttributes().GetId()) | ||||||
| 			testStats(t, s, containerConfigMap[id]) | 			testStats(t, s, containerConfigMap[id]) | ||||||
| 		} | 		} | ||||||
| 	} | 	} | ||||||
| @@ -406,9 +406,9 @@ func testStats(t *testing.T, | |||||||
| 	require.NotEmpty(t, s.GetAttributes().GetId()) | 	require.NotEmpty(t, s.GetAttributes().GetId()) | ||||||
| 	require.NotEmpty(t, s.GetAttributes().GetMetadata()) | 	require.NotEmpty(t, s.GetAttributes().GetMetadata()) | ||||||
| 	require.NotEmpty(t, s.GetAttributes().GetAnnotations()) | 	require.NotEmpty(t, s.GetAttributes().GetAnnotations()) | ||||||
| 	require.Equal(t, s.GetAttributes().GetLabels(), config.Labels) | 	require.Equal(t, config.Labels, s.GetAttributes().GetLabels()) | ||||||
| 	require.Equal(t, s.GetAttributes().GetAnnotations(), config.Annotations) | 	require.Equal(t, config.Annotations, s.GetAttributes().GetAnnotations()) | ||||||
| 	require.Equal(t, s.GetAttributes().GetMetadata().Name, config.Metadata.Name) | 	require.Equal(t, config.Metadata.Name, s.GetAttributes().GetMetadata().Name) | ||||||
| 	require.NotEmpty(t, s.GetAttributes().GetLabels()) | 	require.NotEmpty(t, s.GetAttributes().GetLabels()) | ||||||
| 	require.NotEmpty(t, s.GetCpu().GetTimestamp()) | 	require.NotEmpty(t, s.GetCpu().GetTimestamp()) | ||||||
| 	require.NotEmpty(t, s.GetCpu().GetUsageCoreNanoSeconds().GetValue()) | 	require.NotEmpty(t, s.GetCpu().GetUsageCoreNanoSeconds().GetValue()) | ||||||
|   | |||||||
| @@ -67,7 +67,7 @@ func TestSharedPidMultiProcessContainerStop(t *testing.T) { | |||||||
| 			t.Log("The container state should be exited") | 			t.Log("The container state should be exited") | ||||||
| 			s, err := runtimeService.ContainerStatus(cn) | 			s, err := runtimeService.ContainerStatus(cn) | ||||||
| 			require.NoError(t, err) | 			require.NoError(t, err) | ||||||
| 			assert.Equal(t, s.GetState(), runtime.ContainerState_CONTAINER_EXITED) | 			assert.Equal(t, runtime.ContainerState_CONTAINER_EXITED, s.GetState()) | ||||||
| 		}) | 		}) | ||||||
| 	} | 	} | ||||||
| } | } | ||||||
| @@ -126,5 +126,5 @@ func TestContainerStopCancellation(t *testing.T) { | |||||||
| 	t.Log("The container state should be exited") | 	t.Log("The container state should be exited") | ||||||
| 	s, err := runtimeService.ContainerStatus(cn) | 	s, err := runtimeService.ContainerStatus(cn) | ||||||
| 	require.NoError(t, err) | 	require.NoError(t, err) | ||||||
| 	assert.Equal(t, s.GetState(), runtime.ContainerState_CONTAINER_EXITED) | 	assert.Equal(t, runtime.ContainerState_CONTAINER_EXITED, s.GetState()) | ||||||
| } | } | ||||||
|   | |||||||
| @@ -53,7 +53,7 @@ func TestContainerLifecycleWithoutImageRef(t *testing.T) { | |||||||
| 	t.Log("Container status should be running") | 	t.Log("Container status should be running") | ||||||
| 	status, err := runtimeService.ContainerStatus(cn) | 	status, err := runtimeService.ContainerStatus(cn) | ||||||
| 	require.NoError(t, err) | 	require.NoError(t, err) | ||||||
| 	assert.Equal(t, status.GetState(), runtime.ContainerState_CONTAINER_RUNNING) | 	assert.Equal(t, runtime.ContainerState_CONTAINER_RUNNING, status.GetState()) | ||||||
|  |  | ||||||
| 	t.Logf("Stop container") | 	t.Logf("Stop container") | ||||||
| 	err = runtimeService.StopContainer(cn, 1) | 	err = runtimeService.StopContainer(cn, 1) | ||||||
| @@ -62,5 +62,5 @@ func TestContainerLifecycleWithoutImageRef(t *testing.T) { | |||||||
| 	t.Log("Container status should be exited") | 	t.Log("Container status should be exited") | ||||||
| 	status, err = runtimeService.ContainerStatus(cn) | 	status, err = runtimeService.ContainerStatus(cn) | ||||||
| 	require.NoError(t, err) | 	require.NoError(t, err) | ||||||
| 	assert.Equal(t, status.GetState(), runtime.ContainerState_CONTAINER_EXITED) | 	assert.Equal(t, runtime.ContainerState_CONTAINER_EXITED, status.GetState()) | ||||||
| } | } | ||||||
|   | |||||||
| @@ -117,18 +117,18 @@ func TestContainerdImage(t *testing.T) { | |||||||
| 	t.Logf("the image should be marked as managed") | 	t.Logf("the image should be marked as managed") | ||||||
| 	imgByRef, err := containerdClient.GetImage(ctx, testImage) | 	imgByRef, err := containerdClient.GetImage(ctx, testImage) | ||||||
| 	assert.NoError(t, err) | 	assert.NoError(t, err) | ||||||
| 	assert.Equal(t, imgByRef.Labels()["io.cri-containerd.image"], "managed") | 	assert.Equal(t, "managed", imgByRef.Labels()["io.cri-containerd.image"]) | ||||||
|  |  | ||||||
| 	t.Logf("the image id should be created and managed") | 	t.Logf("the image id should be created and managed") | ||||||
| 	imgByID, err := containerdClient.GetImage(ctx, id) | 	imgByID, err := containerdClient.GetImage(ctx, id) | ||||||
| 	assert.NoError(t, err) | 	assert.NoError(t, err) | ||||||
| 	assert.Equal(t, imgByID.Labels()["io.cri-containerd.image"], "managed") | 	assert.Equal(t, "managed", imgByID.Labels()["io.cri-containerd.image"]) | ||||||
|  |  | ||||||
| 	t.Logf("the image should be labeled") | 	t.Logf("the image should be labeled") | ||||||
| 	img, err := containerdClient.GetImage(ctx, testImage) | 	img, err := containerdClient.GetImage(ctx, testImage) | ||||||
| 	assert.NoError(t, err) | 	assert.NoError(t, err) | ||||||
| 	assert.Equal(t, img.Labels()["foo"], "bar") | 	assert.Equal(t, "bar", img.Labels()["foo"]) | ||||||
| 	assert.Equal(t, img.Labels()[labels.ImageLabelKey], labels.ImageLabelValue) | 	assert.Equal(t, labels.ImageLabelValue, img.Labels()[labels.ImageLabelKey]) | ||||||
|  |  | ||||||
| 	t.Logf("the image should be pinned") | 	t.Logf("the image should be pinned") | ||||||
| 	i, err = imageService.ImageStatus(&runtime.ImageSpec{Image: testImage}) | 	i, err = imageService.ImageStatus(&runtime.ImageSpec{Image: testImage}) | ||||||
| @@ -225,7 +225,7 @@ func TestContainerdSandboxImage(t *testing.T) { | |||||||
| 	pauseImg, err := containerdClient.GetImage(ctx, pauseImage) | 	pauseImg, err := containerdClient.GetImage(ctx, pauseImage) | ||||||
| 	require.NoError(t, err) | 	require.NoError(t, err) | ||||||
| 	t.Log("ensure correct labels are set on pause image") | 	t.Log("ensure correct labels are set on pause image") | ||||||
| 	assert.Equal(t, pauseImg.Labels()["io.cri-containerd.pinned"], "pinned") | 	assert.Equal(t, "pinned", pauseImg.Labels()["io.cri-containerd.pinned"]) | ||||||
|  |  | ||||||
| 	t.Log("pause image should be seen by cri plugin") | 	t.Log("pause image should be seen by cri plugin") | ||||||
| 	pimg, err := imageService.ImageStatus(&runtime.ImageSpec{Image: pauseImage}) | 	pimg, err := imageService.ImageStatus(&runtime.ImageSpec{Image: pauseImage}) | ||||||
|   | |||||||
| @@ -271,12 +271,12 @@ func testCRIImagePullTimeoutByNoDataTransferred(t *testing.T) { | |||||||
| 			}, | 			}, | ||||||
| 		}) | 		}) | ||||||
|  |  | ||||||
| 		assert.Equal(t, errors.Unwrap(err), context.Canceled, "[%v] expected canceled error, but got (%v)", idx, err) | 		assert.Equal(t, context.Canceled, errors.Unwrap(err), "[%v] expected canceled error, but got (%v)", idx, err) | ||||||
| 		assert.Equal(t, mirrorSrv.limiter.clearHitCircuitBreaker(), true, "[%v] expected to hit circuit breaker", idx) | 		assert.True(t, mirrorSrv.limiter.clearHitCircuitBreaker(), "[%v] expected to hit circuit breaker", idx) | ||||||
|  |  | ||||||
| 		// cleanup the temp data by sync delete | 		// cleanup the temp data by sync delete | ||||||
| 		lid, ok := leases.FromContext(dctx) | 		lid, ok := leases.FromContext(dctx) | ||||||
| 		assert.Equal(t, ok, true) | 		assert.True(t, ok) | ||||||
| 		err = cli.LeasesService().Delete(ctx, leases.Lease{ID: lid}, leases.SynchronousDelete) | 		err = cli.LeasesService().Delete(ctx, leases.Lease{ID: lid}, leases.SynchronousDelete) | ||||||
| 		assert.NoError(t, err) | 		assert.NoError(t, err) | ||||||
| 	} | 	} | ||||||
|   | |||||||
| @@ -67,7 +67,7 @@ func TestRunPodSandboxWithSetupCNIFailure(t *testing.T) { | |||||||
| 	t.Logf("Create a sandbox") | 	t.Logf("Create a sandbox") | ||||||
| 	_, err := runtimeService.RunPodSandbox(sbConfig, failpointRuntimeHandler) | 	_, err := runtimeService.RunPodSandbox(sbConfig, failpointRuntimeHandler) | ||||||
| 	require.Error(t, err) | 	require.Error(t, err) | ||||||
| 	require.Equal(t, true, strings.Contains(err.Error(), "you-shall-not-pass!")) | 	require.ErrorContains(t, err, "you-shall-not-pass!") | ||||||
|  |  | ||||||
| 	t.Logf("Retry to create sandbox with same config") | 	t.Logf("Retry to create sandbox with same config") | ||||||
| 	sb, err := runtimeService.RunPodSandbox(sbConfig, failpointRuntimeHandler) | 	sb, err := runtimeService.RunPodSandbox(sbConfig, failpointRuntimeHandler) | ||||||
| @@ -95,7 +95,7 @@ func TestRunPodSandboxWithShimStartFailure(t *testing.T) { | |||||||
| 	t.Logf("Create a sandbox") | 	t.Logf("Create a sandbox") | ||||||
| 	_, err := runtimeService.RunPodSandbox(sbConfig, failpointRuntimeHandler) | 	_, err := runtimeService.RunPodSandbox(sbConfig, failpointRuntimeHandler) | ||||||
| 	require.Error(t, err) | 	require.Error(t, err) | ||||||
| 	require.Equal(t, true, strings.Contains(err.Error(), "no hard feelings")) | 	require.ErrorContains(t, err, "no hard feelings") | ||||||
| } | } | ||||||
|  |  | ||||||
| // TestRunPodSandboxWithShimDeleteFailure should keep the sandbox record if | // TestRunPodSandboxWithShimDeleteFailure should keep the sandbox record if | ||||||
| @@ -130,16 +130,16 @@ func TestRunPodSandboxWithShimDeleteFailure(t *testing.T) { | |||||||
| 			require.Len(t, l, 1) | 			require.Len(t, l, 1) | ||||||
|  |  | ||||||
| 			sb := l[0] | 			sb := l[0] | ||||||
| 			require.Equal(t, sb.State, criapiv1.PodSandboxState_SANDBOX_NOTREADY) | 			require.Equal(t, criapiv1.PodSandboxState_SANDBOX_NOTREADY, sb.State) | ||||||
| 			require.Equal(t, sb.Metadata.Name, sbConfig.Metadata.Name) | 			require.Equal(t, sbConfig.Metadata.Name, sb.Metadata.Name) | ||||||
| 			require.Equal(t, sb.Metadata.Namespace, sbConfig.Metadata.Namespace) | 			require.Equal(t, sbConfig.Metadata.Namespace, sb.Metadata.Namespace) | ||||||
| 			require.Equal(t, sb.Metadata.Uid, sbConfig.Metadata.Uid) | 			require.Equal(t, sbConfig.Metadata.Uid, sb.Metadata.Uid) | ||||||
| 			require.Equal(t, sb.Metadata.Attempt, sbConfig.Metadata.Attempt) | 			require.Equal(t, sbConfig.Metadata.Attempt, sb.Metadata.Attempt) | ||||||
|  |  | ||||||
| 			t.Log("Check PodSandboxStatus") | 			t.Log("Check PodSandboxStatus") | ||||||
| 			sbStatus, err := runtimeService.PodSandboxStatus(sb.Id) | 			sbStatus, err := runtimeService.PodSandboxStatus(sb.Id) | ||||||
| 			require.NoError(t, err) | 			require.NoError(t, err) | ||||||
| 			require.Equal(t, sbStatus.State, criapiv1.PodSandboxState_SANDBOX_NOTREADY) | 			require.Equal(t, criapiv1.PodSandboxState_SANDBOX_NOTREADY, sbStatus.State) | ||||||
| 			require.Greater(t, len(sbStatus.Network.Ip), 0) | 			require.Greater(t, len(sbStatus.Network.Ip), 0) | ||||||
|  |  | ||||||
| 			if restart { | 			if restart { | ||||||
| @@ -150,13 +150,13 @@ func TestRunPodSandboxWithShimDeleteFailure(t *testing.T) { | |||||||
| 				l, err = runtimeService.ListPodSandbox(&criapiv1.PodSandboxFilter{Id: sb.Id}) | 				l, err = runtimeService.ListPodSandbox(&criapiv1.PodSandboxFilter{Id: sb.Id}) | ||||||
| 				require.NoError(t, err) | 				require.NoError(t, err) | ||||||
| 				require.Len(t, l, 1) | 				require.Len(t, l, 1) | ||||||
| 				require.Equal(t, l[0].State, criapiv1.PodSandboxState_SANDBOX_NOTREADY) | 				require.Equal(t, criapiv1.PodSandboxState_SANDBOX_NOTREADY, l[0].State) | ||||||
|  |  | ||||||
| 				t.Log("Check PodSandboxStatus") | 				t.Log("Check PodSandboxStatus") | ||||||
| 				sbStatus, err := runtimeService.PodSandboxStatus(sb.Id) | 				sbStatus, err := runtimeService.PodSandboxStatus(sb.Id) | ||||||
| 				require.NoError(t, err) | 				require.NoError(t, err) | ||||||
| 				t.Log(sbStatus.Network) | 				t.Log(sbStatus.Network) | ||||||
| 				require.Equal(t, sbStatus.State, criapiv1.PodSandboxState_SANDBOX_NOTREADY) | 				require.Equal(t, criapiv1.PodSandboxState_SANDBOX_NOTREADY, sbStatus.State) | ||||||
| 			} | 			} | ||||||
|  |  | ||||||
| 			t.Log("Cleanup leaky sandbox") | 			t.Log("Cleanup leaky sandbox") | ||||||
| @@ -206,11 +206,11 @@ func TestRunPodSandboxWithShimStartAndTeardownCNIFailure(t *testing.T) { | |||||||
| 			require.Len(t, l, 1) | 			require.Len(t, l, 1) | ||||||
|  |  | ||||||
| 			sb := l[0] | 			sb := l[0] | ||||||
| 			require.Equal(t, sb.State, criapiv1.PodSandboxState_SANDBOX_NOTREADY) | 			require.Equal(t, criapiv1.PodSandboxState_SANDBOX_NOTREADY, sb.State) | ||||||
| 			require.Equal(t, sb.Metadata.Name, sbConfig.Metadata.Name) | 			require.Equal(t, sbConfig.Metadata.Name, sb.Metadata.Name) | ||||||
| 			require.Equal(t, sb.Metadata.Namespace, sbConfig.Metadata.Namespace) | 			require.Equal(t, sbConfig.Metadata.Namespace, sb.Metadata.Namespace) | ||||||
| 			require.Equal(t, sb.Metadata.Uid, sbConfig.Metadata.Uid) | 			require.Equal(t, sbConfig.Metadata.Uid, sb.Metadata.Uid) | ||||||
| 			require.Equal(t, sb.Metadata.Attempt, sbConfig.Metadata.Attempt) | 			require.Equal(t, sbConfig.Metadata.Attempt, sb.Metadata.Attempt) | ||||||
|  |  | ||||||
| 			if restart { | 			if restart { | ||||||
| 				t.Log("Restart containerd") | 				t.Log("Restart containerd") | ||||||
| @@ -220,7 +220,7 @@ func TestRunPodSandboxWithShimStartAndTeardownCNIFailure(t *testing.T) { | |||||||
| 				l, err = runtimeService.ListPodSandbox(&criapiv1.PodSandboxFilter{Id: sb.Id}) | 				l, err = runtimeService.ListPodSandbox(&criapiv1.PodSandboxFilter{Id: sb.Id}) | ||||||
| 				require.NoError(t, err) | 				require.NoError(t, err) | ||||||
| 				require.Len(t, l, 1) | 				require.Len(t, l, 1) | ||||||
| 				require.Equal(t, l[0].State, criapiv1.PodSandboxState_SANDBOX_NOTREADY) | 				require.Equal(t, criapiv1.PodSandboxState_SANDBOX_NOTREADY, l[0].State) | ||||||
| 			} | 			} | ||||||
|  |  | ||||||
| 			t.Log("Cleanup leaky sandbox") | 			t.Log("Cleanup leaky sandbox") | ||||||
| @@ -287,11 +287,11 @@ func TestRunPodSandboxAndTeardownCNISlow(t *testing.T) { | |||||||
| 		require.NoError(t, err) | 		require.NoError(t, err) | ||||||
| 	}() | 	}() | ||||||
|  |  | ||||||
| 	assert.Equal(t, sb.State, criapiv1.PodSandboxState_SANDBOX_NOTREADY) | 	assert.Equal(t, criapiv1.PodSandboxState_SANDBOX_NOTREADY, sb.State) | ||||||
| 	assert.Equal(t, sb.Metadata.Name, sbConfig.Metadata.Name) | 	assert.Equal(t, sbConfig.Metadata.Name, sb.Metadata.Name) | ||||||
| 	assert.Equal(t, sb.Metadata.Namespace, sbConfig.Metadata.Namespace) | 	assert.Equal(t, sbConfig.Metadata.Namespace, sb.Metadata.Namespace) | ||||||
| 	assert.Equal(t, sb.Metadata.Uid, sbConfig.Metadata.Uid) | 	assert.Equal(t, sbConfig.Metadata.Uid, sb.Metadata.Uid) | ||||||
| 	assert.Equal(t, sb.Metadata.Attempt, sbConfig.Metadata.Attempt) | 	assert.Equal(t, sbConfig.Metadata.Attempt, sb.Metadata.Attempt) | ||||||
|  |  | ||||||
| 	if os.Getenv("DISABLE_CRI_SANDBOXES") != "" { | 	if os.Getenv("DISABLE_CRI_SANDBOXES") != "" { | ||||||
| 		// non-sbserver | 		// non-sbserver | ||||||
|   | |||||||
| @@ -42,7 +42,7 @@ func TestTruncIndex(t *testing.T) { | |||||||
| 	t.Logf("Get image status by truncindex, truncID: %s", imgTruncID) | 	t.Logf("Get image status by truncindex, truncID: %s", imgTruncID) | ||||||
| 	res, err := imageService.ImageStatus(&runtimeapi.ImageSpec{Image: imgTruncID}) | 	res, err := imageService.ImageStatus(&runtimeapi.ImageSpec{Image: imgTruncID}) | ||||||
| 	require.NoError(t, err) | 	require.NoError(t, err) | ||||||
| 	require.NotEqual(t, nil, res) | 	require.NotNil(t, res) | ||||||
| 	assert.Equal(t, imgID, res.Id) | 	assert.Equal(t, imgID, res.Id) | ||||||
|  |  | ||||||
| 	// TODO(yanxuean): for failure test case where there are two images with the same truncindex. | 	// TODO(yanxuean): for failure test case where there are two images with the same truncindex. | ||||||
|   | |||||||
		Reference in New Issue
	
	Block a user
	 Henry Wang
					Henry Wang