//go:build windows // +build windows /* Copyright The containerd Authors. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. */ package integration import ( "context" "encoding/json" "fmt" "io" "net" "os" "os/exec" "path/filepath" "strconv" "strings" "testing" "time" "github.com/Microsoft/hcsshim/osversion" "github.com/stretchr/testify/assert" "github.com/stretchr/testify/require" "golang.org/x/sys/windows/registry" runtime "k8s.io/cri-api/pkg/apis/runtime/v1" ) // Returns what nanoserver image version to use according to the build number func getTestImage() (string, error) { k, err := registry.OpenKey(registry.LOCAL_MACHINE, `SOFTWARE\Microsoft\Windows NT\CurrentVersion`, registry.QUERY_VALUE) if err != nil { return "", err } defer k.Close() b, _, _ := k.GetStringValue("CurrentBuild") buildNum, _ := strconv.Atoi(b) switch buildNum { case osversion.RS1: return "mcr.microsoft.com/windows/nanoserver:sac2016", nil case osversion.RS3: return "mcr.microsoft.com/windows/nanoserver:1709", nil case osversion.RS4: return "mcr.microsoft.com/windows/nanoserver:1803", nil case osversion.RS5: return "mcr.microsoft.com/windows/nanoserver:1809", nil case osversion.V19H1: return "mcr.microsoft.com/windows/nanoserver:1903", nil case osversion.V19H2: return "mcr.microsoft.com/windows/nanoserver:1909", nil case osversion.V20H1: return "mcr.microsoft.com/windows/nanoserver:2004", nil case osversion.V20H2: return "mcr.microsoft.com/windows/nanoserver:20H2", nil case osversion.V21H2Server: return "mcr.microsoft.com/windows/nanoserver:ltsc2022", nil default: // Due to some efforts in improving down-level compatibility for Windows containers (see // https://techcommunity.microsoft.com/t5/containers/windows-server-2022-and-beyond-for-containers/ba-p/2712487) // the ltsc2022 image should continue to work on builds ws2022 and onwards (Windows 11 for example). With this in mind, // if there's no mapping for the host build just use the Windows Server 2022 image. if buildNum > osversion.V21H2Server { return "mcr.microsoft.com/windows/nanoserver:ltsc2022", nil } return "", fmt.Errorf("No test image defined for Windows build version: %s", b) } } func TestSandboxRemoveWithoutIPLeakage(t *testing.T) { t.Logf("Make sure host-local ipam is in use") config, err := CRIConfig() require.NoError(t, err) fs, err := os.ReadDir(config.NetworkPluginConfDir) require.NoError(t, err) require.NotEmpty(t, fs) f := filepath.Join(config.NetworkPluginConfDir, fs[0].Name()) cniConfig, err := os.ReadFile(f) require.NoError(t, err) if !strings.Contains(string(cniConfig), "azure-vnet-ipam") { t.Skip("azure-vnet ipam is not in use") } t.Logf("Create a sandbox") sbConfig := PodSandboxConfig("sandbox", "remove-without-ip-leakage") sb, err := runtimeService.RunPodSandbox(sbConfig, *runtimeHandler) require.NoError(t, err) defer func() { // Make sure the sandbox is cleaned up in any case. runtimeService.StopPodSandbox(sb) runtimeService.RemovePodSandbox(sb) }() t.Logf("Get pod information") status, info, err := SandboxInfo(sb) require.NoError(t, err) ip := status.GetNetwork().GetIp() require.NotEmpty(t, ip) require.NotNil(t, info.RuntimeSpec.Windows) netNS := info.RuntimeSpec.Windows.Network.NetworkNamespace require.NotEmpty(t, netNS, "network namespace should be set") t.Logf("Should be able to find the pod ip in host-local checkpoint") checkIP := func(ip string) bool { f, err := os.Open("azure-vnet-ipam.json") require.NoError(t, err) defer f.Close() data, err := io.ReadAll(f) require.NoError(t, err) var jsonData map[string]interface{} err = json.Unmarshal(data, &jsonData) require.NoError(t, err) walkJSON := func(initial map[string]interface{}, elementNames ...string) map[string]interface{} { element := initial for _, name := range elementNames { element = element[name].(map[string]interface{}) } return element } pools := walkJSON(jsonData, "IPAM", "AddressSpaces", "local", "Pools") ipAddr := net.ParseIP(ip) var ipPool map[string]interface{} for poolID, pool := range pools { // Each pool will contain its key as its subnet. _, ipnet, _ := net.ParseCIDR(poolID) if ipnet.Contains(ipAddr) { ipPool = pool.(map[string]interface{}) break } } // Search in the IP Pool and see if it's in use or not. for address, details := range walkJSON(ipPool, "Addresses") { if address == ip { d := details.(map[string]interface{}) return d["InUse"].(bool) } } return false } require.True(t, checkIP(ip)) t.Logf("Kill sandbox container") require.NoError(t, KillPid(int(info.Pid))) t.Logf("Delete network namespace") cmd := exec.Command("hnsdiag.exe", "delete", "namespaces", netNS) require.NoError(t, cmd.Run()) t.Logf("Network namespace should be closed") _, info, err = SandboxInfo(sb) require.NoError(t, err) assert.True(t, info.NetNSClosed) t.Logf("Sandbox state should be NOTREADY") assert.NoError(t, Eventually(func() (bool, error) { status, err := runtimeService.PodSandboxStatus(sb) if err != nil { return false, err } return status.GetState() == runtime.PodSandboxState_SANDBOX_NOTREADY, nil }, time.Second, 30*time.Second), "sandbox state should become NOTREADY") t.Logf("Should still be able to find the pod ip in host-local checkpoint") assert.True(t, checkIP(ip)) t.Logf("Should be able to stop and remove the sandbox") assert.NoError(t, runtimeService.StopPodSandbox(sb)) assert.NoError(t, runtimeService.RemovePodSandbox(sb)) t.Logf("Should not be able to find the pod ip in host-local checkpoint") assert.False(t, checkIP(ip), fmt.Sprintf("The IP: %s is still in use in azure-vnet-ipam.json", ip)) } func removePodSandbox(ctx context.Context, t *testing.T, client runtime.RuntimeServiceClient, podID string) { t.Helper() _, err := client.RemovePodSandbox(ctx, &runtime.RemovePodSandboxRequest{ PodSandboxId: podID, }) require.NoError(t, err, "failed RemovePodSandbox for sandbox: %s", podID) } func stopPodSandbox(ctx context.Context, t *testing.T, client runtime.RuntimeServiceClient, podID string) { t.Helper() _, err := client.StopPodSandbox(ctx, &runtime.StopPodSandboxRequest{ PodSandboxId: podID, }) require.NoError(t, err, "failed StopPodSandbox for sandbox: %s", podID) } func stopContainer(ctx context.Context, t *testing.T, client runtime.RuntimeServiceClient, containerID string) { t.Helper() _, err := client.StopContainer(ctx, &runtime.StopContainerRequest{ ContainerId: containerID, Timeout: 0, }) require.NoError(t, err, "failed StopContainer request for container: %s", containerID) } func startContainer(ctx context.Context, t *testing.T, client runtime.RuntimeServiceClient, containerID string) { t.Helper() _, err := client.StartContainer(ctx, &runtime.StartContainerRequest{ ContainerId: containerID, }) require.NoError(t, err, "failed StartContainer request for container: %s", containerID) } func removeContainer(ctx context.Context, t *testing.T, client runtime.RuntimeServiceClient, containerID string) { t.Helper() _, err := client.RemoveContainer(ctx, &runtime.RemoveContainerRequest{ ContainerId: containerID, }) require.NoError(t, err, "failed RemoveContainer request for container: %s", containerID) } // This test checks if create/stop and remove pods and containers work as expected func TestCreateContainer(t *testing.T) { testImage, err := getTestImage() if err != nil { t.Skip("skipping test, error: ", err) } client, err := RawRuntimeClient() require.NoError(t, err, "failed to get raw grpc runtime service client") ctx, cancel := context.WithCancel(context.Background()) t.Cleanup(func() { cancel() }) t.Log("Create a pod sandbox") sbConfig := &runtime.PodSandboxConfig{ Metadata: &runtime.PodSandboxMetadata{ Name: t.Name(), }, } sandboxRequest := &runtime.RunPodSandboxRequest{ Config: sbConfig, RuntimeHandler: "runhcs-wcow-process", } sandBoxResponse, err := client.RunPodSandbox(ctx, sandboxRequest) require.NoError(t, err, "failed RunPodSandbox request") // Make sure the sandbox is cleaned up. t.Cleanup(func() { removePodSandbox(ctx, t, client, sandBoxResponse.PodSandboxId) }) t.Cleanup(func() { stopPodSandbox(ctx, t, client, sandBoxResponse.PodSandboxId) }) EnsureImageExists(t, testImage) t.Log("Create a container") createCtrRequest := &runtime.CreateContainerRequest{ Config: &runtime.ContainerConfig{ Metadata: &runtime.ContainerMetadata{ Name: t.Name() + "-CreateContainerTest", }, Image: &runtime.ImageSpec{ Image: testImage, }, Command: []string{ "cmd", "/c", "ping", "-t", "127.0.0.1", }, Windows: &runtime.WindowsContainerConfig{ Resources: &runtime.WindowsContainerResources{ CpuShares: 500, }, }, }, PodSandboxId: sandBoxResponse.PodSandboxId, SandboxConfig: sandboxRequest.Config, } createCtrResponse, err := client.CreateContainer(ctx, createCtrRequest) require.NoError(t, err, "failed CreateContainer request in sandbox: %s", sandBoxResponse.PodSandboxId) // Make sure the container is cleaned up. t.Cleanup(func() { removeContainer(ctx, t, client, createCtrResponse.ContainerId) }) startContainer(ctx, t, client, createCtrResponse.ContainerId) stopContainer(ctx, t, client, createCtrResponse.ContainerId) }