
This adds a new Label to EndpointSlices that will ensure that multiple controllers or entities can manage subsets of EndpointSlices. This label provides a way to indicate the controller or entity responsible for managing an EndpointSlice. To provide a seamless upgrade from the alpha release of EndpointSlices that did not support this label, a temporary annotation has been added on Services to indicate that this label has been initially set on EndpointSlices. That annotation will be set automatically by the EndpointSlice controller with this commit once appropriate Labels have been added on the corresponding EndpointSlices.
507 lines
18 KiB
Go
507 lines
18 KiB
Go
/*
|
|
Copyright 2019 The Kubernetes Authors.
|
|
|
|
Licensed under the Apache License, Version 2.0 (the "License");
|
|
you may not use this file except in compliance with the License.
|
|
You may obtain a copy of the License at
|
|
|
|
http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
Unless required by applicable law or agreed to in writing, software
|
|
distributed under the License is distributed on an "AS IS" BASIS,
|
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
See the License for the specific language governing permissions and
|
|
limitations under the License.
|
|
*/
|
|
|
|
package endpointslice
|
|
|
|
import (
|
|
"fmt"
|
|
"reflect"
|
|
"testing"
|
|
"time"
|
|
|
|
"github.com/stretchr/testify/assert"
|
|
v1 "k8s.io/api/core/v1"
|
|
discovery "k8s.io/api/discovery/v1alpha1"
|
|
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
|
"k8s.io/apimachinery/pkg/util/intstr"
|
|
"k8s.io/client-go/informers"
|
|
"k8s.io/client-go/kubernetes/fake"
|
|
k8stesting "k8s.io/client-go/testing"
|
|
"k8s.io/client-go/tools/cache"
|
|
"k8s.io/kubernetes/pkg/controller"
|
|
endpointutil "k8s.io/kubernetes/pkg/controller/util/endpoint"
|
|
utilpointer "k8s.io/utils/pointer"
|
|
)
|
|
|
|
// Most of the tests related to EndpointSlice allocation can be found in reconciler_test.go
|
|
// Tests here primarily focus on unique controller functionality before the reconciler begins
|
|
|
|
var alwaysReady = func() bool { return true }
|
|
|
|
type endpointSliceController struct {
|
|
*Controller
|
|
endpointSliceStore cache.Store
|
|
nodeStore cache.Store
|
|
podStore cache.Store
|
|
serviceStore cache.Store
|
|
}
|
|
|
|
func newController(nodeNames []string) (*fake.Clientset, *endpointSliceController) {
|
|
client := newClientset()
|
|
informerFactory := informers.NewSharedInformerFactory(client, controller.NoResyncPeriodFunc())
|
|
nodeInformer := informerFactory.Core().V1().Nodes()
|
|
indexer := nodeInformer.Informer().GetIndexer()
|
|
for _, nodeName := range nodeNames {
|
|
indexer.Add(&v1.Node{ObjectMeta: metav1.ObjectMeta{Name: nodeName}})
|
|
}
|
|
|
|
esController := NewController(
|
|
informerFactory.Core().V1().Pods(),
|
|
informerFactory.Core().V1().Services(),
|
|
nodeInformer,
|
|
informerFactory.Discovery().V1alpha1().EndpointSlices(),
|
|
int32(100),
|
|
client)
|
|
|
|
esController.nodesSynced = alwaysReady
|
|
esController.podsSynced = alwaysReady
|
|
esController.servicesSynced = alwaysReady
|
|
esController.endpointSlicesSynced = alwaysReady
|
|
|
|
return client, &endpointSliceController{
|
|
esController,
|
|
informerFactory.Discovery().V1alpha1().EndpointSlices().Informer().GetStore(),
|
|
informerFactory.Core().V1().Nodes().Informer().GetStore(),
|
|
informerFactory.Core().V1().Pods().Informer().GetStore(),
|
|
informerFactory.Core().V1().Services().Informer().GetStore(),
|
|
}
|
|
}
|
|
|
|
// Ensure SyncService for service with no selector results in no action
|
|
func TestSyncServiceNoSelector(t *testing.T) {
|
|
ns := metav1.NamespaceDefault
|
|
serviceName := "testing-1"
|
|
client, esController := newController([]string{"node-1"})
|
|
esController.serviceStore.Add(&v1.Service{
|
|
ObjectMeta: metav1.ObjectMeta{Name: serviceName, Namespace: ns},
|
|
Spec: v1.ServiceSpec{
|
|
Ports: []v1.ServicePort{{TargetPort: intstr.FromInt(80)}},
|
|
},
|
|
})
|
|
|
|
err := esController.syncService(fmt.Sprintf("%s/%s", ns, serviceName))
|
|
assert.Nil(t, err)
|
|
assert.Len(t, client.Actions(), 0)
|
|
}
|
|
|
|
// Ensure SyncService for service with selector but no pods results in placeholder EndpointSlice
|
|
func TestSyncServiceWithSelector(t *testing.T) {
|
|
ns := metav1.NamespaceDefault
|
|
serviceName := "testing-1"
|
|
client, esController := newController([]string{"node-1"})
|
|
standardSyncService(t, esController, ns, serviceName, "true")
|
|
expectActions(t, client.Actions(), 1, "create", "endpointslices")
|
|
|
|
sliceList, err := client.DiscoveryV1alpha1().EndpointSlices(ns).List(metav1.ListOptions{})
|
|
assert.Nil(t, err, "Expected no error fetching endpoint slices")
|
|
assert.Len(t, sliceList.Items, 1, "Expected 1 endpoint slices")
|
|
slice := sliceList.Items[0]
|
|
assert.Regexp(t, "^"+serviceName, slice.Name)
|
|
assert.Equal(t, serviceName, slice.Labels[discovery.LabelServiceName])
|
|
assert.EqualValues(t, []discovery.EndpointPort{}, slice.Ports)
|
|
assert.EqualValues(t, []discovery.Endpoint{}, slice.Endpoints)
|
|
assert.NotEmpty(t, slice.Annotations["endpoints.kubernetes.io/last-change-trigger-time"])
|
|
}
|
|
|
|
// Ensure SyncService gracefully handles a missing service. This test also
|
|
// populates another existing service to ensure a clean up process doesn't
|
|
// remove too much.
|
|
func TestSyncServiceMissing(t *testing.T) {
|
|
namespace := metav1.NamespaceDefault
|
|
client, esController := newController([]string{"node-1"})
|
|
|
|
// Build up existing service
|
|
existingServiceName := "stillthere"
|
|
existingServiceKey := endpointutil.ServiceKey{Name: existingServiceName, Namespace: namespace}
|
|
esController.triggerTimeTracker.ServiceStates[existingServiceKey] = endpointutil.ServiceState{}
|
|
esController.serviceStore.Add(&v1.Service{
|
|
ObjectMeta: metav1.ObjectMeta{Name: existingServiceName, Namespace: namespace},
|
|
Spec: v1.ServiceSpec{
|
|
Ports: []v1.ServicePort{{TargetPort: intstr.FromInt(80)}},
|
|
Selector: map[string]string{"foo": "bar"},
|
|
},
|
|
})
|
|
|
|
// Add missing service to triggerTimeTracker to ensure the reference is cleaned up
|
|
missingServiceName := "notthere"
|
|
missingServiceKey := endpointutil.ServiceKey{Name: missingServiceName, Namespace: namespace}
|
|
esController.triggerTimeTracker.ServiceStates[missingServiceKey] = endpointutil.ServiceState{}
|
|
|
|
err := esController.syncService(fmt.Sprintf("%s/%s", namespace, missingServiceName))
|
|
|
|
// nil should be returned when the service doesn't exist
|
|
assert.Nil(t, err, "Expected no error syncing service")
|
|
|
|
// That should mean no client actions were performed
|
|
assert.Len(t, client.Actions(), 0)
|
|
|
|
// TriggerTimeTracker should have removed the reference to the missing service
|
|
assert.NotContains(t, esController.triggerTimeTracker.ServiceStates, missingServiceKey)
|
|
|
|
// TriggerTimeTracker should have left the reference to the missing service
|
|
assert.Contains(t, esController.triggerTimeTracker.ServiceStates, existingServiceKey)
|
|
}
|
|
|
|
// Ensure SyncService correctly selects Pods.
|
|
func TestSyncServicePodSelection(t *testing.T) {
|
|
client, esController := newController([]string{"node-1"})
|
|
ns := metav1.NamespaceDefault
|
|
|
|
pod1 := newPod(1, ns, true, 0)
|
|
esController.podStore.Add(pod1)
|
|
|
|
// ensure this pod will not match the selector
|
|
pod2 := newPod(2, ns, true, 0)
|
|
pod2.Labels["foo"] = "boo"
|
|
esController.podStore.Add(pod2)
|
|
|
|
standardSyncService(t, esController, ns, "testing-1", "true")
|
|
expectActions(t, client.Actions(), 1, "create", "endpointslices")
|
|
|
|
// an endpoint slice should be created, it should only reference pod1 (not pod2)
|
|
slices, err := client.DiscoveryV1alpha1().EndpointSlices(ns).List(metav1.ListOptions{})
|
|
assert.Nil(t, err, "Expected no error fetching endpoint slices")
|
|
assert.Len(t, slices.Items, 1, "Expected 1 endpoint slices")
|
|
slice := slices.Items[0]
|
|
assert.Len(t, slice.Endpoints, 1, "Expected 1 endpoint in first slice")
|
|
assert.NotEmpty(t, slice.Annotations["endpoints.kubernetes.io/last-change-trigger-time"])
|
|
endpoint := slice.Endpoints[0]
|
|
assert.EqualValues(t, endpoint.TargetRef, &v1.ObjectReference{Kind: "Pod", Namespace: ns, Name: pod1.Name})
|
|
}
|
|
|
|
// Ensure SyncService correctly selects and labels EndpointSlices.
|
|
func TestSyncServiceEndpointSliceLabelSelection(t *testing.T) {
|
|
client, esController := newController([]string{"node-1"})
|
|
ns := metav1.NamespaceDefault
|
|
serviceName := "testing-1"
|
|
|
|
// 5 slices, 3 with matching labels for our service
|
|
endpointSlices := []*discovery.EndpointSlice{{
|
|
ObjectMeta: metav1.ObjectMeta{
|
|
Name: "matching-1",
|
|
Namespace: ns,
|
|
Labels: map[string]string{
|
|
discovery.LabelServiceName: serviceName,
|
|
discovery.LabelManagedBy: controllerName,
|
|
},
|
|
},
|
|
}, {
|
|
ObjectMeta: metav1.ObjectMeta{
|
|
Name: "matching-2",
|
|
Namespace: ns,
|
|
Labels: map[string]string{
|
|
discovery.LabelServiceName: serviceName,
|
|
discovery.LabelManagedBy: controllerName,
|
|
},
|
|
},
|
|
}, {
|
|
ObjectMeta: metav1.ObjectMeta{
|
|
Name: "partially-matching-1",
|
|
Namespace: ns,
|
|
Labels: map[string]string{
|
|
discovery.LabelServiceName: serviceName,
|
|
},
|
|
},
|
|
}, {
|
|
ObjectMeta: metav1.ObjectMeta{
|
|
Name: "not-matching-1",
|
|
Namespace: ns,
|
|
Labels: map[string]string{
|
|
discovery.LabelServiceName: "something-else",
|
|
discovery.LabelManagedBy: controllerName,
|
|
},
|
|
},
|
|
}, {
|
|
ObjectMeta: metav1.ObjectMeta{
|
|
Name: "not-matching-2",
|
|
Namespace: ns,
|
|
Labels: map[string]string{
|
|
discovery.LabelServiceName: serviceName,
|
|
discovery.LabelManagedBy: "something-else",
|
|
},
|
|
},
|
|
}}
|
|
|
|
// need to add them to both store and fake clientset
|
|
for _, endpointSlice := range endpointSlices {
|
|
err := esController.endpointSliceStore.Add(endpointSlice)
|
|
if err != nil {
|
|
t.Fatalf("Expected no error adding EndpointSlice: %v", err)
|
|
}
|
|
_, err = client.DiscoveryV1alpha1().EndpointSlices(ns).Create(endpointSlice)
|
|
if err != nil {
|
|
t.Fatalf("Expected no error creating EndpointSlice: %v", err)
|
|
}
|
|
}
|
|
|
|
// +1 for extra action involved in Service creation before syncService call.
|
|
numActionsBefore := len(client.Actions()) + 1
|
|
standardSyncService(t, esController, ns, serviceName, "false")
|
|
|
|
if len(client.Actions()) != numActionsBefore+5 {
|
|
t.Errorf("Expected 5 more actions, got %d", len(client.Actions())-numActionsBefore)
|
|
}
|
|
|
|
// endpointslice should have LabelsManagedBy set as part of update.
|
|
expectAction(t, client.Actions(), numActionsBefore, "update", "endpointslices")
|
|
|
|
// service should have managedBySetupAnnotation set as part of update.
|
|
expectAction(t, client.Actions(), numActionsBefore+1, "update", "services")
|
|
|
|
// only 3 slices should match, 2 of those should be deleted, 1 should be updated as a placeholder
|
|
expectAction(t, client.Actions(), numActionsBefore+2, "update", "endpointslices")
|
|
expectAction(t, client.Actions(), numActionsBefore+3, "delete", "endpointslices")
|
|
expectAction(t, client.Actions(), numActionsBefore+4, "delete", "endpointslices")
|
|
}
|
|
|
|
// Ensure SyncService handles a variety of protocols and IPs appropriately.
|
|
func TestSyncServiceFull(t *testing.T) {
|
|
client, esController := newController([]string{"node-1"})
|
|
namespace := metav1.NamespaceDefault
|
|
serviceName := "all-the-protocols"
|
|
|
|
// pod 1 only uses PodIP status attr
|
|
pod1 := newPod(1, namespace, true, 0)
|
|
pod1.Status.PodIP = "1.2.3.4"
|
|
pod1.Status.PodIPs = []v1.PodIP{}
|
|
esController.podStore.Add(pod1)
|
|
|
|
// pod 2 only uses PodIPs status attr
|
|
pod2 := newPod(2, namespace, true, 0)
|
|
pod2.Status.PodIP = ""
|
|
pod2.Status.PodIPs = []v1.PodIP{{IP: "1.2.3.5"}, {IP: "1234::5678:0000:0000:9abc:def0"}}
|
|
esController.podStore.Add(pod2)
|
|
|
|
// create service with all protocols and multiple ports
|
|
serviceCreateTime := time.Now()
|
|
service := &v1.Service{
|
|
ObjectMeta: metav1.ObjectMeta{
|
|
Name: serviceName,
|
|
Namespace: namespace,
|
|
CreationTimestamp: metav1.NewTime(serviceCreateTime),
|
|
},
|
|
Spec: v1.ServiceSpec{
|
|
Ports: []v1.ServicePort{
|
|
{Name: "tcp-example", TargetPort: intstr.FromInt(80), Protocol: v1.ProtocolTCP},
|
|
{Name: "udp-example", TargetPort: intstr.FromInt(161), Protocol: v1.ProtocolUDP},
|
|
{Name: "sctp-example", TargetPort: intstr.FromInt(3456), Protocol: v1.ProtocolSCTP},
|
|
},
|
|
Selector: map[string]string{"foo": "bar"},
|
|
},
|
|
}
|
|
esController.serviceStore.Add(service)
|
|
_, err := esController.client.CoreV1().Services(namespace).Create(service)
|
|
assert.Nil(t, err, "Expected no error creating service")
|
|
|
|
// run through full sync service loop
|
|
err = esController.syncService(fmt.Sprintf("%s/%s", namespace, serviceName))
|
|
assert.Nil(t, err)
|
|
|
|
// last action should be to create endpoint slice
|
|
expectActions(t, client.Actions(), 1, "create", "endpointslices")
|
|
sliceList, err := client.DiscoveryV1alpha1().EndpointSlices(namespace).List(metav1.ListOptions{})
|
|
assert.Nil(t, err, "Expected no error fetching endpoint slices")
|
|
assert.Len(t, sliceList.Items, 1, "Expected 1 endpoint slices")
|
|
|
|
// ensure all attributes of endpoint slice match expected state
|
|
slice := sliceList.Items[0]
|
|
assert.Len(t, slice.Endpoints, 2, "Expected 2 endpoints in first slice")
|
|
assert.Equal(t, slice.Annotations["endpoints.kubernetes.io/last-change-trigger-time"], serviceCreateTime.Format(time.RFC3339Nano))
|
|
assert.EqualValues(t, []discovery.EndpointPort{{
|
|
Name: strPtr("tcp-example"),
|
|
Protocol: protoPtr(v1.ProtocolTCP),
|
|
Port: int32Ptr(int32(80)),
|
|
}, {
|
|
Name: strPtr("udp-example"),
|
|
Protocol: protoPtr(v1.ProtocolUDP),
|
|
Port: int32Ptr(int32(161)),
|
|
}, {
|
|
Name: strPtr("sctp-example"),
|
|
Protocol: protoPtr(v1.ProtocolSCTP),
|
|
Port: int32Ptr(int32(3456)),
|
|
}}, slice.Ports)
|
|
|
|
assert.ElementsMatch(t, []discovery.Endpoint{{
|
|
Conditions: discovery.EndpointConditions{Ready: utilpointer.BoolPtr(true)},
|
|
Addresses: []string{"1.2.3.4"},
|
|
TargetRef: &v1.ObjectReference{Kind: "Pod", Namespace: namespace, Name: pod1.Name},
|
|
Topology: map[string]string{"kubernetes.io/hostname": "node-1"},
|
|
}, {
|
|
Conditions: discovery.EndpointConditions{Ready: utilpointer.BoolPtr(true)},
|
|
Addresses: []string{"1.2.3.5", "1234::5678:0000:0000:9abc:def0"},
|
|
TargetRef: &v1.ObjectReference{Kind: "Pod", Namespace: namespace, Name: pod2.Name},
|
|
Topology: map[string]string{"kubernetes.io/hostname": "node-1"},
|
|
}}, slice.Endpoints)
|
|
}
|
|
|
|
func TestEnsureSetupManagedByAnnotation(t *testing.T) {
|
|
serviceName := "testing-1"
|
|
|
|
testCases := map[string]struct {
|
|
serviceAnnotation string
|
|
startingSliceLabels map[string]string
|
|
expectedSliceLabels map[string]string
|
|
}{
|
|
"already-labeled": {
|
|
serviceAnnotation: "foo",
|
|
startingSliceLabels: map[string]string{
|
|
discovery.LabelServiceName: serviceName,
|
|
discovery.LabelManagedBy: controllerName,
|
|
},
|
|
expectedSliceLabels: map[string]string{
|
|
discovery.LabelServiceName: serviceName,
|
|
discovery.LabelManagedBy: controllerName,
|
|
},
|
|
},
|
|
"already-annotated": {
|
|
serviceAnnotation: managedBySetupCompleteValue,
|
|
startingSliceLabels: map[string]string{
|
|
discovery.LabelServiceName: serviceName,
|
|
discovery.LabelManagedBy: "other-controller",
|
|
},
|
|
expectedSliceLabels: map[string]string{
|
|
discovery.LabelServiceName: serviceName,
|
|
discovery.LabelManagedBy: "other-controller",
|
|
},
|
|
},
|
|
"missing-and-extra-label": {
|
|
serviceAnnotation: "foo",
|
|
startingSliceLabels: map[string]string{
|
|
discovery.LabelServiceName: serviceName,
|
|
"foo": "bar",
|
|
},
|
|
expectedSliceLabels: map[string]string{
|
|
discovery.LabelServiceName: serviceName,
|
|
discovery.LabelManagedBy: controllerName,
|
|
"foo": "bar",
|
|
},
|
|
},
|
|
"different-service": {
|
|
serviceAnnotation: "foo",
|
|
startingSliceLabels: map[string]string{
|
|
discovery.LabelServiceName: "something-else",
|
|
},
|
|
expectedSliceLabels: map[string]string{
|
|
discovery.LabelServiceName: "something-else",
|
|
},
|
|
},
|
|
}
|
|
|
|
for name, testCase := range testCases {
|
|
t.Run(name, func(t *testing.T) {
|
|
client, esController := newController([]string{"node-1"})
|
|
ns := metav1.NamespaceDefault
|
|
service := createService(t, esController, ns, serviceName, testCase.serviceAnnotation)
|
|
|
|
endpointSlice := &discovery.EndpointSlice{
|
|
ObjectMeta: metav1.ObjectMeta{
|
|
Name: "testing",
|
|
Namespace: ns,
|
|
Labels: testCase.startingSliceLabels,
|
|
},
|
|
}
|
|
|
|
err := esController.endpointSliceStore.Add(endpointSlice)
|
|
if err != nil {
|
|
t.Fatalf("Expected no error adding EndpointSlice: %v", err)
|
|
}
|
|
|
|
_, err = client.DiscoveryV1alpha1().EndpointSlices(ns).Create(endpointSlice)
|
|
if err != nil {
|
|
t.Fatalf("Expected no error creating EndpointSlice: %v", err)
|
|
}
|
|
|
|
esController.ensureSetupManagedByAnnotation(service)
|
|
|
|
updatedService, err := client.CoreV1().Services(ns).Get(service.Name, metav1.GetOptions{})
|
|
if err != nil {
|
|
t.Fatalf("Expected no error getting Service: %v", err)
|
|
}
|
|
|
|
if updatedService.Annotations[managedBySetupAnnotation] != managedBySetupCompleteValue {
|
|
t.Errorf("Expected managedBySetupAnnotation: %+v, got: %+v", managedBySetupCompleteValue, updatedService.Annotations[managedBySetupAnnotation])
|
|
}
|
|
|
|
updatedSlice, err := client.DiscoveryV1alpha1().EndpointSlices(ns).Get(endpointSlice.Name, metav1.GetOptions{})
|
|
if err != nil {
|
|
t.Fatalf("Expected no error getting EndpointSlice: %v", err)
|
|
}
|
|
|
|
if !reflect.DeepEqual(updatedSlice.Labels, testCase.expectedSliceLabels) {
|
|
t.Errorf("Expected labels: %+v, got: %+v", updatedSlice.Labels, testCase.expectedSliceLabels)
|
|
}
|
|
})
|
|
}
|
|
}
|
|
|
|
// Test helpers
|
|
|
|
func standardSyncService(t *testing.T, esController *endpointSliceController, namespace, serviceName, managedBySetup string) {
|
|
t.Helper()
|
|
createService(t, esController, namespace, serviceName, managedBySetup)
|
|
|
|
err := esController.syncService(fmt.Sprintf("%s/%s", namespace, serviceName))
|
|
assert.Nil(t, err, "Expected no error syncing service")
|
|
}
|
|
|
|
func createService(t *testing.T, esController *endpointSliceController, namespace, serviceName, managedBySetup string) *v1.Service {
|
|
t.Helper()
|
|
service := &v1.Service{
|
|
ObjectMeta: metav1.ObjectMeta{
|
|
Name: serviceName,
|
|
Namespace: namespace,
|
|
CreationTimestamp: metav1.NewTime(time.Now()),
|
|
Annotations: map[string]string{managedBySetupAnnotation: managedBySetup},
|
|
},
|
|
Spec: v1.ServiceSpec{
|
|
Ports: []v1.ServicePort{{TargetPort: intstr.FromInt(80)}},
|
|
Selector: map[string]string{"foo": "bar"},
|
|
},
|
|
}
|
|
esController.serviceStore.Add(service)
|
|
_, err := esController.client.CoreV1().Services(namespace).Create(service)
|
|
assert.Nil(t, err, "Expected no error creating service")
|
|
return service
|
|
}
|
|
|
|
func expectAction(t *testing.T, actions []k8stesting.Action, index int, verb, resource string) {
|
|
t.Helper()
|
|
if len(actions) <= index {
|
|
t.Fatalf("Expected at least %d actions, got %d", index+1, len(actions))
|
|
}
|
|
|
|
action := actions[index]
|
|
if action.GetVerb() != verb {
|
|
t.Errorf("Expected action %d verb to be %s, got %s", index, verb, action.GetVerb())
|
|
}
|
|
|
|
if action.GetResource().Resource != resource {
|
|
t.Errorf("Expected action %d resource to be %s, got %s", index, resource, action.GetResource().Resource)
|
|
}
|
|
}
|
|
|
|
func strPtr(str string) *string {
|
|
return &str
|
|
}
|
|
|
|
func protoPtr(proto v1.Protocol) *v1.Protocol {
|
|
return &proto
|
|
}
|
|
|
|
func int32Ptr(num int32) *int32 {
|
|
return &num
|
|
}
|