Add scheduler benchmark tests for affinity rules.
This adds new benchmark tests that measure scheduler latency of pods that use affinity rules. Specifically, this tests affinity rules with topologyKey="kubernetes.io/hostname".
This commit is contained in:
		@@ -21,7 +21,10 @@ import (
 | 
				
			|||||||
	"testing"
 | 
						"testing"
 | 
				
			||||||
	"time"
 | 
						"time"
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
						"k8s.io/api/core/v1"
 | 
				
			||||||
 | 
						metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
 | 
				
			||||||
	"k8s.io/apimachinery/pkg/labels"
 | 
						"k8s.io/apimachinery/pkg/labels"
 | 
				
			||||||
 | 
						"k8s.io/kubernetes/pkg/kubelet/apis"
 | 
				
			||||||
	"k8s.io/kubernetes/test/integration/framework"
 | 
						"k8s.io/kubernetes/test/integration/framework"
 | 
				
			||||||
	testutils "k8s.io/kubernetes/test/utils"
 | 
						testutils "k8s.io/kubernetes/test/utils"
 | 
				
			||||||
 | 
					
 | 
				
			||||||
@@ -31,24 +34,80 @@ import (
 | 
				
			|||||||
// BenchmarkScheduling benchmarks the scheduling rate when the cluster has
 | 
					// BenchmarkScheduling benchmarks the scheduling rate when the cluster has
 | 
				
			||||||
// various quantities of nodes and scheduled pods.
 | 
					// various quantities of nodes and scheduled pods.
 | 
				
			||||||
func BenchmarkScheduling(b *testing.B) {
 | 
					func BenchmarkScheduling(b *testing.B) {
 | 
				
			||||||
	tests := []struct{ nodes, pods, minOps int }{
 | 
						tests := []struct{ nodes, existingPods, minPods int }{
 | 
				
			||||||
		{nodes: 100, pods: 0, minOps: 100},
 | 
							{nodes: 100, existingPods: 0, minPods: 100},
 | 
				
			||||||
		{nodes: 100, pods: 1000, minOps: 100},
 | 
							{nodes: 100, existingPods: 1000, minPods: 100},
 | 
				
			||||||
		{nodes: 1000, pods: 0, minOps: 100},
 | 
							{nodes: 1000, existingPods: 0, minPods: 100},
 | 
				
			||||||
		{nodes: 1000, pods: 1000, minOps: 100},
 | 
							{nodes: 1000, existingPods: 1000, minPods: 100},
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
 | 
						setupStrategy := testutils.NewSimpleWithControllerCreatePodStrategy("rc1")
 | 
				
			||||||
 | 
						testStrategy := testutils.NewSimpleWithControllerCreatePodStrategy("rc2")
 | 
				
			||||||
	for _, test := range tests {
 | 
						for _, test := range tests {
 | 
				
			||||||
		name := fmt.Sprintf("%vNodes/%vPods", test.nodes, test.pods)
 | 
							name := fmt.Sprintf("%vNodes/%vPods", test.nodes, test.existingPods)
 | 
				
			||||||
		b.Run(name, func(b *testing.B) { benchmarkScheduling(test.nodes, test.pods, test.minOps, b) })
 | 
							b.Run(name, func(b *testing.B) {
 | 
				
			||||||
 | 
								benchmarkScheduling(test.nodes, test.existingPods, test.minPods, setupStrategy, testStrategy, b)
 | 
				
			||||||
 | 
							})
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					// BenchmarkSchedulingAntiAffinity benchmarks the scheduling rate of pods with
 | 
				
			||||||
 | 
					// PodAntiAffinity rules when the cluster has various quantities of nodes and
 | 
				
			||||||
 | 
					// scheduled pods.
 | 
				
			||||||
 | 
					func BenchmarkSchedulingAntiAffinity(b *testing.B) {
 | 
				
			||||||
 | 
						tests := []struct{ nodes, existingPods, minPods int }{
 | 
				
			||||||
 | 
							{nodes: 500, existingPods: 250, minPods: 250},
 | 
				
			||||||
 | 
							{nodes: 500, existingPods: 5000, minPods: 250},
 | 
				
			||||||
 | 
						}
 | 
				
			||||||
 | 
						// The setup strategy creates pods with no affinity rules.
 | 
				
			||||||
 | 
						setupStrategy := testutils.NewSimpleWithControllerCreatePodStrategy("setup")
 | 
				
			||||||
 | 
						// The test strategy creates pods with anti-affinity for each other.
 | 
				
			||||||
 | 
						testBasePod := makeBasePodWithAntiAffinity(
 | 
				
			||||||
 | 
							map[string]string{"name": "test", "color": "green"},
 | 
				
			||||||
 | 
							map[string]string{"color": "green"})
 | 
				
			||||||
 | 
						testStrategy := testutils.NewCustomCreatePodStrategy(testBasePod)
 | 
				
			||||||
 | 
						for _, test := range tests {
 | 
				
			||||||
 | 
							name := fmt.Sprintf("%vNodes/%vPods", test.nodes, test.existingPods)
 | 
				
			||||||
 | 
							b.Run(name, func(b *testing.B) {
 | 
				
			||||||
 | 
								benchmarkScheduling(test.nodes, test.existingPods, test.minPods, setupStrategy, testStrategy, b)
 | 
				
			||||||
 | 
							})
 | 
				
			||||||
 | 
						}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					// makeBasePodWithAntiAffinity creates a Pod object to be used as a template.
 | 
				
			||||||
 | 
					// The Pod has a PodAntiAffinity requirement against pods with the given labels.
 | 
				
			||||||
 | 
					func makeBasePodWithAntiAffinity(podLabels, affinityLabels map[string]string) *v1.Pod {
 | 
				
			||||||
 | 
						basePod := &v1.Pod{
 | 
				
			||||||
 | 
							ObjectMeta: metav1.ObjectMeta{
 | 
				
			||||||
 | 
								GenerateName: "affinity-pod-",
 | 
				
			||||||
 | 
								Labels:       podLabels,
 | 
				
			||||||
 | 
							},
 | 
				
			||||||
 | 
							Spec: testutils.MakePodSpec(),
 | 
				
			||||||
 | 
						}
 | 
				
			||||||
 | 
						basePod.Spec.Affinity = &v1.Affinity{
 | 
				
			||||||
 | 
							PodAntiAffinity: &v1.PodAntiAffinity{
 | 
				
			||||||
 | 
								RequiredDuringSchedulingIgnoredDuringExecution: []v1.PodAffinityTerm{
 | 
				
			||||||
 | 
									{
 | 
				
			||||||
 | 
										LabelSelector: &metav1.LabelSelector{
 | 
				
			||||||
 | 
											MatchLabels: affinityLabels,
 | 
				
			||||||
 | 
										},
 | 
				
			||||||
 | 
										TopologyKey: apis.LabelHostname,
 | 
				
			||||||
 | 
									},
 | 
				
			||||||
 | 
								},
 | 
				
			||||||
 | 
							},
 | 
				
			||||||
 | 
						}
 | 
				
			||||||
 | 
						return basePod
 | 
				
			||||||
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
// benchmarkScheduling benchmarks scheduling rate with specific number of nodes
 | 
					// benchmarkScheduling benchmarks scheduling rate with specific number of nodes
 | 
				
			||||||
// and specific number of pods already scheduled.
 | 
					// and specific number of pods already scheduled.
 | 
				
			||||||
// Since an operation typically takes more than 1 second, we put a minimum bound on b.N of minOps.
 | 
					// This will schedule numExistingPods pods before the benchmark starts, and at
 | 
				
			||||||
func benchmarkScheduling(numNodes, numScheduledPods, minOps int, b *testing.B) {
 | 
					// least minPods pods during the benchmark.
 | 
				
			||||||
	if b.N < minOps {
 | 
					func benchmarkScheduling(numNodes, numExistingPods, minPods int,
 | 
				
			||||||
		b.N = minOps
 | 
						setupPodStrategy, testPodStrategy testutils.TestPodCreateStrategy,
 | 
				
			||||||
 | 
						b *testing.B) {
 | 
				
			||||||
 | 
						if b.N < minPods {
 | 
				
			||||||
 | 
							b.N = minPods
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
	schedulerConfigFactory, finalFunc := mustSetupScheduler()
 | 
						schedulerConfigFactory, finalFunc := mustSetupScheduler()
 | 
				
			||||||
	defer finalFunc()
 | 
						defer finalFunc()
 | 
				
			||||||
@@ -65,7 +124,7 @@ func benchmarkScheduling(numNodes, numScheduledPods, minOps int, b *testing.B) {
 | 
				
			|||||||
	defer nodePreparer.CleanupNodes()
 | 
						defer nodePreparer.CleanupNodes()
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	config := testutils.NewTestPodCreatorConfig()
 | 
						config := testutils.NewTestPodCreatorConfig()
 | 
				
			||||||
	config.AddStrategy("sched-test", numScheduledPods, testutils.NewSimpleWithControllerCreatePodStrategy("rc1"))
 | 
						config.AddStrategy("sched-test", numExistingPods, setupPodStrategy)
 | 
				
			||||||
	podCreator := testutils.NewTestPodCreator(c, config)
 | 
						podCreator := testutils.NewTestPodCreator(c, config)
 | 
				
			||||||
	podCreator.CreatePods()
 | 
						podCreator.CreatePods()
 | 
				
			||||||
 | 
					
 | 
				
			||||||
@@ -74,7 +133,7 @@ func benchmarkScheduling(numNodes, numScheduledPods, minOps int, b *testing.B) {
 | 
				
			|||||||
		if err != nil {
 | 
							if err != nil {
 | 
				
			||||||
			glog.Fatalf("%v", err)
 | 
								glog.Fatalf("%v", err)
 | 
				
			||||||
		}
 | 
							}
 | 
				
			||||||
		if len(scheduled) >= numScheduledPods {
 | 
							if len(scheduled) >= numExistingPods {
 | 
				
			||||||
			break
 | 
								break
 | 
				
			||||||
		}
 | 
							}
 | 
				
			||||||
		time.Sleep(1 * time.Second)
 | 
							time.Sleep(1 * time.Second)
 | 
				
			||||||
@@ -82,7 +141,7 @@ func benchmarkScheduling(numNodes, numScheduledPods, minOps int, b *testing.B) {
 | 
				
			|||||||
	// start benchmark
 | 
						// start benchmark
 | 
				
			||||||
	b.ResetTimer()
 | 
						b.ResetTimer()
 | 
				
			||||||
	config = testutils.NewTestPodCreatorConfig()
 | 
						config = testutils.NewTestPodCreatorConfig()
 | 
				
			||||||
	config.AddStrategy("sched-test", b.N, testutils.NewSimpleWithControllerCreatePodStrategy("rc2"))
 | 
						config.AddStrategy("sched-test", b.N, testPodStrategy)
 | 
				
			||||||
	podCreator = testutils.NewTestPodCreator(c, config)
 | 
						podCreator = testutils.NewTestPodCreator(c, config)
 | 
				
			||||||
	podCreator.CreatePods()
 | 
						podCreator.CreatePods()
 | 
				
			||||||
	for {
 | 
						for {
 | 
				
			||||||
@@ -92,7 +151,7 @@ func benchmarkScheduling(numNodes, numScheduledPods, minOps int, b *testing.B) {
 | 
				
			|||||||
		if err != nil {
 | 
							if err != nil {
 | 
				
			||||||
			glog.Fatalf("%v", err)
 | 
								glog.Fatalf("%v", err)
 | 
				
			||||||
		}
 | 
							}
 | 
				
			||||||
		if len(scheduled) >= numScheduledPods+b.N {
 | 
							if len(scheduled) >= numExistingPods+b.N {
 | 
				
			||||||
			break
 | 
								break
 | 
				
			||||||
		}
 | 
							}
 | 
				
			||||||
		// Note: This might introduce slight deviation in accuracy of benchmark results.
 | 
							// Note: This might introduce slight deviation in accuracy of benchmark results.
 | 
				
			||||||
 
 | 
				
			|||||||
		Reference in New Issue
	
	Block a user