forked from kubernetes/kubernetes
-
Notifications
You must be signed in to change notification settings - Fork 0
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
Added an end-to-end test measuring autoscaling's efficiency.
- Loading branch information
1 parent
38f175f
commit ce9f3bc
Showing
2 changed files
with
112 additions
and
0 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,111 @@ | ||
/* | ||
Copyright 2017 The Kubernetes Authors. | ||
Licensed under the Apache License, Version 2.0 (the "License"); | ||
you may not use this file except in compliance with the License. | ||
You may obtain a copy of the License at | ||
http://www.apache.org/licenses/LICENSE-2.0 | ||
Unless required by applicable law or agreed to in writing, software | ||
distributed under the License is distributed on an "AS IS" BASIS, | ||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
See the License for the specific language governing permissions and | ||
limitations under the License. | ||
*/ | ||
|
||
package autoscaling | ||
|
||
import ( | ||
"strings" | ||
"time" | ||
|
||
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" | ||
"k8s.io/kubernetes/pkg/api/v1" | ||
"k8s.io/kubernetes/test/e2e/common" | ||
"k8s.io/kubernetes/test/e2e/framework" | ||
|
||
. "github.com/onsi/ginkgo" | ||
. "github.com/onsi/gomega" | ||
) | ||
|
||
var _ = framework.KubeDescribe("[Feature:ClusterSizeAutoscalingScaleUp] [Slow] Autoscaling", func() { | ||
f := framework.NewDefaultFramework("autoscaling") | ||
|
||
framework.KubeDescribe("Autoscaling a service", func() { | ||
BeforeEach(func() { | ||
// Check if Cloud Autoscaler is enabled by trying to get its ConfigMap. | ||
_, err := f.ClientSet.CoreV1().ConfigMaps("kube-system").Get("cluster-autoscaler-status", metav1.GetOptions{}) | ||
if err != nil { | ||
framework.Skipf("test expects Cluster Autoscaler to be enabled") | ||
} | ||
}) | ||
|
||
Context("from 1 pod and 3 nodes to 8 pods and >=4 nodes", func() { | ||
const nodesNum = 3 // Expect there to be 3 nodes before and after the test. | ||
var nodeGroupName string // Set by BeforeEach, used by AfterEach to scale this node group down after the test. | ||
var nodes *v1.NodeList // Set by BeforeEach, used by Measure to calculate CPU request based on node's sizes. | ||
|
||
BeforeEach(func() { | ||
// Make sure there is only 1 node group, otherwise this test becomes useless. | ||
nodeGroups := strings.Split(framework.TestContext.CloudConfig.NodeInstanceGroup, ",") | ||
if len(nodeGroups) != 1 { | ||
framework.Skipf("test expects 1 node group, found %d", len(nodeGroups)) | ||
} | ||
nodeGroupName = nodeGroups[0] | ||
|
||
// Make sure the node group has exactly 'nodesNum' nodes, otherwise this test becomes useless. | ||
nodeGroupSize, err := framework.GroupSize(nodeGroupName) | ||
framework.ExpectNoError(err) | ||
if nodeGroupSize != nodesNum { | ||
framework.Skipf("test expects %d nodes, found %d", nodesNum, nodeGroupSize) | ||
} | ||
|
||
// Make sure all nodes are schedulable, otherwise we are in some kind of a problem state. | ||
nodes = framework.GetReadySchedulableNodesOrDie(f.ClientSet) | ||
schedulableCount := len(nodes.Items) | ||
Expect(schedulableCount).To(Equal(nodeGroupSize), "not all nodes are schedulable") | ||
}) | ||
|
||
AfterEach(func() { | ||
// Scale down back to only 'nodesNum' nodes, as expected at the start of the test. | ||
framework.ExpectNoError(framework.ResizeGroup(nodeGroupName, nodesNum)) | ||
framework.ExpectNoError(framework.WaitForClusterSize(f.ClientSet, nodesNum, 15*time.Minute)) | ||
}) | ||
|
||
Measure("takes less than 15 minutes", func(b Benchmarker) { | ||
// Measured over multiple samples, scaling takes 10 +/- 2 minutes, so 15 minutes should be fully sufficient. | ||
const timeToWait = 15 * time.Minute | ||
|
||
// Calculate the CPU request of the service. | ||
// This test expects that 8 pods will not fit in 'nodesNum' nodes, but will fit in >='nodesNum'+1 nodes. | ||
// Make it so that 'nodesNum' pods fit perfectly per node (in practice other things take space, so less than that will fit). | ||
nodeCpus := nodes.Items[0].Status.Capacity[v1.ResourceCPU] | ||
nodeCpuMillis := (&nodeCpus).MilliValue() | ||
cpuRequestMillis := int64(nodeCpuMillis / nodesNum) | ||
|
||
// Start the service we want to scale and wait for it to be up and running. | ||
nodeMemoryBytes := nodes.Items[0].Status.Capacity[v1.ResourceMemory] | ||
nodeMemoryMB := (&nodeMemoryBytes).Value() / 1024 / 1024 | ||
memRequestMB := nodeMemoryMB / 10 // Ensure each pod takes not more than 10% of node's total memory. | ||
replicas := 1 | ||
resourceConsumer := common.NewDynamicResourceConsumer("resource-consumer", common.KindDeployment, replicas, 0, 0, 0, cpuRequestMillis, memRequestMB, f) | ||
defer resourceConsumer.CleanUp() | ||
resourceConsumer.WaitForReplicas(replicas, 1*time.Minute) // Should finish ~immediately, so 1 minute is more than enough. | ||
|
||
// Enable Horizontal Pod Autoscaler with 50% target utilization and | ||
// scale up the CPU usage to trigger autoscaling to 8 pods for target to be satisfied. | ||
targetCpuUtilizationPercent := int32(50) | ||
hpa := common.CreateCPUHorizontalPodAutoscaler(resourceConsumer, targetCpuUtilizationPercent, 1, 10) | ||
defer common.DeleteHorizontalPodAutoscaler(resourceConsumer, hpa.Name) | ||
cpuLoad := 8 * cpuRequestMillis * int64(targetCpuUtilizationPercent) / 100 // 8 pods utilized to the target level | ||
resourceConsumer.ConsumeCPU(int(cpuLoad)) | ||
|
||
// Measure the time it takes for the service to scale to 8 pods with 50% CPU utilization each. | ||
b.Time("total scale-up time", func() { | ||
resourceConsumer.WaitForReplicas(8, timeToWait) | ||
}) | ||
}, 1) // Increase to run the test more than once. | ||
}) | ||
}) | ||
}) |