From 12d49b6bfbd0f3746b4a215cc8fc43548bc07a40 Mon Sep 17 00:00:00 2001 From: sivchari Date: Fri, 26 Aug 2022 00:44:31 +0900 Subject: [PATCH] fix: rename --- pkg/kubelet/cm/cgroup_manager_linux.go | 20 +++++++++---------- pkg/kubelet/cm/helpers_linux.go | 14 ++++++------- .../cm/node_container_manager_linux.go | 6 +++--- pkg/kubelet/cm/qos_container_manager_linux.go | 6 +++--- 4 files changed, 23 insertions(+), 23 deletions(-) diff --git a/pkg/kubelet/cm/cgroup_manager_linux.go b/pkg/kubelet/cm/cgroup_manager_linux.go index fa36d6bc1d7..ef5560f4b3d 100644 --- a/pkg/kubelet/cm/cgroup_manager_linux.go +++ b/pkg/kubelet/cm/cgroup_manager_linux.go @@ -307,8 +307,8 @@ func (m *cgroupManagerImpl) Destroy(cgroupConfig *CgroupConfig) error { return nil } -// getCpuWeight converts from the range [2, 262144] to [1, 10000] -func getCpuWeight(cpuShares *uint64) uint64 { +// getCPUWeight converts from the range [2, 262144] to [1, 10000] +func getCPUWeight(cpuShares *uint64) uint64 { if cpuShares == nil { return 0 } @@ -360,18 +360,18 @@ func (m *cgroupManagerImpl) toResources(resourceConfig *ResourceConfig) *libcont if resourceConfig.Memory != nil { resources.Memory = *resourceConfig.Memory } - if resourceConfig.CpuShares != nil { + if resourceConfig.CPUShares != nil { if libcontainercgroups.IsCgroup2UnifiedMode() { - resources.CpuWeight = getCpuWeight(resourceConfig.CpuShares) + resources.CpuWeight = getCPUWeight(resourceConfig.CPUShares) } else { - resources.CpuShares = *resourceConfig.CpuShares + resources.CpuShares = *resourceConfig.CPUShares } } - if resourceConfig.CpuQuota != nil { - resources.CpuQuota = *resourceConfig.CpuQuota + if resourceConfig.CPUQuota != nil { + resources.CpuQuota = *resourceConfig.CPUQuota } - if resourceConfig.CpuPeriod != nil { - resources.CpuPeriod = *resourceConfig.CpuPeriod + if resourceConfig.CPUPeriod != nil { + resources.CpuPeriod = *resourceConfig.CPUPeriod } if resourceConfig.PidsLimit != nil { resources.PidsLimit = *resourceConfig.PidsLimit @@ -531,7 +531,7 @@ func (m *cgroupManagerImpl) ReduceCPULimits(cgroupName CgroupName) error { // Set lowest possible CpuShares value for the cgroup minimumCPUShares := uint64(MinShares) resources := &ResourceConfig{ - CpuShares: &minimumCPUShares, + CPUShares: &minimumCPUShares, } containerConfig := &CgroupConfig{ Name: cgroupName, diff --git a/pkg/kubelet/cm/helpers_linux.go b/pkg/kubelet/cm/helpers_linux.go index 25ff3f13b82..96546f47be0 100644 --- a/pkg/kubelet/cm/helpers_linux.go +++ b/pkg/kubelet/cm/helpers_linux.go @@ -182,22 +182,22 @@ func ResourceConfigForPod(pod *v1.Pod, enforceCPULimits bool, cpuPeriod uint64, // build the result result := &ResourceConfig{} if qosClass == v1.PodQOSGuaranteed { - result.CpuShares = &cpuShares - result.CpuQuota = &cpuQuota - result.CpuPeriod = &cpuPeriod + result.CPUShares = &cpuShares + result.CPUQuota = &cpuQuota + result.CPUPeriod = &cpuPeriod result.Memory = &memoryLimits } else if qosClass == v1.PodQOSBurstable { - result.CpuShares = &cpuShares + result.CPUShares = &cpuShares if cpuLimitsDeclared { - result.CpuQuota = &cpuQuota - result.CpuPeriod = &cpuPeriod + result.CPUQuota = &cpuQuota + result.CPUPeriod = &cpuPeriod } if memoryLimitsDeclared { result.Memory = &memoryLimits } } else { shares := uint64(MinShares) - result.CpuShares = &shares + result.CPUShares = &shares } result.HugePageLimit = hugePageLimits diff --git a/pkg/kubelet/cm/node_container_manager_linux.go b/pkg/kubelet/cm/node_container_manager_linux.go index 4b35d3c4fe3..74221c67047 100644 --- a/pkg/kubelet/cm/node_container_manager_linux.go +++ b/pkg/kubelet/cm/node_container_manager_linux.go @@ -40,7 +40,7 @@ const ( defaultNodeAllocatableCgroupName = "kubepods" ) -//createNodeAllocatableCgroups creates Node Allocatable Cgroup when CgroupsPerQOS flag is specified as true +// createNodeAllocatableCgroups creates Node Allocatable Cgroup when CgroupsPerQOS flag is specified as true func (cm *containerManagerImpl) createNodeAllocatableCgroups() error { nodeAllocatable := cm.internalCapacity // Use Node Allocatable limits instead of capacity if the user requested enforcing node allocatable. @@ -155,7 +155,7 @@ func enforceExistingCgroup(cgroupManager CgroupManager, cName CgroupName, rl v1. Name: cName, ResourceParameters: rp, } - klog.V(4).InfoS("Enforcing limits on cgroup", "cgroupName", cName, "cpuShares", cgroupConfig.ResourceParameters.CpuShares, "memory", cgroupConfig.ResourceParameters.Memory, "pidsLimit", cgroupConfig.ResourceParameters.PidsLimit) + klog.V(4).InfoS("Enforcing limits on cgroup", "cgroupName", cName, "cpuShares", cgroupConfig.ResourceParameters.CPUShares, "memory", cgroupConfig.ResourceParameters.Memory, "pidsLimit", cgroupConfig.ResourceParameters.PidsLimit) if err := cgroupManager.Validate(cgroupConfig.Name); err != nil { return err } @@ -180,7 +180,7 @@ func getCgroupConfig(rl v1.ResourceList) *ResourceConfig { if q, exists := rl[v1.ResourceCPU]; exists { // CPU is defined in milli-cores. val := MilliCPUToShares(q.MilliValue()) - rc.CpuShares = &val + rc.CPUShares = &val } if q, exists := rl[pidlimit.PIDs]; exists { val := q.Value() diff --git a/pkg/kubelet/cm/qos_container_manager_linux.go b/pkg/kubelet/cm/qos_container_manager_linux.go index 0ddd44ac234..1525087a33e 100644 --- a/pkg/kubelet/cm/qos_container_manager_linux.go +++ b/pkg/kubelet/cm/qos_container_manager_linux.go @@ -98,7 +98,7 @@ func (m *qosContainerManagerImpl) Start(getNodeAllocatable func() v1.ResourceLis // the BestEffort QoS class has a statically configured minShares value if qosClass == v1.PodQOSBestEffort { minShares := uint64(MinShares) - resourceParameters.CpuShares = &minShares + resourceParameters.CPUShares = &minShares } // containerConfig object stores the cgroup specifications @@ -184,11 +184,11 @@ func (m *qosContainerManagerImpl) setCPUCgroupConfig(configs map[v1.PodQOSClass] // make sure best effort is always 2 shares bestEffortCPUShares := uint64(MinShares) - configs[v1.PodQOSBestEffort].ResourceParameters.CpuShares = &bestEffortCPUShares + configs[v1.PodQOSBestEffort].ResourceParameters.CPUShares = &bestEffortCPUShares // set burstable shares based on current observe state burstableCPUShares := MilliCPUToShares(burstablePodCPURequest) - configs[v1.PodQOSBurstable].ResourceParameters.CpuShares = &burstableCPUShares + configs[v1.PodQOSBurstable].ResourceParameters.CPUShares = &burstableCPUShares return nil }