mirror of
https://github.com/kubernetes-sigs/descheduler.git
synced 2026-01-28 14:41:10 +01:00
Compare commits
1 Commits
deschedule
...
deschedule
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
6bfc76b6b2 |
6
.github/ci/ct.yaml
vendored
6
.github/ci/ct.yaml
vendored
@@ -1,6 +0,0 @@
|
|||||||
chart-dirs:
|
|
||||||
- charts
|
|
||||||
helm-extra-args: "--timeout=5m"
|
|
||||||
check-version-increment: false
|
|
||||||
helm-extra-set-args: "--set=kind=Deployment"
|
|
||||||
target-branch: master
|
|
||||||
67
.github/workflows/helm.yaml
vendored
67
.github/workflows/helm.yaml
vendored
@@ -1,67 +0,0 @@
|
|||||||
name: Helm
|
|
||||||
|
|
||||||
on:
|
|
||||||
push:
|
|
||||||
branches:
|
|
||||||
- master
|
|
||||||
- release-*
|
|
||||||
paths:
|
|
||||||
- 'charts/**'
|
|
||||||
- '.github/workflows/helm.yaml'
|
|
||||||
pull_request:
|
|
||||||
paths:
|
|
||||||
- 'charts/**'
|
|
||||||
- '.github/workflows/helm.yaml'
|
|
||||||
|
|
||||||
jobs:
|
|
||||||
lint-and-test:
|
|
||||||
runs-on: ubuntu-latest
|
|
||||||
steps:
|
|
||||||
- name: Checkout
|
|
||||||
uses: actions/checkout@v3
|
|
||||||
with:
|
|
||||||
fetch-depth: 0
|
|
||||||
|
|
||||||
- name: Set up Helm
|
|
||||||
uses: azure/setup-helm@v2.1
|
|
||||||
with:
|
|
||||||
version: v3.9.2
|
|
||||||
|
|
||||||
- uses: actions/setup-python@v3.1.2
|
|
||||||
with:
|
|
||||||
python-version: 3.7
|
|
||||||
|
|
||||||
- uses: actions/setup-go@v3
|
|
||||||
with:
|
|
||||||
go-version: '1.19.0'
|
|
||||||
|
|
||||||
- name: Set up chart-testing
|
|
||||||
uses: helm/chart-testing-action@v2.2.1
|
|
||||||
with:
|
|
||||||
version: v3.7.0
|
|
||||||
|
|
||||||
- name: Run chart-testing (list-changed)
|
|
||||||
id: list-changed
|
|
||||||
run: |
|
|
||||||
changed=$(ct list-changed --config=.github/ci/ct.yaml)
|
|
||||||
if [[ -n "$changed" ]]; then
|
|
||||||
echo "::set-output name=changed::true"
|
|
||||||
fi
|
|
||||||
|
|
||||||
- name: Run chart-testing (lint)
|
|
||||||
run: ct lint --config=.github/ci/ct.yaml --validate-maintainers=false
|
|
||||||
|
|
||||||
# Need a multi node cluster so descheduler runs until evictions
|
|
||||||
- name: Create multi node Kind cluster
|
|
||||||
run: make kind-multi-node
|
|
||||||
|
|
||||||
# helm-extra-set-args only available after ct 3.6.0
|
|
||||||
- name: Run chart-testing (install)
|
|
||||||
run: ct install --config=.github/ci/ct.yaml
|
|
||||||
|
|
||||||
- name: E2E after chart install
|
|
||||||
env:
|
|
||||||
KUBERNETES_VERSION: "v1.25.0"
|
|
||||||
KIND_E2E: true
|
|
||||||
SKIP_INSTALL: true
|
|
||||||
run: make test-e2e
|
|
||||||
47
.github/workflows/security.yaml
vendored
47
.github/workflows/security.yaml
vendored
@@ -1,47 +0,0 @@
|
|||||||
name: "Security"
|
|
||||||
|
|
||||||
on:
|
|
||||||
push:
|
|
||||||
branches:
|
|
||||||
- main
|
|
||||||
- master
|
|
||||||
- release-*
|
|
||||||
schedule:
|
|
||||||
- cron: '30 1 * * 0'
|
|
||||||
|
|
||||||
jobs:
|
|
||||||
analyze:
|
|
||||||
name: Analyze
|
|
||||||
runs-on: ubuntu-latest
|
|
||||||
permissions:
|
|
||||||
actions: read
|
|
||||||
contents: read
|
|
||||||
security-events: write
|
|
||||||
|
|
||||||
strategy:
|
|
||||||
fail-fast: false
|
|
||||||
steps:
|
|
||||||
- name: Checkout
|
|
||||||
uses: actions/checkout@v3
|
|
||||||
with:
|
|
||||||
fetch-depth: 0
|
|
||||||
|
|
||||||
- name: Build image
|
|
||||||
run: |
|
|
||||||
IMAGE_REPO=${HELM_IMAGE_REPO:-descheduler}
|
|
||||||
IMAGE_TAG=${HELM_IMAGE_TAG:-security-test}
|
|
||||||
VERSION=security-test make image
|
|
||||||
- name: Run Trivy vulnerability scanner
|
|
||||||
uses: aquasecurity/trivy-action@master
|
|
||||||
with:
|
|
||||||
image-ref: 'descheduler:security-test'
|
|
||||||
format: 'sarif'
|
|
||||||
exit-code: '0'
|
|
||||||
severity: 'CRITICAL,HIGH'
|
|
||||||
output: 'trivy-results.sarif'
|
|
||||||
|
|
||||||
- name: Upload Trivy scan results to GitHub Security tab
|
|
||||||
uses: github/codeql-action/upload-sarif@v2
|
|
||||||
with:
|
|
||||||
sarif_file: 'trivy-results.sarif'
|
|
||||||
exit-code: '0'
|
|
||||||
3
.gitignore
vendored
3
.gitignore
vendored
@@ -4,5 +4,4 @@ vendordiff.patch
|
|||||||
.idea/
|
.idea/
|
||||||
*.code-workspace
|
*.code-workspace
|
||||||
.vscode/
|
.vscode/
|
||||||
kind
|
kind
|
||||||
bin/
|
|
||||||
@@ -11,7 +11,7 @@
|
|||||||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
# See the License for the specific language governing permissions and
|
# See the License for the specific language governing permissions and
|
||||||
# limitations under the License.
|
# limitations under the License.
|
||||||
FROM golang:1.19.0
|
FROM golang:1.16.7
|
||||||
|
|
||||||
WORKDIR /go/src/sigs.k8s.io/descheduler
|
WORKDIR /go/src/sigs.k8s.io/descheduler
|
||||||
COPY . .
|
COPY . .
|
||||||
|
|||||||
51
Makefile
51
Makefile
@@ -14,10 +14,8 @@
|
|||||||
|
|
||||||
.PHONY: test
|
.PHONY: test
|
||||||
|
|
||||||
export CONTAINER_ENGINE ?= docker
|
|
||||||
|
|
||||||
# VERSION is based on a date stamp plus the last commit
|
# VERSION is based on a date stamp plus the last commit
|
||||||
VERSION?=v$(shell date +%Y%m%d)-$(shell git describe --tags)
|
VERSION?=v$(shell date +%Y%m%d)-$(shell git describe --tags --match "v*")
|
||||||
BRANCH?=$(shell git branch --show-current)
|
BRANCH?=$(shell git branch --show-current)
|
||||||
SHA1?=$(shell git rev-parse HEAD)
|
SHA1?=$(shell git rev-parse HEAD)
|
||||||
BUILD=$(shell date +%FT%T%z)
|
BUILD=$(shell date +%FT%T%z)
|
||||||
@@ -26,7 +24,7 @@ ARCHS = amd64 arm arm64
|
|||||||
|
|
||||||
LDFLAGS=-ldflags "-X ${LDFLAG_LOCATION}.version=${VERSION} -X ${LDFLAG_LOCATION}.buildDate=${BUILD} -X ${LDFLAG_LOCATION}.gitbranch=${BRANCH} -X ${LDFLAG_LOCATION}.gitsha1=${SHA1}"
|
LDFLAGS=-ldflags "-X ${LDFLAG_LOCATION}.version=${VERSION} -X ${LDFLAG_LOCATION}.buildDate=${BUILD} -X ${LDFLAG_LOCATION}.gitbranch=${BRANCH} -X ${LDFLAG_LOCATION}.gitsha1=${SHA1}"
|
||||||
|
|
||||||
GOLANGCI_VERSION := v1.49.0
|
GOLANGCI_VERSION := v1.30.0
|
||||||
HAS_GOLANGCI := $(shell ls _output/bin/golangci-lint 2> /dev/null)
|
HAS_GOLANGCI := $(shell ls _output/bin/golangci-lint 2> /dev/null)
|
||||||
|
|
||||||
# REGISTRY is the container registry to push
|
# REGISTRY is the container registry to push
|
||||||
@@ -62,36 +60,36 @@ build.arm64:
|
|||||||
CGO_ENABLED=0 GOOS=linux GOARCH=arm64 go build ${LDFLAGS} -o _output/bin/descheduler sigs.k8s.io/descheduler/cmd/descheduler
|
CGO_ENABLED=0 GOOS=linux GOARCH=arm64 go build ${LDFLAGS} -o _output/bin/descheduler sigs.k8s.io/descheduler/cmd/descheduler
|
||||||
|
|
||||||
dev-image: build
|
dev-image: build
|
||||||
$(CONTAINER_ENGINE) build -f Dockerfile.dev -t $(IMAGE) .
|
docker build -f Dockerfile.dev -t $(IMAGE) .
|
||||||
|
|
||||||
image:
|
image:
|
||||||
$(CONTAINER_ENGINE) build --build-arg VERSION="$(VERSION)" --build-arg ARCH="amd64" -t $(IMAGE) .
|
docker build --build-arg VERSION="$(VERSION)" --build-arg ARCH="amd64" -t $(IMAGE) .
|
||||||
|
|
||||||
image.amd64:
|
image.amd64:
|
||||||
$(CONTAINER_ENGINE) build --build-arg VERSION="$(VERSION)" --build-arg ARCH="amd64" -t $(IMAGE)-amd64 .
|
docker build --build-arg VERSION="$(VERSION)" --build-arg ARCH="amd64" -t $(IMAGE)-amd64 .
|
||||||
|
|
||||||
image.arm:
|
image.arm:
|
||||||
$(CONTAINER_ENGINE) build --build-arg VERSION="$(VERSION)" --build-arg ARCH="arm" -t $(IMAGE)-arm .
|
docker build --build-arg VERSION="$(VERSION)" --build-arg ARCH="arm" -t $(IMAGE)-arm .
|
||||||
|
|
||||||
image.arm64:
|
image.arm64:
|
||||||
$(CONTAINER_ENGINE) build --build-arg VERSION="$(VERSION)" --build-arg ARCH="arm64" -t $(IMAGE)-arm64 .
|
docker build --build-arg VERSION="$(VERSION)" --build-arg ARCH="arm64" -t $(IMAGE)-arm64 .
|
||||||
|
|
||||||
push: image
|
push: image
|
||||||
gcloud auth configure-docker
|
gcloud auth configure-docker
|
||||||
$(CONTAINER_ENGINE) tag $(IMAGE) $(IMAGE_GCLOUD)
|
docker tag $(IMAGE) $(IMAGE_GCLOUD)
|
||||||
$(CONTAINER_ENGINE) push $(IMAGE_GCLOUD)
|
docker push $(IMAGE_GCLOUD)
|
||||||
|
|
||||||
push-all: image.amd64 image.arm image.arm64
|
push-all: image.amd64 image.arm image.arm64
|
||||||
gcloud auth configure-docker
|
gcloud auth configure-docker
|
||||||
for arch in $(ARCHS); do \
|
for arch in $(ARCHS); do \
|
||||||
$(CONTAINER_ENGINE) tag $(IMAGE)-$${arch} $(IMAGE_GCLOUD)-$${arch} ;\
|
docker tag $(IMAGE)-$${arch} $(IMAGE_GCLOUD)-$${arch} ;\
|
||||||
$(CONTAINER_ENGINE) push $(IMAGE_GCLOUD)-$${arch} ;\
|
docker push $(IMAGE_GCLOUD)-$${arch} ;\
|
||||||
done
|
done
|
||||||
DOCKER_CLI_EXPERIMENTAL=enabled $(CONTAINER_ENGINE) manifest create $(IMAGE_GCLOUD) $(addprefix --amend $(IMAGE_GCLOUD)-, $(ARCHS))
|
DOCKER_CLI_EXPERIMENTAL=enabled docker manifest create $(IMAGE_GCLOUD) $(addprefix --amend $(IMAGE_GCLOUD)-, $(ARCHS))
|
||||||
for arch in $(ARCHS); do \
|
for arch in $(ARCHS); do \
|
||||||
DOCKER_CLI_EXPERIMENTAL=enabled $(CONTAINER_ENGINE) manifest annotate --arch $${arch} $(IMAGE_GCLOUD) $(IMAGE_GCLOUD)-$${arch} ;\
|
DOCKER_CLI_EXPERIMENTAL=enabled docker manifest annotate --arch $${arch} $(IMAGE_GCLOUD) $(IMAGE_GCLOUD)-$${arch} ;\
|
||||||
done
|
done
|
||||||
DOCKER_CLI_EXPERIMENTAL=enabled $(CONTAINER_ENGINE) manifest push $(IMAGE_GCLOUD) ;\
|
DOCKER_CLI_EXPERIMENTAL=enabled docker manifest push $(IMAGE_GCLOUD) ;\
|
||||||
|
|
||||||
clean:
|
clean:
|
||||||
rm -rf _output
|
rm -rf _output
|
||||||
@@ -133,28 +131,17 @@ verify-gen:
|
|||||||
|
|
||||||
lint:
|
lint:
|
||||||
ifndef HAS_GOLANGCI
|
ifndef HAS_GOLANGCI
|
||||||
curl -sSfL https://raw.githubusercontent.com/golangci/golangci-lint/master/install.sh | sh -s -- -b ./_output/bin ${GOLANGCI_VERSION}
|
curl -sfL https://install.goreleaser.com/github.com/golangci/golangci-lint.sh | sh -s -- -b ./_output/bin ${GOLANGCI_VERSION}
|
||||||
endif
|
endif
|
||||||
./_output/bin/golangci-lint run
|
./_output/bin/golangci-lint run
|
||||||
|
|
||||||
# helm
|
|
||||||
|
|
||||||
ensure-helm-install:
|
|
||||||
ifndef HAS_HELM
|
|
||||||
curl -fsSL -o get_helm.sh https://raw.githubusercontent.com/helm/helm/master/scripts/get-helm-3 && chmod 700 ./get_helm.sh && ./get_helm.sh
|
|
||||||
endif
|
|
||||||
|
|
||||||
lint-chart: ensure-helm-install
|
lint-chart: ensure-helm-install
|
||||||
helm lint ./charts/descheduler
|
helm lint ./charts/descheduler
|
||||||
|
|
||||||
build-helm:
|
|
||||||
helm package ./charts/descheduler --dependency-update --destination ./bin/chart
|
|
||||||
|
|
||||||
test-helm: ensure-helm-install
|
test-helm: ensure-helm-install
|
||||||
./test/run-helm-tests.sh
|
./test/run-helm-tests.sh
|
||||||
|
|
||||||
kind-multi-node:
|
ensure-helm-install:
|
||||||
kind create cluster --name kind --config ./hack/kind_config.yaml --wait 2m
|
ifndef HAS_HELM
|
||||||
|
curl -fsSL -o get_helm.sh https://raw.githubusercontent.com/helm/helm/master/scripts/get-helm-3 && chmod 700 ./get_helm.sh && ./get_helm.sh
|
||||||
ct-helm:
|
endif
|
||||||
./hack/verify-chart.sh
|
|
||||||
6
OWNERS
6
OWNERS
@@ -2,14 +2,14 @@ approvers:
|
|||||||
- damemi
|
- damemi
|
||||||
- ingvagabund
|
- ingvagabund
|
||||||
- seanmalloy
|
- seanmalloy
|
||||||
- a7i
|
|
||||||
reviewers:
|
reviewers:
|
||||||
|
- aveshagarwal
|
||||||
|
- k82cn
|
||||||
|
- ravisantoshgudimetla
|
||||||
- damemi
|
- damemi
|
||||||
- seanmalloy
|
- seanmalloy
|
||||||
- ingvagabund
|
- ingvagabund
|
||||||
- lixiang233
|
- lixiang233
|
||||||
- a7i
|
|
||||||
- janeliul
|
|
||||||
emeritus_approvers:
|
emeritus_approvers:
|
||||||
- aveshagarwal
|
- aveshagarwal
|
||||||
- k82cn
|
- k82cn
|
||||||
|
|||||||
162
README.md
162
README.md
@@ -50,8 +50,6 @@ Table of Contents
|
|||||||
- [Node Fit filtering](#node-fit-filtering)
|
- [Node Fit filtering](#node-fit-filtering)
|
||||||
- [Pod Evictions](#pod-evictions)
|
- [Pod Evictions](#pod-evictions)
|
||||||
- [Pod Disruption Budget (PDB)](#pod-disruption-budget-pdb)
|
- [Pod Disruption Budget (PDB)](#pod-disruption-budget-pdb)
|
||||||
- [High Availability](#high-availability)
|
|
||||||
- [Configure HA Mode](#configure-ha-mode)
|
|
||||||
- [Metrics](#metrics)
|
- [Metrics](#metrics)
|
||||||
- [Compatibility Matrix](#compatibility-matrix)
|
- [Compatibility Matrix](#compatibility-matrix)
|
||||||
- [Getting Involved and Contributing](#getting-involved-and-contributing)
|
- [Getting Involved and Contributing](#getting-involved-and-contributing)
|
||||||
@@ -105,17 +103,17 @@ See the [resources | Kustomize](https://kubectl.docs.kubernetes.io/references/ku
|
|||||||
|
|
||||||
Run As A Job
|
Run As A Job
|
||||||
```
|
```
|
||||||
kustomize build 'github.com/kubernetes-sigs/descheduler/kubernetes/job?ref=v0.25.1' | kubectl apply -f -
|
kustomize build 'github.com/kubernetes-sigs/descheduler/kubernetes/job?ref=v0.22.0' | kubectl apply -f -
|
||||||
```
|
```
|
||||||
|
|
||||||
Run As A CronJob
|
Run As A CronJob
|
||||||
```
|
```
|
||||||
kustomize build 'github.com/kubernetes-sigs/descheduler/kubernetes/cronjob?ref=v0.25.1' | kubectl apply -f -
|
kustomize build 'github.com/kubernetes-sigs/descheduler/kubernetes/cronjob?ref=v0.22.0' | kubectl apply -f -
|
||||||
```
|
```
|
||||||
|
|
||||||
Run As A Deployment
|
Run As A Deployment
|
||||||
```
|
```
|
||||||
kustomize build 'github.com/kubernetes-sigs/descheduler/kubernetes/deployment?ref=v0.25.1' | kubectl apply -f -
|
kustomize build 'github.com/kubernetes-sigs/descheduler/kubernetes/deployment?ref=v0.22.0' | kubectl apply -f -
|
||||||
```
|
```
|
||||||
|
|
||||||
## User Guide
|
## User Guide
|
||||||
@@ -134,8 +132,6 @@ The policy includes a common configuration that applies to all the strategies:
|
|||||||
| `evictSystemCriticalPods` | `false` | [Warning: Will evict Kubernetes system pods] allows eviction of pods with any priority, including system pods like kube-dns |
|
| `evictSystemCriticalPods` | `false` | [Warning: Will evict Kubernetes system pods] allows eviction of pods with any priority, including system pods like kube-dns |
|
||||||
| `ignorePvcPods` | `false` | set whether PVC pods should be evicted or ignored |
|
| `ignorePvcPods` | `false` | set whether PVC pods should be evicted or ignored |
|
||||||
| `maxNoOfPodsToEvictPerNode` | `nil` | maximum number of pods evicted from each node (summed through all strategies) |
|
| `maxNoOfPodsToEvictPerNode` | `nil` | maximum number of pods evicted from each node (summed through all strategies) |
|
||||||
| `maxNoOfPodsToEvictPerNamespace` | `nil` | maximum number of pods evicted from each namespace (summed through all strategies) |
|
|
||||||
| `evictFailedBarePods` | `false` | allow eviction of pods without owner references and in failed phase |
|
|
||||||
|
|
||||||
As part of the policy, the parameters associated with each strategy can be configured.
|
As part of the policy, the parameters associated with each strategy can be configured.
|
||||||
See each strategy for details on available parameters.
|
See each strategy for details on available parameters.
|
||||||
@@ -146,7 +142,6 @@ See each strategy for details on available parameters.
|
|||||||
apiVersion: "descheduler/v1alpha1"
|
apiVersion: "descheduler/v1alpha1"
|
||||||
kind: "DeschedulerPolicy"
|
kind: "DeschedulerPolicy"
|
||||||
nodeSelector: prod=dev
|
nodeSelector: prod=dev
|
||||||
evictFailedBarePods: false
|
|
||||||
evictLocalStoragePods: true
|
evictLocalStoragePods: true
|
||||||
evictSystemCriticalPods: true
|
evictSystemCriticalPods: true
|
||||||
maxNoOfPodsToEvictPerNode: 40
|
maxNoOfPodsToEvictPerNode: 40
|
||||||
@@ -221,17 +216,6 @@ These thresholds, `thresholds` and `targetThresholds`, could be tuned as per you
|
|||||||
strategy evicts pods from `overutilized nodes` (those with usage above `targetThresholds`) to `underutilized nodes`
|
strategy evicts pods from `overutilized nodes` (those with usage above `targetThresholds`) to `underutilized nodes`
|
||||||
(those with usage below `thresholds`), it will abort if any number of `underutilized nodes` or `overutilized nodes` is zero.
|
(those with usage below `thresholds`), it will abort if any number of `underutilized nodes` or `overutilized nodes` is zero.
|
||||||
|
|
||||||
Additionally, the strategy accepts a `useDeviationThresholds` parameter.
|
|
||||||
If that parameter is set to `true`, the thresholds are considered as percentage deviations from mean resource usage.
|
|
||||||
`thresholds` will be deducted from the mean among all nodes and `targetThresholds` will be added to the mean.
|
|
||||||
A resource consumption above (resp. below) this window is considered as overutilization (resp. underutilization).
|
|
||||||
|
|
||||||
**NOTE:** Node resource consumption is determined by the requests and limits of pods, not actual usage.
|
|
||||||
This approach is chosen in order to maintain consistency with the kube-scheduler, which follows the same
|
|
||||||
design for scheduling pods onto nodes. This means that resource usage as reported by Kubelet (or commands
|
|
||||||
like `kubectl top`) may differ from the calculated consumption, due to these components reporting
|
|
||||||
actual usage metrics. Implementing metrics-based descheduling is currently TODO for the project.
|
|
||||||
|
|
||||||
**Parameters:**
|
**Parameters:**
|
||||||
|
|
||||||
|Name|Type|
|
|Name|Type|
|
||||||
@@ -239,7 +223,6 @@ actual usage metrics. Implementing metrics-based descheduling is currently TODO
|
|||||||
|`thresholds`|map(string:int)|
|
|`thresholds`|map(string:int)|
|
||||||
|`targetThresholds`|map(string:int)|
|
|`targetThresholds`|map(string:int)|
|
||||||
|`numberOfNodes`|int|
|
|`numberOfNodes`|int|
|
||||||
|`useDeviationThresholds`|bool|
|
|
||||||
|`thresholdPriority`|int (see [priority filtering](#priority-filtering))|
|
|`thresholdPriority`|int (see [priority filtering](#priority-filtering))|
|
||||||
|`thresholdPriorityClassName`|string (see [priority filtering](#priority-filtering))|
|
|`thresholdPriorityClassName`|string (see [priority filtering](#priority-filtering))|
|
||||||
|`nodeFit`|bool (see [node fit filtering](#node-fit-filtering))|
|
|`nodeFit`|bool (see [node fit filtering](#node-fit-filtering))|
|
||||||
@@ -280,10 +263,10 @@ under utilized frequently or for a short period of time. By default, `numberOfNo
|
|||||||
|
|
||||||
### HighNodeUtilization
|
### HighNodeUtilization
|
||||||
|
|
||||||
This strategy finds nodes that are under utilized and evicts pods from the nodes in the hope that these pods will be
|
This strategy finds nodes that are under utilized and evicts pods from the nodes in the hope that these pods will be
|
||||||
scheduled compactly into fewer nodes. Used in conjunction with node auto-scaling, this strategy is intended to help
|
scheduled compactly into fewer nodes. Used in conjunction with node auto-scaling, this strategy is intended to help
|
||||||
trigger down scaling of under utilized nodes.
|
trigger down scaling of under utilized nodes.
|
||||||
This strategy **must** be used with the scheduler scoring strategy `MostAllocated`. The parameters of this strategy are
|
This strategy **must** be used with the scheduler strategy `MostRequestedPriority`. The parameters of this strategy are
|
||||||
configured under `nodeResourceUtilizationThresholds`.
|
configured under `nodeResourceUtilizationThresholds`.
|
||||||
|
|
||||||
The under utilization of nodes is determined by a configurable threshold `thresholds`. The threshold
|
The under utilization of nodes is determined by a configurable threshold `thresholds`. The threshold
|
||||||
@@ -300,12 +283,6 @@ strategy evicts pods from `underutilized nodes` (those with usage below `thresho
|
|||||||
so that they can be recreated in appropriately utilized nodes.
|
so that they can be recreated in appropriately utilized nodes.
|
||||||
The strategy will abort if any number of `underutilized nodes` or `appropriately utilized nodes` is zero.
|
The strategy will abort if any number of `underutilized nodes` or `appropriately utilized nodes` is zero.
|
||||||
|
|
||||||
**NOTE:** Node resource consumption is determined by the requests and limits of pods, not actual usage.
|
|
||||||
This approach is chosen in order to maintain consistency with the kube-scheduler, which follows the same
|
|
||||||
design for scheduling pods onto nodes. This means that resource usage as reported by Kubelet (or commands
|
|
||||||
like `kubectl top`) may differ from the calculated consumption, due to these components reporting
|
|
||||||
actual usage metrics. Implementing metrics-based descheduling is currently TODO for the project.
|
|
||||||
|
|
||||||
**Parameters:**
|
**Parameters:**
|
||||||
|
|
||||||
|Name|Type|
|
|Name|Type|
|
||||||
@@ -420,17 +397,10 @@ pod "podA" with a toleration to tolerate a taint ``key=value:NoSchedule`` schedu
|
|||||||
node. If the node's taint is subsequently updated/removed, taint is no longer satisfied by its pods' tolerations
|
node. If the node's taint is subsequently updated/removed, taint is no longer satisfied by its pods' tolerations
|
||||||
and will be evicted.
|
and will be evicted.
|
||||||
|
|
||||||
Node taints can be excluded from consideration by specifying a list of excludedTaints. If a node taint key **or**
|
|
||||||
key=value matches an excludedTaints entry, the taint will be ignored.
|
|
||||||
|
|
||||||
For example, excludedTaints entry "dedicated" would match all taints with key "dedicated", regardless of value.
|
|
||||||
excludedTaints entry "dedicated=special-user" would match taints with key "dedicated" and value "special-user".
|
|
||||||
|
|
||||||
**Parameters:**
|
**Parameters:**
|
||||||
|
|
||||||
|Name|Type|
|
|Name|Type|
|
||||||
|---|---|
|
|---|---|
|
||||||
|`excludedTaints`|list(string)|
|
|
||||||
|`thresholdPriority`|int (see [priority filtering](#priority-filtering))|
|
|`thresholdPriority`|int (see [priority filtering](#priority-filtering))|
|
||||||
|`thresholdPriorityClassName`|string (see [priority filtering](#priority-filtering))|
|
|`thresholdPriorityClassName`|string (see [priority filtering](#priority-filtering))|
|
||||||
|`namespaces`|(see [namespace filtering](#namespace-filtering))|
|
|`namespaces`|(see [namespace filtering](#namespace-filtering))|
|
||||||
@@ -445,10 +415,6 @@ kind: "DeschedulerPolicy"
|
|||||||
strategies:
|
strategies:
|
||||||
"RemovePodsViolatingNodeTaints":
|
"RemovePodsViolatingNodeTaints":
|
||||||
enabled: true
|
enabled: true
|
||||||
params:
|
|
||||||
excludedTaints:
|
|
||||||
- dedicated=special-user # exclude taints with key "dedicated" and value "special-user"
|
|
||||||
- reserved # exclude all taints with key "reserved"
|
|
||||||
````
|
````
|
||||||
|
|
||||||
### RemovePodsViolatingTopologySpreadConstraint
|
### RemovePodsViolatingTopologySpreadConstraint
|
||||||
@@ -490,9 +456,9 @@ strategies:
|
|||||||
|
|
||||||
This strategy makes sure that pods having too many restarts are removed from nodes. For example a pod with EBS/PD that
|
This strategy makes sure that pods having too many restarts are removed from nodes. For example a pod with EBS/PD that
|
||||||
can't get the volume/disk attached to the instance, then the pod should be re-scheduled to other nodes. Its parameters
|
can't get the volume/disk attached to the instance, then the pod should be re-scheduled to other nodes. Its parameters
|
||||||
include `podRestartThreshold`, which is the number of restarts (summed over all eligible containers) at which a pod
|
include `podRestartThreshold`, which is the number of restarts at which a pod should be evicted, and `includingInitContainers`,
|
||||||
should be evicted, and `includingInitContainers`, which determines whether init container restarts should be factored
|
which determines whether init container restarts should be factored into that calculation.
|
||||||
into that calculation.
|
|`labelSelector`|(see [label filtering](#label-filtering))|
|
||||||
|
|
||||||
**Parameters:**
|
**Parameters:**
|
||||||
|
|
||||||
@@ -503,7 +469,6 @@ into that calculation.
|
|||||||
|`thresholdPriority`|int (see [priority filtering](#priority-filtering))|
|
|`thresholdPriority`|int (see [priority filtering](#priority-filtering))|
|
||||||
|`thresholdPriorityClassName`|string (see [priority filtering](#priority-filtering))|
|
|`thresholdPriorityClassName`|string (see [priority filtering](#priority-filtering))|
|
||||||
|`namespaces`|(see [namespace filtering](#namespace-filtering))|
|
|`namespaces`|(see [namespace filtering](#namespace-filtering))|
|
||||||
|`labelSelector`|(see [label filtering](#label-filtering))|
|
|
||||||
|`nodeFit`|bool (see [node fit filtering](#node-fit-filtering))|
|
|`nodeFit`|bool (see [node fit filtering](#node-fit-filtering))|
|
||||||
|
|
||||||
**Example:**
|
**Example:**
|
||||||
@@ -524,24 +489,19 @@ strategies:
|
|||||||
|
|
||||||
This strategy evicts pods that are older than `maxPodLifeTimeSeconds`.
|
This strategy evicts pods that are older than `maxPodLifeTimeSeconds`.
|
||||||
|
|
||||||
You can also specify `states` parameter to **only** evict pods matching the following conditions:
|
You can also specify `podStatusPhases` to `only` evict pods with specific `StatusPhases`, currently this parameter is limited
|
||||||
- [Pod Phase](https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#pod-phase) status of: `Running`, `Pending`
|
to `Running` and `Pending`.
|
||||||
- [Container State Waiting](https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#container-state-waiting) condition of: `PodInitializing`, `ContainerCreating`
|
|
||||||
|
|
||||||
If a value for `states` or `podStatusPhases` is not specified,
|
|
||||||
Pods in any state (even `Running`) are considered for eviction.
|
|
||||||
|
|
||||||
**Parameters:**
|
**Parameters:**
|
||||||
|
|
||||||
|Name|Type|Notes|
|
|Name|Type|
|
||||||
|---|---|---|
|
|---|---|
|
||||||
|`maxPodLifeTimeSeconds`|int||
|
|`maxPodLifeTimeSeconds`|int|
|
||||||
|`podStatusPhases`|list(string)|Deprecated in v0.25+ Use `states` instead|
|
|`podStatusPhases`|list(string)|
|
||||||
|`states`|list(string)|Only supported in v0.25+|
|
|`thresholdPriority`|int (see [priority filtering](#priority-filtering))|
|
||||||
|`thresholdPriority`|int (see [priority filtering](#priority-filtering))||
|
|`thresholdPriorityClassName`|string (see [priority filtering](#priority-filtering))|
|
||||||
|`thresholdPriorityClassName`|string (see [priority filtering](#priority-filtering))||
|
|`namespaces`|(see [namespace filtering](#namespace-filtering))|
|
||||||
|`namespaces`|(see [namespace filtering](#namespace-filtering))||
|
|`labelSelector`|(see [label filtering](#label-filtering))|
|
||||||
|`labelSelector`|(see [label filtering](#label-filtering))||
|
|
||||||
|
|
||||||
**Example:**
|
**Example:**
|
||||||
|
|
||||||
@@ -554,9 +514,8 @@ strategies:
|
|||||||
params:
|
params:
|
||||||
podLifeTime:
|
podLifeTime:
|
||||||
maxPodLifeTimeSeconds: 86400
|
maxPodLifeTimeSeconds: 86400
|
||||||
states:
|
podStatusPhases:
|
||||||
- "Pending"
|
- "Pending"
|
||||||
- "PodInitializing"
|
|
||||||
```
|
```
|
||||||
|
|
||||||
### RemoveFailedPods
|
### RemoveFailedPods
|
||||||
@@ -564,7 +523,7 @@ strategies:
|
|||||||
This strategy evicts pods that are in failed status phase.
|
This strategy evicts pods that are in failed status phase.
|
||||||
You can provide an optional parameter to filter by failed `reasons`.
|
You can provide an optional parameter to filter by failed `reasons`.
|
||||||
`reasons` can be expanded to include reasons of InitContainers as well by setting the optional parameter `includingInitContainers` to `true`.
|
`reasons` can be expanded to include reasons of InitContainers as well by setting the optional parameter `includingInitContainers` to `true`.
|
||||||
You can specify an optional parameter `minPodLifetimeSeconds` to evict pods that are older than specified seconds.
|
You can specify an optional parameter `minPodLifeTimeSeconds` to evict pods that are older than specified seconds.
|
||||||
Lastly, you can specify the optional parameter `excludeOwnerKinds` and if a pod
|
Lastly, you can specify the optional parameter `excludeOwnerKinds` and if a pod
|
||||||
has any of these `Kind`s listed as an `OwnerRef`, that pod will not be considered for eviction.
|
has any of these `Kind`s listed as an `OwnerRef`, that pod will not be considered for eviction.
|
||||||
|
|
||||||
@@ -572,7 +531,7 @@ has any of these `Kind`s listed as an `OwnerRef`, that pod will not be considere
|
|||||||
|
|
||||||
|Name|Type|
|
|Name|Type|
|
||||||
|---|---|
|
|---|---|
|
||||||
|`minPodLifetimeSeconds`|uint|
|
|`minPodLifeTimeSeconds`|uint|
|
||||||
|`excludeOwnerKinds`|list(string)|
|
|`excludeOwnerKinds`|list(string)|
|
||||||
|`reasons`|list(string)|
|
|`reasons`|list(string)|
|
||||||
|`includingInitContainers`|bool|
|
|`includingInitContainers`|bool|
|
||||||
@@ -597,7 +556,7 @@ strategies:
|
|||||||
includingInitContainers: true
|
includingInitContainers: true
|
||||||
excludeOwnerKinds:
|
excludeOwnerKinds:
|
||||||
- "Job"
|
- "Job"
|
||||||
minPodLifetimeSeconds: 3600
|
minPodLifeTimeSeconds: 3600
|
||||||
```
|
```
|
||||||
|
|
||||||
## Filter Pods
|
## Filter Pods
|
||||||
@@ -695,7 +654,7 @@ does not exist, descheduler won't create it and will throw an error.
|
|||||||
|
|
||||||
### Label filtering
|
### Label filtering
|
||||||
|
|
||||||
The following strategies can configure a [standard kubernetes labelSelector](https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.25/#labelselector-v1-meta)
|
The following strategies can configure a [standard kubernetes labelSelector](https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.22/#labelselector-v1-meta)
|
||||||
to filter pods by their labels:
|
to filter pods by their labels:
|
||||||
|
|
||||||
* `PodLifeTime`
|
* `PodLifeTime`
|
||||||
@@ -743,9 +702,8 @@ The following strategies accept a `nodeFit` boolean parameter which can optimize
|
|||||||
|
|
||||||
If set to `true` the descheduler will consider whether or not the pods that meet eviction criteria will fit on other nodes before evicting them. If a pod cannot be rescheduled to another node, it will not be evicted. Currently the following criteria are considered when setting `nodeFit` to `true`:
|
If set to `true` the descheduler will consider whether or not the pods that meet eviction criteria will fit on other nodes before evicting them. If a pod cannot be rescheduled to another node, it will not be evicted. Currently the following criteria are considered when setting `nodeFit` to `true`:
|
||||||
- A `nodeSelector` on the pod
|
- A `nodeSelector` on the pod
|
||||||
- Any `tolerations` on the pod and any `taints` on the other nodes
|
- Any `Tolerations` on the pod and any `Taints` on the other nodes
|
||||||
- `nodeAffinity` on the pod
|
- `nodeAffinity` on the pod
|
||||||
- Resource `requests` made by the pod and the resources available on other nodes
|
|
||||||
- Whether any of the other nodes are marked as `unschedulable`
|
- Whether any of the other nodes are marked as `unschedulable`
|
||||||
|
|
||||||
E.g.
|
E.g.
|
||||||
@@ -755,18 +713,18 @@ apiVersion: "descheduler/v1alpha1"
|
|||||||
kind: "DeschedulerPolicy"
|
kind: "DeschedulerPolicy"
|
||||||
strategies:
|
strategies:
|
||||||
"LowNodeUtilization":
|
"LowNodeUtilization":
|
||||||
enabled: true
|
enabled: true
|
||||||
params:
|
params:
|
||||||
nodeFit: true
|
nodeResourceUtilizationThresholds:
|
||||||
nodeResourceUtilizationThresholds:
|
thresholds:
|
||||||
thresholds:
|
"cpu" : 20
|
||||||
"cpu": 20
|
"memory": 20
|
||||||
"memory": 20
|
"pods": 20
|
||||||
"pods": 20
|
targetThresholds:
|
||||||
targetThresholds:
|
"cpu" : 50
|
||||||
"cpu": 50
|
"memory": 50
|
||||||
"memory": 50
|
"pods": 50
|
||||||
"pods": 50
|
nodeFit: true
|
||||||
```
|
```
|
||||||
|
|
||||||
Note that node fit filtering references the current pod spec, and not that of it's owner.
|
Note that node fit filtering references the current pod spec, and not that of it's owner.
|
||||||
@@ -781,8 +739,8 @@ Using Deployments instead of ReplicationControllers provides an automated rollou
|
|||||||
When the descheduler decides to evict pods from a node, it employs the following general mechanism:
|
When the descheduler decides to evict pods from a node, it employs the following general mechanism:
|
||||||
|
|
||||||
* [Critical pods](https://kubernetes.io/docs/tasks/administer-cluster/guaranteed-scheduling-critical-addon-pods/) (with priorityClassName set to system-cluster-critical or system-node-critical) are never evicted (unless `evictSystemCriticalPods: true` is set).
|
* [Critical pods](https://kubernetes.io/docs/tasks/administer-cluster/guaranteed-scheduling-critical-addon-pods/) (with priorityClassName set to system-cluster-critical or system-node-critical) are never evicted (unless `evictSystemCriticalPods: true` is set).
|
||||||
* Pods (static or mirrored pods or standalone pods) not part of an ReplicationController, ReplicaSet(Deployment), StatefulSet, or Job are
|
* Pods (static or mirrored pods or stand alone pods) not part of an ReplicationController, ReplicaSet(Deployment), StatefulSet, or Job are
|
||||||
never evicted because these pods won't be recreated. (Standalone pods in failed status phase can be evicted by setting `evictFailedBarePods: true`)
|
never evicted because these pods won't be recreated.
|
||||||
* Pods associated with DaemonSets are never evicted.
|
* Pods associated with DaemonSets are never evicted.
|
||||||
* Pods with local storage are never evicted (unless `evictLocalStoragePods: true` is set).
|
* Pods with local storage are never evicted (unless `evictLocalStoragePods: true` is set).
|
||||||
* Pods with PVCs are evicted (unless `ignorePvcPods: true` is set).
|
* Pods with PVCs are evicted (unless `ignorePvcPods: true` is set).
|
||||||
@@ -791,7 +749,6 @@ best effort pods are evicted before burstable and guaranteed pods.
|
|||||||
* All types of pods with the annotation `descheduler.alpha.kubernetes.io/evict` are eligible for eviction. This
|
* All types of pods with the annotation `descheduler.alpha.kubernetes.io/evict` are eligible for eviction. This
|
||||||
annotation is used to override checks which prevent eviction and users can select which pod is evicted.
|
annotation is used to override checks which prevent eviction and users can select which pod is evicted.
|
||||||
Users should know how and if the pod will be recreated.
|
Users should know how and if the pod will be recreated.
|
||||||
* Pods with a non-nil DeletionTimestamp are not evicted by default.
|
|
||||||
|
|
||||||
Setting `--v=4` or greater on the Descheduler will log all reasons why any pod is not evictable.
|
Setting `--v=4` or greater on the Descheduler will log all reasons why any pod is not evictable.
|
||||||
|
|
||||||
@@ -800,23 +757,6 @@ Setting `--v=4` or greater on the Descheduler will log all reasons why any pod i
|
|||||||
Pods subject to a Pod Disruption Budget(PDB) are not evicted if descheduling violates its PDB. The pods
|
Pods subject to a Pod Disruption Budget(PDB) are not evicted if descheduling violates its PDB. The pods
|
||||||
are evicted by using the eviction subresource to handle PDB.
|
are evicted by using the eviction subresource to handle PDB.
|
||||||
|
|
||||||
## High Availability
|
|
||||||
|
|
||||||
In High Availability mode, Descheduler starts [leader election](https://github.com/kubernetes/client-go/tree/master/tools/leaderelection) process in Kubernetes. You can activate HA mode
|
|
||||||
if you choose to deploy your application as Deployment.
|
|
||||||
|
|
||||||
Deployment starts with 1 replica by default. If you want to use more than 1 replica, you must consider
|
|
||||||
enable High Availability mode since we don't want to run descheduler pods simultaneously.
|
|
||||||
|
|
||||||
### Configure HA Mode
|
|
||||||
|
|
||||||
The leader election process can be enabled by setting `--leader-elect` in the CLI. You can also set
|
|
||||||
`--set=leaderElection.enabled=true` flag if you are using Helm.
|
|
||||||
|
|
||||||
To get best results from HA mode some additional configurations might require:
|
|
||||||
* Configure a [podAntiAffinity](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node) rule if you want to schedule onto a node only if that node is in the same zone as at least one already-running descheduler
|
|
||||||
* Set the replica count greater than 1
|
|
||||||
|
|
||||||
## Metrics
|
## Metrics
|
||||||
|
|
||||||
| name | type | description |
|
| name | type | description |
|
||||||
@@ -836,19 +776,17 @@ v0.18 should work with k8s v1.18, v1.17, and v1.16.
|
|||||||
Starting with descheduler release v0.18 the minor version of descheduler matches the minor version of the k8s client
|
Starting with descheduler release v0.18 the minor version of descheduler matches the minor version of the k8s client
|
||||||
packages that it is compiled with.
|
packages that it is compiled with.
|
||||||
|
|
||||||
| Descheduler | Supported Kubernetes Version |
|
Descheduler | Supported Kubernetes Version
|
||||||
|-------------|------------------------------|
|
-------------|-----------------------------
|
||||||
| v0.25 | v1.25 |
|
v0.22 | v1.22
|
||||||
| v0.24 | v1.24 |
|
v0.21 | v1.21
|
||||||
| v0.23 | v1.23 |
|
v0.20 | v1.20
|
||||||
| v0.22 | v1.22 |
|
v0.19 | v1.19
|
||||||
| v0.21 | v1.21 |
|
v0.18 | v1.18
|
||||||
| v0.20 | v1.20 |
|
v0.10 | v1.17
|
||||||
| v0.19 | v1.19 |
|
v0.4-v0.9 | v1.9+
|
||||||
| v0.18 | v1.18 |
|
v0.1-v0.3 | v1.7-v1.8
|
||||||
| v0.10 | v1.17 |
|
|
||||||
| v0.4-v0.9 | v1.9+ |
|
|
||||||
| v0.1-v0.3 | v1.7-v1.8 |
|
|
||||||
|
|
||||||
## Getting Involved and Contributing
|
## Getting Involved and Contributing
|
||||||
|
|
||||||
|
|||||||
@@ -1,7 +1,7 @@
|
|||||||
apiVersion: v1
|
apiVersion: v1
|
||||||
name: descheduler
|
name: descheduler
|
||||||
version: 0.25.2
|
version: 0.22.0
|
||||||
appVersion: 0.25.1
|
appVersion: 0.22.0
|
||||||
description: Descheduler for Kubernetes is used to rebalance clusters by evicting pods that can potentially be scheduled on better nodes. In the current implementation, descheduler does not schedule replacement of evicted pods but relies on the default scheduler for that.
|
description: Descheduler for Kubernetes is used to rebalance clusters by evicting pods that can potentially be scheduled on better nodes. In the current implementation, descheduler does not schedule replacement of evicted pods but relies on the default scheduler for that.
|
||||||
keywords:
|
keywords:
|
||||||
- kubernetes
|
- kubernetes
|
||||||
|
|||||||
@@ -43,45 +43,28 @@ The command removes all the Kubernetes components associated with the chart and
|
|||||||
|
|
||||||
The following table lists the configurable parameters of the _descheduler_ chart and their default values.
|
The following table lists the configurable parameters of the _descheduler_ chart and their default values.
|
||||||
|
|
||||||
| Parameter | Description | Default |
|
| Parameter | Description | Default |
|
||||||
|-------------------------------------|-----------------------------------------------------------------------------------------------------------------------|--------------------------------------|
|
| ------------------------------ | --------------------------------------------------------------------------------------------------------------------- | ------------------------------------ |
|
||||||
| `kind` | Use as CronJob or Deployment | `CronJob` |
|
| `kind` | Use as CronJob or Deployment | `CronJob` |
|
||||||
| `image.repository` | Docker repository to use | `k8s.gcr.io/descheduler/descheduler` |
|
| `image.repository` | Docker repository to use | `k8s.gcr.io/descheduler/descheduler` |
|
||||||
| `image.tag` | Docker tag to use | `v[chart appVersion]` |
|
| `image.tag` | Docker tag to use | `v[chart appVersion]` |
|
||||||
| `image.pullPolicy` | Docker image pull policy | `IfNotPresent` |
|
| `image.pullPolicy` | Docker image pull policy | `IfNotPresent` |
|
||||||
| `imagePullSecrets` | Docker repository secrets | `[]` |
|
| `imagePullSecrets` | Docker repository secrets | `[]` |
|
||||||
| `nameOverride` | String to partially override `descheduler.fullname` template (will prepend the release name) | `""` |
|
| `nameOverride` | String to partially override `descheduler.fullname` template (will prepend the release name) | `""` |
|
||||||
| `fullnameOverride` | String to fully override `descheduler.fullname` template | `""` |
|
| `fullnameOverride` | String to fully override `descheduler.fullname` template | `""` |
|
||||||
| `cronJobApiVersion` | CronJob API Group Version | `"batch/v1"` |
|
| `cronJobApiVersion` | CronJob API Group Version | `"batch/v1"` |
|
||||||
| `schedule` | The cron schedule to run the _descheduler_ job on | `"*/2 * * * *"` |
|
| `schedule` | The cron schedule to run the _descheduler_ job on | `"*/2 * * * *"` |
|
||||||
| `startingDeadlineSeconds` | If set, configure `startingDeadlineSeconds` for the _descheduler_ job | `nil` |
|
| `startingDeadlineSeconds` | If set, configure `startingDeadlineSeconds` for the _descheduler_ job | `nil` |
|
||||||
| `successfulJobsHistoryLimit` | If set, configure `successfulJobsHistoryLimit` for the _descheduler_ job | `nil` |
|
| `successfulJobsHistoryLimit` | If set, configure `successfulJobsHistoryLimit` for the _descheduler_ job | `nil` |
|
||||||
| `failedJobsHistoryLimit` | If set, configure `failedJobsHistoryLimit` for the _descheduler_ job | `nil` |
|
| `failedJobsHistoryLimit` | If set, configure `failedJobsHistoryLimit` for the _descheduler_ job | `nil` |
|
||||||
| `deschedulingInterval` | If using kind:Deployment, sets time between consecutive descheduler executions. | `5m` |
|
| `deschedulingInterval` | If using kind:Deployment, sets time between consecutive descheduler executions. | `5m` |
|
||||||
| `replicas` | The replica count for Deployment | `1` |
|
| `cmdOptions` | The options to pass to the _descheduler_ command | _see values.yaml_ |
|
||||||
| `leaderElection` | The options for high availability when running replicated components | _see values.yaml_ |
|
| `deschedulerPolicy.strategies` | The _descheduler_ strategies to apply | _see values.yaml_ |
|
||||||
| `cmdOptions` | The options to pass to the _descheduler_ command | _see values.yaml_ |
|
| `priorityClassName` | The name of the priority class to add to pods | `system-cluster-critical` |
|
||||||
| `deschedulerPolicy.strategies` | The _descheduler_ strategies to apply | _see values.yaml_ |
|
| `rbac.create` | If `true`, create & use RBAC resources | `true` |
|
||||||
| `priorityClassName` | The name of the priority class to add to pods | `system-cluster-critical` |
|
| `podSecurityPolicy.create` | If `true`, create PodSecurityPolicy | `true` |
|
||||||
| `rbac.create` | If `true`, create & use RBAC resources | `true` |
|
| `resources` | Descheduler container CPU and memory requests/limits | _see values.yaml_ |
|
||||||
| `resources` | Descheduler container CPU and memory requests/limits | _see values.yaml_ |
|
| `serviceAccount.create` | If `true`, create a service account for the cron job | `true` |
|
||||||
| `serviceAccount.create` | If `true`, create a service account for the cron job | `true` |
|
| `serviceAccount.name` | The name of the service account to use, if not set and create is true a name is generated using the fullname template | `nil` |
|
||||||
| `serviceAccount.name` | The name of the service account to use, if not set and create is true a name is generated using the fullname template | `nil` |
|
| `nodeSelector` | Node selectors to run the descheduler cronjob on specific nodes | `nil` |
|
||||||
| `serviceAccount.annotations` | Specifies custom annotations for the serviceAccount | `{}` |
|
| `tolerations` | tolerations to run the descheduler cronjob on specific nodes | `nil` |
|
||||||
| `podAnnotations` | Annotations to add to the descheduler Pods | `{}` |
|
|
||||||
| `podLabels` | Labels to add to the descheduler Pods | `{}` |
|
|
||||||
| `nodeSelector` | Node selectors to run the descheduler cronjob/deployment on specific nodes | `nil` |
|
|
||||||
| `service.enabled` | If `true`, create a service for deployment | `false` |
|
|
||||||
| `serviceMonitor.enabled` | If `true`, create a ServiceMonitor for deployment | `false` |
|
|
||||||
| `serviceMonitor.namespace` | The namespace where Prometheus expects to find service monitors | `nil` |
|
|
||||||
| `serviceMonitor.interval` | The scrape interval. If not set, the Prometheus default scrape interval is used | `nil` |
|
|
||||||
| `serviceMonitor.honorLabels` | Keeps the scraped data's labels when labels are on collisions with target labels. | `true` |
|
|
||||||
| `serviceMonitor.insecureSkipVerify` | Skip TLS certificate validation when scraping | `true` |
|
|
||||||
| `serviceMonitor.serverName` | Name of the server to use when validating TLS certificate | `nil` |
|
|
||||||
| `serviceMonitor.metricRelabelings` | MetricRelabelConfigs to apply to samples after scraping, but before ingestion | `[]` |
|
|
||||||
| `serviceMonitor.relabelings` | RelabelConfigs to apply to samples before scraping | `[]` |
|
|
||||||
| `affinity` | Node affinity to run the descheduler cronjob/deployment on specific nodes | `nil` |
|
|
||||||
| `tolerations` | tolerations to run the descheduler cronjob/deployment on specific nodes | `nil` |
|
|
||||||
| `suspend` | Set spec.suspend in descheduler cronjob | `false` |
|
|
||||||
| `commonLabels` | Labels to apply to all resources | `{}` |
|
|
||||||
| `livenessProbe` | Liveness probe configuration for the descheduler container | _see values.yaml_ |
|
|
||||||
@@ -1,7 +1 @@
|
|||||||
Descheduler installed as a {{ .Values.kind }}.
|
Descheduler installed as a cron job.
|
||||||
|
|
||||||
{{- if eq .Values.kind "Deployment" }}
|
|
||||||
{{- if eq .Values.replicas 1.0}}
|
|
||||||
WARNING: You set replica count as 1 and workload kind as Deployment however leaderElection is not enabled. Consider enabling Leader Election for HA mode.
|
|
||||||
{{- end}}
|
|
||||||
{{- end}}
|
|
||||||
|
|||||||
@@ -42,9 +42,6 @@ app.kubernetes.io/instance: {{ .Release.Name }}
|
|||||||
app.kubernetes.io/version: {{ .Chart.AppVersion | quote }}
|
app.kubernetes.io/version: {{ .Chart.AppVersion | quote }}
|
||||||
{{- end }}
|
{{- end }}
|
||||||
app.kubernetes.io/managed-by: {{ .Release.Service }}
|
app.kubernetes.io/managed-by: {{ .Release.Service }}
|
||||||
{{- if .Values.commonLabels}}
|
|
||||||
{{ toYaml .Values.commonLabels }}
|
|
||||||
{{- end }}
|
|
||||||
{{- end -}}
|
{{- end -}}
|
||||||
|
|
||||||
{{/*
|
{{/*
|
||||||
@@ -65,30 +62,3 @@ Create the name of the service account to use
|
|||||||
{{ default "default" .Values.serviceAccount.name }}
|
{{ default "default" .Values.serviceAccount.name }}
|
||||||
{{- end -}}
|
{{- end -}}
|
||||||
{{- end -}}
|
{{- end -}}
|
||||||
|
|
||||||
{{/*
|
|
||||||
Leader Election
|
|
||||||
*/}}
|
|
||||||
{{- define "descheduler.leaderElection"}}
|
|
||||||
{{- if .Values.leaderElection -}}
|
|
||||||
- --leader-elect={{ .Values.leaderElection.enabled }}
|
|
||||||
{{- if .Values.leaderElection.leaseDuration }}
|
|
||||||
- --leader-elect-lease-duration={{ .Values.leaderElection.leaseDuration }}
|
|
||||||
{{- end }}
|
|
||||||
{{- if .Values.leaderElection.renewDeadline }}
|
|
||||||
- --leader-elect-renew-deadline={{ .Values.leaderElection.renewDeadline }}
|
|
||||||
{{- end }}
|
|
||||||
{{- if .Values.leaderElection.retryPeriod }}
|
|
||||||
- --leader-elect-retry-period={{ .Values.leaderElection.retryPeriod }}
|
|
||||||
{{- end }}
|
|
||||||
{{- if .Values.leaderElection.resourceLock }}
|
|
||||||
- --leader-elect-resource-lock={{ .Values.leaderElection.resourceLock }}
|
|
||||||
{{- end }}
|
|
||||||
{{- if .Values.leaderElection.resourceName }}
|
|
||||||
- --leader-elect-resource-name={{ .Values.leaderElection.resourceName }}
|
|
||||||
{{- end }}
|
|
||||||
{{- if .Values.leaderElection.resourceNamescape }}
|
|
||||||
- --leader-elect-resource-namespace={{ .Values.leaderElection.resourceNamescape }}
|
|
||||||
{{- end -}}
|
|
||||||
{{- end }}
|
|
||||||
{{- end }}
|
|
||||||
|
|||||||
@@ -6,7 +6,7 @@ metadata:
|
|||||||
labels:
|
labels:
|
||||||
{{- include "descheduler.labels" . | nindent 4 }}
|
{{- include "descheduler.labels" . | nindent 4 }}
|
||||||
rules:
|
rules:
|
||||||
- apiGroups: ["events.k8s.io"]
|
- apiGroups: [""]
|
||||||
resources: ["events"]
|
resources: ["events"]
|
||||||
verbs: ["create", "update"]
|
verbs: ["create", "update"]
|
||||||
- apiGroups: [""]
|
- apiGroups: [""]
|
||||||
@@ -14,7 +14,7 @@ rules:
|
|||||||
verbs: ["get", "watch", "list"]
|
verbs: ["get", "watch", "list"]
|
||||||
- apiGroups: [""]
|
- apiGroups: [""]
|
||||||
resources: ["namespaces"]
|
resources: ["namespaces"]
|
||||||
verbs: ["get", "watch", "list"]
|
verbs: ["get", "list"]
|
||||||
- apiGroups: [""]
|
- apiGroups: [""]
|
||||||
resources: ["pods"]
|
resources: ["pods"]
|
||||||
verbs: ["get", "watch", "list", "delete"]
|
verbs: ["get", "watch", "list", "delete"]
|
||||||
@@ -24,13 +24,11 @@ rules:
|
|||||||
- apiGroups: ["scheduling.k8s.io"]
|
- apiGroups: ["scheduling.k8s.io"]
|
||||||
resources: ["priorityclasses"]
|
resources: ["priorityclasses"]
|
||||||
verbs: ["get", "watch", "list"]
|
verbs: ["get", "watch", "list"]
|
||||||
{{- if .Values.leaderElection.enabled }}
|
{{- if .Values.podSecurityPolicy.create }}
|
||||||
- apiGroups: ["coordination.k8s.io"]
|
- apiGroups: ['policy']
|
||||||
resources: ["leases"]
|
resources: ['podsecuritypolicies']
|
||||||
verbs: ["create", "update"]
|
verbs: ['use']
|
||||||
- apiGroups: ["coordination.k8s.io"]
|
resourceNames:
|
||||||
resources: ["leases"]
|
- {{ template "descheduler.fullname" . }}
|
||||||
resourceNames: ["{{ .Values.leaderElection.resourceName | default "descheduler" }}"]
|
|
||||||
verbs: ["get", "patch", "delete"]
|
|
||||||
{{- end }}
|
{{- end }}
|
||||||
{{- end -}}
|
{{- end -}}
|
||||||
|
|||||||
@@ -2,7 +2,6 @@ apiVersion: v1
|
|||||||
kind: ConfigMap
|
kind: ConfigMap
|
||||||
metadata:
|
metadata:
|
||||||
name: {{ template "descheduler.fullname" . }}
|
name: {{ template "descheduler.fullname" . }}
|
||||||
namespace: {{ .Release.Namespace }}
|
|
||||||
labels:
|
labels:
|
||||||
{{- include "descheduler.labels" . | nindent 4 }}
|
{{- include "descheduler.labels" . | nindent 4 }}
|
||||||
data:
|
data:
|
||||||
|
|||||||
@@ -3,14 +3,10 @@ apiVersion: {{ .Values.cronJobApiVersion | default "batch/v1" }}
|
|||||||
kind: CronJob
|
kind: CronJob
|
||||||
metadata:
|
metadata:
|
||||||
name: {{ template "descheduler.fullname" . }}
|
name: {{ template "descheduler.fullname" . }}
|
||||||
namespace: {{ .Release.Namespace }}
|
|
||||||
labels:
|
labels:
|
||||||
{{- include "descheduler.labels" . | nindent 4 }}
|
{{- include "descheduler.labels" . | nindent 4 }}
|
||||||
spec:
|
spec:
|
||||||
schedule: {{ .Values.schedule | quote }}
|
schedule: {{ .Values.schedule | quote }}
|
||||||
{{- if .Values.suspend }}
|
|
||||||
suspend: {{ .Values.suspend }}
|
|
||||||
{{- end }}
|
|
||||||
concurrencyPolicy: "Forbid"
|
concurrencyPolicy: "Forbid"
|
||||||
{{- if .Values.startingDeadlineSeconds }}
|
{{- if .Values.startingDeadlineSeconds }}
|
||||||
startingDeadlineSeconds: {{ .Values.startingDeadlineSeconds }}
|
startingDeadlineSeconds: {{ .Values.startingDeadlineSeconds }}
|
||||||
@@ -41,10 +37,6 @@ spec:
|
|||||||
nodeSelector:
|
nodeSelector:
|
||||||
{{- toYaml . | nindent 12 }}
|
{{- toYaml . | nindent 12 }}
|
||||||
{{- end }}
|
{{- end }}
|
||||||
{{- with .Values.affinity }}
|
|
||||||
affinity:
|
|
||||||
{{- toYaml . | nindent 12 }}
|
|
||||||
{{- end }}
|
|
||||||
{{- with .Values.tolerations }}
|
{{- with .Values.tolerations }}
|
||||||
tolerations:
|
tolerations:
|
||||||
{{- toYaml . | nindent 12 }}
|
{{- toYaml . | nindent 12 }}
|
||||||
@@ -73,8 +65,6 @@ spec:
|
|||||||
- {{ $value | quote }}
|
- {{ $value | quote }}
|
||||||
{{- end }}
|
{{- end }}
|
||||||
{{- end }}
|
{{- end }}
|
||||||
livenessProbe:
|
|
||||||
{{- toYaml .Values.livenessProbe | nindent 16 }}
|
|
||||||
resources:
|
resources:
|
||||||
{{- toYaml .Values.resources | nindent 16 }}
|
{{- toYaml .Values.resources | nindent 16 }}
|
||||||
securityContext:
|
securityContext:
|
||||||
|
|||||||
@@ -3,18 +3,10 @@ apiVersion: apps/v1
|
|||||||
kind: Deployment
|
kind: Deployment
|
||||||
metadata:
|
metadata:
|
||||||
name: {{ template "descheduler.fullname" . }}
|
name: {{ template "descheduler.fullname" . }}
|
||||||
namespace: {{ .Release.Namespace }}
|
|
||||||
labels:
|
labels:
|
||||||
{{- include "descheduler.labels" . | nindent 4 }}
|
{{- include "descheduler.labels" . | nindent 4 }}
|
||||||
spec:
|
spec:
|
||||||
{{- if gt .Values.replicas 1.0}}
|
|
||||||
{{- if not .Values.leaderElection.enabled }}
|
|
||||||
{{- fail "You must set leaderElection to use more than 1 replica"}}
|
|
||||||
{{- end}}
|
|
||||||
replicas: {{ required "leaderElection required for running more than one replica" .Values.replicas }}
|
|
||||||
{{- else }}
|
|
||||||
replicas: 1
|
replicas: 1
|
||||||
{{- end }}
|
|
||||||
selector:
|
selector:
|
||||||
matchLabels:
|
matchLabels:
|
||||||
{{- include "descheduler.selectorLabels" . | nindent 6 }}
|
{{- include "descheduler.selectorLabels" . | nindent 6 }}
|
||||||
@@ -35,10 +27,6 @@ spec:
|
|||||||
priorityClassName: {{ .Values.priorityClassName }}
|
priorityClassName: {{ .Values.priorityClassName }}
|
||||||
{{- end }}
|
{{- end }}
|
||||||
serviceAccountName: {{ template "descheduler.serviceAccountName" . }}
|
serviceAccountName: {{ template "descheduler.serviceAccountName" . }}
|
||||||
{{- with .Values.imagePullSecrets }}
|
|
||||||
imagePullSecrets:
|
|
||||||
{{- toYaml . | nindent 10 }}
|
|
||||||
{{- end }}
|
|
||||||
containers:
|
containers:
|
||||||
- name: {{ .Chart.Name }}
|
- name: {{ .Chart.Name }}
|
||||||
image: "{{ .Values.image.repository }}:{{ .Values.image.tag | default (printf "v%s" .Chart.AppVersion) }}"
|
image: "{{ .Values.image.repository }}:{{ .Values.image.tag | default (printf "v%s" .Chart.AppVersion) }}"
|
||||||
@@ -56,12 +44,9 @@ spec:
|
|||||||
- {{ $value | quote }}
|
- {{ $value | quote }}
|
||||||
{{- end }}
|
{{- end }}
|
||||||
{{- end }}
|
{{- end }}
|
||||||
{{- include "descheduler.leaderElection" . | nindent 12 }}
|
|
||||||
ports:
|
ports:
|
||||||
- containerPort: 10258
|
- containerPort: 10258
|
||||||
protocol: TCP
|
protocol: TCP
|
||||||
livenessProbe:
|
|
||||||
{{- toYaml .Values.livenessProbe | nindent 12 }}
|
|
||||||
resources:
|
resources:
|
||||||
{{- toYaml .Values.resources | nindent 12 }}
|
{{- toYaml .Values.resources | nindent 12 }}
|
||||||
securityContext:
|
securityContext:
|
||||||
@@ -83,10 +68,6 @@ spec:
|
|||||||
nodeSelector:
|
nodeSelector:
|
||||||
{{- toYaml . | nindent 8 }}
|
{{- toYaml . | nindent 8 }}
|
||||||
{{- end }}
|
{{- end }}
|
||||||
{{- with .Values.affinity }}
|
|
||||||
affinity:
|
|
||||||
{{- toYaml . | nindent 8 }}
|
|
||||||
{{- end }}
|
|
||||||
{{- with .Values.tolerations }}
|
{{- with .Values.tolerations }}
|
||||||
tolerations:
|
tolerations:
|
||||||
{{- toYaml . | nindent 8 }}
|
{{- toYaml . | nindent 8 }}
|
||||||
|
|||||||
38
charts/descheduler/templates/podsecuritypolicy.yaml
Normal file
38
charts/descheduler/templates/podsecuritypolicy.yaml
Normal file
@@ -0,0 +1,38 @@
|
|||||||
|
{{- if .Values.podSecurityPolicy.create -}}
|
||||||
|
apiVersion: policy/v1beta1
|
||||||
|
kind: PodSecurityPolicy
|
||||||
|
metadata:
|
||||||
|
name: {{ template "descheduler.fullname" . }}
|
||||||
|
annotations:
|
||||||
|
seccomp.security.alpha.kubernetes.io/allowedProfileNames: 'docker/default,runtime/default'
|
||||||
|
seccomp.security.alpha.kubernetes.io/defaultProfileName: 'runtime/default'
|
||||||
|
spec:
|
||||||
|
privileged: false
|
||||||
|
allowPrivilegeEscalation: false
|
||||||
|
requiredDropCapabilities:
|
||||||
|
- ALL
|
||||||
|
volumes:
|
||||||
|
- 'configMap'
|
||||||
|
- 'secret'
|
||||||
|
hostNetwork: false
|
||||||
|
hostIPC: false
|
||||||
|
hostPID: false
|
||||||
|
runAsUser:
|
||||||
|
rule: 'MustRunAs'
|
||||||
|
ranges:
|
||||||
|
- min: 1
|
||||||
|
max: 65535
|
||||||
|
seLinux:
|
||||||
|
rule: 'RunAsAny'
|
||||||
|
supplementalGroups:
|
||||||
|
rule: 'MustRunAs'
|
||||||
|
ranges:
|
||||||
|
- min: 1
|
||||||
|
max: 65535
|
||||||
|
fsGroup:
|
||||||
|
rule: 'MustRunAs'
|
||||||
|
ranges:
|
||||||
|
- min: 1
|
||||||
|
max: 65535
|
||||||
|
readOnlyRootFilesystem: true
|
||||||
|
{{- end -}}
|
||||||
@@ -1,21 +0,0 @@
|
|||||||
{{- if eq .Values.kind "Deployment" }}
|
|
||||||
{{- if eq .Values.service.enabled true }}
|
|
||||||
apiVersion: v1
|
|
||||||
kind: Service
|
|
||||||
metadata:
|
|
||||||
labels:
|
|
||||||
{{- include "descheduler.labels" . | nindent 4 }}
|
|
||||||
name: {{ template "descheduler.fullname" . }}
|
|
||||||
namespace: {{ .Release.Namespace }}
|
|
||||||
spec:
|
|
||||||
clusterIP: None
|
|
||||||
ports:
|
|
||||||
- name: http-metrics
|
|
||||||
port: 10258
|
|
||||||
protocol: TCP
|
|
||||||
targetPort: 10258
|
|
||||||
selector:
|
|
||||||
{{- include "descheduler.selectorLabels" . | nindent 4 }}
|
|
||||||
type: ClusterIP
|
|
||||||
{{- end }}
|
|
||||||
{{- end }}
|
|
||||||
@@ -3,10 +3,6 @@ apiVersion: v1
|
|||||||
kind: ServiceAccount
|
kind: ServiceAccount
|
||||||
metadata:
|
metadata:
|
||||||
name: {{ template "descheduler.serviceAccountName" . }}
|
name: {{ template "descheduler.serviceAccountName" . }}
|
||||||
namespace: {{ .Release.Namespace }}
|
|
||||||
labels:
|
labels:
|
||||||
{{- include "descheduler.labels" . | nindent 4 }}
|
{{- include "descheduler.labels" . | nindent 4 }}
|
||||||
{{- if .Values.serviceAccount.annotations }}
|
|
||||||
annotations: {{ toYaml .Values.serviceAccount.annotations | nindent 4 }}
|
|
||||||
{{- end }}
|
|
||||||
{{- end -}}
|
{{- end -}}
|
||||||
|
|||||||
@@ -1,41 +0,0 @@
|
|||||||
{{- if eq .Values.kind "Deployment" }}
|
|
||||||
{{- if eq .Values.serviceMonitor.enabled true }}
|
|
||||||
apiVersion: monitoring.coreos.com/v1
|
|
||||||
kind: ServiceMonitor
|
|
||||||
metadata:
|
|
||||||
name: {{ template "descheduler.fullname" . }}-servicemonitor
|
|
||||||
namespace: {{ .Values.serviceMonitor.namespace | default .Release.Namespace }}
|
|
||||||
labels:
|
|
||||||
{{- include "descheduler.labels" . | nindent 4 }}
|
|
||||||
spec:
|
|
||||||
jobLabel: jobLabel
|
|
||||||
namespaceSelector:
|
|
||||||
matchNames:
|
|
||||||
- {{ .Release.Namespace }}
|
|
||||||
selector:
|
|
||||||
matchLabels:
|
|
||||||
{{- include "descheduler.selectorLabels" . | nindent 6 }}
|
|
||||||
endpoints:
|
|
||||||
- honorLabels: {{ .Values.serviceMonitor.honorLabels | default true }}
|
|
||||||
port: http-metrics
|
|
||||||
{{- if .Values.serviceMonitor.interval }}
|
|
||||||
interval: {{ .Values.serviceMonitor.interval }}
|
|
||||||
{{- end }}
|
|
||||||
scheme: https
|
|
||||||
tlsConfig:
|
|
||||||
{{- if eq .Values.serviceMonitor.insecureSkipVerify true }}
|
|
||||||
insecureSkipVerify: true
|
|
||||||
{{- end }}
|
|
||||||
{{- if .Values.serviceMonitor.serverName }}
|
|
||||||
serverName: {{ .Values.serviceMonitor.serverName }}
|
|
||||||
{{- end}}
|
|
||||||
{{- if .Values.serviceMonitor.metricRelabelings }}
|
|
||||||
metricRelabelings:
|
|
||||||
{{ tpl (toYaml .Values.serviceMonitor.metricRelabelings | indent 4) . }}
|
|
||||||
{{- end }}
|
|
||||||
{{- if .Values.serviceMonitor.relabelings }}
|
|
||||||
relabelings:
|
|
||||||
{{ tpl (toYaml .Values.serviceMonitor.relabelings | indent 4) . }}
|
|
||||||
{{- end }}
|
|
||||||
{{- end }}
|
|
||||||
{{- end }}
|
|
||||||
29
charts/descheduler/templates/tests/test-descheduler-pod.yaml
Normal file
29
charts/descheduler/templates/tests/test-descheduler-pod.yaml
Normal file
@@ -0,0 +1,29 @@
|
|||||||
|
apiVersion: v1
|
||||||
|
kind: Pod
|
||||||
|
metadata:
|
||||||
|
name: descheduler-test-pod
|
||||||
|
annotations:
|
||||||
|
"helm.sh/hook": test
|
||||||
|
spec:
|
||||||
|
restartPolicy: Never
|
||||||
|
serviceAccountName: descheduler-ci
|
||||||
|
containers:
|
||||||
|
- name: descheduler-test-container
|
||||||
|
image: alpine:latest
|
||||||
|
imagePullPolicy: IfNotPresent
|
||||||
|
securityContext:
|
||||||
|
allowPrivilegeEscalation: false
|
||||||
|
capabilities:
|
||||||
|
drop:
|
||||||
|
- All
|
||||||
|
privileged: false
|
||||||
|
runAsNonRoot: false
|
||||||
|
command: ["/bin/ash"]
|
||||||
|
args:
|
||||||
|
- -c
|
||||||
|
- >-
|
||||||
|
apk --no-cache add curl &&
|
||||||
|
curl -LO https://storage.googleapis.com/kubernetes-release/release/$(curl -s https://storage.googleapis.com/kubernetes-release/release/stable.txt)/bin/linux/amd64/kubectl &&
|
||||||
|
chmod +x ./kubectl &&
|
||||||
|
mv ./kubectl /usr/local/bin/kubectl &&
|
||||||
|
/usr/local/bin/kubectl get pods --namespace kube-system --token "$(cat /var/run/secrets/kubernetes.io/serviceaccount/token)" | grep "descheduler" | grep "Completed"
|
||||||
@@ -11,8 +11,7 @@ image:
|
|||||||
tag: ""
|
tag: ""
|
||||||
pullPolicy: IfNotPresent
|
pullPolicy: IfNotPresent
|
||||||
|
|
||||||
imagePullSecrets:
|
imagePullSecrets: []
|
||||||
# - name: container-registry-secret
|
|
||||||
|
|
||||||
resources:
|
resources:
|
||||||
requests:
|
requests:
|
||||||
@@ -25,54 +24,25 @@ resources:
|
|||||||
nameOverride: ""
|
nameOverride: ""
|
||||||
fullnameOverride: ""
|
fullnameOverride: ""
|
||||||
|
|
||||||
# labels that'll be applied to all resources
|
cronJobApiVersion: "batch/v1" # Use "batch/v1beta1" for k8s version < 1.21.0. TODO(@7i) remove with 1.23 release
|
||||||
commonLabels: {}
|
|
||||||
|
|
||||||
cronJobApiVersion: "batch/v1"
|
|
||||||
schedule: "*/2 * * * *"
|
schedule: "*/2 * * * *"
|
||||||
suspend: false
|
#startingDeadlineSeconds: 200
|
||||||
# startingDeadlineSeconds: 200
|
#successfulJobsHistoryLimit: 1
|
||||||
# successfulJobsHistoryLimit: 1
|
#failedJobsHistoryLimit: 1
|
||||||
# failedJobsHistoryLimit: 1
|
|
||||||
|
|
||||||
# Required when running as a Deployment
|
# Required when running as a Deployment
|
||||||
deschedulingInterval: 5m
|
deschedulingInterval: 5m
|
||||||
|
|
||||||
# Specifies the replica count for Deployment
|
|
||||||
# Set leaderElection if you want to use more than 1 replica
|
|
||||||
# Set affinity.podAntiAffinity rule if you want to schedule onto a node
|
|
||||||
# only if that node is in the same zone as at least one already-running descheduler
|
|
||||||
replicas: 1
|
|
||||||
|
|
||||||
# Specifies whether Leader Election resources should be created
|
|
||||||
# Required when running as a Deployment
|
|
||||||
leaderElection: {}
|
|
||||||
# enabled: true
|
|
||||||
# leaseDuration: 15s
|
|
||||||
# renewDeadline: 10s
|
|
||||||
# retryPeriod: 2s
|
|
||||||
# resourceLock: "leases"
|
|
||||||
# resourceName: "descheduler"
|
|
||||||
# resourceNamescape: "kube-system"
|
|
||||||
|
|
||||||
cmdOptions:
|
cmdOptions:
|
||||||
v: 3
|
v: 3
|
||||||
|
# evict-local-storage-pods:
|
||||||
|
# max-pods-to-evict-per-node: 10
|
||||||
|
# node-selector: "key1=value1,key2=value2"
|
||||||
|
|
||||||
deschedulerPolicy:
|
deschedulerPolicy:
|
||||||
# nodeSelector: "key1=value1,key2=value2"
|
|
||||||
# maxNoOfPodsToEvictPerNode: 10
|
|
||||||
# maxNoOfPodsToEvictPerNamespace: 10
|
|
||||||
# ignorePvcPods: true
|
|
||||||
# evictLocalStoragePods: true
|
|
||||||
strategies:
|
strategies:
|
||||||
RemoveDuplicates:
|
RemoveDuplicates:
|
||||||
enabled: true
|
enabled: true
|
||||||
RemovePodsHavingTooManyRestarts:
|
|
||||||
enabled: true
|
|
||||||
params:
|
|
||||||
podsHavingTooManyRestarts:
|
|
||||||
podRestartThreshold: 100
|
|
||||||
includingInitContainers: true
|
|
||||||
RemovePodsViolatingNodeTaints:
|
RemovePodsViolatingNodeTaints:
|
||||||
enabled: true
|
enabled: true
|
||||||
RemovePodsViolatingNodeAffinity:
|
RemovePodsViolatingNodeAffinity:
|
||||||
@@ -82,10 +52,6 @@ deschedulerPolicy:
|
|||||||
- requiredDuringSchedulingIgnoredDuringExecution
|
- requiredDuringSchedulingIgnoredDuringExecution
|
||||||
RemovePodsViolatingInterPodAntiAffinity:
|
RemovePodsViolatingInterPodAntiAffinity:
|
||||||
enabled: true
|
enabled: true
|
||||||
RemovePodsViolatingTopologySpreadConstraint:
|
|
||||||
enabled: true
|
|
||||||
params:
|
|
||||||
includeSoftConstraints: false
|
|
||||||
LowNodeUtilization:
|
LowNodeUtilization:
|
||||||
enabled: true
|
enabled: true
|
||||||
params:
|
params:
|
||||||
@@ -104,25 +70,6 @@ priorityClassName: system-cluster-critical
|
|||||||
nodeSelector: {}
|
nodeSelector: {}
|
||||||
# foo: bar
|
# foo: bar
|
||||||
|
|
||||||
affinity: {}
|
|
||||||
# nodeAffinity:
|
|
||||||
# requiredDuringSchedulingIgnoredDuringExecution:
|
|
||||||
# nodeSelectorTerms:
|
|
||||||
# - matchExpressions:
|
|
||||||
# - key: kubernetes.io/e2e-az-name
|
|
||||||
# operator: In
|
|
||||||
# values:
|
|
||||||
# - e2e-az1
|
|
||||||
# - e2e-az2
|
|
||||||
# podAntiAffinity:
|
|
||||||
# requiredDuringSchedulingIgnoredDuringExecution:
|
|
||||||
# - labelSelector:
|
|
||||||
# matchExpressions:
|
|
||||||
# - key: app.kubernetes.io/name
|
|
||||||
# operator: In
|
|
||||||
# values:
|
|
||||||
# - descheduler
|
|
||||||
# topologyKey: "kubernetes.io/hostname"
|
|
||||||
tolerations: []
|
tolerations: []
|
||||||
# - key: 'management'
|
# - key: 'management'
|
||||||
# operator: 'Equal'
|
# operator: 'Equal'
|
||||||
@@ -133,47 +80,13 @@ rbac:
|
|||||||
# Specifies whether RBAC resources should be created
|
# Specifies whether RBAC resources should be created
|
||||||
create: true
|
create: true
|
||||||
|
|
||||||
|
podSecurityPolicy:
|
||||||
|
# Specifies whether PodSecurityPolicy should be created.
|
||||||
|
create: true
|
||||||
|
|
||||||
serviceAccount:
|
serviceAccount:
|
||||||
# Specifies whether a ServiceAccount should be created
|
# Specifies whether a ServiceAccount should be created
|
||||||
create: true
|
create: true
|
||||||
# The name of the ServiceAccount to use.
|
# The name of the ServiceAccount to use.
|
||||||
# If not set and create is true, a name is generated using the fullname template
|
# If not set and create is true, a name is generated using the fullname template
|
||||||
name:
|
name:
|
||||||
# Specifies custom annotations for the serviceAccount
|
|
||||||
annotations: {}
|
|
||||||
|
|
||||||
podAnnotations: {}
|
|
||||||
|
|
||||||
podLabels: {}
|
|
||||||
|
|
||||||
livenessProbe:
|
|
||||||
failureThreshold: 3
|
|
||||||
httpGet:
|
|
||||||
path: /healthz
|
|
||||||
port: 10258
|
|
||||||
scheme: HTTPS
|
|
||||||
initialDelaySeconds: 3
|
|
||||||
periodSeconds: 10
|
|
||||||
|
|
||||||
service:
|
|
||||||
enabled: false
|
|
||||||
|
|
||||||
serviceMonitor:
|
|
||||||
enabled: false
|
|
||||||
# The namespace where Prometheus expects to find service monitors.
|
|
||||||
# namespace: ""
|
|
||||||
interval: ""
|
|
||||||
# honorLabels: true
|
|
||||||
insecureSkipVerify: true
|
|
||||||
serverName: null
|
|
||||||
metricRelabelings: []
|
|
||||||
# - action: keep
|
|
||||||
# regex: 'descheduler_(build_info|pods_evicted)'
|
|
||||||
# sourceLabels: [__name__]
|
|
||||||
relabelings: []
|
|
||||||
# - sourceLabels: [__meta_kubernetes_pod_node_name]
|
|
||||||
# separator: ;
|
|
||||||
# regex: ^(.*)$
|
|
||||||
# targetLabel: nodename
|
|
||||||
# replacement: $1
|
|
||||||
# action: replace
|
|
||||||
|
|||||||
@@ -1,13 +1,13 @@
|
|||||||
# See https://cloud.google.com/cloud-build/docs/build-config
|
# See https://cloud.google.com/cloud-build/docs/build-config
|
||||||
|
|
||||||
# this must be specified in seconds. If omitted, defaults to 600s (10 mins)
|
# this must be specified in seconds. If omitted, defaults to 600s (10 mins)
|
||||||
timeout: 1500s
|
timeout: 1200s
|
||||||
# this prevents errors if you don't use both _GIT_TAG and _PULL_BASE_REF,
|
# this prevents errors if you don't use both _GIT_TAG and _PULL_BASE_REF,
|
||||||
# or any new substitutions added in the future.
|
# or any new substitutions added in the future.
|
||||||
options:
|
options:
|
||||||
substitution_option: ALLOW_LOOSE
|
substitution_option: ALLOW_LOOSE
|
||||||
steps:
|
steps:
|
||||||
- name: 'gcr.io/k8s-staging-test-infra/gcb-docker-gcloud:v20211118-2f2d816b90'
|
- name: 'gcr.io/k8s-testimages/gcb-docker-gcloud:v20190906-745fed4'
|
||||||
entrypoint: make
|
entrypoint: make
|
||||||
env:
|
env:
|
||||||
- DOCKER_CLI_EXPERIMENTAL=enabled
|
- DOCKER_CLI_EXPERIMENTAL=enabled
|
||||||
|
|||||||
@@ -18,14 +18,13 @@ limitations under the License.
|
|||||||
package options
|
package options
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"time"
|
|
||||||
|
|
||||||
"github.com/spf13/pflag"
|
"github.com/spf13/pflag"
|
||||||
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
|
||||||
|
utilerrors "k8s.io/apimachinery/pkg/util/errors"
|
||||||
apiserveroptions "k8s.io/apiserver/pkg/server/options"
|
apiserveroptions "k8s.io/apiserver/pkg/server/options"
|
||||||
clientset "k8s.io/client-go/kubernetes"
|
clientset "k8s.io/client-go/kubernetes"
|
||||||
componentbaseconfig "k8s.io/component-base/config"
|
"k8s.io/component-base/logs"
|
||||||
componentbaseoptions "k8s.io/component-base/config/options"
|
|
||||||
"sigs.k8s.io/descheduler/pkg/apis/componentconfig"
|
"sigs.k8s.io/descheduler/pkg/apis/componentconfig"
|
||||||
"sigs.k8s.io/descheduler/pkg/apis/componentconfig/v1alpha1"
|
"sigs.k8s.io/descheduler/pkg/apis/componentconfig/v1alpha1"
|
||||||
deschedulerscheme "sigs.k8s.io/descheduler/pkg/descheduler/scheme"
|
deschedulerscheme "sigs.k8s.io/descheduler/pkg/descheduler/scheme"
|
||||||
@@ -40,7 +39,7 @@ type DeschedulerServer struct {
|
|||||||
componentconfig.DeschedulerConfiguration
|
componentconfig.DeschedulerConfiguration
|
||||||
|
|
||||||
Client clientset.Interface
|
Client clientset.Interface
|
||||||
EventClient clientset.Interface
|
Logs *logs.Options
|
||||||
SecureServing *apiserveroptions.SecureServingOptionsWithLoopback
|
SecureServing *apiserveroptions.SecureServingOptionsWithLoopback
|
||||||
DisableMetrics bool
|
DisableMetrics bool
|
||||||
}
|
}
|
||||||
@@ -57,22 +56,20 @@ func NewDeschedulerServer() (*DeschedulerServer, error) {
|
|||||||
|
|
||||||
return &DeschedulerServer{
|
return &DeschedulerServer{
|
||||||
DeschedulerConfiguration: *cfg,
|
DeschedulerConfiguration: *cfg,
|
||||||
|
Logs: logs.NewOptions(),
|
||||||
SecureServing: secureServing,
|
SecureServing: secureServing,
|
||||||
}, nil
|
}, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Validation checks for DeschedulerServer.
|
||||||
|
func (s *DeschedulerServer) Validate() error {
|
||||||
|
var errs []error
|
||||||
|
errs = append(errs, s.Logs.Validate()...)
|
||||||
|
return utilerrors.NewAggregate(errs)
|
||||||
|
}
|
||||||
|
|
||||||
func newDefaultComponentConfig() (*componentconfig.DeschedulerConfiguration, error) {
|
func newDefaultComponentConfig() (*componentconfig.DeschedulerConfiguration, error) {
|
||||||
versionedCfg := v1alpha1.DeschedulerConfiguration{
|
versionedCfg := v1alpha1.DeschedulerConfiguration{}
|
||||||
LeaderElection: componentbaseconfig.LeaderElectionConfiguration{
|
|
||||||
LeaderElect: false,
|
|
||||||
LeaseDuration: metav1.Duration{Duration: 137 * time.Second},
|
|
||||||
RenewDeadline: metav1.Duration{Duration: 107 * time.Second},
|
|
||||||
RetryPeriod: metav1.Duration{Duration: 26 * time.Second},
|
|
||||||
ResourceLock: "leases",
|
|
||||||
ResourceName: "descheduler",
|
|
||||||
ResourceNamespace: "kube-system",
|
|
||||||
},
|
|
||||||
}
|
|
||||||
deschedulerscheme.Scheme.Default(&versionedCfg)
|
deschedulerscheme.Scheme.Default(&versionedCfg)
|
||||||
cfg := componentconfig.DeschedulerConfiguration{}
|
cfg := componentconfig.DeschedulerConfiguration{}
|
||||||
if err := deschedulerscheme.Scheme.Convert(&versionedCfg, &cfg, nil); err != nil {
|
if err := deschedulerscheme.Scheme.Convert(&versionedCfg, &cfg, nil); err != nil {
|
||||||
@@ -83,14 +80,18 @@ func newDefaultComponentConfig() (*componentconfig.DeschedulerConfiguration, err
|
|||||||
|
|
||||||
// AddFlags adds flags for a specific SchedulerServer to the specified FlagSet
|
// AddFlags adds flags for a specific SchedulerServer to the specified FlagSet
|
||||||
func (rs *DeschedulerServer) AddFlags(fs *pflag.FlagSet) {
|
func (rs *DeschedulerServer) AddFlags(fs *pflag.FlagSet) {
|
||||||
fs.StringVar(&rs.Logging.Format, "logging-format", "text", `Sets the log format. Permitted formats: "text", "json". Non-default formats don't honor these flags: --add-dir-header, --alsologtostderr, --log-backtrace-at, --log_dir, --log_file, --log_file_max_size, --logtostderr, --skip-headers, --skip-log-headers, --stderrthreshold, --log-flush-frequency.\nNon-default choices are currently alpha and subject to change without warning.`)
|
fs.StringVar(&rs.Logging.Format, "logging-format", "text", `Sets the log format. Permitted formats: "text", "json". Non-default formats don't honor these flags: --add-dir-header, --alsologtostderr, --log-backtrace-at, --log-dir, --log-file, --log-file-max-size, --logtostderr, --skip-headers, --skip-log-headers, --stderrthreshold, --log-flush-frequency.\nNon-default choices are currently alpha and subject to change without warning.`)
|
||||||
fs.DurationVar(&rs.DeschedulingInterval, "descheduling-interval", rs.DeschedulingInterval, "Time interval between two consecutive descheduler executions. Setting this value instructs the descheduler to run in a continuous loop at the interval specified.")
|
fs.DurationVar(&rs.DeschedulingInterval, "descheduling-interval", rs.DeschedulingInterval, "Time interval between two consecutive descheduler executions. Setting this value instructs the descheduler to run in a continuous loop at the interval specified.")
|
||||||
fs.StringVar(&rs.KubeconfigFile, "kubeconfig", rs.KubeconfigFile, "File with kube configuration.")
|
fs.StringVar(&rs.KubeconfigFile, "kubeconfig", rs.KubeconfigFile, "File with kube configuration.")
|
||||||
fs.StringVar(&rs.PolicyConfigFile, "policy-config-file", rs.PolicyConfigFile, "File with descheduler policy configuration.")
|
fs.StringVar(&rs.PolicyConfigFile, "policy-config-file", rs.PolicyConfigFile, "File with descheduler policy configuration.")
|
||||||
fs.BoolVar(&rs.DryRun, "dry-run", rs.DryRun, "execute descheduler in dry run mode.")
|
fs.BoolVar(&rs.DryRun, "dry-run", rs.DryRun, "execute descheduler in dry run mode.")
|
||||||
|
// node-selector query causes descheduler to run only on nodes that matches the node labels in the query
|
||||||
|
fs.StringVar(&rs.NodeSelector, "node-selector", rs.NodeSelector, "DEPRECATED: selector (label query) to filter on, supports '=', '==', and '!='.(e.g. -l key1=value1,key2=value2)")
|
||||||
|
// max-no-pods-to-evict limits the maximum number of pods to be evicted per node by descheduler.
|
||||||
|
fs.IntVar(&rs.MaxNoOfPodsToEvictPerNode, "max-pods-to-evict-per-node", rs.MaxNoOfPodsToEvictPerNode, "DEPRECATED: limits the maximum number of pods to be evicted per node by descheduler")
|
||||||
|
// evict-local-storage-pods allows eviction of pods that are using local storage. This is false by default.
|
||||||
|
fs.BoolVar(&rs.EvictLocalStoragePods, "evict-local-storage-pods", rs.EvictLocalStoragePods, "DEPRECATED: enables evicting pods using local storage by descheduler")
|
||||||
fs.BoolVar(&rs.DisableMetrics, "disable-metrics", rs.DisableMetrics, "Disables metrics. The metrics are by default served through https://localhost:10258/metrics. Secure address, resp. port can be changed through --bind-address, resp. --secure-port flags.")
|
fs.BoolVar(&rs.DisableMetrics, "disable-metrics", rs.DisableMetrics, "Disables metrics. The metrics are by default served through https://localhost:10258/metrics. Secure address, resp. port can be changed through --bind-address, resp. --secure-port flags.")
|
||||||
|
|
||||||
componentbaseoptions.BindLeaderElectionFlags(&rs.LeaderElection, fs)
|
|
||||||
|
|
||||||
rs.SecureServing.AddFlags(fs)
|
rs.SecureServing.AddFlags(fs)
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -19,11 +19,8 @@ package app
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
|
"flag"
|
||||||
"io"
|
"io"
|
||||||
"os/signal"
|
|
||||||
"syscall"
|
|
||||||
|
|
||||||
"k8s.io/apiserver/pkg/server/healthz"
|
|
||||||
|
|
||||||
"sigs.k8s.io/descheduler/cmd/descheduler/app/options"
|
"sigs.k8s.io/descheduler/cmd/descheduler/app/options"
|
||||||
"sigs.k8s.io/descheduler/pkg/descheduler"
|
"sigs.k8s.io/descheduler/pkg/descheduler"
|
||||||
@@ -33,8 +30,7 @@ import (
|
|||||||
apiserver "k8s.io/apiserver/pkg/server"
|
apiserver "k8s.io/apiserver/pkg/server"
|
||||||
"k8s.io/apiserver/pkg/server/mux"
|
"k8s.io/apiserver/pkg/server/mux"
|
||||||
restclient "k8s.io/client-go/rest"
|
restclient "k8s.io/client-go/rest"
|
||||||
registry "k8s.io/component-base/logs/api/v1"
|
aflag "k8s.io/component-base/cli/flag"
|
||||||
_ "k8s.io/component-base/logs/json/register"
|
|
||||||
"k8s.io/component-base/metrics/legacyregistry"
|
"k8s.io/component-base/metrics/legacyregistry"
|
||||||
"k8s.io/klog/v2"
|
"k8s.io/klog/v2"
|
||||||
)
|
)
|
||||||
@@ -52,7 +48,8 @@ func NewDeschedulerCommand(out io.Writer) *cobra.Command {
|
|||||||
Short: "descheduler",
|
Short: "descheduler",
|
||||||
Long: `The descheduler evicts pods which may be bound to less desired nodes`,
|
Long: `The descheduler evicts pods which may be bound to less desired nodes`,
|
||||||
Run: func(cmd *cobra.Command, args []string) {
|
Run: func(cmd *cobra.Command, args []string) {
|
||||||
// s.Logs.Config.Format = s.Logging.Format
|
s.Logs.Config.Format = s.Logging.Format
|
||||||
|
s.Logs.Apply()
|
||||||
|
|
||||||
// LoopbackClientConfig is a config for a privileged loopback connection
|
// LoopbackClientConfig is a config for a privileged loopback connection
|
||||||
var LoopbackClientConfig *restclient.Config
|
var LoopbackClientConfig *restclient.Config
|
||||||
@@ -61,49 +58,37 @@ func NewDeschedulerCommand(out io.Writer) *cobra.Command {
|
|||||||
klog.ErrorS(err, "failed to apply secure server configuration")
|
klog.ErrorS(err, "failed to apply secure server configuration")
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
var factory registry.LogFormatFactory
|
|
||||||
if factory == nil {
|
|
||||||
klog.ClearLogger()
|
|
||||||
} else {
|
|
||||||
log, logrFlush := factory.Create(registry.LoggingConfiguration{
|
|
||||||
Format: s.Logging.Format,
|
|
||||||
})
|
|
||||||
|
|
||||||
defer logrFlush()
|
if err := s.Validate(); err != nil {
|
||||||
klog.SetLogger(log)
|
klog.ErrorS(err, "failed to validate server configuration")
|
||||||
}
|
|
||||||
|
|
||||||
ctx, done := signal.NotifyContext(context.Background(), syscall.SIGINT, syscall.SIGTERM)
|
|
||||||
|
|
||||||
pathRecorderMux := mux.NewPathRecorderMux("descheduler")
|
|
||||||
if !s.DisableMetrics {
|
|
||||||
pathRecorderMux.Handle("/metrics", legacyregistry.HandlerWithReset())
|
|
||||||
}
|
|
||||||
|
|
||||||
healthz.InstallHandler(pathRecorderMux, healthz.NamedCheck("Descheduler", healthz.PingHealthz.Check))
|
|
||||||
|
|
||||||
stoppedCh, _, err := SecureServing.Serve(pathRecorderMux, 0, ctx.Done())
|
|
||||||
if err != nil {
|
|
||||||
klog.Fatalf("failed to start secure server: %v", err)
|
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
err = Run(ctx, s)
|
if !s.DisableMetrics {
|
||||||
|
ctx := context.TODO()
|
||||||
|
pathRecorderMux := mux.NewPathRecorderMux("descheduler")
|
||||||
|
pathRecorderMux.Handle("/metrics", legacyregistry.HandlerWithReset())
|
||||||
|
|
||||||
|
if _, err := SecureServing.Serve(pathRecorderMux, 0, ctx.Done()); err != nil {
|
||||||
|
klog.Fatalf("failed to start secure server: %v", err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
err := Run(s)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
klog.ErrorS(err, "descheduler server")
|
klog.ErrorS(err, "descheduler server")
|
||||||
}
|
}
|
||||||
|
|
||||||
done()
|
|
||||||
// wait for metrics server to close
|
|
||||||
<-stoppedCh
|
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
cmd.SetOut(out)
|
cmd.SetOut(out)
|
||||||
flags := cmd.Flags()
|
flags := cmd.Flags()
|
||||||
|
flags.SetNormalizeFunc(aflag.WordSepNormalizeFunc)
|
||||||
|
flags.AddGoFlagSet(flag.CommandLine)
|
||||||
s.AddFlags(flags)
|
s.AddFlags(flags)
|
||||||
return cmd
|
return cmd
|
||||||
}
|
}
|
||||||
|
|
||||||
func Run(ctx context.Context, rs *options.DeschedulerServer) error {
|
func Run(rs *options.DeschedulerServer) error {
|
||||||
return descheduler.Run(ctx, rs)
|
return descheduler.Run(rs)
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -17,23 +17,22 @@ limitations under the License.
|
|||||||
package main
|
package main
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
"fmt"
|
||||||
|
"k8s.io/component-base/logs"
|
||||||
"os"
|
"os"
|
||||||
|
|
||||||
"k8s.io/component-base/cli"
|
|
||||||
"k8s.io/klog/v2"
|
|
||||||
"sigs.k8s.io/descheduler/cmd/descheduler/app"
|
"sigs.k8s.io/descheduler/cmd/descheduler/app"
|
||||||
)
|
)
|
||||||
|
|
||||||
func init() {
|
|
||||||
klog.SetOutput(os.Stdout)
|
|
||||||
klog.InitFlags(nil)
|
|
||||||
}
|
|
||||||
|
|
||||||
func main() {
|
func main() {
|
||||||
out := os.Stdout
|
out := os.Stdout
|
||||||
cmd := app.NewDeschedulerCommand(out)
|
cmd := app.NewDeschedulerCommand(out)
|
||||||
cmd.AddCommand(app.NewVersionCommand())
|
cmd.AddCommand(app.NewVersionCommand())
|
||||||
|
|
||||||
code := cli.Run(cmd)
|
logs.InitLogs()
|
||||||
os.Exit(code)
|
defer logs.FlushLogs()
|
||||||
|
|
||||||
|
if err := cmd.Execute(); err != nil {
|
||||||
|
fmt.Println(err)
|
||||||
|
os.Exit(1)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -31,7 +31,7 @@ View all CLI options.
|
|||||||
## Run Tests
|
## Run Tests
|
||||||
```
|
```
|
||||||
GOOS=linux make dev-image
|
GOOS=linux make dev-image
|
||||||
make kind-multi-node
|
kind create cluster --config hack/kind_config.yaml
|
||||||
kind load docker-image <image name>
|
kind load docker-image <image name>
|
||||||
kind get kubeconfig > /tmp/admin.conf
|
kind get kubeconfig > /tmp/admin.conf
|
||||||
export KUBECONFIG=/tmp/admin.conf
|
export KUBECONFIG=/tmp/admin.conf
|
||||||
@@ -39,31 +39,17 @@ make test-unit
|
|||||||
make test-e2e
|
make test-e2e
|
||||||
```
|
```
|
||||||
|
|
||||||
|
## Run Helm Tests
|
||||||
## Build Helm Package locally
|
Run the helm test for a particular descheduler release by setting below variables,
|
||||||
|
```
|
||||||
If you made some changes in the chart, and just want to check if templating is ok, or if the chart is buildable, you can run this command to have a package built from the `./charts` directory.
|
HELM_IMAGE_REPO="descheduler"
|
||||||
|
HELM_IMAGE_TAG="helm-test"
|
||||||
|
HELM_CHART_LOCATION="./charts/descheduler"
|
||||||
|
```
|
||||||
|
The helm tests runs as part of descheduler CI. But, to run it manually from the descheduler root,
|
||||||
|
|
||||||
```
|
```
|
||||||
make build-helm
|
make test-helm
|
||||||
```
|
|
||||||
|
|
||||||
## Lint Helm Chart locally
|
|
||||||
|
|
||||||
To check linting of your changes in the helm chart locally you can run:
|
|
||||||
|
|
||||||
```
|
|
||||||
make lint-helm
|
|
||||||
```
|
|
||||||
|
|
||||||
## Test helm changes locally with kind and ct
|
|
||||||
|
|
||||||
You will need kind and docker (or equivalent) installed. We can use ct public image to avoid installing ct and all its dependencies.
|
|
||||||
|
|
||||||
|
|
||||||
```
|
|
||||||
make kind-multi-node
|
|
||||||
make ct-helm
|
|
||||||
```
|
```
|
||||||
|
|
||||||
### Miscellaneous
|
### Miscellaneous
|
||||||
|
|||||||
@@ -1,16 +0,0 @@
|
|||||||
# Proposals
|
|
||||||
This document walk you through about all the enhancements proposals for descheduler.
|
|
||||||
|
|
||||||
## Descheduler v1alpha2 Design Proposal
|
|
||||||
```yaml
|
|
||||||
title: Descheduler v1alpha2 Design Proposal
|
|
||||||
authors:
|
|
||||||
- "@damemi"
|
|
||||||
link:
|
|
||||||
- https://docs.google.com/document/d/1S1JCh-0F-QCJvBBG-kbmXiHAJFF8doArhDIAKbOj93I/edit#heading=h.imbp1ctnc8lx
|
|
||||||
- https://github.com/kubernetes-sigs/descheduler/issues/679
|
|
||||||
owning-sig: sig-scheduling
|
|
||||||
creation-date: 2021-05-01
|
|
||||||
status: implementable
|
|
||||||
```
|
|
||||||
|
|
||||||
@@ -1,82 +1,36 @@
|
|||||||
# Release Guide
|
# Release Guide
|
||||||
|
|
||||||
The process for publishing each Descheduler release includes a mixture of manual and automatic steps. Over
|
## Container Image
|
||||||
time, it would be good to automate as much of this process as possible. However, due to current limitations there
|
|
||||||
is care that must be taken to perform each manual step precisely so that the automated steps execute properly.
|
|
||||||
|
|
||||||
## Pre-release Code Changes
|
### Semi-automatic
|
||||||
|
|
||||||
Before publishing each release, the following code updates must be made:
|
1. Make sure your repo is clean by git's standards
|
||||||
|
2. Create a release branch `git checkout -b release-1.18` (not required for patch releases)
|
||||||
|
3. Push the release branch to the descheuler repo and ensure branch protection is enabled (not required for patch releases)
|
||||||
|
4. Tag the repository from the `master` branch (from the `release-1.18` branch for a patch release) and push the tag `VERSION=v0.18.0 git tag -m $VERSION $VERSION; git push origin $VERSION`
|
||||||
|
5. Publish a draft release using the tag you just created
|
||||||
|
6. Perform the [image promotion process](https://github.com/kubernetes/k8s.io/tree/main/k8s.gcr.io#image-promoter)
|
||||||
|
7. Publish release
|
||||||
|
8. Email `kubernetes-sig-scheduling@googlegroups.com` to announce the release
|
||||||
|
|
||||||
- [ ] (Optional, but recommended) Bump `k8s.io` dependencies to the `-rc` tags. These tags are usually published around upstream code freeze. [Example](https://github.com/kubernetes-sigs/descheduler/pull/539)
|
### Manual
|
||||||
- [ ] Bump `k8s.io` dependencies to GA tags once they are published (following the upstream release). [Example](https://github.com/kubernetes-sigs/descheduler/pull/615)
|
|
||||||
- [ ] Ensure that Go is updated to the same version as upstream. [Example](https://github.com/kubernetes-sigs/descheduler/pull/801)
|
|
||||||
- [ ] Make CI changes in [github.com/kubernetes/test-infra](https://github.com/kubernetes/test-infra) to add the new version's tests (note, this may also include a Go bump). [Example](https://github.com/kubernetes/test-infra/pull/25833)
|
|
||||||
- [ ] Update local CI versions for utils (such as golang-ci), kind, and go. [Example - e2e](https://github.com/kubernetes-sigs/descheduler/commit/ac4d576df8831c0c399ee8fff1e85469e90b8c44), [Example - helm](https://github.com/kubernetes-sigs/descheduler/pull/821)
|
|
||||||
- [ ] Update version references in docs and Readme. [Example](https://github.com/kubernetes-sigs/descheduler/pull/617)
|
|
||||||
|
|
||||||
## Release Process
|
1. Make sure your repo is clean by git's standards
|
||||||
|
2. Create a release branch `git checkout -b release-1.18` (not required for patch releases)
|
||||||
|
3. Push the release branch to the descheuler repo and ensure branch protection is enabled (not required for patch releases)
|
||||||
|
4. Tag the repository from the `master` branch (from the `release-1.18` branch for a patch release) and push the tag `VERSION=v0.18.0 git tag -m $VERSION $VERSION; git push origin $VERSION`
|
||||||
|
5. Checkout the tag you just created and make sure your repo is clean by git's standards `git checkout $VERSION`
|
||||||
|
6. Build and push the container image to the staging registry `VERSION=$VERSION make push-all`
|
||||||
|
7. Publish a draft release using the tag you just created
|
||||||
|
8. Perform the [image promotion process](https://github.com/kubernetes/k8s.io/tree/main/k8s.gcr.io#image-promoter)
|
||||||
|
9. Publish release
|
||||||
|
10. Email `kubernetes-sig-scheduling@googlegroups.com` to announce the release
|
||||||
|
|
||||||
When the above pre-release steps are complete and the release is ready to be cut, perform the following steps **in order**
|
### Notes
|
||||||
(the flowchart below demonstrates these steps):
|
It's important to create the tag on the master branch after creating the release-* branch so that the [Helm releaser-action](#helm-chart) can work.
|
||||||
|
It compares the changes in the action-triggering branch to the latest tag on that branch, so if you tag before creating the new branch there
|
||||||
**Version release**
|
will be nothing to compare and it will fail (creating a new release branch usually involves no code changes). For this same reason, you should
|
||||||
1. Create the `git tag` on `master` for the release, eg `v0.24.0`
|
also tag patch releases (on the release-* branch) *after* pushing changes (if those changes involve bumping the Helm chart version).
|
||||||
2. Merge Helm chart version update to `master` (see [Helm chart](#helm-chart) below). [Example](https://github.com/kubernetes-sigs/descheduler/pull/709)
|
|
||||||
3. Perform the [image promotion process](https://github.com/kubernetes/k8s.io/tree/main/k8s.gcr.io#image-promoter). [Example](https://github.com/kubernetes/k8s.io/pull/3344)
|
|
||||||
4. Cut release branch from `master`, eg `release-1.24`
|
|
||||||
5. Publish release using Github's release process from the git tag you created
|
|
||||||
6. Email `kubernetes-sig-scheduling@googlegroups.com` to announce the release
|
|
||||||
|
|
||||||
**Patch release**
|
|
||||||
1. Pick relevant code change commits to the matching release branch, eg `release-1.24`
|
|
||||||
2. Create the patch tag on the release branch, eg `v0.24.1` on `release-1.24`
|
|
||||||
3. Merge Helm chart version update to release branch
|
|
||||||
4. Perform the image promotion process for the patch version
|
|
||||||
5. Publish release using Github's release process from the git tag you created
|
|
||||||
6. Email `kubernetes-sig-scheduling@googlegroups.com` to announce the release
|
|
||||||
|
|
||||||
### Flowchart
|
|
||||||
|
|
||||||

|
|
||||||
|
|
||||||
### Image promotion process
|
|
||||||
|
|
||||||
Every merge to any branch triggers an [image build and push](https://github.com/kubernetes/test-infra/blob/c36b8e5/config/jobs/image-pushing/k8s-staging-descheduler.yaml) to a `gcr.io` repository.
|
|
||||||
These automated image builds are snapshots of the code in place at the time of every PR merge and
|
|
||||||
tagged with the latest git SHA at the time of the build. To create a final release image, the desired
|
|
||||||
auto-built image SHA is added to a [file upstream](https://github.com/kubernetes/k8s.io/blob/e9e971c/k8s.gcr.io/images/k8s-staging-descheduler/images.yaml) which
|
|
||||||
copies that image to a public registry.
|
|
||||||
|
|
||||||
Automatic builds can be monitored and re-triggered with the [`post-descheduler-push-images` job](https://prow.k8s.io/?job=post-descheduler-push-images) on prow.k8s.io.
|
|
||||||
|
|
||||||
Note that images can also be manually built and pushed using `VERSION=$VERSION make push-all` by [users with access](https://github.com/kubernetes/k8s.io/blob/fbee8f67b70304241e613a672c625ad972998ad7/groups/sig-scheduling/groups.yaml#L33-L43).
|
|
||||||
|
|
||||||
## Helm Chart
|
|
||||||
We currently use the [chart-releaser-action GitHub Action](https://github.com/helm/chart-releaser-action) to automatically
|
|
||||||
publish [Helm chart releases](https://github.com/kubernetes-sigs/descheduler/blob/022e07c/.github/workflows/release.yaml).
|
|
||||||
This action is triggered when it detects any changes to [`Chart.yaml`](https://github.com/kubernetes-sigs/descheduler/blob/022e07c27853fade6d1304adc0a6ebe02642386c/charts/descheduler/Chart.yaml) on
|
|
||||||
a `release-*` branch.
|
|
||||||
|
|
||||||
Helm chart releases are managed by a separate set of git tags that are prefixed with `descheduler-helm-chart-*`. Example git tag name is `descheduler-helm-chart-0.18.0`.
|
|
||||||
Released versions of the helm charts are stored in the `gh-pages` branch of this repo.
|
|
||||||
|
|
||||||
The major and minor version of the chart matches the descheduler major and minor versions. For example descheduler helm chart version helm-descheduler-chart-0.18.0 corresponds
|
|
||||||
to descheduler version v0.18.0. The patch version of the descheduler helm chart and the patcher version of the descheduler will not necessarily match. The patch
|
|
||||||
version of the descheduler helm chart is used to version changes specific to the helm chart.
|
|
||||||
|
|
||||||
1. Merge all helm chart changes into the master branch before the release is tagged/cut
|
|
||||||
1. Ensure that `appVersion` in file `charts/descheduler/Chart.yaml` matches the descheduler version(no `v` prefix)
|
|
||||||
2. Ensure that `version` in file `charts/descheduler/Chart.yaml` has been incremented. This is the chart version.
|
|
||||||
2. Make sure your repo is clean by git's standards
|
|
||||||
3. Follow the release-branch or patch release tagging pattern from the above section.
|
|
||||||
4. Verify the new helm artifact has been successfully pushed to the `gh-pages` branch
|
|
||||||
|
|
||||||
## Notes
|
|
||||||
The Helm releaser-action compares the changes in the action-triggering branch to the latest tag on that branch, so if you tag before creating the new branch there
|
|
||||||
will be nothing to compare and it will fail. This is why it's necessary to tag, eg, `v0.24.0` *before* making the changes to the
|
|
||||||
Helm chart version, so that there is a new diff for the action to find. (Tagging *after* making the Helm chart changes would
|
|
||||||
also work, but then the code that gets built into the promoted image will be tagged as `descheduler-helm-chart-xxx` rather than `v0.xx.0`).
|
|
||||||
|
|
||||||
See [post-descheduler-push-images dashboard](https://testgrid.k8s.io/sig-scheduling#post-descheduler-push-images) for staging registry image build job status.
|
See [post-descheduler-push-images dashboard](https://testgrid.k8s.io/sig-scheduling#post-descheduler-push-images) for staging registry image build job status.
|
||||||
|
|
||||||
@@ -102,3 +56,19 @@ Pull image from the staging registry.
|
|||||||
```
|
```
|
||||||
docker pull gcr.io/k8s-staging-descheduler/descheduler:v20200206-0.9.0-94-ge2a23f284
|
docker pull gcr.io/k8s-staging-descheduler/descheduler:v20200206-0.9.0-94-ge2a23f284
|
||||||
```
|
```
|
||||||
|
|
||||||
|
## Helm Chart
|
||||||
|
Helm chart releases are managed by a separate set of git tags that are prefixed with `descheduler-helm-chart-*`. Example git tag name is `descheduler-helm-chart-0.18.0`.
|
||||||
|
Released versions of the helm charts are stored in the `gh-pages` branch of this repo. The [chart-releaser-action GitHub Action](https://github.com/helm/chart-releaser-action)
|
||||||
|
is setup to build and push the helm charts to the `gh-pages` branch when changes are pushed to a `release-*` branch.
|
||||||
|
|
||||||
|
The major and minor version of the chart matches the descheduler major and minor versions. For example descheduler helm chart version helm-descheduler-chart-0.18.0 corresponds
|
||||||
|
to descheduler version v0.18.0. The patch version of the descheduler helm chart and the patcher version of the descheduler will not necessarily match. The patch
|
||||||
|
version of the descheduler helm chart is used to version changes specific to the helm chart.
|
||||||
|
|
||||||
|
1. Merge all helm chart changes into the master branch before the release is tagged/cut
|
||||||
|
1. Ensure that `appVersion` in file `charts/descheduler/Chart.yaml` matches the descheduler version(no `v` prefix)
|
||||||
|
2. Ensure that `version` in file `charts/descheduler/Chart.yaml` has been incremented. This is the chart version.
|
||||||
|
2. Make sure your repo is clean by git's standards
|
||||||
|
3. Follow the release-branch or patch release tagging pattern from the above section.
|
||||||
|
4. Verify the new helm artifact has been successfully pushed to the `gh-pages` branch
|
||||||
Binary file not shown.
|
Before Width: | Height: | Size: 121 KiB |
@@ -2,19 +2,14 @@
|
|||||||
|
|
||||||
Starting with descheduler release v0.10.0 container images are available in the official k8s container registry.
|
Starting with descheduler release v0.10.0 container images are available in the official k8s container registry.
|
||||||
|
|
||||||
Descheduler Version | Container Image | Architectures |
|
Descheduler Version | Container Image | Architectures |
|
||||||
------------------- |--------------------------------------------|-------------------------|
|
------------------- |-----------------------------------------------------|-------------------------|
|
||||||
v0.25.1 | k8s.gcr.io/descheduler/descheduler:v0.25.1 | AMD64<br>ARM64<br>ARMv7 |
|
v0.22.0 | k8s.gcr.io/descheduler/descheduler:v0.22.0 | AMD64<br>ARM64<br>ARMv7 |
|
||||||
v0.25.0 | k8s.gcr.io/descheduler/descheduler:v0.25.0 | AMD64<br>ARM64<br>ARMv7 |
|
v0.21.0 | k8s.gcr.io/descheduler/descheduler:v0.21.0 | AMD64<br>ARM64<br>ARMv7 |
|
||||||
v0.24.1 | k8s.gcr.io/descheduler/descheduler:v0.24.1 | AMD64<br>ARM64<br>ARMv7 |
|
v0.20.0 | k8s.gcr.io/descheduler/descheduler:v0.20.0 | AMD64<br>ARM64 |
|
||||||
v0.24.0 | k8s.gcr.io/descheduler/descheduler:v0.24.0 | AMD64<br>ARM64<br>ARMv7 |
|
v0.19.0 | k8s.gcr.io/descheduler/descheduler:v0.19.0 | AMD64 |
|
||||||
v0.23.1 | k8s.gcr.io/descheduler/descheduler:v0.23.1 | AMD64<br>ARM64<br>ARMv7 |
|
v0.18.0 | k8s.gcr.io/descheduler/descheduler:v0.18.0 | AMD64 |
|
||||||
v0.22.0 | k8s.gcr.io/descheduler/descheduler:v0.22.0 | AMD64<br>ARM64<br>ARMv7 |
|
v0.10.0 | k8s.gcr.io/descheduler/descheduler:v0.10.0 | AMD64 |
|
||||||
v0.21.0 | k8s.gcr.io/descheduler/descheduler:v0.21.0 | AMD64<br>ARM64<br>ARMv7 |
|
|
||||||
v0.20.0 | k8s.gcr.io/descheduler/descheduler:v0.20.0 | AMD64<br>ARM64 |
|
|
||||||
v0.19.0 | k8s.gcr.io/descheduler/descheduler:v0.19.0 | AMD64 |
|
|
||||||
v0.18.0 | k8s.gcr.io/descheduler/descheduler:v0.18.0 | AMD64 |
|
|
||||||
v0.10.0 | k8s.gcr.io/descheduler/descheduler:v0.10.0 | AMD64 |
|
|
||||||
|
|
||||||
Note that multi-arch container images cannot be pulled by [kind](https://kind.sigs.k8s.io) from a registry. Therefore
|
Note that multi-arch container images cannot be pulled by [kind](https://kind.sigs.k8s.io) from a registry. Therefore
|
||||||
starting with descheduler release v0.20.0 use the below process to download the official descheduler
|
starting with descheduler release v0.20.0 use the below process to download the official descheduler
|
||||||
@@ -39,52 +34,31 @@ Usage:
|
|||||||
descheduler [command]
|
descheduler [command]
|
||||||
|
|
||||||
Available Commands:
|
Available Commands:
|
||||||
completion generate the autocompletion script for the specified shell
|
|
||||||
help Help about any command
|
help Help about any command
|
||||||
version Version of descheduler
|
version Version of descheduler
|
||||||
|
|
||||||
Flags:
|
Flags:
|
||||||
--add-dir-header If true, adds the file directory to the header of the log messages (DEPRECATED: will be removed in a future release, see https://github.com/kubernetes/enhancements/tree/master/keps/sig-instrumentation/2845-deprecate-klog-specific-flags-in-k8s-components)
|
--add-dir-header If true, adds the file directory to the header of the log messages
|
||||||
--alsologtostderr log to standard error as well as files (DEPRECATED: will be removed in a future release, see https://github.com/kubernetes/enhancements/tree/master/keps/sig-instrumentation/2845-deprecate-klog-specific-flags-in-k8s-components)
|
--alsologtostderr log to standard error as well as files
|
||||||
--bind-address ip The IP address on which to listen for the --secure-port port. The associated interface(s) must be reachable by the rest of the cluster, and by CLI/web clients. If blank or an unspecified address (0.0.0.0 or ::), all interfaces will be used. (default 0.0.0.0)
|
--descheduling-interval duration Time interval between two consecutive descheduler executions. Setting this value instructs the descheduler to run in a continuous loop at the interval specified.
|
||||||
--cert-dir string The directory where the TLS certs are located. If --tls-cert-file and --tls-private-key-file are provided, this flag will be ignored. (default "apiserver.local.config/certificates")
|
--dry-run execute descheduler in dry run mode.
|
||||||
--descheduling-interval duration Time interval between two consecutive descheduler executions. Setting this value instructs the descheduler to run in a continuous loop at the interval specified.
|
--evict-local-storage-pods DEPRECATED: enables evicting pods using local storage by descheduler
|
||||||
--disable-metrics Disables metrics. The metrics are by default served through https://localhost:10258/metrics. Secure address, resp. port can be changed through --bind-address, resp. --secure-port flags.
|
-h, --help help for descheduler
|
||||||
--dry-run execute descheduler in dry run mode.
|
--kubeconfig string File with kube configuration.
|
||||||
-h, --help help for descheduler
|
--log-backtrace-at traceLocation when logging hits line file:N, emit a stack trace (default :0)
|
||||||
--http2-max-streams-per-connection int The limit that the server gives to clients for the maximum number of streams in an HTTP/2 connection. Zero means to use golang's default.
|
--log-dir string If non-empty, write log files in this directory
|
||||||
--kubeconfig string File with kube configuration.
|
--log-file string If non-empty, use this log file
|
||||||
--leader-elect Start a leader election client and gain leadership before executing the main loop. Enable this when running replicated components for high availability.
|
--log-file-max-size uint Defines the maximum size a log file can grow to. Unit is megabytes. If the value is 0, the maximum file size is unlimited. (default 1800)
|
||||||
--leader-elect-lease-duration duration The duration that non-leader candidates will wait after observing a leadership renewal until attempting to acquire leadership of a led but unrenewed leader slot. This is effectively the maximum duration that a leader can be stopped before it is replaced by another candidate. This is only applicable if leader election is enabled. (default 15s)
|
--log-flush-frequency duration Maximum number of seconds between log flushes (default 5s)
|
||||||
--leader-elect-renew-deadline duration The interval between attempts by the acting master to renew a leadership slot before it stops leading. This must be less than or equal to the lease duration. This is only applicable if leader election is enabled. (default 10s)
|
--logtostderr log to standard error instead of files (default true)
|
||||||
--leader-elect-resource-lock string The type of resource object that is used for locking during leader election. Supported options are 'endpoints', 'configmaps', 'leases', 'endpointsleases' and 'configmapsleases'. (default "leases")
|
--max-pods-to-evict-per-node int DEPRECATED: limits the maximum number of pods to be evicted per node by descheduler
|
||||||
--leader-elect-resource-name string The name of resource object that is used for locking during leader election. (default "descheduler")
|
--node-selector string DEPRECATED: selector (label query) to filter on, supports '=', '==', and '!='.(e.g. -l key1=value1,key2=value2)
|
||||||
--leader-elect-resource-namespace string The namespace of resource object that is used for locking during leader election. (default "kube-system")
|
--policy-config-file string File with descheduler policy configuration.
|
||||||
--leader-elect-retry-period duration The duration the clients should wait between attempting acquisition and renewal of a leadership. This is only applicable if leader election is enabled. (default 2s)
|
--skip-headers If true, avoid header prefixes in the log messages
|
||||||
--log-backtrace-at traceLocation when logging hits line file:N, emit a stack trace (default :0) (DEPRECATED: will be removed in a future release, see https://github.com/kubernetes/enhancements/tree/master/keps/sig-instrumentation/2845-deprecate-klog-specific-flags-in-k8s-components)
|
--skip-log-headers If true, avoid headers when opening log files
|
||||||
--log_dir string If non-empty, write log files in this directory (DEPRECATED: will be removed in a future release, see https://github.com/kubernetes/enhancements/tree/master/keps/sig-instrumentation/2845-deprecate-klog-specific-flags-in-k8s-components)
|
--stderrthreshold severity logs at or above this threshold go to stderr (default 2)
|
||||||
--log_file string If non-empty, use this log file (DEPRECATED: will be removed in a future release, see https://github.com/kubernetes/enhancements/tree/master/keps/sig-instrumentation/2845-deprecate-klog-specific-flags-in-k8s-components)
|
-v, --v Level number for the log level verbosity
|
||||||
--log_file_max_size uint Defines the maximum size a log file can grow to. Unit is megabytes. If the value is 0, the maximum file size is unlimited. (default 1800) (DEPRECATED: will be removed in a future release, see https://github.com/kubernetes/enhancements/tree/master/keps/sig-instrumentation/2845-deprecate-klog-specific-flags-in-k8s-components)
|
--vmodule moduleSpec comma-separated list of pattern=N settings for file-filtered logging
|
||||||
--log-flush-frequency duration Maximum number of seconds between log flushes (default 5s)
|
|
||||||
--logging-format string Sets the log format. Permitted formats: "text", "json". Non-default formats don't honor these flags: --add-dir-header, --alsologtostderr, --log-backtrace-at, --log_dir, --log_file, --log_file_max_size, --logtostderr, --skip-headers, --skip-log-headers, --stderrthreshold, --log-flush-frequency.\nNon-default choices are currently alpha and subject to change without warning. (default "text")
|
|
||||||
--logtostderr log to standard error instead of files (default true) (DEPRECATED: will be removed in a future release, see https://github.com/kubernetes/enhancements/tree/master/keps/sig-instrumentation/2845-deprecate-klog-specific-flags-in-k8s-components)
|
|
||||||
--one-output If true, only write logs to their native severity level (vs also writing to each lower severity level) (DEPRECATED: will be removed in a future release, see https://github.com/kubernetes/enhancements/tree/master/keps/sig-instrumentation/2845-deprecate-klog-specific-flags-in-k8s-components)
|
|
||||||
--permit-address-sharing If true, SO_REUSEADDR will be used when binding the port. This allows binding to wildcard IPs like 0.0.0.0 and specific IPs in parallel, and it avoids waiting for the kernel to release sockets in TIME_WAIT state. [default=false]
|
|
||||||
--permit-port-sharing If true, SO_REUSEPORT will be used when binding the port, which allows more than one instance to bind on the same address and port. [default=false]
|
|
||||||
--policy-config-file string File with descheduler policy configuration.
|
|
||||||
--secure-port int The port on which to serve HTTPS with authentication and authorization. If 0, don't serve HTTPS at all. (default 10258)
|
|
||||||
--skip-headers If true, avoid header prefixes in the log messages (DEPRECATED: will be removed in a future release, see https://github.com/kubernetes/enhancements/tree/master/keps/sig-instrumentation/2845-deprecate-klog-specific-flags-in-k8s-components)
|
|
||||||
--skip-log-headers If true, avoid headers when opening log files (DEPRECATED: will be removed in a future release, see https://github.com/kubernetes/enhancements/tree/master/keps/sig-instrumentation/2845-deprecate-klog-specific-flags-in-k8s-components)
|
|
||||||
--stderrthreshold severity logs at or above this threshold go to stderr (default 2) (DEPRECATED: will be removed in a future release, see https://github.com/kubernetes/enhancements/tree/master/keps/sig-instrumentation/2845-deprecate-klog-specific-flags-in-k8s-components)
|
|
||||||
--tls-cert-file string File containing the default x509 Certificate for HTTPS. (CA cert, if any, concatenated after server cert). If HTTPS serving is enabled, and --tls-cert-file and --tls-private-key-file are not provided, a self-signed certificate and key are generated for the public address and saved to the directory specified by --cert-dir.
|
|
||||||
--tls-cipher-suites strings Comma-separated list of cipher suites for the server. If omitted, the default Go cipher suites will be used.
|
|
||||||
Preferred values: TLS_AES_128_GCM_SHA256, TLS_AES_256_GCM_SHA384, TLS_CHACHA20_POLY1305_SHA256, TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA, TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256, TLS_ECDHE_ECDSA_WITH_AES_256_CBC_SHA, TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384, TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305, TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256, TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA, TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256, TLS_ECDHE_RSA_WITH_AES_256_CBC_SHA, TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384, TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305, TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256, TLS_RSA_WITH_AES_128_CBC_SHA, TLS_RSA_WITH_AES_128_GCM_SHA256, TLS_RSA_WITH_AES_256_CBC_SHA, TLS_RSA_WITH_AES_256_GCM_SHA384.
|
|
||||||
Insecure values: TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256, TLS_ECDHE_ECDSA_WITH_RC4_128_SHA, TLS_ECDHE_RSA_WITH_3DES_EDE_CBC_SHA, TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256, TLS_ECDHE_RSA_WITH_RC4_128_SHA, TLS_RSA_WITH_3DES_EDE_CBC_SHA, TLS_RSA_WITH_AES_128_CBC_SHA256, TLS_RSA_WITH_RC4_128_SHA.
|
|
||||||
--tls-min-version string Minimum TLS version supported. Possible values: VersionTLS10, VersionTLS11, VersionTLS12, VersionTLS13
|
|
||||||
--tls-private-key-file string File containing the default x509 private key matching --tls-cert-file.
|
|
||||||
--tls-sni-cert-key namedCertKey A pair of x509 certificate and private key file paths, optionally suffixed with a list of domain patterns which are fully qualified domain names, possibly with prefixed wildcard segments. The domain patterns also allow IP addresses, but IPs should only be used if the apiserver has visibility to the IP address requested by a client. If no domain patterns are provided, the names of the certificate are extracted. Non-wildcard matches trump over wildcard matches, explicit domain patterns trump over extracted names. For multiple key/certificate pairs, use the --tls-sni-cert-key multiple times. Examples: "example.crt,example.key" or "foo.crt,foo.key:*.foo.com,foo.com". (default [])
|
|
||||||
-v, --v Level number for the log level verbosity
|
|
||||||
--vmodule moduleSpec comma-separated list of pattern=N settings for file-filtered logging
|
|
||||||
|
|
||||||
Use "descheduler [command] --help" for more information about a command.
|
Use "descheduler [command] --help" for more information about a command.
|
||||||
```
|
```
|
||||||
@@ -115,8 +89,7 @@ strategies:
|
|||||||
"PodLifeTime":
|
"PodLifeTime":
|
||||||
enabled: true
|
enabled: true
|
||||||
params:
|
params:
|
||||||
podLifeTime:
|
maxPodLifeTimeSeconds: 604800 # pods run for a maximum of 7 days
|
||||||
maxPodLifeTimeSeconds: 604800 # pods run for a maximum of 7 days
|
|
||||||
```
|
```
|
||||||
|
|
||||||
### Balance Cluster By Node Memory Utilization
|
### Balance Cluster By Node Memory Utilization
|
||||||
@@ -144,7 +117,7 @@ strategies:
|
|||||||
|
|
||||||
#### Balance low utilization nodes
|
#### Balance low utilization nodes
|
||||||
Using `HighNodeUtilization`, descheduler will rebalance the cluster based on memory by evicting pods
|
Using `HighNodeUtilization`, descheduler will rebalance the cluster based on memory by evicting pods
|
||||||
from nodes with memory utilization lower than 20%. This should be use `NodeResourcesFit` with the `MostAllocated` scoring strategy based on these [doc](https://kubernetes.io/docs/reference/scheduling/config/#scheduling-plugins).
|
from nodes with memory utilization lower than 20%. This should be used along with scheduler strategy `MostRequestedPriority`.
|
||||||
The evicted pods will be compacted into minimal set of nodes.
|
The evicted pods will be compacted into minimal set of nodes.
|
||||||
|
|
||||||
```
|
```
|
||||||
@@ -163,14 +136,7 @@ strategies:
|
|||||||
Descheduler's `RemovePodsViolatingNodeTaints` strategy can be combined with
|
Descheduler's `RemovePodsViolatingNodeTaints` strategy can be combined with
|
||||||
[Node Problem Detector](https://github.com/kubernetes/node-problem-detector/) and
|
[Node Problem Detector](https://github.com/kubernetes/node-problem-detector/) and
|
||||||
[Cluster Autoscaler](https://github.com/kubernetes/autoscaler/tree/master/cluster-autoscaler) to automatically remove
|
[Cluster Autoscaler](https://github.com/kubernetes/autoscaler/tree/master/cluster-autoscaler) to automatically remove
|
||||||
Nodes which have problems. Node Problem Detector can detect specific Node problems and report them to the API server.
|
Nodes which have problems. Node Problem Detector can detect specific Node problems and taint any Nodes which have those
|
||||||
There is a feature called TaintNodeByCondition of the node controller that takes some conditions and turns them into taints. Currently, this only works for the default node conditions: PIDPressure, MemoryPressure, DiskPressure, Ready, and some cloud provider specific conditions.
|
problems. The Descheduler will then deschedule workloads from those Nodes. Finally, if the descheduled Node's resource
|
||||||
The Descheduler will then deschedule workloads from those Nodes. Finally, if the descheduled Node's resource
|
|
||||||
allocation falls below the Cluster Autoscaler's scale down threshold, the Node will become a scale down candidate
|
allocation falls below the Cluster Autoscaler's scale down threshold, the Node will become a scale down candidate
|
||||||
and can be removed by Cluster Autoscaler. These three components form an autohealing cycle for Node problems.
|
and can be removed by Cluster Autoscaler. These three components form an autohealing cycle for Node problems.
|
||||||
---
|
|
||||||
**NOTE**
|
|
||||||
|
|
||||||
Once [kubernetes/node-problem-detector#565](https://github.com/kubernetes/node-problem-detector/pull/565) is available in NPD, we need to update this section.
|
|
||||||
|
|
||||||
---
|
|
||||||
|
|||||||
@@ -11,4 +11,4 @@ strategies:
|
|||||||
includingInitContainers: true
|
includingInitContainers: true
|
||||||
excludeOwnerKinds:
|
excludeOwnerKinds:
|
||||||
- "Job"
|
- "Job"
|
||||||
minPodLifetimeSeconds: 3600 # 1 hour
|
minPodLifeTimeSeconds: 3600 # 1 hour
|
||||||
|
|||||||
@@ -6,6 +6,3 @@ strategies:
|
|||||||
params:
|
params:
|
||||||
podLifeTime:
|
podLifeTime:
|
||||||
maxPodLifeTimeSeconds: 604800 # 7 days
|
maxPodLifeTimeSeconds: 604800 # 7 days
|
||||||
states:
|
|
||||||
- "Pending"
|
|
||||||
- "PodInitializing"
|
|
||||||
|
|||||||
@@ -4,5 +4,4 @@ strategies:
|
|||||||
"RemovePodsViolatingTopologySpreadConstraint":
|
"RemovePodsViolatingTopologySpreadConstraint":
|
||||||
enabled: true
|
enabled: true
|
||||||
params:
|
params:
|
||||||
nodeFit: true
|
|
||||||
includeSoftConstraints: true # Include 'ScheduleAnyways' constraints
|
includeSoftConstraints: true # Include 'ScheduleAnyways' constraints
|
||||||
|
|||||||
118
go.mod
118
go.mod
@@ -1,115 +1,19 @@
|
|||||||
module sigs.k8s.io/descheduler
|
module sigs.k8s.io/descheduler
|
||||||
|
|
||||||
go 1.19
|
go 1.16
|
||||||
|
|
||||||
require (
|
require (
|
||||||
github.com/client9/misspell v0.3.4
|
github.com/client9/misspell v0.3.4
|
||||||
github.com/spf13/cobra v1.4.0
|
github.com/spf13/cobra v1.1.3
|
||||||
github.com/spf13/pflag v1.0.5
|
github.com/spf13/pflag v1.0.5
|
||||||
k8s.io/api v0.25.0
|
k8s.io/api v0.22.0
|
||||||
k8s.io/apimachinery v0.25.0
|
k8s.io/apimachinery v0.22.0
|
||||||
k8s.io/apiserver v0.25.0
|
k8s.io/apiserver v0.22.0
|
||||||
k8s.io/client-go v0.25.0
|
k8s.io/client-go v0.22.0
|
||||||
k8s.io/code-generator v0.25.0
|
k8s.io/code-generator v0.22.0
|
||||||
k8s.io/component-base v0.25.0
|
k8s.io/component-base v0.22.0
|
||||||
k8s.io/component-helpers v0.25.0
|
k8s.io/component-helpers v0.22.0
|
||||||
k8s.io/klog/v2 v2.70.1
|
k8s.io/klog/v2 v2.9.0
|
||||||
k8s.io/utils v0.0.0-20220823124924-e9cbc92d1a73
|
k8s.io/kubectl v0.20.5
|
||||||
sigs.k8s.io/mdtoc v1.0.1
|
sigs.k8s.io/mdtoc v1.0.1
|
||||||
)
|
)
|
||||||
|
|
||||||
require (
|
|
||||||
cloud.google.com/go v0.97.0 // indirect
|
|
||||||
github.com/Azure/go-autorest v14.2.0+incompatible // indirect
|
|
||||||
github.com/Azure/go-autorest/autorest v0.11.27 // indirect
|
|
||||||
github.com/Azure/go-autorest/autorest/adal v0.9.20 // indirect
|
|
||||||
github.com/Azure/go-autorest/autorest/date v0.3.0 // indirect
|
|
||||||
github.com/Azure/go-autorest/logger v0.2.1 // indirect
|
|
||||||
github.com/Azure/go-autorest/tracing v0.6.0 // indirect
|
|
||||||
github.com/BurntSushi/toml v0.3.1 // indirect
|
|
||||||
github.com/NYTimes/gziphandler v1.1.1 // indirect
|
|
||||||
github.com/PuerkitoBio/purell v1.1.1 // indirect
|
|
||||||
github.com/PuerkitoBio/urlesc v0.0.0-20170810143723-de5bf2ad4578 // indirect
|
|
||||||
github.com/beorn7/perks v1.0.1 // indirect
|
|
||||||
github.com/blang/semver/v4 v4.0.0 // indirect
|
|
||||||
github.com/cespare/xxhash/v2 v2.1.2 // indirect
|
|
||||||
github.com/coreos/go-semver v0.3.0 // indirect
|
|
||||||
github.com/coreos/go-systemd/v22 v22.3.2 // indirect
|
|
||||||
github.com/davecgh/go-spew v1.1.1 // indirect
|
|
||||||
github.com/emicklei/go-restful/v3 v3.8.0 // indirect
|
|
||||||
github.com/evanphx/json-patch v4.12.0+incompatible // indirect
|
|
||||||
github.com/felixge/httpsnoop v1.0.1 // indirect
|
|
||||||
github.com/fsnotify/fsnotify v1.4.9 // indirect
|
|
||||||
github.com/go-logr/logr v1.2.3 // indirect
|
|
||||||
github.com/go-logr/zapr v1.2.3 // indirect
|
|
||||||
github.com/go-openapi/jsonpointer v0.19.5 // indirect
|
|
||||||
github.com/go-openapi/jsonreference v0.19.5 // indirect
|
|
||||||
github.com/go-openapi/swag v0.19.14 // indirect
|
|
||||||
github.com/gogo/protobuf v1.3.2 // indirect
|
|
||||||
github.com/golang-jwt/jwt/v4 v4.2.0 // indirect
|
|
||||||
github.com/golang/groupcache v0.0.0-20210331224755-41bb18bfe9da // indirect
|
|
||||||
github.com/golang/protobuf v1.5.2 // indirect
|
|
||||||
github.com/gomarkdown/markdown v0.0.0-20200824053859-8c8b3816f167 // indirect
|
|
||||||
github.com/google/gnostic v0.5.7-v3refs // indirect
|
|
||||||
github.com/google/go-cmp v0.5.6 // indirect
|
|
||||||
github.com/google/gofuzz v1.1.0 // indirect
|
|
||||||
github.com/google/uuid v1.1.2 // indirect
|
|
||||||
github.com/grpc-ecosystem/go-grpc-prometheus v1.2.0 // indirect
|
|
||||||
github.com/grpc-ecosystem/grpc-gateway v1.16.0 // indirect
|
|
||||||
github.com/imdario/mergo v0.3.6 // indirect
|
|
||||||
github.com/inconshreveable/mousetrap v1.0.0 // indirect
|
|
||||||
github.com/josharian/intern v1.0.0 // indirect
|
|
||||||
github.com/json-iterator/go v1.1.12 // indirect
|
|
||||||
github.com/mailru/easyjson v0.7.6 // indirect
|
|
||||||
github.com/matttproud/golang_protobuf_extensions v1.0.2-0.20181231171920-c182affec369 // indirect
|
|
||||||
github.com/mmarkdown/mmark v2.0.40+incompatible // indirect
|
|
||||||
github.com/modern-go/concurrent v0.0.0-20180306012644-bacd9c7ef1dd // indirect
|
|
||||||
github.com/modern-go/reflect2 v1.0.2 // indirect
|
|
||||||
github.com/munnerz/goautoneg v0.0.0-20191010083416-a7dc8b61c822 // indirect
|
|
||||||
github.com/pkg/errors v0.9.1 // indirect
|
|
||||||
github.com/prometheus/client_golang v1.12.1 // indirect
|
|
||||||
github.com/prometheus/client_model v0.2.0 // indirect
|
|
||||||
github.com/prometheus/common v0.32.1 // indirect
|
|
||||||
github.com/prometheus/procfs v0.7.3 // indirect
|
|
||||||
go.etcd.io/etcd/api/v3 v3.5.4 // indirect
|
|
||||||
go.etcd.io/etcd/client/pkg/v3 v3.5.4 // indirect
|
|
||||||
go.etcd.io/etcd/client/v3 v3.5.4 // indirect
|
|
||||||
go.opentelemetry.io/contrib v0.20.0 // indirect
|
|
||||||
go.opentelemetry.io/contrib/instrumentation/google.golang.org/grpc/otelgrpc v0.20.0 // indirect
|
|
||||||
go.opentelemetry.io/contrib/instrumentation/net/http/otelhttp v0.20.0 // indirect
|
|
||||||
go.opentelemetry.io/otel v0.20.0 // indirect
|
|
||||||
go.opentelemetry.io/otel/exporters/otlp v0.20.0 // indirect
|
|
||||||
go.opentelemetry.io/otel/metric v0.20.0 // indirect
|
|
||||||
go.opentelemetry.io/otel/sdk v0.20.0 // indirect
|
|
||||||
go.opentelemetry.io/otel/sdk/export/metric v0.20.0 // indirect
|
|
||||||
go.opentelemetry.io/otel/sdk/metric v0.20.0 // indirect
|
|
||||||
go.opentelemetry.io/otel/trace v0.20.0 // indirect
|
|
||||||
go.opentelemetry.io/proto/otlp v0.7.0 // indirect
|
|
||||||
go.uber.org/atomic v1.7.0 // indirect
|
|
||||||
go.uber.org/multierr v1.6.0 // indirect
|
|
||||||
go.uber.org/zap v1.19.0 // indirect
|
|
||||||
golang.org/x/crypto v0.0.0-20220518034528-6f7dac969898 // indirect
|
|
||||||
golang.org/x/mod v0.6.0-dev.0.20220419223038-86c51ed26bb4 // indirect
|
|
||||||
golang.org/x/net v0.0.0-20220722155237-a158d28d115b // indirect
|
|
||||||
golang.org/x/oauth2 v0.0.0-20211104180415-d3ed0bb246c8 // indirect
|
|
||||||
golang.org/x/sync v0.0.0-20220722155255-886fb9371eb4 // indirect
|
|
||||||
golang.org/x/sys v0.0.0-20220722155257-8c9f86f7a55f // indirect
|
|
||||||
golang.org/x/term v0.0.0-20210927222741-03fcf44c2211 // indirect
|
|
||||||
golang.org/x/text v0.3.7 // indirect
|
|
||||||
golang.org/x/time v0.0.0-20220210224613-90d013bbcef8 // indirect
|
|
||||||
golang.org/x/tools v0.1.12 // indirect
|
|
||||||
google.golang.org/appengine v1.6.7 // indirect
|
|
||||||
google.golang.org/genproto v0.0.0-20220502173005-c8bf987b8c21 // indirect
|
|
||||||
google.golang.org/grpc v1.47.0 // indirect
|
|
||||||
google.golang.org/protobuf v1.28.0 // indirect
|
|
||||||
gopkg.in/inf.v0 v0.9.1 // indirect
|
|
||||||
gopkg.in/natefinch/lumberjack.v2 v2.0.0 // indirect
|
|
||||||
gopkg.in/yaml.v2 v2.4.0 // indirect
|
|
||||||
gopkg.in/yaml.v3 v3.0.1 // indirect
|
|
||||||
k8s.io/gengo v0.0.0-20211129171323-c02415ce4185 // indirect
|
|
||||||
k8s.io/kube-openapi v0.0.0-20220803162953-67bda5d908f1 // indirect
|
|
||||||
sigs.k8s.io/apiserver-network-proxy/konnectivity-client v0.0.32 // indirect
|
|
||||||
sigs.k8s.io/json v0.0.0-20220713155537-f223a00ba0e2 // indirect
|
|
||||||
sigs.k8s.io/structured-merge-diff/v4 v4.2.3 // indirect
|
|
||||||
sigs.k8s.io/yaml v1.2.0 // indirect
|
|
||||||
)
|
|
||||||
|
|||||||
@@ -1,4 +1,3 @@
|
|||||||
//go:build tools
|
|
||||||
// +build tools
|
// +build tools
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
|||||||
@@ -5,6 +5,6 @@ go build -o "${OS_OUTPUT_BINPATH}/deepcopy-gen" "k8s.io/code-generator/cmd/deepc
|
|||||||
|
|
||||||
${OS_OUTPUT_BINPATH}/deepcopy-gen \
|
${OS_OUTPUT_BINPATH}/deepcopy-gen \
|
||||||
--go-header-file "hack/boilerplate/boilerplate.go.txt" \
|
--go-header-file "hack/boilerplate/boilerplate.go.txt" \
|
||||||
--input-dirs "${PRJ_PREFIX}/pkg/apis/componentconfig,${PRJ_PREFIX}/pkg/apis/componentconfig/v1alpha1,${PRJ_PREFIX}/pkg/api,${PRJ_PREFIX}/pkg/api/v1alpha1,${PRJ_PREFIX}/pkg/framework/plugins/defaultevictor/" \
|
--input-dirs "${PRJ_PREFIX}/pkg/apis/componentconfig,${PRJ_PREFIX}/pkg/apis/componentconfig/v1alpha1,${PRJ_PREFIX}/pkg/api,${PRJ_PREFIX}/pkg/api/v1alpha1" \
|
||||||
--output-file-base zz_generated.deepcopy
|
--output-file-base zz_generated.deepcopy
|
||||||
|
|
||||||
|
|||||||
@@ -23,7 +23,7 @@ DESCHEDULER_ROOT=$(dirname "${BASH_SOURCE}")/..
|
|||||||
|
|
||||||
GO_VERSION=($(go version))
|
GO_VERSION=($(go version))
|
||||||
|
|
||||||
if [[ -z $(echo "${GO_VERSION[2]}" | grep -E 'go1.17|go1.18|go1.19') ]]; then
|
if [[ -z $(echo "${GO_VERSION[2]}" | grep -E 'go1.14|go1.15|go1.16') ]]; then
|
||||||
echo "Unknown go version '${GO_VERSION[2]}', skipping gofmt."
|
echo "Unknown go version '${GO_VERSION[2]}', skipping gofmt."
|
||||||
exit 1
|
exit 1
|
||||||
fi
|
fi
|
||||||
|
|||||||
@@ -1 +0,0 @@
|
|||||||
${CONTAINER_ENGINE:-docker} run -it --rm --network host --workdir=/data --volume ~/.kube/config:/root/.kube/config:ro --volume $(pwd):/data quay.io/helmpack/chart-testing:v3.7.0 /bin/bash -c "git config --global --add safe.directory /data; ct install --config=.github/ci/ct.yaml --helm-extra-set-args=\"--set=kind=Deployment\""
|
|
||||||
@@ -20,7 +20,7 @@ go build -o "${OS_OUTPUT_BINPATH}/deepcopy-gen" "k8s.io/code-generator/cmd/deepc
|
|||||||
|
|
||||||
${OS_OUTPUT_BINPATH}/deepcopy-gen \
|
${OS_OUTPUT_BINPATH}/deepcopy-gen \
|
||||||
--go-header-file "hack/boilerplate/boilerplate.go.txt" \
|
--go-header-file "hack/boilerplate/boilerplate.go.txt" \
|
||||||
--input-dirs "./pkg/apis/componentconfig,./pkg/apis/componentconfig/v1alpha1,./pkg/api,./pkg/api/v1alpha1,./pkg/framework/plugins/defaultevictor/" \
|
--input-dirs "./pkg/apis/componentconfig,./pkg/apis/componentconfig/v1alpha1,./pkg/api,./pkg/api/v1alpha1" \
|
||||||
--output-file-base zz_generated.deepcopy
|
--output-file-base zz_generated.deepcopy
|
||||||
popd > /dev/null 2>&1
|
popd > /dev/null 2>&1
|
||||||
|
|
||||||
|
|||||||
@@ -23,7 +23,7 @@ DESCHEDULER_ROOT=$(dirname "${BASH_SOURCE}")/..
|
|||||||
|
|
||||||
GO_VERSION=($(go version))
|
GO_VERSION=($(go version))
|
||||||
|
|
||||||
if [[ -z $(echo "${GO_VERSION[2]}" | grep -E 'go1.17|go1.18|go1.19') ]]; then
|
if [[ -z $(echo "${GO_VERSION[2]}" | grep -E 'go1.14|go1.15|go1.16') ]]; then
|
||||||
echo "Unknown go version '${GO_VERSION[2]}', skipping gofmt."
|
echo "Unknown go version '${GO_VERSION[2]}', skipping gofmt."
|
||||||
exit 1
|
exit 1
|
||||||
fi
|
fi
|
||||||
|
|||||||
@@ -4,7 +4,7 @@ apiVersion: rbac.authorization.k8s.io/v1
|
|||||||
metadata:
|
metadata:
|
||||||
name: descheduler-cluster-role
|
name: descheduler-cluster-role
|
||||||
rules:
|
rules:
|
||||||
- apiGroups: ["events.k8s.io"]
|
- apiGroups: [""]
|
||||||
resources: ["events"]
|
resources: ["events"]
|
||||||
verbs: ["create", "update"]
|
verbs: ["create", "update"]
|
||||||
- apiGroups: [""]
|
- apiGroups: [""]
|
||||||
@@ -12,7 +12,7 @@ rules:
|
|||||||
verbs: ["get", "watch", "list"]
|
verbs: ["get", "watch", "list"]
|
||||||
- apiGroups: [""]
|
- apiGroups: [""]
|
||||||
resources: ["namespaces"]
|
resources: ["namespaces"]
|
||||||
verbs: ["get", "watch", "list"]
|
verbs: ["get", "list"]
|
||||||
- apiGroups: [""]
|
- apiGroups: [""]
|
||||||
resources: ["pods"]
|
resources: ["pods"]
|
||||||
verbs: ["get", "watch", "list", "delete"]
|
verbs: ["get", "watch", "list", "delete"]
|
||||||
@@ -22,13 +22,6 @@ rules:
|
|||||||
- apiGroups: ["scheduling.k8s.io"]
|
- apiGroups: ["scheduling.k8s.io"]
|
||||||
resources: ["priorityclasses"]
|
resources: ["priorityclasses"]
|
||||||
verbs: ["get", "watch", "list"]
|
verbs: ["get", "watch", "list"]
|
||||||
- apiGroups: ["coordination.k8s.io"]
|
|
||||||
resources: ["leases"]
|
|
||||||
verbs: ["create"]
|
|
||||||
- apiGroups: ["coordination.k8s.io"]
|
|
||||||
resources: ["leases"]
|
|
||||||
resourceNames: ["descheduler"]
|
|
||||||
verbs: ["get", "patch", "delete"]
|
|
||||||
---
|
---
|
||||||
apiVersion: v1
|
apiVersion: v1
|
||||||
kind: ServiceAccount
|
kind: ServiceAccount
|
||||||
@@ -48,3 +41,4 @@ subjects:
|
|||||||
- name: descheduler-sa
|
- name: descheduler-sa
|
||||||
kind: ServiceAccount
|
kind: ServiceAccount
|
||||||
namespace: kube-system
|
namespace: kube-system
|
||||||
|
|
||||||
|
|||||||
@@ -1,5 +1,5 @@
|
|||||||
---
|
---
|
||||||
apiVersion: batch/v1
|
apiVersion: batch/v1 # for k8s version < 1.21.0, use batch/v1beta1
|
||||||
kind: CronJob
|
kind: CronJob
|
||||||
metadata:
|
metadata:
|
||||||
name: descheduler-cronjob
|
name: descheduler-cronjob
|
||||||
@@ -16,7 +16,7 @@ spec:
|
|||||||
priorityClassName: system-cluster-critical
|
priorityClassName: system-cluster-critical
|
||||||
containers:
|
containers:
|
||||||
- name: descheduler
|
- name: descheduler
|
||||||
image: k8s.gcr.io/descheduler/descheduler:v0.25.1
|
image: k8s.gcr.io/descheduler/descheduler:v0.22.0
|
||||||
volumeMounts:
|
volumeMounts:
|
||||||
- mountPath: /policy-dir
|
- mountPath: /policy-dir
|
||||||
name: policy-volume
|
name: policy-volume
|
||||||
@@ -31,14 +31,6 @@ spec:
|
|||||||
requests:
|
requests:
|
||||||
cpu: "500m"
|
cpu: "500m"
|
||||||
memory: "256Mi"
|
memory: "256Mi"
|
||||||
livenessProbe:
|
|
||||||
failureThreshold: 3
|
|
||||||
httpGet:
|
|
||||||
path: /healthz
|
|
||||||
port: 10258
|
|
||||||
scheme: HTTPS
|
|
||||||
initialDelaySeconds: 3
|
|
||||||
periodSeconds: 10
|
|
||||||
securityContext:
|
securityContext:
|
||||||
allowPrivilegeEscalation: false
|
allowPrivilegeEscalation: false
|
||||||
capabilities:
|
capabilities:
|
||||||
|
|||||||
@@ -19,7 +19,7 @@ spec:
|
|||||||
serviceAccountName: descheduler-sa
|
serviceAccountName: descheduler-sa
|
||||||
containers:
|
containers:
|
||||||
- name: descheduler
|
- name: descheduler
|
||||||
image: k8s.gcr.io/descheduler/descheduler:v0.25.1
|
image: k8s.gcr.io/descheduler/descheduler:v0.22.0
|
||||||
imagePullPolicy: IfNotPresent
|
imagePullPolicy: IfNotPresent
|
||||||
command:
|
command:
|
||||||
- "/bin/descheduler"
|
- "/bin/descheduler"
|
||||||
@@ -33,14 +33,6 @@ spec:
|
|||||||
ports:
|
ports:
|
||||||
- containerPort: 10258
|
- containerPort: 10258
|
||||||
protocol: TCP
|
protocol: TCP
|
||||||
livenessProbe:
|
|
||||||
failureThreshold: 3
|
|
||||||
httpGet:
|
|
||||||
path: /healthz
|
|
||||||
port: 10258
|
|
||||||
scheme: HTTPS
|
|
||||||
initialDelaySeconds: 3
|
|
||||||
periodSeconds: 10
|
|
||||||
resources:
|
resources:
|
||||||
requests:
|
requests:
|
||||||
cpu: 500m
|
cpu: 500m
|
||||||
|
|||||||
@@ -14,7 +14,7 @@ spec:
|
|||||||
priorityClassName: system-cluster-critical
|
priorityClassName: system-cluster-critical
|
||||||
containers:
|
containers:
|
||||||
- name: descheduler
|
- name: descheduler
|
||||||
image: k8s.gcr.io/descheduler/descheduler:v0.25.1
|
image: k8s.gcr.io/descheduler/descheduler:v0.22.0
|
||||||
volumeMounts:
|
volumeMounts:
|
||||||
- mountPath: /policy-dir
|
- mountPath: /policy-dir
|
||||||
name: policy-volume
|
name: policy-volume
|
||||||
@@ -29,14 +29,6 @@ spec:
|
|||||||
requests:
|
requests:
|
||||||
cpu: "500m"
|
cpu: "500m"
|
||||||
memory: "256Mi"
|
memory: "256Mi"
|
||||||
livenessProbe:
|
|
||||||
failureThreshold: 3
|
|
||||||
httpGet:
|
|
||||||
path: /healthz
|
|
||||||
port: 10258
|
|
||||||
scheme: HTTPS
|
|
||||||
initialDelaySeconds: 3
|
|
||||||
periodSeconds: 10
|
|
||||||
securityContext:
|
securityContext:
|
||||||
allowPrivilegeEscalation: false
|
allowPrivilegeEscalation: false
|
||||||
capabilities:
|
capabilities:
|
||||||
|
|||||||
@@ -34,16 +34,16 @@ var (
|
|||||||
&metrics.CounterOpts{
|
&metrics.CounterOpts{
|
||||||
Subsystem: DeschedulerSubsystem,
|
Subsystem: DeschedulerSubsystem,
|
||||||
Name: "pods_evicted",
|
Name: "pods_evicted",
|
||||||
Help: "Number of evicted pods, by the result, by the strategy, by the namespace, by the node name. 'error' result means a pod could not be evicted",
|
Help: "Number of evicted pods, by the result, by the strategy, by the namespace. 'failed' result means a pod could not be evicted",
|
||||||
StabilityLevel: metrics.ALPHA,
|
StabilityLevel: metrics.ALPHA,
|
||||||
}, []string{"result", "strategy", "namespace", "node"})
|
}, []string{"result", "strategy", "namespace"})
|
||||||
|
|
||||||
buildInfo = metrics.NewGauge(
|
buildInfo = metrics.NewGauge(
|
||||||
&metrics.GaugeOpts{
|
&metrics.GaugeOpts{
|
||||||
Subsystem: DeschedulerSubsystem,
|
Subsystem: DeschedulerSubsystem,
|
||||||
Name: "build_info",
|
Name: "build_info",
|
||||||
Help: "Build info about descheduler, including Go version, Descheduler version, Git SHA, Git branch",
|
Help: "Build info about descheduler, including Go version, Descheduler version, Git SHA, Git branch",
|
||||||
ConstLabels: map[string]string{"GoVersion": version.Get().GoVersion, "AppVersion": version.Get().Major + "." + version.Get().Minor, "DeschedulerVersion": version.Get().GitVersion, "GitBranch": version.Get().GitBranch, "GitSha1": version.Get().GitSha1},
|
ConstLabels: map[string]string{"GoVersion": version.Get().GoVersion, "DeschedulerVersion": version.Get().GitVersion, "GitBranch": version.Get().GitBranch, "GitSha1": version.Get().GitSha1},
|
||||||
StabilityLevel: metrics.ALPHA,
|
StabilityLevel: metrics.ALPHA,
|
||||||
},
|
},
|
||||||
)
|
)
|
||||||
|
|||||||
@@ -32,9 +32,6 @@ type DeschedulerPolicy struct {
|
|||||||
// NodeSelector for a set of nodes to operate over
|
// NodeSelector for a set of nodes to operate over
|
||||||
NodeSelector *string
|
NodeSelector *string
|
||||||
|
|
||||||
// EvictFailedBarePods allows pods without ownerReferences and in failed phase to be evicted.
|
|
||||||
EvictFailedBarePods *bool
|
|
||||||
|
|
||||||
// EvictLocalStoragePods allows pods using local storage to be evicted.
|
// EvictLocalStoragePods allows pods using local storage to be evicted.
|
||||||
EvictLocalStoragePods *bool
|
EvictLocalStoragePods *bool
|
||||||
|
|
||||||
@@ -45,10 +42,7 @@ type DeschedulerPolicy struct {
|
|||||||
IgnorePVCPods *bool
|
IgnorePVCPods *bool
|
||||||
|
|
||||||
// MaxNoOfPodsToEvictPerNode restricts maximum of pods to be evicted per node.
|
// MaxNoOfPodsToEvictPerNode restricts maximum of pods to be evicted per node.
|
||||||
MaxNoOfPodsToEvictPerNode *uint
|
MaxNoOfPodsToEvictPerNode *int
|
||||||
|
|
||||||
// MaxNoOfPodsToEvictPerNamespace restricts maximum of pods to be evicted per namespace.
|
|
||||||
MaxNoOfPodsToEvictPerNamespace *uint
|
|
||||||
}
|
}
|
||||||
|
|
||||||
type StrategyName string
|
type StrategyName string
|
||||||
@@ -74,8 +68,7 @@ type Namespaces struct {
|
|||||||
|
|
||||||
// Besides Namespaces only one of its members may be specified
|
// Besides Namespaces only one of its members may be specified
|
||||||
// TODO(jchaloup): move Namespaces ThresholdPriority and ThresholdPriorityClassName to individual strategies
|
// TODO(jchaloup): move Namespaces ThresholdPriority and ThresholdPriorityClassName to individual strategies
|
||||||
//
|
// once the policy version is bumped to v1alpha2
|
||||||
// once the policy version is bumped to v1alpha2
|
|
||||||
type StrategyParameters struct {
|
type StrategyParameters struct {
|
||||||
NodeResourceUtilizationThresholds *NodeResourceUtilizationThresholds
|
NodeResourceUtilizationThresholds *NodeResourceUtilizationThresholds
|
||||||
NodeAffinityType []string
|
NodeAffinityType []string
|
||||||
@@ -89,18 +82,15 @@ type StrategyParameters struct {
|
|||||||
ThresholdPriorityClassName string
|
ThresholdPriorityClassName string
|
||||||
LabelSelector *metav1.LabelSelector
|
LabelSelector *metav1.LabelSelector
|
||||||
NodeFit bool
|
NodeFit bool
|
||||||
IncludePreferNoSchedule bool
|
|
||||||
ExcludedTaints []string
|
|
||||||
}
|
}
|
||||||
|
|
||||||
type Percentage float64
|
type Percentage float64
|
||||||
type ResourceThresholds map[v1.ResourceName]Percentage
|
type ResourceThresholds map[v1.ResourceName]Percentage
|
||||||
|
|
||||||
type NodeResourceUtilizationThresholds struct {
|
type NodeResourceUtilizationThresholds struct {
|
||||||
UseDeviationThresholds bool
|
Thresholds ResourceThresholds
|
||||||
Thresholds ResourceThresholds
|
TargetThresholds ResourceThresholds
|
||||||
TargetThresholds ResourceThresholds
|
NumberOfNodes int
|
||||||
NumberOfNodes int
|
|
||||||
}
|
}
|
||||||
|
|
||||||
type PodsHavingTooManyRestarts struct {
|
type PodsHavingTooManyRestarts struct {
|
||||||
@@ -114,10 +104,7 @@ type RemoveDuplicates struct {
|
|||||||
|
|
||||||
type PodLifeTime struct {
|
type PodLifeTime struct {
|
||||||
MaxPodLifeTimeSeconds *uint
|
MaxPodLifeTimeSeconds *uint
|
||||||
States []string
|
PodStatusPhases []string
|
||||||
|
|
||||||
// Deprecated: Use States instead.
|
|
||||||
PodStatusPhases []string
|
|
||||||
}
|
}
|
||||||
|
|
||||||
type FailedPods struct {
|
type FailedPods struct {
|
||||||
@@ -126,8 +113,3 @@ type FailedPods struct {
|
|||||||
Reasons []string
|
Reasons []string
|
||||||
IncludingInitContainers bool
|
IncludingInitContainers bool
|
||||||
}
|
}
|
||||||
|
|
||||||
type PriorityThreshold struct {
|
|
||||||
Value *int32
|
|
||||||
Name string
|
|
||||||
}
|
|
||||||
|
|||||||
@@ -32,9 +32,6 @@ type DeschedulerPolicy struct {
|
|||||||
// NodeSelector for a set of nodes to operate over
|
// NodeSelector for a set of nodes to operate over
|
||||||
NodeSelector *string `json:"nodeSelector,omitempty"`
|
NodeSelector *string `json:"nodeSelector,omitempty"`
|
||||||
|
|
||||||
// EvictFailedBarePods allows pods without ownerReferences and in failed phase to be evicted.
|
|
||||||
EvictFailedBarePods *bool `json:"evictFailedBarePods,omitempty"`
|
|
||||||
|
|
||||||
// EvictLocalStoragePods allows pods using local storage to be evicted.
|
// EvictLocalStoragePods allows pods using local storage to be evicted.
|
||||||
EvictLocalStoragePods *bool `json:"evictLocalStoragePods,omitempty"`
|
EvictLocalStoragePods *bool `json:"evictLocalStoragePods,omitempty"`
|
||||||
|
|
||||||
@@ -46,9 +43,6 @@ type DeschedulerPolicy struct {
|
|||||||
|
|
||||||
// MaxNoOfPodsToEvictPerNode restricts maximum of pods to be evicted per node.
|
// MaxNoOfPodsToEvictPerNode restricts maximum of pods to be evicted per node.
|
||||||
MaxNoOfPodsToEvictPerNode *int `json:"maxNoOfPodsToEvictPerNode,omitempty"`
|
MaxNoOfPodsToEvictPerNode *int `json:"maxNoOfPodsToEvictPerNode,omitempty"`
|
||||||
|
|
||||||
// MaxNoOfPodsToEvictPerNamespace restricts maximum of pods to be evicted per namespace.
|
|
||||||
MaxNoOfPodsToEvictPerNamespace *int `json:"maxNoOfPodsToEvictPerNamespace,omitempty"`
|
|
||||||
}
|
}
|
||||||
|
|
||||||
type StrategyName string
|
type StrategyName string
|
||||||
@@ -86,18 +80,15 @@ type StrategyParameters struct {
|
|||||||
ThresholdPriorityClassName string `json:"thresholdPriorityClassName"`
|
ThresholdPriorityClassName string `json:"thresholdPriorityClassName"`
|
||||||
LabelSelector *metav1.LabelSelector `json:"labelSelector"`
|
LabelSelector *metav1.LabelSelector `json:"labelSelector"`
|
||||||
NodeFit bool `json:"nodeFit"`
|
NodeFit bool `json:"nodeFit"`
|
||||||
IncludePreferNoSchedule bool `json:"includePreferNoSchedule"`
|
|
||||||
ExcludedTaints []string `json:"excludedTaints,omitempty"`
|
|
||||||
}
|
}
|
||||||
|
|
||||||
type Percentage float64
|
type Percentage float64
|
||||||
type ResourceThresholds map[v1.ResourceName]Percentage
|
type ResourceThresholds map[v1.ResourceName]Percentage
|
||||||
|
|
||||||
type NodeResourceUtilizationThresholds struct {
|
type NodeResourceUtilizationThresholds struct {
|
||||||
UseDeviationThresholds bool `json:"useDeviationThresholds,omitempty"`
|
Thresholds ResourceThresholds `json:"thresholds,omitempty"`
|
||||||
Thresholds ResourceThresholds `json:"thresholds,omitempty"`
|
TargetThresholds ResourceThresholds `json:"targetThresholds,omitempty"`
|
||||||
TargetThresholds ResourceThresholds `json:"targetThresholds,omitempty"`
|
NumberOfNodes int `json:"numberOfNodes,omitempty"`
|
||||||
NumberOfNodes int `json:"numberOfNodes,omitempty"`
|
|
||||||
}
|
}
|
||||||
|
|
||||||
type PodsHavingTooManyRestarts struct {
|
type PodsHavingTooManyRestarts struct {
|
||||||
@@ -111,10 +102,7 @@ type RemoveDuplicates struct {
|
|||||||
|
|
||||||
type PodLifeTime struct {
|
type PodLifeTime struct {
|
||||||
MaxPodLifeTimeSeconds *uint `json:"maxPodLifeTimeSeconds,omitempty"`
|
MaxPodLifeTimeSeconds *uint `json:"maxPodLifeTimeSeconds,omitempty"`
|
||||||
States []string `json:"states,omitempty"`
|
PodStatusPhases []string `json:"podStatusPhases,omitempty"`
|
||||||
|
|
||||||
// Deprecated: Use States instead.
|
|
||||||
PodStatusPhases []string `json:"podStatusPhases,omitempty"`
|
|
||||||
}
|
}
|
||||||
|
|
||||||
type FailedPods struct {
|
type FailedPods struct {
|
||||||
|
|||||||
@@ -1,8 +1,7 @@
|
|||||||
//go:build !ignore_autogenerated
|
|
||||||
// +build !ignore_autogenerated
|
// +build !ignore_autogenerated
|
||||||
|
|
||||||
/*
|
/*
|
||||||
Copyright 2022 The Kubernetes Authors.
|
Copyright 2021 The Kubernetes Authors.
|
||||||
|
|
||||||
Licensed under the Apache License, Version 2.0 (the "License");
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
you may not use this file except in compliance with the License.
|
you may not use this file except in compliance with the License.
|
||||||
@@ -133,24 +132,10 @@ func RegisterConversions(s *runtime.Scheme) error {
|
|||||||
func autoConvert_v1alpha1_DeschedulerPolicy_To_api_DeschedulerPolicy(in *DeschedulerPolicy, out *api.DeschedulerPolicy, s conversion.Scope) error {
|
func autoConvert_v1alpha1_DeschedulerPolicy_To_api_DeschedulerPolicy(in *DeschedulerPolicy, out *api.DeschedulerPolicy, s conversion.Scope) error {
|
||||||
out.Strategies = *(*api.StrategyList)(unsafe.Pointer(&in.Strategies))
|
out.Strategies = *(*api.StrategyList)(unsafe.Pointer(&in.Strategies))
|
||||||
out.NodeSelector = (*string)(unsafe.Pointer(in.NodeSelector))
|
out.NodeSelector = (*string)(unsafe.Pointer(in.NodeSelector))
|
||||||
out.EvictFailedBarePods = (*bool)(unsafe.Pointer(in.EvictFailedBarePods))
|
|
||||||
out.EvictLocalStoragePods = (*bool)(unsafe.Pointer(in.EvictLocalStoragePods))
|
out.EvictLocalStoragePods = (*bool)(unsafe.Pointer(in.EvictLocalStoragePods))
|
||||||
out.EvictSystemCriticalPods = (*bool)(unsafe.Pointer(in.EvictSystemCriticalPods))
|
out.EvictSystemCriticalPods = (*bool)(unsafe.Pointer(in.EvictSystemCriticalPods))
|
||||||
out.IgnorePVCPods = (*bool)(unsafe.Pointer(in.IgnorePVCPods))
|
out.IgnorePVCPods = (*bool)(unsafe.Pointer(in.IgnorePVCPods))
|
||||||
if in.MaxNoOfPodsToEvictPerNode != nil {
|
out.MaxNoOfPodsToEvictPerNode = (*int)(unsafe.Pointer(in.MaxNoOfPodsToEvictPerNode))
|
||||||
in, out := &in.MaxNoOfPodsToEvictPerNode, &out.MaxNoOfPodsToEvictPerNode
|
|
||||||
*out = new(uint)
|
|
||||||
**out = uint(**in)
|
|
||||||
} else {
|
|
||||||
out.MaxNoOfPodsToEvictPerNode = nil
|
|
||||||
}
|
|
||||||
if in.MaxNoOfPodsToEvictPerNamespace != nil {
|
|
||||||
in, out := &in.MaxNoOfPodsToEvictPerNamespace, &out.MaxNoOfPodsToEvictPerNamespace
|
|
||||||
*out = new(uint)
|
|
||||||
**out = uint(**in)
|
|
||||||
} else {
|
|
||||||
out.MaxNoOfPodsToEvictPerNamespace = nil
|
|
||||||
}
|
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -162,24 +147,10 @@ func Convert_v1alpha1_DeschedulerPolicy_To_api_DeschedulerPolicy(in *Descheduler
|
|||||||
func autoConvert_api_DeschedulerPolicy_To_v1alpha1_DeschedulerPolicy(in *api.DeschedulerPolicy, out *DeschedulerPolicy, s conversion.Scope) error {
|
func autoConvert_api_DeschedulerPolicy_To_v1alpha1_DeschedulerPolicy(in *api.DeschedulerPolicy, out *DeschedulerPolicy, s conversion.Scope) error {
|
||||||
out.Strategies = *(*StrategyList)(unsafe.Pointer(&in.Strategies))
|
out.Strategies = *(*StrategyList)(unsafe.Pointer(&in.Strategies))
|
||||||
out.NodeSelector = (*string)(unsafe.Pointer(in.NodeSelector))
|
out.NodeSelector = (*string)(unsafe.Pointer(in.NodeSelector))
|
||||||
out.EvictFailedBarePods = (*bool)(unsafe.Pointer(in.EvictFailedBarePods))
|
|
||||||
out.EvictLocalStoragePods = (*bool)(unsafe.Pointer(in.EvictLocalStoragePods))
|
out.EvictLocalStoragePods = (*bool)(unsafe.Pointer(in.EvictLocalStoragePods))
|
||||||
out.EvictSystemCriticalPods = (*bool)(unsafe.Pointer(in.EvictSystemCriticalPods))
|
out.EvictSystemCriticalPods = (*bool)(unsafe.Pointer(in.EvictSystemCriticalPods))
|
||||||
out.IgnorePVCPods = (*bool)(unsafe.Pointer(in.IgnorePVCPods))
|
out.IgnorePVCPods = (*bool)(unsafe.Pointer(in.IgnorePVCPods))
|
||||||
if in.MaxNoOfPodsToEvictPerNode != nil {
|
out.MaxNoOfPodsToEvictPerNode = (*int)(unsafe.Pointer(in.MaxNoOfPodsToEvictPerNode))
|
||||||
in, out := &in.MaxNoOfPodsToEvictPerNode, &out.MaxNoOfPodsToEvictPerNode
|
|
||||||
*out = new(int)
|
|
||||||
**out = int(**in)
|
|
||||||
} else {
|
|
||||||
out.MaxNoOfPodsToEvictPerNode = nil
|
|
||||||
}
|
|
||||||
if in.MaxNoOfPodsToEvictPerNamespace != nil {
|
|
||||||
in, out := &in.MaxNoOfPodsToEvictPerNamespace, &out.MaxNoOfPodsToEvictPerNamespace
|
|
||||||
*out = new(int)
|
|
||||||
**out = int(**in)
|
|
||||||
} else {
|
|
||||||
out.MaxNoOfPodsToEvictPerNamespace = nil
|
|
||||||
}
|
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -261,7 +232,6 @@ func Convert_api_Namespaces_To_v1alpha1_Namespaces(in *api.Namespaces, out *Name
|
|||||||
}
|
}
|
||||||
|
|
||||||
func autoConvert_v1alpha1_NodeResourceUtilizationThresholds_To_api_NodeResourceUtilizationThresholds(in *NodeResourceUtilizationThresholds, out *api.NodeResourceUtilizationThresholds, s conversion.Scope) error {
|
func autoConvert_v1alpha1_NodeResourceUtilizationThresholds_To_api_NodeResourceUtilizationThresholds(in *NodeResourceUtilizationThresholds, out *api.NodeResourceUtilizationThresholds, s conversion.Scope) error {
|
||||||
out.UseDeviationThresholds = in.UseDeviationThresholds
|
|
||||||
out.Thresholds = *(*api.ResourceThresholds)(unsafe.Pointer(&in.Thresholds))
|
out.Thresholds = *(*api.ResourceThresholds)(unsafe.Pointer(&in.Thresholds))
|
||||||
out.TargetThresholds = *(*api.ResourceThresholds)(unsafe.Pointer(&in.TargetThresholds))
|
out.TargetThresholds = *(*api.ResourceThresholds)(unsafe.Pointer(&in.TargetThresholds))
|
||||||
out.NumberOfNodes = in.NumberOfNodes
|
out.NumberOfNodes = in.NumberOfNodes
|
||||||
@@ -274,7 +244,6 @@ func Convert_v1alpha1_NodeResourceUtilizationThresholds_To_api_NodeResourceUtili
|
|||||||
}
|
}
|
||||||
|
|
||||||
func autoConvert_api_NodeResourceUtilizationThresholds_To_v1alpha1_NodeResourceUtilizationThresholds(in *api.NodeResourceUtilizationThresholds, out *NodeResourceUtilizationThresholds, s conversion.Scope) error {
|
func autoConvert_api_NodeResourceUtilizationThresholds_To_v1alpha1_NodeResourceUtilizationThresholds(in *api.NodeResourceUtilizationThresholds, out *NodeResourceUtilizationThresholds, s conversion.Scope) error {
|
||||||
out.UseDeviationThresholds = in.UseDeviationThresholds
|
|
||||||
out.Thresholds = *(*ResourceThresholds)(unsafe.Pointer(&in.Thresholds))
|
out.Thresholds = *(*ResourceThresholds)(unsafe.Pointer(&in.Thresholds))
|
||||||
out.TargetThresholds = *(*ResourceThresholds)(unsafe.Pointer(&in.TargetThresholds))
|
out.TargetThresholds = *(*ResourceThresholds)(unsafe.Pointer(&in.TargetThresholds))
|
||||||
out.NumberOfNodes = in.NumberOfNodes
|
out.NumberOfNodes = in.NumberOfNodes
|
||||||
@@ -288,7 +257,6 @@ func Convert_api_NodeResourceUtilizationThresholds_To_v1alpha1_NodeResourceUtili
|
|||||||
|
|
||||||
func autoConvert_v1alpha1_PodLifeTime_To_api_PodLifeTime(in *PodLifeTime, out *api.PodLifeTime, s conversion.Scope) error {
|
func autoConvert_v1alpha1_PodLifeTime_To_api_PodLifeTime(in *PodLifeTime, out *api.PodLifeTime, s conversion.Scope) error {
|
||||||
out.MaxPodLifeTimeSeconds = (*uint)(unsafe.Pointer(in.MaxPodLifeTimeSeconds))
|
out.MaxPodLifeTimeSeconds = (*uint)(unsafe.Pointer(in.MaxPodLifeTimeSeconds))
|
||||||
out.States = *(*[]string)(unsafe.Pointer(&in.States))
|
|
||||||
out.PodStatusPhases = *(*[]string)(unsafe.Pointer(&in.PodStatusPhases))
|
out.PodStatusPhases = *(*[]string)(unsafe.Pointer(&in.PodStatusPhases))
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
@@ -300,7 +268,6 @@ func Convert_v1alpha1_PodLifeTime_To_api_PodLifeTime(in *PodLifeTime, out *api.P
|
|||||||
|
|
||||||
func autoConvert_api_PodLifeTime_To_v1alpha1_PodLifeTime(in *api.PodLifeTime, out *PodLifeTime, s conversion.Scope) error {
|
func autoConvert_api_PodLifeTime_To_v1alpha1_PodLifeTime(in *api.PodLifeTime, out *PodLifeTime, s conversion.Scope) error {
|
||||||
out.MaxPodLifeTimeSeconds = (*uint)(unsafe.Pointer(in.MaxPodLifeTimeSeconds))
|
out.MaxPodLifeTimeSeconds = (*uint)(unsafe.Pointer(in.MaxPodLifeTimeSeconds))
|
||||||
out.States = *(*[]string)(unsafe.Pointer(&in.States))
|
|
||||||
out.PodStatusPhases = *(*[]string)(unsafe.Pointer(&in.PodStatusPhases))
|
out.PodStatusPhases = *(*[]string)(unsafe.Pointer(&in.PodStatusPhases))
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
@@ -365,8 +332,6 @@ func autoConvert_v1alpha1_StrategyParameters_To_api_StrategyParameters(in *Strat
|
|||||||
out.ThresholdPriorityClassName = in.ThresholdPriorityClassName
|
out.ThresholdPriorityClassName = in.ThresholdPriorityClassName
|
||||||
out.LabelSelector = (*v1.LabelSelector)(unsafe.Pointer(in.LabelSelector))
|
out.LabelSelector = (*v1.LabelSelector)(unsafe.Pointer(in.LabelSelector))
|
||||||
out.NodeFit = in.NodeFit
|
out.NodeFit = in.NodeFit
|
||||||
out.IncludePreferNoSchedule = in.IncludePreferNoSchedule
|
|
||||||
out.ExcludedTaints = *(*[]string)(unsafe.Pointer(&in.ExcludedTaints))
|
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -388,8 +353,6 @@ func autoConvert_api_StrategyParameters_To_v1alpha1_StrategyParameters(in *api.S
|
|||||||
out.ThresholdPriorityClassName = in.ThresholdPriorityClassName
|
out.ThresholdPriorityClassName = in.ThresholdPriorityClassName
|
||||||
out.LabelSelector = (*v1.LabelSelector)(unsafe.Pointer(in.LabelSelector))
|
out.LabelSelector = (*v1.LabelSelector)(unsafe.Pointer(in.LabelSelector))
|
||||||
out.NodeFit = in.NodeFit
|
out.NodeFit = in.NodeFit
|
||||||
out.IncludePreferNoSchedule = in.IncludePreferNoSchedule
|
|
||||||
out.ExcludedTaints = *(*[]string)(unsafe.Pointer(&in.ExcludedTaints))
|
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -1,8 +1,7 @@
|
|||||||
//go:build !ignore_autogenerated
|
|
||||||
// +build !ignore_autogenerated
|
// +build !ignore_autogenerated
|
||||||
|
|
||||||
/*
|
/*
|
||||||
Copyright 2022 The Kubernetes Authors.
|
Copyright 2021 The Kubernetes Authors.
|
||||||
|
|
||||||
Licensed under the Apache License, Version 2.0 (the "License");
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
you may not use this file except in compliance with the License.
|
you may not use this file except in compliance with the License.
|
||||||
@@ -42,11 +41,6 @@ func (in *DeschedulerPolicy) DeepCopyInto(out *DeschedulerPolicy) {
|
|||||||
*out = new(string)
|
*out = new(string)
|
||||||
**out = **in
|
**out = **in
|
||||||
}
|
}
|
||||||
if in.EvictFailedBarePods != nil {
|
|
||||||
in, out := &in.EvictFailedBarePods, &out.EvictFailedBarePods
|
|
||||||
*out = new(bool)
|
|
||||||
**out = **in
|
|
||||||
}
|
|
||||||
if in.EvictLocalStoragePods != nil {
|
if in.EvictLocalStoragePods != nil {
|
||||||
in, out := &in.EvictLocalStoragePods, &out.EvictLocalStoragePods
|
in, out := &in.EvictLocalStoragePods, &out.EvictLocalStoragePods
|
||||||
*out = new(bool)
|
*out = new(bool)
|
||||||
@@ -67,11 +61,6 @@ func (in *DeschedulerPolicy) DeepCopyInto(out *DeschedulerPolicy) {
|
|||||||
*out = new(int)
|
*out = new(int)
|
||||||
**out = **in
|
**out = **in
|
||||||
}
|
}
|
||||||
if in.MaxNoOfPodsToEvictPerNamespace != nil {
|
|
||||||
in, out := &in.MaxNoOfPodsToEvictPerNamespace, &out.MaxNoOfPodsToEvictPerNamespace
|
|
||||||
*out = new(int)
|
|
||||||
**out = **in
|
|
||||||
}
|
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -209,11 +198,6 @@ func (in *PodLifeTime) DeepCopyInto(out *PodLifeTime) {
|
|||||||
*out = new(uint)
|
*out = new(uint)
|
||||||
**out = **in
|
**out = **in
|
||||||
}
|
}
|
||||||
if in.States != nil {
|
|
||||||
in, out := &in.States, &out.States
|
|
||||||
*out = make([]string, len(*in))
|
|
||||||
copy(*out, *in)
|
|
||||||
}
|
|
||||||
if in.PodStatusPhases != nil {
|
if in.PodStatusPhases != nil {
|
||||||
in, out := &in.PodStatusPhases, &out.PodStatusPhases
|
in, out := &in.PodStatusPhases, &out.PodStatusPhases
|
||||||
*out = make([]string, len(*in))
|
*out = make([]string, len(*in))
|
||||||
@@ -361,11 +345,6 @@ func (in *StrategyParameters) DeepCopyInto(out *StrategyParameters) {
|
|||||||
*out = new(v1.LabelSelector)
|
*out = new(v1.LabelSelector)
|
||||||
(*in).DeepCopyInto(*out)
|
(*in).DeepCopyInto(*out)
|
||||||
}
|
}
|
||||||
if in.ExcludedTaints != nil {
|
|
||||||
in, out := &in.ExcludedTaints, &out.ExcludedTaints
|
|
||||||
*out = make([]string, len(*in))
|
|
||||||
copy(*out, *in)
|
|
||||||
}
|
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -1,8 +1,7 @@
|
|||||||
//go:build !ignore_autogenerated
|
|
||||||
// +build !ignore_autogenerated
|
// +build !ignore_autogenerated
|
||||||
|
|
||||||
/*
|
/*
|
||||||
Copyright 2022 The Kubernetes Authors.
|
Copyright 2021 The Kubernetes Authors.
|
||||||
|
|
||||||
Licensed under the Apache License, Version 2.0 (the "License");
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
you may not use this file except in compliance with the License.
|
you may not use this file except in compliance with the License.
|
||||||
|
|||||||
@@ -1,8 +1,7 @@
|
|||||||
//go:build !ignore_autogenerated
|
|
||||||
// +build !ignore_autogenerated
|
// +build !ignore_autogenerated
|
||||||
|
|
||||||
/*
|
/*
|
||||||
Copyright 2022 The Kubernetes Authors.
|
Copyright 2021 The Kubernetes Authors.
|
||||||
|
|
||||||
Licensed under the Apache License, Version 2.0 (the "License");
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
you may not use this file except in compliance with the License.
|
you may not use this file except in compliance with the License.
|
||||||
@@ -42,11 +41,6 @@ func (in *DeschedulerPolicy) DeepCopyInto(out *DeschedulerPolicy) {
|
|||||||
*out = new(string)
|
*out = new(string)
|
||||||
**out = **in
|
**out = **in
|
||||||
}
|
}
|
||||||
if in.EvictFailedBarePods != nil {
|
|
||||||
in, out := &in.EvictFailedBarePods, &out.EvictFailedBarePods
|
|
||||||
*out = new(bool)
|
|
||||||
**out = **in
|
|
||||||
}
|
|
||||||
if in.EvictLocalStoragePods != nil {
|
if in.EvictLocalStoragePods != nil {
|
||||||
in, out := &in.EvictLocalStoragePods, &out.EvictLocalStoragePods
|
in, out := &in.EvictLocalStoragePods, &out.EvictLocalStoragePods
|
||||||
*out = new(bool)
|
*out = new(bool)
|
||||||
@@ -64,12 +58,7 @@ func (in *DeschedulerPolicy) DeepCopyInto(out *DeschedulerPolicy) {
|
|||||||
}
|
}
|
||||||
if in.MaxNoOfPodsToEvictPerNode != nil {
|
if in.MaxNoOfPodsToEvictPerNode != nil {
|
||||||
in, out := &in.MaxNoOfPodsToEvictPerNode, &out.MaxNoOfPodsToEvictPerNode
|
in, out := &in.MaxNoOfPodsToEvictPerNode, &out.MaxNoOfPodsToEvictPerNode
|
||||||
*out = new(uint)
|
*out = new(int)
|
||||||
**out = **in
|
|
||||||
}
|
|
||||||
if in.MaxNoOfPodsToEvictPerNamespace != nil {
|
|
||||||
in, out := &in.MaxNoOfPodsToEvictPerNamespace, &out.MaxNoOfPodsToEvictPerNamespace
|
|
||||||
*out = new(uint)
|
|
||||||
**out = **in
|
**out = **in
|
||||||
}
|
}
|
||||||
return
|
return
|
||||||
@@ -209,11 +198,6 @@ func (in *PodLifeTime) DeepCopyInto(out *PodLifeTime) {
|
|||||||
*out = new(uint)
|
*out = new(uint)
|
||||||
**out = **in
|
**out = **in
|
||||||
}
|
}
|
||||||
if in.States != nil {
|
|
||||||
in, out := &in.States, &out.States
|
|
||||||
*out = make([]string, len(*in))
|
|
||||||
copy(*out, *in)
|
|
||||||
}
|
|
||||||
if in.PodStatusPhases != nil {
|
if in.PodStatusPhases != nil {
|
||||||
in, out := &in.PodStatusPhases, &out.PodStatusPhases
|
in, out := &in.PodStatusPhases, &out.PodStatusPhases
|
||||||
*out = make([]string, len(*in))
|
*out = make([]string, len(*in))
|
||||||
@@ -248,27 +232,6 @@ func (in *PodsHavingTooManyRestarts) DeepCopy() *PodsHavingTooManyRestarts {
|
|||||||
return out
|
return out
|
||||||
}
|
}
|
||||||
|
|
||||||
// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
|
|
||||||
func (in *PriorityThreshold) DeepCopyInto(out *PriorityThreshold) {
|
|
||||||
*out = *in
|
|
||||||
if in.Value != nil {
|
|
||||||
in, out := &in.Value, &out.Value
|
|
||||||
*out = new(int32)
|
|
||||||
**out = **in
|
|
||||||
}
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PriorityThreshold.
|
|
||||||
func (in *PriorityThreshold) DeepCopy() *PriorityThreshold {
|
|
||||||
if in == nil {
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
out := new(PriorityThreshold)
|
|
||||||
in.DeepCopyInto(out)
|
|
||||||
return out
|
|
||||||
}
|
|
||||||
|
|
||||||
// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
|
// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
|
||||||
func (in *RemoveDuplicates) DeepCopyInto(out *RemoveDuplicates) {
|
func (in *RemoveDuplicates) DeepCopyInto(out *RemoveDuplicates) {
|
||||||
*out = *in
|
*out = *in
|
||||||
@@ -382,11 +345,6 @@ func (in *StrategyParameters) DeepCopyInto(out *StrategyParameters) {
|
|||||||
*out = new(v1.LabelSelector)
|
*out = new(v1.LabelSelector)
|
||||||
(*in).DeepCopyInto(*out)
|
(*in).DeepCopyInto(*out)
|
||||||
}
|
}
|
||||||
if in.ExcludedTaints != nil {
|
|
||||||
in, out := &in.ExcludedTaints, &out.ExcludedTaints
|
|
||||||
*out = make([]string, len(*in))
|
|
||||||
copy(*out, *in)
|
|
||||||
}
|
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -21,7 +21,6 @@ import (
|
|||||||
|
|
||||||
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
||||||
componentbaseconfig "k8s.io/component-base/config"
|
componentbaseconfig "k8s.io/component-base/config"
|
||||||
registry "k8s.io/component-base/logs/api/v1"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
|
// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
|
||||||
@@ -54,10 +53,7 @@ type DeschedulerConfiguration struct {
|
|||||||
// IgnorePVCPods sets whether PVC pods should be allowed to be evicted
|
// IgnorePVCPods sets whether PVC pods should be allowed to be evicted
|
||||||
IgnorePVCPods bool
|
IgnorePVCPods bool
|
||||||
|
|
||||||
// LeaderElection starts Deployment using leader election loop
|
|
||||||
LeaderElection componentbaseconfig.LeaderElectionConfiguration
|
|
||||||
|
|
||||||
// Logging specifies the options of logging.
|
// Logging specifies the options of logging.
|
||||||
// Refer [Logs Options](https://github.com/kubernetes/component-base/blob/master/logs/api/v1/options.go) for more information.
|
// Refer [Logs Options](https://github.com/kubernetes/component-base/blob/master/logs/options.go) for more information.
|
||||||
Logging registry.LoggingConfiguration
|
Logging componentbaseconfig.LoggingConfiguration
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,133 +0,0 @@
|
|||||||
/*
|
|
||||||
Copyright 2022 The Kubernetes Authors.
|
|
||||||
|
|
||||||
Licensed under the Apache License, Version 2.0 (the "License");
|
|
||||||
you may not use this file except in compliance with the License.
|
|
||||||
You may obtain a copy of the License at
|
|
||||||
|
|
||||||
http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
|
|
||||||
Unless required by applicable law or agreed to in writing, software
|
|
||||||
distributed under the License is distributed on an "AS IS" BASIS,
|
|
||||||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
||||||
See the License for the specific language governing permissions and
|
|
||||||
limitations under the License.
|
|
||||||
*/
|
|
||||||
|
|
||||||
package componentconfig
|
|
||||||
|
|
||||||
import (
|
|
||||||
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
|
||||||
"sigs.k8s.io/descheduler/pkg/api"
|
|
||||||
)
|
|
||||||
|
|
||||||
// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
|
|
||||||
|
|
||||||
// RemovePodsViolatingNodeTaintsArgs holds arguments used to configure the RemovePodsViolatingNodeTaints plugin.
|
|
||||||
type RemovePodsViolatingNodeTaintsArgs struct {
|
|
||||||
metav1.TypeMeta
|
|
||||||
|
|
||||||
Namespaces *api.Namespaces
|
|
||||||
LabelSelector *metav1.LabelSelector
|
|
||||||
IncludePreferNoSchedule bool
|
|
||||||
ExcludedTaints []string
|
|
||||||
}
|
|
||||||
|
|
||||||
// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
|
|
||||||
|
|
||||||
// RemoveFailedPodsArgs holds arguments used to configure RemoveFailedPods plugin.
|
|
||||||
type RemoveFailedPodsArgs struct {
|
|
||||||
metav1.TypeMeta
|
|
||||||
|
|
||||||
Namespaces *api.Namespaces
|
|
||||||
LabelSelector *metav1.LabelSelector
|
|
||||||
ExcludeOwnerKinds []string
|
|
||||||
MinPodLifetimeSeconds *uint
|
|
||||||
Reasons []string
|
|
||||||
IncludingInitContainers bool
|
|
||||||
}
|
|
||||||
|
|
||||||
// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
|
|
||||||
|
|
||||||
// RemovePodsViolatingNodeAffinityArgs holds arguments used to configure RemovePodsViolatingNodeAffinity plugin.
|
|
||||||
type RemovePodsViolatingNodeAffinityArgs struct {
|
|
||||||
metav1.TypeMeta
|
|
||||||
|
|
||||||
Namespaces *api.Namespaces
|
|
||||||
LabelSelector *metav1.LabelSelector
|
|
||||||
NodeAffinityType []string
|
|
||||||
}
|
|
||||||
|
|
||||||
// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
|
|
||||||
|
|
||||||
// RemovePodsHavingTooManyRestartsArgs holds arguments used to configure RemovePodsHavingTooManyRestarts plugin.
|
|
||||||
type RemovePodsHavingTooManyRestartsArgs struct {
|
|
||||||
metav1.TypeMeta
|
|
||||||
|
|
||||||
Namespaces *api.Namespaces
|
|
||||||
LabelSelector *metav1.LabelSelector
|
|
||||||
PodRestartThreshold int32
|
|
||||||
IncludingInitContainers bool
|
|
||||||
}
|
|
||||||
|
|
||||||
// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
|
|
||||||
|
|
||||||
type RemoveDuplicatesArgs struct {
|
|
||||||
metav1.TypeMeta
|
|
||||||
|
|
||||||
Namespaces *api.Namespaces
|
|
||||||
ExcludeOwnerKinds []string
|
|
||||||
}
|
|
||||||
|
|
||||||
// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
|
|
||||||
|
|
||||||
// PodLifeTimeArgs holds arguments used to configure PodLifeTime plugin.
|
|
||||||
type PodLifeTimeArgs struct {
|
|
||||||
metav1.TypeMeta
|
|
||||||
|
|
||||||
Namespaces *api.Namespaces
|
|
||||||
LabelSelector *metav1.LabelSelector
|
|
||||||
MaxPodLifeTimeSeconds *uint
|
|
||||||
States []string
|
|
||||||
}
|
|
||||||
|
|
||||||
// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
|
|
||||||
|
|
||||||
// RemovePodsViolatingTopologySpreadConstraintArgs holds arguments used to configure RemovePodsViolatingTopologySpreadConstraint plugin.
|
|
||||||
type RemovePodsViolatingTopologySpreadConstraintArgs struct {
|
|
||||||
metav1.TypeMeta
|
|
||||||
|
|
||||||
Namespaces *api.Namespaces
|
|
||||||
LabelSelector *metav1.LabelSelector
|
|
||||||
IncludeSoftConstraints bool
|
|
||||||
}
|
|
||||||
|
|
||||||
// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
|
|
||||||
|
|
||||||
// RemovePodsViolatingInterPodAntiAffinity holds arguments used to configure RemovePodsViolatingInterPodAntiAffinity plugin.
|
|
||||||
type RemovePodsViolatingInterPodAntiAffinityArgs struct {
|
|
||||||
metav1.TypeMeta
|
|
||||||
|
|
||||||
Namespaces *api.Namespaces
|
|
||||||
LabelSelector *metav1.LabelSelector
|
|
||||||
}
|
|
||||||
|
|
||||||
// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
|
|
||||||
|
|
||||||
type LowNodeUtilizationArgs struct {
|
|
||||||
metav1.TypeMeta
|
|
||||||
|
|
||||||
UseDeviationThresholds bool
|
|
||||||
Thresholds api.ResourceThresholds
|
|
||||||
TargetThresholds api.ResourceThresholds
|
|
||||||
NumberOfNodes int
|
|
||||||
}
|
|
||||||
|
|
||||||
// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
|
|
||||||
|
|
||||||
type HighNodeUtilizationArgs struct {
|
|
||||||
metav1.TypeMeta
|
|
||||||
|
|
||||||
Thresholds api.ResourceThresholds
|
|
||||||
NumberOfNodes int
|
|
||||||
}
|
|
||||||
@@ -17,11 +17,10 @@ limitations under the License.
|
|||||||
package v1alpha1
|
package v1alpha1
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
componentbaseconfig "k8s.io/component-base/config"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
||||||
componentbaseconfig "k8s.io/component-base/config"
|
|
||||||
registry "k8s.io/component-base/logs/api/v1"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
|
// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
|
||||||
@@ -54,10 +53,7 @@ type DeschedulerConfiguration struct {
|
|||||||
// IgnorePVCPods sets whether PVC pods should be allowed to be evicted
|
// IgnorePVCPods sets whether PVC pods should be allowed to be evicted
|
||||||
IgnorePVCPods bool `json:"ignorePvcPods,omitempty"`
|
IgnorePVCPods bool `json:"ignorePvcPods,omitempty"`
|
||||||
|
|
||||||
// LeaderElection starts Deployment using leader election loop
|
|
||||||
LeaderElection componentbaseconfig.LeaderElectionConfiguration `json:"leaderElection,omitempty"`
|
|
||||||
|
|
||||||
// Logging specifies the options of logging.
|
// Logging specifies the options of logging.
|
||||||
// Refer [Logs Options](https://github.com/kubernetes/component-base/blob/master/logs/api/v1/options.go) for more information.
|
// Refer [Logs Options](https://github.com/kubernetes/component-base/blob/master/logs/options.go) for more information.
|
||||||
Logging registry.LoggingConfiguration `json:"logging,omitempty"`
|
Logging componentbaseconfig.LoggingConfiguration `json:"logging,omitempty"`
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,8 +1,7 @@
|
|||||||
//go:build !ignore_autogenerated
|
|
||||||
// +build !ignore_autogenerated
|
// +build !ignore_autogenerated
|
||||||
|
|
||||||
/*
|
/*
|
||||||
Copyright 2022 The Kubernetes Authors.
|
Copyright 2021 The Kubernetes Authors.
|
||||||
|
|
||||||
Licensed under the Apache License, Version 2.0 (the "License");
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
you may not use this file except in compliance with the License.
|
you may not use this file except in compliance with the License.
|
||||||
@@ -58,7 +57,6 @@ func autoConvert_v1alpha1_DeschedulerConfiguration_To_componentconfig_Deschedule
|
|||||||
out.MaxNoOfPodsToEvictPerNode = in.MaxNoOfPodsToEvictPerNode
|
out.MaxNoOfPodsToEvictPerNode = in.MaxNoOfPodsToEvictPerNode
|
||||||
out.EvictLocalStoragePods = in.EvictLocalStoragePods
|
out.EvictLocalStoragePods = in.EvictLocalStoragePods
|
||||||
out.IgnorePVCPods = in.IgnorePVCPods
|
out.IgnorePVCPods = in.IgnorePVCPods
|
||||||
out.LeaderElection = in.LeaderElection
|
|
||||||
out.Logging = in.Logging
|
out.Logging = in.Logging
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
@@ -77,7 +75,6 @@ func autoConvert_componentconfig_DeschedulerConfiguration_To_v1alpha1_Deschedule
|
|||||||
out.MaxNoOfPodsToEvictPerNode = in.MaxNoOfPodsToEvictPerNode
|
out.MaxNoOfPodsToEvictPerNode = in.MaxNoOfPodsToEvictPerNode
|
||||||
out.EvictLocalStoragePods = in.EvictLocalStoragePods
|
out.EvictLocalStoragePods = in.EvictLocalStoragePods
|
||||||
out.IgnorePVCPods = in.IgnorePVCPods
|
out.IgnorePVCPods = in.IgnorePVCPods
|
||||||
out.LeaderElection = in.LeaderElection
|
|
||||||
out.Logging = in.Logging
|
out.Logging = in.Logging
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,8 +1,7 @@
|
|||||||
//go:build !ignore_autogenerated
|
|
||||||
// +build !ignore_autogenerated
|
// +build !ignore_autogenerated
|
||||||
|
|
||||||
/*
|
/*
|
||||||
Copyright 2022 The Kubernetes Authors.
|
Copyright 2021 The Kubernetes Authors.
|
||||||
|
|
||||||
Licensed under the Apache License, Version 2.0 (the "License");
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
you may not use this file except in compliance with the License.
|
you may not use this file except in compliance with the License.
|
||||||
@@ -29,8 +28,7 @@ import (
|
|||||||
func (in *DeschedulerConfiguration) DeepCopyInto(out *DeschedulerConfiguration) {
|
func (in *DeschedulerConfiguration) DeepCopyInto(out *DeschedulerConfiguration) {
|
||||||
*out = *in
|
*out = *in
|
||||||
out.TypeMeta = in.TypeMeta
|
out.TypeMeta = in.TypeMeta
|
||||||
out.LeaderElection = in.LeaderElection
|
out.Logging = in.Logging
|
||||||
in.Logging.DeepCopyInto(&out.Logging)
|
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -1,8 +1,7 @@
|
|||||||
//go:build !ignore_autogenerated
|
|
||||||
// +build !ignore_autogenerated
|
// +build !ignore_autogenerated
|
||||||
|
|
||||||
/*
|
/*
|
||||||
Copyright 2022 The Kubernetes Authors.
|
Copyright 2021 The Kubernetes Authors.
|
||||||
|
|
||||||
Licensed under the Apache License, Version 2.0 (the "License");
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
you may not use this file except in compliance with the License.
|
you may not use this file except in compliance with the License.
|
||||||
|
|||||||
@@ -1,203 +0,0 @@
|
|||||||
/*
|
|
||||||
Copyright 2022 The Kubernetes Authors.
|
|
||||||
|
|
||||||
Licensed under the Apache License, Version 2.0 (the "License");
|
|
||||||
you may not use this file except in compliance with the License.
|
|
||||||
You may obtain a copy of the License at
|
|
||||||
|
|
||||||
http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
|
|
||||||
Unless required by applicable law or agreed to in writing, software
|
|
||||||
distributed under the License is distributed on an "AS IS" BASIS,
|
|
||||||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
||||||
See the License for the specific language governing permissions and
|
|
||||||
limitations under the License.
|
|
||||||
*/
|
|
||||||
|
|
||||||
package validation
|
|
||||||
|
|
||||||
import (
|
|
||||||
"fmt"
|
|
||||||
|
|
||||||
v1 "k8s.io/api/core/v1"
|
|
||||||
utilerrors "k8s.io/apimachinery/pkg/util/errors"
|
|
||||||
"k8s.io/apimachinery/pkg/util/sets"
|
|
||||||
|
|
||||||
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
|
||||||
"sigs.k8s.io/descheduler/pkg/api"
|
|
||||||
"sigs.k8s.io/descheduler/pkg/apis/componentconfig"
|
|
||||||
)
|
|
||||||
|
|
||||||
const (
|
|
||||||
// MinResourcePercentage is the minimum value of a resource's percentage
|
|
||||||
MinResourcePercentage = 0
|
|
||||||
// MaxResourcePercentage is the maximum value of a resource's percentage
|
|
||||||
MaxResourcePercentage = 100
|
|
||||||
)
|
|
||||||
|
|
||||||
// ValidateRemovePodsViolatingNodeTaintsArgs validates RemovePodsViolatingNodeTaints arguments
|
|
||||||
func ValidateRemovePodsViolatingNodeTaintsArgs(args *componentconfig.RemovePodsViolatingNodeTaintsArgs) error {
|
|
||||||
return errorsAggregate(
|
|
||||||
validateNamespaceArgs(args.Namespaces),
|
|
||||||
validateLabelSelectorArgs(args.LabelSelector),
|
|
||||||
)
|
|
||||||
}
|
|
||||||
|
|
||||||
// ValidateRemoveFailedPodsArgs validates RemoveFailedPods arguments
|
|
||||||
func ValidateRemoveFailedPodsArgs(args *componentconfig.RemoveFailedPodsArgs) error {
|
|
||||||
return errorsAggregate(
|
|
||||||
validateNamespaceArgs(args.Namespaces),
|
|
||||||
validateLabelSelectorArgs(args.LabelSelector),
|
|
||||||
)
|
|
||||||
}
|
|
||||||
|
|
||||||
// ValidateRemovePodsHavingTooManyRestartsArgs validates RemovePodsHavingTooManyRestarts arguments
|
|
||||||
func ValidateRemovePodsHavingTooManyRestartsArgs(args *componentconfig.RemovePodsHavingTooManyRestartsArgs) error {
|
|
||||||
return errorsAggregate(
|
|
||||||
validateNamespaceArgs(args.Namespaces),
|
|
||||||
validateLabelSelectorArgs(args.LabelSelector),
|
|
||||||
validatePodRestartThreshold(args.PodRestartThreshold),
|
|
||||||
)
|
|
||||||
}
|
|
||||||
|
|
||||||
// ValidateRemovePodsViolatingNodeAffinityArgs validates RemovePodsViolatingNodeAffinity arguments
|
|
||||||
func ValidateRemovePodsViolatingNodeAffinityArgs(args *componentconfig.RemovePodsViolatingNodeAffinityArgs) error {
|
|
||||||
var err error
|
|
||||||
if args == nil || len(args.NodeAffinityType) == 0 {
|
|
||||||
err = fmt.Errorf("nodeAffinityType needs to be set")
|
|
||||||
}
|
|
||||||
|
|
||||||
return errorsAggregate(
|
|
||||||
err,
|
|
||||||
validateNamespaceArgs(args.Namespaces),
|
|
||||||
validateLabelSelectorArgs(args.LabelSelector),
|
|
||||||
)
|
|
||||||
}
|
|
||||||
|
|
||||||
// ValidatePodLifeTimeArgs validates PodLifeTime arguments
|
|
||||||
func ValidatePodLifeTimeArgs(args *componentconfig.PodLifeTimeArgs) error {
|
|
||||||
var err error
|
|
||||||
if args.MaxPodLifeTimeSeconds == nil {
|
|
||||||
err = fmt.Errorf("MaxPodLifeTimeSeconds not set")
|
|
||||||
}
|
|
||||||
|
|
||||||
return errorsAggregate(
|
|
||||||
err,
|
|
||||||
validateNamespaceArgs(args.Namespaces),
|
|
||||||
validateLabelSelectorArgs(args.LabelSelector),
|
|
||||||
validatePodLifeTimeStates(args.States),
|
|
||||||
)
|
|
||||||
}
|
|
||||||
|
|
||||||
func ValidateRemoveDuplicatesArgs(args *componentconfig.RemoveDuplicatesArgs) error {
|
|
||||||
return validateNamespaceArgs(args.Namespaces)
|
|
||||||
}
|
|
||||||
|
|
||||||
// ValidateRemovePodsViolatingTopologySpreadConstraintArgs validates RemovePodsViolatingTopologySpreadConstraint arguments
|
|
||||||
func ValidateRemovePodsViolatingTopologySpreadConstraintArgs(args *componentconfig.RemovePodsViolatingTopologySpreadConstraintArgs) error {
|
|
||||||
return errorsAggregate(
|
|
||||||
validateNamespaceArgs(args.Namespaces),
|
|
||||||
validateLabelSelectorArgs(args.LabelSelector),
|
|
||||||
)
|
|
||||||
}
|
|
||||||
|
|
||||||
// ValidateRemovePodsViolatingInterPodAntiAffinityArgs validates ValidateRemovePodsViolatingInterPodAntiAffinity arguments
|
|
||||||
func ValidateRemovePodsViolatingInterPodAntiAffinityArgs(args *componentconfig.RemovePodsViolatingInterPodAntiAffinityArgs) error {
|
|
||||||
return errorsAggregate(
|
|
||||||
validateNamespaceArgs(args.Namespaces),
|
|
||||||
validateLabelSelectorArgs(args.LabelSelector),
|
|
||||||
)
|
|
||||||
}
|
|
||||||
|
|
||||||
// errorsAggregate converts all arg validation errors to a single error interface.
|
|
||||||
// if no errors, it will return nil.
|
|
||||||
func errorsAggregate(errors ...error) error {
|
|
||||||
return utilerrors.NewAggregate(errors)
|
|
||||||
}
|
|
||||||
|
|
||||||
func validateNamespaceArgs(namespaces *api.Namespaces) error {
|
|
||||||
// At most one of include/exclude can be set
|
|
||||||
if namespaces != nil && len(namespaces.Include) > 0 && len(namespaces.Exclude) > 0 {
|
|
||||||
return fmt.Errorf("only one of Include/Exclude namespaces can be set")
|
|
||||||
}
|
|
||||||
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func validateLabelSelectorArgs(labelSelector *metav1.LabelSelector) error {
|
|
||||||
if labelSelector != nil {
|
|
||||||
if _, err := metav1.LabelSelectorAsSelector(labelSelector); err != nil {
|
|
||||||
return fmt.Errorf("failed to get label selectors from strategy's params: %+v", err)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func validatePodRestartThreshold(podRestartThreshold int32) error {
|
|
||||||
if podRestartThreshold < 1 {
|
|
||||||
return fmt.Errorf("PodsHavingTooManyRestarts threshold not set")
|
|
||||||
}
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func validatePodLifeTimeStates(states []string) error {
|
|
||||||
podLifeTimeAllowedStates := sets.NewString(
|
|
||||||
string(v1.PodRunning),
|
|
||||||
string(v1.PodPending),
|
|
||||||
|
|
||||||
// Container state reasons: https://github.com/kubernetes/kubernetes/blob/release-1.24/pkg/kubelet/kubelet_pods.go#L76-L79
|
|
||||||
"PodInitializing",
|
|
||||||
"ContainerCreating",
|
|
||||||
)
|
|
||||||
|
|
||||||
if !podLifeTimeAllowedStates.HasAll(states...) {
|
|
||||||
return fmt.Errorf("states must be one of %v", podLifeTimeAllowedStates.List())
|
|
||||||
}
|
|
||||||
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func ValidateHighNodeUtilizationArgs(args *componentconfig.HighNodeUtilizationArgs) error {
|
|
||||||
return validateThresholds(args.Thresholds)
|
|
||||||
}
|
|
||||||
|
|
||||||
func ValidateLowNodeUtilizationArgs(args *componentconfig.LowNodeUtilizationArgs) error {
|
|
||||||
return validateLowNodeUtilizationThresholds(args.Thresholds, args.TargetThresholds, args.UseDeviationThresholds)
|
|
||||||
}
|
|
||||||
|
|
||||||
func validateLowNodeUtilizationThresholds(thresholds, targetThresholds api.ResourceThresholds, useDeviationThresholds bool) error {
|
|
||||||
// validate thresholds and targetThresholds config
|
|
||||||
if err := validateThresholds(thresholds); err != nil {
|
|
||||||
return fmt.Errorf("thresholds config is not valid: %v", err)
|
|
||||||
}
|
|
||||||
if err := validateThresholds(targetThresholds); err != nil {
|
|
||||||
return fmt.Errorf("targetThresholds config is not valid: %v", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
// validate if thresholds and targetThresholds have same resources configured
|
|
||||||
if len(thresholds) != len(targetThresholds) {
|
|
||||||
return fmt.Errorf("thresholds and targetThresholds configured different resources")
|
|
||||||
}
|
|
||||||
for resourceName, value := range thresholds {
|
|
||||||
if targetValue, ok := targetThresholds[resourceName]; !ok {
|
|
||||||
return fmt.Errorf("thresholds and targetThresholds configured different resources")
|
|
||||||
} else if value > targetValue && !useDeviationThresholds {
|
|
||||||
return fmt.Errorf("thresholds' %v percentage is greater than targetThresholds'", resourceName)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// validateThresholds checks if thresholds have valid resource name and resource percentage configured
|
|
||||||
func validateThresholds(thresholds api.ResourceThresholds) error {
|
|
||||||
if len(thresholds) == 0 {
|
|
||||||
return fmt.Errorf("no resource threshold is configured")
|
|
||||||
}
|
|
||||||
for name, percent := range thresholds {
|
|
||||||
if percent < MinResourcePercentage || percent > MaxResourcePercentage {
|
|
||||||
return fmt.Errorf("%v threshold not in [%v, %v] range", name, MinResourcePercentage, MaxResourcePercentage)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
@@ -1,331 +0,0 @@
|
|||||||
/*
|
|
||||||
Copyright 2022 The Kubernetes Authors.
|
|
||||||
|
|
||||||
Licensed under the Apache License, Version 2.0 (the "License");
|
|
||||||
you may not use this file except in compliance with the License.
|
|
||||||
You may obtain a copy of the License at
|
|
||||||
|
|
||||||
http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
|
|
||||||
Unless required by applicable law or agreed to in writing, software
|
|
||||||
distributed under the License is distributed on an "AS IS" BASIS,
|
|
||||||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
||||||
See the License for the specific language governing permissions and
|
|
||||||
limitations under the License.
|
|
||||||
*/
|
|
||||||
|
|
||||||
package validation
|
|
||||||
|
|
||||||
import (
|
|
||||||
"fmt"
|
|
||||||
"testing"
|
|
||||||
|
|
||||||
v1 "k8s.io/api/core/v1"
|
|
||||||
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
|
||||||
"sigs.k8s.io/descheduler/pkg/api"
|
|
||||||
"sigs.k8s.io/descheduler/pkg/apis/componentconfig"
|
|
||||||
)
|
|
||||||
|
|
||||||
func TestValidateRemovePodsViolatingNodeTaintsArgs(t *testing.T) {
|
|
||||||
testCases := []struct {
|
|
||||||
description string
|
|
||||||
args *componentconfig.RemovePodsViolatingNodeTaintsArgs
|
|
||||||
expectError bool
|
|
||||||
}{
|
|
||||||
{
|
|
||||||
description: "valid namespace args, no errors",
|
|
||||||
args: &componentconfig.RemovePodsViolatingNodeTaintsArgs{
|
|
||||||
Namespaces: &api.Namespaces{
|
|
||||||
Include: []string{"default"},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
expectError: false,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
description: "invalid namespaces args, expects error",
|
|
||||||
args: &componentconfig.RemovePodsViolatingNodeTaintsArgs{
|
|
||||||
Namespaces: &api.Namespaces{
|
|
||||||
Include: []string{"default"},
|
|
||||||
Exclude: []string{"kube-system"},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
expectError: true,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
description: "valid label selector args, no errors",
|
|
||||||
args: &componentconfig.RemovePodsViolatingNodeTaintsArgs{
|
|
||||||
LabelSelector: &metav1.LabelSelector{
|
|
||||||
MatchLabels: map[string]string{"role.kubernetes.io/node": ""},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
expectError: false,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
description: "invalid label selector args, expects errors",
|
|
||||||
args: &componentconfig.RemovePodsViolatingNodeTaintsArgs{
|
|
||||||
LabelSelector: &metav1.LabelSelector{
|
|
||||||
MatchExpressions: []metav1.LabelSelectorRequirement{
|
|
||||||
{
|
|
||||||
Operator: metav1.LabelSelectorOpIn,
|
|
||||||
},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
expectError: true,
|
|
||||||
},
|
|
||||||
}
|
|
||||||
|
|
||||||
for _, tc := range testCases {
|
|
||||||
t.Run(tc.description, func(t *testing.T) {
|
|
||||||
err := ValidateRemovePodsViolatingNodeTaintsArgs(tc.args)
|
|
||||||
|
|
||||||
hasError := err != nil
|
|
||||||
if tc.expectError != hasError {
|
|
||||||
t.Error("unexpected arg validation behavior")
|
|
||||||
}
|
|
||||||
})
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestValidateRemovePodsViolatingNodeAffinityArgs(t *testing.T) {
|
|
||||||
testCases := []struct {
|
|
||||||
description string
|
|
||||||
args *componentconfig.RemovePodsViolatingNodeAffinityArgs
|
|
||||||
expectError bool
|
|
||||||
}{
|
|
||||||
{
|
|
||||||
description: "nil NodeAffinityType args, expects errors",
|
|
||||||
args: &componentconfig.RemovePodsViolatingNodeAffinityArgs{
|
|
||||||
NodeAffinityType: nil,
|
|
||||||
},
|
|
||||||
expectError: true,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
description: "empty NodeAffinityType args, expects errors",
|
|
||||||
args: &componentconfig.RemovePodsViolatingNodeAffinityArgs{
|
|
||||||
NodeAffinityType: []string{},
|
|
||||||
},
|
|
||||||
expectError: true,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
description: "valid NodeAffinityType args, no errors",
|
|
||||||
args: &componentconfig.RemovePodsViolatingNodeAffinityArgs{
|
|
||||||
NodeAffinityType: []string{"requiredDuringSchedulingIgnoredDuringExecution"},
|
|
||||||
},
|
|
||||||
expectError: false,
|
|
||||||
},
|
|
||||||
}
|
|
||||||
|
|
||||||
for _, tc := range testCases {
|
|
||||||
t.Run(tc.description, func(t *testing.T) {
|
|
||||||
err := ValidateRemovePodsViolatingNodeAffinityArgs(tc.args)
|
|
||||||
|
|
||||||
hasError := err != nil
|
|
||||||
if tc.expectError != hasError {
|
|
||||||
t.Error("unexpected arg validation behavior")
|
|
||||||
}
|
|
||||||
})
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestValidateRemovePodLifeTimeArgs(t *testing.T) {
|
|
||||||
testCases := []struct {
|
|
||||||
description string
|
|
||||||
args *componentconfig.PodLifeTimeArgs
|
|
||||||
expectError bool
|
|
||||||
}{
|
|
||||||
{
|
|
||||||
description: "valid arg, no errors",
|
|
||||||
args: &componentconfig.PodLifeTimeArgs{
|
|
||||||
MaxPodLifeTimeSeconds: func(i uint) *uint { return &i }(1),
|
|
||||||
States: []string{string(v1.PodRunning)},
|
|
||||||
},
|
|
||||||
expectError: false,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
description: "nil MaxPodLifeTimeSeconds arg, expects errors",
|
|
||||||
args: &componentconfig.PodLifeTimeArgs{
|
|
||||||
MaxPodLifeTimeSeconds: nil,
|
|
||||||
},
|
|
||||||
expectError: true,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
description: "invalid pod state arg, expects errors",
|
|
||||||
args: &componentconfig.PodLifeTimeArgs{
|
|
||||||
States: []string{string(v1.NodeRunning)},
|
|
||||||
},
|
|
||||||
expectError: true,
|
|
||||||
},
|
|
||||||
}
|
|
||||||
|
|
||||||
for _, tc := range testCases {
|
|
||||||
t.Run(tc.description, func(t *testing.T) {
|
|
||||||
err := ValidatePodLifeTimeArgs(tc.args)
|
|
||||||
|
|
||||||
hasError := err != nil
|
|
||||||
if tc.expectError != hasError {
|
|
||||||
t.Error("unexpected arg validation behavior")
|
|
||||||
}
|
|
||||||
})
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestValidateLowNodeUtilizationPluginConfig(t *testing.T) {
|
|
||||||
var extendedResource = v1.ResourceName("example.com/foo")
|
|
||||||
tests := []struct {
|
|
||||||
name string
|
|
||||||
thresholds api.ResourceThresholds
|
|
||||||
targetThresholds api.ResourceThresholds
|
|
||||||
errInfo error
|
|
||||||
}{
|
|
||||||
{
|
|
||||||
name: "passing invalid thresholds",
|
|
||||||
thresholds: api.ResourceThresholds{
|
|
||||||
v1.ResourceCPU: 20,
|
|
||||||
v1.ResourceMemory: 120,
|
|
||||||
},
|
|
||||||
targetThresholds: api.ResourceThresholds{
|
|
||||||
v1.ResourceCPU: 80,
|
|
||||||
v1.ResourceMemory: 80,
|
|
||||||
},
|
|
||||||
errInfo: fmt.Errorf("thresholds config is not valid: %v", fmt.Errorf(
|
|
||||||
"%v threshold not in [%v, %v] range", v1.ResourceMemory, MinResourcePercentage, MaxResourcePercentage)),
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "thresholds and targetThresholds configured different num of resources",
|
|
||||||
thresholds: api.ResourceThresholds{
|
|
||||||
v1.ResourceCPU: 20,
|
|
||||||
v1.ResourceMemory: 20,
|
|
||||||
},
|
|
||||||
targetThresholds: api.ResourceThresholds{
|
|
||||||
v1.ResourceCPU: 80,
|
|
||||||
v1.ResourceMemory: 80,
|
|
||||||
v1.ResourcePods: 80,
|
|
||||||
},
|
|
||||||
errInfo: fmt.Errorf("thresholds and targetThresholds configured different resources"),
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "thresholds and targetThresholds configured different resources",
|
|
||||||
thresholds: api.ResourceThresholds{
|
|
||||||
v1.ResourceCPU: 20,
|
|
||||||
v1.ResourceMemory: 20,
|
|
||||||
},
|
|
||||||
targetThresholds: api.ResourceThresholds{
|
|
||||||
v1.ResourceCPU: 80,
|
|
||||||
v1.ResourcePods: 80,
|
|
||||||
},
|
|
||||||
errInfo: fmt.Errorf("thresholds and targetThresholds configured different resources"),
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "thresholds' CPU config value is greater than targetThresholds'",
|
|
||||||
thresholds: api.ResourceThresholds{
|
|
||||||
v1.ResourceCPU: 90,
|
|
||||||
v1.ResourceMemory: 20,
|
|
||||||
},
|
|
||||||
targetThresholds: api.ResourceThresholds{
|
|
||||||
v1.ResourceCPU: 80,
|
|
||||||
v1.ResourceMemory: 80,
|
|
||||||
},
|
|
||||||
errInfo: fmt.Errorf("thresholds' %v percentage is greater than targetThresholds'", v1.ResourceCPU),
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "only thresholds configured extended resource",
|
|
||||||
thresholds: api.ResourceThresholds{
|
|
||||||
v1.ResourceCPU: 20,
|
|
||||||
v1.ResourceMemory: 20,
|
|
||||||
extendedResource: 20,
|
|
||||||
},
|
|
||||||
targetThresholds: api.ResourceThresholds{
|
|
||||||
v1.ResourceCPU: 80,
|
|
||||||
v1.ResourceMemory: 80,
|
|
||||||
},
|
|
||||||
errInfo: fmt.Errorf("thresholds and targetThresholds configured different resources"),
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "only targetThresholds configured extended resource",
|
|
||||||
thresholds: api.ResourceThresholds{
|
|
||||||
v1.ResourceCPU: 20,
|
|
||||||
v1.ResourceMemory: 20,
|
|
||||||
},
|
|
||||||
targetThresholds: api.ResourceThresholds{
|
|
||||||
v1.ResourceCPU: 80,
|
|
||||||
v1.ResourceMemory: 80,
|
|
||||||
extendedResource: 80,
|
|
||||||
},
|
|
||||||
errInfo: fmt.Errorf("thresholds and targetThresholds configured different resources"),
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "thresholds and targetThresholds configured different extended resources",
|
|
||||||
thresholds: api.ResourceThresholds{
|
|
||||||
v1.ResourceCPU: 20,
|
|
||||||
v1.ResourceMemory: 20,
|
|
||||||
extendedResource: 20,
|
|
||||||
},
|
|
||||||
targetThresholds: api.ResourceThresholds{
|
|
||||||
v1.ResourceCPU: 80,
|
|
||||||
v1.ResourceMemory: 80,
|
|
||||||
"example.com/bar": 80,
|
|
||||||
},
|
|
||||||
errInfo: fmt.Errorf("thresholds and targetThresholds configured different resources"),
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "thresholds' extended resource config value is greater than targetThresholds'",
|
|
||||||
thresholds: api.ResourceThresholds{
|
|
||||||
v1.ResourceCPU: 20,
|
|
||||||
v1.ResourceMemory: 20,
|
|
||||||
extendedResource: 90,
|
|
||||||
},
|
|
||||||
targetThresholds: api.ResourceThresholds{
|
|
||||||
v1.ResourceCPU: 80,
|
|
||||||
v1.ResourceMemory: 80,
|
|
||||||
extendedResource: 20,
|
|
||||||
},
|
|
||||||
errInfo: fmt.Errorf("thresholds' %v percentage is greater than targetThresholds'", extendedResource),
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "passing valid plugin config",
|
|
||||||
thresholds: api.ResourceThresholds{
|
|
||||||
v1.ResourceCPU: 20,
|
|
||||||
v1.ResourceMemory: 20,
|
|
||||||
},
|
|
||||||
targetThresholds: api.ResourceThresholds{
|
|
||||||
v1.ResourceCPU: 80,
|
|
||||||
v1.ResourceMemory: 80,
|
|
||||||
},
|
|
||||||
errInfo: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "passing valid plugin config with extended resource",
|
|
||||||
thresholds: api.ResourceThresholds{
|
|
||||||
v1.ResourceCPU: 20,
|
|
||||||
v1.ResourceMemory: 20,
|
|
||||||
extendedResource: 20,
|
|
||||||
},
|
|
||||||
targetThresholds: api.ResourceThresholds{
|
|
||||||
v1.ResourceCPU: 80,
|
|
||||||
v1.ResourceMemory: 80,
|
|
||||||
extendedResource: 80,
|
|
||||||
},
|
|
||||||
errInfo: nil,
|
|
||||||
},
|
|
||||||
}
|
|
||||||
|
|
||||||
for _, testCase := range tests {
|
|
||||||
args := &componentconfig.LowNodeUtilizationArgs{
|
|
||||||
|
|
||||||
Thresholds: testCase.thresholds,
|
|
||||||
TargetThresholds: testCase.targetThresholds,
|
|
||||||
}
|
|
||||||
validateErr := validateLowNodeUtilizationThresholds(args.Thresholds, args.TargetThresholds, false)
|
|
||||||
|
|
||||||
if validateErr == nil || testCase.errInfo == nil {
|
|
||||||
if validateErr != testCase.errInfo {
|
|
||||||
t.Errorf("expected validity of plugin config: thresholds %#v targetThresholds %#v to be %v but got %v instead",
|
|
||||||
testCase.thresholds, testCase.targetThresholds, testCase.errInfo, validateErr)
|
|
||||||
}
|
|
||||||
} else if validateErr.Error() != testCase.errInfo.Error() {
|
|
||||||
t.Errorf("expected validity of plugin config: thresholds %#v targetThresholds %#v to be %v but got %v instead",
|
|
||||||
testCase.thresholds, testCase.targetThresholds, testCase.errInfo, validateErr)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
@@ -1,8 +1,7 @@
|
|||||||
//go:build !ignore_autogenerated
|
|
||||||
// +build !ignore_autogenerated
|
// +build !ignore_autogenerated
|
||||||
|
|
||||||
/*
|
/*
|
||||||
Copyright 2022 The Kubernetes Authors.
|
Copyright 2021 The Kubernetes Authors.
|
||||||
|
|
||||||
Licensed under the Apache License, Version 2.0 (the "License");
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
you may not use this file except in compliance with the License.
|
you may not use this file except in compliance with the License.
|
||||||
@@ -22,17 +21,14 @@ limitations under the License.
|
|||||||
package componentconfig
|
package componentconfig
|
||||||
|
|
||||||
import (
|
import (
|
||||||
v1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
|
||||||
runtime "k8s.io/apimachinery/pkg/runtime"
|
runtime "k8s.io/apimachinery/pkg/runtime"
|
||||||
api "sigs.k8s.io/descheduler/pkg/api"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
|
// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
|
||||||
func (in *DeschedulerConfiguration) DeepCopyInto(out *DeschedulerConfiguration) {
|
func (in *DeschedulerConfiguration) DeepCopyInto(out *DeschedulerConfiguration) {
|
||||||
*out = *in
|
*out = *in
|
||||||
out.TypeMeta = in.TypeMeta
|
out.TypeMeta = in.TypeMeta
|
||||||
out.LeaderElection = in.LeaderElection
|
out.Logging = in.Logging
|
||||||
in.Logging.DeepCopyInto(&out.Logging)
|
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -53,389 +49,3 @@ func (in *DeschedulerConfiguration) DeepCopyObject() runtime.Object {
|
|||||||
}
|
}
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
|
|
||||||
func (in *HighNodeUtilizationArgs) DeepCopyInto(out *HighNodeUtilizationArgs) {
|
|
||||||
*out = *in
|
|
||||||
out.TypeMeta = in.TypeMeta
|
|
||||||
if in.Thresholds != nil {
|
|
||||||
in, out := &in.Thresholds, &out.Thresholds
|
|
||||||
*out = make(api.ResourceThresholds, len(*in))
|
|
||||||
for key, val := range *in {
|
|
||||||
(*out)[key] = val
|
|
||||||
}
|
|
||||||
}
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new HighNodeUtilizationArgs.
|
|
||||||
func (in *HighNodeUtilizationArgs) DeepCopy() *HighNodeUtilizationArgs {
|
|
||||||
if in == nil {
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
out := new(HighNodeUtilizationArgs)
|
|
||||||
in.DeepCopyInto(out)
|
|
||||||
return out
|
|
||||||
}
|
|
||||||
|
|
||||||
// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
|
|
||||||
func (in *HighNodeUtilizationArgs) DeepCopyObject() runtime.Object {
|
|
||||||
if c := in.DeepCopy(); c != nil {
|
|
||||||
return c
|
|
||||||
}
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
|
|
||||||
func (in *LowNodeUtilizationArgs) DeepCopyInto(out *LowNodeUtilizationArgs) {
|
|
||||||
*out = *in
|
|
||||||
out.TypeMeta = in.TypeMeta
|
|
||||||
if in.Thresholds != nil {
|
|
||||||
in, out := &in.Thresholds, &out.Thresholds
|
|
||||||
*out = make(api.ResourceThresholds, len(*in))
|
|
||||||
for key, val := range *in {
|
|
||||||
(*out)[key] = val
|
|
||||||
}
|
|
||||||
}
|
|
||||||
if in.TargetThresholds != nil {
|
|
||||||
in, out := &in.TargetThresholds, &out.TargetThresholds
|
|
||||||
*out = make(api.ResourceThresholds, len(*in))
|
|
||||||
for key, val := range *in {
|
|
||||||
(*out)[key] = val
|
|
||||||
}
|
|
||||||
}
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new LowNodeUtilizationArgs.
|
|
||||||
func (in *LowNodeUtilizationArgs) DeepCopy() *LowNodeUtilizationArgs {
|
|
||||||
if in == nil {
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
out := new(LowNodeUtilizationArgs)
|
|
||||||
in.DeepCopyInto(out)
|
|
||||||
return out
|
|
||||||
}
|
|
||||||
|
|
||||||
// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
|
|
||||||
func (in *LowNodeUtilizationArgs) DeepCopyObject() runtime.Object {
|
|
||||||
if c := in.DeepCopy(); c != nil {
|
|
||||||
return c
|
|
||||||
}
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
|
|
||||||
func (in *PodLifeTimeArgs) DeepCopyInto(out *PodLifeTimeArgs) {
|
|
||||||
*out = *in
|
|
||||||
out.TypeMeta = in.TypeMeta
|
|
||||||
if in.Namespaces != nil {
|
|
||||||
in, out := &in.Namespaces, &out.Namespaces
|
|
||||||
*out = new(api.Namespaces)
|
|
||||||
(*in).DeepCopyInto(*out)
|
|
||||||
}
|
|
||||||
if in.LabelSelector != nil {
|
|
||||||
in, out := &in.LabelSelector, &out.LabelSelector
|
|
||||||
*out = new(v1.LabelSelector)
|
|
||||||
(*in).DeepCopyInto(*out)
|
|
||||||
}
|
|
||||||
if in.MaxPodLifeTimeSeconds != nil {
|
|
||||||
in, out := &in.MaxPodLifeTimeSeconds, &out.MaxPodLifeTimeSeconds
|
|
||||||
*out = new(uint)
|
|
||||||
**out = **in
|
|
||||||
}
|
|
||||||
if in.States != nil {
|
|
||||||
in, out := &in.States, &out.States
|
|
||||||
*out = make([]string, len(*in))
|
|
||||||
copy(*out, *in)
|
|
||||||
}
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PodLifeTimeArgs.
|
|
||||||
func (in *PodLifeTimeArgs) DeepCopy() *PodLifeTimeArgs {
|
|
||||||
if in == nil {
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
out := new(PodLifeTimeArgs)
|
|
||||||
in.DeepCopyInto(out)
|
|
||||||
return out
|
|
||||||
}
|
|
||||||
|
|
||||||
// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
|
|
||||||
func (in *PodLifeTimeArgs) DeepCopyObject() runtime.Object {
|
|
||||||
if c := in.DeepCopy(); c != nil {
|
|
||||||
return c
|
|
||||||
}
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
|
|
||||||
func (in *RemoveDuplicatesArgs) DeepCopyInto(out *RemoveDuplicatesArgs) {
|
|
||||||
*out = *in
|
|
||||||
out.TypeMeta = in.TypeMeta
|
|
||||||
if in.Namespaces != nil {
|
|
||||||
in, out := &in.Namespaces, &out.Namespaces
|
|
||||||
*out = new(api.Namespaces)
|
|
||||||
(*in).DeepCopyInto(*out)
|
|
||||||
}
|
|
||||||
if in.ExcludeOwnerKinds != nil {
|
|
||||||
in, out := &in.ExcludeOwnerKinds, &out.ExcludeOwnerKinds
|
|
||||||
*out = make([]string, len(*in))
|
|
||||||
copy(*out, *in)
|
|
||||||
}
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new RemoveDuplicatesArgs.
|
|
||||||
func (in *RemoveDuplicatesArgs) DeepCopy() *RemoveDuplicatesArgs {
|
|
||||||
if in == nil {
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
out := new(RemoveDuplicatesArgs)
|
|
||||||
in.DeepCopyInto(out)
|
|
||||||
return out
|
|
||||||
}
|
|
||||||
|
|
||||||
// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
|
|
||||||
func (in *RemoveDuplicatesArgs) DeepCopyObject() runtime.Object {
|
|
||||||
if c := in.DeepCopy(); c != nil {
|
|
||||||
return c
|
|
||||||
}
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
|
|
||||||
func (in *RemoveFailedPodsArgs) DeepCopyInto(out *RemoveFailedPodsArgs) {
|
|
||||||
*out = *in
|
|
||||||
out.TypeMeta = in.TypeMeta
|
|
||||||
if in.Namespaces != nil {
|
|
||||||
in, out := &in.Namespaces, &out.Namespaces
|
|
||||||
*out = new(api.Namespaces)
|
|
||||||
(*in).DeepCopyInto(*out)
|
|
||||||
}
|
|
||||||
if in.LabelSelector != nil {
|
|
||||||
in, out := &in.LabelSelector, &out.LabelSelector
|
|
||||||
*out = new(v1.LabelSelector)
|
|
||||||
(*in).DeepCopyInto(*out)
|
|
||||||
}
|
|
||||||
if in.ExcludeOwnerKinds != nil {
|
|
||||||
in, out := &in.ExcludeOwnerKinds, &out.ExcludeOwnerKinds
|
|
||||||
*out = make([]string, len(*in))
|
|
||||||
copy(*out, *in)
|
|
||||||
}
|
|
||||||
if in.MinPodLifetimeSeconds != nil {
|
|
||||||
in, out := &in.MinPodLifetimeSeconds, &out.MinPodLifetimeSeconds
|
|
||||||
*out = new(uint)
|
|
||||||
**out = **in
|
|
||||||
}
|
|
||||||
if in.Reasons != nil {
|
|
||||||
in, out := &in.Reasons, &out.Reasons
|
|
||||||
*out = make([]string, len(*in))
|
|
||||||
copy(*out, *in)
|
|
||||||
}
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new RemoveFailedPodsArgs.
|
|
||||||
func (in *RemoveFailedPodsArgs) DeepCopy() *RemoveFailedPodsArgs {
|
|
||||||
if in == nil {
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
out := new(RemoveFailedPodsArgs)
|
|
||||||
in.DeepCopyInto(out)
|
|
||||||
return out
|
|
||||||
}
|
|
||||||
|
|
||||||
// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
|
|
||||||
func (in *RemoveFailedPodsArgs) DeepCopyObject() runtime.Object {
|
|
||||||
if c := in.DeepCopy(); c != nil {
|
|
||||||
return c
|
|
||||||
}
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
|
|
||||||
func (in *RemovePodsHavingTooManyRestartsArgs) DeepCopyInto(out *RemovePodsHavingTooManyRestartsArgs) {
|
|
||||||
*out = *in
|
|
||||||
out.TypeMeta = in.TypeMeta
|
|
||||||
if in.Namespaces != nil {
|
|
||||||
in, out := &in.Namespaces, &out.Namespaces
|
|
||||||
*out = new(api.Namespaces)
|
|
||||||
(*in).DeepCopyInto(*out)
|
|
||||||
}
|
|
||||||
if in.LabelSelector != nil {
|
|
||||||
in, out := &in.LabelSelector, &out.LabelSelector
|
|
||||||
*out = new(v1.LabelSelector)
|
|
||||||
(*in).DeepCopyInto(*out)
|
|
||||||
}
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new RemovePodsHavingTooManyRestartsArgs.
|
|
||||||
func (in *RemovePodsHavingTooManyRestartsArgs) DeepCopy() *RemovePodsHavingTooManyRestartsArgs {
|
|
||||||
if in == nil {
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
out := new(RemovePodsHavingTooManyRestartsArgs)
|
|
||||||
in.DeepCopyInto(out)
|
|
||||||
return out
|
|
||||||
}
|
|
||||||
|
|
||||||
// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
|
|
||||||
func (in *RemovePodsHavingTooManyRestartsArgs) DeepCopyObject() runtime.Object {
|
|
||||||
if c := in.DeepCopy(); c != nil {
|
|
||||||
return c
|
|
||||||
}
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
|
|
||||||
func (in *RemovePodsViolatingInterPodAntiAffinityArgs) DeepCopyInto(out *RemovePodsViolatingInterPodAntiAffinityArgs) {
|
|
||||||
*out = *in
|
|
||||||
out.TypeMeta = in.TypeMeta
|
|
||||||
if in.Namespaces != nil {
|
|
||||||
in, out := &in.Namespaces, &out.Namespaces
|
|
||||||
*out = new(api.Namespaces)
|
|
||||||
(*in).DeepCopyInto(*out)
|
|
||||||
}
|
|
||||||
if in.LabelSelector != nil {
|
|
||||||
in, out := &in.LabelSelector, &out.LabelSelector
|
|
||||||
*out = new(v1.LabelSelector)
|
|
||||||
(*in).DeepCopyInto(*out)
|
|
||||||
}
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new RemovePodsViolatingInterPodAntiAffinityArgs.
|
|
||||||
func (in *RemovePodsViolatingInterPodAntiAffinityArgs) DeepCopy() *RemovePodsViolatingInterPodAntiAffinityArgs {
|
|
||||||
if in == nil {
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
out := new(RemovePodsViolatingInterPodAntiAffinityArgs)
|
|
||||||
in.DeepCopyInto(out)
|
|
||||||
return out
|
|
||||||
}
|
|
||||||
|
|
||||||
// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
|
|
||||||
func (in *RemovePodsViolatingInterPodAntiAffinityArgs) DeepCopyObject() runtime.Object {
|
|
||||||
if c := in.DeepCopy(); c != nil {
|
|
||||||
return c
|
|
||||||
}
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
|
|
||||||
func (in *RemovePodsViolatingNodeAffinityArgs) DeepCopyInto(out *RemovePodsViolatingNodeAffinityArgs) {
|
|
||||||
*out = *in
|
|
||||||
out.TypeMeta = in.TypeMeta
|
|
||||||
if in.Namespaces != nil {
|
|
||||||
in, out := &in.Namespaces, &out.Namespaces
|
|
||||||
*out = new(api.Namespaces)
|
|
||||||
(*in).DeepCopyInto(*out)
|
|
||||||
}
|
|
||||||
if in.LabelSelector != nil {
|
|
||||||
in, out := &in.LabelSelector, &out.LabelSelector
|
|
||||||
*out = new(v1.LabelSelector)
|
|
||||||
(*in).DeepCopyInto(*out)
|
|
||||||
}
|
|
||||||
if in.NodeAffinityType != nil {
|
|
||||||
in, out := &in.NodeAffinityType, &out.NodeAffinityType
|
|
||||||
*out = make([]string, len(*in))
|
|
||||||
copy(*out, *in)
|
|
||||||
}
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new RemovePodsViolatingNodeAffinityArgs.
|
|
||||||
func (in *RemovePodsViolatingNodeAffinityArgs) DeepCopy() *RemovePodsViolatingNodeAffinityArgs {
|
|
||||||
if in == nil {
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
out := new(RemovePodsViolatingNodeAffinityArgs)
|
|
||||||
in.DeepCopyInto(out)
|
|
||||||
return out
|
|
||||||
}
|
|
||||||
|
|
||||||
// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
|
|
||||||
func (in *RemovePodsViolatingNodeAffinityArgs) DeepCopyObject() runtime.Object {
|
|
||||||
if c := in.DeepCopy(); c != nil {
|
|
||||||
return c
|
|
||||||
}
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
|
|
||||||
func (in *RemovePodsViolatingNodeTaintsArgs) DeepCopyInto(out *RemovePodsViolatingNodeTaintsArgs) {
|
|
||||||
*out = *in
|
|
||||||
out.TypeMeta = in.TypeMeta
|
|
||||||
if in.Namespaces != nil {
|
|
||||||
in, out := &in.Namespaces, &out.Namespaces
|
|
||||||
*out = new(api.Namespaces)
|
|
||||||
(*in).DeepCopyInto(*out)
|
|
||||||
}
|
|
||||||
if in.LabelSelector != nil {
|
|
||||||
in, out := &in.LabelSelector, &out.LabelSelector
|
|
||||||
*out = new(v1.LabelSelector)
|
|
||||||
(*in).DeepCopyInto(*out)
|
|
||||||
}
|
|
||||||
if in.ExcludedTaints != nil {
|
|
||||||
in, out := &in.ExcludedTaints, &out.ExcludedTaints
|
|
||||||
*out = make([]string, len(*in))
|
|
||||||
copy(*out, *in)
|
|
||||||
}
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new RemovePodsViolatingNodeTaintsArgs.
|
|
||||||
func (in *RemovePodsViolatingNodeTaintsArgs) DeepCopy() *RemovePodsViolatingNodeTaintsArgs {
|
|
||||||
if in == nil {
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
out := new(RemovePodsViolatingNodeTaintsArgs)
|
|
||||||
in.DeepCopyInto(out)
|
|
||||||
return out
|
|
||||||
}
|
|
||||||
|
|
||||||
// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
|
|
||||||
func (in *RemovePodsViolatingNodeTaintsArgs) DeepCopyObject() runtime.Object {
|
|
||||||
if c := in.DeepCopy(); c != nil {
|
|
||||||
return c
|
|
||||||
}
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
|
|
||||||
func (in *RemovePodsViolatingTopologySpreadConstraintArgs) DeepCopyInto(out *RemovePodsViolatingTopologySpreadConstraintArgs) {
|
|
||||||
*out = *in
|
|
||||||
out.TypeMeta = in.TypeMeta
|
|
||||||
if in.Namespaces != nil {
|
|
||||||
in, out := &in.Namespaces, &out.Namespaces
|
|
||||||
*out = new(api.Namespaces)
|
|
||||||
(*in).DeepCopyInto(*out)
|
|
||||||
}
|
|
||||||
if in.LabelSelector != nil {
|
|
||||||
in, out := &in.LabelSelector, &out.LabelSelector
|
|
||||||
*out = new(v1.LabelSelector)
|
|
||||||
(*in).DeepCopyInto(*out)
|
|
||||||
}
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new RemovePodsViolatingTopologySpreadConstraintArgs.
|
|
||||||
func (in *RemovePodsViolatingTopologySpreadConstraintArgs) DeepCopy() *RemovePodsViolatingTopologySpreadConstraintArgs {
|
|
||||||
if in == nil {
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
out := new(RemovePodsViolatingTopologySpreadConstraintArgs)
|
|
||||||
in.DeepCopyInto(out)
|
|
||||||
return out
|
|
||||||
}
|
|
||||||
|
|
||||||
// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
|
|
||||||
func (in *RemovePodsViolatingTopologySpreadConstraintArgs) DeepCopyObject() runtime.Object {
|
|
||||||
if c := in.DeepCopy(); c != nil {
|
|
||||||
return c
|
|
||||||
}
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|||||||
@@ -26,7 +26,7 @@ import (
|
|||||||
"k8s.io/client-go/tools/clientcmd"
|
"k8s.io/client-go/tools/clientcmd"
|
||||||
)
|
)
|
||||||
|
|
||||||
func CreateClient(kubeconfig string, userAgt string) (clientset.Interface, error) {
|
func CreateClient(kubeconfig string) (clientset.Interface, error) {
|
||||||
var cfg *rest.Config
|
var cfg *rest.Config
|
||||||
if len(kubeconfig) != 0 {
|
if len(kubeconfig) != 0 {
|
||||||
master, err := GetMasterFromKubeconfig(kubeconfig)
|
master, err := GetMasterFromKubeconfig(kubeconfig)
|
||||||
@@ -47,11 +47,7 @@ func CreateClient(kubeconfig string, userAgt string) (clientset.Interface, error
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
if len(userAgt) != 0 {
|
return clientset.NewForConfig(cfg)
|
||||||
return clientset.NewForConfig(rest.AddUserAgent(cfg, userAgt))
|
|
||||||
} else {
|
|
||||||
return clientset.NewForConfig(cfg)
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func GetMasterFromKubeconfig(filename string) (string, error) {
|
func GetMasterFromKubeconfig(filename string) (string, error) {
|
||||||
|
|||||||
@@ -19,22 +19,14 @@ package descheduler
|
|||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
"fmt"
|
"fmt"
|
||||||
|
"sigs.k8s.io/descheduler/pkg/descheduler/strategies/nodeutilization"
|
||||||
"k8s.io/klog/v2"
|
|
||||||
|
|
||||||
v1 "k8s.io/api/core/v1"
|
v1 "k8s.io/api/core/v1"
|
||||||
policy "k8s.io/api/policy/v1"
|
clientset "k8s.io/client-go/kubernetes"
|
||||||
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
"k8s.io/klog/v2"
|
||||||
"k8s.io/apimachinery/pkg/labels"
|
|
||||||
"k8s.io/apimachinery/pkg/runtime"
|
|
||||||
"k8s.io/apimachinery/pkg/util/wait"
|
"k8s.io/apimachinery/pkg/util/wait"
|
||||||
"k8s.io/client-go/informers"
|
"k8s.io/client-go/informers"
|
||||||
corev1informers "k8s.io/client-go/informers/core/v1"
|
|
||||||
schedulingv1informers "k8s.io/client-go/informers/scheduling/v1"
|
|
||||||
clientset "k8s.io/client-go/kubernetes"
|
|
||||||
fakeclientset "k8s.io/client-go/kubernetes/fake"
|
|
||||||
core "k8s.io/client-go/testing"
|
|
||||||
|
|
||||||
"sigs.k8s.io/descheduler/cmd/descheduler/app/options"
|
"sigs.k8s.io/descheduler/cmd/descheduler/app/options"
|
||||||
"sigs.k8s.io/descheduler/metrics"
|
"sigs.k8s.io/descheduler/metrics"
|
||||||
"sigs.k8s.io/descheduler/pkg/api"
|
"sigs.k8s.io/descheduler/pkg/api"
|
||||||
@@ -42,21 +34,18 @@ import (
|
|||||||
"sigs.k8s.io/descheduler/pkg/descheduler/evictions"
|
"sigs.k8s.io/descheduler/pkg/descheduler/evictions"
|
||||||
eutils "sigs.k8s.io/descheduler/pkg/descheduler/evictions/utils"
|
eutils "sigs.k8s.io/descheduler/pkg/descheduler/evictions/utils"
|
||||||
nodeutil "sigs.k8s.io/descheduler/pkg/descheduler/node"
|
nodeutil "sigs.k8s.io/descheduler/pkg/descheduler/node"
|
||||||
podutil "sigs.k8s.io/descheduler/pkg/descheduler/pod"
|
"sigs.k8s.io/descheduler/pkg/descheduler/strategies"
|
||||||
"sigs.k8s.io/descheduler/pkg/framework"
|
|
||||||
"sigs.k8s.io/descheduler/pkg/framework/plugins/defaultevictor"
|
|
||||||
"sigs.k8s.io/descheduler/pkg/utils"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
func Run(ctx context.Context, rs *options.DeschedulerServer) error {
|
func Run(rs *options.DeschedulerServer) error {
|
||||||
metrics.Register()
|
metrics.Register()
|
||||||
|
|
||||||
rsclient, eventClient, err := createClients(rs.KubeconfigFile)
|
ctx := context.Background()
|
||||||
|
rsclient, err := client.CreateClient(rs.KubeconfigFile)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
rs.Client = rsclient
|
rs.Client = rsclient
|
||||||
rs.EventClient = eventClient
|
|
||||||
|
|
||||||
deschedulerPolicy, err := LoadPolicyConfig(rs.PolicyConfigFile)
|
deschedulerPolicy, err := LoadPolicyConfig(rs.PolicyConfigFile)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@@ -71,208 +60,42 @@ func Run(ctx context.Context, rs *options.DeschedulerServer) error {
|
|||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
runFn := func() error {
|
stopChannel := make(chan struct{})
|
||||||
return RunDeschedulerStrategies(ctx, rs, deschedulerPolicy, evictionPolicyGroupVersion)
|
return RunDeschedulerStrategies(ctx, rs, deschedulerPolicy, evictionPolicyGroupVersion, stopChannel)
|
||||||
}
|
|
||||||
|
|
||||||
if rs.LeaderElection.LeaderElect && rs.DeschedulingInterval.Seconds() == 0 {
|
|
||||||
return fmt.Errorf("leaderElection must be used with deschedulingInterval")
|
|
||||||
}
|
|
||||||
|
|
||||||
if rs.LeaderElection.LeaderElect && !rs.DryRun {
|
|
||||||
if err := NewLeaderElection(runFn, rsclient, &rs.LeaderElection, ctx); err != nil {
|
|
||||||
return fmt.Errorf("leaderElection: %w", err)
|
|
||||||
}
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
return runFn()
|
|
||||||
}
|
}
|
||||||
|
|
||||||
type strategyFunction func(ctx context.Context, client clientset.Interface, strategy api.DeschedulerStrategy, nodes []*v1.Node, podEvictor *evictions.PodEvictor, evictorFilter framework.EvictorPlugin, getPodsAssignedToNode podutil.GetPodsAssignedToNodeFunc)
|
type strategyFunction func(ctx context.Context, client clientset.Interface, strategy api.DeschedulerStrategy, nodes []*v1.Node, podEvictor *evictions.PodEvictor)
|
||||||
|
|
||||||
func cachedClient(
|
func RunDeschedulerStrategies(ctx context.Context, rs *options.DeschedulerServer, deschedulerPolicy *api.DeschedulerPolicy, evictionPolicyGroupVersion string, stopChannel chan struct{}) error {
|
||||||
realClient clientset.Interface,
|
|
||||||
podInformer corev1informers.PodInformer,
|
|
||||||
nodeInformer corev1informers.NodeInformer,
|
|
||||||
namespaceInformer corev1informers.NamespaceInformer,
|
|
||||||
priorityClassInformer schedulingv1informers.PriorityClassInformer,
|
|
||||||
) (clientset.Interface, error) {
|
|
||||||
fakeClient := fakeclientset.NewSimpleClientset()
|
|
||||||
// simulate a pod eviction by deleting a pod
|
|
||||||
fakeClient.PrependReactor("create", "pods", func(action core.Action) (bool, runtime.Object, error) {
|
|
||||||
if action.GetSubresource() == "eviction" {
|
|
||||||
createAct, matched := action.(core.CreateActionImpl)
|
|
||||||
if !matched {
|
|
||||||
return false, nil, fmt.Errorf("unable to convert action to core.CreateActionImpl")
|
|
||||||
}
|
|
||||||
eviction, matched := createAct.Object.(*policy.Eviction)
|
|
||||||
if !matched {
|
|
||||||
return false, nil, fmt.Errorf("unable to convert action object into *policy.Eviction")
|
|
||||||
}
|
|
||||||
if err := fakeClient.Tracker().Delete(action.GetResource(), eviction.GetNamespace(), eviction.GetName()); err != nil {
|
|
||||||
return false, nil, fmt.Errorf("unable to delete pod %v/%v: %v", eviction.GetNamespace(), eviction.GetName(), err)
|
|
||||||
}
|
|
||||||
return true, nil, nil
|
|
||||||
}
|
|
||||||
// fallback to the default reactor
|
|
||||||
return false, nil, nil
|
|
||||||
})
|
|
||||||
|
|
||||||
klog.V(3).Infof("Pulling resources for the cached client from the cluster")
|
|
||||||
pods, err := podInformer.Lister().List(labels.Everything())
|
|
||||||
if err != nil {
|
|
||||||
return nil, fmt.Errorf("unable to list pods: %v", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
for _, item := range pods {
|
|
||||||
if _, err := fakeClient.CoreV1().Pods(item.Namespace).Create(context.TODO(), item, metav1.CreateOptions{}); err != nil {
|
|
||||||
return nil, fmt.Errorf("unable to copy pod: %v", err)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
nodes, err := nodeInformer.Lister().List(labels.Everything())
|
|
||||||
if err != nil {
|
|
||||||
return nil, fmt.Errorf("unable to list nodes: %v", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
for _, item := range nodes {
|
|
||||||
if _, err := fakeClient.CoreV1().Nodes().Create(context.TODO(), item, metav1.CreateOptions{}); err != nil {
|
|
||||||
return nil, fmt.Errorf("unable to copy node: %v", err)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
namespaces, err := namespaceInformer.Lister().List(labels.Everything())
|
|
||||||
if err != nil {
|
|
||||||
return nil, fmt.Errorf("unable to list namespaces: %v", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
for _, item := range namespaces {
|
|
||||||
if _, err := fakeClient.CoreV1().Namespaces().Create(context.TODO(), item, metav1.CreateOptions{}); err != nil {
|
|
||||||
return nil, fmt.Errorf("unable to copy node: %v", err)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
priorityClasses, err := priorityClassInformer.Lister().List(labels.Everything())
|
|
||||||
if err != nil {
|
|
||||||
return nil, fmt.Errorf("unable to list priorityclasses: %v", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
for _, item := range priorityClasses {
|
|
||||||
if _, err := fakeClient.SchedulingV1().PriorityClasses().Create(context.TODO(), item, metav1.CreateOptions{}); err != nil {
|
|
||||||
return nil, fmt.Errorf("unable to copy priorityclass: %v", err)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
return fakeClient, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// evictorImpl implements the Evictor interface so plugins
|
|
||||||
// can evict a pod without importing a specific pod evictor
|
|
||||||
type evictorImpl struct {
|
|
||||||
podEvictor *evictions.PodEvictor
|
|
||||||
evictorFilter framework.EvictorPlugin
|
|
||||||
}
|
|
||||||
|
|
||||||
var _ framework.Evictor = &evictorImpl{}
|
|
||||||
|
|
||||||
// Filter checks if a pod can be evicted
|
|
||||||
func (ei *evictorImpl) Filter(pod *v1.Pod) bool {
|
|
||||||
return ei.evictorFilter.Filter(pod)
|
|
||||||
}
|
|
||||||
|
|
||||||
// Evict evicts a pod (no pre-check performed)
|
|
||||||
func (ei *evictorImpl) Evict(ctx context.Context, pod *v1.Pod, opts evictions.EvictOptions) bool {
|
|
||||||
return ei.podEvictor.EvictPod(ctx, pod, opts)
|
|
||||||
}
|
|
||||||
|
|
||||||
func (ei *evictorImpl) NodeLimitExceeded(node *v1.Node) bool {
|
|
||||||
return ei.podEvictor.NodeLimitExceeded(node)
|
|
||||||
}
|
|
||||||
|
|
||||||
// handleImpl implements the framework handle which gets passed to plugins
|
|
||||||
type handleImpl struct {
|
|
||||||
clientSet clientset.Interface
|
|
||||||
getPodsAssignedToNodeFunc podutil.GetPodsAssignedToNodeFunc
|
|
||||||
sharedInformerFactory informers.SharedInformerFactory
|
|
||||||
evictor *evictorImpl
|
|
||||||
}
|
|
||||||
|
|
||||||
var _ framework.Handle = &handleImpl{}
|
|
||||||
|
|
||||||
// ClientSet retrieves kube client set
|
|
||||||
func (hi *handleImpl) ClientSet() clientset.Interface {
|
|
||||||
return hi.clientSet
|
|
||||||
}
|
|
||||||
|
|
||||||
// GetPodsAssignedToNodeFunc retrieves GetPodsAssignedToNodeFunc implementation
|
|
||||||
func (hi *handleImpl) GetPodsAssignedToNodeFunc() podutil.GetPodsAssignedToNodeFunc {
|
|
||||||
return hi.getPodsAssignedToNodeFunc
|
|
||||||
}
|
|
||||||
|
|
||||||
// SharedInformerFactory retrieves shared informer factory
|
|
||||||
func (hi *handleImpl) SharedInformerFactory() informers.SharedInformerFactory {
|
|
||||||
return hi.sharedInformerFactory
|
|
||||||
}
|
|
||||||
|
|
||||||
// Evictor retrieves evictor so plugins can filter and evict pods
|
|
||||||
func (hi *handleImpl) Evictor() framework.Evictor {
|
|
||||||
return hi.evictor
|
|
||||||
}
|
|
||||||
|
|
||||||
func RunDeschedulerStrategies(ctx context.Context, rs *options.DeschedulerServer, deschedulerPolicy *api.DeschedulerPolicy, evictionPolicyGroupVersion string) error {
|
|
||||||
sharedInformerFactory := informers.NewSharedInformerFactory(rs.Client, 0)
|
sharedInformerFactory := informers.NewSharedInformerFactory(rs.Client, 0)
|
||||||
nodeInformer := sharedInformerFactory.Core().V1().Nodes()
|
nodeInformer := sharedInformerFactory.Core().V1().Nodes()
|
||||||
podInformer := sharedInformerFactory.Core().V1().Pods()
|
|
||||||
namespaceInformer := sharedInformerFactory.Core().V1().Namespaces()
|
|
||||||
priorityClassInformer := sharedInformerFactory.Scheduling().V1().PriorityClasses()
|
|
||||||
|
|
||||||
ctx, cancel := context.WithCancel(ctx)
|
sharedInformerFactory.Start(stopChannel)
|
||||||
defer cancel()
|
sharedInformerFactory.WaitForCacheSync(stopChannel)
|
||||||
|
|
||||||
// create the informers
|
|
||||||
namespaceInformer.Informer()
|
|
||||||
priorityClassInformer.Informer()
|
|
||||||
|
|
||||||
getPodsAssignedToNode, err := podutil.BuildGetPodsAssignedToNodeFunc(podInformer)
|
|
||||||
if err != nil {
|
|
||||||
return fmt.Errorf("build get pods assigned to node function error: %v", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
sharedInformerFactory.Start(ctx.Done())
|
|
||||||
sharedInformerFactory.WaitForCacheSync(ctx.Done())
|
|
||||||
|
|
||||||
strategyFuncs := map[api.StrategyName]strategyFunction{
|
strategyFuncs := map[api.StrategyName]strategyFunction{
|
||||||
"RemoveDuplicates": nil,
|
"RemoveDuplicates": strategies.RemoveDuplicatePods,
|
||||||
"LowNodeUtilization": nil,
|
"LowNodeUtilization": nodeutilization.LowNodeUtilization,
|
||||||
"HighNodeUtilization": nil,
|
"HighNodeUtilization": nodeutilization.HighNodeUtilization,
|
||||||
"RemovePodsViolatingInterPodAntiAffinity": nil,
|
"RemovePodsViolatingInterPodAntiAffinity": strategies.RemovePodsViolatingInterPodAntiAffinity,
|
||||||
"RemovePodsViolatingNodeAffinity": nil,
|
"RemovePodsViolatingNodeAffinity": strategies.RemovePodsViolatingNodeAffinity,
|
||||||
"RemovePodsViolatingNodeTaints": nil,
|
"RemovePodsViolatingNodeTaints": strategies.RemovePodsViolatingNodeTaints,
|
||||||
"RemovePodsHavingTooManyRestarts": nil,
|
"RemovePodsHavingTooManyRestarts": strategies.RemovePodsHavingTooManyRestarts,
|
||||||
"PodLifeTime": nil,
|
"PodLifeTime": strategies.PodLifeTime,
|
||||||
"RemovePodsViolatingTopologySpreadConstraint": nil,
|
"RemovePodsViolatingTopologySpreadConstraint": strategies.RemovePodsViolatingTopologySpreadConstraint,
|
||||||
"RemoveFailedPods": nil,
|
"RemoveFailedPods": strategies.RemoveFailedPods,
|
||||||
}
|
}
|
||||||
|
|
||||||
var nodeSelector string
|
nodeSelector := rs.NodeSelector
|
||||||
if deschedulerPolicy.NodeSelector != nil {
|
if deschedulerPolicy.NodeSelector != nil {
|
||||||
nodeSelector = *deschedulerPolicy.NodeSelector
|
nodeSelector = *deschedulerPolicy.NodeSelector
|
||||||
}
|
}
|
||||||
|
|
||||||
var evictLocalStoragePods bool
|
evictLocalStoragePods := rs.EvictLocalStoragePods
|
||||||
if deschedulerPolicy.EvictLocalStoragePods != nil {
|
if deschedulerPolicy.EvictLocalStoragePods != nil {
|
||||||
evictLocalStoragePods = *deschedulerPolicy.EvictLocalStoragePods
|
evictLocalStoragePods = *deschedulerPolicy.EvictLocalStoragePods
|
||||||
}
|
}
|
||||||
|
|
||||||
evictBarePods := false
|
|
||||||
if deschedulerPolicy.EvictFailedBarePods != nil {
|
|
||||||
evictBarePods = *deschedulerPolicy.EvictFailedBarePods
|
|
||||||
if evictBarePods {
|
|
||||||
klog.V(1).InfoS("Warning: EvictFailedBarePods is set to True. This could cause eviction of pods without ownerReferences.")
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
evictSystemCriticalPods := false
|
evictSystemCriticalPods := false
|
||||||
if deschedulerPolicy.EvictSystemCriticalPods != nil {
|
if deschedulerPolicy.EvictSystemCriticalPods != nil {
|
||||||
evictSystemCriticalPods = *deschedulerPolicy.EvictSystemCriticalPods
|
evictSystemCriticalPods = *deschedulerPolicy.EvictSystemCriticalPods
|
||||||
@@ -286,137 +109,40 @@ func RunDeschedulerStrategies(ctx context.Context, rs *options.DeschedulerServer
|
|||||||
ignorePvcPods = *deschedulerPolicy.IgnorePVCPods
|
ignorePvcPods = *deschedulerPolicy.IgnorePVCPods
|
||||||
}
|
}
|
||||||
|
|
||||||
var eventClient clientset.Interface
|
maxNoOfPodsToEvictPerNode := rs.MaxNoOfPodsToEvictPerNode
|
||||||
if rs.DryRun {
|
if deschedulerPolicy.MaxNoOfPodsToEvictPerNode != nil {
|
||||||
eventClient = fakeclientset.NewSimpleClientset()
|
maxNoOfPodsToEvictPerNode = *deschedulerPolicy.MaxNoOfPodsToEvictPerNode
|
||||||
} else {
|
|
||||||
eventClient = rs.Client
|
|
||||||
}
|
}
|
||||||
|
|
||||||
eventBroadcaster, eventRecorder := utils.GetRecorderAndBroadcaster(ctx, eventClient)
|
wait.Until(func() {
|
||||||
defer eventBroadcaster.Shutdown()
|
|
||||||
|
|
||||||
wait.NonSlidingUntil(func() {
|
|
||||||
nodes, err := nodeutil.ReadyNodes(ctx, rs.Client, nodeInformer, nodeSelector)
|
nodes, err := nodeutil.ReadyNodes(ctx, rs.Client, nodeInformer, nodeSelector)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
klog.V(1).InfoS("Unable to get ready nodes", "err", err)
|
klog.V(1).InfoS("Unable to get ready nodes", "err", err)
|
||||||
cancel()
|
close(stopChannel)
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
if len(nodes) <= 1 {
|
if len(nodes) <= 1 {
|
||||||
klog.V(1).InfoS("The cluster size is 0 or 1 meaning eviction causes service disruption or degradation. So aborting..")
|
klog.V(1).InfoS("The cluster size is 0 or 1 meaning eviction causes service disruption or degradation. So aborting..")
|
||||||
cancel()
|
close(stopChannel)
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
var podEvictorClient clientset.Interface
|
|
||||||
// When the dry mode is enable, collect all the relevant objects (mostly pods) under a fake client.
|
|
||||||
// So when evicting pods while running multiple strategies in a row have the cummulative effect
|
|
||||||
// as is when evicting pods for real.
|
|
||||||
if rs.DryRun {
|
|
||||||
klog.V(3).Infof("Building a cached client from the cluster for the dry run")
|
|
||||||
// Create a new cache so we start from scratch without any leftovers
|
|
||||||
fakeClient, err := cachedClient(rs.Client, podInformer, nodeInformer, namespaceInformer, priorityClassInformer)
|
|
||||||
if err != nil {
|
|
||||||
klog.Error(err)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
fakeSharedInformerFactory := informers.NewSharedInformerFactory(fakeClient, 0)
|
|
||||||
getPodsAssignedToNode, err = podutil.BuildGetPodsAssignedToNodeFunc(fakeSharedInformerFactory.Core().V1().Pods())
|
|
||||||
if err != nil {
|
|
||||||
klog.Errorf("build get pods assigned to node function error: %v", err)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
fakeCtx, cncl := context.WithCancel(context.TODO())
|
|
||||||
defer cncl()
|
|
||||||
fakeSharedInformerFactory.Start(fakeCtx.Done())
|
|
||||||
fakeSharedInformerFactory.WaitForCacheSync(fakeCtx.Done())
|
|
||||||
|
|
||||||
podEvictorClient = fakeClient
|
|
||||||
} else {
|
|
||||||
podEvictorClient = rs.Client
|
|
||||||
}
|
|
||||||
|
|
||||||
klog.V(3).Infof("Building a pod evictor")
|
|
||||||
podEvictor := evictions.NewPodEvictor(
|
podEvictor := evictions.NewPodEvictor(
|
||||||
podEvictorClient,
|
rs.Client,
|
||||||
evictionPolicyGroupVersion,
|
evictionPolicyGroupVersion,
|
||||||
rs.DryRun,
|
rs.DryRun,
|
||||||
deschedulerPolicy.MaxNoOfPodsToEvictPerNode,
|
maxNoOfPodsToEvictPerNode,
|
||||||
deschedulerPolicy.MaxNoOfPodsToEvictPerNamespace,
|
|
||||||
nodes,
|
nodes,
|
||||||
!rs.DisableMetrics,
|
evictLocalStoragePods,
|
||||||
eventRecorder,
|
evictSystemCriticalPods,
|
||||||
|
ignorePvcPods,
|
||||||
)
|
)
|
||||||
|
|
||||||
for name, strategy := range deschedulerPolicy.Strategies {
|
for name, strategy := range deschedulerPolicy.Strategies {
|
||||||
if f, ok := strategyFuncs[name]; ok {
|
if f, ok := strategyFuncs[name]; ok {
|
||||||
if strategy.Enabled {
|
if strategy.Enabled {
|
||||||
params := strategy.Params
|
f(ctx, rs.Client, strategy, nodes, podEvictor)
|
||||||
if params == nil {
|
|
||||||
params = &api.StrategyParameters{}
|
|
||||||
}
|
|
||||||
|
|
||||||
nodeFit := false
|
|
||||||
if name != "PodLifeTime" {
|
|
||||||
nodeFit = params.NodeFit
|
|
||||||
}
|
|
||||||
|
|
||||||
// TODO(jchaloup): once all strategies are migrated move this check under
|
|
||||||
// the default evictor args validation
|
|
||||||
if params.ThresholdPriority != nil && params.ThresholdPriorityClassName != "" {
|
|
||||||
klog.V(1).ErrorS(fmt.Errorf("priority threshold misconfigured"), "only one of priorityThreshold fields can be set", "pluginName", name)
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
thresholdPriority, err := utils.GetPriorityFromStrategyParams(ctx, rs.Client, strategy.Params)
|
|
||||||
if err != nil {
|
|
||||||
klog.ErrorS(err, "Failed to get threshold priority from strategy's params")
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
|
|
||||||
defaultevictorArgs := &defaultevictor.DefaultEvictorArgs{
|
|
||||||
EvictLocalStoragePods: evictLocalStoragePods,
|
|
||||||
EvictSystemCriticalPods: evictSystemCriticalPods,
|
|
||||||
IgnorePvcPods: ignorePvcPods,
|
|
||||||
EvictFailedBarePods: evictBarePods,
|
|
||||||
NodeFit: nodeFit,
|
|
||||||
PriorityThreshold: &api.PriorityThreshold{
|
|
||||||
Value: &thresholdPriority,
|
|
||||||
},
|
|
||||||
}
|
|
||||||
|
|
||||||
evictorFilter, _ := defaultevictor.New(
|
|
||||||
defaultevictorArgs,
|
|
||||||
&handleImpl{
|
|
||||||
clientSet: rs.Client,
|
|
||||||
getPodsAssignedToNodeFunc: getPodsAssignedToNode,
|
|
||||||
sharedInformerFactory: sharedInformerFactory,
|
|
||||||
},
|
|
||||||
)
|
|
||||||
|
|
||||||
handle := &handleImpl{
|
|
||||||
clientSet: rs.Client,
|
|
||||||
getPodsAssignedToNodeFunc: getPodsAssignedToNode,
|
|
||||||
sharedInformerFactory: sharedInformerFactory,
|
|
||||||
evictor: &evictorImpl{
|
|
||||||
podEvictor: podEvictor,
|
|
||||||
evictorFilter: evictorFilter.(framework.EvictorPlugin),
|
|
||||||
},
|
|
||||||
}
|
|
||||||
|
|
||||||
// TODO: strategyName should be accessible from within the strategy using a framework
|
|
||||||
// handle or function which the Evictor has access to. For migration/in-progress framework
|
|
||||||
// work, we are currently passing this via context. To be removed
|
|
||||||
// (See discussion thread https://github.com/kubernetes-sigs/descheduler/pull/885#discussion_r919962292)
|
|
||||||
childCtx := context.WithValue(ctx, "strategyName", string(name))
|
|
||||||
if pgFnc, exists := pluginsMap[string(name)]; exists {
|
|
||||||
pgFnc(childCtx, nodes, params, handle)
|
|
||||||
} else {
|
|
||||||
f(childCtx, rs.Client, strategy, nodes, podEvictor, evictorFilter.(framework.EvictorPlugin), getPodsAssignedToNode)
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
klog.ErrorS(fmt.Errorf("unknown strategy name"), "skipping strategy", "strategy", name)
|
klog.ErrorS(fmt.Errorf("unknown strategy name"), "skipping strategy", "strategy", name)
|
||||||
@@ -427,23 +153,9 @@ func RunDeschedulerStrategies(ctx context.Context, rs *options.DeschedulerServer
|
|||||||
|
|
||||||
// If there was no interval specified, send a signal to the stopChannel to end the wait.Until loop after 1 iteration
|
// If there was no interval specified, send a signal to the stopChannel to end the wait.Until loop after 1 iteration
|
||||||
if rs.DeschedulingInterval.Seconds() == 0 {
|
if rs.DeschedulingInterval.Seconds() == 0 {
|
||||||
cancel()
|
close(stopChannel)
|
||||||
}
|
}
|
||||||
}, rs.DeschedulingInterval, ctx.Done())
|
}, rs.DeschedulingInterval, stopChannel)
|
||||||
|
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func createClients(kubeconfig string) (clientset.Interface, clientset.Interface, error) {
|
|
||||||
kClient, err := client.CreateClient(kubeconfig, "descheduler")
|
|
||||||
if err != nil {
|
|
||||||
return nil, nil, err
|
|
||||||
}
|
|
||||||
|
|
||||||
eventClient, err := client.CreateClient(kubeconfig, "")
|
|
||||||
if err != nil {
|
|
||||||
return nil, nil, err
|
|
||||||
}
|
|
||||||
|
|
||||||
return kClient, eventClient, nil
|
|
||||||
}
|
|
||||||
|
|||||||
@@ -3,15 +3,14 @@ package descheduler
|
|||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
"fmt"
|
"fmt"
|
||||||
|
"strings"
|
||||||
"testing"
|
"testing"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
v1 "k8s.io/api/core/v1"
|
v1 "k8s.io/api/core/v1"
|
||||||
policy "k8s.io/api/policy/v1"
|
|
||||||
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
||||||
"k8s.io/apimachinery/pkg/runtime"
|
"k8s.io/apimachinery/pkg/util/wait"
|
||||||
fakeclientset "k8s.io/client-go/kubernetes/fake"
|
fakeclientset "k8s.io/client-go/kubernetes/fake"
|
||||||
core "k8s.io/client-go/testing"
|
|
||||||
"sigs.k8s.io/descheduler/cmd/descheduler/app/options"
|
"sigs.k8s.io/descheduler/cmd/descheduler/app/options"
|
||||||
"sigs.k8s.io/descheduler/pkg/api"
|
"sigs.k8s.io/descheduler/pkg/api"
|
||||||
"sigs.k8s.io/descheduler/test"
|
"sigs.k8s.io/descheduler/test"
|
||||||
@@ -28,7 +27,6 @@ func TestTaintsUpdated(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
client := fakeclientset.NewSimpleClientset(n1, n2, p1)
|
client := fakeclientset.NewSimpleClientset(n1, n2, p1)
|
||||||
eventClient := fakeclientset.NewSimpleClientset(n1, n2, p1)
|
|
||||||
dp := &api.DeschedulerPolicy{
|
dp := &api.DeschedulerPolicy{
|
||||||
Strategies: api.StrategyList{
|
Strategies: api.StrategyList{
|
||||||
"RemovePodsViolatingNodeTaints": api.DeschedulerStrategy{
|
"RemovePodsViolatingNodeTaints": api.DeschedulerStrategy{
|
||||||
@@ -37,12 +35,29 @@ func TestTaintsUpdated(t *testing.T) {
|
|||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
|
stopChannel := make(chan struct{})
|
||||||
|
defer close(stopChannel)
|
||||||
|
|
||||||
rs, err := options.NewDeschedulerServer()
|
rs, err := options.NewDeschedulerServer()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
t.Fatalf("Unable to initialize server: %v", err)
|
t.Fatalf("Unable to initialize server: %v", err)
|
||||||
}
|
}
|
||||||
rs.Client = client
|
rs.Client = client
|
||||||
rs.EventClient = eventClient
|
rs.DeschedulingInterval = 100 * time.Millisecond
|
||||||
|
errChan := make(chan error, 1)
|
||||||
|
defer close(errChan)
|
||||||
|
go func() {
|
||||||
|
err := RunDeschedulerStrategies(ctx, rs, dp, "v1beta1", stopChannel)
|
||||||
|
errChan <- err
|
||||||
|
}()
|
||||||
|
select {
|
||||||
|
case err := <-errChan:
|
||||||
|
if err != nil {
|
||||||
|
t.Fatalf("Unable to run descheduler strategies: %v", err)
|
||||||
|
}
|
||||||
|
case <-time.After(300 * time.Millisecond):
|
||||||
|
// Wait for few cycles and then verify the only pod still exists
|
||||||
|
}
|
||||||
|
|
||||||
pods, err := client.CoreV1().Pods(p1.Namespace).List(ctx, metav1.ListOptions{})
|
pods, err := client.CoreV1().Pods(p1.Namespace).List(ctx, metav1.ListOptions{})
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@@ -65,154 +80,24 @@ func TestTaintsUpdated(t *testing.T) {
|
|||||||
t.Fatalf("Unable to update node: %v\n", err)
|
t.Fatalf("Unable to update node: %v\n", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
var evictedPods []string
|
if err := wait.PollImmediate(100*time.Millisecond, time.Second, func() (bool, error) {
|
||||||
client.PrependReactor("create", "pods", podEvictionReactionFuc(&evictedPods))
|
// Get over evicted pod result in panic
|
||||||
|
//pods, err := client.CoreV1().Pods(p1.Namespace).Get(p1.Name, metav1.GetOptions{})
|
||||||
if err := RunDeschedulerStrategies(ctx, rs, dp, "v1"); err != nil {
|
// List is better, it does not panic.
|
||||||
t.Fatalf("Unable to run descheduler strategies: %v", err)
|
// Though once the pod is evicted, List starts to error with "can't assign or convert v1beta1.Eviction into v1.Pod"
|
||||||
}
|
pods, err := client.CoreV1().Pods(p1.Namespace).List(ctx, metav1.ListOptions{})
|
||||||
|
if err == nil {
|
||||||
if len(evictedPods) != 1 {
|
if len(pods.Items) > 0 {
|
||||||
t.Fatalf("Unable to evict pod, node taint did not get propagated to descheduler strategies %v\n", err)
|
return false, nil
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestDuplicate(t *testing.T) {
|
|
||||||
ctx := context.Background()
|
|
||||||
node1 := test.BuildTestNode("n1", 2000, 3000, 10, nil)
|
|
||||||
node2 := test.BuildTestNode("n2", 2000, 3000, 10, nil)
|
|
||||||
|
|
||||||
p1 := test.BuildTestPod("p1", 100, 0, node1.Name, nil)
|
|
||||||
p1.Namespace = "dev"
|
|
||||||
p2 := test.BuildTestPod("p2", 100, 0, node1.Name, nil)
|
|
||||||
p2.Namespace = "dev"
|
|
||||||
p3 := test.BuildTestPod("p3", 100, 0, node1.Name, nil)
|
|
||||||
p3.Namespace = "dev"
|
|
||||||
|
|
||||||
ownerRef1 := test.GetReplicaSetOwnerRefList()
|
|
||||||
p1.ObjectMeta.OwnerReferences = ownerRef1
|
|
||||||
p2.ObjectMeta.OwnerReferences = ownerRef1
|
|
||||||
p3.ObjectMeta.OwnerReferences = ownerRef1
|
|
||||||
|
|
||||||
client := fakeclientset.NewSimpleClientset(node1, node2, p1, p2, p3)
|
|
||||||
eventClient := fakeclientset.NewSimpleClientset(node1, node2, p1, p2, p3)
|
|
||||||
dp := &api.DeschedulerPolicy{
|
|
||||||
Strategies: api.StrategyList{
|
|
||||||
"RemoveDuplicates": api.DeschedulerStrategy{
|
|
||||||
Enabled: true,
|
|
||||||
},
|
|
||||||
},
|
|
||||||
}
|
|
||||||
|
|
||||||
rs, err := options.NewDeschedulerServer()
|
|
||||||
if err != nil {
|
|
||||||
t.Fatalf("Unable to initialize server: %v", err)
|
|
||||||
}
|
|
||||||
rs.Client = client
|
|
||||||
rs.EventClient = eventClient
|
|
||||||
|
|
||||||
pods, err := client.CoreV1().Pods(p1.Namespace).List(ctx, metav1.ListOptions{})
|
|
||||||
if err != nil {
|
|
||||||
t.Errorf("Unable to list pods: %v", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
if len(pods.Items) != 3 {
|
|
||||||
t.Errorf("Pods number should be 3 before evict")
|
|
||||||
}
|
|
||||||
|
|
||||||
var evictedPods []string
|
|
||||||
client.PrependReactor("create", "pods", podEvictionReactionFuc(&evictedPods))
|
|
||||||
|
|
||||||
if err := RunDeschedulerStrategies(ctx, rs, dp, "v1"); err != nil {
|
|
||||||
t.Fatalf("Unable to run descheduler strategies: %v", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
if len(evictedPods) == 0 {
|
|
||||||
t.Fatalf("Unable to evict pod, node taint did not get propagated to descheduler strategies %v\n", err)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestRootCancel(t *testing.T) {
|
|
||||||
ctx, cancel := context.WithCancel(context.Background())
|
|
||||||
n1 := test.BuildTestNode("n1", 2000, 3000, 10, nil)
|
|
||||||
n2 := test.BuildTestNode("n2", 2000, 3000, 10, nil)
|
|
||||||
client := fakeclientset.NewSimpleClientset(n1, n2)
|
|
||||||
eventClient := fakeclientset.NewSimpleClientset(n1, n2)
|
|
||||||
dp := &api.DeschedulerPolicy{
|
|
||||||
Strategies: api.StrategyList{}, // no strategies needed for this test
|
|
||||||
}
|
|
||||||
|
|
||||||
rs, err := options.NewDeschedulerServer()
|
|
||||||
if err != nil {
|
|
||||||
t.Fatalf("Unable to initialize server: %v", err)
|
|
||||||
}
|
|
||||||
rs.Client = client
|
|
||||||
rs.EventClient = eventClient
|
|
||||||
rs.DeschedulingInterval = 100 * time.Millisecond
|
|
||||||
errChan := make(chan error, 1)
|
|
||||||
defer close(errChan)
|
|
||||||
|
|
||||||
go func() {
|
|
||||||
err := RunDeschedulerStrategies(ctx, rs, dp, "v1")
|
|
||||||
errChan <- err
|
|
||||||
}()
|
|
||||||
cancel()
|
|
||||||
select {
|
|
||||||
case err := <-errChan:
|
|
||||||
if err != nil {
|
|
||||||
t.Fatalf("Unable to run descheduler strategies: %v", err)
|
|
||||||
}
|
|
||||||
case <-time.After(1 * time.Second):
|
|
||||||
t.Fatal("Root ctx should have canceled immediately")
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestRootCancelWithNoInterval(t *testing.T) {
|
|
||||||
ctx, cancel := context.WithCancel(context.Background())
|
|
||||||
n1 := test.BuildTestNode("n1", 2000, 3000, 10, nil)
|
|
||||||
n2 := test.BuildTestNode("n2", 2000, 3000, 10, nil)
|
|
||||||
client := fakeclientset.NewSimpleClientset(n1, n2)
|
|
||||||
eventClient := fakeclientset.NewSimpleClientset(n1, n2)
|
|
||||||
dp := &api.DeschedulerPolicy{
|
|
||||||
Strategies: api.StrategyList{}, // no strategies needed for this test
|
|
||||||
}
|
|
||||||
|
|
||||||
rs, err := options.NewDeschedulerServer()
|
|
||||||
if err != nil {
|
|
||||||
t.Fatalf("Unable to initialize server: %v", err)
|
|
||||||
}
|
|
||||||
rs.Client = client
|
|
||||||
rs.EventClient = eventClient
|
|
||||||
rs.DeschedulingInterval = 0
|
|
||||||
errChan := make(chan error, 1)
|
|
||||||
defer close(errChan)
|
|
||||||
|
|
||||||
go func() {
|
|
||||||
err := RunDeschedulerStrategies(ctx, rs, dp, "v1")
|
|
||||||
errChan <- err
|
|
||||||
}()
|
|
||||||
cancel()
|
|
||||||
select {
|
|
||||||
case err := <-errChan:
|
|
||||||
if err != nil {
|
|
||||||
t.Fatalf("Unable to run descheduler strategies: %v", err)
|
|
||||||
}
|
|
||||||
case <-time.After(1 * time.Second):
|
|
||||||
t.Fatal("Root ctx should have canceled immediately")
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func podEvictionReactionFuc(evictedPods *[]string) func(action core.Action) (bool, runtime.Object, error) {
|
|
||||||
return func(action core.Action) (bool, runtime.Object, error) {
|
|
||||||
if action.GetSubresource() == "eviction" {
|
|
||||||
createAct, matched := action.(core.CreateActionImpl)
|
|
||||||
if !matched {
|
|
||||||
return false, nil, fmt.Errorf("unable to convert action to core.CreateActionImpl")
|
|
||||||
}
|
|
||||||
if eviction, matched := createAct.Object.(*policy.Eviction); matched {
|
|
||||||
*evictedPods = append(*evictedPods, eviction.GetName())
|
|
||||||
}
|
}
|
||||||
|
return true, nil
|
||||||
}
|
}
|
||||||
return false, nil, nil // fallback to the default reactor
|
if strings.Contains(err.Error(), "can't assign or convert v1beta1.Eviction into v1.Pod") {
|
||||||
|
return true, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
return false, nil
|
||||||
|
}); err != nil {
|
||||||
|
t.Fatalf("Unable to evict pod, node taint did not get propagated to descheduler strategies")
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -19,158 +19,129 @@ package evictions
|
|||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
"fmt"
|
"fmt"
|
||||||
|
"strings"
|
||||||
|
|
||||||
v1 "k8s.io/api/core/v1"
|
v1 "k8s.io/api/core/v1"
|
||||||
policy "k8s.io/api/policy/v1"
|
policy "k8s.io/api/policy/v1beta1"
|
||||||
apierrors "k8s.io/apimachinery/pkg/api/errors"
|
apierrors "k8s.io/apimachinery/pkg/api/errors"
|
||||||
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
||||||
|
"k8s.io/apimachinery/pkg/labels"
|
||||||
|
"k8s.io/apimachinery/pkg/util/errors"
|
||||||
clientset "k8s.io/client-go/kubernetes"
|
clientset "k8s.io/client-go/kubernetes"
|
||||||
"k8s.io/client-go/tools/events"
|
"k8s.io/client-go/kubernetes/scheme"
|
||||||
|
clientcorev1 "k8s.io/client-go/kubernetes/typed/core/v1"
|
||||||
|
"k8s.io/client-go/tools/record"
|
||||||
"k8s.io/klog/v2"
|
"k8s.io/klog/v2"
|
||||||
"sigs.k8s.io/descheduler/metrics"
|
"sigs.k8s.io/descheduler/metrics"
|
||||||
|
nodeutil "sigs.k8s.io/descheduler/pkg/descheduler/node"
|
||||||
|
podutil "sigs.k8s.io/descheduler/pkg/descheduler/pod"
|
||||||
|
"sigs.k8s.io/descheduler/pkg/utils"
|
||||||
|
|
||||||
eutils "sigs.k8s.io/descheduler/pkg/descheduler/evictions/utils"
|
eutils "sigs.k8s.io/descheduler/pkg/descheduler/evictions/utils"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
const (
|
||||||
|
evictPodAnnotationKey = "descheduler.alpha.kubernetes.io/evict"
|
||||||
|
)
|
||||||
|
|
||||||
// nodePodEvictedCount keeps count of pods evicted on node
|
// nodePodEvictedCount keeps count of pods evicted on node
|
||||||
type nodePodEvictedCount map[string]uint
|
type nodePodEvictedCount map[*v1.Node]int
|
||||||
type namespacePodEvictCount map[string]uint
|
|
||||||
|
|
||||||
type PodEvictor struct {
|
type PodEvictor struct {
|
||||||
client clientset.Interface
|
client clientset.Interface
|
||||||
nodes []*v1.Node
|
nodes []*v1.Node
|
||||||
policyGroupVersion string
|
policyGroupVersion string
|
||||||
dryRun bool
|
dryRun bool
|
||||||
maxPodsToEvictPerNode *uint
|
maxPodsToEvictPerNode int
|
||||||
maxPodsToEvictPerNamespace *uint
|
nodepodCount nodePodEvictedCount
|
||||||
nodepodCount nodePodEvictedCount
|
evictLocalStoragePods bool
|
||||||
namespacePodCount namespacePodEvictCount
|
evictSystemCriticalPods bool
|
||||||
metricsEnabled bool
|
ignorePvcPods bool
|
||||||
eventRecorder events.EventRecorder
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func NewPodEvictor(
|
func NewPodEvictor(
|
||||||
client clientset.Interface,
|
client clientset.Interface,
|
||||||
policyGroupVersion string,
|
policyGroupVersion string,
|
||||||
dryRun bool,
|
dryRun bool,
|
||||||
maxPodsToEvictPerNode *uint,
|
maxPodsToEvictPerNode int,
|
||||||
maxPodsToEvictPerNamespace *uint,
|
|
||||||
nodes []*v1.Node,
|
nodes []*v1.Node,
|
||||||
metricsEnabled bool,
|
evictLocalStoragePods bool,
|
||||||
eventRecorder events.EventRecorder,
|
evictSystemCriticalPods bool,
|
||||||
|
ignorePvcPods bool,
|
||||||
) *PodEvictor {
|
) *PodEvictor {
|
||||||
var nodePodCount = make(nodePodEvictedCount)
|
var nodePodCount = make(nodePodEvictedCount)
|
||||||
var namespacePodCount = make(namespacePodEvictCount)
|
|
||||||
for _, node := range nodes {
|
for _, node := range nodes {
|
||||||
// Initialize podsEvicted till now with 0.
|
// Initialize podsEvicted till now with 0.
|
||||||
nodePodCount[node.Name] = 0
|
nodePodCount[node] = 0
|
||||||
}
|
}
|
||||||
|
|
||||||
return &PodEvictor{
|
return &PodEvictor{
|
||||||
client: client,
|
client: client,
|
||||||
nodes: nodes,
|
nodes: nodes,
|
||||||
policyGroupVersion: policyGroupVersion,
|
policyGroupVersion: policyGroupVersion,
|
||||||
dryRun: dryRun,
|
dryRun: dryRun,
|
||||||
maxPodsToEvictPerNode: maxPodsToEvictPerNode,
|
maxPodsToEvictPerNode: maxPodsToEvictPerNode,
|
||||||
maxPodsToEvictPerNamespace: maxPodsToEvictPerNamespace,
|
nodepodCount: nodePodCount,
|
||||||
nodepodCount: nodePodCount,
|
evictLocalStoragePods: evictLocalStoragePods,
|
||||||
namespacePodCount: namespacePodCount,
|
evictSystemCriticalPods: evictSystemCriticalPods,
|
||||||
metricsEnabled: metricsEnabled,
|
ignorePvcPods: ignorePvcPods,
|
||||||
eventRecorder: eventRecorder,
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// NodeEvicted gives a number of pods evicted for node
|
// NodeEvicted gives a number of pods evicted for node
|
||||||
func (pe *PodEvictor) NodeEvicted(node *v1.Node) uint {
|
func (pe *PodEvictor) NodeEvicted(node *v1.Node) int {
|
||||||
return pe.nodepodCount[node.Name]
|
return pe.nodepodCount[node]
|
||||||
}
|
}
|
||||||
|
|
||||||
// TotalEvicted gives a number of pods evicted through all nodes
|
// TotalEvicted gives a number of pods evicted through all nodes
|
||||||
func (pe *PodEvictor) TotalEvicted() uint {
|
func (pe *PodEvictor) TotalEvicted() int {
|
||||||
var total uint
|
var total int
|
||||||
for _, count := range pe.nodepodCount {
|
for _, count := range pe.nodepodCount {
|
||||||
total += count
|
total += count
|
||||||
}
|
}
|
||||||
return total
|
return total
|
||||||
}
|
}
|
||||||
|
|
||||||
// NodeLimitExceeded checks if the number of evictions for a node was exceeded
|
// EvictPod returns non-nil error only when evicting a pod on a node is not
|
||||||
func (pe *PodEvictor) NodeLimitExceeded(node *v1.Node) bool {
|
// possible (due to maxPodsToEvictPerNode constraint). Success is true when the pod
|
||||||
if pe.maxPodsToEvictPerNode != nil {
|
// is evicted on the server side.
|
||||||
return pe.nodepodCount[node.Name] == *pe.maxPodsToEvictPerNode
|
func (pe *PodEvictor) EvictPod(ctx context.Context, pod *v1.Pod, node *v1.Node, strategy string, reasons ...string) (bool, error) {
|
||||||
|
reason := strategy
|
||||||
|
if len(reasons) > 0 {
|
||||||
|
reason += " (" + strings.Join(reasons, ", ") + ")"
|
||||||
}
|
}
|
||||||
return false
|
if pe.maxPodsToEvictPerNode > 0 && pe.nodepodCount[node]+1 > pe.maxPodsToEvictPerNode {
|
||||||
}
|
metrics.PodsEvicted.With(map[string]string{"result": "maximum number reached", "strategy": strategy, "namespace": pod.Namespace}).Inc()
|
||||||
|
return false, fmt.Errorf("Maximum number %v of evicted pods per %q node reached", pe.maxPodsToEvictPerNode, node.Name)
|
||||||
// EvictOptions provides a handle for passing additional info to EvictPod
|
|
||||||
type EvictOptions struct {
|
|
||||||
// Reason allows for passing details about the specific eviction for logging.
|
|
||||||
Reason string
|
|
||||||
}
|
|
||||||
|
|
||||||
// EvictPod evicts a pod while exercising eviction limits.
|
|
||||||
// Returns true when the pod is evicted on the server side.
|
|
||||||
func (pe *PodEvictor) EvictPod(ctx context.Context, pod *v1.Pod, opts EvictOptions) bool {
|
|
||||||
// TODO: Replace context-propagated Strategy name with a defined framework handle for accessing Strategy info
|
|
||||||
strategy := ""
|
|
||||||
if ctx.Value("strategyName") != nil {
|
|
||||||
strategy = ctx.Value("strategyName").(string)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
if pod.Spec.NodeName != "" {
|
err := evictPod(ctx, pe.client, pod, pe.policyGroupVersion, pe.dryRun)
|
||||||
if pe.maxPodsToEvictPerNode != nil && pe.nodepodCount[pod.Spec.NodeName]+1 > *pe.maxPodsToEvictPerNode {
|
|
||||||
if pe.metricsEnabled {
|
|
||||||
metrics.PodsEvicted.With(map[string]string{"result": "maximum number of pods per node reached", "strategy": strategy, "namespace": pod.Namespace, "node": pod.Spec.NodeName}).Inc()
|
|
||||||
}
|
|
||||||
klog.ErrorS(fmt.Errorf("Maximum number of evicted pods per node reached"), "limit", *pe.maxPodsToEvictPerNode, "node", pod.Spec.NodeName)
|
|
||||||
return false
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
if pe.maxPodsToEvictPerNamespace != nil && pe.namespacePodCount[pod.Namespace]+1 > *pe.maxPodsToEvictPerNamespace {
|
|
||||||
if pe.metricsEnabled {
|
|
||||||
metrics.PodsEvicted.With(map[string]string{"result": "maximum number of pods per namespace reached", "strategy": strategy, "namespace": pod.Namespace, "node": pod.Spec.NodeName}).Inc()
|
|
||||||
}
|
|
||||||
klog.ErrorS(fmt.Errorf("Maximum number of evicted pods per namespace reached"), "limit", *pe.maxPodsToEvictPerNamespace, "namespace", pod.Namespace)
|
|
||||||
return false
|
|
||||||
}
|
|
||||||
|
|
||||||
err := evictPod(ctx, pe.client, pod, pe.policyGroupVersion)
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
// err is used only for logging purposes
|
// err is used only for logging purposes
|
||||||
klog.ErrorS(err, "Error evicting pod", "pod", klog.KObj(pod), "reason", opts.Reason)
|
klog.ErrorS(err, "Error evicting pod", "pod", klog.KObj(pod), "reason", reason)
|
||||||
if pe.metricsEnabled {
|
metrics.PodsEvicted.With(map[string]string{"result": "error", "strategy": strategy, "namespace": pod.Namespace}).Inc()
|
||||||
metrics.PodsEvicted.With(map[string]string{"result": "error", "strategy": strategy, "namespace": pod.Namespace, "node": pod.Spec.NodeName}).Inc()
|
return false, nil
|
||||||
}
|
|
||||||
return false
|
|
||||||
}
|
|
||||||
|
|
||||||
if pod.Spec.NodeName != "" {
|
|
||||||
pe.nodepodCount[pod.Spec.NodeName]++
|
|
||||||
}
|
|
||||||
pe.namespacePodCount[pod.Namespace]++
|
|
||||||
|
|
||||||
if pe.metricsEnabled {
|
|
||||||
metrics.PodsEvicted.With(map[string]string{"result": "success", "strategy": strategy, "namespace": pod.Namespace, "node": pod.Spec.NodeName}).Inc()
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
pe.nodepodCount[node]++
|
||||||
if pe.dryRun {
|
if pe.dryRun {
|
||||||
klog.V(1).InfoS("Evicted pod in dry run mode", "pod", klog.KObj(pod), "reason", opts.Reason, "strategy", strategy, "node", pod.Spec.NodeName)
|
klog.V(1).InfoS("Evicted pod in dry run mode", "pod", klog.KObj(pod), "reason", reason)
|
||||||
} else {
|
} else {
|
||||||
klog.V(1).InfoS("Evicted pod", "pod", klog.KObj(pod), "reason", opts.Reason, "strategy", strategy, "node", pod.Spec.NodeName)
|
klog.V(1).InfoS("Evicted pod", "pod", klog.KObj(pod), "reason", reason)
|
||||||
reason := opts.Reason
|
eventBroadcaster := record.NewBroadcaster()
|
||||||
if len(reason) == 0 {
|
eventBroadcaster.StartStructuredLogging(3)
|
||||||
reason = strategy
|
eventBroadcaster.StartRecordingToSink(&clientcorev1.EventSinkImpl{Interface: pe.client.CoreV1().Events(pod.Namespace)})
|
||||||
if len(reason) == 0 {
|
r := eventBroadcaster.NewRecorder(scheme.Scheme, v1.EventSource{Component: "sigs.k8s.io.descheduler"})
|
||||||
reason = "NotSet"
|
r.Event(pod, v1.EventTypeNormal, "Descheduled", fmt.Sprintf("pod evicted by sigs.k8s.io/descheduler%s", reason))
|
||||||
}
|
metrics.PodsEvicted.With(map[string]string{"result": "success", "strategy": strategy, "namespace": pod.Namespace}).Inc()
|
||||||
}
|
|
||||||
pe.eventRecorder.Eventf(pod, nil, v1.EventTypeNormal, reason, "Descheduled", "pod evicted by sigs.k8s.io/descheduler")
|
|
||||||
}
|
}
|
||||||
return true
|
return true, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func evictPod(ctx context.Context, client clientset.Interface, pod *v1.Pod, policyGroupVersion string) error {
|
func evictPod(ctx context.Context, client clientset.Interface, pod *v1.Pod, policyGroupVersion string, dryRun bool) error {
|
||||||
|
if dryRun {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
deleteOptions := &metav1.DeleteOptions{}
|
deleteOptions := &metav1.DeleteOptions{}
|
||||||
// GracePeriodSeconds ?
|
// GracePeriodSeconds ?
|
||||||
eviction := &policy.Eviction{
|
eviction := &policy.Eviction{
|
||||||
@@ -184,7 +155,7 @@ func evictPod(ctx context.Context, client clientset.Interface, pod *v1.Pod, poli
|
|||||||
},
|
},
|
||||||
DeleteOptions: deleteOptions,
|
DeleteOptions: deleteOptions,
|
||||||
}
|
}
|
||||||
err := client.PolicyV1().Evictions(eviction.Namespace).Evict(ctx, eviction)
|
err := client.PolicyV1beta1().Evictions(eviction.Namespace).Evict(ctx, eviction)
|
||||||
|
|
||||||
if apierrors.IsTooManyRequests(err) {
|
if apierrors.IsTooManyRequests(err) {
|
||||||
return fmt.Errorf("error when evicting pod (ignoring) %q: %v", pod.Name, err)
|
return fmt.Errorf("error when evicting pod (ignoring) %q: %v", pod.Name, err)
|
||||||
@@ -194,3 +165,152 @@ func evictPod(ctx context.Context, client clientset.Interface, pod *v1.Pod, poli
|
|||||||
}
|
}
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
|
type Options struct {
|
||||||
|
priority *int32
|
||||||
|
nodeFit bool
|
||||||
|
labelSelector labels.Selector
|
||||||
|
}
|
||||||
|
|
||||||
|
// WithPriorityThreshold sets a threshold for pod's priority class.
|
||||||
|
// Any pod whose priority class is lower is evictable.
|
||||||
|
func WithPriorityThreshold(priority int32) func(opts *Options) {
|
||||||
|
return func(opts *Options) {
|
||||||
|
var p int32 = priority
|
||||||
|
opts.priority = &p
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// WithNodeFit sets whether or not to consider taints, node selectors,
|
||||||
|
// and pod affinity when evicting. A pod whose tolerations, node selectors,
|
||||||
|
// and affinity match a node other than the one it is currently running on
|
||||||
|
// is evictable.
|
||||||
|
func WithNodeFit(nodeFit bool) func(opts *Options) {
|
||||||
|
return func(opts *Options) {
|
||||||
|
opts.nodeFit = nodeFit
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// WithLabelSelector sets whether or not to apply label filtering when evicting.
|
||||||
|
// Any pod matching the label selector is considered evictable.
|
||||||
|
func WithLabelSelector(labelSelector labels.Selector) func(opts *Options) {
|
||||||
|
return func(opts *Options) {
|
||||||
|
opts.labelSelector = labelSelector
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
type constraint func(pod *v1.Pod) error
|
||||||
|
|
||||||
|
type evictable struct {
|
||||||
|
constraints []constraint
|
||||||
|
}
|
||||||
|
|
||||||
|
// Evictable provides an implementation of IsEvictable(IsEvictable(pod *v1.Pod) bool).
|
||||||
|
// The method accepts a list of options which allow to extend constraints
|
||||||
|
// which decides when a pod is considered evictable.
|
||||||
|
func (pe *PodEvictor) Evictable(opts ...func(opts *Options)) *evictable {
|
||||||
|
options := &Options{}
|
||||||
|
for _, opt := range opts {
|
||||||
|
opt(options)
|
||||||
|
}
|
||||||
|
|
||||||
|
ev := &evictable{}
|
||||||
|
if !pe.evictSystemCriticalPods {
|
||||||
|
ev.constraints = append(ev.constraints, func(pod *v1.Pod) error {
|
||||||
|
// Moved from IsEvictable function to allow for disabling
|
||||||
|
if utils.IsCriticalPriorityPod(pod) {
|
||||||
|
return fmt.Errorf("pod has system critical priority")
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
})
|
||||||
|
|
||||||
|
if options.priority != nil {
|
||||||
|
ev.constraints = append(ev.constraints, func(pod *v1.Pod) error {
|
||||||
|
if IsPodEvictableBasedOnPriority(pod, *options.priority) {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
return fmt.Errorf("pod has higher priority than specified priority class threshold")
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if !pe.evictLocalStoragePods {
|
||||||
|
ev.constraints = append(ev.constraints, func(pod *v1.Pod) error {
|
||||||
|
if utils.IsPodWithLocalStorage(pod) {
|
||||||
|
return fmt.Errorf("pod has local storage and descheduler is not configured with evictLocalStoragePods")
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
})
|
||||||
|
}
|
||||||
|
if pe.ignorePvcPods {
|
||||||
|
ev.constraints = append(ev.constraints, func(pod *v1.Pod) error {
|
||||||
|
if utils.IsPodWithPVC(pod) {
|
||||||
|
return fmt.Errorf("pod has a PVC and descheduler is configured to ignore PVC pods")
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
})
|
||||||
|
}
|
||||||
|
if options.nodeFit {
|
||||||
|
ev.constraints = append(ev.constraints, func(pod *v1.Pod) error {
|
||||||
|
if !nodeutil.PodFitsAnyOtherNode(pod, pe.nodes) {
|
||||||
|
return fmt.Errorf("pod does not fit on any other node because of nodeSelector(s), Taint(s), or nodes marked as unschedulable")
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
})
|
||||||
|
}
|
||||||
|
if options.labelSelector != nil && !options.labelSelector.Empty() {
|
||||||
|
ev.constraints = append(ev.constraints, func(pod *v1.Pod) error {
|
||||||
|
if !options.labelSelector.Matches(labels.Set(pod.Labels)) {
|
||||||
|
return fmt.Errorf("pod labels do not match the labelSelector filter in the policy parameter")
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
return ev
|
||||||
|
}
|
||||||
|
|
||||||
|
// IsEvictable decides when a pod is evictable
|
||||||
|
func (ev *evictable) IsEvictable(pod *v1.Pod) bool {
|
||||||
|
checkErrs := []error{}
|
||||||
|
|
||||||
|
ownerRefList := podutil.OwnerRef(pod)
|
||||||
|
if utils.IsDaemonsetPod(ownerRefList) {
|
||||||
|
checkErrs = append(checkErrs, fmt.Errorf("pod is a DaemonSet pod"))
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(ownerRefList) == 0 {
|
||||||
|
checkErrs = append(checkErrs, fmt.Errorf("pod does not have any ownerrefs"))
|
||||||
|
}
|
||||||
|
|
||||||
|
if utils.IsMirrorPod(pod) {
|
||||||
|
checkErrs = append(checkErrs, fmt.Errorf("pod is a mirror pod"))
|
||||||
|
}
|
||||||
|
|
||||||
|
if utils.IsStaticPod(pod) {
|
||||||
|
checkErrs = append(checkErrs, fmt.Errorf("pod is a static pod"))
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, c := range ev.constraints {
|
||||||
|
if err := c(pod); err != nil {
|
||||||
|
checkErrs = append(checkErrs, err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(checkErrs) > 0 && !HaveEvictAnnotation(pod) {
|
||||||
|
klog.V(4).InfoS("Pod lacks an eviction annotation and fails the following checks", "pod", klog.KObj(pod), "checks", errors.NewAggregate(checkErrs).Error())
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
|
||||||
|
// HaveEvictAnnotation checks if the pod have evict annotation
|
||||||
|
func HaveEvictAnnotation(pod *v1.Pod) bool {
|
||||||
|
_, found := pod.ObjectMeta.Annotations[evictPodAnnotationKey]
|
||||||
|
return found
|
||||||
|
}
|
||||||
|
|
||||||
|
// IsPodEvictableBasedOnPriority checks if the given pod is evictable based on priority resolved from pod Spec.
|
||||||
|
func IsPodEvictableBasedOnPriority(pod *v1.Pod, priority int32) bool {
|
||||||
|
return pod.Spec.Priority == nil || *pod.Spec.Priority < priority
|
||||||
|
}
|
||||||
|
|||||||
@@ -62,13 +62,381 @@ func TestEvictPod(t *testing.T) {
|
|||||||
fakeClient.Fake.AddReactor("list", "pods", func(action core.Action) (bool, runtime.Object, error) {
|
fakeClient.Fake.AddReactor("list", "pods", func(action core.Action) (bool, runtime.Object, error) {
|
||||||
return true, &v1.PodList{Items: test.pods}, nil
|
return true, &v1.PodList{Items: test.pods}, nil
|
||||||
})
|
})
|
||||||
got := evictPod(ctx, fakeClient, test.pod, "v1")
|
got := evictPod(ctx, fakeClient, test.pod, "v1", false)
|
||||||
if got != test.want {
|
if got != test.want {
|
||||||
t.Errorf("Test error for Desc: %s. Expected %v pod eviction to be %v, got %v", test.description, test.pod.Name, test.want, got)
|
t.Errorf("Test error for Desc: %s. Expected %v pod eviction to be %v, got %v", test.description, test.pod.Name, test.want, got)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func TestIsEvictable(t *testing.T) {
|
||||||
|
n1 := test.BuildTestNode("node1", 1000, 2000, 13, nil)
|
||||||
|
lowPriority := int32(800)
|
||||||
|
highPriority := int32(900)
|
||||||
|
|
||||||
|
nodeTaintKey := "hardware"
|
||||||
|
nodeTaintValue := "gpu"
|
||||||
|
|
||||||
|
nodeLabelKey := "datacenter"
|
||||||
|
nodeLabelValue := "east"
|
||||||
|
type testCase struct {
|
||||||
|
pod *v1.Pod
|
||||||
|
nodes []*v1.Node
|
||||||
|
runBefore func(*v1.Pod, []*v1.Node)
|
||||||
|
evictLocalStoragePods bool
|
||||||
|
evictSystemCriticalPods bool
|
||||||
|
priorityThreshold *int32
|
||||||
|
nodeFit bool
|
||||||
|
result bool
|
||||||
|
}
|
||||||
|
|
||||||
|
testCases := []testCase{
|
||||||
|
{ // Normal pod eviction with normal ownerRefs
|
||||||
|
pod: test.BuildTestPod("p1", 400, 0, n1.Name, nil),
|
||||||
|
runBefore: func(pod *v1.Pod, nodes []*v1.Node) {
|
||||||
|
pod.ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
||||||
|
},
|
||||||
|
evictLocalStoragePods: false,
|
||||||
|
evictSystemCriticalPods: false,
|
||||||
|
result: true,
|
||||||
|
}, { // Normal pod eviction with normal ownerRefs and descheduler.alpha.kubernetes.io/evict annotation
|
||||||
|
pod: test.BuildTestPod("p2", 400, 0, n1.Name, nil),
|
||||||
|
runBefore: func(pod *v1.Pod, nodes []*v1.Node) {
|
||||||
|
pod.Annotations = map[string]string{"descheduler.alpha.kubernetes.io/evict": "true"}
|
||||||
|
pod.ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
||||||
|
},
|
||||||
|
evictLocalStoragePods: false,
|
||||||
|
evictSystemCriticalPods: false,
|
||||||
|
result: true,
|
||||||
|
}, { // Normal pod eviction with replicaSet ownerRefs
|
||||||
|
pod: test.BuildTestPod("p3", 400, 0, n1.Name, nil),
|
||||||
|
runBefore: func(pod *v1.Pod, nodes []*v1.Node) {
|
||||||
|
pod.ObjectMeta.OwnerReferences = test.GetReplicaSetOwnerRefList()
|
||||||
|
},
|
||||||
|
evictLocalStoragePods: false,
|
||||||
|
evictSystemCriticalPods: false,
|
||||||
|
result: true,
|
||||||
|
}, { // Normal pod eviction with replicaSet ownerRefs and descheduler.alpha.kubernetes.io/evict annotation
|
||||||
|
pod: test.BuildTestPod("p4", 400, 0, n1.Name, nil),
|
||||||
|
runBefore: func(pod *v1.Pod, nodes []*v1.Node) {
|
||||||
|
pod.Annotations = map[string]string{"descheduler.alpha.kubernetes.io/evict": "true"}
|
||||||
|
pod.ObjectMeta.OwnerReferences = test.GetReplicaSetOwnerRefList()
|
||||||
|
},
|
||||||
|
evictLocalStoragePods: false,
|
||||||
|
evictSystemCriticalPods: false,
|
||||||
|
result: true,
|
||||||
|
}, { // Normal pod eviction with statefulSet ownerRefs
|
||||||
|
pod: test.BuildTestPod("p18", 400, 0, n1.Name, nil),
|
||||||
|
runBefore: func(pod *v1.Pod, nodes []*v1.Node) {
|
||||||
|
pod.ObjectMeta.OwnerReferences = test.GetStatefulSetOwnerRefList()
|
||||||
|
},
|
||||||
|
evictLocalStoragePods: false,
|
||||||
|
evictSystemCriticalPods: false,
|
||||||
|
result: true,
|
||||||
|
}, { // Normal pod eviction with statefulSet ownerRefs and descheduler.alpha.kubernetes.io/evict annotation
|
||||||
|
pod: test.BuildTestPod("p19", 400, 0, n1.Name, nil),
|
||||||
|
runBefore: func(pod *v1.Pod, nodes []*v1.Node) {
|
||||||
|
pod.Annotations = map[string]string{"descheduler.alpha.kubernetes.io/evict": "true"}
|
||||||
|
pod.ObjectMeta.OwnerReferences = test.GetStatefulSetOwnerRefList()
|
||||||
|
},
|
||||||
|
evictLocalStoragePods: false,
|
||||||
|
evictSystemCriticalPods: false,
|
||||||
|
result: true,
|
||||||
|
}, { // Pod not evicted because it is bound to a PV and evictLocalStoragePods = false
|
||||||
|
pod: test.BuildTestPod("p5", 400, 0, n1.Name, nil),
|
||||||
|
runBefore: func(pod *v1.Pod, nodes []*v1.Node) {
|
||||||
|
pod.ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
||||||
|
pod.Spec.Volumes = []v1.Volume{
|
||||||
|
{
|
||||||
|
Name: "sample",
|
||||||
|
VolumeSource: v1.VolumeSource{
|
||||||
|
HostPath: &v1.HostPathVolumeSource{Path: "somePath"},
|
||||||
|
EmptyDir: &v1.EmptyDirVolumeSource{
|
||||||
|
SizeLimit: resource.NewQuantity(int64(10), resource.BinarySI)},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
},
|
||||||
|
evictLocalStoragePods: false,
|
||||||
|
evictSystemCriticalPods: false,
|
||||||
|
result: false,
|
||||||
|
}, { // Pod is evicted because it is bound to a PV and evictLocalStoragePods = true
|
||||||
|
pod: test.BuildTestPod("p6", 400, 0, n1.Name, nil),
|
||||||
|
runBefore: func(pod *v1.Pod, nodes []*v1.Node) {
|
||||||
|
pod.ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
||||||
|
pod.Spec.Volumes = []v1.Volume{
|
||||||
|
{
|
||||||
|
Name: "sample",
|
||||||
|
VolumeSource: v1.VolumeSource{
|
||||||
|
HostPath: &v1.HostPathVolumeSource{Path: "somePath"},
|
||||||
|
EmptyDir: &v1.EmptyDirVolumeSource{
|
||||||
|
SizeLimit: resource.NewQuantity(int64(10), resource.BinarySI)},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
},
|
||||||
|
evictLocalStoragePods: true,
|
||||||
|
evictSystemCriticalPods: false,
|
||||||
|
result: true,
|
||||||
|
}, { // Pod is evicted because it is bound to a PV and evictLocalStoragePods = false, but it has scheduler.alpha.kubernetes.io/evict annotation
|
||||||
|
pod: test.BuildTestPod("p7", 400, 0, n1.Name, nil),
|
||||||
|
runBefore: func(pod *v1.Pod, nodes []*v1.Node) {
|
||||||
|
pod.Annotations = map[string]string{"descheduler.alpha.kubernetes.io/evict": "true"}
|
||||||
|
pod.ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
||||||
|
pod.Spec.Volumes = []v1.Volume{
|
||||||
|
{
|
||||||
|
Name: "sample",
|
||||||
|
VolumeSource: v1.VolumeSource{
|
||||||
|
HostPath: &v1.HostPathVolumeSource{Path: "somePath"},
|
||||||
|
EmptyDir: &v1.EmptyDirVolumeSource{
|
||||||
|
SizeLimit: resource.NewQuantity(int64(10), resource.BinarySI)},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
},
|
||||||
|
evictLocalStoragePods: false,
|
||||||
|
evictSystemCriticalPods: false,
|
||||||
|
result: true,
|
||||||
|
}, { // Pod not evicted becasuse it is part of a daemonSet
|
||||||
|
pod: test.BuildTestPod("p8", 400, 0, n1.Name, nil),
|
||||||
|
runBefore: func(pod *v1.Pod, nodes []*v1.Node) {
|
||||||
|
pod.ObjectMeta.OwnerReferences = test.GetDaemonSetOwnerRefList()
|
||||||
|
},
|
||||||
|
evictLocalStoragePods: false,
|
||||||
|
evictSystemCriticalPods: false,
|
||||||
|
result: false,
|
||||||
|
}, { // Pod is evicted becasuse it is part of a daemonSet, but it has scheduler.alpha.kubernetes.io/evict annotation
|
||||||
|
pod: test.BuildTestPod("p9", 400, 0, n1.Name, nil),
|
||||||
|
runBefore: func(pod *v1.Pod, nodes []*v1.Node) {
|
||||||
|
pod.Annotations = map[string]string{"descheduler.alpha.kubernetes.io/evict": "true"}
|
||||||
|
pod.ObjectMeta.OwnerReferences = test.GetDaemonSetOwnerRefList()
|
||||||
|
},
|
||||||
|
evictLocalStoragePods: false,
|
||||||
|
evictSystemCriticalPods: false,
|
||||||
|
result: true,
|
||||||
|
}, { // Pod not evicted becasuse it is a mirror pod
|
||||||
|
pod: test.BuildTestPod("p10", 400, 0, n1.Name, nil),
|
||||||
|
runBefore: func(pod *v1.Pod, nodes []*v1.Node) {
|
||||||
|
pod.ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
||||||
|
pod.Annotations = test.GetMirrorPodAnnotation()
|
||||||
|
},
|
||||||
|
evictLocalStoragePods: false,
|
||||||
|
evictSystemCriticalPods: false,
|
||||||
|
result: false,
|
||||||
|
}, { // Pod is evicted becasuse it is a mirror pod, but it has scheduler.alpha.kubernetes.io/evict annotation
|
||||||
|
pod: test.BuildTestPod("p11", 400, 0, n1.Name, nil),
|
||||||
|
runBefore: func(pod *v1.Pod, nodes []*v1.Node) {
|
||||||
|
pod.ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
||||||
|
pod.Annotations = test.GetMirrorPodAnnotation()
|
||||||
|
pod.Annotations["descheduler.alpha.kubernetes.io/evict"] = "true"
|
||||||
|
},
|
||||||
|
evictLocalStoragePods: false,
|
||||||
|
evictSystemCriticalPods: false,
|
||||||
|
result: true,
|
||||||
|
}, { // Pod not evicted becasuse it has system critical priority
|
||||||
|
pod: test.BuildTestPod("p12", 400, 0, n1.Name, nil),
|
||||||
|
runBefore: func(pod *v1.Pod, nodes []*v1.Node) {
|
||||||
|
pod.ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
||||||
|
priority := utils.SystemCriticalPriority
|
||||||
|
pod.Spec.Priority = &priority
|
||||||
|
},
|
||||||
|
evictLocalStoragePods: false,
|
||||||
|
evictSystemCriticalPods: false,
|
||||||
|
result: false,
|
||||||
|
}, { // Pod is evicted becasuse it has system critical priority, but it has scheduler.alpha.kubernetes.io/evict annotation
|
||||||
|
pod: test.BuildTestPod("p13", 400, 0, n1.Name, nil),
|
||||||
|
runBefore: func(pod *v1.Pod, nodes []*v1.Node) {
|
||||||
|
pod.ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
||||||
|
priority := utils.SystemCriticalPriority
|
||||||
|
pod.Spec.Priority = &priority
|
||||||
|
pod.Annotations = map[string]string{
|
||||||
|
"descheduler.alpha.kubernetes.io/evict": "true",
|
||||||
|
}
|
||||||
|
},
|
||||||
|
evictLocalStoragePods: false,
|
||||||
|
evictSystemCriticalPods: false,
|
||||||
|
result: true,
|
||||||
|
}, { // Pod not evicted becasuse it has a priority higher than the configured priority threshold
|
||||||
|
pod: test.BuildTestPod("p14", 400, 0, n1.Name, nil),
|
||||||
|
runBefore: func(pod *v1.Pod, nodes []*v1.Node) {
|
||||||
|
pod.ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
||||||
|
pod.Spec.Priority = &highPriority
|
||||||
|
},
|
||||||
|
evictLocalStoragePods: false,
|
||||||
|
evictSystemCriticalPods: false,
|
||||||
|
priorityThreshold: &lowPriority,
|
||||||
|
result: false,
|
||||||
|
}, { // Pod is evicted becasuse it has a priority higher than the configured priority threshold, but it has scheduler.alpha.kubernetes.io/evict annotation
|
||||||
|
pod: test.BuildTestPod("p15", 400, 0, n1.Name, nil),
|
||||||
|
runBefore: func(pod *v1.Pod, nodes []*v1.Node) {
|
||||||
|
pod.ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
||||||
|
pod.Annotations = map[string]string{"descheduler.alpha.kubernetes.io/evict": "true"}
|
||||||
|
pod.Spec.Priority = &highPriority
|
||||||
|
},
|
||||||
|
evictLocalStoragePods: false,
|
||||||
|
evictSystemCriticalPods: false,
|
||||||
|
priorityThreshold: &lowPriority,
|
||||||
|
result: true,
|
||||||
|
}, { // Pod is evicted becasuse it has system critical priority, but evictSystemCriticalPods = true
|
||||||
|
pod: test.BuildTestPod("p16", 400, 0, n1.Name, nil),
|
||||||
|
runBefore: func(pod *v1.Pod, nodes []*v1.Node) {
|
||||||
|
pod.ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
||||||
|
priority := utils.SystemCriticalPriority
|
||||||
|
pod.Spec.Priority = &priority
|
||||||
|
},
|
||||||
|
evictLocalStoragePods: false,
|
||||||
|
evictSystemCriticalPods: true,
|
||||||
|
result: true,
|
||||||
|
}, { // Pod is evicted becasuse it has system critical priority, but evictSystemCriticalPods = true and it has scheduler.alpha.kubernetes.io/evict annotation
|
||||||
|
pod: test.BuildTestPod("p16", 400, 0, n1.Name, nil),
|
||||||
|
runBefore: func(pod *v1.Pod, nodes []*v1.Node) {
|
||||||
|
pod.ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
||||||
|
pod.Annotations = map[string]string{"descheduler.alpha.kubernetes.io/evict": "true"}
|
||||||
|
priority := utils.SystemCriticalPriority
|
||||||
|
pod.Spec.Priority = &priority
|
||||||
|
},
|
||||||
|
evictLocalStoragePods: false,
|
||||||
|
evictSystemCriticalPods: true,
|
||||||
|
result: true,
|
||||||
|
}, { // Pod is evicted becasuse it has a priority higher than the configured priority threshold, but evictSystemCriticalPods = true
|
||||||
|
pod: test.BuildTestPod("p17", 400, 0, n1.Name, nil),
|
||||||
|
runBefore: func(pod *v1.Pod, nodes []*v1.Node) {
|
||||||
|
pod.ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
||||||
|
pod.Spec.Priority = &highPriority
|
||||||
|
},
|
||||||
|
evictLocalStoragePods: false,
|
||||||
|
evictSystemCriticalPods: true,
|
||||||
|
priorityThreshold: &lowPriority,
|
||||||
|
result: true,
|
||||||
|
}, { // Pod is evicted becasuse it has a priority higher than the configured priority threshold, but evictSystemCriticalPods = true and it has scheduler.alpha.kubernetes.io/evict annotation
|
||||||
|
pod: test.BuildTestPod("p17", 400, 0, n1.Name, nil),
|
||||||
|
runBefore: func(pod *v1.Pod, nodes []*v1.Node) {
|
||||||
|
pod.ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
||||||
|
pod.Annotations = map[string]string{"descheduler.alpha.kubernetes.io/evict": "true"}
|
||||||
|
pod.Spec.Priority = &highPriority
|
||||||
|
},
|
||||||
|
evictLocalStoragePods: false,
|
||||||
|
evictSystemCriticalPods: true,
|
||||||
|
priorityThreshold: &lowPriority,
|
||||||
|
result: true,
|
||||||
|
}, { // Pod with no tolerations running on normal node, all other nodes tainted
|
||||||
|
pod: test.BuildTestPod("p1", 400, 0, n1.Name, nil),
|
||||||
|
nodes: []*v1.Node{test.BuildTestNode("node2", 1000, 2000, 13, nil), test.BuildTestNode("node3", 1000, 2000, 13, nil)},
|
||||||
|
runBefore: func(pod *v1.Pod, nodes []*v1.Node) {
|
||||||
|
pod.ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
||||||
|
|
||||||
|
for _, node := range nodes {
|
||||||
|
node.Spec.Taints = []v1.Taint{
|
||||||
|
{
|
||||||
|
Key: nodeTaintKey,
|
||||||
|
Value: nodeTaintValue,
|
||||||
|
Effect: v1.TaintEffectNoSchedule,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
}
|
||||||
|
},
|
||||||
|
evictLocalStoragePods: false,
|
||||||
|
evictSystemCriticalPods: false,
|
||||||
|
nodeFit: true,
|
||||||
|
result: false,
|
||||||
|
}, { // Pod with correct tolerations running on normal node, all other nodes tainted
|
||||||
|
pod: test.BuildTestPod("p1", 400, 0, n1.Name, func(pod *v1.Pod) {
|
||||||
|
pod.Spec.Tolerations = []v1.Toleration{
|
||||||
|
{
|
||||||
|
Key: nodeTaintKey,
|
||||||
|
Value: nodeTaintValue,
|
||||||
|
Effect: v1.TaintEffectNoSchedule,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
}),
|
||||||
|
nodes: []*v1.Node{test.BuildTestNode("node2", 1000, 2000, 13, nil), test.BuildTestNode("node3", 1000, 2000, 13, nil)},
|
||||||
|
runBefore: func(pod *v1.Pod, nodes []*v1.Node) {
|
||||||
|
pod.ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
||||||
|
|
||||||
|
for _, node := range nodes {
|
||||||
|
node.Spec.Taints = []v1.Taint{
|
||||||
|
{
|
||||||
|
Key: nodeTaintKey,
|
||||||
|
Value: nodeTaintValue,
|
||||||
|
Effect: v1.TaintEffectNoSchedule,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
}
|
||||||
|
},
|
||||||
|
evictLocalStoragePods: false,
|
||||||
|
evictSystemCriticalPods: false,
|
||||||
|
nodeFit: true,
|
||||||
|
result: true,
|
||||||
|
}, { // Pod with incorrect node selector
|
||||||
|
pod: test.BuildTestPod("p1", 400, 0, n1.Name, func(pod *v1.Pod) {
|
||||||
|
pod.Spec.NodeSelector = map[string]string{
|
||||||
|
nodeLabelKey: "fail",
|
||||||
|
}
|
||||||
|
}),
|
||||||
|
nodes: []*v1.Node{test.BuildTestNode("node2", 1000, 2000, 13, nil), test.BuildTestNode("node3", 1000, 2000, 13, nil)},
|
||||||
|
runBefore: func(pod *v1.Pod, nodes []*v1.Node) {
|
||||||
|
pod.ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
||||||
|
|
||||||
|
for _, node := range nodes {
|
||||||
|
node.ObjectMeta.Labels = map[string]string{
|
||||||
|
nodeLabelKey: nodeLabelValue,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
},
|
||||||
|
evictLocalStoragePods: false,
|
||||||
|
evictSystemCriticalPods: false,
|
||||||
|
nodeFit: true,
|
||||||
|
result: false,
|
||||||
|
}, { // Pod with correct node selector
|
||||||
|
pod: test.BuildTestPod("p1", 400, 0, n1.Name, func(pod *v1.Pod) {
|
||||||
|
pod.Spec.NodeSelector = map[string]string{
|
||||||
|
nodeLabelKey: nodeLabelValue,
|
||||||
|
}
|
||||||
|
}),
|
||||||
|
nodes: []*v1.Node{test.BuildTestNode("node2", 1000, 2000, 13, nil), test.BuildTestNode("node3", 1000, 2000, 13, nil)},
|
||||||
|
runBefore: func(pod *v1.Pod, nodes []*v1.Node) {
|
||||||
|
pod.ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
||||||
|
|
||||||
|
for _, node := range nodes {
|
||||||
|
node.ObjectMeta.Labels = map[string]string{
|
||||||
|
nodeLabelKey: nodeLabelValue,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
},
|
||||||
|
evictLocalStoragePods: false,
|
||||||
|
evictSystemCriticalPods: false,
|
||||||
|
nodeFit: true,
|
||||||
|
result: true,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, test := range testCases {
|
||||||
|
test.runBefore(test.pod, test.nodes)
|
||||||
|
nodes := append(test.nodes, n1)
|
||||||
|
|
||||||
|
podEvictor := &PodEvictor{
|
||||||
|
evictLocalStoragePods: test.evictLocalStoragePods,
|
||||||
|
evictSystemCriticalPods: test.evictSystemCriticalPods,
|
||||||
|
nodes: nodes,
|
||||||
|
}
|
||||||
|
|
||||||
|
evictable := podEvictor.Evictable()
|
||||||
|
var opts []func(opts *Options)
|
||||||
|
if test.priorityThreshold != nil {
|
||||||
|
opts = append(opts, WithPriorityThreshold(*test.priorityThreshold))
|
||||||
|
}
|
||||||
|
if test.nodeFit {
|
||||||
|
opts = append(opts, WithNodeFit(true))
|
||||||
|
}
|
||||||
|
evictable = podEvictor.Evictable(opts...)
|
||||||
|
|
||||||
|
result := evictable.IsEvictable(test.pod)
|
||||||
|
if result != test.result {
|
||||||
|
t.Errorf("IsEvictable should return for pod %s %t, but it returns %t", test.pod.Name, test.result, result)
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
}
|
||||||
func TestPodTypes(t *testing.T) {
|
func TestPodTypes(t *testing.T) {
|
||||||
n1 := test.BuildTestNode("node1", 1000, 2000, 9, nil)
|
n1 := test.BuildTestNode("node1", 1000, 2000, 9, nil)
|
||||||
p1 := test.BuildTestPod("p1", 400, 0, n1.Name, nil)
|
p1 := test.BuildTestPod("p1", 400, 0, n1.Name, nil)
|
||||||
|
|||||||
@@ -1,99 +0,0 @@
|
|||||||
/*
|
|
||||||
Copyright 2022 The Kubernetes Authors.
|
|
||||||
|
|
||||||
Licensed under the Apache License, Version 2.0 (the "License");
|
|
||||||
you may not use this file except in compliance with the License.
|
|
||||||
You may obtain a copy of the License at
|
|
||||||
|
|
||||||
http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
|
|
||||||
Unless required by applicable law or agreed to in writing, software
|
|
||||||
distributed under the License is distributed on an "AS IS" BASIS,
|
|
||||||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
||||||
See the License for the specific language governing permissions and
|
|
||||||
limitations under the License.
|
|
||||||
*/
|
|
||||||
|
|
||||||
package descheduler
|
|
||||||
|
|
||||||
import (
|
|
||||||
"context"
|
|
||||||
"fmt"
|
|
||||||
"k8s.io/apimachinery/pkg/util/uuid"
|
|
||||||
clientset "k8s.io/client-go/kubernetes"
|
|
||||||
"k8s.io/client-go/tools/leaderelection"
|
|
||||||
"k8s.io/client-go/tools/leaderelection/resourcelock"
|
|
||||||
componentbaseconfig "k8s.io/component-base/config"
|
|
||||||
"k8s.io/klog/v2"
|
|
||||||
"os"
|
|
||||||
)
|
|
||||||
|
|
||||||
// NewLeaderElection starts the leader election code loop
|
|
||||||
func NewLeaderElection(
|
|
||||||
run func() error,
|
|
||||||
client clientset.Interface,
|
|
||||||
LeaderElectionConfig *componentbaseconfig.LeaderElectionConfiguration,
|
|
||||||
ctx context.Context,
|
|
||||||
) error {
|
|
||||||
var id string
|
|
||||||
|
|
||||||
if hostname, err := os.Hostname(); err != nil {
|
|
||||||
// on errors, make sure we're unique
|
|
||||||
id = string(uuid.NewUUID())
|
|
||||||
} else {
|
|
||||||
// add a uniquifier so that two processes on the same host don't accidentally both become active
|
|
||||||
id = hostname + "_" + string(uuid.NewUUID())
|
|
||||||
}
|
|
||||||
|
|
||||||
klog.V(3).Infof("Assigned unique lease holder id: %s", id)
|
|
||||||
|
|
||||||
if len(LeaderElectionConfig.ResourceNamespace) == 0 {
|
|
||||||
return fmt.Errorf("namespace may not be empty")
|
|
||||||
}
|
|
||||||
|
|
||||||
if len(LeaderElectionConfig.ResourceName) == 0 {
|
|
||||||
return fmt.Errorf("name may not be empty")
|
|
||||||
}
|
|
||||||
|
|
||||||
lock, err := resourcelock.New(
|
|
||||||
LeaderElectionConfig.ResourceLock,
|
|
||||||
LeaderElectionConfig.ResourceNamespace,
|
|
||||||
LeaderElectionConfig.ResourceName,
|
|
||||||
client.CoreV1(),
|
|
||||||
client.CoordinationV1(),
|
|
||||||
resourcelock.ResourceLockConfig{
|
|
||||||
Identity: id,
|
|
||||||
},
|
|
||||||
)
|
|
||||||
if err != nil {
|
|
||||||
return fmt.Errorf("unable to create leader election lock: %v", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
leaderelection.RunOrDie(ctx, leaderelection.LeaderElectionConfig{
|
|
||||||
Lock: lock,
|
|
||||||
ReleaseOnCancel: true,
|
|
||||||
LeaseDuration: LeaderElectionConfig.LeaseDuration.Duration,
|
|
||||||
RenewDeadline: LeaderElectionConfig.RenewDeadline.Duration,
|
|
||||||
RetryPeriod: LeaderElectionConfig.RetryPeriod.Duration,
|
|
||||||
Callbacks: leaderelection.LeaderCallbacks{
|
|
||||||
OnStartedLeading: func(ctx context.Context) {
|
|
||||||
klog.V(1).InfoS("Started leading")
|
|
||||||
err := run()
|
|
||||||
if err != nil {
|
|
||||||
klog.Error(err)
|
|
||||||
}
|
|
||||||
},
|
|
||||||
OnStoppedLeading: func() {
|
|
||||||
klog.V(1).InfoS("Leader lost")
|
|
||||||
},
|
|
||||||
OnNewLeader: func(identity string) {
|
|
||||||
// Just got the lock
|
|
||||||
if identity == id {
|
|
||||||
return
|
|
||||||
}
|
|
||||||
klog.V(1).Infof("New leader elected: %v", identity)
|
|
||||||
},
|
|
||||||
},
|
|
||||||
})
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
@@ -18,16 +18,13 @@ package node
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
"fmt"
|
|
||||||
|
|
||||||
v1 "k8s.io/api/core/v1"
|
v1 "k8s.io/api/core/v1"
|
||||||
"k8s.io/apimachinery/pkg/api/resource"
|
|
||||||
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
||||||
"k8s.io/apimachinery/pkg/labels"
|
"k8s.io/apimachinery/pkg/labels"
|
||||||
coreinformers "k8s.io/client-go/informers/core/v1"
|
coreinformers "k8s.io/client-go/informers/core/v1"
|
||||||
clientset "k8s.io/client-go/kubernetes"
|
clientset "k8s.io/client-go/kubernetes"
|
||||||
"k8s.io/klog/v2"
|
"k8s.io/klog/v2"
|
||||||
podutil "sigs.k8s.io/descheduler/pkg/descheduler/pod"
|
|
||||||
"sigs.k8s.io/descheduler/pkg/utils"
|
"sigs.k8s.io/descheduler/pkg/utils"
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -99,92 +96,32 @@ func IsReady(node *v1.Node) bool {
|
|||||||
return true
|
return true
|
||||||
}
|
}
|
||||||
|
|
||||||
// NodeFit returns true if the provided pod can be scheduled onto the provided node.
|
// PodFitsAnyOtherNode checks if the given pod fits any of the given nodes, besides the node
|
||||||
// This function is used when the NodeFit pod filtering feature of the Descheduler is enabled.
|
// the pod is already running on. The node fit is based on multiple criteria, like, pod node selector
|
||||||
// This function currently considers a subset of the Kubernetes Scheduler's predicates when
|
// matching the node label (including affinity), the taints on the node, and the node being schedulable or not.
|
||||||
// deciding if a pod would fit on a node, but more predicates may be added in the future.
|
func PodFitsAnyOtherNode(pod *v1.Pod, nodes []*v1.Node) bool {
|
||||||
func NodeFit(nodeIndexer podutil.GetPodsAssignedToNodeFunc, pod *v1.Pod, node *v1.Node) []error {
|
|
||||||
// Check node selector and required affinity
|
|
||||||
var errors []error
|
|
||||||
if ok, err := utils.PodMatchNodeSelector(pod, node); err != nil {
|
|
||||||
errors = append(errors, err)
|
|
||||||
} else if !ok {
|
|
||||||
errors = append(errors, fmt.Errorf("pod node selector does not match the node label"))
|
|
||||||
}
|
|
||||||
// Check taints (we only care about NoSchedule and NoExecute taints)
|
|
||||||
ok := utils.TolerationsTolerateTaintsWithFilter(pod.Spec.Tolerations, node.Spec.Taints, func(taint *v1.Taint) bool {
|
|
||||||
return taint.Effect == v1.TaintEffectNoSchedule || taint.Effect == v1.TaintEffectNoExecute
|
|
||||||
})
|
|
||||||
if !ok {
|
|
||||||
errors = append(errors, fmt.Errorf("pod does not tolerate taints on the node"))
|
|
||||||
}
|
|
||||||
// Check if the pod can fit on a node based off it's requests
|
|
||||||
if pod.Spec.NodeName == "" || pod.Spec.NodeName != node.Name {
|
|
||||||
ok, reqErrors := fitsRequest(nodeIndexer, pod, node)
|
|
||||||
if !ok {
|
|
||||||
errors = append(errors, reqErrors...)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
// Check if node is schedulable
|
|
||||||
if IsNodeUnschedulable(node) {
|
|
||||||
errors = append(errors, fmt.Errorf("node is not schedulable"))
|
|
||||||
}
|
|
||||||
|
|
||||||
return errors
|
|
||||||
}
|
|
||||||
|
|
||||||
// PodFitsAnyOtherNode checks if the given pod will fit any of the given nodes, besides the node
|
|
||||||
// the pod is already running on. The predicates used to determine if the pod will fit can be found in the NodeFit function.
|
|
||||||
func PodFitsAnyOtherNode(nodeIndexer podutil.GetPodsAssignedToNodeFunc, pod *v1.Pod, nodes []*v1.Node) bool {
|
|
||||||
for _, node := range nodes {
|
for _, node := range nodes {
|
||||||
// Skip node pod is already on
|
// Skip node pod is already on
|
||||||
if node.Name == pod.Spec.NodeName {
|
if node.Name == pod.Spec.NodeName {
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
|
// Check node selector and required affinity
|
||||||
errors := NodeFit(nodeIndexer, pod, node)
|
ok, err := utils.PodMatchNodeSelector(pod, node)
|
||||||
if len(errors) == 0 {
|
if err != nil || !ok {
|
||||||
klog.V(4).InfoS("Pod fits on node", "pod", klog.KObj(pod), "node", klog.KObj(node))
|
continue
|
||||||
return true
|
|
||||||
} else {
|
|
||||||
klog.V(4).InfoS("Pod does not fit on node", "pod", klog.KObj(pod), "node", klog.KObj(node))
|
|
||||||
for _, err := range errors {
|
|
||||||
klog.V(4).InfoS(err.Error())
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}
|
// Check taints (we only care about NoSchedule and NoExecute taints)
|
||||||
return false
|
ok = utils.TolerationsTolerateTaintsWithFilter(pod.Spec.Tolerations, node.Spec.Taints, func(taint *v1.Taint) bool {
|
||||||
}
|
return taint.Effect == v1.TaintEffectNoSchedule || taint.Effect == v1.TaintEffectNoExecute
|
||||||
|
})
|
||||||
// PodFitsAnyNode checks if the given pod will fit any of the given nodes. The predicates used
|
if !ok {
|
||||||
// to determine if the pod will fit can be found in the NodeFit function.
|
continue
|
||||||
func PodFitsAnyNode(nodeIndexer podutil.GetPodsAssignedToNodeFunc, pod *v1.Pod, nodes []*v1.Node) bool {
|
|
||||||
for _, node := range nodes {
|
|
||||||
errors := NodeFit(nodeIndexer, pod, node)
|
|
||||||
if len(errors) == 0 {
|
|
||||||
klog.V(4).InfoS("Pod fits on node", "pod", klog.KObj(pod), "node", klog.KObj(node))
|
|
||||||
return true
|
|
||||||
} else {
|
|
||||||
klog.V(4).InfoS("Pod does not fit on node", "pod", klog.KObj(pod), "node", klog.KObj(node))
|
|
||||||
for _, err := range errors {
|
|
||||||
klog.V(4).InfoS(err.Error())
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}
|
// Check if node is schedulable
|
||||||
return false
|
if !IsNodeUnschedulable(node) {
|
||||||
}
|
klog.V(2).InfoS("Pod can possibly be scheduled on a different node", "pod", klog.KObj(pod), "node", klog.KObj(node))
|
||||||
|
return true
|
||||||
// PodFitsCurrentNode checks if the given pod will fit onto the given node. The predicates used
|
|
||||||
// to determine if the pod will fit can be found in the NodeFit function.
|
|
||||||
func PodFitsCurrentNode(nodeIndexer podutil.GetPodsAssignedToNodeFunc, pod *v1.Pod, node *v1.Node) bool {
|
|
||||||
errors := NodeFit(nodeIndexer, pod, node)
|
|
||||||
if len(errors) == 0 {
|
|
||||||
klog.V(4).InfoS("Pod fits on node", "pod", klog.KObj(pod), "node", klog.KObj(node))
|
|
||||||
return true
|
|
||||||
} else {
|
|
||||||
klog.V(4).InfoS("Pod does not fit on node", "pod", klog.KObj(pod), "node", klog.KObj(node))
|
|
||||||
for _, err := range errors {
|
|
||||||
klog.V(4).InfoS(err.Error())
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
return false
|
return false
|
||||||
@@ -196,95 +133,39 @@ func IsNodeUnschedulable(node *v1.Node) bool {
|
|||||||
return node.Spec.Unschedulable
|
return node.Spec.Unschedulable
|
||||||
}
|
}
|
||||||
|
|
||||||
// fitsRequest determines if a pod can fit on a node based on its resource requests. It returns true if
|
// PodFitsAnyNode checks if the given pod fits any of the given nodes, based on
|
||||||
// the pod will fit.
|
// multiple criteria, like, pod node selector matching the node label, node
|
||||||
func fitsRequest(nodeIndexer podutil.GetPodsAssignedToNodeFunc, pod *v1.Pod, node *v1.Node) (bool, []error) {
|
// being schedulable or not.
|
||||||
var insufficientResources []error
|
func PodFitsAnyNode(pod *v1.Pod, nodes []*v1.Node) bool {
|
||||||
|
for _, node := range nodes {
|
||||||
|
|
||||||
// Get pod requests
|
ok, err := utils.PodMatchNodeSelector(pod, node)
|
||||||
podRequests, _ := utils.PodRequestsAndLimits(pod)
|
if err != nil || !ok {
|
||||||
resourceNames := make([]v1.ResourceName, 0, len(podRequests))
|
continue
|
||||||
for name := range podRequests {
|
}
|
||||||
resourceNames = append(resourceNames, name)
|
if !IsNodeUnschedulable(node) {
|
||||||
|
klog.V(2).InfoS("Pod can possibly be scheduled on a different node", "pod", klog.KObj(pod), "node", klog.KObj(node))
|
||||||
|
return true
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
|
// PodFitsCurrentNode checks if the given pod fits on the given node if the pod
|
||||||
|
// node selector matches the node label.
|
||||||
|
func PodFitsCurrentNode(pod *v1.Pod, node *v1.Node) bool {
|
||||||
|
ok, err := utils.PodMatchNodeSelector(pod, node)
|
||||||
|
|
||||||
availableResources, err := nodeAvailableResources(nodeIndexer, node, resourceNames)
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return false, []error{err}
|
klog.ErrorS(err, "Failed to match node selector")
|
||||||
}
|
|
||||||
|
|
||||||
podFitsOnNode := true
|
|
||||||
for _, resource := range resourceNames {
|
|
||||||
podResourceRequest := podRequests[resource]
|
|
||||||
availableResource, ok := availableResources[resource]
|
|
||||||
if !ok || podResourceRequest.MilliValue() > availableResource.MilliValue() {
|
|
||||||
insufficientResources = append(insufficientResources, fmt.Errorf("insufficient %v", resource))
|
|
||||||
podFitsOnNode = false
|
|
||||||
}
|
|
||||||
}
|
|
||||||
return podFitsOnNode, insufficientResources
|
|
||||||
}
|
|
||||||
|
|
||||||
// nodeAvailableResources returns resources mapped to the quanitity available on the node.
|
|
||||||
func nodeAvailableResources(nodeIndexer podutil.GetPodsAssignedToNodeFunc, node *v1.Node, resourceNames []v1.ResourceName) (map[v1.ResourceName]*resource.Quantity, error) {
|
|
||||||
podsOnNode, err := podutil.ListPodsOnANode(node.Name, nodeIndexer, nil)
|
|
||||||
if err != nil {
|
|
||||||
return nil, err
|
|
||||||
}
|
|
||||||
nodeUtilization := NodeUtilization(podsOnNode, resourceNames)
|
|
||||||
remainingResources := map[v1.ResourceName]*resource.Quantity{
|
|
||||||
v1.ResourceCPU: resource.NewMilliQuantity(node.Status.Allocatable.Cpu().MilliValue()-nodeUtilization[v1.ResourceCPU].MilliValue(), resource.DecimalSI),
|
|
||||||
v1.ResourceMemory: resource.NewQuantity(node.Status.Allocatable.Memory().Value()-nodeUtilization[v1.ResourceMemory].Value(), resource.BinarySI),
|
|
||||||
v1.ResourcePods: resource.NewQuantity(node.Status.Allocatable.Pods().Value()-nodeUtilization[v1.ResourcePods].Value(), resource.DecimalSI),
|
|
||||||
}
|
|
||||||
for _, name := range resourceNames {
|
|
||||||
if !IsBasicResource(name) {
|
|
||||||
if _, exists := node.Status.Allocatable[name]; exists {
|
|
||||||
allocatableResource := node.Status.Allocatable[name]
|
|
||||||
remainingResources[name] = resource.NewQuantity(allocatableResource.Value()-nodeUtilization[name].Value(), resource.DecimalSI)
|
|
||||||
} else {
|
|
||||||
remainingResources[name] = resource.NewQuantity(0, resource.DecimalSI)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
return remainingResources, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// NodeUtilization returns the resources requested by the given pods. Only resources supplied in the resourceNames parameter are calculated.
|
|
||||||
func NodeUtilization(pods []*v1.Pod, resourceNames []v1.ResourceName) map[v1.ResourceName]*resource.Quantity {
|
|
||||||
totalReqs := map[v1.ResourceName]*resource.Quantity{
|
|
||||||
v1.ResourceCPU: resource.NewMilliQuantity(0, resource.DecimalSI),
|
|
||||||
v1.ResourceMemory: resource.NewQuantity(0, resource.BinarySI),
|
|
||||||
v1.ResourcePods: resource.NewQuantity(int64(len(pods)), resource.DecimalSI),
|
|
||||||
}
|
|
||||||
for _, name := range resourceNames {
|
|
||||||
if !IsBasicResource(name) {
|
|
||||||
totalReqs[name] = resource.NewQuantity(0, resource.DecimalSI)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
for _, pod := range pods {
|
|
||||||
req, _ := utils.PodRequestsAndLimits(pod)
|
|
||||||
for _, name := range resourceNames {
|
|
||||||
quantity, ok := req[name]
|
|
||||||
if ok && name != v1.ResourcePods {
|
|
||||||
// As Quantity.Add says: Add adds the provided y quantity to the current value. If the current value is zero,
|
|
||||||
// the format of the quantity will be updated to the format of y.
|
|
||||||
totalReqs[name].Add(quantity)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
return totalReqs
|
|
||||||
}
|
|
||||||
|
|
||||||
// IsBasicResource checks if resource is basic native.
|
|
||||||
func IsBasicResource(name v1.ResourceName) bool {
|
|
||||||
switch name {
|
|
||||||
case v1.ResourceCPU, v1.ResourceMemory, v1.ResourcePods:
|
|
||||||
return true
|
|
||||||
default:
|
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if !ok {
|
||||||
|
klog.V(2).InfoS("Pod does not fit on node", "pod", klog.KObj(pod), "node", klog.KObj(node))
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
|
klog.V(2).InfoS("Pod fits on node", "pod", klog.KObj(pod), "node", klog.KObj(node))
|
||||||
|
return true
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -21,12 +21,9 @@ import (
|
|||||||
"testing"
|
"testing"
|
||||||
|
|
||||||
v1 "k8s.io/api/core/v1"
|
v1 "k8s.io/api/core/v1"
|
||||||
"k8s.io/apimachinery/pkg/api/resource"
|
|
||||||
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
||||||
"k8s.io/apimachinery/pkg/runtime"
|
|
||||||
"k8s.io/client-go/informers"
|
"k8s.io/client-go/informers"
|
||||||
"k8s.io/client-go/kubernetes/fake"
|
"k8s.io/client-go/kubernetes/fake"
|
||||||
podutil "sigs.k8s.io/descheduler/pkg/descheduler/pod"
|
|
||||||
"sigs.k8s.io/descheduler/test"
|
"sigs.k8s.io/descheduler/test"
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -150,13 +147,13 @@ func TestPodFitsCurrentNode(t *testing.T) {
|
|||||||
},
|
},
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
node: test.BuildTestNode("node1", 64000, 128*1000*1000*1000, 200, func(node *v1.Node) {
|
node: &v1.Node{
|
||||||
node.ObjectMeta.Labels = map[string]string{
|
ObjectMeta: metav1.ObjectMeta{
|
||||||
nodeLabelKey: nodeLabelValue,
|
Labels: map[string]string{
|
||||||
}
|
nodeLabelKey: nodeLabelValue,
|
||||||
|
},
|
||||||
node.Status.Allocatable[v1.ResourceEphemeralStorage] = *resource.NewQuantity(1000*1000*1000*1000, resource.DecimalSI)
|
},
|
||||||
}),
|
},
|
||||||
success: true,
|
success: true,
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
@@ -184,48 +181,27 @@ func TestPodFitsCurrentNode(t *testing.T) {
|
|||||||
},
|
},
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
node: test.BuildTestNode("node1", 64000, 128*1000*1000*1000, 200, func(node *v1.Node) {
|
node: &v1.Node{
|
||||||
node.ObjectMeta.Labels = map[string]string{
|
ObjectMeta: metav1.ObjectMeta{
|
||||||
nodeLabelKey: "no",
|
Labels: map[string]string{
|
||||||
}
|
nodeLabelKey: "no",
|
||||||
|
},
|
||||||
node.Status.Allocatable[v1.ResourceEphemeralStorage] = *resource.NewQuantity(1000*1000*1000*1000, resource.DecimalSI)
|
},
|
||||||
}),
|
},
|
||||||
success: false,
|
success: false,
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
for _, tc := range tests {
|
for _, tc := range tests {
|
||||||
t.Run(tc.description, func(t *testing.T) {
|
actual := PodFitsCurrentNode(tc.pod, tc.node)
|
||||||
ctx, cancel := context.WithCancel(context.Background())
|
if actual != tc.success {
|
||||||
defer cancel()
|
t.Errorf("Test %#v failed", tc.description)
|
||||||
|
}
|
||||||
var objs []runtime.Object
|
|
||||||
objs = append(objs, tc.node)
|
|
||||||
objs = append(objs, tc.pod)
|
|
||||||
|
|
||||||
fakeClient := fake.NewSimpleClientset(objs...)
|
|
||||||
|
|
||||||
sharedInformerFactory := informers.NewSharedInformerFactory(fakeClient, 0)
|
|
||||||
podInformer := sharedInformerFactory.Core().V1().Pods()
|
|
||||||
|
|
||||||
getPodsAssignedToNode, err := podutil.BuildGetPodsAssignedToNodeFunc(podInformer)
|
|
||||||
if err != nil {
|
|
||||||
t.Errorf("Build get pods assigned to node function error: %v", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
sharedInformerFactory.Start(ctx.Done())
|
|
||||||
sharedInformerFactory.WaitForCacheSync(ctx.Done())
|
|
||||||
|
|
||||||
actual := PodFitsCurrentNode(getPodsAssignedToNode, tc.pod, tc.node)
|
|
||||||
if actual != tc.success {
|
|
||||||
t.Errorf("Test %#v failed", tc.description)
|
|
||||||
}
|
|
||||||
})
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestPodFitsAnyOtherNode(t *testing.T) {
|
func TestPodFitsAnyOtherNode(t *testing.T) {
|
||||||
|
|
||||||
nodeLabelKey := "kubernetes.io/desiredNode"
|
nodeLabelKey := "kubernetes.io/desiredNode"
|
||||||
nodeLabelValue := "yes"
|
nodeLabelValue := "yes"
|
||||||
nodeTaintKey := "hardware"
|
nodeTaintKey := "hardware"
|
||||||
@@ -239,527 +215,238 @@ func TestPodFitsAnyOtherNode(t *testing.T) {
|
|||||||
pod *v1.Pod
|
pod *v1.Pod
|
||||||
nodes []*v1.Node
|
nodes []*v1.Node
|
||||||
success bool
|
success bool
|
||||||
podsOnNodes []*v1.Pod
|
|
||||||
}{
|
}{
|
||||||
{
|
{
|
||||||
description: "Pod fits another node matching node affinity",
|
description: "Pod fits another node matching node affinity",
|
||||||
pod: test.BuildTestPod("p1", 0, 0, nodeNames[2], func(pod *v1.Pod) {
|
pod: createPodManifest(nodeNames[2], nodeLabelKey, nodeLabelValue),
|
||||||
pod.Spec.NodeSelector = map[string]string{
|
|
||||||
nodeLabelKey: nodeLabelValue,
|
|
||||||
}
|
|
||||||
}),
|
|
||||||
nodes: []*v1.Node{
|
nodes: []*v1.Node{
|
||||||
test.BuildTestNode(nodeNames[0], 64000, 128*1000*1000*1000, 200, func(node *v1.Node) {
|
{
|
||||||
node.ObjectMeta.Labels = map[string]string{
|
ObjectMeta: metav1.ObjectMeta{
|
||||||
nodeLabelKey: nodeLabelValue,
|
Name: nodeNames[0],
|
||||||
}
|
Labels: map[string]string{
|
||||||
|
nodeLabelKey: nodeLabelValue,
|
||||||
node.Status.Allocatable[v1.ResourceEphemeralStorage] = *resource.NewQuantity(1000*1000*1000*1000, resource.DecimalSI)
|
},
|
||||||
}),
|
},
|
||||||
test.BuildTestNode(nodeNames[1], 64000, 128*1000*1000*1000, 200, func(node *v1.Node) {
|
},
|
||||||
node.ObjectMeta.Labels = map[string]string{
|
{
|
||||||
nodeLabelKey: "no",
|
ObjectMeta: metav1.ObjectMeta{
|
||||||
}
|
Name: nodeNames[1],
|
||||||
|
Labels: map[string]string{
|
||||||
node.Status.Allocatable[v1.ResourceEphemeralStorage] = *resource.NewQuantity(1000*1000*1000*1000, resource.DecimalSI)
|
nodeLabelKey: "no",
|
||||||
}),
|
},
|
||||||
test.BuildTestNode(nodeNames[2], 0, 0, 0, nil),
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
ObjectMeta: metav1.ObjectMeta{
|
||||||
|
Name: nodeNames[2],
|
||||||
|
},
|
||||||
|
},
|
||||||
},
|
},
|
||||||
podsOnNodes: []*v1.Pod{},
|
success: true,
|
||||||
success: true,
|
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
description: "Pod expected to fit one of the nodes",
|
description: "Pod expected to fit one of the nodes",
|
||||||
pod: test.BuildTestPod("p1", 0, 0, nodeNames[2], func(pod *v1.Pod) {
|
pod: createPodManifest(nodeNames[2], nodeLabelKey, nodeLabelValue),
|
||||||
pod.Spec.NodeSelector = map[string]string{
|
|
||||||
nodeLabelKey: nodeLabelValue,
|
|
||||||
}
|
|
||||||
}),
|
|
||||||
nodes: []*v1.Node{
|
nodes: []*v1.Node{
|
||||||
test.BuildTestNode(nodeNames[0], 64000, 128*1000*1000*1000, 200, func(node *v1.Node) {
|
{
|
||||||
node.ObjectMeta.Labels = map[string]string{
|
ObjectMeta: metav1.ObjectMeta{
|
||||||
nodeLabelKey: "no",
|
Name: nodeNames[0],
|
||||||
}
|
Labels: map[string]string{
|
||||||
|
nodeLabelKey: "no",
|
||||||
node.Status.Allocatable[v1.ResourceEphemeralStorage] = *resource.NewQuantity(1000*1000*1000*1000, resource.DecimalSI)
|
},
|
||||||
}),
|
},
|
||||||
test.BuildTestNode(nodeNames[1], 64000, 128*1000*1000*1000, 200, func(node *v1.Node) {
|
},
|
||||||
node.ObjectMeta.Labels = map[string]string{
|
{
|
||||||
nodeLabelKey: nodeLabelValue,
|
ObjectMeta: metav1.ObjectMeta{
|
||||||
}
|
Name: nodeNames[1],
|
||||||
|
Labels: map[string]string{
|
||||||
node.Status.Allocatable[v1.ResourceEphemeralStorage] = *resource.NewQuantity(1000*1000*1000*1000, resource.DecimalSI)
|
nodeLabelKey: nodeLabelValue,
|
||||||
}),
|
},
|
||||||
test.BuildTestNode(nodeNames[2], 0, 0, 0, nil),
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
ObjectMeta: metav1.ObjectMeta{
|
||||||
|
Name: nodeNames[2],
|
||||||
|
},
|
||||||
|
},
|
||||||
},
|
},
|
||||||
podsOnNodes: []*v1.Pod{},
|
success: true,
|
||||||
success: true,
|
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
description: "Pod expected to fit none of the nodes",
|
description: "Pod expected to fit none of the nodes",
|
||||||
pod: test.BuildTestPod("p1", 0, 0, nodeNames[2], func(pod *v1.Pod) {
|
pod: createPodManifest(nodeNames[2], nodeLabelKey, nodeLabelValue),
|
||||||
pod.Spec.NodeSelector = map[string]string{
|
|
||||||
nodeLabelKey: nodeLabelValue,
|
|
||||||
}
|
|
||||||
}),
|
|
||||||
nodes: []*v1.Node{
|
nodes: []*v1.Node{
|
||||||
test.BuildTestNode(nodeNames[0], 64000, 128*1000*1000*1000, 200, func(node *v1.Node) {
|
{
|
||||||
node.ObjectMeta.Labels = map[string]string{
|
ObjectMeta: metav1.ObjectMeta{
|
||||||
nodeLabelKey: "unfit1",
|
Name: nodeNames[0],
|
||||||
}
|
Labels: map[string]string{
|
||||||
|
nodeLabelKey: "unfit1",
|
||||||
node.Status.Allocatable[v1.ResourceEphemeralStorage] = *resource.NewQuantity(1000*1000*1000*1000, resource.DecimalSI)
|
},
|
||||||
}),
|
},
|
||||||
test.BuildTestNode(nodeNames[1], 64000, 128*1000*1000*1000, 200, func(node *v1.Node) {
|
},
|
||||||
node.ObjectMeta.Labels = map[string]string{
|
{
|
||||||
nodeLabelKey: "unfit2",
|
ObjectMeta: metav1.ObjectMeta{
|
||||||
}
|
Name: nodeNames[1],
|
||||||
|
Labels: map[string]string{
|
||||||
node.Status.Allocatable[v1.ResourceEphemeralStorage] = *resource.NewQuantity(1000*1000*1000*1000, resource.DecimalSI)
|
nodeLabelKey: "unfit2",
|
||||||
}),
|
},
|
||||||
test.BuildTestNode(nodeNames[2], 0, 0, 0, nil),
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
ObjectMeta: metav1.ObjectMeta{
|
||||||
|
Name: nodeNames[2],
|
||||||
|
},
|
||||||
|
},
|
||||||
},
|
},
|
||||||
podsOnNodes: []*v1.Pod{},
|
success: false,
|
||||||
success: false,
|
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
description: "Nodes are unschedulable but labels match, should fail",
|
description: "Nodes are unschedulable but labels match, should fail",
|
||||||
pod: test.BuildTestPod("p1", 0, 0, nodeNames[2], func(pod *v1.Pod) {
|
pod: createPodManifest(nodeNames[2], nodeLabelKey, nodeLabelValue),
|
||||||
pod.Spec.NodeSelector = map[string]string{
|
|
||||||
nodeLabelKey: nodeLabelValue,
|
|
||||||
}
|
|
||||||
}),
|
|
||||||
nodes: []*v1.Node{
|
nodes: []*v1.Node{
|
||||||
test.BuildTestNode(nodeNames[0], 64000, 128*1000*1000*1000, 200, func(node *v1.Node) {
|
{
|
||||||
node.ObjectMeta.Labels = map[string]string{
|
ObjectMeta: metav1.ObjectMeta{
|
||||||
nodeLabelKey: nodeLabelValue,
|
Name: nodeNames[0],
|
||||||
}
|
Labels: map[string]string{
|
||||||
|
nodeLabelKey: nodeLabelValue,
|
||||||
node.Status.Allocatable[v1.ResourceEphemeralStorage] = *resource.NewQuantity(1000*1000*1000*1000, resource.DecimalSI)
|
},
|
||||||
node.Spec.Unschedulable = true
|
},
|
||||||
}),
|
Spec: v1.NodeSpec{
|
||||||
test.BuildTestNode(nodeNames[1], 64000, 128*1000*1000*1000, 200, func(node *v1.Node) {
|
Unschedulable: true,
|
||||||
node.ObjectMeta.Labels = map[string]string{
|
},
|
||||||
nodeLabelKey: "no",
|
},
|
||||||
}
|
{
|
||||||
|
ObjectMeta: metav1.ObjectMeta{
|
||||||
node.Status.Allocatable[v1.ResourceEphemeralStorage] = *resource.NewQuantity(1000*1000*1000*1000, resource.DecimalSI)
|
Name: nodeNames[1],
|
||||||
}),
|
Labels: map[string]string{
|
||||||
test.BuildTestNode(nodeNames[2], 0, 0, 0, nil),
|
nodeLabelKey: "no",
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
ObjectMeta: metav1.ObjectMeta{
|
||||||
|
Name: nodeNames[2],
|
||||||
|
},
|
||||||
|
},
|
||||||
},
|
},
|
||||||
podsOnNodes: []*v1.Pod{},
|
success: false,
|
||||||
success: false,
|
},
|
||||||
|
{
|
||||||
|
description: "Two nodes matches node selector, one of them is tained, should pass",
|
||||||
|
pod: createPodManifest(nodeNames[2], nodeLabelKey, nodeLabelValue),
|
||||||
|
nodes: []*v1.Node{
|
||||||
|
{
|
||||||
|
ObjectMeta: metav1.ObjectMeta{
|
||||||
|
Name: nodeNames[0],
|
||||||
|
Labels: map[string]string{
|
||||||
|
nodeLabelKey: nodeLabelValue,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
Spec: v1.NodeSpec{
|
||||||
|
Taints: []v1.Taint{
|
||||||
|
{
|
||||||
|
Key: nodeTaintKey,
|
||||||
|
Value: nodeTaintValue,
|
||||||
|
Effect: v1.TaintEffectNoSchedule,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
ObjectMeta: metav1.ObjectMeta{
|
||||||
|
Name: nodeNames[1],
|
||||||
|
Labels: map[string]string{
|
||||||
|
nodeLabelKey: nodeLabelValue,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
ObjectMeta: metav1.ObjectMeta{
|
||||||
|
Name: nodeNames[2],
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
success: true,
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
description: "Both nodes are tained, should fail",
|
description: "Both nodes are tained, should fail",
|
||||||
pod: test.BuildTestPod("p1", 2000, 2*1000*1000*1000, nodeNames[2], func(pod *v1.Pod) {
|
pod: createPodManifest(nodeNames[2], nodeLabelKey, nodeLabelValue),
|
||||||
pod.Spec.NodeSelector = map[string]string{
|
|
||||||
nodeLabelKey: nodeLabelValue,
|
|
||||||
}
|
|
||||||
pod.Spec.Containers[0].Resources.Requests[v1.ResourceEphemeralStorage] = *resource.NewQuantity(10*1000*1000*1000, resource.DecimalSI)
|
|
||||||
}),
|
|
||||||
nodes: []*v1.Node{
|
nodes: []*v1.Node{
|
||||||
test.BuildTestNode(nodeNames[0], 64000, 128*1000*1000*1000, 200, func(node *v1.Node) {
|
{
|
||||||
node.ObjectMeta.Labels = map[string]string{
|
ObjectMeta: metav1.ObjectMeta{
|
||||||
nodeLabelKey: nodeLabelValue,
|
Name: nodeNames[0],
|
||||||
}
|
|
||||||
|
|
||||||
node.Status.Allocatable[v1.ResourceEphemeralStorage] = *resource.NewQuantity(1000*1000*1000*1000, resource.DecimalSI)
|
|
||||||
node.Spec.Taints = []v1.Taint{
|
|
||||||
{
|
|
||||||
Key: nodeTaintKey,
|
|
||||||
Value: nodeTaintValue,
|
|
||||||
Effect: v1.TaintEffectNoSchedule,
|
|
||||||
},
|
|
||||||
}
|
|
||||||
}),
|
|
||||||
test.BuildTestNode(nodeNames[1], 64000, 128*1000*1000*1000, 200, func(node *v1.Node) {
|
|
||||||
node.ObjectMeta.Labels = map[string]string{
|
|
||||||
nodeLabelKey: nodeLabelValue,
|
|
||||||
}
|
|
||||||
|
|
||||||
node.Status.Allocatable[v1.ResourceEphemeralStorage] = *resource.NewQuantity(1000*1000*1000*1000, resource.DecimalSI)
|
|
||||||
node.Spec.Taints = []v1.Taint{
|
|
||||||
{
|
|
||||||
Key: nodeTaintKey,
|
|
||||||
Value: nodeTaintValue,
|
|
||||||
Effect: v1.TaintEffectNoSchedule,
|
|
||||||
},
|
|
||||||
}
|
|
||||||
}),
|
|
||||||
test.BuildTestNode(nodeNames[2], 0, 0, 0, nil),
|
|
||||||
},
|
|
||||||
podsOnNodes: []*v1.Pod{},
|
|
||||||
success: false,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
description: "Two nodes matches node selector, one of them is tained, there is a pod on the available node, and requests are low, should pass",
|
|
||||||
pod: test.BuildTestPod("p1", 2000, 2*1000*1000*1000, nodeNames[2], func(pod *v1.Pod) {
|
|
||||||
pod.Spec.NodeSelector = map[string]string{
|
|
||||||
nodeLabelKey: nodeLabelValue,
|
|
||||||
}
|
|
||||||
pod.Spec.Containers[0].Resources.Requests[v1.ResourceEphemeralStorage] = *resource.NewQuantity(10*1000*1000*1000, resource.DecimalSI)
|
|
||||||
}),
|
|
||||||
nodes: []*v1.Node{
|
|
||||||
test.BuildTestNode(nodeNames[0], 64000, 128*1000*1000*1000, 200, func(node *v1.Node) {
|
|
||||||
node.ObjectMeta.Labels = map[string]string{
|
|
||||||
nodeLabelKey: nodeLabelValue,
|
|
||||||
}
|
|
||||||
|
|
||||||
node.Status.Allocatable[v1.ResourceEphemeralStorage] = *resource.NewQuantity(1000*1000*1000*1000, resource.DecimalSI)
|
|
||||||
node.Spec.Taints = []v1.Taint{
|
|
||||||
{
|
|
||||||
Key: nodeTaintKey,
|
|
||||||
Value: nodeTaintValue,
|
|
||||||
Effect: v1.TaintEffectNoSchedule,
|
|
||||||
},
|
|
||||||
}
|
|
||||||
}),
|
|
||||||
test.BuildTestNode(nodeNames[1], 64000, 128*1000*1000*1000, 200, func(node *v1.Node) {
|
|
||||||
node.ObjectMeta.Labels = map[string]string{
|
|
||||||
nodeLabelKey: nodeLabelValue,
|
|
||||||
}
|
|
||||||
|
|
||||||
node.Status.Allocatable[v1.ResourceEphemeralStorage] = *resource.NewQuantity(1000*1000*1000*1000, resource.DecimalSI)
|
|
||||||
}),
|
|
||||||
test.BuildTestNode(nodeNames[2], 0, 0, 0, nil),
|
|
||||||
},
|
|
||||||
podsOnNodes: []*v1.Pod{
|
|
||||||
test.BuildTestPod("test-pod", 12*1000, 20*1000*1000*1000, nodeNames[1], func(pod *v1.Pod) {
|
|
||||||
pod.ObjectMeta = metav1.ObjectMeta{
|
|
||||||
Namespace: "test",
|
|
||||||
Labels: map[string]string{
|
Labels: map[string]string{
|
||||||
"test": "true",
|
nodeLabelKey: nodeLabelValue,
|
||||||
},
|
},
|
||||||
}
|
},
|
||||||
pod.Spec.Containers[0].Resources.Requests[v1.ResourceEphemeralStorage] = *resource.NewQuantity(40*1000*1000*1000, resource.DecimalSI)
|
Spec: v1.NodeSpec{
|
||||||
}),
|
Taints: []v1.Taint{
|
||||||
},
|
{
|
||||||
success: true,
|
Key: nodeTaintKey,
|
||||||
},
|
Value: nodeTaintValue,
|
||||||
{
|
Effect: v1.TaintEffectNoSchedule,
|
||||||
description: "Two nodes matches node selector, one of them is tained, but CPU requests are too big, should fail",
|
},
|
||||||
pod: test.BuildTestPod("p1", 2000, 2*1000*1000*1000, nodeNames[2], func(pod *v1.Pod) {
|
|
||||||
pod.Spec.NodeSelector = map[string]string{
|
|
||||||
nodeLabelKey: nodeLabelValue,
|
|
||||||
}
|
|
||||||
pod.Spec.Containers[0].Resources.Requests[v1.ResourceEphemeralStorage] = *resource.NewQuantity(10*1000*1000*1000, resource.DecimalSI)
|
|
||||||
}),
|
|
||||||
nodes: []*v1.Node{
|
|
||||||
test.BuildTestNode(nodeNames[0], 64000, 128*1000*1000*1000, 200, func(node *v1.Node) {
|
|
||||||
node.ObjectMeta.Labels = map[string]string{
|
|
||||||
nodeLabelKey: nodeLabelValue,
|
|
||||||
}
|
|
||||||
|
|
||||||
node.Status.Allocatable[v1.ResourceEphemeralStorage] = *resource.NewQuantity(1000*1000*1000*1000, resource.DecimalSI)
|
|
||||||
node.Spec.Taints = []v1.Taint{
|
|
||||||
{
|
|
||||||
Key: nodeTaintKey,
|
|
||||||
Value: nodeTaintValue,
|
|
||||||
Effect: v1.TaintEffectNoSchedule,
|
|
||||||
},
|
},
|
||||||
}
|
},
|
||||||
}),
|
},
|
||||||
// Notice that this node only has 4 cores, the pod already on the node below requests 3 cores, and the pod above requests 2 cores
|
{
|
||||||
test.BuildTestNode(nodeNames[1], 4000, 8*1000*1000*1000, 12, func(node *v1.Node) {
|
ObjectMeta: metav1.ObjectMeta{
|
||||||
node.ObjectMeta.Labels = map[string]string{
|
Name: nodeNames[1],
|
||||||
nodeLabelKey: nodeLabelValue,
|
|
||||||
}
|
|
||||||
|
|
||||||
node.Status.Allocatable[v1.ResourceEphemeralStorage] = *resource.NewQuantity(200*1000*1000*1000, resource.DecimalSI)
|
|
||||||
}),
|
|
||||||
test.BuildTestNode(nodeNames[2], 0, 0, 0, nil),
|
|
||||||
},
|
|
||||||
podsOnNodes: []*v1.Pod{
|
|
||||||
test.BuildTestPod("3-core-pod", 3000, 4*1000*1000*1000, nodeNames[1], func(pod *v1.Pod) {
|
|
||||||
pod.ObjectMeta = metav1.ObjectMeta{
|
|
||||||
Namespace: "test",
|
|
||||||
Labels: map[string]string{
|
Labels: map[string]string{
|
||||||
"test": "true",
|
nodeLabelKey: nodeLabelValue,
|
||||||
},
|
},
|
||||||
}
|
},
|
||||||
pod.Spec.Containers[0].Resources.Requests[v1.ResourceEphemeralStorage] = *resource.NewQuantity(10*1000*1000*1000, resource.DecimalSI)
|
Spec: v1.NodeSpec{
|
||||||
}),
|
Taints: []v1.Taint{
|
||||||
},
|
{
|
||||||
success: false,
|
Key: nodeTaintKey,
|
||||||
},
|
Value: nodeTaintValue,
|
||||||
{
|
Effect: v1.TaintEffectNoExecute,
|
||||||
description: "Two nodes matches node selector, one of them is tained, but memory requests are too big, should fail",
|
},
|
||||||
pod: test.BuildTestPod("p1", 2000, 5*1000*1000*1000, nodeNames[2], func(pod *v1.Pod) {
|
|
||||||
pod.Spec.NodeSelector = map[string]string{
|
|
||||||
nodeLabelKey: nodeLabelValue,
|
|
||||||
}
|
|
||||||
pod.Spec.Containers[0].Resources.Requests[v1.ResourceEphemeralStorage] = *resource.NewQuantity(10*1000*1000*1000, resource.DecimalSI)
|
|
||||||
}),
|
|
||||||
nodes: []*v1.Node{
|
|
||||||
test.BuildTestNode(nodeNames[0], 64000, 128*1000*1000*1000, 200, func(node *v1.Node) {
|
|
||||||
node.ObjectMeta.Labels = map[string]string{
|
|
||||||
nodeLabelKey: nodeLabelValue,
|
|
||||||
}
|
|
||||||
|
|
||||||
node.Status.Allocatable[v1.ResourceEphemeralStorage] = *resource.NewQuantity(1000*1000*1000*1000, resource.DecimalSI)
|
|
||||||
node.Spec.Taints = []v1.Taint{
|
|
||||||
{
|
|
||||||
Key: nodeTaintKey,
|
|
||||||
Value: nodeTaintValue,
|
|
||||||
Effect: v1.TaintEffectNoSchedule,
|
|
||||||
},
|
},
|
||||||
}
|
},
|
||||||
}),
|
},
|
||||||
// Notice that this node only has 8GB of memory, the pod already on the node below requests 4GB, and the pod above requests 5GB
|
{
|
||||||
test.BuildTestNode(nodeNames[1], 10*1000, 8*1000*1000*1000, 12, func(node *v1.Node) {
|
ObjectMeta: metav1.ObjectMeta{
|
||||||
node.ObjectMeta.Labels = map[string]string{
|
Name: nodeNames[2],
|
||||||
nodeLabelKey: nodeLabelValue,
|
},
|
||||||
}
|
},
|
||||||
|
|
||||||
node.Status.Allocatable[v1.ResourceEphemeralStorage] = *resource.NewQuantity(200*1000*1000*1000, resource.DecimalSI)
|
|
||||||
}),
|
|
||||||
test.BuildTestNode(nodeNames[2], 0, 0, 0, nil),
|
|
||||||
},
|
|
||||||
podsOnNodes: []*v1.Pod{
|
|
||||||
test.BuildTestPod("4GB-mem-pod", 2000, 4*1000*1000*1000, nodeNames[1], func(pod *v1.Pod) {
|
|
||||||
pod.ObjectMeta = metav1.ObjectMeta{
|
|
||||||
Namespace: "test",
|
|
||||||
Labels: map[string]string{
|
|
||||||
"test": "true",
|
|
||||||
},
|
|
||||||
}
|
|
||||||
pod.Spec.Containers[0].Resources.Requests[v1.ResourceEphemeralStorage] = *resource.NewQuantity(10*1000*1000*1000, resource.DecimalSI)
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
success: false,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
description: "Two nodes matches node selector, one of them is tained, but ephemeral storage requests are too big, should fail",
|
|
||||||
pod: test.BuildTestPod("p1", 2000, 4*1000*1000*1000, nodeNames[2], func(pod *v1.Pod) {
|
|
||||||
pod.Spec.NodeSelector = map[string]string{
|
|
||||||
nodeLabelKey: nodeLabelValue,
|
|
||||||
}
|
|
||||||
pod.Spec.Containers[0].Resources.Requests[v1.ResourceEphemeralStorage] = *resource.NewQuantity(10*1000*1000*1000, resource.DecimalSI)
|
|
||||||
}),
|
|
||||||
nodes: []*v1.Node{
|
|
||||||
test.BuildTestNode(nodeNames[0], 64000, 128*1000*1000*1000, 200, func(node *v1.Node) {
|
|
||||||
node.ObjectMeta.Labels = map[string]string{
|
|
||||||
nodeLabelKey: nodeLabelValue,
|
|
||||||
}
|
|
||||||
|
|
||||||
node.Status.Allocatable[v1.ResourceEphemeralStorage] = *resource.NewQuantity(1000*1000*1000*1000, resource.DecimalSI)
|
|
||||||
node.Spec.Taints = []v1.Taint{
|
|
||||||
{
|
|
||||||
Key: nodeTaintKey,
|
|
||||||
Value: nodeTaintValue,
|
|
||||||
Effect: v1.TaintEffectNoSchedule,
|
|
||||||
},
|
|
||||||
}
|
|
||||||
}),
|
|
||||||
// Notice that this node only has 20GB of storage, the pod already on the node below requests 11GB, and the pod above requests 10GB
|
|
||||||
test.BuildTestNode(nodeNames[1], 10*1000, 8*1000*1000*1000, 12, func(node *v1.Node) {
|
|
||||||
node.ObjectMeta.Labels = map[string]string{
|
|
||||||
nodeLabelKey: nodeLabelValue,
|
|
||||||
}
|
|
||||||
|
|
||||||
node.Status.Allocatable[v1.ResourceEphemeralStorage] = *resource.NewQuantity(20*1000*1000*1000, resource.DecimalSI)
|
|
||||||
}),
|
|
||||||
test.BuildTestNode(nodeNames[2], 0, 0, 0, nil),
|
|
||||||
},
|
|
||||||
podsOnNodes: []*v1.Pod{
|
|
||||||
test.BuildTestPod("11GB-storage-pod", 2000, 4*1000*1000*1000, nodeNames[1], func(pod *v1.Pod) {
|
|
||||||
pod.ObjectMeta = metav1.ObjectMeta{
|
|
||||||
Namespace: "test",
|
|
||||||
Labels: map[string]string{
|
|
||||||
"test": "true",
|
|
||||||
},
|
|
||||||
}
|
|
||||||
pod.Spec.Containers[0].Resources.Requests[v1.ResourceEphemeralStorage] = *resource.NewQuantity(11*1000*1000*1000, resource.DecimalSI)
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
success: false,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
description: "Two nodes matches node selector, one of them is tained, but custom resource requests are too big, should fail",
|
|
||||||
pod: test.BuildTestPod("p1", 2000, 2*1000*1000*1000, nodeNames[2], func(pod *v1.Pod) {
|
|
||||||
pod.Spec.NodeSelector = map[string]string{
|
|
||||||
nodeLabelKey: nodeLabelValue,
|
|
||||||
}
|
|
||||||
pod.Spec.Containers[0].Resources.Requests[v1.ResourceEphemeralStorage] = *resource.NewQuantity(10*1000*1000*1000, resource.DecimalSI)
|
|
||||||
pod.Spec.Containers[0].Resources.Requests["example.com/custom-resource"] = *resource.NewQuantity(10, resource.DecimalSI)
|
|
||||||
}),
|
|
||||||
nodes: []*v1.Node{
|
|
||||||
test.BuildTestNode(nodeNames[0], 64000, 128*1000*1000*1000, 200, func(node *v1.Node) {
|
|
||||||
node.ObjectMeta.Labels = map[string]string{
|
|
||||||
nodeLabelKey: nodeLabelValue,
|
|
||||||
}
|
|
||||||
|
|
||||||
node.Status.Allocatable[v1.ResourceEphemeralStorage] = *resource.NewQuantity(1000*1000*1000*1000, resource.DecimalSI)
|
|
||||||
node.Spec.Taints = []v1.Taint{
|
|
||||||
{
|
|
||||||
Key: nodeTaintKey,
|
|
||||||
Value: nodeTaintValue,
|
|
||||||
Effect: v1.TaintEffectNoSchedule,
|
|
||||||
},
|
|
||||||
}
|
|
||||||
node.Status.Allocatable["example.com/custom-resource"] = *resource.NewQuantity(15, resource.DecimalSI)
|
|
||||||
}),
|
|
||||||
// Notice that this node only has 15 of the custom resource, the pod already on the node below requests 10, and the pod above requests 10
|
|
||||||
test.BuildTestNode(nodeNames[1], 10*1000, 8*1000*1000*1000, 12, func(node *v1.Node) {
|
|
||||||
node.ObjectMeta.Labels = map[string]string{
|
|
||||||
nodeLabelKey: nodeLabelValue,
|
|
||||||
}
|
|
||||||
|
|
||||||
node.Status.Allocatable[v1.ResourceEphemeralStorage] = *resource.NewQuantity(200*1000*1000*1000, resource.DecimalSI)
|
|
||||||
node.Status.Allocatable["example.com/custom-resource"] = *resource.NewQuantity(15, resource.DecimalSI)
|
|
||||||
}),
|
|
||||||
test.BuildTestNode(nodeNames[2], 0, 0, 0, nil),
|
|
||||||
},
|
|
||||||
podsOnNodes: []*v1.Pod{
|
|
||||||
test.BuildTestPod("10-custom-resource-pod", 0, 0, nodeNames[1], func(pod *v1.Pod) {
|
|
||||||
pod.ObjectMeta = metav1.ObjectMeta{
|
|
||||||
Namespace: "test",
|
|
||||||
Labels: map[string]string{
|
|
||||||
"test": "true",
|
|
||||||
},
|
|
||||||
}
|
|
||||||
pod.Spec.Containers[0].Resources.Requests["example.com/custom-resource"] = *resource.NewQuantity(10, resource.DecimalSI)
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
success: false,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
description: "Two nodes matches node selector, one of them is tained, CPU requests will fit, and pod Overhead is low enough, should pass",
|
|
||||||
pod: test.BuildTestPod("p1", 1000, 2*1000*1000*1000, nodeNames[2], func(pod *v1.Pod) {
|
|
||||||
pod.Spec.NodeSelector = map[string]string{
|
|
||||||
nodeLabelKey: nodeLabelValue,
|
|
||||||
}
|
|
||||||
pod.Spec.Containers[0].Resources.Requests[v1.ResourceEphemeralStorage] = *resource.NewQuantity(10*1000*1000*1000, resource.DecimalSI)
|
|
||||||
}),
|
|
||||||
nodes: []*v1.Node{
|
|
||||||
test.BuildTestNode(nodeNames[0], 64000, 128*1000*1000*1000, 200, func(node *v1.Node) {
|
|
||||||
node.ObjectMeta.Labels = map[string]string{
|
|
||||||
nodeLabelKey: nodeLabelValue,
|
|
||||||
}
|
|
||||||
|
|
||||||
node.Status.Allocatable[v1.ResourceEphemeralStorage] = *resource.NewQuantity(1000*1000*1000*1000, resource.DecimalSI)
|
|
||||||
node.Spec.Taints = []v1.Taint{
|
|
||||||
{
|
|
||||||
Key: nodeTaintKey,
|
|
||||||
Value: nodeTaintValue,
|
|
||||||
Effect: v1.TaintEffectNoSchedule,
|
|
||||||
},
|
|
||||||
}
|
|
||||||
}),
|
|
||||||
// Notice that this node has 5 CPU cores, the pod below requests 2 cores, and has CPU overhead of 1 cores, and the pod above requests 1 core
|
|
||||||
test.BuildTestNode(nodeNames[1], 5000, 8*1000*1000*1000, 12, func(node *v1.Node) {
|
|
||||||
node.ObjectMeta.Labels = map[string]string{
|
|
||||||
nodeLabelKey: nodeLabelValue,
|
|
||||||
}
|
|
||||||
|
|
||||||
node.Status.Allocatable[v1.ResourceEphemeralStorage] = *resource.NewQuantity(200*1000*1000*1000, resource.DecimalSI)
|
|
||||||
}),
|
|
||||||
test.BuildTestNode(nodeNames[2], 0, 0, 0, nil),
|
|
||||||
},
|
|
||||||
podsOnNodes: []*v1.Pod{
|
|
||||||
test.BuildTestPod("3-core-pod", 2000, 4*1000*1000*1000, nodeNames[1], func(pod *v1.Pod) {
|
|
||||||
pod.ObjectMeta = metav1.ObjectMeta{
|
|
||||||
Namespace: "test",
|
|
||||||
Labels: map[string]string{
|
|
||||||
"test": "true",
|
|
||||||
},
|
|
||||||
}
|
|
||||||
pod.Spec.Containers[0].Resources.Requests[v1.ResourceEphemeralStorage] = *resource.NewQuantity(10*1000*1000*1000, resource.DecimalSI)
|
|
||||||
pod.Spec.Overhead = createResourceList(1000, 1000*1000*1000, 1000*1000*1000)
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
success: true,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
description: "Two nodes matches node selector, one of them is tained, CPU requests will fit, but pod Overhead is too high, should fail",
|
|
||||||
pod: test.BuildTestPod("p1", 2000, 2*1000*1000*1000, nodeNames[2], func(pod *v1.Pod) {
|
|
||||||
pod.Spec.NodeSelector = map[string]string{
|
|
||||||
nodeLabelKey: nodeLabelValue,
|
|
||||||
}
|
|
||||||
pod.Spec.Containers[0].Resources.Requests[v1.ResourceEphemeralStorage] = *resource.NewQuantity(10*1000*1000*1000, resource.DecimalSI)
|
|
||||||
}),
|
|
||||||
nodes: []*v1.Node{
|
|
||||||
test.BuildTestNode(nodeNames[0], 64000, 128*1000*1000*1000, 200, func(node *v1.Node) {
|
|
||||||
node.ObjectMeta.Labels = map[string]string{
|
|
||||||
nodeLabelKey: nodeLabelValue,
|
|
||||||
}
|
|
||||||
|
|
||||||
node.Status.Allocatable[v1.ResourceEphemeralStorage] = *resource.NewQuantity(1000*1000*1000*1000, resource.DecimalSI)
|
|
||||||
node.Spec.Taints = []v1.Taint{
|
|
||||||
{
|
|
||||||
Key: nodeTaintKey,
|
|
||||||
Value: nodeTaintValue,
|
|
||||||
Effect: v1.TaintEffectNoSchedule,
|
|
||||||
},
|
|
||||||
}
|
|
||||||
}),
|
|
||||||
// Notice that this node only has 5 CPU cores, the pod below requests 2 cores, but has CPU overhead of 2 cores, and the pod above requests 2 cores
|
|
||||||
test.BuildTestNode(nodeNames[1], 5000, 8*1000*1000*1000, 12, func(node *v1.Node) {
|
|
||||||
node.ObjectMeta.Labels = map[string]string{
|
|
||||||
nodeLabelKey: nodeLabelValue,
|
|
||||||
}
|
|
||||||
|
|
||||||
node.Status.Allocatable[v1.ResourceEphemeralStorage] = *resource.NewQuantity(200*1000*1000*1000, resource.DecimalSI)
|
|
||||||
}),
|
|
||||||
test.BuildTestNode(nodeNames[2], 0, 0, 0, nil),
|
|
||||||
},
|
|
||||||
podsOnNodes: []*v1.Pod{
|
|
||||||
test.BuildTestPod("3-core-pod", 2000, 4*1000*1000*1000, nodeNames[1], func(pod *v1.Pod) {
|
|
||||||
pod.ObjectMeta = metav1.ObjectMeta{
|
|
||||||
Namespace: "test",
|
|
||||||
Labels: map[string]string{
|
|
||||||
"test": "true",
|
|
||||||
},
|
|
||||||
}
|
|
||||||
pod.Spec.Containers[0].Resources.Requests[v1.ResourceEphemeralStorage] = *resource.NewQuantity(10*1000*1000*1000, resource.DecimalSI)
|
|
||||||
pod.Spec.Overhead = createResourceList(2000, 1000*1000*1000, 1000*1000*1000)
|
|
||||||
}),
|
|
||||||
},
|
},
|
||||||
success: false,
|
success: false,
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
for _, tc := range tests {
|
for _, tc := range tests {
|
||||||
t.Run(tc.description, func(t *testing.T) {
|
actual := PodFitsAnyOtherNode(tc.pod, tc.nodes)
|
||||||
ctx, cancel := context.WithCancel(context.Background())
|
if actual != tc.success {
|
||||||
defer cancel()
|
t.Errorf("Test %#v failed", tc.description)
|
||||||
|
}
|
||||||
var objs []runtime.Object
|
|
||||||
for _, node := range tc.nodes {
|
|
||||||
objs = append(objs, node)
|
|
||||||
}
|
|
||||||
for _, pod := range tc.podsOnNodes {
|
|
||||||
objs = append(objs, pod)
|
|
||||||
}
|
|
||||||
objs = append(objs, tc.pod)
|
|
||||||
|
|
||||||
fakeClient := fake.NewSimpleClientset(objs...)
|
|
||||||
|
|
||||||
sharedInformerFactory := informers.NewSharedInformerFactory(fakeClient, 0)
|
|
||||||
podInformer := sharedInformerFactory.Core().V1().Pods()
|
|
||||||
|
|
||||||
getPodsAssignedToNode, err := podutil.BuildGetPodsAssignedToNodeFunc(podInformer)
|
|
||||||
if err != nil {
|
|
||||||
t.Errorf("Build get pods assigned to node function error: %v", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
sharedInformerFactory.Start(ctx.Done())
|
|
||||||
sharedInformerFactory.WaitForCacheSync(ctx.Done())
|
|
||||||
|
|
||||||
actual := PodFitsAnyOtherNode(getPodsAssignedToNode, tc.pod, tc.nodes)
|
|
||||||
if actual != tc.success {
|
|
||||||
t.Errorf("Test %#v failed", tc.description)
|
|
||||||
}
|
|
||||||
})
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// createResourceList builds a small resource list of core resources
|
func createPodManifest(nodeName string, nodeSelectorKey string, nodeSelectorValue string) *v1.Pod {
|
||||||
func createResourceList(cpu int64, memory int64, ephemeralStorage int64) v1.ResourceList {
|
return (&v1.Pod{
|
||||||
resourceList := make(map[v1.ResourceName]resource.Quantity)
|
Spec: v1.PodSpec{
|
||||||
resourceList[v1.ResourceCPU] = *resource.NewMilliQuantity(cpu, resource.DecimalSI)
|
NodeName: nodeName,
|
||||||
resourceList[v1.ResourceMemory] = *resource.NewQuantity(memory, resource.DecimalSI)
|
Affinity: &v1.Affinity{
|
||||||
resourceList[v1.ResourceEphemeralStorage] = *resource.NewQuantity(ephemeralStorage, resource.DecimalSI)
|
NodeAffinity: &v1.NodeAffinity{
|
||||||
return resourceList
|
RequiredDuringSchedulingIgnoredDuringExecution: &v1.NodeSelector{
|
||||||
|
NodeSelectorTerms: []v1.NodeSelectorTerm{
|
||||||
|
{
|
||||||
|
MatchExpressions: []v1.NodeSelectorRequirement{
|
||||||
|
{
|
||||||
|
Key: nodeSelectorKey,
|
||||||
|
Operator: "In",
|
||||||
|
Values: []string{
|
||||||
|
nodeSelectorValue,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
})
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -17,174 +17,150 @@ limitations under the License.
|
|||||||
package pod
|
package pod
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
"context"
|
||||||
"sort"
|
"sort"
|
||||||
|
|
||||||
v1 "k8s.io/api/core/v1"
|
v1 "k8s.io/api/core/v1"
|
||||||
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
||||||
"k8s.io/apimachinery/pkg/labels"
|
"k8s.io/apimachinery/pkg/fields"
|
||||||
"k8s.io/apimachinery/pkg/util/sets"
|
clientset "k8s.io/client-go/kubernetes"
|
||||||
coreinformers "k8s.io/client-go/informers/core/v1"
|
|
||||||
"k8s.io/client-go/tools/cache"
|
|
||||||
|
|
||||||
"sigs.k8s.io/descheduler/pkg/utils"
|
"sigs.k8s.io/descheduler/pkg/utils"
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
|
||||||
nodeNameKeyIndex = "spec.nodeName"
|
|
||||||
)
|
|
||||||
|
|
||||||
// FilterFunc is a filter for a pod.
|
|
||||||
type FilterFunc func(*v1.Pod) bool
|
|
||||||
|
|
||||||
// GetPodsAssignedToNodeFunc is a function which accept a node name and a pod filter function
|
|
||||||
// as input and returns the pods that assigned to the node.
|
|
||||||
type GetPodsAssignedToNodeFunc func(string, FilterFunc) ([]*v1.Pod, error)
|
|
||||||
|
|
||||||
// WrapFilterFuncs wraps a set of FilterFunc in one.
|
|
||||||
func WrapFilterFuncs(filters ...FilterFunc) FilterFunc {
|
|
||||||
return func(pod *v1.Pod) bool {
|
|
||||||
for _, filter := range filters {
|
|
||||||
if filter != nil && !filter(pod) {
|
|
||||||
return false
|
|
||||||
}
|
|
||||||
}
|
|
||||||
return true
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
type Options struct {
|
type Options struct {
|
||||||
filter FilterFunc
|
filter func(pod *v1.Pod) bool
|
||||||
includedNamespaces sets.String
|
includedNamespaces []string
|
||||||
excludedNamespaces sets.String
|
excludedNamespaces []string
|
||||||
labelSelector *metav1.LabelSelector
|
labelSelector *metav1.LabelSelector
|
||||||
}
|
}
|
||||||
|
|
||||||
// NewOptions returns an empty Options.
|
|
||||||
func NewOptions() *Options {
|
|
||||||
return &Options{}
|
|
||||||
}
|
|
||||||
|
|
||||||
// WithFilter sets a pod filter.
|
// WithFilter sets a pod filter.
|
||||||
// The filter function should return true if the pod should be returned from ListPodsOnANode
|
// The filter function should return true if the pod should be returned from ListPodsOnANode
|
||||||
func (o *Options) WithFilter(filter FilterFunc) *Options {
|
func WithFilter(filter func(pod *v1.Pod) bool) func(opts *Options) {
|
||||||
o.filter = filter
|
return func(opts *Options) {
|
||||||
return o
|
opts.filter = filter
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// WithNamespaces sets included namespaces
|
// WithNamespaces sets included namespaces
|
||||||
func (o *Options) WithNamespaces(namespaces sets.String) *Options {
|
func WithNamespaces(namespaces []string) func(opts *Options) {
|
||||||
o.includedNamespaces = namespaces
|
return func(opts *Options) {
|
||||||
return o
|
opts.includedNamespaces = namespaces
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// WithoutNamespaces sets excluded namespaces
|
// WithoutNamespaces sets excluded namespaces
|
||||||
func (o *Options) WithoutNamespaces(namespaces sets.String) *Options {
|
func WithoutNamespaces(namespaces []string) func(opts *Options) {
|
||||||
o.excludedNamespaces = namespaces
|
return func(opts *Options) {
|
||||||
return o
|
opts.excludedNamespaces = namespaces
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// WithLabelSelector sets a pod label selector
|
// WithLabelSelector sets a pod label selector
|
||||||
func (o *Options) WithLabelSelector(labelSelector *metav1.LabelSelector) *Options {
|
func WithLabelSelector(labelSelector *metav1.LabelSelector) func(opts *Options) {
|
||||||
o.labelSelector = labelSelector
|
return func(opts *Options) {
|
||||||
return o
|
opts.labelSelector = labelSelector
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// BuildFilterFunc builds a final FilterFunc based on Options.
|
// ListPodsOnANode lists all of the pods on a node
|
||||||
func (o *Options) BuildFilterFunc() (FilterFunc, error) {
|
// It also accepts an optional "filter" function which can be used to further limit the pods that are returned.
|
||||||
var s labels.Selector
|
// (Usually this is podEvictor.Evictable().IsEvictable, in order to only list the evictable pods on a node, but can
|
||||||
var err error
|
// be used by strategies to extend it if there are further restrictions, such as with NodeAffinity).
|
||||||
if o.labelSelector != nil {
|
func ListPodsOnANode(
|
||||||
s, err = metav1.LabelSelectorAsSelector(o.labelSelector)
|
ctx context.Context,
|
||||||
if err != nil {
|
client clientset.Interface,
|
||||||
return nil, err
|
node *v1.Node,
|
||||||
}
|
opts ...func(opts *Options),
|
||||||
}
|
) ([]*v1.Pod, error) {
|
||||||
return func(pod *v1.Pod) bool {
|
fieldSelectorString := "spec.nodeName=" + node.Name + ",status.phase!=" + string(v1.PodSucceeded) + ",status.phase!=" + string(v1.PodFailed)
|
||||||
if o.filter != nil && !o.filter(pod) {
|
|
||||||
return false
|
return ListPodsOnANodeWithFieldSelector(ctx, client, node, fieldSelectorString, opts...)
|
||||||
}
|
|
||||||
if len(o.includedNamespaces) > 0 && !o.includedNamespaces.Has(pod.Namespace) {
|
|
||||||
return false
|
|
||||||
}
|
|
||||||
if len(o.excludedNamespaces) > 0 && o.excludedNamespaces.Has(pod.Namespace) {
|
|
||||||
return false
|
|
||||||
}
|
|
||||||
if s != nil && !s.Matches(labels.Set(pod.GetLabels())) {
|
|
||||||
return false
|
|
||||||
}
|
|
||||||
return true
|
|
||||||
}, nil
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// BuildGetPodsAssignedToNodeFunc establishes an indexer to map the pods and their assigned nodes.
|
// ListPodsOnANodeWithFieldSelector lists all of the pods on a node using the filter selectors
|
||||||
// It returns a function to help us get all the pods that assigned to a node based on the indexer.
|
func ListPodsOnANodeWithFieldSelector(
|
||||||
func BuildGetPodsAssignedToNodeFunc(podInformer coreinformers.PodInformer) (GetPodsAssignedToNodeFunc, error) {
|
ctx context.Context,
|
||||||
// Establish an indexer to map the pods and their assigned nodes.
|
client clientset.Interface,
|
||||||
err := podInformer.Informer().AddIndexers(cache.Indexers{
|
node *v1.Node,
|
||||||
nodeNameKeyIndex: func(obj interface{}) ([]string, error) {
|
fieldSelectorString string,
|
||||||
pod, ok := obj.(*v1.Pod)
|
opts ...func(opts *Options),
|
||||||
if !ok {
|
) ([]*v1.Pod, error) {
|
||||||
return []string{}, nil
|
options := &Options{}
|
||||||
}
|
for _, opt := range opts {
|
||||||
if len(pod.Spec.NodeName) == 0 {
|
opt(options)
|
||||||
return []string{}, nil
|
|
||||||
}
|
|
||||||
return []string{pod.Spec.NodeName}, nil
|
|
||||||
},
|
|
||||||
})
|
|
||||||
if err != nil {
|
|
||||||
return nil, err
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// The indexer helps us get all the pods that assigned to a node.
|
pods := make([]*v1.Pod, 0)
|
||||||
podIndexer := podInformer.Informer().GetIndexer()
|
|
||||||
getPodsAssignedToNode := func(nodeName string, filter FilterFunc) ([]*v1.Pod, error) {
|
labelSelectorString := ""
|
||||||
objs, err := podIndexer.ByIndex(nodeNameKeyIndex, nodeName)
|
if options.labelSelector != nil {
|
||||||
|
selector, err := metav1.LabelSelectorAsSelector(options.labelSelector)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return []*v1.Pod{}, err
|
||||||
}
|
}
|
||||||
pods := make([]*v1.Pod, 0, len(objs))
|
labelSelectorString = selector.String()
|
||||||
for _, obj := range objs {
|
}
|
||||||
pod, ok := obj.(*v1.Pod)
|
|
||||||
if !ok {
|
if len(options.includedNamespaces) > 0 {
|
||||||
continue
|
fieldSelector, err := fields.ParseSelector(fieldSelectorString)
|
||||||
|
if err != nil {
|
||||||
|
return []*v1.Pod{}, err
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, namespace := range options.includedNamespaces {
|
||||||
|
podList, err := client.CoreV1().Pods(namespace).List(ctx,
|
||||||
|
metav1.ListOptions{
|
||||||
|
FieldSelector: fieldSelector.String(),
|
||||||
|
LabelSelector: labelSelectorString,
|
||||||
|
})
|
||||||
|
if err != nil {
|
||||||
|
return []*v1.Pod{}, err
|
||||||
}
|
}
|
||||||
if filter(pod) {
|
for i := range podList.Items {
|
||||||
pods = append(pods, pod)
|
if options.filter != nil && !options.filter(&podList.Items[i]) {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
pods = append(pods, &podList.Items[i])
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
return pods, nil
|
return pods, nil
|
||||||
}
|
}
|
||||||
return getPodsAssignedToNode, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// ListPodsOnANode lists all pods on a node.
|
if len(options.excludedNamespaces) > 0 {
|
||||||
// It also accepts a "filter" function which can be used to further limit the pods that are returned.
|
for _, namespace := range options.excludedNamespaces {
|
||||||
// (Usually this is podEvictor.Evictable().IsEvictable, in order to only list the evictable pods on a node, but can
|
fieldSelectorString += ",metadata.namespace!=" + namespace
|
||||||
// be used by strategies to extend it if there are further restrictions, such as with NodeAffinity).
|
}
|
||||||
func ListPodsOnANode(
|
|
||||||
nodeName string,
|
|
||||||
getPodsAssignedToNode GetPodsAssignedToNodeFunc,
|
|
||||||
filter FilterFunc,
|
|
||||||
) ([]*v1.Pod, error) {
|
|
||||||
// Succeeded and failed pods are not considered because they don't occupy any resource.
|
|
||||||
f := func(pod *v1.Pod) bool {
|
|
||||||
return pod.Status.Phase != v1.PodSucceeded && pod.Status.Phase != v1.PodFailed
|
|
||||||
}
|
}
|
||||||
return ListAllPodsOnANode(nodeName, getPodsAssignedToNode, WrapFilterFuncs(f, filter))
|
|
||||||
}
|
|
||||||
|
|
||||||
// ListAllPodsOnANode lists all the pods on a node no matter what the phase of the pod is.
|
fieldSelector, err := fields.ParseSelector(fieldSelectorString)
|
||||||
func ListAllPodsOnANode(
|
|
||||||
nodeName string,
|
|
||||||
getPodsAssignedToNode GetPodsAssignedToNodeFunc,
|
|
||||||
filter FilterFunc,
|
|
||||||
) ([]*v1.Pod, error) {
|
|
||||||
pods, err := getPodsAssignedToNode(nodeName, filter)
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return []*v1.Pod{}, err
|
return []*v1.Pod{}, err
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// INFO(jchaloup): field selectors do not work properly with listers
|
||||||
|
// Once the descheduler switches to pod listers (through informers),
|
||||||
|
// We need to flip to client-side filtering.
|
||||||
|
podList, err := client.CoreV1().Pods(v1.NamespaceAll).List(ctx,
|
||||||
|
metav1.ListOptions{
|
||||||
|
FieldSelector: fieldSelector.String(),
|
||||||
|
LabelSelector: labelSelectorString,
|
||||||
|
})
|
||||||
|
if err != nil {
|
||||||
|
return []*v1.Pod{}, err
|
||||||
|
}
|
||||||
|
|
||||||
|
for i := range podList.Items {
|
||||||
|
// fake client does not support field selectors
|
||||||
|
// so let's filter based on the node name as well (quite cheap)
|
||||||
|
if podList.Items[i].Spec.NodeName != node.Name {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
if options.filter != nil && !options.filter(&podList.Items[i]) {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
pods = append(pods, &podList.Items[i])
|
||||||
|
}
|
||||||
return pods, nil
|
return pods, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -228,10 +204,3 @@ func SortPodsBasedOnPriorityLowToHigh(pods []*v1.Pod) {
|
|||||||
return *pods[i].Spec.Priority < *pods[j].Spec.Priority
|
return *pods[i].Spec.Priority < *pods[j].Spec.Priority
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
// SortPodsBasedOnAge sorts Pods from oldest to most recent in place
|
|
||||||
func SortPodsBasedOnAge(pods []*v1.Pod) {
|
|
||||||
sort.Slice(pods, func(i, j int) bool {
|
|
||||||
return pods[i].CreationTimestamp.Before(&pods[j].CreationTimestamp)
|
|
||||||
})
|
|
||||||
}
|
|
||||||
|
|||||||
@@ -20,15 +20,14 @@ import (
|
|||||||
"context"
|
"context"
|
||||||
"fmt"
|
"fmt"
|
||||||
"reflect"
|
"reflect"
|
||||||
|
"strings"
|
||||||
"testing"
|
"testing"
|
||||||
"time"
|
|
||||||
|
|
||||||
v1 "k8s.io/api/core/v1"
|
v1 "k8s.io/api/core/v1"
|
||||||
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
||||||
"k8s.io/apimachinery/pkg/runtime"
|
"k8s.io/apimachinery/pkg/runtime"
|
||||||
"k8s.io/client-go/informers"
|
|
||||||
"k8s.io/client-go/kubernetes/fake"
|
"k8s.io/client-go/kubernetes/fake"
|
||||||
|
core "k8s.io/client-go/testing"
|
||||||
"sigs.k8s.io/descheduler/test"
|
"sigs.k8s.io/descheduler/test"
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -40,17 +39,19 @@ var (
|
|||||||
func TestListPodsOnANode(t *testing.T) {
|
func TestListPodsOnANode(t *testing.T) {
|
||||||
testCases := []struct {
|
testCases := []struct {
|
||||||
name string
|
name string
|
||||||
pods []*v1.Pod
|
pods map[string][]v1.Pod
|
||||||
node *v1.Node
|
node *v1.Node
|
||||||
labelSelector *metav1.LabelSelector
|
labelSelector *metav1.LabelSelector
|
||||||
expectedPodCount int
|
expectedPodCount int
|
||||||
}{
|
}{
|
||||||
{
|
{
|
||||||
name: "test listing pods on a node",
|
name: "test listing pods on a node",
|
||||||
pods: []*v1.Pod{
|
pods: map[string][]v1.Pod{
|
||||||
test.BuildTestPod("pod1", 100, 0, "n1", nil),
|
"n1": {
|
||||||
test.BuildTestPod("pod2", 100, 0, "n1", nil),
|
*test.BuildTestPod("pod1", 100, 0, "n1", nil),
|
||||||
test.BuildTestPod("pod3", 100, 0, "n2", nil),
|
*test.BuildTestPod("pod2", 100, 0, "n1", nil),
|
||||||
|
},
|
||||||
|
"n2": {*test.BuildTestPod("pod3", 100, 0, "n2", nil)},
|
||||||
},
|
},
|
||||||
node: test.BuildTestNode("n1", 2000, 3000, 10, nil),
|
node: test.BuildTestNode("n1", 2000, 3000, 10, nil),
|
||||||
labelSelector: nil,
|
labelSelector: nil,
|
||||||
@@ -58,15 +59,17 @@ func TestListPodsOnANode(t *testing.T) {
|
|||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "test listing pods with label selector",
|
name: "test listing pods with label selector",
|
||||||
pods: []*v1.Pod{
|
pods: map[string][]v1.Pod{
|
||||||
test.BuildTestPod("pod1", 100, 0, "n1", nil),
|
"n1": {
|
||||||
test.BuildTestPod("pod2", 100, 0, "n1", func(pod *v1.Pod) {
|
*test.BuildTestPod("pod1", 100, 0, "n1", nil),
|
||||||
pod.Labels = map[string]string{"foo": "bar"}
|
*test.BuildTestPod("pod2", 100, 0, "n1", func(pod *v1.Pod) {
|
||||||
}),
|
pod.Labels = map[string]string{"foo": "bar"}
|
||||||
test.BuildTestPod("pod3", 100, 0, "n1", func(pod *v1.Pod) {
|
}),
|
||||||
pod.Labels = map[string]string{"foo": "bar1"}
|
*test.BuildTestPod("pod3", 100, 0, "n1", func(pod *v1.Pod) {
|
||||||
}),
|
pod.Labels = map[string]string{"foo": "bar1"}
|
||||||
test.BuildTestPod("pod4", 100, 0, "n2", nil),
|
}),
|
||||||
|
},
|
||||||
|
"n2": {*test.BuildTestPod("pod4", 100, 0, "n2", nil)},
|
||||||
},
|
},
|
||||||
node: test.BuildTestNode("n1", 2000, 3000, 10, nil),
|
node: test.BuildTestNode("n1", 2000, 3000, 10, nil),
|
||||||
labelSelector: &metav1.LabelSelector{
|
labelSelector: &metav1.LabelSelector{
|
||||||
@@ -82,38 +85,21 @@ func TestListPodsOnANode(t *testing.T) {
|
|||||||
},
|
},
|
||||||
}
|
}
|
||||||
for _, testCase := range testCases {
|
for _, testCase := range testCases {
|
||||||
t.Run(testCase.name, func(t *testing.T) {
|
fakeClient := &fake.Clientset{}
|
||||||
ctx, cancel := context.WithCancel(context.Background())
|
fakeClient.Fake.AddReactor("list", "pods", func(action core.Action) (bool, runtime.Object, error) {
|
||||||
defer cancel()
|
list := action.(core.ListAction)
|
||||||
|
fieldString := list.GetListRestrictions().Fields.String()
|
||||||
var objs []runtime.Object
|
if strings.Contains(fieldString, "n1") {
|
||||||
objs = append(objs, testCase.node)
|
return true, &v1.PodList{Items: testCase.pods["n1"]}, nil
|
||||||
for _, pod := range testCase.pods {
|
} else if strings.Contains(fieldString, "n2") {
|
||||||
objs = append(objs, pod)
|
return true, &v1.PodList{Items: testCase.pods["n2"]}, nil
|
||||||
}
|
|
||||||
fakeClient := fake.NewSimpleClientset(objs...)
|
|
||||||
|
|
||||||
sharedInformerFactory := informers.NewSharedInformerFactory(fakeClient, 0)
|
|
||||||
podInformer := sharedInformerFactory.Core().V1().Pods()
|
|
||||||
|
|
||||||
getPodsAssignedToNode, err := BuildGetPodsAssignedToNodeFunc(podInformer)
|
|
||||||
if err != nil {
|
|
||||||
t.Errorf("Build get pods assigned to node function error: %v", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
sharedInformerFactory.Start(ctx.Done())
|
|
||||||
sharedInformerFactory.WaitForCacheSync(ctx.Done())
|
|
||||||
|
|
||||||
filter, err := NewOptions().WithLabelSelector(testCase.labelSelector).BuildFilterFunc()
|
|
||||||
if err != nil {
|
|
||||||
t.Errorf("Build filter function error: %v", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
pods, _ := ListPodsOnANode(testCase.node.Name, getPodsAssignedToNode, filter)
|
|
||||||
if len(pods) != testCase.expectedPodCount {
|
|
||||||
t.Errorf("Expected %v pods on node %v, got %+v", testCase.expectedPodCount, testCase.node.Name, len(pods))
|
|
||||||
}
|
}
|
||||||
|
return true, nil, fmt.Errorf("Failed to list: %v", list)
|
||||||
})
|
})
|
||||||
|
pods, _ := ListPodsOnANode(context.TODO(), fakeClient, testCase.node, WithLabelSelector(testCase.labelSelector))
|
||||||
|
if len(pods) != testCase.expectedPodCount {
|
||||||
|
t.Errorf("expected %v pods on node %v, got %+v", testCase.expectedPodCount, testCase.node.Name, len(pods))
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -156,23 +142,3 @@ func TestSortPodsBasedOnPriorityLowToHigh(t *testing.T) {
|
|||||||
t.Errorf("Expected last pod in sorted list to be %v which of highest priority and guaranteed but got %v", p4, podList[len(podList)-1])
|
t.Errorf("Expected last pod in sorted list to be %v which of highest priority and guaranteed but got %v", p4, podList[len(podList)-1])
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestSortPodsBasedOnAge(t *testing.T) {
|
|
||||||
podList := make([]*v1.Pod, 9)
|
|
||||||
n1 := test.BuildTestNode("n1", 4000, 3000, int64(len(podList)), nil)
|
|
||||||
|
|
||||||
for i := 0; i < len(podList); i++ {
|
|
||||||
podList[i] = test.BuildTestPod(fmt.Sprintf("p%d", i), 1, 32, n1.Name, func(pod *v1.Pod) {
|
|
||||||
creationTimestamp := metav1.Now().Add(time.Minute * time.Duration(-i))
|
|
||||||
pod.ObjectMeta.SetCreationTimestamp(metav1.NewTime(creationTimestamp))
|
|
||||||
})
|
|
||||||
}
|
|
||||||
|
|
||||||
SortPodsBasedOnAge(podList)
|
|
||||||
|
|
||||||
for i := 0; i < len(podList)-1; i++ {
|
|
||||||
if podList[i+1].CreationTimestamp.Before(&podList[i].CreationTimestamp) {
|
|
||||||
t.Errorf("Expected pods to be sorted by age but pod at index %d was older than %d", i+1, i)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|||||||
@@ -1,5 +1,5 @@
|
|||||||
/*
|
/*
|
||||||
Copyright 2022 The Kubernetes Authors.
|
Copyright 2017 The Kubernetes Authors.
|
||||||
|
|
||||||
Licensed under the Apache License, Version 2.0 (the "License");
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
you may not use this file except in compliance with the License.
|
you may not use this file except in compliance with the License.
|
||||||
@@ -14,7 +14,7 @@ See the License for the specific language governing permissions and
|
|||||||
limitations under the License.
|
limitations under the License.
|
||||||
*/
|
*/
|
||||||
|
|
||||||
package removeduplicates
|
package strategies
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
@@ -24,76 +24,72 @@ import (
|
|||||||
"sort"
|
"sort"
|
||||||
"strings"
|
"strings"
|
||||||
|
|
||||||
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
|
||||||
"sigs.k8s.io/descheduler/pkg/utils"
|
|
||||||
|
|
||||||
v1 "k8s.io/api/core/v1"
|
v1 "k8s.io/api/core/v1"
|
||||||
"k8s.io/apimachinery/pkg/runtime"
|
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
||||||
"k8s.io/apimachinery/pkg/util/sets"
|
"k8s.io/apimachinery/pkg/util/sets"
|
||||||
|
clientset "k8s.io/client-go/kubernetes"
|
||||||
"k8s.io/klog/v2"
|
"k8s.io/klog/v2"
|
||||||
|
|
||||||
"sigs.k8s.io/descheduler/pkg/apis/componentconfig"
|
"sigs.k8s.io/descheduler/pkg/api"
|
||||||
"sigs.k8s.io/descheduler/pkg/descheduler/evictions"
|
"sigs.k8s.io/descheduler/pkg/descheduler/evictions"
|
||||||
podutil "sigs.k8s.io/descheduler/pkg/descheduler/pod"
|
podutil "sigs.k8s.io/descheduler/pkg/descheduler/pod"
|
||||||
"sigs.k8s.io/descheduler/pkg/framework"
|
"sigs.k8s.io/descheduler/pkg/utils"
|
||||||
)
|
)
|
||||||
|
|
||||||
const PluginName = "RemoveDuplicates"
|
func validateRemoveDuplicatePodsParams(params *api.StrategyParameters) error {
|
||||||
|
if params == nil {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
// At most one of include/exclude can be set
|
||||||
|
if params.Namespaces != nil && len(params.Namespaces.Include) > 0 && len(params.Namespaces.Exclude) > 0 {
|
||||||
|
return fmt.Errorf("only one of Include/Exclude namespaces can be set")
|
||||||
|
}
|
||||||
|
if params.ThresholdPriority != nil && params.ThresholdPriorityClassName != "" {
|
||||||
|
return fmt.Errorf("only one of thresholdPriority and thresholdPriorityClassName can be set")
|
||||||
|
}
|
||||||
|
|
||||||
// RemoveDuplicatePods removes the duplicate pods on node. This plugin evicts all duplicate pods on node.
|
return nil
|
||||||
// A pod is said to be a duplicate of other if both of them are from same creator, kind and are within the same
|
|
||||||
// namespace, and have at least one container with the same image.
|
|
||||||
// As of now, this plugin won't evict daemonsets, mirror pods, critical pods and pods with local storages.
|
|
||||||
|
|
||||||
type RemoveDuplicates struct {
|
|
||||||
handle framework.Handle
|
|
||||||
args *componentconfig.RemoveDuplicatesArgs
|
|
||||||
podFilter podutil.FilterFunc
|
|
||||||
}
|
}
|
||||||
|
|
||||||
var _ framework.BalancePlugin = &RemoveDuplicates{}
|
|
||||||
|
|
||||||
type podOwner struct {
|
type podOwner struct {
|
||||||
namespace, kind, name string
|
namespace, kind, name string
|
||||||
imagesHash string
|
imagesHash string
|
||||||
}
|
}
|
||||||
|
|
||||||
// New builds plugin from its arguments while passing a handle
|
// RemoveDuplicatePods removes the duplicate pods on node. This strategy evicts all duplicate pods on node.
|
||||||
func New(args runtime.Object, handle framework.Handle) (framework.Plugin, error) {
|
// A pod is said to be a duplicate of other if both of them are from same creator, kind and are within the same
|
||||||
removeDuplicatesArgs, ok := args.(*componentconfig.RemoveDuplicatesArgs)
|
// namespace, and have at least one container with the same image.
|
||||||
if !ok {
|
// As of now, this strategy won't evict daemonsets, mirror pods, critical pods and pods with local storages.
|
||||||
return nil, fmt.Errorf("want args to be of type RemoveDuplicatesArgs, got %T", args)
|
func RemoveDuplicatePods(
|
||||||
|
ctx context.Context,
|
||||||
|
client clientset.Interface,
|
||||||
|
strategy api.DeschedulerStrategy,
|
||||||
|
nodes []*v1.Node,
|
||||||
|
podEvictor *evictions.PodEvictor,
|
||||||
|
) {
|
||||||
|
if err := validateRemoveDuplicatePodsParams(strategy.Params); err != nil {
|
||||||
|
klog.ErrorS(err, "Invalid RemoveDuplicatePods parameters")
|
||||||
|
return
|
||||||
}
|
}
|
||||||
|
thresholdPriority, err := utils.GetPriorityFromStrategyParams(ctx, client, strategy.Params)
|
||||||
var includedNamespaces, excludedNamespaces sets.String
|
|
||||||
if removeDuplicatesArgs.Namespaces != nil {
|
|
||||||
includedNamespaces = sets.NewString(removeDuplicatesArgs.Namespaces.Include...)
|
|
||||||
excludedNamespaces = sets.NewString(removeDuplicatesArgs.Namespaces.Exclude...)
|
|
||||||
}
|
|
||||||
|
|
||||||
podFilter, err := podutil.NewOptions().
|
|
||||||
WithFilter(handle.Evictor().Filter).
|
|
||||||
WithNamespaces(includedNamespaces).
|
|
||||||
WithoutNamespaces(excludedNamespaces).
|
|
||||||
BuildFilterFunc()
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, fmt.Errorf("error initializing pod filter function: %v", err)
|
klog.ErrorS(err, "Failed to get threshold priority from strategy's params")
|
||||||
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
return &RemoveDuplicates{
|
var includedNamespaces, excludedNamespaces []string
|
||||||
handle: handle,
|
if strategy.Params != nil && strategy.Params.Namespaces != nil {
|
||||||
args: removeDuplicatesArgs,
|
includedNamespaces = strategy.Params.Namespaces.Include
|
||||||
podFilter: podFilter,
|
excludedNamespaces = strategy.Params.Namespaces.Exclude
|
||||||
}, nil
|
}
|
||||||
}
|
|
||||||
|
|
||||||
// Name retrieves the plugin name
|
nodeFit := false
|
||||||
func (r *RemoveDuplicates) Name() string {
|
if strategy.Params != nil {
|
||||||
return PluginName
|
nodeFit = strategy.Params.NodeFit
|
||||||
}
|
}
|
||||||
|
|
||||||
|
evictable := podEvictor.Evictable(evictions.WithPriorityThreshold(thresholdPriority), evictions.WithNodeFit(nodeFit))
|
||||||
|
|
||||||
// Balance extension point implementation for the plugin
|
|
||||||
func (r *RemoveDuplicates) Balance(ctx context.Context, nodes []*v1.Node) *framework.Status {
|
|
||||||
duplicatePods := make(map[podOwner]map[string][]*v1.Pod)
|
duplicatePods := make(map[podOwner]map[string][]*v1.Pod)
|
||||||
ownerKeyOccurence := make(map[podOwner]int32)
|
ownerKeyOccurence := make(map[podOwner]int32)
|
||||||
nodeCount := 0
|
nodeCount := 0
|
||||||
@@ -101,7 +97,13 @@ func (r *RemoveDuplicates) Balance(ctx context.Context, nodes []*v1.Node) *frame
|
|||||||
|
|
||||||
for _, node := range nodes {
|
for _, node := range nodes {
|
||||||
klog.V(1).InfoS("Processing node", "node", klog.KObj(node))
|
klog.V(1).InfoS("Processing node", "node", klog.KObj(node))
|
||||||
pods, err := podutil.ListPodsOnANode(node.Name, r.handle.GetPodsAssignedToNodeFunc(), r.podFilter)
|
pods, err := podutil.ListPodsOnANode(ctx,
|
||||||
|
client,
|
||||||
|
node,
|
||||||
|
podutil.WithFilter(evictable.IsEvictable),
|
||||||
|
podutil.WithNamespaces(includedNamespaces),
|
||||||
|
podutil.WithoutNamespaces(excludedNamespaces),
|
||||||
|
)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
klog.ErrorS(err, "Error listing evictable pods on node", "node", klog.KObj(node))
|
klog.ErrorS(err, "Error listing evictable pods on node", "node", klog.KObj(node))
|
||||||
continue
|
continue
|
||||||
@@ -124,8 +126,7 @@ func (r *RemoveDuplicates) Balance(ctx context.Context, nodes []*v1.Node) *frame
|
|||||||
duplicateKeysMap := map[string][][]string{}
|
duplicateKeysMap := map[string][][]string{}
|
||||||
for _, pod := range pods {
|
for _, pod := range pods {
|
||||||
ownerRefList := podutil.OwnerRef(pod)
|
ownerRefList := podutil.OwnerRef(pod)
|
||||||
|
if hasExcludedOwnerRefKind(ownerRefList, strategy) || len(ownerRefList) == 0 {
|
||||||
if len(ownerRefList) == 0 || hasExcludedOwnerRefKind(ownerRefList, r.args.ExcludeOwnerKinds) {
|
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
podContainerKeys := make([]string, 0, len(ownerRefList)*len(pod.Spec.Containers))
|
podContainerKeys := make([]string, 0, len(ownerRefList)*len(pod.Spec.Containers))
|
||||||
@@ -188,7 +189,6 @@ func (r *RemoveDuplicates) Balance(ctx context.Context, nodes []*v1.Node) *frame
|
|||||||
|
|
||||||
// 1. how many pods can be evicted to respect uniform placement of pods among viable nodes?
|
// 1. how many pods can be evicted to respect uniform placement of pods among viable nodes?
|
||||||
for ownerKey, podNodes := range duplicatePods {
|
for ownerKey, podNodes := range duplicatePods {
|
||||||
|
|
||||||
targetNodes := getTargetNodes(podNodes, nodes)
|
targetNodes := getTargetNodes(podNodes, nodes)
|
||||||
|
|
||||||
klog.V(2).InfoS("Adjusting feasible nodes", "owner", ownerKey, "from", nodeCount, "to", len(targetNodes))
|
klog.V(2).InfoS("Adjusting feasible nodes", "owner", ownerKey, "from", nodeCount, "to", len(targetNodes))
|
||||||
@@ -198,7 +198,6 @@ func (r *RemoveDuplicates) Balance(ctx context.Context, nodes []*v1.Node) *frame
|
|||||||
}
|
}
|
||||||
|
|
||||||
upperAvg := int(math.Ceil(float64(ownerKeyOccurence[ownerKey]) / float64(len(targetNodes))))
|
upperAvg := int(math.Ceil(float64(ownerKeyOccurence[ownerKey]) / float64(len(targetNodes))))
|
||||||
loop:
|
|
||||||
for nodeName, pods := range podNodes {
|
for nodeName, pods := range podNodes {
|
||||||
klog.V(2).InfoS("Average occurrence per node", "node", klog.KObj(nodeMap[nodeName]), "ownerKey", ownerKey, "avg", upperAvg)
|
klog.V(2).InfoS("Average occurrence per node", "node", klog.KObj(nodeMap[nodeName]), "ownerKey", ownerKey, "avg", upperAvg)
|
||||||
// list of duplicated pods does not contain the original referential pod
|
// list of duplicated pods does not contain the original referential pod
|
||||||
@@ -206,15 +205,24 @@ func (r *RemoveDuplicates) Balance(ctx context.Context, nodes []*v1.Node) *frame
|
|||||||
// It's assumed all duplicated pods are in the same priority class
|
// It's assumed all duplicated pods are in the same priority class
|
||||||
// TODO(jchaloup): check if the pod has a different node to lend to
|
// TODO(jchaloup): check if the pod has a different node to lend to
|
||||||
for _, pod := range pods[upperAvg-1:] {
|
for _, pod := range pods[upperAvg-1:] {
|
||||||
r.handle.Evictor().Evict(ctx, pod, evictions.EvictOptions{})
|
if _, err := podEvictor.EvictPod(ctx, pod, nodeMap[nodeName], "RemoveDuplicatePods"); err != nil {
|
||||||
if r.handle.Evictor().NodeLimitExceeded(nodeMap[nodeName]) {
|
klog.ErrorS(err, "Error evicting pod", "pod", klog.KObj(pod))
|
||||||
continue loop
|
break
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
return nil
|
}
|
||||||
|
|
||||||
|
func getNodeAffinityNodeSelector(pod *v1.Pod) *v1.NodeSelector {
|
||||||
|
if pod.Spec.Affinity == nil {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
if pod.Spec.Affinity.NodeAffinity == nil {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
return pod.Spec.Affinity.NodeAffinity.RequiredDuringSchedulingIgnoredDuringExecution
|
||||||
}
|
}
|
||||||
|
|
||||||
func getTargetNodes(podNodes map[string][]*v1.Pod, nodes []*v1.Node) []*v1.Node {
|
func getTargetNodes(podNodes map[string][]*v1.Pod, nodes []*v1.Node) []*v1.Node {
|
||||||
@@ -274,27 +282,15 @@ func getTargetNodes(podNodes map[string][]*v1.Pod, nodes []*v1.Node) []*v1.Node
|
|||||||
return targetNodes
|
return targetNodes
|
||||||
}
|
}
|
||||||
|
|
||||||
func hasExcludedOwnerRefKind(ownerRefs []metav1.OwnerReference, excludeOwnerKinds []string) bool {
|
func hasExcludedOwnerRefKind(ownerRefs []metav1.OwnerReference, strategy api.DeschedulerStrategy) bool {
|
||||||
if len(excludeOwnerKinds) == 0 {
|
if strategy.Params == nil || strategy.Params.RemoveDuplicates == nil {
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
|
exclude := sets.NewString(strategy.Params.RemoveDuplicates.ExcludeOwnerKinds...)
|
||||||
exclude := sets.NewString(excludeOwnerKinds...)
|
|
||||||
for _, owner := range ownerRefs {
|
for _, owner := range ownerRefs {
|
||||||
if exclude.Has(owner.Kind) {
|
if exclude.Has(owner.Kind) {
|
||||||
return true
|
return true
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
|
|
||||||
func getNodeAffinityNodeSelector(pod *v1.Pod) *v1.NodeSelector {
|
|
||||||
if pod.Spec.Affinity == nil {
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
if pod.Spec.Affinity.NodeAffinity == nil {
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
return pod.Spec.Affinity.NodeAffinity.RequiredDuringSchedulingIgnoredDuringExecution
|
|
||||||
}
|
|
||||||
@@ -1,5 +1,5 @@
|
|||||||
/*
|
/*
|
||||||
Copyright 2022 The Kubernetes Authors.
|
Copyright 2017 The Kubernetes Authors.
|
||||||
|
|
||||||
Licensed under the Apache License, Version 2.0 (the "License");
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
you may not use this file except in compliance with the License.
|
you may not use this file except in compliance with the License.
|
||||||
@@ -14,28 +14,21 @@ See the License for the specific language governing permissions and
|
|||||||
limitations under the License.
|
limitations under the License.
|
||||||
*/
|
*/
|
||||||
|
|
||||||
package removeduplicates
|
package strategies
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
"testing"
|
"testing"
|
||||||
|
|
||||||
"k8s.io/client-go/tools/events"
|
|
||||||
"sigs.k8s.io/descheduler/pkg/apis/componentconfig"
|
|
||||||
"sigs.k8s.io/descheduler/pkg/framework"
|
|
||||||
frameworkfake "sigs.k8s.io/descheduler/pkg/framework/fake"
|
|
||||||
"sigs.k8s.io/descheduler/pkg/framework/plugins/defaultevictor"
|
|
||||||
|
|
||||||
v1 "k8s.io/api/core/v1"
|
v1 "k8s.io/api/core/v1"
|
||||||
policyv1 "k8s.io/api/policy/v1"
|
policyv1 "k8s.io/api/policy/v1"
|
||||||
"k8s.io/apimachinery/pkg/api/resource"
|
"k8s.io/apimachinery/pkg/api/resource"
|
||||||
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
||||||
"k8s.io/apimachinery/pkg/runtime"
|
"k8s.io/apimachinery/pkg/runtime"
|
||||||
"k8s.io/client-go/informers"
|
|
||||||
"k8s.io/client-go/kubernetes/fake"
|
"k8s.io/client-go/kubernetes/fake"
|
||||||
|
core "k8s.io/client-go/testing"
|
||||||
|
"sigs.k8s.io/descheduler/pkg/api"
|
||||||
"sigs.k8s.io/descheduler/pkg/descheduler/evictions"
|
"sigs.k8s.io/descheduler/pkg/descheduler/evictions"
|
||||||
podutil "sigs.k8s.io/descheduler/pkg/descheduler/pod"
|
|
||||||
"sigs.k8s.io/descheduler/pkg/utils"
|
"sigs.k8s.io/descheduler/pkg/utils"
|
||||||
"sigs.k8s.io/descheduler/test"
|
"sigs.k8s.io/descheduler/test"
|
||||||
)
|
)
|
||||||
@@ -50,6 +43,7 @@ func buildTestPodWithImage(podName, node, image string) *v1.Pod {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func TestFindDuplicatePods(t *testing.T) {
|
func TestFindDuplicatePods(t *testing.T) {
|
||||||
|
ctx := context.Background()
|
||||||
// first setup pods
|
// first setup pods
|
||||||
node1 := test.BuildTestNode("n1", 2000, 3000, 10, nil)
|
node1 := test.BuildTestNode("n1", 2000, 3000, 10, nil)
|
||||||
node2 := test.BuildTestNode("n2", 2000, 3000, 10, nil)
|
node2 := test.BuildTestNode("n2", 2000, 3000, 10, nil)
|
||||||
@@ -72,7 +66,6 @@ func TestFindDuplicatePods(t *testing.T) {
|
|||||||
Unschedulable: true,
|
Unschedulable: true,
|
||||||
}
|
}
|
||||||
})
|
})
|
||||||
node6 := test.BuildTestNode("n6", 200, 200, 10, nil)
|
|
||||||
|
|
||||||
p1 := test.BuildTestPod("p1", 100, 0, node1.Name, nil)
|
p1 := test.BuildTestPod("p1", 100, 0, node1.Name, nil)
|
||||||
p1.Namespace = "dev"
|
p1.Namespace = "dev"
|
||||||
@@ -108,14 +101,6 @@ func TestFindDuplicatePods(t *testing.T) {
|
|||||||
p18 := test.BuildTestPod("TARGET", 100, 0, node1.Name, nil)
|
p18 := test.BuildTestPod("TARGET", 100, 0, node1.Name, nil)
|
||||||
p18.Namespace = "node-fit"
|
p18.Namespace = "node-fit"
|
||||||
|
|
||||||
// This pod sits on node6 and is used to take up CPU requests on the node
|
|
||||||
p19 := test.BuildTestPod("CPU-eater", 150, 150, node6.Name, nil)
|
|
||||||
p19.Namespace = "test"
|
|
||||||
|
|
||||||
// Dummy pod for node6 used to do the opposite of p19
|
|
||||||
p20 := test.BuildTestPod("CPU-saver", 100, 150, node6.Name, nil)
|
|
||||||
p20.Namespace = "test"
|
|
||||||
|
|
||||||
// ### Evictable Pods ###
|
// ### Evictable Pods ###
|
||||||
|
|
||||||
// Three Pods in the "default" Namespace, bound to same ReplicaSet. 2 should be evicted.
|
// Three Pods in the "default" Namespace, bound to same ReplicaSet. 2 should be evicted.
|
||||||
@@ -188,182 +173,131 @@ func TestFindDuplicatePods(t *testing.T) {
|
|||||||
|
|
||||||
testCases := []struct {
|
testCases := []struct {
|
||||||
description string
|
description string
|
||||||
pods []*v1.Pod
|
maxPodsToEvictPerNode int
|
||||||
|
pods []v1.Pod
|
||||||
nodes []*v1.Node
|
nodes []*v1.Node
|
||||||
expectedEvictedPodCount uint
|
expectedEvictedPodCount int
|
||||||
excludeOwnerKinds []string
|
strategy api.DeschedulerStrategy
|
||||||
nodefit bool
|
|
||||||
}{
|
}{
|
||||||
{
|
{
|
||||||
description: "Three pods in the `dev` Namespace, bound to same ReplicaSet. 1 should be evicted.",
|
description: "Three pods in the `dev` Namespace, bound to same ReplicaSet. 1 should be evicted.",
|
||||||
pods: []*v1.Pod{p1, p2, p3},
|
maxPodsToEvictPerNode: 5,
|
||||||
|
pods: []v1.Pod{*p1, *p2, *p3},
|
||||||
nodes: []*v1.Node{node1, node2},
|
nodes: []*v1.Node{node1, node2},
|
||||||
expectedEvictedPodCount: 1,
|
expectedEvictedPodCount: 1,
|
||||||
|
strategy: api.DeschedulerStrategy{},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
description: "Three pods in the `dev` Namespace, bound to same ReplicaSet, but ReplicaSet kind is excluded. 0 should be evicted.",
|
description: "Three pods in the `dev` Namespace, bound to same ReplicaSet, but ReplicaSet kind is excluded. 0 should be evicted.",
|
||||||
pods: []*v1.Pod{p1, p2, p3},
|
maxPodsToEvictPerNode: 5,
|
||||||
|
pods: []v1.Pod{*p1, *p2, *p3},
|
||||||
nodes: []*v1.Node{node1, node2},
|
nodes: []*v1.Node{node1, node2},
|
||||||
expectedEvictedPodCount: 0,
|
expectedEvictedPodCount: 0,
|
||||||
excludeOwnerKinds: []string{"ReplicaSet"},
|
strategy: api.DeschedulerStrategy{Params: &api.StrategyParameters{RemoveDuplicates: &api.RemoveDuplicates{ExcludeOwnerKinds: []string{"ReplicaSet"}}}},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
description: "Three Pods in the `test` Namespace, bound to same ReplicaSet. 1 should be evicted.",
|
description: "Three Pods in the `test` Namespace, bound to same ReplicaSet. 1 should be evicted.",
|
||||||
pods: []*v1.Pod{p8, p9, p10},
|
maxPodsToEvictPerNode: 5,
|
||||||
|
pods: []v1.Pod{*p8, *p9, *p10},
|
||||||
nodes: []*v1.Node{node1, node2},
|
nodes: []*v1.Node{node1, node2},
|
||||||
expectedEvictedPodCount: 1,
|
expectedEvictedPodCount: 1,
|
||||||
|
strategy: api.DeschedulerStrategy{},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
description: "Three Pods in the `dev` Namespace, three Pods in the `test` Namespace. Bound to ReplicaSet with same name. 4 should be evicted.",
|
description: "Three Pods in the `dev` Namespace, three Pods in the `test` Namespace. Bound to ReplicaSet with same name. 4 should be evicted.",
|
||||||
pods: []*v1.Pod{p1, p2, p3, p8, p9, p10},
|
maxPodsToEvictPerNode: 5,
|
||||||
|
pods: []v1.Pod{*p1, *p2, *p3, *p8, *p9, *p10},
|
||||||
nodes: []*v1.Node{node1, node2},
|
nodes: []*v1.Node{node1, node2},
|
||||||
expectedEvictedPodCount: 2,
|
expectedEvictedPodCount: 2,
|
||||||
|
strategy: api.DeschedulerStrategy{},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
description: "Pods are: part of DaemonSet, with local storage, mirror pod annotation, critical pod annotation - none should be evicted.",
|
description: "Pods are: part of DaemonSet, with local storage, mirror pod annotation, critical pod annotation - none should be evicted.",
|
||||||
pods: []*v1.Pod{p4, p5, p6, p7},
|
maxPodsToEvictPerNode: 2,
|
||||||
|
pods: []v1.Pod{*p4, *p5, *p6, *p7},
|
||||||
nodes: []*v1.Node{node1, node2},
|
nodes: []*v1.Node{node1, node2},
|
||||||
expectedEvictedPodCount: 0,
|
expectedEvictedPodCount: 0,
|
||||||
|
strategy: api.DeschedulerStrategy{},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
description: "Test all Pods: 4 should be evicted.",
|
description: "Test all Pods: 4 should be evicted.",
|
||||||
pods: []*v1.Pod{p1, p2, p3, p4, p5, p6, p7, p8, p9, p10},
|
maxPodsToEvictPerNode: 5,
|
||||||
|
pods: []v1.Pod{*p1, *p2, *p3, *p4, *p5, *p6, *p7, *p8, *p9, *p10},
|
||||||
nodes: []*v1.Node{node1, node2},
|
nodes: []*v1.Node{node1, node2},
|
||||||
expectedEvictedPodCount: 2,
|
expectedEvictedPodCount: 2,
|
||||||
|
strategy: api.DeschedulerStrategy{},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
description: "Pods with the same owner but different images should not be evicted",
|
description: "Pods with the same owner but different images should not be evicted",
|
||||||
pods: []*v1.Pod{p11, p12},
|
maxPodsToEvictPerNode: 5,
|
||||||
|
pods: []v1.Pod{*p11, *p12},
|
||||||
nodes: []*v1.Node{node1, node2},
|
nodes: []*v1.Node{node1, node2},
|
||||||
expectedEvictedPodCount: 0,
|
expectedEvictedPodCount: 0,
|
||||||
|
strategy: api.DeschedulerStrategy{},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
description: "Pods with multiple containers should not match themselves",
|
description: "Pods with multiple containers should not match themselves",
|
||||||
pods: []*v1.Pod{p13},
|
maxPodsToEvictPerNode: 5,
|
||||||
|
pods: []v1.Pod{*p13},
|
||||||
nodes: []*v1.Node{node1, node2},
|
nodes: []*v1.Node{node1, node2},
|
||||||
expectedEvictedPodCount: 0,
|
expectedEvictedPodCount: 0,
|
||||||
|
strategy: api.DeschedulerStrategy{},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
description: "Pods with matching ownerrefs and at not all matching image should not trigger an eviction",
|
description: "Pods with matching ownerrefs and at not all matching image should not trigger an eviction",
|
||||||
pods: []*v1.Pod{p11, p13},
|
maxPodsToEvictPerNode: 5,
|
||||||
|
pods: []v1.Pod{*p11, *p13},
|
||||||
nodes: []*v1.Node{node1, node2},
|
nodes: []*v1.Node{node1, node2},
|
||||||
expectedEvictedPodCount: 0,
|
expectedEvictedPodCount: 0,
|
||||||
|
strategy: api.DeschedulerStrategy{},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
description: "Three pods in the `dev` Namespace, bound to same ReplicaSet. Only node available has a taint, and nodeFit set to true. 0 should be evicted.",
|
description: "Three pods in the `dev` Namespace, bound to same ReplicaSet. Only node available has a taint, and nodeFit set to true. 0 should be evicted.",
|
||||||
pods: []*v1.Pod{p1, p2, p3},
|
maxPodsToEvictPerNode: 5,
|
||||||
|
pods: []v1.Pod{*p1, *p2, *p3},
|
||||||
nodes: []*v1.Node{node1, node3},
|
nodes: []*v1.Node{node1, node3},
|
||||||
expectedEvictedPodCount: 0,
|
expectedEvictedPodCount: 0,
|
||||||
nodefit: true,
|
strategy: api.DeschedulerStrategy{Params: &api.StrategyParameters{NodeFit: true}},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
description: "Three pods in the `node-fit` Namespace, bound to same ReplicaSet, all with a nodeSelector. Only node available has an incorrect node label, and nodeFit set to true. 0 should be evicted.",
|
description: "Three pods in the `node-fit` Namespace, bound to same ReplicaSet, all with a nodeSelector. Only node available has an incorrect node label, and nodeFit set to true. 0 should be evicted.",
|
||||||
pods: []*v1.Pod{p15, p16, p17},
|
maxPodsToEvictPerNode: 5,
|
||||||
|
pods: []v1.Pod{*p15, *p16, *p17},
|
||||||
nodes: []*v1.Node{node1, node4},
|
nodes: []*v1.Node{node1, node4},
|
||||||
expectedEvictedPodCount: 0,
|
expectedEvictedPodCount: 0,
|
||||||
nodefit: true,
|
strategy: api.DeschedulerStrategy{Params: &api.StrategyParameters{NodeFit: true}},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
description: "Three pods in the `node-fit` Namespace, bound to same ReplicaSet. Only node available is not schedulable, and nodeFit set to true. 0 should be evicted.",
|
description: "Three pods in the `node-fit` Namespace, bound to same ReplicaSet. Only node available is not schedulable, and nodeFit set to true. 0 should be evicted.",
|
||||||
pods: []*v1.Pod{p1, p2, p3},
|
maxPodsToEvictPerNode: 5,
|
||||||
|
pods: []v1.Pod{*p1, *p2, *p3},
|
||||||
nodes: []*v1.Node{node1, node5},
|
nodes: []*v1.Node{node1, node5},
|
||||||
expectedEvictedPodCount: 0,
|
expectedEvictedPodCount: 0,
|
||||||
nodefit: true,
|
strategy: api.DeschedulerStrategy{Params: &api.StrategyParameters{NodeFit: true}},
|
||||||
},
|
|
||||||
{
|
|
||||||
description: "Three pods in the `node-fit` Namespace, bound to same ReplicaSet. Only node available does not have enough CPU, and nodeFit set to true. 0 should be evicted.",
|
|
||||||
pods: []*v1.Pod{p1, p2, p3, p19},
|
|
||||||
nodes: []*v1.Node{node1, node6},
|
|
||||||
expectedEvictedPodCount: 0,
|
|
||||||
nodefit: true,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
description: "Three pods in the `node-fit` Namespace, bound to same ReplicaSet. Only node available has enough CPU, and nodeFit set to true. 1 should be evicted.",
|
|
||||||
pods: []*v1.Pod{p1, p2, p3, p20},
|
|
||||||
nodes: []*v1.Node{node1, node6},
|
|
||||||
expectedEvictedPodCount: 1,
|
|
||||||
nodefit: true,
|
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
for _, testCase := range testCases {
|
for _, testCase := range testCases {
|
||||||
t.Run(testCase.description, func(t *testing.T) {
|
t.Run(testCase.description, func(t *testing.T) {
|
||||||
ctx, cancel := context.WithCancel(context.Background())
|
fakeClient := &fake.Clientset{}
|
||||||
defer cancel()
|
fakeClient.Fake.AddReactor("list", "pods", func(action core.Action) (bool, runtime.Object, error) {
|
||||||
|
return true, &v1.PodList{Items: testCase.pods}, nil
|
||||||
var objs []runtime.Object
|
})
|
||||||
for _, node := range testCase.nodes {
|
|
||||||
objs = append(objs, node)
|
|
||||||
}
|
|
||||||
for _, pod := range testCase.pods {
|
|
||||||
objs = append(objs, pod)
|
|
||||||
}
|
|
||||||
fakeClient := fake.NewSimpleClientset(objs...)
|
|
||||||
|
|
||||||
sharedInformerFactory := informers.NewSharedInformerFactory(fakeClient, 0)
|
|
||||||
podInformer := sharedInformerFactory.Core().V1().Pods()
|
|
||||||
|
|
||||||
getPodsAssignedToNode, err := podutil.BuildGetPodsAssignedToNodeFunc(podInformer)
|
|
||||||
if err != nil {
|
|
||||||
t.Errorf("Build get pods assigned to node function error: %v", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
sharedInformerFactory.Start(ctx.Done())
|
|
||||||
sharedInformerFactory.WaitForCacheSync(ctx.Done())
|
|
||||||
|
|
||||||
eventRecorder := &events.FakeRecorder{}
|
|
||||||
|
|
||||||
podEvictor := evictions.NewPodEvictor(
|
podEvictor := evictions.NewPodEvictor(
|
||||||
fakeClient,
|
fakeClient,
|
||||||
"v1",
|
"v1",
|
||||||
false,
|
false,
|
||||||
nil,
|
testCase.maxPodsToEvictPerNode,
|
||||||
nil,
|
|
||||||
testCase.nodes,
|
testCase.nodes,
|
||||||
false,
|
false,
|
||||||
eventRecorder,
|
false,
|
||||||
|
false,
|
||||||
)
|
)
|
||||||
|
|
||||||
nodeFit := testCase.nodefit
|
RemoveDuplicatePods(ctx, fakeClient, testCase.strategy, testCase.nodes, podEvictor)
|
||||||
|
podsEvicted := podEvictor.TotalEvicted()
|
||||||
defaultEvictorFilterArgs := &defaultevictor.DefaultEvictorArgs{
|
if podsEvicted != testCase.expectedEvictedPodCount {
|
||||||
EvictLocalStoragePods: false,
|
t.Errorf("Test error for description: %s. Expected evicted pods count %v, got %v", testCase.description, testCase.expectedEvictedPodCount, podsEvicted)
|
||||||
EvictSystemCriticalPods: false,
|
|
||||||
IgnorePvcPods: false,
|
|
||||||
EvictFailedBarePods: false,
|
|
||||||
NodeFit: nodeFit,
|
|
||||||
}
|
|
||||||
|
|
||||||
evictorFilter, _ := defaultevictor.New(
|
|
||||||
defaultEvictorFilterArgs,
|
|
||||||
&frameworkfake.HandleImpl{
|
|
||||||
ClientsetImpl: fakeClient,
|
|
||||||
GetPodsAssignedToNodeFuncImpl: getPodsAssignedToNode,
|
|
||||||
SharedInformerFactoryImpl: sharedInformerFactory,
|
|
||||||
},
|
|
||||||
)
|
|
||||||
|
|
||||||
handle := &frameworkfake.HandleImpl{
|
|
||||||
ClientsetImpl: fakeClient,
|
|
||||||
GetPodsAssignedToNodeFuncImpl: getPodsAssignedToNode,
|
|
||||||
PodEvictorImpl: podEvictor,
|
|
||||||
EvictorFilterImpl: evictorFilter.(framework.EvictorPlugin),
|
|
||||||
SharedInformerFactoryImpl: sharedInformerFactory,
|
|
||||||
}
|
|
||||||
|
|
||||||
plugin, err := New(&componentconfig.RemoveDuplicatesArgs{
|
|
||||||
ExcludeOwnerKinds: testCase.excludeOwnerKinds,
|
|
||||||
},
|
|
||||||
handle,
|
|
||||||
)
|
|
||||||
if err != nil {
|
|
||||||
t.Fatalf("Unable to initialize the plugin: %v", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
plugin.(framework.BalancePlugin).Balance(ctx, testCase.nodes)
|
|
||||||
actualEvictedPodCount := podEvictor.TotalEvicted()
|
|
||||||
if actualEvictedPodCount != testCase.expectedEvictedPodCount {
|
|
||||||
t.Errorf("Test %#v failed, Unexpected no of pods evicted: pods evicted: %d, expected: %d", testCase.description, actualEvictedPodCount, testCase.expectedEvictedPodCount)
|
|
||||||
}
|
}
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
@@ -371,6 +305,8 @@ func TestFindDuplicatePods(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func TestRemoveDuplicatesUniformly(t *testing.T) {
|
func TestRemoveDuplicatesUniformly(t *testing.T) {
|
||||||
|
ctx := context.Background()
|
||||||
|
|
||||||
setRSOwnerRef2 := func(pod *v1.Pod) {
|
setRSOwnerRef2 := func(pod *v1.Pod) {
|
||||||
pod.ObjectMeta.OwnerReferences = []metav1.OwnerReference{
|
pod.ObjectMeta.OwnerReferences = []metav1.OwnerReference{
|
||||||
{Kind: "ReplicaSet", APIVersion: "v1", Name: "replicaset-2"},
|
{Kind: "ReplicaSet", APIVersion: "v1", Name: "replicaset-2"},
|
||||||
@@ -410,7 +346,7 @@ func TestRemoveDuplicatesUniformly(t *testing.T) {
|
|||||||
node.Spec.Taints = []v1.Taint{
|
node.Spec.Taints = []v1.Taint{
|
||||||
{
|
{
|
||||||
Effect: v1.TaintEffectNoSchedule,
|
Effect: v1.TaintEffectNoSchedule,
|
||||||
Key: "node-role.kubernetes.io/control-plane",
|
Key: "node-role.kubernetes.io/master",
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -419,7 +355,7 @@ func TestRemoveDuplicatesUniformly(t *testing.T) {
|
|||||||
if node.ObjectMeta.Labels == nil {
|
if node.ObjectMeta.Labels == nil {
|
||||||
node.ObjectMeta.Labels = map[string]string{}
|
node.ObjectMeta.Labels = map[string]string{}
|
||||||
}
|
}
|
||||||
node.ObjectMeta.Labels["node-role.kubernetes.io/control-plane"] = ""
|
node.ObjectMeta.Labels["node-role.kubernetes.io/master"] = ""
|
||||||
}
|
}
|
||||||
|
|
||||||
setWorkerLabel := func(node *v1.Node) {
|
setWorkerLabel := func(node *v1.Node) {
|
||||||
@@ -439,7 +375,7 @@ func TestRemoveDuplicatesUniformly(t *testing.T) {
|
|||||||
{
|
{
|
||||||
MatchExpressions: []v1.NodeSelectorRequirement{
|
MatchExpressions: []v1.NodeSelectorRequirement{
|
||||||
{
|
{
|
||||||
Key: "node-role.kubernetes.io/control-plane",
|
Key: "node-role.kubernetes.io/master",
|
||||||
Operator: v1.NodeSelectorOpDoesNotExist,
|
Operator: v1.NodeSelectorOpDoesNotExist,
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
@@ -463,7 +399,7 @@ func TestRemoveDuplicatesUniformly(t *testing.T) {
|
|||||||
{
|
{
|
||||||
MatchExpressions: []v1.NodeSelectorRequirement{
|
MatchExpressions: []v1.NodeSelectorRequirement{
|
||||||
{
|
{
|
||||||
Key: "node-role.kubernetes.io/control-plane",
|
Key: "node-role.kubernetes.io/master",
|
||||||
Operator: v1.NodeSelectorOpDoesNotExist,
|
Operator: v1.NodeSelectorOpDoesNotExist,
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
@@ -496,23 +432,25 @@ func TestRemoveDuplicatesUniformly(t *testing.T) {
|
|||||||
|
|
||||||
testCases := []struct {
|
testCases := []struct {
|
||||||
description string
|
description string
|
||||||
pods []*v1.Pod
|
maxPodsToEvictPerNode int
|
||||||
|
pods []v1.Pod
|
||||||
nodes []*v1.Node
|
nodes []*v1.Node
|
||||||
expectedEvictedPodCount uint
|
expectedEvictedPodCount int
|
||||||
|
strategy api.DeschedulerStrategy
|
||||||
}{
|
}{
|
||||||
{
|
{
|
||||||
description: "Evict pods uniformly",
|
description: "Evict pods uniformly",
|
||||||
pods: []*v1.Pod{
|
pods: []v1.Pod{
|
||||||
// (5,3,1) -> (3,3,3) -> 2 evictions
|
// (5,3,1) -> (3,3,3) -> 2 evictions
|
||||||
test.BuildTestPod("p1", 100, 0, "n1", test.SetRSOwnerRef),
|
*test.BuildTestPod("p1", 100, 0, "n1", test.SetRSOwnerRef),
|
||||||
test.BuildTestPod("p2", 100, 0, "n1", test.SetRSOwnerRef),
|
*test.BuildTestPod("p2", 100, 0, "n1", test.SetRSOwnerRef),
|
||||||
test.BuildTestPod("p3", 100, 0, "n1", test.SetRSOwnerRef),
|
*test.BuildTestPod("p3", 100, 0, "n1", test.SetRSOwnerRef),
|
||||||
test.BuildTestPod("p4", 100, 0, "n1", test.SetRSOwnerRef),
|
*test.BuildTestPod("p4", 100, 0, "n1", test.SetRSOwnerRef),
|
||||||
test.BuildTestPod("p5", 100, 0, "n1", test.SetRSOwnerRef),
|
*test.BuildTestPod("p5", 100, 0, "n1", test.SetRSOwnerRef),
|
||||||
test.BuildTestPod("p6", 100, 0, "n2", test.SetRSOwnerRef),
|
*test.BuildTestPod("p6", 100, 0, "n2", test.SetRSOwnerRef),
|
||||||
test.BuildTestPod("p7", 100, 0, "n2", test.SetRSOwnerRef),
|
*test.BuildTestPod("p7", 100, 0, "n2", test.SetRSOwnerRef),
|
||||||
test.BuildTestPod("p8", 100, 0, "n2", test.SetRSOwnerRef),
|
*test.BuildTestPod("p8", 100, 0, "n2", test.SetRSOwnerRef),
|
||||||
test.BuildTestPod("p9", 100, 0, "n3", test.SetRSOwnerRef),
|
*test.BuildTestPod("p9", 100, 0, "n3", test.SetRSOwnerRef),
|
||||||
},
|
},
|
||||||
expectedEvictedPodCount: 2,
|
expectedEvictedPodCount: 2,
|
||||||
nodes: []*v1.Node{
|
nodes: []*v1.Node{
|
||||||
@@ -520,40 +458,42 @@ func TestRemoveDuplicatesUniformly(t *testing.T) {
|
|||||||
test.BuildTestNode("n2", 2000, 3000, 10, nil),
|
test.BuildTestNode("n2", 2000, 3000, 10, nil),
|
||||||
test.BuildTestNode("n3", 2000, 3000, 10, nil),
|
test.BuildTestNode("n3", 2000, 3000, 10, nil),
|
||||||
},
|
},
|
||||||
|
strategy: api.DeschedulerStrategy{},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
description: "Evict pods uniformly with one node left out",
|
description: "Evict pods uniformly with one node left out",
|
||||||
pods: []*v1.Pod{
|
pods: []v1.Pod{
|
||||||
// (5,3,1) -> (4,4,1) -> 1 eviction
|
// (5,3,1) -> (4,4,1) -> 1 eviction
|
||||||
test.BuildTestPod("p1", 100, 0, "n1", test.SetRSOwnerRef),
|
*test.BuildTestPod("p1", 100, 0, "n1", test.SetRSOwnerRef),
|
||||||
test.BuildTestPod("p2", 100, 0, "n1", test.SetRSOwnerRef),
|
*test.BuildTestPod("p2", 100, 0, "n1", test.SetRSOwnerRef),
|
||||||
test.BuildTestPod("p3", 100, 0, "n1", test.SetRSOwnerRef),
|
*test.BuildTestPod("p3", 100, 0, "n1", test.SetRSOwnerRef),
|
||||||
test.BuildTestPod("p4", 100, 0, "n1", test.SetRSOwnerRef),
|
*test.BuildTestPod("p4", 100, 0, "n1", test.SetRSOwnerRef),
|
||||||
test.BuildTestPod("p5", 100, 0, "n1", test.SetRSOwnerRef),
|
*test.BuildTestPod("p5", 100, 0, "n1", test.SetRSOwnerRef),
|
||||||
test.BuildTestPod("p6", 100, 0, "n2", test.SetRSOwnerRef),
|
*test.BuildTestPod("p6", 100, 0, "n2", test.SetRSOwnerRef),
|
||||||
test.BuildTestPod("p7", 100, 0, "n2", test.SetRSOwnerRef),
|
*test.BuildTestPod("p7", 100, 0, "n2", test.SetRSOwnerRef),
|
||||||
test.BuildTestPod("p8", 100, 0, "n2", test.SetRSOwnerRef),
|
*test.BuildTestPod("p8", 100, 0, "n2", test.SetRSOwnerRef),
|
||||||
test.BuildTestPod("p9", 100, 0, "n3", test.SetRSOwnerRef),
|
*test.BuildTestPod("p9", 100, 0, "n3", test.SetRSOwnerRef),
|
||||||
},
|
},
|
||||||
expectedEvictedPodCount: 1,
|
expectedEvictedPodCount: 1,
|
||||||
nodes: []*v1.Node{
|
nodes: []*v1.Node{
|
||||||
test.BuildTestNode("n1", 2000, 3000, 10, nil),
|
test.BuildTestNode("n1", 2000, 3000, 10, nil),
|
||||||
test.BuildTestNode("n2", 2000, 3000, 10, nil),
|
test.BuildTestNode("n2", 2000, 3000, 10, nil),
|
||||||
},
|
},
|
||||||
|
strategy: api.DeschedulerStrategy{},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
description: "Evict pods uniformly with two replica sets",
|
description: "Evict pods uniformly with two replica sets",
|
||||||
pods: []*v1.Pod{
|
pods: []v1.Pod{
|
||||||
// (5,3,1) -> (3,3,3) -> 2 evictions
|
// (5,3,1) -> (3,3,3) -> 2 evictions
|
||||||
test.BuildTestPod("p11", 100, 0, "n1", setTwoRSOwnerRef),
|
*test.BuildTestPod("p11", 100, 0, "n1", setTwoRSOwnerRef),
|
||||||
test.BuildTestPod("p12", 100, 0, "n1", setTwoRSOwnerRef),
|
*test.BuildTestPod("p12", 100, 0, "n1", setTwoRSOwnerRef),
|
||||||
test.BuildTestPod("p13", 100, 0, "n1", setTwoRSOwnerRef),
|
*test.BuildTestPod("p13", 100, 0, "n1", setTwoRSOwnerRef),
|
||||||
test.BuildTestPod("p14", 100, 0, "n1", setTwoRSOwnerRef),
|
*test.BuildTestPod("p14", 100, 0, "n1", setTwoRSOwnerRef),
|
||||||
test.BuildTestPod("p15", 100, 0, "n1", setTwoRSOwnerRef),
|
*test.BuildTestPod("p15", 100, 0, "n1", setTwoRSOwnerRef),
|
||||||
test.BuildTestPod("p16", 100, 0, "n2", setTwoRSOwnerRef),
|
*test.BuildTestPod("p16", 100, 0, "n2", setTwoRSOwnerRef),
|
||||||
test.BuildTestPod("p17", 100, 0, "n2", setTwoRSOwnerRef),
|
*test.BuildTestPod("p17", 100, 0, "n2", setTwoRSOwnerRef),
|
||||||
test.BuildTestPod("p18", 100, 0, "n2", setTwoRSOwnerRef),
|
*test.BuildTestPod("p18", 100, 0, "n2", setTwoRSOwnerRef),
|
||||||
test.BuildTestPod("p19", 100, 0, "n3", setTwoRSOwnerRef),
|
*test.BuildTestPod("p19", 100, 0, "n3", setTwoRSOwnerRef),
|
||||||
},
|
},
|
||||||
expectedEvictedPodCount: 4,
|
expectedEvictedPodCount: 4,
|
||||||
nodes: []*v1.Node{
|
nodes: []*v1.Node{
|
||||||
@@ -561,30 +501,31 @@ func TestRemoveDuplicatesUniformly(t *testing.T) {
|
|||||||
test.BuildTestNode("n2", 2000, 3000, 10, nil),
|
test.BuildTestNode("n2", 2000, 3000, 10, nil),
|
||||||
test.BuildTestNode("n3", 2000, 3000, 10, nil),
|
test.BuildTestNode("n3", 2000, 3000, 10, nil),
|
||||||
},
|
},
|
||||||
|
strategy: api.DeschedulerStrategy{},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
description: "Evict pods uniformly with two owner references",
|
description: "Evict pods uniformly with two owner references",
|
||||||
pods: []*v1.Pod{
|
pods: []v1.Pod{
|
||||||
// (5,3,1) -> (3,3,3) -> 2 evictions
|
// (5,3,1) -> (3,3,3) -> 2 evictions
|
||||||
test.BuildTestPod("p11", 100, 0, "n1", test.SetRSOwnerRef),
|
*test.BuildTestPod("p11", 100, 0, "n1", test.SetRSOwnerRef),
|
||||||
test.BuildTestPod("p12", 100, 0, "n1", test.SetRSOwnerRef),
|
*test.BuildTestPod("p12", 100, 0, "n1", test.SetRSOwnerRef),
|
||||||
test.BuildTestPod("p13", 100, 0, "n1", test.SetRSOwnerRef),
|
*test.BuildTestPod("p13", 100, 0, "n1", test.SetRSOwnerRef),
|
||||||
test.BuildTestPod("p14", 100, 0, "n1", test.SetRSOwnerRef),
|
*test.BuildTestPod("p14", 100, 0, "n1", test.SetRSOwnerRef),
|
||||||
test.BuildTestPod("p15", 100, 0, "n1", test.SetRSOwnerRef),
|
*test.BuildTestPod("p15", 100, 0, "n1", test.SetRSOwnerRef),
|
||||||
test.BuildTestPod("p16", 100, 0, "n2", test.SetRSOwnerRef),
|
*test.BuildTestPod("p16", 100, 0, "n2", test.SetRSOwnerRef),
|
||||||
test.BuildTestPod("p17", 100, 0, "n2", test.SetRSOwnerRef),
|
*test.BuildTestPod("p17", 100, 0, "n2", test.SetRSOwnerRef),
|
||||||
test.BuildTestPod("p18", 100, 0, "n2", test.SetRSOwnerRef),
|
*test.BuildTestPod("p18", 100, 0, "n2", test.SetRSOwnerRef),
|
||||||
test.BuildTestPod("p19", 100, 0, "n3", test.SetRSOwnerRef),
|
*test.BuildTestPod("p19", 100, 0, "n3", test.SetRSOwnerRef),
|
||||||
// (1,3,5) -> (3,3,3) -> 2 evictions
|
// (1,3,5) -> (3,3,3) -> 2 evictions
|
||||||
test.BuildTestPod("p21", 100, 0, "n1", setRSOwnerRef2),
|
*test.BuildTestPod("p21", 100, 0, "n1", setRSOwnerRef2),
|
||||||
test.BuildTestPod("p22", 100, 0, "n2", setRSOwnerRef2),
|
*test.BuildTestPod("p22", 100, 0, "n2", setRSOwnerRef2),
|
||||||
test.BuildTestPod("p23", 100, 0, "n2", setRSOwnerRef2),
|
*test.BuildTestPod("p23", 100, 0, "n2", setRSOwnerRef2),
|
||||||
test.BuildTestPod("p24", 100, 0, "n2", setRSOwnerRef2),
|
*test.BuildTestPod("p24", 100, 0, "n2", setRSOwnerRef2),
|
||||||
test.BuildTestPod("p25", 100, 0, "n3", setRSOwnerRef2),
|
*test.BuildTestPod("p25", 100, 0, "n3", setRSOwnerRef2),
|
||||||
test.BuildTestPod("p26", 100, 0, "n3", setRSOwnerRef2),
|
*test.BuildTestPod("p26", 100, 0, "n3", setRSOwnerRef2),
|
||||||
test.BuildTestPod("p27", 100, 0, "n3", setRSOwnerRef2),
|
*test.BuildTestPod("p27", 100, 0, "n3", setRSOwnerRef2),
|
||||||
test.BuildTestPod("p28", 100, 0, "n3", setRSOwnerRef2),
|
*test.BuildTestPod("p28", 100, 0, "n3", setRSOwnerRef2),
|
||||||
test.BuildTestPod("p29", 100, 0, "n3", setRSOwnerRef2),
|
*test.BuildTestPod("p29", 100, 0, "n3", setRSOwnerRef2),
|
||||||
},
|
},
|
||||||
expectedEvictedPodCount: 4,
|
expectedEvictedPodCount: 4,
|
||||||
nodes: []*v1.Node{
|
nodes: []*v1.Node{
|
||||||
@@ -592,13 +533,14 @@ func TestRemoveDuplicatesUniformly(t *testing.T) {
|
|||||||
test.BuildTestNode("n2", 2000, 3000, 10, nil),
|
test.BuildTestNode("n2", 2000, 3000, 10, nil),
|
||||||
test.BuildTestNode("n3", 2000, 3000, 10, nil),
|
test.BuildTestNode("n3", 2000, 3000, 10, nil),
|
||||||
},
|
},
|
||||||
|
strategy: api.DeschedulerStrategy{},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
description: "Evict pods with number of pods less than nodes",
|
description: "Evict pods with number of pods less than nodes",
|
||||||
pods: []*v1.Pod{
|
pods: []v1.Pod{
|
||||||
// (2,0,0) -> (1,1,0) -> 1 eviction
|
// (2,0,0) -> (1,1,0) -> 1 eviction
|
||||||
test.BuildTestPod("p1", 100, 0, "n1", test.SetRSOwnerRef),
|
*test.BuildTestPod("p1", 100, 0, "n1", test.SetRSOwnerRef),
|
||||||
test.BuildTestPod("p2", 100, 0, "n1", test.SetRSOwnerRef),
|
*test.BuildTestPod("p2", 100, 0, "n1", test.SetRSOwnerRef),
|
||||||
},
|
},
|
||||||
expectedEvictedPodCount: 1,
|
expectedEvictedPodCount: 1,
|
||||||
nodes: []*v1.Node{
|
nodes: []*v1.Node{
|
||||||
@@ -606,17 +548,18 @@ func TestRemoveDuplicatesUniformly(t *testing.T) {
|
|||||||
test.BuildTestNode("n2", 2000, 3000, 10, nil),
|
test.BuildTestNode("n2", 2000, 3000, 10, nil),
|
||||||
test.BuildTestNode("n3", 2000, 3000, 10, nil),
|
test.BuildTestNode("n3", 2000, 3000, 10, nil),
|
||||||
},
|
},
|
||||||
|
strategy: api.DeschedulerStrategy{},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
description: "Evict pods with number of pods less than nodes, but ignore different pods with the same ownerref",
|
description: "Evict pods with number of pods less than nodes, but ignore different pods with the same ownerref",
|
||||||
pods: []*v1.Pod{
|
pods: []v1.Pod{
|
||||||
// (1, 0, 0) for "bar","baz" images -> no eviction, even with a matching ownerKey
|
// (1, 0, 0) for "bar","baz" images -> no eviction, even with a matching ownerKey
|
||||||
// (2, 0, 0) for "foo" image -> (1,1,0) - 1 eviction
|
// (2, 0, 0) for "foo" image -> (1,1,0) - 1 eviction
|
||||||
// In this case the only "real" duplicates are p1 and p4, so one of those should be evicted
|
// In this case the only "real" duplicates are p1 and p4, so one of those should be evicted
|
||||||
buildTestPodWithImage("p1", "n1", "foo"),
|
*buildTestPodWithImage("p1", "n1", "foo"),
|
||||||
buildTestPodWithImage("p2", "n1", "bar"),
|
*buildTestPodWithImage("p2", "n1", "bar"),
|
||||||
buildTestPodWithImage("p3", "n1", "baz"),
|
*buildTestPodWithImage("p3", "n1", "baz"),
|
||||||
buildTestPodWithImage("p4", "n1", "foo"),
|
*buildTestPodWithImage("p4", "n1", "foo"),
|
||||||
},
|
},
|
||||||
expectedEvictedPodCount: 1,
|
expectedEvictedPodCount: 1,
|
||||||
nodes: []*v1.Node{
|
nodes: []*v1.Node{
|
||||||
@@ -624,12 +567,13 @@ func TestRemoveDuplicatesUniformly(t *testing.T) {
|
|||||||
test.BuildTestNode("n2", 2000, 3000, 10, nil),
|
test.BuildTestNode("n2", 2000, 3000, 10, nil),
|
||||||
test.BuildTestNode("n3", 2000, 3000, 10, nil),
|
test.BuildTestNode("n3", 2000, 3000, 10, nil),
|
||||||
},
|
},
|
||||||
|
strategy: api.DeschedulerStrategy{},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
description: "Evict pods with a single pod with three nodes",
|
description: "Evict pods with a single pod with three nodes",
|
||||||
pods: []*v1.Pod{
|
pods: []v1.Pod{
|
||||||
// (2,0,0) -> (1,1,0) -> 1 eviction
|
// (2,0,0) -> (1,1,0) -> 1 eviction
|
||||||
test.BuildTestPod("p1", 100, 0, "n1", test.SetRSOwnerRef),
|
*test.BuildTestPod("p1", 100, 0, "n1", test.SetRSOwnerRef),
|
||||||
},
|
},
|
||||||
expectedEvictedPodCount: 0,
|
expectedEvictedPodCount: 0,
|
||||||
nodes: []*v1.Node{
|
nodes: []*v1.Node{
|
||||||
@@ -637,20 +581,21 @@ func TestRemoveDuplicatesUniformly(t *testing.T) {
|
|||||||
test.BuildTestNode("n2", 2000, 3000, 10, nil),
|
test.BuildTestNode("n2", 2000, 3000, 10, nil),
|
||||||
test.BuildTestNode("n3", 2000, 3000, 10, nil),
|
test.BuildTestNode("n3", 2000, 3000, 10, nil),
|
||||||
},
|
},
|
||||||
|
strategy: api.DeschedulerStrategy{},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
description: "Evict pods uniformly respecting taints",
|
description: "Evict pods uniformly respecting taints",
|
||||||
pods: []*v1.Pod{
|
pods: []v1.Pod{
|
||||||
// (5,3,1,0,0,0) -> (3,3,3,0,0,0) -> 2 evictions
|
// (5,3,1,0,0,0) -> (3,3,3,0,0,0) -> 2 evictions
|
||||||
test.BuildTestPod("p1", 100, 0, "worker1", setTolerationsK1),
|
*test.BuildTestPod("p1", 100, 0, "worker1", setTolerationsK1),
|
||||||
test.BuildTestPod("p2", 100, 0, "worker1", setTolerationsK2),
|
*test.BuildTestPod("p2", 100, 0, "worker1", setTolerationsK2),
|
||||||
test.BuildTestPod("p3", 100, 0, "worker1", setTolerationsK1),
|
*test.BuildTestPod("p3", 100, 0, "worker1", setTolerationsK1),
|
||||||
test.BuildTestPod("p4", 100, 0, "worker1", setTolerationsK2),
|
*test.BuildTestPod("p4", 100, 0, "worker1", setTolerationsK2),
|
||||||
test.BuildTestPod("p5", 100, 0, "worker1", setTolerationsK1),
|
*test.BuildTestPod("p5", 100, 0, "worker1", setTolerationsK1),
|
||||||
test.BuildTestPod("p6", 100, 0, "worker2", setTolerationsK2),
|
*test.BuildTestPod("p6", 100, 0, "worker2", setTolerationsK2),
|
||||||
test.BuildTestPod("p7", 100, 0, "worker2", setTolerationsK1),
|
*test.BuildTestPod("p7", 100, 0, "worker2", setTolerationsK1),
|
||||||
test.BuildTestPod("p8", 100, 0, "worker2", setTolerationsK2),
|
*test.BuildTestPod("p8", 100, 0, "worker2", setTolerationsK2),
|
||||||
test.BuildTestPod("p9", 100, 0, "worker3", setTolerationsK1),
|
*test.BuildTestPod("p9", 100, 0, "worker3", setTolerationsK1),
|
||||||
},
|
},
|
||||||
expectedEvictedPodCount: 2,
|
expectedEvictedPodCount: 2,
|
||||||
nodes: []*v1.Node{
|
nodes: []*v1.Node{
|
||||||
@@ -661,20 +606,21 @@ func TestRemoveDuplicatesUniformly(t *testing.T) {
|
|||||||
test.BuildTestNode("master2", 2000, 3000, 10, setMasterNoScheduleTaint),
|
test.BuildTestNode("master2", 2000, 3000, 10, setMasterNoScheduleTaint),
|
||||||
test.BuildTestNode("master3", 2000, 3000, 10, setMasterNoScheduleTaint),
|
test.BuildTestNode("master3", 2000, 3000, 10, setMasterNoScheduleTaint),
|
||||||
},
|
},
|
||||||
|
strategy: api.DeschedulerStrategy{},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
description: "Evict pods uniformly respecting RequiredDuringSchedulingIgnoredDuringExecution node affinity",
|
description: "Evict pods uniformly respecting RequiredDuringSchedulingIgnoredDuringExecution node affinity",
|
||||||
pods: []*v1.Pod{
|
pods: []v1.Pod{
|
||||||
// (5,3,1,0,0,0) -> (3,3,3,0,0,0) -> 2 evictions
|
// (5,3,1,0,0,0) -> (3,3,3,0,0,0) -> 2 evictions
|
||||||
test.BuildTestPod("p1", 100, 0, "worker1", setNotMasterNodeSelectorK1),
|
*test.BuildTestPod("p1", 100, 0, "worker1", setNotMasterNodeSelectorK1),
|
||||||
test.BuildTestPod("p2", 100, 0, "worker1", setNotMasterNodeSelectorK2),
|
*test.BuildTestPod("p2", 100, 0, "worker1", setNotMasterNodeSelectorK2),
|
||||||
test.BuildTestPod("p3", 100, 0, "worker1", setNotMasterNodeSelectorK1),
|
*test.BuildTestPod("p3", 100, 0, "worker1", setNotMasterNodeSelectorK1),
|
||||||
test.BuildTestPod("p4", 100, 0, "worker1", setNotMasterNodeSelectorK2),
|
*test.BuildTestPod("p4", 100, 0, "worker1", setNotMasterNodeSelectorK2),
|
||||||
test.BuildTestPod("p5", 100, 0, "worker1", setNotMasterNodeSelectorK1),
|
*test.BuildTestPod("p5", 100, 0, "worker1", setNotMasterNodeSelectorK1),
|
||||||
test.BuildTestPod("p6", 100, 0, "worker2", setNotMasterNodeSelectorK2),
|
*test.BuildTestPod("p6", 100, 0, "worker2", setNotMasterNodeSelectorK2),
|
||||||
test.BuildTestPod("p7", 100, 0, "worker2", setNotMasterNodeSelectorK1),
|
*test.BuildTestPod("p7", 100, 0, "worker2", setNotMasterNodeSelectorK1),
|
||||||
test.BuildTestPod("p8", 100, 0, "worker2", setNotMasterNodeSelectorK2),
|
*test.BuildTestPod("p8", 100, 0, "worker2", setNotMasterNodeSelectorK2),
|
||||||
test.BuildTestPod("p9", 100, 0, "worker3", setNotMasterNodeSelectorK1),
|
*test.BuildTestPod("p9", 100, 0, "worker3", setNotMasterNodeSelectorK1),
|
||||||
},
|
},
|
||||||
expectedEvictedPodCount: 2,
|
expectedEvictedPodCount: 2,
|
||||||
nodes: []*v1.Node{
|
nodes: []*v1.Node{
|
||||||
@@ -685,20 +631,21 @@ func TestRemoveDuplicatesUniformly(t *testing.T) {
|
|||||||
test.BuildTestNode("master2", 2000, 3000, 10, setMasterNoScheduleLabel),
|
test.BuildTestNode("master2", 2000, 3000, 10, setMasterNoScheduleLabel),
|
||||||
test.BuildTestNode("master3", 2000, 3000, 10, setMasterNoScheduleLabel),
|
test.BuildTestNode("master3", 2000, 3000, 10, setMasterNoScheduleLabel),
|
||||||
},
|
},
|
||||||
|
strategy: api.DeschedulerStrategy{},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
description: "Evict pods uniformly respecting node selector",
|
description: "Evict pods uniformly respecting node selector",
|
||||||
pods: []*v1.Pod{
|
pods: []v1.Pod{
|
||||||
// (5,3,1,0,0,0) -> (3,3,3,0,0,0) -> 2 evictions
|
// (5,3,1,0,0,0) -> (3,3,3,0,0,0) -> 2 evictions
|
||||||
test.BuildTestPod("p1", 100, 0, "worker1", setWorkerLabelSelectorK1),
|
*test.BuildTestPod("p1", 100, 0, "worker1", setWorkerLabelSelectorK1),
|
||||||
test.BuildTestPod("p2", 100, 0, "worker1", setWorkerLabelSelectorK2),
|
*test.BuildTestPod("p2", 100, 0, "worker1", setWorkerLabelSelectorK2),
|
||||||
test.BuildTestPod("p3", 100, 0, "worker1", setWorkerLabelSelectorK1),
|
*test.BuildTestPod("p3", 100, 0, "worker1", setWorkerLabelSelectorK1),
|
||||||
test.BuildTestPod("p4", 100, 0, "worker1", setWorkerLabelSelectorK2),
|
*test.BuildTestPod("p4", 100, 0, "worker1", setWorkerLabelSelectorK2),
|
||||||
test.BuildTestPod("p5", 100, 0, "worker1", setWorkerLabelSelectorK1),
|
*test.BuildTestPod("p5", 100, 0, "worker1", setWorkerLabelSelectorK1),
|
||||||
test.BuildTestPod("p6", 100, 0, "worker2", setWorkerLabelSelectorK2),
|
*test.BuildTestPod("p6", 100, 0, "worker2", setWorkerLabelSelectorK2),
|
||||||
test.BuildTestPod("p7", 100, 0, "worker2", setWorkerLabelSelectorK1),
|
*test.BuildTestPod("p7", 100, 0, "worker2", setWorkerLabelSelectorK1),
|
||||||
test.BuildTestPod("p8", 100, 0, "worker2", setWorkerLabelSelectorK2),
|
*test.BuildTestPod("p8", 100, 0, "worker2", setWorkerLabelSelectorK2),
|
||||||
test.BuildTestPod("p9", 100, 0, "worker3", setWorkerLabelSelectorK1),
|
*test.BuildTestPod("p9", 100, 0, "worker3", setWorkerLabelSelectorK1),
|
||||||
},
|
},
|
||||||
expectedEvictedPodCount: 2,
|
expectedEvictedPodCount: 2,
|
||||||
nodes: []*v1.Node{
|
nodes: []*v1.Node{
|
||||||
@@ -709,20 +656,21 @@ func TestRemoveDuplicatesUniformly(t *testing.T) {
|
|||||||
test.BuildTestNode("master2", 2000, 3000, 10, nil),
|
test.BuildTestNode("master2", 2000, 3000, 10, nil),
|
||||||
test.BuildTestNode("master3", 2000, 3000, 10, nil),
|
test.BuildTestNode("master3", 2000, 3000, 10, nil),
|
||||||
},
|
},
|
||||||
|
strategy: api.DeschedulerStrategy{},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
description: "Evict pods uniformly respecting node selector with zero target nodes",
|
description: "Evict pods uniformly respecting node selector with zero target nodes",
|
||||||
pods: []*v1.Pod{
|
pods: []v1.Pod{
|
||||||
// (5,3,1,0,0,0) -> (3,3,3,0,0,0) -> 2 evictions
|
// (5,3,1,0,0,0) -> (3,3,3,0,0,0) -> 2 evictions
|
||||||
test.BuildTestPod("p1", 100, 0, "worker1", setWorkerLabelSelectorK1),
|
*test.BuildTestPod("p1", 100, 0, "worker1", setWorkerLabelSelectorK1),
|
||||||
test.BuildTestPod("p2", 100, 0, "worker1", setWorkerLabelSelectorK2),
|
*test.BuildTestPod("p2", 100, 0, "worker1", setWorkerLabelSelectorK2),
|
||||||
test.BuildTestPod("p3", 100, 0, "worker1", setWorkerLabelSelectorK1),
|
*test.BuildTestPod("p3", 100, 0, "worker1", setWorkerLabelSelectorK1),
|
||||||
test.BuildTestPod("p4", 100, 0, "worker1", setWorkerLabelSelectorK2),
|
*test.BuildTestPod("p4", 100, 0, "worker1", setWorkerLabelSelectorK2),
|
||||||
test.BuildTestPod("p5", 100, 0, "worker1", setWorkerLabelSelectorK1),
|
*test.BuildTestPod("p5", 100, 0, "worker1", setWorkerLabelSelectorK1),
|
||||||
test.BuildTestPod("p6", 100, 0, "worker2", setWorkerLabelSelectorK2),
|
*test.BuildTestPod("p6", 100, 0, "worker2", setWorkerLabelSelectorK2),
|
||||||
test.BuildTestPod("p7", 100, 0, "worker2", setWorkerLabelSelectorK1),
|
*test.BuildTestPod("p7", 100, 0, "worker2", setWorkerLabelSelectorK1),
|
||||||
test.BuildTestPod("p8", 100, 0, "worker2", setWorkerLabelSelectorK2),
|
*test.BuildTestPod("p8", 100, 0, "worker2", setWorkerLabelSelectorK2),
|
||||||
test.BuildTestPod("p9", 100, 0, "worker3", setWorkerLabelSelectorK1),
|
*test.BuildTestPod("p9", 100, 0, "worker3", setWorkerLabelSelectorK1),
|
||||||
},
|
},
|
||||||
expectedEvictedPodCount: 0,
|
expectedEvictedPodCount: 0,
|
||||||
nodes: []*v1.Node{
|
nodes: []*v1.Node{
|
||||||
@@ -733,86 +681,31 @@ func TestRemoveDuplicatesUniformly(t *testing.T) {
|
|||||||
test.BuildTestNode("master2", 2000, 3000, 10, nil),
|
test.BuildTestNode("master2", 2000, 3000, 10, nil),
|
||||||
test.BuildTestNode("master3", 2000, 3000, 10, nil),
|
test.BuildTestNode("master3", 2000, 3000, 10, nil),
|
||||||
},
|
},
|
||||||
|
strategy: api.DeschedulerStrategy{},
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
for _, testCase := range testCases {
|
for _, testCase := range testCases {
|
||||||
t.Run(testCase.description, func(t *testing.T) {
|
t.Run(testCase.description, func(t *testing.T) {
|
||||||
ctx, cancel := context.WithCancel(context.Background())
|
fakeClient := &fake.Clientset{}
|
||||||
defer cancel()
|
fakeClient.Fake.AddReactor("list", "pods", func(action core.Action) (bool, runtime.Object, error) {
|
||||||
|
return true, &v1.PodList{Items: testCase.pods}, nil
|
||||||
var objs []runtime.Object
|
})
|
||||||
for _, node := range testCase.nodes {
|
|
||||||
objs = append(objs, node)
|
|
||||||
}
|
|
||||||
for _, pod := range testCase.pods {
|
|
||||||
objs = append(objs, pod)
|
|
||||||
}
|
|
||||||
fakeClient := fake.NewSimpleClientset(objs...)
|
|
||||||
|
|
||||||
sharedInformerFactory := informers.NewSharedInformerFactory(fakeClient, 0)
|
|
||||||
podInformer := sharedInformerFactory.Core().V1().Pods()
|
|
||||||
|
|
||||||
getPodsAssignedToNode, err := podutil.BuildGetPodsAssignedToNodeFunc(podInformer)
|
|
||||||
if err != nil {
|
|
||||||
t.Errorf("Build get pods assigned to node function error: %v", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
sharedInformerFactory.Start(ctx.Done())
|
|
||||||
sharedInformerFactory.WaitForCacheSync(ctx.Done())
|
|
||||||
|
|
||||||
eventRecorder := &events.FakeRecorder{}
|
|
||||||
|
|
||||||
podEvictor := evictions.NewPodEvictor(
|
podEvictor := evictions.NewPodEvictor(
|
||||||
fakeClient,
|
fakeClient,
|
||||||
policyv1.SchemeGroupVersion.String(),
|
policyv1.SchemeGroupVersion.String(),
|
||||||
false,
|
false,
|
||||||
nil,
|
testCase.maxPodsToEvictPerNode,
|
||||||
nil,
|
|
||||||
testCase.nodes,
|
testCase.nodes,
|
||||||
false,
|
false,
|
||||||
eventRecorder,
|
false,
|
||||||
|
false,
|
||||||
)
|
)
|
||||||
|
|
||||||
defaultEvictorFilterArgs := &defaultevictor.DefaultEvictorArgs{
|
RemoveDuplicatePods(ctx, fakeClient, testCase.strategy, testCase.nodes, podEvictor)
|
||||||
EvictLocalStoragePods: false,
|
podsEvicted := podEvictor.TotalEvicted()
|
||||||
EvictSystemCriticalPods: false,
|
if podsEvicted != testCase.expectedEvictedPodCount {
|
||||||
IgnorePvcPods: false,
|
t.Errorf("Test error for description: %s. Expected evicted pods count %v, got %v", testCase.description, testCase.expectedEvictedPodCount, podsEvicted)
|
||||||
EvictFailedBarePods: false,
|
|
||||||
}
|
|
||||||
|
|
||||||
evictorFilter, err := defaultevictor.New(
|
|
||||||
defaultEvictorFilterArgs,
|
|
||||||
&frameworkfake.HandleImpl{
|
|
||||||
ClientsetImpl: fakeClient,
|
|
||||||
GetPodsAssignedToNodeFuncImpl: getPodsAssignedToNode,
|
|
||||||
SharedInformerFactoryImpl: sharedInformerFactory,
|
|
||||||
},
|
|
||||||
)
|
|
||||||
|
|
||||||
if err != nil {
|
|
||||||
t.Fatalf("Unable to initialize the plugin: %v", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
handle := &frameworkfake.HandleImpl{
|
|
||||||
ClientsetImpl: fakeClient,
|
|
||||||
GetPodsAssignedToNodeFuncImpl: getPodsAssignedToNode,
|
|
||||||
PodEvictorImpl: podEvictor,
|
|
||||||
EvictorFilterImpl: evictorFilter.(framework.EvictorPlugin),
|
|
||||||
SharedInformerFactoryImpl: sharedInformerFactory,
|
|
||||||
}
|
|
||||||
|
|
||||||
plugin, err := New(&componentconfig.RemoveDuplicatesArgs{},
|
|
||||||
handle,
|
|
||||||
)
|
|
||||||
if err != nil {
|
|
||||||
t.Fatalf("Unable to initialize the plugin: %v", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
plugin.(framework.BalancePlugin).Balance(ctx, testCase.nodes)
|
|
||||||
actualEvictedPodCount := podEvictor.TotalEvicted()
|
|
||||||
if actualEvictedPodCount != testCase.expectedEvictedPodCount {
|
|
||||||
t.Errorf("Test %#v failed, Unexpected no of pods evicted: pods evicted: %d, expected: %d", testCase.description, actualEvictedPodCount, testCase.expectedEvictedPodCount)
|
|
||||||
}
|
}
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
163
pkg/descheduler/strategies/failedpods.go
Normal file
163
pkg/descheduler/strategies/failedpods.go
Normal file
@@ -0,0 +1,163 @@
|
|||||||
|
package strategies
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"fmt"
|
||||||
|
"k8s.io/apimachinery/pkg/util/sets"
|
||||||
|
|
||||||
|
v1 "k8s.io/api/core/v1"
|
||||||
|
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
||||||
|
utilerrors "k8s.io/apimachinery/pkg/util/errors"
|
||||||
|
clientset "k8s.io/client-go/kubernetes"
|
||||||
|
"k8s.io/klog/v2"
|
||||||
|
|
||||||
|
"sigs.k8s.io/descheduler/pkg/api"
|
||||||
|
"sigs.k8s.io/descheduler/pkg/descheduler/evictions"
|
||||||
|
podutil "sigs.k8s.io/descheduler/pkg/descheduler/pod"
|
||||||
|
"sigs.k8s.io/descheduler/pkg/descheduler/strategies/validation"
|
||||||
|
)
|
||||||
|
|
||||||
|
// validatedFailedPodsStrategyParams contains validated strategy parameters
|
||||||
|
type validatedFailedPodsStrategyParams struct {
|
||||||
|
*validation.ValidatedStrategyParams
|
||||||
|
includingInitContainers bool
|
||||||
|
reasons sets.String
|
||||||
|
excludeOwnerKinds sets.String
|
||||||
|
minPodLifetimeSeconds *uint
|
||||||
|
}
|
||||||
|
|
||||||
|
// RemoveFailedPods removes Pods that are in failed status phase.
|
||||||
|
func RemoveFailedPods(
|
||||||
|
ctx context.Context,
|
||||||
|
client clientset.Interface,
|
||||||
|
strategy api.DeschedulerStrategy,
|
||||||
|
nodes []*v1.Node,
|
||||||
|
podEvictor *evictions.PodEvictor,
|
||||||
|
) {
|
||||||
|
strategyParams, err := validateAndParseRemoveFailedPodsParams(ctx, client, strategy.Params)
|
||||||
|
if err != nil {
|
||||||
|
klog.ErrorS(err, "Invalid RemoveFailedPods parameters")
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
evictable := podEvictor.Evictable(
|
||||||
|
evictions.WithPriorityThreshold(strategyParams.ThresholdPriority),
|
||||||
|
evictions.WithNodeFit(strategyParams.NodeFit),
|
||||||
|
evictions.WithLabelSelector(strategyParams.LabelSelector),
|
||||||
|
)
|
||||||
|
|
||||||
|
for _, node := range nodes {
|
||||||
|
klog.V(1).InfoS("Processing node", "node", klog.KObj(node))
|
||||||
|
fieldSelectorString := "spec.nodeName=" + node.Name + ",status.phase=" + string(v1.PodFailed)
|
||||||
|
pods, err := podutil.ListPodsOnANodeWithFieldSelector(
|
||||||
|
ctx,
|
||||||
|
client,
|
||||||
|
node,
|
||||||
|
fieldSelectorString,
|
||||||
|
podutil.WithFilter(evictable.IsEvictable),
|
||||||
|
podutil.WithNamespaces(strategyParams.IncludedNamespaces.UnsortedList()),
|
||||||
|
podutil.WithoutNamespaces(strategyParams.ExcludedNamespaces.UnsortedList()),
|
||||||
|
podutil.WithLabelSelector(strategy.Params.LabelSelector),
|
||||||
|
)
|
||||||
|
if err != nil {
|
||||||
|
klog.ErrorS(err, "Error listing a nodes failed pods", "node", klog.KObj(node))
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
|
for i, pod := range pods {
|
||||||
|
if err = validateFailedPodShouldEvict(pod, *strategyParams); err != nil {
|
||||||
|
klog.V(4).InfoS(fmt.Sprintf("ignoring pod for eviction due to: %s", err.Error()), "pod", klog.KObj(pod))
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
|
if _, err = podEvictor.EvictPod(ctx, pods[i], node, "FailedPod"); err != nil {
|
||||||
|
klog.ErrorS(err, "Error evicting pod", "pod", klog.KObj(pod))
|
||||||
|
break
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func validateAndParseRemoveFailedPodsParams(
|
||||||
|
ctx context.Context,
|
||||||
|
client clientset.Interface,
|
||||||
|
params *api.StrategyParameters,
|
||||||
|
) (*validatedFailedPodsStrategyParams, error) {
|
||||||
|
if params == nil {
|
||||||
|
return &validatedFailedPodsStrategyParams{}, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
strategyParams, err := validation.ValidateAndParseStrategyParams(ctx, client, params)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
var reasons, excludeOwnerKinds sets.String
|
||||||
|
var includingInitContainers bool
|
||||||
|
var minPodLifetimeSeconds *uint
|
||||||
|
if params.FailedPods != nil {
|
||||||
|
reasons = sets.NewString(params.FailedPods.Reasons...)
|
||||||
|
includingInitContainers = params.FailedPods.IncludingInitContainers
|
||||||
|
excludeOwnerKinds = sets.NewString(params.FailedPods.ExcludeOwnerKinds...)
|
||||||
|
minPodLifetimeSeconds = params.FailedPods.MinPodLifetimeSeconds
|
||||||
|
}
|
||||||
|
|
||||||
|
return &validatedFailedPodsStrategyParams{
|
||||||
|
ValidatedStrategyParams: strategyParams,
|
||||||
|
includingInitContainers: includingInitContainers,
|
||||||
|
reasons: reasons,
|
||||||
|
excludeOwnerKinds: excludeOwnerKinds,
|
||||||
|
minPodLifetimeSeconds: minPodLifetimeSeconds,
|
||||||
|
}, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// validateFailedPodShouldEvict looks at strategy params settings to see if the Pod
|
||||||
|
// should be evicted given the params in the PodFailed policy.
|
||||||
|
func validateFailedPodShouldEvict(pod *v1.Pod, strategyParams validatedFailedPodsStrategyParams) error {
|
||||||
|
var errs []error
|
||||||
|
|
||||||
|
if strategyParams.minPodLifetimeSeconds != nil {
|
||||||
|
podAgeSeconds := uint(metav1.Now().Sub(pod.GetCreationTimestamp().Local()).Seconds())
|
||||||
|
if podAgeSeconds < *strategyParams.minPodLifetimeSeconds {
|
||||||
|
errs = append(errs, fmt.Errorf("pod does not exceed the min age seconds of %d", *strategyParams.minPodLifetimeSeconds))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(strategyParams.excludeOwnerKinds) > 0 {
|
||||||
|
ownerRefList := podutil.OwnerRef(pod)
|
||||||
|
for _, owner := range ownerRefList {
|
||||||
|
if strategyParams.excludeOwnerKinds.Has(owner.Kind) {
|
||||||
|
errs = append(errs, fmt.Errorf("pod's owner kind of %s is excluded", owner.Kind))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(strategyParams.reasons) > 0 {
|
||||||
|
reasons := getFailedContainerStatusReasons(pod.Status.ContainerStatuses)
|
||||||
|
|
||||||
|
if strategyParams.includingInitContainers {
|
||||||
|
reasons = append(reasons, getFailedContainerStatusReasons(pod.Status.InitContainerStatuses)...)
|
||||||
|
}
|
||||||
|
|
||||||
|
if !strategyParams.reasons.HasAny(reasons...) {
|
||||||
|
errs = append(errs, fmt.Errorf("pod does not match any of the reasons"))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return utilerrors.NewAggregate(errs)
|
||||||
|
}
|
||||||
|
|
||||||
|
func getFailedContainerStatusReasons(containerStatuses []v1.ContainerStatus) []string {
|
||||||
|
reasons := make([]string, 0)
|
||||||
|
|
||||||
|
for _, containerStatus := range containerStatuses {
|
||||||
|
if containerStatus.State.Waiting != nil && containerStatus.State.Waiting.Reason != "" {
|
||||||
|
reasons = append(reasons, containerStatus.State.Waiting.Reason)
|
||||||
|
}
|
||||||
|
if containerStatus.State.Terminated != nil && containerStatus.State.Terminated.Reason != "" {
|
||||||
|
reasons = append(reasons, containerStatus.State.Terminated.Reason)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return reasons
|
||||||
|
}
|
||||||
274
pkg/descheduler/strategies/failedpods_test.go
Normal file
274
pkg/descheduler/strategies/failedpods_test.go
Normal file
@@ -0,0 +1,274 @@
|
|||||||
|
package strategies
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
||||||
|
"testing"
|
||||||
|
|
||||||
|
v1 "k8s.io/api/core/v1"
|
||||||
|
policyv1 "k8s.io/api/policy/v1"
|
||||||
|
"k8s.io/apimachinery/pkg/runtime"
|
||||||
|
"k8s.io/client-go/kubernetes/fake"
|
||||||
|
core "k8s.io/client-go/testing"
|
||||||
|
"sigs.k8s.io/descheduler/pkg/api"
|
||||||
|
"sigs.k8s.io/descheduler/pkg/descheduler/evictions"
|
||||||
|
"sigs.k8s.io/descheduler/test"
|
||||||
|
)
|
||||||
|
|
||||||
|
var (
|
||||||
|
OneHourInSeconds uint = 3600
|
||||||
|
)
|
||||||
|
|
||||||
|
func TestRemoveFailedPods(t *testing.T) {
|
||||||
|
ctx := context.Background()
|
||||||
|
|
||||||
|
createStrategy := func(enabled, includingInitContainers bool, reasons, excludeKinds []string, minAgeSeconds *uint, nodeFit bool) api.DeschedulerStrategy {
|
||||||
|
return api.DeschedulerStrategy{
|
||||||
|
Enabled: enabled,
|
||||||
|
Params: &api.StrategyParameters{
|
||||||
|
FailedPods: &api.FailedPods{
|
||||||
|
Reasons: reasons,
|
||||||
|
IncludingInitContainers: includingInitContainers,
|
||||||
|
ExcludeOwnerKinds: excludeKinds,
|
||||||
|
MinPodLifetimeSeconds: minAgeSeconds,
|
||||||
|
},
|
||||||
|
NodeFit: nodeFit,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
tests := []struct {
|
||||||
|
description string
|
||||||
|
nodes []*v1.Node
|
||||||
|
strategy api.DeschedulerStrategy
|
||||||
|
expectedEvictedPodCount int
|
||||||
|
pods []v1.Pod
|
||||||
|
}{
|
||||||
|
{
|
||||||
|
description: "0 failures, 0 evictions",
|
||||||
|
strategy: createStrategy(true, false, nil, nil, nil, false),
|
||||||
|
nodes: []*v1.Node{test.BuildTestNode("node1", 2000, 3000, 10, nil)},
|
||||||
|
expectedEvictedPodCount: 0,
|
||||||
|
pods: []v1.Pod{}, // no pods come back with field selector phase=Failed
|
||||||
|
},
|
||||||
|
{
|
||||||
|
description: "1 container terminated with reason NodeAffinity, 1 eviction",
|
||||||
|
strategy: createStrategy(true, false, nil, nil, nil, false),
|
||||||
|
nodes: []*v1.Node{test.BuildTestNode("node1", 2000, 3000, 10, nil)},
|
||||||
|
expectedEvictedPodCount: 1,
|
||||||
|
pods: []v1.Pod{
|
||||||
|
buildTestPod("p1", "node1", nil, &v1.ContainerState{
|
||||||
|
Terminated: &v1.ContainerStateTerminated{Reason: "NodeAffinity"},
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
description: "1 init container terminated with reason NodeAffinity, 1 eviction",
|
||||||
|
strategy: createStrategy(true, true, nil, nil, nil, false),
|
||||||
|
nodes: []*v1.Node{test.BuildTestNode("node1", 2000, 3000, 10, nil)},
|
||||||
|
expectedEvictedPodCount: 1,
|
||||||
|
pods: []v1.Pod{
|
||||||
|
buildTestPod("p1", "node1", &v1.ContainerState{
|
||||||
|
Terminated: &v1.ContainerStateTerminated{Reason: "NodeAffinity"},
|
||||||
|
}, nil),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
description: "1 init container waiting with reason CreateContainerConfigError, 1 eviction",
|
||||||
|
strategy: createStrategy(true, true, nil, nil, nil, false),
|
||||||
|
nodes: []*v1.Node{test.BuildTestNode("node1", 2000, 3000, 10, nil)},
|
||||||
|
expectedEvictedPodCount: 1,
|
||||||
|
pods: []v1.Pod{
|
||||||
|
buildTestPod("p1", "node1", &v1.ContainerState{
|
||||||
|
Waiting: &v1.ContainerStateWaiting{Reason: "CreateContainerConfigError"},
|
||||||
|
}, nil),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
description: "2 init container waiting with reason CreateContainerConfigError, 2 nodes, 2 evictions",
|
||||||
|
strategy: createStrategy(true, true, nil, nil, nil, false),
|
||||||
|
nodes: []*v1.Node{
|
||||||
|
test.BuildTestNode("node1", 2000, 3000, 10, nil),
|
||||||
|
test.BuildTestNode("node2", 2000, 3000, 10, nil),
|
||||||
|
},
|
||||||
|
expectedEvictedPodCount: 2,
|
||||||
|
pods: []v1.Pod{
|
||||||
|
buildTestPod("p1", "node1", &v1.ContainerState{
|
||||||
|
Terminated: &v1.ContainerStateTerminated{Reason: "CreateContainerConfigError"},
|
||||||
|
}, nil),
|
||||||
|
buildTestPod("p2", "node2", &v1.ContainerState{
|
||||||
|
Terminated: &v1.ContainerStateTerminated{Reason: "CreateContainerConfigError"},
|
||||||
|
}, nil),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
description: "include reason=CreateContainerConfigError, 1 container terminated with reason CreateContainerConfigError, 1 eviction",
|
||||||
|
strategy: createStrategy(true, false, []string{"CreateContainerConfigError"}, nil, nil, false),
|
||||||
|
nodes: []*v1.Node{test.BuildTestNode("node1", 2000, 3000, 10, nil)},
|
||||||
|
expectedEvictedPodCount: 1,
|
||||||
|
pods: []v1.Pod{
|
||||||
|
buildTestPod("p1", "node1", nil, &v1.ContainerState{
|
||||||
|
Terminated: &v1.ContainerStateTerminated{Reason: "CreateContainerConfigError"},
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
description: "include reason=CreateContainerConfigError+NodeAffinity, 1 container terminated with reason CreateContainerConfigError, 1 eviction",
|
||||||
|
strategy: createStrategy(true, false, []string{"CreateContainerConfigError", "NodeAffinity"}, nil, nil, false),
|
||||||
|
nodes: []*v1.Node{test.BuildTestNode("node1", 2000, 3000, 10, nil)},
|
||||||
|
expectedEvictedPodCount: 1,
|
||||||
|
pods: []v1.Pod{
|
||||||
|
buildTestPod("p1", "node1", nil, &v1.ContainerState{
|
||||||
|
Terminated: &v1.ContainerStateTerminated{Reason: "CreateContainerConfigError"},
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
description: "include reason=CreateContainerConfigError, 1 container terminated with reason NodeAffinity, 0 eviction",
|
||||||
|
strategy: createStrategy(true, false, []string{"CreateContainerConfigError"}, nil, nil, false),
|
||||||
|
nodes: []*v1.Node{test.BuildTestNode("node1", 2000, 3000, 10, nil)},
|
||||||
|
expectedEvictedPodCount: 0,
|
||||||
|
pods: []v1.Pod{
|
||||||
|
buildTestPod("p1", "node1", nil, &v1.ContainerState{
|
||||||
|
Terminated: &v1.ContainerStateTerminated{Reason: "NodeAffinity"},
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
description: "include init container=false, 1 init container waiting with reason CreateContainerConfigError, 0 eviction",
|
||||||
|
strategy: createStrategy(true, false, []string{"CreateContainerConfigError"}, nil, nil, false),
|
||||||
|
nodes: []*v1.Node{test.BuildTestNode("node1", 2000, 3000, 10, nil)},
|
||||||
|
expectedEvictedPodCount: 0,
|
||||||
|
pods: []v1.Pod{
|
||||||
|
buildTestPod("p1", "node1", &v1.ContainerState{
|
||||||
|
Waiting: &v1.ContainerStateWaiting{Reason: "CreateContainerConfigError"},
|
||||||
|
}, nil),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
description: "lifetime 1 hour, 1 container terminated with reason NodeAffinity, 0 eviction",
|
||||||
|
strategy: createStrategy(true, false, nil, nil, &OneHourInSeconds, false),
|
||||||
|
nodes: []*v1.Node{test.BuildTestNode("node1", 2000, 3000, 10, nil)},
|
||||||
|
expectedEvictedPodCount: 0,
|
||||||
|
pods: []v1.Pod{
|
||||||
|
buildTestPod("p1", "node1", nil, &v1.ContainerState{
|
||||||
|
Terminated: &v1.ContainerStateTerminated{Reason: "NodeAffinity"},
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
description: "nodeFit=true, 1 unschedulable node, 1 container terminated with reason NodeAffinity, 0 eviction",
|
||||||
|
strategy: createStrategy(true, false, nil, nil, nil, true),
|
||||||
|
nodes: []*v1.Node{test.BuildTestNode("node1", 2000, 3000, 10, func(node *v1.Node) {
|
||||||
|
node.Spec.Unschedulable = true
|
||||||
|
})},
|
||||||
|
expectedEvictedPodCount: 0,
|
||||||
|
pods: []v1.Pod{
|
||||||
|
buildTestPod("p1", "node1", nil, &v1.ContainerState{
|
||||||
|
Terminated: &v1.ContainerStateTerminated{Reason: "NodeAffinity"},
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
description: "excluded owner kind=ReplicaSet, 1 init container terminated with owner kind=ReplicaSet, 0 eviction",
|
||||||
|
strategy: createStrategy(true, true, nil, []string{"ReplicaSet"}, nil, false),
|
||||||
|
nodes: []*v1.Node{test.BuildTestNode("node1", 2000, 3000, 10, nil)},
|
||||||
|
expectedEvictedPodCount: 0,
|
||||||
|
pods: []v1.Pod{
|
||||||
|
buildTestPod("p1", "node1", &v1.ContainerState{
|
||||||
|
Terminated: &v1.ContainerStateTerminated{Reason: "NodeAffinity"},
|
||||||
|
}, nil),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
description: "excluded owner kind=DaemonSet, 1 init container terminated with owner kind=ReplicaSet, 1 eviction",
|
||||||
|
strategy: createStrategy(true, true, nil, []string{"DaemonSet"}, nil, false),
|
||||||
|
nodes: []*v1.Node{test.BuildTestNode("node1", 2000, 3000, 10, nil)},
|
||||||
|
expectedEvictedPodCount: 1,
|
||||||
|
pods: []v1.Pod{
|
||||||
|
buildTestPod("p1", "node1", &v1.ContainerState{
|
||||||
|
Terminated: &v1.ContainerStateTerminated{Reason: "NodeAffinity"},
|
||||||
|
}, nil),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
for _, tc := range tests {
|
||||||
|
fakeClient := &fake.Clientset{}
|
||||||
|
fakeClient.Fake.AddReactor("list", "pods", func(action core.Action) (bool, runtime.Object, error) {
|
||||||
|
return true, &v1.PodList{Items: tc.pods}, nil
|
||||||
|
})
|
||||||
|
|
||||||
|
podEvictor := evictions.NewPodEvictor(
|
||||||
|
fakeClient,
|
||||||
|
policyv1.SchemeGroupVersion.String(),
|
||||||
|
false,
|
||||||
|
100,
|
||||||
|
tc.nodes,
|
||||||
|
false,
|
||||||
|
false,
|
||||||
|
false,
|
||||||
|
)
|
||||||
|
|
||||||
|
RemoveFailedPods(ctx, fakeClient, tc.strategy, tc.nodes, podEvictor)
|
||||||
|
actualEvictedPodCount := podEvictor.TotalEvicted()
|
||||||
|
if actualEvictedPodCount != tc.expectedEvictedPodCount {
|
||||||
|
t.Errorf("Test %#v failed, expected %v pod evictions, but got %v pod evictions\n", tc.description, tc.expectedEvictedPodCount, actualEvictedPodCount)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestValidRemoveFailedPodsParams(t *testing.T) {
|
||||||
|
ctx := context.Background()
|
||||||
|
fakeClient := &fake.Clientset{}
|
||||||
|
testCases := []struct {
|
||||||
|
name string
|
||||||
|
params *api.StrategyParameters
|
||||||
|
}{
|
||||||
|
{name: "validate nil params", params: nil},
|
||||||
|
{name: "validate reasons params", params: &api.StrategyParameters{FailedPods: &api.FailedPods{
|
||||||
|
Reasons: []string{"CreateContainerConfigError"},
|
||||||
|
}}},
|
||||||
|
{name: "validate includingInitContainers params", params: &api.StrategyParameters{FailedPods: &api.FailedPods{
|
||||||
|
IncludingInitContainers: true,
|
||||||
|
}}},
|
||||||
|
{name: "validate excludeOwnerKinds params", params: &api.StrategyParameters{FailedPods: &api.FailedPods{
|
||||||
|
ExcludeOwnerKinds: []string{"Job"},
|
||||||
|
}}},
|
||||||
|
{name: "validate excludeOwnerKinds params", params: &api.StrategyParameters{FailedPods: &api.FailedPods{
|
||||||
|
MinPodLifetimeSeconds: &OneHourInSeconds,
|
||||||
|
}}},
|
||||||
|
}
|
||||||
|
for _, tc := range testCases {
|
||||||
|
t.Run(tc.name, func(t *testing.T) {
|
||||||
|
t.Parallel()
|
||||||
|
params, err := validateAndParseRemoveFailedPodsParams(ctx, fakeClient, tc.params)
|
||||||
|
if err != nil {
|
||||||
|
t.Errorf("strategy params should be valid but got err: %v", err.Error())
|
||||||
|
}
|
||||||
|
if params == nil {
|
||||||
|
t.Errorf("strategy params should return a ValidatedFailedPodsStrategyParams but got nil")
|
||||||
|
}
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func buildTestPod(podName, nodeName string, initContainerState, containerState *v1.ContainerState) v1.Pod {
|
||||||
|
pod := test.BuildTestPod(podName, 1, 1, nodeName, func(p *v1.Pod) {
|
||||||
|
ps := v1.PodStatus{}
|
||||||
|
|
||||||
|
if initContainerState != nil {
|
||||||
|
ps.InitContainerStatuses = []v1.ContainerStatus{{State: *initContainerState}}
|
||||||
|
ps.Phase = v1.PodFailed
|
||||||
|
}
|
||||||
|
|
||||||
|
if containerState != nil {
|
||||||
|
ps.ContainerStatuses = []v1.ContainerStatus{{State: *containerState}}
|
||||||
|
ps.Phase = v1.PodFailed
|
||||||
|
}
|
||||||
|
|
||||||
|
p.Status = ps
|
||||||
|
})
|
||||||
|
pod.ObjectMeta.OwnerReferences = test.GetReplicaSetOwnerRefList()
|
||||||
|
pod.ObjectMeta.SetCreationTimestamp(metav1.Now())
|
||||||
|
return *pod
|
||||||
|
}
|
||||||
113
pkg/descheduler/strategies/node_affinity.go
Normal file
113
pkg/descheduler/strategies/node_affinity.go
Normal file
@@ -0,0 +1,113 @@
|
|||||||
|
/*
|
||||||
|
Copyright 2017 The Kubernetes Authors.
|
||||||
|
|
||||||
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
you may not use this file except in compliance with the License.
|
||||||
|
You may obtain a copy of the License at
|
||||||
|
|
||||||
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
|
Unless required by applicable law or agreed to in writing, software
|
||||||
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
See the License for the specific language governing permissions and
|
||||||
|
limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package strategies
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"fmt"
|
||||||
|
|
||||||
|
v1 "k8s.io/api/core/v1"
|
||||||
|
clientset "k8s.io/client-go/kubernetes"
|
||||||
|
"k8s.io/klog/v2"
|
||||||
|
|
||||||
|
"sigs.k8s.io/descheduler/pkg/api"
|
||||||
|
"sigs.k8s.io/descheduler/pkg/descheduler/evictions"
|
||||||
|
nodeutil "sigs.k8s.io/descheduler/pkg/descheduler/node"
|
||||||
|
podutil "sigs.k8s.io/descheduler/pkg/descheduler/pod"
|
||||||
|
"sigs.k8s.io/descheduler/pkg/utils"
|
||||||
|
)
|
||||||
|
|
||||||
|
func validatePodsViolatingNodeAffinityParams(params *api.StrategyParameters) error {
|
||||||
|
if params == nil || len(params.NodeAffinityType) == 0 {
|
||||||
|
return fmt.Errorf("NodeAffinityType is empty")
|
||||||
|
}
|
||||||
|
// At most one of include/exclude can be set
|
||||||
|
if params.Namespaces != nil && len(params.Namespaces.Include) > 0 && len(params.Namespaces.Exclude) > 0 {
|
||||||
|
return fmt.Errorf("only one of Include/Exclude namespaces can be set")
|
||||||
|
}
|
||||||
|
if params.ThresholdPriority != nil && params.ThresholdPriorityClassName != "" {
|
||||||
|
return fmt.Errorf("only one of thresholdPriority and thresholdPriorityClassName can be set")
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// RemovePodsViolatingNodeAffinity evicts pods on nodes which violate node affinity
|
||||||
|
func RemovePodsViolatingNodeAffinity(ctx context.Context, client clientset.Interface, strategy api.DeschedulerStrategy, nodes []*v1.Node, podEvictor *evictions.PodEvictor) {
|
||||||
|
if err := validatePodsViolatingNodeAffinityParams(strategy.Params); err != nil {
|
||||||
|
klog.ErrorS(err, "Invalid RemovePodsViolatingNodeAffinity parameters")
|
||||||
|
return
|
||||||
|
}
|
||||||
|
thresholdPriority, err := utils.GetPriorityFromStrategyParams(ctx, client, strategy.Params)
|
||||||
|
if err != nil {
|
||||||
|
klog.ErrorS(err, "Failed to get threshold priority from strategy's params")
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
var includedNamespaces, excludedNamespaces []string
|
||||||
|
if strategy.Params.Namespaces != nil {
|
||||||
|
includedNamespaces = strategy.Params.Namespaces.Include
|
||||||
|
excludedNamespaces = strategy.Params.Namespaces.Exclude
|
||||||
|
}
|
||||||
|
|
||||||
|
nodeFit := false
|
||||||
|
if strategy.Params != nil {
|
||||||
|
nodeFit = strategy.Params.NodeFit
|
||||||
|
}
|
||||||
|
|
||||||
|
evictable := podEvictor.Evictable(evictions.WithPriorityThreshold(thresholdPriority), evictions.WithNodeFit(nodeFit))
|
||||||
|
|
||||||
|
for _, nodeAffinity := range strategy.Params.NodeAffinityType {
|
||||||
|
klog.V(2).InfoS("Executing for nodeAffinityType", "nodeAffinity", nodeAffinity)
|
||||||
|
|
||||||
|
switch nodeAffinity {
|
||||||
|
case "requiredDuringSchedulingIgnoredDuringExecution":
|
||||||
|
for _, node := range nodes {
|
||||||
|
klog.V(1).InfoS("Processing node", "node", klog.KObj(node))
|
||||||
|
|
||||||
|
pods, err := podutil.ListPodsOnANode(
|
||||||
|
ctx,
|
||||||
|
client,
|
||||||
|
node,
|
||||||
|
podutil.WithFilter(func(pod *v1.Pod) bool {
|
||||||
|
return evictable.IsEvictable(pod) &&
|
||||||
|
!nodeutil.PodFitsCurrentNode(pod, node) &&
|
||||||
|
nodeutil.PodFitsAnyNode(pod, nodes)
|
||||||
|
}),
|
||||||
|
podutil.WithNamespaces(includedNamespaces),
|
||||||
|
podutil.WithoutNamespaces(excludedNamespaces),
|
||||||
|
podutil.WithLabelSelector(strategy.Params.LabelSelector),
|
||||||
|
)
|
||||||
|
if err != nil {
|
||||||
|
klog.ErrorS(err, "Failed to get pods", "node", klog.KObj(node))
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, pod := range pods {
|
||||||
|
if pod.Spec.Affinity != nil && pod.Spec.Affinity.NodeAffinity != nil && pod.Spec.Affinity.NodeAffinity.RequiredDuringSchedulingIgnoredDuringExecution != nil {
|
||||||
|
klog.V(1).InfoS("Evicting pod", "pod", klog.KObj(pod))
|
||||||
|
if _, err := podEvictor.EvictPod(ctx, pod, node, "NodeAffinity"); err != nil {
|
||||||
|
klog.ErrorS(err, "Error evicting pod")
|
||||||
|
break
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
default:
|
||||||
|
klog.ErrorS(nil, "Invalid nodeAffinityType", "nodeAffinity", nodeAffinity)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
190
pkg/descheduler/strategies/node_affinity_test.go
Normal file
190
pkg/descheduler/strategies/node_affinity_test.go
Normal file
@@ -0,0 +1,190 @@
|
|||||||
|
/*
|
||||||
|
Copyright 2017 The Kubernetes Authors.
|
||||||
|
|
||||||
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
you may not use this file except in compliance with the License.
|
||||||
|
You may obtain a copy of the License at
|
||||||
|
|
||||||
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
|
Unless required by applicable law or agreed to in writing, software
|
||||||
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
See the License for the specific language governing permissions and
|
||||||
|
limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package strategies
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"testing"
|
||||||
|
|
||||||
|
v1 "k8s.io/api/core/v1"
|
||||||
|
policyv1 "k8s.io/api/policy/v1"
|
||||||
|
"k8s.io/apimachinery/pkg/runtime"
|
||||||
|
"k8s.io/client-go/kubernetes/fake"
|
||||||
|
core "k8s.io/client-go/testing"
|
||||||
|
"sigs.k8s.io/descheduler/pkg/api"
|
||||||
|
"sigs.k8s.io/descheduler/pkg/descheduler/evictions"
|
||||||
|
"sigs.k8s.io/descheduler/test"
|
||||||
|
)
|
||||||
|
|
||||||
|
func TestRemovePodsViolatingNodeAffinity(t *testing.T) {
|
||||||
|
ctx := context.Background()
|
||||||
|
requiredDuringSchedulingIgnoredDuringExecutionStrategy := api.DeschedulerStrategy{
|
||||||
|
Enabled: true,
|
||||||
|
Params: &api.StrategyParameters{
|
||||||
|
NodeAffinityType: []string{
|
||||||
|
"requiredDuringSchedulingIgnoredDuringExecution",
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
requiredDuringSchedulingIgnoredDuringExecutionWithNodeFitStrategy := api.DeschedulerStrategy{
|
||||||
|
Enabled: true,
|
||||||
|
Params: &api.StrategyParameters{
|
||||||
|
NodeAffinityType: []string{
|
||||||
|
"requiredDuringSchedulingIgnoredDuringExecution",
|
||||||
|
},
|
||||||
|
NodeFit: true,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
nodeLabelKey := "kubernetes.io/desiredNode"
|
||||||
|
nodeLabelValue := "yes"
|
||||||
|
nodeWithLabels := test.BuildTestNode("nodeWithLabels", 2000, 3000, 10, nil)
|
||||||
|
nodeWithLabels.Labels[nodeLabelKey] = nodeLabelValue
|
||||||
|
|
||||||
|
nodeWithoutLabels := test.BuildTestNode("nodeWithoutLabels", 2000, 3000, 10, nil)
|
||||||
|
|
||||||
|
unschedulableNodeWithLabels := test.BuildTestNode("unschedulableNodeWithLabels", 2000, 3000, 10, nil)
|
||||||
|
nodeWithLabels.Labels[nodeLabelKey] = nodeLabelValue
|
||||||
|
unschedulableNodeWithLabels.Spec.Unschedulable = true
|
||||||
|
|
||||||
|
addPodsToNode := func(node *v1.Node) []v1.Pod {
|
||||||
|
podWithNodeAffinity := test.BuildTestPod("podWithNodeAffinity", 100, 0, node.Name, nil)
|
||||||
|
podWithNodeAffinity.Spec.Affinity = &v1.Affinity{
|
||||||
|
NodeAffinity: &v1.NodeAffinity{
|
||||||
|
RequiredDuringSchedulingIgnoredDuringExecution: &v1.NodeSelector{
|
||||||
|
NodeSelectorTerms: []v1.NodeSelectorTerm{
|
||||||
|
{
|
||||||
|
MatchExpressions: []v1.NodeSelectorRequirement{
|
||||||
|
{
|
||||||
|
Key: nodeLabelKey,
|
||||||
|
Operator: "In",
|
||||||
|
Values: []string{
|
||||||
|
nodeLabelValue,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
pod1 := test.BuildTestPod("pod1", 100, 0, node.Name, nil)
|
||||||
|
pod2 := test.BuildTestPod("pod2", 100, 0, node.Name, nil)
|
||||||
|
|
||||||
|
podWithNodeAffinity.ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
||||||
|
pod1.ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
||||||
|
pod2.ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
||||||
|
|
||||||
|
return []v1.Pod{
|
||||||
|
*podWithNodeAffinity,
|
||||||
|
*pod1,
|
||||||
|
*pod2,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
tests := []struct {
|
||||||
|
description string
|
||||||
|
nodes []*v1.Node
|
||||||
|
pods []v1.Pod
|
||||||
|
strategy api.DeschedulerStrategy
|
||||||
|
expectedEvictedPodCount int
|
||||||
|
maxPodsToEvictPerNode int
|
||||||
|
}{
|
||||||
|
{
|
||||||
|
description: "Invalid strategy type, should not evict any pods",
|
||||||
|
strategy: api.DeschedulerStrategy{
|
||||||
|
Enabled: true,
|
||||||
|
Params: &api.StrategyParameters{
|
||||||
|
NodeAffinityType: []string{
|
||||||
|
"requiredDuringSchedulingRequiredDuringExecution",
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
expectedEvictedPodCount: 0,
|
||||||
|
pods: addPodsToNode(nodeWithoutLabels),
|
||||||
|
nodes: []*v1.Node{nodeWithoutLabels, nodeWithLabels},
|
||||||
|
maxPodsToEvictPerNode: 0,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
description: "Pod is correctly scheduled on node, no eviction expected",
|
||||||
|
strategy: requiredDuringSchedulingIgnoredDuringExecutionStrategy,
|
||||||
|
expectedEvictedPodCount: 0,
|
||||||
|
pods: addPodsToNode(nodeWithLabels),
|
||||||
|
nodes: []*v1.Node{nodeWithLabels},
|
||||||
|
maxPodsToEvictPerNode: 0,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
description: "Pod is scheduled on node without matching labels, another schedulable node available, should be evicted",
|
||||||
|
expectedEvictedPodCount: 1,
|
||||||
|
strategy: requiredDuringSchedulingIgnoredDuringExecutionStrategy,
|
||||||
|
pods: addPodsToNode(nodeWithoutLabels),
|
||||||
|
nodes: []*v1.Node{nodeWithoutLabels, nodeWithLabels},
|
||||||
|
maxPodsToEvictPerNode: 0,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
description: "Pod is scheduled on node without matching labels, another schedulable node available, maxPodsToEvictPerNode set to 1, should not be evicted",
|
||||||
|
expectedEvictedPodCount: 1,
|
||||||
|
strategy: requiredDuringSchedulingIgnoredDuringExecutionStrategy,
|
||||||
|
pods: addPodsToNode(nodeWithoutLabels),
|
||||||
|
nodes: []*v1.Node{nodeWithoutLabels, nodeWithLabels},
|
||||||
|
maxPodsToEvictPerNode: 1,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
description: "Pod is scheduled on node without matching labels, but no node where pod fits is available, should not evict",
|
||||||
|
expectedEvictedPodCount: 0,
|
||||||
|
strategy: requiredDuringSchedulingIgnoredDuringExecutionWithNodeFitStrategy,
|
||||||
|
pods: addPodsToNode(nodeWithoutLabels),
|
||||||
|
nodes: []*v1.Node{nodeWithoutLabels, unschedulableNodeWithLabels},
|
||||||
|
maxPodsToEvictPerNode: 0,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
description: "Pod is scheduled on node without matching labels, and node where pod fits is available, should evict",
|
||||||
|
expectedEvictedPodCount: 0,
|
||||||
|
strategy: requiredDuringSchedulingIgnoredDuringExecutionWithNodeFitStrategy,
|
||||||
|
pods: addPodsToNode(nodeWithoutLabels),
|
||||||
|
nodes: []*v1.Node{nodeWithLabels, unschedulableNodeWithLabels},
|
||||||
|
maxPodsToEvictPerNode: 1,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, tc := range tests {
|
||||||
|
|
||||||
|
fakeClient := &fake.Clientset{}
|
||||||
|
fakeClient.Fake.AddReactor("list", "pods", func(action core.Action) (bool, runtime.Object, error) {
|
||||||
|
return true, &v1.PodList{Items: tc.pods}, nil
|
||||||
|
})
|
||||||
|
|
||||||
|
podEvictor := evictions.NewPodEvictor(
|
||||||
|
fakeClient,
|
||||||
|
policyv1.SchemeGroupVersion.String(),
|
||||||
|
false,
|
||||||
|
tc.maxPodsToEvictPerNode,
|
||||||
|
tc.nodes,
|
||||||
|
false,
|
||||||
|
false,
|
||||||
|
false,
|
||||||
|
)
|
||||||
|
|
||||||
|
RemovePodsViolatingNodeAffinity(ctx, fakeClient, tc.strategy, tc.nodes, podEvictor)
|
||||||
|
actualEvictedPodCount := podEvictor.TotalEvicted()
|
||||||
|
if actualEvictedPodCount != tc.expectedEvictedPodCount {
|
||||||
|
t.Errorf("Test %#v failed, expected %v pod evictions, but got %v pod evictions\n", tc.description, tc.expectedEvictedPodCount, actualEvictedPodCount)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
110
pkg/descheduler/strategies/node_taint.go
Normal file
110
pkg/descheduler/strategies/node_taint.go
Normal file
@@ -0,0 +1,110 @@
|
|||||||
|
/*
|
||||||
|
Copyright 2017 The Kubernetes Authors.
|
||||||
|
|
||||||
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
you may not use this file except in compliance with the License.
|
||||||
|
You may obtain a copy of the License at
|
||||||
|
|
||||||
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
|
Unless required by applicable law or agreed to in writing, software
|
||||||
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
See the License for the specific language governing permissions and
|
||||||
|
limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package strategies
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"fmt"
|
||||||
|
|
||||||
|
"sigs.k8s.io/descheduler/pkg/api"
|
||||||
|
"sigs.k8s.io/descheduler/pkg/descheduler/evictions"
|
||||||
|
podutil "sigs.k8s.io/descheduler/pkg/descheduler/pod"
|
||||||
|
"sigs.k8s.io/descheduler/pkg/utils"
|
||||||
|
|
||||||
|
v1 "k8s.io/api/core/v1"
|
||||||
|
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
||||||
|
clientset "k8s.io/client-go/kubernetes"
|
||||||
|
"k8s.io/klog/v2"
|
||||||
|
)
|
||||||
|
|
||||||
|
func validateRemovePodsViolatingNodeTaintsParams(params *api.StrategyParameters) error {
|
||||||
|
if params == nil {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// At most one of include/exclude can be set
|
||||||
|
if params.Namespaces != nil && len(params.Namespaces.Include) > 0 && len(params.Namespaces.Exclude) > 0 {
|
||||||
|
return fmt.Errorf("only one of Include/Exclude namespaces can be set")
|
||||||
|
}
|
||||||
|
if params.ThresholdPriority != nil && params.ThresholdPriorityClassName != "" {
|
||||||
|
return fmt.Errorf("only one of thresholdPriority and thresholdPriorityClassName can be set")
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// RemovePodsViolatingNodeTaints evicts pods on the node which violate NoSchedule Taints on nodes
|
||||||
|
func RemovePodsViolatingNodeTaints(ctx context.Context, client clientset.Interface, strategy api.DeschedulerStrategy, nodes []*v1.Node, podEvictor *evictions.PodEvictor) {
|
||||||
|
if err := validateRemovePodsViolatingNodeTaintsParams(strategy.Params); err != nil {
|
||||||
|
klog.ErrorS(err, "Invalid RemovePodsViolatingNodeTaints parameters")
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
var includedNamespaces, excludedNamespaces []string
|
||||||
|
var labelSelector *metav1.LabelSelector
|
||||||
|
if strategy.Params != nil {
|
||||||
|
if strategy.Params.Namespaces != nil {
|
||||||
|
includedNamespaces = strategy.Params.Namespaces.Include
|
||||||
|
excludedNamespaces = strategy.Params.Namespaces.Exclude
|
||||||
|
}
|
||||||
|
labelSelector = strategy.Params.LabelSelector
|
||||||
|
}
|
||||||
|
|
||||||
|
thresholdPriority, err := utils.GetPriorityFromStrategyParams(ctx, client, strategy.Params)
|
||||||
|
if err != nil {
|
||||||
|
klog.ErrorS(err, "Failed to get threshold priority from strategy's params")
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
nodeFit := false
|
||||||
|
if strategy.Params != nil {
|
||||||
|
nodeFit = strategy.Params.NodeFit
|
||||||
|
}
|
||||||
|
|
||||||
|
evictable := podEvictor.Evictable(evictions.WithPriorityThreshold(thresholdPriority), evictions.WithNodeFit(nodeFit))
|
||||||
|
|
||||||
|
for _, node := range nodes {
|
||||||
|
klog.V(1).InfoS("Processing node", "node", klog.KObj(node))
|
||||||
|
pods, err := podutil.ListPodsOnANode(
|
||||||
|
ctx,
|
||||||
|
client,
|
||||||
|
node,
|
||||||
|
podutil.WithFilter(evictable.IsEvictable),
|
||||||
|
podutil.WithNamespaces(includedNamespaces),
|
||||||
|
podutil.WithoutNamespaces(excludedNamespaces),
|
||||||
|
podutil.WithLabelSelector(labelSelector),
|
||||||
|
)
|
||||||
|
if err != nil {
|
||||||
|
//no pods evicted as error encountered retrieving evictable Pods
|
||||||
|
return
|
||||||
|
}
|
||||||
|
totalPods := len(pods)
|
||||||
|
for i := 0; i < totalPods; i++ {
|
||||||
|
if !utils.TolerationsTolerateTaintsWithFilter(
|
||||||
|
pods[i].Spec.Tolerations,
|
||||||
|
node.Spec.Taints,
|
||||||
|
func(taint *v1.Taint) bool { return taint.Effect == v1.TaintEffectNoSchedule },
|
||||||
|
) {
|
||||||
|
klog.V(2).InfoS("Not all taints with NoSchedule effect are tolerated after update for pod on node", "pod", klog.KObj(pods[i]), "node", klog.KObj(node))
|
||||||
|
if _, err := podEvictor.EvictPod(ctx, pods[i], node, "NodeTaint"); err != nil {
|
||||||
|
klog.ErrorS(err, "Error evicting pod")
|
||||||
|
break
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
@@ -1,20 +1,4 @@
|
|||||||
/*
|
package strategies
|
||||||
Copyright 2022 The Kubernetes Authors.
|
|
||||||
|
|
||||||
Licensed under the Apache License, Version 2.0 (the "License");
|
|
||||||
you may not use this file except in compliance with the License.
|
|
||||||
You may obtain a copy of the License at
|
|
||||||
|
|
||||||
http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
|
|
||||||
Unless required by applicable law or agreed to in writing, software
|
|
||||||
distributed under the License is distributed on an "AS IS" BASIS,
|
|
||||||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
||||||
See the License for the specific language governing permissions and
|
|
||||||
limitations under the License.
|
|
||||||
*/
|
|
||||||
|
|
||||||
package removepodsviolatingnodetaints
|
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
@@ -25,16 +9,10 @@ import (
|
|||||||
policyv1 "k8s.io/api/policy/v1"
|
policyv1 "k8s.io/api/policy/v1"
|
||||||
"k8s.io/apimachinery/pkg/api/resource"
|
"k8s.io/apimachinery/pkg/api/resource"
|
||||||
"k8s.io/apimachinery/pkg/runtime"
|
"k8s.io/apimachinery/pkg/runtime"
|
||||||
"k8s.io/client-go/informers"
|
|
||||||
"k8s.io/client-go/kubernetes/fake"
|
"k8s.io/client-go/kubernetes/fake"
|
||||||
"k8s.io/client-go/tools/events"
|
core "k8s.io/client-go/testing"
|
||||||
|
"sigs.k8s.io/descheduler/pkg/api"
|
||||||
"sigs.k8s.io/descheduler/pkg/apis/componentconfig"
|
|
||||||
"sigs.k8s.io/descheduler/pkg/descheduler/evictions"
|
"sigs.k8s.io/descheduler/pkg/descheduler/evictions"
|
||||||
podutil "sigs.k8s.io/descheduler/pkg/descheduler/pod"
|
|
||||||
"sigs.k8s.io/descheduler/pkg/framework"
|
|
||||||
frameworkfake "sigs.k8s.io/descheduler/pkg/framework/fake"
|
|
||||||
"sigs.k8s.io/descheduler/pkg/framework/plugins/defaultevictor"
|
|
||||||
"sigs.k8s.io/descheduler/pkg/utils"
|
"sigs.k8s.io/descheduler/pkg/utils"
|
||||||
"sigs.k8s.io/descheduler/test"
|
"sigs.k8s.io/descheduler/test"
|
||||||
)
|
)
|
||||||
@@ -47,14 +25,6 @@ func createNoScheduleTaint(key, value string, index int) v1.Taint {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func createPreferNoScheduleTaint(key, value string, index int) v1.Taint {
|
|
||||||
return v1.Taint{
|
|
||||||
Key: "testTaint" + fmt.Sprintf("%v", index),
|
|
||||||
Value: "test" + fmt.Sprintf("%v", index),
|
|
||||||
Effect: v1.TaintEffectPreferNoSchedule,
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func addTaintsToNode(node *v1.Node, key, value string, indices []int) *v1.Node {
|
func addTaintsToNode(node *v1.Node, key, value string, indices []int) *v1.Node {
|
||||||
taints := []v1.Taint{}
|
taints := []v1.Taint{}
|
||||||
for _, index := range indices {
|
for _, index := range indices {
|
||||||
@@ -64,21 +34,22 @@ func addTaintsToNode(node *v1.Node, key, value string, indices []int) *v1.Node {
|
|||||||
return node
|
return node
|
||||||
}
|
}
|
||||||
|
|
||||||
func addTolerationToPod(pod *v1.Pod, key, value string, index int, effect v1.TaintEffect) *v1.Pod {
|
func addTolerationToPod(pod *v1.Pod, key, value string, index int) *v1.Pod {
|
||||||
if pod.Annotations == nil {
|
if pod.Annotations == nil {
|
||||||
pod.Annotations = map[string]string{}
|
pod.Annotations = map[string]string{}
|
||||||
}
|
}
|
||||||
|
|
||||||
pod.Spec.Tolerations = []v1.Toleration{{Key: key + fmt.Sprintf("%v", index), Value: value + fmt.Sprintf("%v", index), Effect: effect}}
|
pod.Spec.Tolerations = []v1.Toleration{{Key: key + fmt.Sprintf("%v", index), Value: value + fmt.Sprintf("%v", index), Effect: v1.TaintEffectNoSchedule}}
|
||||||
|
|
||||||
return pod
|
return pod
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestDeletePodsViolatingNodeTaints(t *testing.T) {
|
func TestDeletePodsViolatingNodeTaints(t *testing.T) {
|
||||||
|
ctx := context.Background()
|
||||||
node1 := test.BuildTestNode("n1", 2000, 3000, 10, nil)
|
node1 := test.BuildTestNode("n1", 2000, 3000, 10, nil)
|
||||||
node1 = addTaintsToNode(node1, "testTaint", "test", []int{1})
|
node1 = addTaintsToNode(node1, "testTaint", "test", []int{1})
|
||||||
node2 := test.BuildTestNode("n2", 2000, 3000, 10, nil)
|
node2 := test.BuildTestNode("n2", 2000, 3000, 10, nil)
|
||||||
node2 = addTaintsToNode(node2, "testingTaint", "testing", []int{1})
|
node1 = addTaintsToNode(node2, "testingTaint", "testing", []int{1})
|
||||||
|
|
||||||
node3 := test.BuildTestNode("n3", 2000, 3000, 10, func(node *v1.Node) {
|
node3 := test.BuildTestNode("n3", 2000, 3000, 10, func(node *v1.Node) {
|
||||||
node.ObjectMeta.Labels = map[string]string{
|
node.ObjectMeta.Labels = map[string]string{
|
||||||
@@ -91,16 +62,6 @@ func TestDeletePodsViolatingNodeTaints(t *testing.T) {
|
|||||||
}
|
}
|
||||||
})
|
})
|
||||||
|
|
||||||
node5 := test.BuildTestNode("n5", 2000, 3000, 10, nil)
|
|
||||||
node5.Spec.Taints = []v1.Taint{
|
|
||||||
createPreferNoScheduleTaint("testTaint", "test", 1),
|
|
||||||
}
|
|
||||||
|
|
||||||
node6 := test.BuildTestNode("n6", 1, 1, 1, nil)
|
|
||||||
node6.Spec.Taints = []v1.Taint{
|
|
||||||
createPreferNoScheduleTaint("testTaint", "test", 1),
|
|
||||||
}
|
|
||||||
|
|
||||||
p1 := test.BuildTestPod("p1", 100, 0, node1.Name, nil)
|
p1 := test.BuildTestPod("p1", 100, 0, node1.Name, nil)
|
||||||
p2 := test.BuildTestPod("p2", 100, 0, node1.Name, nil)
|
p2 := test.BuildTestPod("p2", 100, 0, node1.Name, nil)
|
||||||
p3 := test.BuildTestPod("p3", 100, 0, node1.Name, nil)
|
p3 := test.BuildTestPod("p3", 100, 0, node1.Name, nil)
|
||||||
@@ -147,269 +108,152 @@ func TestDeletePodsViolatingNodeTaints(t *testing.T) {
|
|||||||
// A Mirror Pod.
|
// A Mirror Pod.
|
||||||
p10.Annotations = test.GetMirrorPodAnnotation()
|
p10.Annotations = test.GetMirrorPodAnnotation()
|
||||||
|
|
||||||
p1 = addTolerationToPod(p1, "testTaint", "test", 1, v1.TaintEffectNoSchedule)
|
p1 = addTolerationToPod(p1, "testTaint", "test", 1)
|
||||||
p3 = addTolerationToPod(p3, "testTaint", "test", 1, v1.TaintEffectNoSchedule)
|
p3 = addTolerationToPod(p3, "testTaint", "test", 1)
|
||||||
p4 = addTolerationToPod(p4, "testTaintX", "testX", 1, v1.TaintEffectNoSchedule)
|
p4 = addTolerationToPod(p4, "testTaintX", "testX", 1)
|
||||||
|
|
||||||
p12.Spec.NodeSelector = map[string]string{
|
p12.Spec.NodeSelector = map[string]string{
|
||||||
"datacenter": "west",
|
"datacenter": "west",
|
||||||
}
|
}
|
||||||
|
|
||||||
p13 := test.BuildTestPod("p13", 100, 0, node5.Name, nil)
|
|
||||||
p13.ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
|
||||||
// node5 has PreferNoSchedule:testTaint1=test1, so the p13 has to have
|
|
||||||
// PreferNoSchedule:testTaint0=test0 so the pod is not tolarated
|
|
||||||
p13 = addTolerationToPod(p13, "testTaint", "test", 0, v1.TaintEffectPreferNoSchedule)
|
|
||||||
|
|
||||||
var uint1 uint = 1
|
|
||||||
|
|
||||||
tests := []struct {
|
tests := []struct {
|
||||||
description string
|
description string
|
||||||
nodes []*v1.Node
|
nodes []*v1.Node
|
||||||
pods []*v1.Pod
|
pods []v1.Pod
|
||||||
evictLocalStoragePods bool
|
evictLocalStoragePods bool
|
||||||
evictSystemCriticalPods bool
|
evictSystemCriticalPods bool
|
||||||
maxPodsToEvictPerNode *uint
|
maxPodsToEvictPerNode int
|
||||||
maxNoOfPodsToEvictPerNamespace *uint
|
expectedEvictedPodCount int
|
||||||
expectedEvictedPodCount uint
|
nodeFit bool
|
||||||
nodeFit bool
|
|
||||||
includePreferNoSchedule bool
|
|
||||||
excludedTaints []string
|
|
||||||
}{
|
}{
|
||||||
|
|
||||||
{
|
{
|
||||||
description: "Pods not tolerating node taint should be evicted",
|
description: "Pods not tolerating node taint should be evicted",
|
||||||
pods: []*v1.Pod{p1, p2, p3},
|
pods: []v1.Pod{*p1, *p2, *p3},
|
||||||
nodes: []*v1.Node{node1},
|
nodes: []*v1.Node{node1},
|
||||||
evictLocalStoragePods: false,
|
evictLocalStoragePods: false,
|
||||||
evictSystemCriticalPods: false,
|
evictSystemCriticalPods: false,
|
||||||
|
maxPodsToEvictPerNode: 0,
|
||||||
expectedEvictedPodCount: 1, //p2 gets evicted
|
expectedEvictedPodCount: 1, //p2 gets evicted
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
description: "Pods with tolerations but not tolerating node taint should be evicted",
|
description: "Pods with tolerations but not tolerating node taint should be evicted",
|
||||||
pods: []*v1.Pod{p1, p3, p4},
|
pods: []v1.Pod{*p1, *p3, *p4},
|
||||||
nodes: []*v1.Node{node1},
|
nodes: []*v1.Node{node1},
|
||||||
evictLocalStoragePods: false,
|
evictLocalStoragePods: false,
|
||||||
evictSystemCriticalPods: false,
|
evictSystemCriticalPods: false,
|
||||||
|
maxPodsToEvictPerNode: 0,
|
||||||
expectedEvictedPodCount: 1, //p4 gets evicted
|
expectedEvictedPodCount: 1, //p4 gets evicted
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
description: "Only <maxPodsToEvictPerNode> number of Pods not tolerating node taint should be evicted",
|
description: "Only <maxPodsToEvictPerNode> number of Pods not tolerating node taint should be evicted",
|
||||||
pods: []*v1.Pod{p1, p5, p6},
|
pods: []v1.Pod{*p1, *p5, *p6},
|
||||||
nodes: []*v1.Node{node1},
|
nodes: []*v1.Node{node1},
|
||||||
evictLocalStoragePods: false,
|
evictLocalStoragePods: false,
|
||||||
evictSystemCriticalPods: false,
|
evictSystemCriticalPods: false,
|
||||||
maxPodsToEvictPerNode: &uint1,
|
maxPodsToEvictPerNode: 1,
|
||||||
expectedEvictedPodCount: 1, //p5 or p6 gets evicted
|
expectedEvictedPodCount: 1, //p5 or p6 gets evicted
|
||||||
},
|
},
|
||||||
{
|
|
||||||
description: "Only <maxNoOfPodsToEvictPerNamespace> number of Pods not tolerating node taint should be evicted",
|
|
||||||
pods: []*v1.Pod{p1, p5, p6},
|
|
||||||
nodes: []*v1.Node{node1},
|
|
||||||
evictLocalStoragePods: false,
|
|
||||||
evictSystemCriticalPods: false,
|
|
||||||
maxNoOfPodsToEvictPerNamespace: &uint1,
|
|
||||||
expectedEvictedPodCount: 1, //p5 or p6 gets evicted
|
|
||||||
},
|
|
||||||
{
|
{
|
||||||
description: "Critical pods not tolerating node taint should not be evicted",
|
description: "Critical pods not tolerating node taint should not be evicted",
|
||||||
pods: []*v1.Pod{p7, p8, p9, p10},
|
pods: []v1.Pod{*p7, *p8, *p9, *p10},
|
||||||
nodes: []*v1.Node{node2},
|
nodes: []*v1.Node{node2},
|
||||||
evictLocalStoragePods: false,
|
evictLocalStoragePods: false,
|
||||||
evictSystemCriticalPods: false,
|
evictSystemCriticalPods: false,
|
||||||
|
maxPodsToEvictPerNode: 0,
|
||||||
expectedEvictedPodCount: 0, //nothing is evicted
|
expectedEvictedPodCount: 0, //nothing is evicted
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
description: "Critical pods except storage pods not tolerating node taint should not be evicted",
|
description: "Critical pods except storage pods not tolerating node taint should not be evicted",
|
||||||
pods: []*v1.Pod{p7, p8, p9, p10},
|
pods: []v1.Pod{*p7, *p8, *p9, *p10},
|
||||||
nodes: []*v1.Node{node2},
|
nodes: []*v1.Node{node2},
|
||||||
evictLocalStoragePods: true,
|
evictLocalStoragePods: true,
|
||||||
evictSystemCriticalPods: false,
|
evictSystemCriticalPods: false,
|
||||||
|
maxPodsToEvictPerNode: 0,
|
||||||
expectedEvictedPodCount: 1, //p9 gets evicted
|
expectedEvictedPodCount: 1, //p9 gets evicted
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
description: "Critical and non critical pods, only non critical pods not tolerating node taint should be evicted",
|
description: "Critical and non critical pods, only non critical pods not tolerating node taint should be evicted",
|
||||||
pods: []*v1.Pod{p7, p8, p10, p11},
|
pods: []v1.Pod{*p7, *p8, *p10, *p11},
|
||||||
nodes: []*v1.Node{node2},
|
nodes: []*v1.Node{node2},
|
||||||
evictLocalStoragePods: false,
|
evictLocalStoragePods: false,
|
||||||
evictSystemCriticalPods: false,
|
evictSystemCriticalPods: false,
|
||||||
|
maxPodsToEvictPerNode: 0,
|
||||||
expectedEvictedPodCount: 1, //p11 gets evicted
|
expectedEvictedPodCount: 1, //p11 gets evicted
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
description: "Critical and non critical pods, pods not tolerating node taint should be evicted even if they are critical",
|
description: "Critical and non critical pods, pods not tolerating node taint should be evicted even if they are critical",
|
||||||
pods: []*v1.Pod{p2, p7, p9, p10},
|
pods: []v1.Pod{*p2, *p7, *p9, *p10},
|
||||||
nodes: []*v1.Node{node1, node2},
|
nodes: []*v1.Node{node2},
|
||||||
evictLocalStoragePods: false,
|
evictLocalStoragePods: false,
|
||||||
evictSystemCriticalPods: true,
|
evictSystemCriticalPods: true,
|
||||||
|
maxPodsToEvictPerNode: 0,
|
||||||
expectedEvictedPodCount: 2, //p2 and p7 are evicted
|
expectedEvictedPodCount: 2, //p2 and p7 are evicted
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
description: "Pod p2 doesn't tolerate taint on it's node, but also doesn't tolerate taints on other nodes",
|
description: "Pod p2 doesn't tolerate taint on it's node, but also doesn't tolerate taints on other nodes",
|
||||||
pods: []*v1.Pod{p1, p2, p3},
|
pods: []v1.Pod{*p1, *p2, *p3},
|
||||||
nodes: []*v1.Node{node1, node2},
|
nodes: []*v1.Node{node1, node2},
|
||||||
evictLocalStoragePods: false,
|
evictLocalStoragePods: false,
|
||||||
evictSystemCriticalPods: false,
|
evictSystemCriticalPods: false,
|
||||||
|
maxPodsToEvictPerNode: 0,
|
||||||
expectedEvictedPodCount: 0, //p2 gets evicted
|
expectedEvictedPodCount: 0, //p2 gets evicted
|
||||||
nodeFit: true,
|
nodeFit: true,
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
description: "Pod p12 doesn't tolerate taint on it's node, but other nodes don't match it's selector",
|
description: "Pod p12 doesn't tolerate taint on it's node, but other nodes don't match it's selector",
|
||||||
pods: []*v1.Pod{p1, p3, p12},
|
pods: []v1.Pod{*p1, *p3, *p12},
|
||||||
nodes: []*v1.Node{node1, node3},
|
nodes: []*v1.Node{node1, node3},
|
||||||
evictLocalStoragePods: false,
|
evictLocalStoragePods: false,
|
||||||
evictSystemCriticalPods: false,
|
evictSystemCriticalPods: false,
|
||||||
|
maxPodsToEvictPerNode: 0,
|
||||||
expectedEvictedPodCount: 0, //p2 gets evicted
|
expectedEvictedPodCount: 0, //p2 gets evicted
|
||||||
nodeFit: true,
|
nodeFit: true,
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
description: "Pod p2 doesn't tolerate taint on it's node, but other nodes are unschedulable",
|
description: "Pod p2 doesn't tolerate taint on it's node, but other nodes are unschedulable",
|
||||||
pods: []*v1.Pod{p1, p2, p3},
|
pods: []v1.Pod{*p1, *p2, *p3},
|
||||||
nodes: []*v1.Node{node1, node4},
|
nodes: []*v1.Node{node1, node4},
|
||||||
evictLocalStoragePods: false,
|
evictLocalStoragePods: false,
|
||||||
evictSystemCriticalPods: false,
|
evictSystemCriticalPods: false,
|
||||||
|
maxPodsToEvictPerNode: 0,
|
||||||
expectedEvictedPodCount: 0, //p2 gets evicted
|
expectedEvictedPodCount: 0, //p2 gets evicted
|
||||||
nodeFit: true,
|
nodeFit: true,
|
||||||
},
|
},
|
||||||
{
|
|
||||||
description: "Pods not tolerating PreferNoSchedule node taint should not be evicted when not enabled",
|
|
||||||
pods: []*v1.Pod{p13},
|
|
||||||
nodes: []*v1.Node{node5},
|
|
||||||
evictLocalStoragePods: false,
|
|
||||||
evictSystemCriticalPods: false,
|
|
||||||
expectedEvictedPodCount: 0,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
description: "Pods not tolerating PreferNoSchedule node taint should be evicted when enabled",
|
|
||||||
pods: []*v1.Pod{p13},
|
|
||||||
nodes: []*v1.Node{node5},
|
|
||||||
evictLocalStoragePods: false,
|
|
||||||
evictSystemCriticalPods: false,
|
|
||||||
includePreferNoSchedule: true,
|
|
||||||
expectedEvictedPodCount: 1, // p13 gets evicted
|
|
||||||
},
|
|
||||||
{
|
|
||||||
description: "Pods not tolerating excluded node taints (by key) should not be evicted",
|
|
||||||
pods: []*v1.Pod{p2},
|
|
||||||
nodes: []*v1.Node{node1},
|
|
||||||
evictLocalStoragePods: false,
|
|
||||||
evictSystemCriticalPods: false,
|
|
||||||
excludedTaints: []string{"excludedTaint1", "testTaint1"},
|
|
||||||
expectedEvictedPodCount: 0, // nothing gets evicted, as one of the specified excludedTaints matches the key of node1's taint
|
|
||||||
},
|
|
||||||
{
|
|
||||||
description: "Pods not tolerating excluded node taints (by key and value) should not be evicted",
|
|
||||||
pods: []*v1.Pod{p2},
|
|
||||||
nodes: []*v1.Node{node1},
|
|
||||||
evictLocalStoragePods: false,
|
|
||||||
evictSystemCriticalPods: false,
|
|
||||||
excludedTaints: []string{"testTaint1=test1"},
|
|
||||||
expectedEvictedPodCount: 0, // nothing gets evicted, as both the key and value of the excluded taint match node1's taint
|
|
||||||
},
|
|
||||||
{
|
|
||||||
description: "The excluded taint matches the key of node1's taint, but does not match the value",
|
|
||||||
pods: []*v1.Pod{p2},
|
|
||||||
nodes: []*v1.Node{node1},
|
|
||||||
evictLocalStoragePods: false,
|
|
||||||
evictSystemCriticalPods: false,
|
|
||||||
excludedTaints: []string{"testTaint1=test2"},
|
|
||||||
expectedEvictedPodCount: 1, // pod gets evicted, as excluded taint value does not match node1's taint value
|
|
||||||
},
|
|
||||||
{
|
|
||||||
description: "Critical and non critical pods, pods not tolerating node taint can't be evicted because the only available node does not have enough resources.",
|
|
||||||
pods: []*v1.Pod{p2, p7, p9, p10},
|
|
||||||
nodes: []*v1.Node{node1, node6},
|
|
||||||
evictLocalStoragePods: false,
|
|
||||||
evictSystemCriticalPods: true,
|
|
||||||
expectedEvictedPodCount: 0, //p2 and p7 can't be evicted
|
|
||||||
nodeFit: true,
|
|
||||||
},
|
|
||||||
}
|
}
|
||||||
|
|
||||||
for _, tc := range tests {
|
for _, tc := range tests {
|
||||||
t.Run(tc.description, func(t *testing.T) {
|
|
||||||
|
|
||||||
ctx, cancel := context.WithCancel(context.Background())
|
// create fake client
|
||||||
defer cancel()
|
fakeClient := &fake.Clientset{}
|
||||||
|
fakeClient.Fake.AddReactor("list", "pods", func(action core.Action) (bool, runtime.Object, error) {
|
||||||
var objs []runtime.Object
|
return true, &v1.PodList{Items: tc.pods}, nil
|
||||||
for _, node := range tc.nodes {
|
|
||||||
objs = append(objs, node)
|
|
||||||
}
|
|
||||||
for _, pod := range tc.pods {
|
|
||||||
objs = append(objs, pod)
|
|
||||||
}
|
|
||||||
fakeClient := fake.NewSimpleClientset(objs...)
|
|
||||||
|
|
||||||
sharedInformerFactory := informers.NewSharedInformerFactory(fakeClient, 0)
|
|
||||||
podInformer := sharedInformerFactory.Core().V1().Pods()
|
|
||||||
|
|
||||||
getPodsAssignedToNode, err := podutil.BuildGetPodsAssignedToNodeFunc(podInformer)
|
|
||||||
if err != nil {
|
|
||||||
t.Errorf("Build get pods assigned to node function error: %v", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
sharedInformerFactory.Start(ctx.Done())
|
|
||||||
sharedInformerFactory.WaitForCacheSync(ctx.Done())
|
|
||||||
|
|
||||||
eventRecorder := &events.FakeRecorder{}
|
|
||||||
|
|
||||||
podEvictor := evictions.NewPodEvictor(
|
|
||||||
fakeClient,
|
|
||||||
policyv1.SchemeGroupVersion.String(),
|
|
||||||
false,
|
|
||||||
tc.maxPodsToEvictPerNode,
|
|
||||||
tc.maxNoOfPodsToEvictPerNamespace,
|
|
||||||
tc.nodes,
|
|
||||||
false,
|
|
||||||
eventRecorder,
|
|
||||||
)
|
|
||||||
|
|
||||||
defaultevictorArgs := &defaultevictor.DefaultEvictorArgs{
|
|
||||||
EvictLocalStoragePods: tc.evictLocalStoragePods,
|
|
||||||
EvictSystemCriticalPods: tc.evictSystemCriticalPods,
|
|
||||||
IgnorePvcPods: false,
|
|
||||||
EvictFailedBarePods: false,
|
|
||||||
NodeFit: tc.nodeFit,
|
|
||||||
}
|
|
||||||
|
|
||||||
evictorFilter, err := defaultevictor.New(
|
|
||||||
defaultevictorArgs,
|
|
||||||
&frameworkfake.HandleImpl{
|
|
||||||
ClientsetImpl: fakeClient,
|
|
||||||
GetPodsAssignedToNodeFuncImpl: getPodsAssignedToNode,
|
|
||||||
SharedInformerFactoryImpl: sharedInformerFactory,
|
|
||||||
},
|
|
||||||
)
|
|
||||||
|
|
||||||
if err != nil {
|
|
||||||
t.Fatalf("Unable to initialize the plugin: %v", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
handle := &frameworkfake.HandleImpl{
|
|
||||||
ClientsetImpl: fakeClient,
|
|
||||||
GetPodsAssignedToNodeFuncImpl: getPodsAssignedToNode,
|
|
||||||
PodEvictorImpl: podEvictor,
|
|
||||||
EvictorFilterImpl: evictorFilter.(framework.EvictorPlugin),
|
|
||||||
SharedInformerFactoryImpl: sharedInformerFactory,
|
|
||||||
}
|
|
||||||
|
|
||||||
plugin, err := New(&componentconfig.RemovePodsViolatingNodeTaintsArgs{
|
|
||||||
IncludePreferNoSchedule: tc.includePreferNoSchedule,
|
|
||||||
ExcludedTaints: tc.excludedTaints,
|
|
||||||
},
|
|
||||||
handle,
|
|
||||||
)
|
|
||||||
if err != nil {
|
|
||||||
t.Fatalf("Unable to initialize the plugin: %v", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
plugin.(framework.DeschedulePlugin).Deschedule(ctx, tc.nodes)
|
|
||||||
actualEvictedPodCount := podEvictor.TotalEvicted()
|
|
||||||
if actualEvictedPodCount != tc.expectedEvictedPodCount {
|
|
||||||
t.Errorf("Test %#v failed, Unexpected no of pods evicted: pods evicted: %d, expected: %d", tc.description, actualEvictedPodCount, tc.expectedEvictedPodCount)
|
|
||||||
}
|
|
||||||
})
|
})
|
||||||
|
|
||||||
|
podEvictor := evictions.NewPodEvictor(
|
||||||
|
fakeClient,
|
||||||
|
policyv1.SchemeGroupVersion.String(),
|
||||||
|
false,
|
||||||
|
tc.maxPodsToEvictPerNode,
|
||||||
|
tc.nodes,
|
||||||
|
tc.evictLocalStoragePods,
|
||||||
|
tc.evictSystemCriticalPods,
|
||||||
|
false,
|
||||||
|
)
|
||||||
|
|
||||||
|
strategy := api.DeschedulerStrategy{
|
||||||
|
Params: &api.StrategyParameters{
|
||||||
|
NodeFit: tc.nodeFit,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
RemovePodsViolatingNodeTaints(ctx, fakeClient, strategy, tc.nodes, podEvictor)
|
||||||
|
actualEvictedPodCount := podEvictor.TotalEvicted()
|
||||||
|
if actualEvictedPodCount != tc.expectedEvictedPodCount {
|
||||||
|
t.Errorf("Test %#v failed, Unexpected no of pods evicted: pods evicted: %d, expected: %d", tc.description, actualEvictedPodCount, tc.expectedEvictedPodCount)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestToleratesTaint(t *testing.T) {
|
func TestToleratesTaint(t *testing.T) {
|
||||||
@@ -1,11 +1,11 @@
|
|||||||
/*
|
/*
|
||||||
Copyright 2022 The Kubernetes Authors.
|
Copyright 2021 The Kubernetes Authors.
|
||||||
|
|
||||||
Licensed under the Apache License, Version 2.0 (the "License");
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
you may not use this file except in compliance with the License.
|
you may not use this file except in compliance with the License.
|
||||||
You may obtain a copy of the License at
|
You may obtain a copy of the License at
|
||||||
|
|
||||||
http://www.apache.org/licenses/LICENSE-2.0
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
Unless required by applicable law or agreed to in writing, software
|
Unless required by applicable law or agreed to in writing, software
|
||||||
distributed under the License is distributed on an "AS IS" BASIS,
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
@@ -22,75 +22,55 @@ import (
|
|||||||
|
|
||||||
v1 "k8s.io/api/core/v1"
|
v1 "k8s.io/api/core/v1"
|
||||||
"k8s.io/apimachinery/pkg/api/resource"
|
"k8s.io/apimachinery/pkg/api/resource"
|
||||||
"k8s.io/apimachinery/pkg/runtime"
|
clientset "k8s.io/client-go/kubernetes"
|
||||||
"k8s.io/klog/v2"
|
"k8s.io/klog/v2"
|
||||||
"sigs.k8s.io/descheduler/pkg/api"
|
"sigs.k8s.io/descheduler/pkg/api"
|
||||||
|
"sigs.k8s.io/descheduler/pkg/descheduler/evictions"
|
||||||
nodeutil "sigs.k8s.io/descheduler/pkg/descheduler/node"
|
nodeutil "sigs.k8s.io/descheduler/pkg/descheduler/node"
|
||||||
|
"sigs.k8s.io/descheduler/pkg/utils"
|
||||||
"sigs.k8s.io/descheduler/pkg/apis/componentconfig"
|
|
||||||
podutil "sigs.k8s.io/descheduler/pkg/descheduler/pod"
|
|
||||||
"sigs.k8s.io/descheduler/pkg/framework"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
const HighNodeUtilizationPluginName = "HighNodeUtilization"
|
// HighNodeUtilization evicts pods from under utilized nodes so that scheduler can schedule according to its strategy.
|
||||||
|
|
||||||
// HighNodeUtilization evicts pods from under utilized nodes so that scheduler can schedule according to its plugin.
|
|
||||||
// Note that CPU/Memory requests are used to calculate nodes' utilization and not the actual resource usage.
|
// Note that CPU/Memory requests are used to calculate nodes' utilization and not the actual resource usage.
|
||||||
|
func HighNodeUtilization(ctx context.Context, client clientset.Interface, strategy api.DeschedulerStrategy, nodes []*v1.Node, podEvictor *evictions.PodEvictor) {
|
||||||
type HighNodeUtilization struct {
|
if err := validateNodeUtilizationParams(strategy.Params); err != nil {
|
||||||
handle framework.Handle
|
klog.ErrorS(err, "Invalid HighNodeUtilization parameters")
|
||||||
args *componentconfig.HighNodeUtilizationArgs
|
return
|
||||||
podFilter func(pod *v1.Pod) bool
|
|
||||||
}
|
|
||||||
|
|
||||||
var _ framework.BalancePlugin = &HighNodeUtilization{}
|
|
||||||
|
|
||||||
// NewHighNodeUtilization builds plugin from its arguments while passing a handle
|
|
||||||
func NewHighNodeUtilization(args runtime.Object, handle framework.Handle) (framework.Plugin, error) {
|
|
||||||
highNodeUtilizatioArgs, ok := args.(*componentconfig.HighNodeUtilizationArgs)
|
|
||||||
if !ok {
|
|
||||||
return nil, fmt.Errorf("want args to be of type HighNodeUtilizationArgs, got %T", args)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
podFilter, err := podutil.NewOptions().
|
nodeFit := false
|
||||||
WithFilter(handle.Evictor().Filter).
|
if strategy.Params != nil {
|
||||||
BuildFilterFunc()
|
nodeFit = strategy.Params.NodeFit
|
||||||
|
}
|
||||||
|
|
||||||
|
thresholdPriority, err := utils.GetPriorityFromStrategyParams(ctx, client, strategy.Params)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, fmt.Errorf("error initializing pod filter function: %v", err)
|
klog.ErrorS(err, "Failed to get threshold priority from strategy's params")
|
||||||
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
return &HighNodeUtilization{
|
thresholds := strategy.Params.NodeResourceUtilizationThresholds.Thresholds
|
||||||
handle: handle,
|
targetThresholds := strategy.Params.NodeResourceUtilizationThresholds.TargetThresholds
|
||||||
args: highNodeUtilizatioArgs,
|
if err := validateHighUtilizationStrategyConfig(thresholds, targetThresholds); err != nil {
|
||||||
podFilter: podFilter,
|
klog.ErrorS(err, "HighNodeUtilization config is not valid")
|
||||||
}, nil
|
return
|
||||||
}
|
}
|
||||||
|
targetThresholds = make(api.ResourceThresholds)
|
||||||
// Name retrieves the plugin name
|
|
||||||
func (h *HighNodeUtilization) Name() string {
|
|
||||||
return HighNodeUtilizationPluginName
|
|
||||||
}
|
|
||||||
|
|
||||||
// Balance extension point implementation for the plugin
|
|
||||||
func (h *HighNodeUtilization) Balance(ctx context.Context, nodes []*v1.Node) *framework.Status {
|
|
||||||
thresholds := h.args.Thresholds
|
|
||||||
targetThresholds := make(api.ResourceThresholds)
|
|
||||||
|
|
||||||
setDefaultForThresholds(thresholds, targetThresholds)
|
setDefaultForThresholds(thresholds, targetThresholds)
|
||||||
resourceNames := getResourceNames(targetThresholds)
|
resourceNames := getResourceNames(targetThresholds)
|
||||||
|
|
||||||
sourceNodes, highNodes := classifyNodes(
|
sourceNodes, highNodes := classifyNodes(
|
||||||
getNodeUsage(nodes, resourceNames, h.handle.GetPodsAssignedToNodeFunc()),
|
getNodeUsage(ctx, client, nodes, thresholds, targetThresholds, resourceNames),
|
||||||
getNodeThresholds(nodes, thresholds, targetThresholds, resourceNames, h.handle.GetPodsAssignedToNodeFunc(), false),
|
func(node *v1.Node, usage NodeUsage) bool {
|
||||||
func(node *v1.Node, usage NodeUsage, threshold NodeThresholds) bool {
|
return isNodeWithLowUtilization(usage)
|
||||||
return isNodeWithLowUtilization(usage, threshold.lowResourceThreshold)
|
|
||||||
},
|
},
|
||||||
func(node *v1.Node, usage NodeUsage, threshold NodeThresholds) bool {
|
func(node *v1.Node, usage NodeUsage) bool {
|
||||||
if nodeutil.IsNodeUnschedulable(node) {
|
if nodeutil.IsNodeUnschedulable(node) {
|
||||||
klog.V(2).InfoS("Node is unschedulable", "node", klog.KObj(node))
|
klog.V(2).InfoS("Node is unschedulable", "node", klog.KObj(node))
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
return !isNodeWithLowUtilization(usage, threshold.lowResourceThreshold)
|
return !isNodeWithLowUtilization(usage)
|
||||||
})
|
})
|
||||||
|
|
||||||
// log message in one line
|
// log message in one line
|
||||||
@@ -100,7 +80,7 @@ func (h *HighNodeUtilization) Balance(ctx context.Context, nodes []*v1.Node) *fr
|
|||||||
"Pods", thresholds[v1.ResourcePods],
|
"Pods", thresholds[v1.ResourcePods],
|
||||||
}
|
}
|
||||||
for name := range thresholds {
|
for name := range thresholds {
|
||||||
if !nodeutil.IsBasicResource(name) {
|
if !isBasicResource(name) {
|
||||||
keysAndValues = append(keysAndValues, string(name), int64(thresholds[name]))
|
keysAndValues = append(keysAndValues, string(name), int64(thresholds[name]))
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -110,23 +90,25 @@ func (h *HighNodeUtilization) Balance(ctx context.Context, nodes []*v1.Node) *fr
|
|||||||
|
|
||||||
if len(sourceNodes) == 0 {
|
if len(sourceNodes) == 0 {
|
||||||
klog.V(1).InfoS("No node is underutilized, nothing to do here, you might tune your thresholds further")
|
klog.V(1).InfoS("No node is underutilized, nothing to do here, you might tune your thresholds further")
|
||||||
return nil
|
return
|
||||||
}
|
}
|
||||||
if len(sourceNodes) <= h.args.NumberOfNodes {
|
if len(sourceNodes) <= strategy.Params.NodeResourceUtilizationThresholds.NumberOfNodes {
|
||||||
klog.V(1).InfoS("Number of nodes underutilized is less or equal than NumberOfNodes, nothing to do here", "underutilizedNodes", len(sourceNodes), "numberOfNodes", h.args.NumberOfNodes)
|
klog.V(1).InfoS("Number of nodes underutilized is less or equal than NumberOfNodes, nothing to do here", "underutilizedNodes", len(sourceNodes), "numberOfNodes", strategy.Params.NodeResourceUtilizationThresholds.NumberOfNodes)
|
||||||
return nil
|
return
|
||||||
}
|
}
|
||||||
if len(sourceNodes) == len(nodes) {
|
if len(sourceNodes) == len(nodes) {
|
||||||
klog.V(1).InfoS("All nodes are underutilized, nothing to do here")
|
klog.V(1).InfoS("All nodes are underutilized, nothing to do here")
|
||||||
return nil
|
return
|
||||||
}
|
}
|
||||||
if len(highNodes) == 0 {
|
if len(highNodes) == 0 {
|
||||||
klog.V(1).InfoS("No node is available to schedule the pods, nothing to do here")
|
klog.V(1).InfoS("No node is available to schedule the pods, nothing to do here")
|
||||||
return nil
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
|
evictable := podEvictor.Evictable(evictions.WithPriorityThreshold(thresholdPriority), evictions.WithNodeFit(nodeFit))
|
||||||
|
|
||||||
// stop if the total available usage has dropped to zero - no more pods can be scheduled
|
// stop if the total available usage has dropped to zero - no more pods can be scheduled
|
||||||
continueEvictionCond := func(nodeInfo NodeInfo, totalAvailableUsage map[v1.ResourceName]*resource.Quantity) bool {
|
continueEvictionCond := func(nodeUsage NodeUsage, totalAvailableUsage map[v1.ResourceName]*resource.Quantity) bool {
|
||||||
for name := range totalAvailableUsage {
|
for name := range totalAvailableUsage {
|
||||||
if totalAvailableUsage[name].CmpInt64(0) < 1 {
|
if totalAvailableUsage[name].CmpInt64(0) < 1 {
|
||||||
return false
|
return false
|
||||||
@@ -135,19 +117,25 @@ func (h *HighNodeUtilization) Balance(ctx context.Context, nodes []*v1.Node) *fr
|
|||||||
|
|
||||||
return true
|
return true
|
||||||
}
|
}
|
||||||
|
|
||||||
// Sort the nodes by the usage in ascending order
|
|
||||||
sortNodesByUsage(sourceNodes, true)
|
|
||||||
|
|
||||||
evictPodsFromSourceNodes(
|
evictPodsFromSourceNodes(
|
||||||
ctx,
|
ctx,
|
||||||
sourceNodes,
|
sourceNodes,
|
||||||
highNodes,
|
highNodes,
|
||||||
h.handle.Evictor(),
|
podEvictor,
|
||||||
h.podFilter,
|
evictable.IsEvictable,
|
||||||
resourceNames,
|
resourceNames,
|
||||||
|
"HighNodeUtilization",
|
||||||
continueEvictionCond)
|
continueEvictionCond)
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
func validateHighUtilizationStrategyConfig(thresholds, targetThresholds api.ResourceThresholds) error {
|
||||||
|
if targetThresholds != nil {
|
||||||
|
return fmt.Errorf("targetThresholds is not applicable for HighNodeUtilization")
|
||||||
|
}
|
||||||
|
if err := validateThresholds(thresholds); err != nil {
|
||||||
|
return fmt.Errorf("thresholds config is not valid: %v", err)
|
||||||
|
}
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -169,7 +157,7 @@ func setDefaultForThresholds(thresholds, targetThresholds api.ResourceThresholds
|
|||||||
targetThresholds[v1.ResourceMemory] = MaxResourcePercentage
|
targetThresholds[v1.ResourceMemory] = MaxResourcePercentage
|
||||||
|
|
||||||
for name := range thresholds {
|
for name := range thresholds {
|
||||||
if !nodeutil.IsBasicResource(name) {
|
if !isBasicResource(name) {
|
||||||
targetThresholds[name] = MaxResourcePercentage
|
targetThresholds[name] = MaxResourcePercentage
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -0,0 +1,784 @@
|
|||||||
|
/*
|
||||||
|
Copyright 2021 The Kubernetes Authors.
|
||||||
|
|
||||||
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
you may not use this file except in compliance with the License.
|
||||||
|
You may obtain a copy of the License at
|
||||||
|
|
||||||
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
|
Unless required by applicable law or agreed to in writing, software
|
||||||
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
See the License for the specific language governing permissions and
|
||||||
|
limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package nodeutilization
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"fmt"
|
||||||
|
"strings"
|
||||||
|
"testing"
|
||||||
|
|
||||||
|
v1 "k8s.io/api/core/v1"
|
||||||
|
"k8s.io/api/policy/v1beta1"
|
||||||
|
"k8s.io/apimachinery/pkg/api/resource"
|
||||||
|
"k8s.io/apimachinery/pkg/runtime"
|
||||||
|
"k8s.io/client-go/kubernetes/fake"
|
||||||
|
core "k8s.io/client-go/testing"
|
||||||
|
"sigs.k8s.io/descheduler/pkg/api"
|
||||||
|
"sigs.k8s.io/descheduler/pkg/descheduler/evictions"
|
||||||
|
"sigs.k8s.io/descheduler/pkg/utils"
|
||||||
|
"sigs.k8s.io/descheduler/test"
|
||||||
|
)
|
||||||
|
|
||||||
|
func TestHighNodeUtilization(t *testing.T) {
|
||||||
|
ctx := context.Background()
|
||||||
|
n1NodeName := "n1"
|
||||||
|
n2NodeName := "n2"
|
||||||
|
n3NodeName := "n3"
|
||||||
|
|
||||||
|
nodeSelectorKey := "datacenter"
|
||||||
|
nodeSelectorValue := "west"
|
||||||
|
|
||||||
|
testCases := []struct {
|
||||||
|
name string
|
||||||
|
thresholds api.ResourceThresholds
|
||||||
|
nodes map[string]*v1.Node
|
||||||
|
pods map[string]*v1.PodList
|
||||||
|
maxPodsToEvictPerNode int
|
||||||
|
expectedPodsEvicted int
|
||||||
|
evictedPods []string
|
||||||
|
}{
|
||||||
|
{
|
||||||
|
name: "no node below threshold usage",
|
||||||
|
thresholds: api.ResourceThresholds{
|
||||||
|
v1.ResourceCPU: 20,
|
||||||
|
v1.ResourcePods: 20,
|
||||||
|
},
|
||||||
|
nodes: map[string]*v1.Node{
|
||||||
|
n1NodeName: test.BuildTestNode(n1NodeName, 4000, 3000, 10, nil),
|
||||||
|
n2NodeName: test.BuildTestNode(n2NodeName, 4000, 3000, 10, nil),
|
||||||
|
n3NodeName: test.BuildTestNode(n3NodeName, 4000, 3000, 10, nil),
|
||||||
|
},
|
||||||
|
pods: map[string]*v1.PodList{
|
||||||
|
n1NodeName: {
|
||||||
|
Items: []v1.Pod{
|
||||||
|
// These won't be evicted.
|
||||||
|
*test.BuildTestPod("p1", 400, 0, n1NodeName, test.SetRSOwnerRef),
|
||||||
|
*test.BuildTestPod("p2", 400, 0, n1NodeName, test.SetRSOwnerRef),
|
||||||
|
*test.BuildTestPod("p3", 400, 0, n1NodeName, test.SetRSOwnerRef),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
n2NodeName: {
|
||||||
|
Items: []v1.Pod{
|
||||||
|
// These won't be evicted.
|
||||||
|
*test.BuildTestPod("p4", 400, 0, n2NodeName, test.SetRSOwnerRef),
|
||||||
|
*test.BuildTestPod("p5", 400, 0, n2NodeName, test.SetRSOwnerRef),
|
||||||
|
*test.BuildTestPod("p6", 400, 0, n2NodeName, test.SetRSOwnerRef),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
n3NodeName: {
|
||||||
|
Items: []v1.Pod{
|
||||||
|
// These won't be evicted.
|
||||||
|
*test.BuildTestPod("p7", 400, 0, n3NodeName, test.SetRSOwnerRef),
|
||||||
|
*test.BuildTestPod("p8", 400, 0, n3NodeName, test.SetRSOwnerRef),
|
||||||
|
*test.BuildTestPod("p9", 400, 0, n3NodeName, test.SetRSOwnerRef),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
maxPodsToEvictPerNode: 0,
|
||||||
|
expectedPodsEvicted: 0,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "no evictable pods",
|
||||||
|
thresholds: api.ResourceThresholds{
|
||||||
|
v1.ResourceCPU: 40,
|
||||||
|
v1.ResourcePods: 40,
|
||||||
|
},
|
||||||
|
nodes: map[string]*v1.Node{
|
||||||
|
n1NodeName: test.BuildTestNode(n1NodeName, 4000, 3000, 9, nil),
|
||||||
|
n2NodeName: test.BuildTestNode(n2NodeName, 4000, 3000, 10, nil),
|
||||||
|
n3NodeName: test.BuildTestNode(n3NodeName, 4000, 3000, 10, nil),
|
||||||
|
},
|
||||||
|
pods: map[string]*v1.PodList{
|
||||||
|
n1NodeName: {
|
||||||
|
Items: []v1.Pod{
|
||||||
|
// These won't be evicted.
|
||||||
|
*test.BuildTestPod("p1", 400, 0, n1NodeName, func(pod *v1.Pod) {
|
||||||
|
// A pod with local storage.
|
||||||
|
test.SetNormalOwnerRef(pod)
|
||||||
|
pod.Spec.Volumes = []v1.Volume{
|
||||||
|
{
|
||||||
|
Name: "sample",
|
||||||
|
VolumeSource: v1.VolumeSource{
|
||||||
|
HostPath: &v1.HostPathVolumeSource{Path: "somePath"},
|
||||||
|
EmptyDir: &v1.EmptyDirVolumeSource{
|
||||||
|
SizeLimit: resource.NewQuantity(int64(10), resource.BinarySI)},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
// A Mirror Pod.
|
||||||
|
pod.Annotations = test.GetMirrorPodAnnotation()
|
||||||
|
}),
|
||||||
|
*test.BuildTestPod("p2", 400, 0, n1NodeName, func(pod *v1.Pod) {
|
||||||
|
// A Critical Pod.
|
||||||
|
pod.Namespace = "kube-system"
|
||||||
|
priority := utils.SystemCriticalPriority
|
||||||
|
pod.Spec.Priority = &priority
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
n2NodeName: {
|
||||||
|
Items: []v1.Pod{
|
||||||
|
// These won't be evicted.
|
||||||
|
*test.BuildTestPod("p3", 400, 0, n2NodeName, test.SetDSOwnerRef),
|
||||||
|
*test.BuildTestPod("p4", 400, 0, n2NodeName, test.SetDSOwnerRef),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
n3NodeName: {
|
||||||
|
Items: []v1.Pod{
|
||||||
|
*test.BuildTestPod("p5", 400, 0, n3NodeName, test.SetRSOwnerRef),
|
||||||
|
*test.BuildTestPod("p6", 400, 0, n3NodeName, test.SetRSOwnerRef),
|
||||||
|
*test.BuildTestPod("p7", 400, 0, n3NodeName, test.SetRSOwnerRef),
|
||||||
|
*test.BuildTestPod("p8", 400, 0, n3NodeName, test.SetRSOwnerRef),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
maxPodsToEvictPerNode: 0,
|
||||||
|
expectedPodsEvicted: 0,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "no node to schedule evicted pods",
|
||||||
|
thresholds: api.ResourceThresholds{
|
||||||
|
v1.ResourceCPU: 20,
|
||||||
|
v1.ResourcePods: 20,
|
||||||
|
},
|
||||||
|
nodes: map[string]*v1.Node{
|
||||||
|
n1NodeName: test.BuildTestNode(n1NodeName, 4000, 3000, 10, nil),
|
||||||
|
n2NodeName: test.BuildTestNode(n2NodeName, 4000, 3000, 10, nil),
|
||||||
|
n3NodeName: test.BuildTestNode(n3NodeName, 4000, 3000, 10, test.SetNodeUnschedulable),
|
||||||
|
},
|
||||||
|
pods: map[string]*v1.PodList{
|
||||||
|
n1NodeName: {
|
||||||
|
Items: []v1.Pod{
|
||||||
|
// These can't be evicted.
|
||||||
|
*test.BuildTestPod("p1", 400, 0, n1NodeName, test.SetRSOwnerRef),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
n2NodeName: {
|
||||||
|
Items: []v1.Pod{
|
||||||
|
// These can't be evicted.
|
||||||
|
*test.BuildTestPod("p2", 400, 0, n2NodeName, test.SetRSOwnerRef),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
n3NodeName: {
|
||||||
|
Items: []v1.Pod{
|
||||||
|
*test.BuildTestPod("p3", 400, 0, n3NodeName, test.SetRSOwnerRef),
|
||||||
|
*test.BuildTestPod("p4", 400, 0, n3NodeName, test.SetRSOwnerRef),
|
||||||
|
*test.BuildTestPod("p5", 400, 0, n3NodeName, test.SetRSOwnerRef),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
maxPodsToEvictPerNode: 0,
|
||||||
|
expectedPodsEvicted: 0,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "without priorities",
|
||||||
|
thresholds: api.ResourceThresholds{
|
||||||
|
v1.ResourceCPU: 30,
|
||||||
|
v1.ResourcePods: 30,
|
||||||
|
},
|
||||||
|
nodes: map[string]*v1.Node{
|
||||||
|
n1NodeName: test.BuildTestNode(n1NodeName, 4000, 3000, 10, nil),
|
||||||
|
n2NodeName: test.BuildTestNode(n2NodeName, 4000, 3000, 10, nil),
|
||||||
|
n3NodeName: test.BuildTestNode(n3NodeName, 4000, 3000, 10, test.SetNodeUnschedulable),
|
||||||
|
},
|
||||||
|
pods: map[string]*v1.PodList{
|
||||||
|
n1NodeName: {
|
||||||
|
Items: []v1.Pod{
|
||||||
|
*test.BuildTestPod("p1", 400, 0, n1NodeName, test.SetRSOwnerRef),
|
||||||
|
// These won't be evicted.
|
||||||
|
*test.BuildTestPod("p2", 400, 0, n1NodeName, func(pod *v1.Pod) {
|
||||||
|
// A Critical Pod.
|
||||||
|
pod.Namespace = "kube-system"
|
||||||
|
priority := utils.SystemCriticalPriority
|
||||||
|
pod.Spec.Priority = &priority
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
n2NodeName: {
|
||||||
|
Items: []v1.Pod{
|
||||||
|
// These won't be evicted.
|
||||||
|
*test.BuildTestPod("p3", 400, 0, n2NodeName, test.SetRSOwnerRef),
|
||||||
|
*test.BuildTestPod("p4", 400, 0, n2NodeName, test.SetRSOwnerRef),
|
||||||
|
*test.BuildTestPod("p5", 400, 0, n2NodeName, test.SetRSOwnerRef),
|
||||||
|
*test.BuildTestPod("p6", 400, 0, n2NodeName, test.SetRSOwnerRef),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
n3NodeName: {
|
||||||
|
Items: []v1.Pod{
|
||||||
|
*test.BuildTestPod("p7", 400, 0, n3NodeName, test.SetRSOwnerRef),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
maxPodsToEvictPerNode: 0,
|
||||||
|
expectedPodsEvicted: 2,
|
||||||
|
evictedPods: []string{"p1", "p7"},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "without priorities stop when resource capacity is depleted",
|
||||||
|
thresholds: api.ResourceThresholds{
|
||||||
|
v1.ResourceCPU: 30,
|
||||||
|
v1.ResourcePods: 30,
|
||||||
|
},
|
||||||
|
nodes: map[string]*v1.Node{
|
||||||
|
n1NodeName: test.BuildTestNode(n1NodeName, 2000, 3000, 10, nil),
|
||||||
|
n2NodeName: test.BuildTestNode(n2NodeName, 2000, 3000, 10, nil),
|
||||||
|
n3NodeName: test.BuildTestNode(n3NodeName, 2000, 3000, 10, test.SetNodeUnschedulable),
|
||||||
|
},
|
||||||
|
pods: map[string]*v1.PodList{
|
||||||
|
n1NodeName: {
|
||||||
|
Items: []v1.Pod{
|
||||||
|
*test.BuildTestPod("p1", 400, 0, n1NodeName, test.SetRSOwnerRef),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
n2NodeName: {
|
||||||
|
Items: []v1.Pod{
|
||||||
|
// These won't be evicted.
|
||||||
|
*test.BuildTestPod("p2", 400, 0, n2NodeName, test.SetRSOwnerRef),
|
||||||
|
*test.BuildTestPod("p3", 400, 0, n2NodeName, test.SetRSOwnerRef),
|
||||||
|
*test.BuildTestPod("p4", 400, 0, n2NodeName, test.SetRSOwnerRef),
|
||||||
|
*test.BuildTestPod("p5", 400, 0, n2NodeName, test.SetRSOwnerRef),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
n3NodeName: {
|
||||||
|
Items: []v1.Pod{
|
||||||
|
*test.BuildTestPod("p6", 400, 0, n3NodeName, test.SetRSOwnerRef),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
maxPodsToEvictPerNode: 0,
|
||||||
|
expectedPodsEvicted: 1,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "with priorities",
|
||||||
|
thresholds: api.ResourceThresholds{
|
||||||
|
v1.ResourceCPU: 30,
|
||||||
|
},
|
||||||
|
nodes: map[string]*v1.Node{
|
||||||
|
n1NodeName: test.BuildTestNode(n1NodeName, 4000, 3000, 10, nil),
|
||||||
|
n2NodeName: test.BuildTestNode(n2NodeName, 2000, 3000, 10, nil),
|
||||||
|
n3NodeName: test.BuildTestNode(n3NodeName, 2000, 3000, 10, test.SetNodeUnschedulable),
|
||||||
|
},
|
||||||
|
pods: map[string]*v1.PodList{
|
||||||
|
n1NodeName: {
|
||||||
|
Items: []v1.Pod{
|
||||||
|
*test.BuildTestPod("p1", 400, 0, n1NodeName, func(pod *v1.Pod) {
|
||||||
|
test.SetRSOwnerRef(pod)
|
||||||
|
test.SetPodPriority(pod, lowPriority)
|
||||||
|
}),
|
||||||
|
*test.BuildTestPod("p2", 400, 0, n1NodeName, func(pod *v1.Pod) {
|
||||||
|
test.SetRSOwnerRef(pod)
|
||||||
|
test.SetPodPriority(pod, highPriority)
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
n2NodeName: {
|
||||||
|
Items: []v1.Pod{
|
||||||
|
// These won't be evicted.
|
||||||
|
*test.BuildTestPod("p5", 400, 0, n2NodeName, test.SetRSOwnerRef),
|
||||||
|
*test.BuildTestPod("p6", 400, 0, n2NodeName, test.SetRSOwnerRef),
|
||||||
|
*test.BuildTestPod("p7", 400, 0, n2NodeName, test.SetRSOwnerRef),
|
||||||
|
*test.BuildTestPod("p8", 400, 0, n2NodeName, test.SetRSOwnerRef),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
n3NodeName: {
|
||||||
|
Items: []v1.Pod{
|
||||||
|
// These won't be evicted.
|
||||||
|
*test.BuildTestPod("p9", 400, 0, n3NodeName, test.SetDSOwnerRef),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
maxPodsToEvictPerNode: 0,
|
||||||
|
expectedPodsEvicted: 1,
|
||||||
|
evictedPods: []string{"p1"},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "without priorities evicting best-effort pods only",
|
||||||
|
thresholds: api.ResourceThresholds{
|
||||||
|
v1.ResourceCPU: 30,
|
||||||
|
},
|
||||||
|
nodes: map[string]*v1.Node{
|
||||||
|
n1NodeName: test.BuildTestNode(n1NodeName, 3000, 3000, 10, nil),
|
||||||
|
n2NodeName: test.BuildTestNode(n2NodeName, 3000, 3000, 5, nil),
|
||||||
|
n3NodeName: test.BuildTestNode(n3NodeName, 3000, 3000, 10, test.SetNodeUnschedulable),
|
||||||
|
},
|
||||||
|
// All pods are assumed to be burstable (test.BuildTestNode always sets both cpu/memory resource requests to some value)
|
||||||
|
pods: map[string]*v1.PodList{
|
||||||
|
n1NodeName: {
|
||||||
|
Items: []v1.Pod{
|
||||||
|
*test.BuildTestPod("p1", 400, 0, n1NodeName, func(pod *v1.Pod) {
|
||||||
|
test.SetRSOwnerRef(pod)
|
||||||
|
test.MakeBestEffortPod(pod)
|
||||||
|
}),
|
||||||
|
*test.BuildTestPod("p2", 400, 0, n1NodeName, func(pod *v1.Pod) {
|
||||||
|
test.SetRSOwnerRef(pod)
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
n2NodeName: {
|
||||||
|
Items: []v1.Pod{
|
||||||
|
// These won't be evicted.
|
||||||
|
*test.BuildTestPod("p3", 400, 0, n2NodeName, test.SetRSOwnerRef),
|
||||||
|
*test.BuildTestPod("p4", 400, 0, n2NodeName, test.SetRSOwnerRef),
|
||||||
|
*test.BuildTestPod("p5", 400, 0, n2NodeName, test.SetRSOwnerRef),
|
||||||
|
*test.BuildTestPod("p6", 400, 0, n2NodeName, test.SetRSOwnerRef),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
n3NodeName: {
|
||||||
|
Items: []v1.Pod{},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
maxPodsToEvictPerNode: 0,
|
||||||
|
expectedPodsEvicted: 1,
|
||||||
|
evictedPods: []string{"p1"},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "with extended resource",
|
||||||
|
thresholds: api.ResourceThresholds{
|
||||||
|
v1.ResourceCPU: 20,
|
||||||
|
extendedResource: 40,
|
||||||
|
},
|
||||||
|
nodes: map[string]*v1.Node{
|
||||||
|
n1NodeName: test.BuildTestNode(n1NodeName, 4000, 3000, 10, func(node *v1.Node) {
|
||||||
|
test.SetNodeExtendedResource(node, extendedResource, 8)
|
||||||
|
}),
|
||||||
|
n2NodeName: test.BuildTestNode(n2NodeName, 4000, 3000, 10, func(node *v1.Node) {
|
||||||
|
test.SetNodeExtendedResource(node, extendedResource, 8)
|
||||||
|
}),
|
||||||
|
n3NodeName: test.BuildTestNode(n3NodeName, 4000, 3000, 10, test.SetNodeUnschedulable),
|
||||||
|
},
|
||||||
|
pods: map[string]*v1.PodList{
|
||||||
|
n1NodeName: {
|
||||||
|
Items: []v1.Pod{
|
||||||
|
*test.BuildTestPod("p1", 100, 0, n1NodeName, func(pod *v1.Pod) {
|
||||||
|
test.SetRSOwnerRef(pod)
|
||||||
|
test.SetPodExtendedResourceRequest(pod, extendedResource, 1)
|
||||||
|
}),
|
||||||
|
*test.BuildTestPod("p2", 100, 0, n1NodeName, func(pod *v1.Pod) {
|
||||||
|
test.SetRSOwnerRef(pod)
|
||||||
|
test.SetPodExtendedResourceRequest(pod, extendedResource, 1)
|
||||||
|
}),
|
||||||
|
// These won't be evicted
|
||||||
|
*test.BuildTestPod("p2", 100, 0, n1NodeName, func(pod *v1.Pod) {
|
||||||
|
test.SetDSOwnerRef(pod)
|
||||||
|
test.SetPodExtendedResourceRequest(pod, extendedResource, 1)
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
n2NodeName: {
|
||||||
|
Items: []v1.Pod{
|
||||||
|
*test.BuildTestPod("p3", 500, 0, n2NodeName, func(pod *v1.Pod) {
|
||||||
|
test.SetRSOwnerRef(pod)
|
||||||
|
test.SetPodExtendedResourceRequest(pod, extendedResource, 1)
|
||||||
|
}),
|
||||||
|
*test.BuildTestPod("p4", 500, 0, n2NodeName, func(pod *v1.Pod) {
|
||||||
|
test.SetRSOwnerRef(pod)
|
||||||
|
test.SetPodExtendedResourceRequest(pod, extendedResource, 1)
|
||||||
|
}),
|
||||||
|
*test.BuildTestPod("p5", 500, 0, n2NodeName, func(pod *v1.Pod) {
|
||||||
|
test.SetRSOwnerRef(pod)
|
||||||
|
test.SetPodExtendedResourceRequest(pod, extendedResource, 1)
|
||||||
|
}),
|
||||||
|
*test.BuildTestPod("p6", 500, 0, n2NodeName, func(pod *v1.Pod) {
|
||||||
|
test.SetRSOwnerRef(pod)
|
||||||
|
test.SetPodExtendedResourceRequest(pod, extendedResource, 1)
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
n3NodeName: {
|
||||||
|
Items: []v1.Pod{},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
maxPodsToEvictPerNode: 0,
|
||||||
|
expectedPodsEvicted: 2,
|
||||||
|
evictedPods: []string{"p1", "p2"},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "with extended resource in some of nodes",
|
||||||
|
thresholds: api.ResourceThresholds{
|
||||||
|
v1.ResourceCPU: 40,
|
||||||
|
extendedResource: 40,
|
||||||
|
},
|
||||||
|
nodes: map[string]*v1.Node{
|
||||||
|
n1NodeName: test.BuildTestNode(n1NodeName, 4000, 3000, 10, func(node *v1.Node) {
|
||||||
|
test.SetNodeExtendedResource(node, extendedResource, 8)
|
||||||
|
}),
|
||||||
|
n2NodeName: test.BuildTestNode(n2NodeName, 4000, 3000, 10, nil),
|
||||||
|
n3NodeName: test.BuildTestNode(n3NodeName, 4000, 3000, 10, test.SetNodeUnschedulable),
|
||||||
|
},
|
||||||
|
pods: map[string]*v1.PodList{
|
||||||
|
n1NodeName: {
|
||||||
|
Items: []v1.Pod{
|
||||||
|
//These won't be evicted
|
||||||
|
*test.BuildTestPod("p1", 100, 0, n1NodeName, func(pod *v1.Pod) {
|
||||||
|
test.SetRSOwnerRef(pod)
|
||||||
|
test.SetPodExtendedResourceRequest(pod, extendedResource, 1)
|
||||||
|
}),
|
||||||
|
*test.BuildTestPod("p2", 100, 0, n1NodeName, func(pod *v1.Pod) {
|
||||||
|
test.SetRSOwnerRef(pod)
|
||||||
|
test.SetPodExtendedResourceRequest(pod, extendedResource, 1)
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
n2NodeName: {
|
||||||
|
Items: []v1.Pod{
|
||||||
|
*test.BuildTestPod("p3", 500, 0, n2NodeName, test.SetRSOwnerRef),
|
||||||
|
*test.BuildTestPod("p4", 500, 0, n2NodeName, test.SetRSOwnerRef),
|
||||||
|
*test.BuildTestPod("p5", 500, 0, n2NodeName, test.SetRSOwnerRef),
|
||||||
|
*test.BuildTestPod("p6", 500, 0, n2NodeName, test.SetRSOwnerRef),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
n3NodeName: {
|
||||||
|
Items: []v1.Pod{},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
maxPodsToEvictPerNode: 0,
|
||||||
|
expectedPodsEvicted: 0,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "Other node match pod node selector",
|
||||||
|
thresholds: api.ResourceThresholds{
|
||||||
|
v1.ResourceCPU: 30,
|
||||||
|
v1.ResourcePods: 30,
|
||||||
|
},
|
||||||
|
nodes: map[string]*v1.Node{
|
||||||
|
n1NodeName: test.BuildTestNode(n1NodeName, 4000, 3000, 9, func(node *v1.Node) {
|
||||||
|
node.ObjectMeta.Labels = map[string]string{
|
||||||
|
nodeSelectorKey: nodeSelectorValue,
|
||||||
|
}
|
||||||
|
}),
|
||||||
|
n2NodeName: test.BuildTestNode(n2NodeName, 4000, 3000, 10, nil),
|
||||||
|
},
|
||||||
|
pods: map[string]*v1.PodList{
|
||||||
|
n1NodeName: {
|
||||||
|
Items: []v1.Pod{
|
||||||
|
*test.BuildTestPod("p1", 400, 0, n1NodeName, test.SetRSOwnerRef),
|
||||||
|
*test.BuildTestPod("p2", 400, 0, n1NodeName, test.SetRSOwnerRef),
|
||||||
|
*test.BuildTestPod("p3", 400, 0, n1NodeName, test.SetRSOwnerRef),
|
||||||
|
*test.BuildTestPod("p4", 400, 0, n1NodeName, test.SetDSOwnerRef),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
n2NodeName: {
|
||||||
|
Items: []v1.Pod{
|
||||||
|
*test.BuildTestPod("p5", 400, 0, n2NodeName, func(pod *v1.Pod) {
|
||||||
|
// A pod selecting nodes in the "west" datacenter
|
||||||
|
test.SetRSOwnerRef(pod)
|
||||||
|
pod.Spec.NodeSelector = map[string]string{
|
||||||
|
nodeSelectorKey: nodeSelectorValue,
|
||||||
|
}
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
maxPodsToEvictPerNode: 0,
|
||||||
|
expectedPodsEvicted: 1,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "Other node does not match pod node selector",
|
||||||
|
thresholds: api.ResourceThresholds{
|
||||||
|
v1.ResourceCPU: 30,
|
||||||
|
v1.ResourcePods: 30,
|
||||||
|
},
|
||||||
|
nodes: map[string]*v1.Node{
|
||||||
|
n1NodeName: test.BuildTestNode(n1NodeName, 4000, 3000, 9, nil),
|
||||||
|
n2NodeName: test.BuildTestNode(n2NodeName, 4000, 3000, 10, nil),
|
||||||
|
},
|
||||||
|
pods: map[string]*v1.PodList{
|
||||||
|
n1NodeName: {
|
||||||
|
Items: []v1.Pod{
|
||||||
|
*test.BuildTestPod("p1", 400, 0, n1NodeName, test.SetRSOwnerRef),
|
||||||
|
*test.BuildTestPod("p2", 400, 0, n1NodeName, test.SetRSOwnerRef),
|
||||||
|
*test.BuildTestPod("p3", 400, 0, n1NodeName, test.SetRSOwnerRef),
|
||||||
|
*test.BuildTestPod("p4", 400, 0, n1NodeName, test.SetDSOwnerRef),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
n2NodeName: {
|
||||||
|
Items: []v1.Pod{
|
||||||
|
*test.BuildTestPod("p5", 400, 0, n2NodeName, func(pod *v1.Pod) {
|
||||||
|
// A pod selecting nodes in the "west" datacenter
|
||||||
|
test.SetRSOwnerRef(pod)
|
||||||
|
pod.Spec.NodeSelector = map[string]string{
|
||||||
|
nodeSelectorKey: nodeSelectorValue,
|
||||||
|
}
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
maxPodsToEvictPerNode: 0,
|
||||||
|
expectedPodsEvicted: 0,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, test := range testCases {
|
||||||
|
t.Run(test.name, func(t *testing.T) {
|
||||||
|
fakeClient := &fake.Clientset{}
|
||||||
|
fakeClient.Fake.AddReactor("list", "pods", func(action core.Action) (bool, runtime.Object, error) {
|
||||||
|
list := action.(core.ListAction)
|
||||||
|
fieldString := list.GetListRestrictions().Fields.String()
|
||||||
|
if strings.Contains(fieldString, n1NodeName) {
|
||||||
|
return true, test.pods[n1NodeName], nil
|
||||||
|
}
|
||||||
|
if strings.Contains(fieldString, n2NodeName) {
|
||||||
|
return true, test.pods[n2NodeName], nil
|
||||||
|
}
|
||||||
|
if strings.Contains(fieldString, n3NodeName) {
|
||||||
|
return true, test.pods[n3NodeName], nil
|
||||||
|
}
|
||||||
|
return true, nil, fmt.Errorf("Failed to list: %v", list)
|
||||||
|
})
|
||||||
|
fakeClient.Fake.AddReactor("get", "nodes", func(action core.Action) (bool, runtime.Object, error) {
|
||||||
|
getAction := action.(core.GetAction)
|
||||||
|
if node, exists := test.nodes[getAction.GetName()]; exists {
|
||||||
|
return true, node, nil
|
||||||
|
}
|
||||||
|
return true, nil, fmt.Errorf("Wrong node: %v", getAction.GetName())
|
||||||
|
})
|
||||||
|
podsForEviction := make(map[string]struct{})
|
||||||
|
for _, pod := range test.evictedPods {
|
||||||
|
podsForEviction[pod] = struct{}{}
|
||||||
|
}
|
||||||
|
|
||||||
|
evictionFailed := false
|
||||||
|
if len(test.evictedPods) > 0 {
|
||||||
|
fakeClient.Fake.AddReactor("create", "pods", func(action core.Action) (bool, runtime.Object, error) {
|
||||||
|
getAction := action.(core.CreateAction)
|
||||||
|
obj := getAction.GetObject()
|
||||||
|
if eviction, ok := obj.(*v1beta1.Eviction); ok {
|
||||||
|
if _, exists := podsForEviction[eviction.Name]; exists {
|
||||||
|
return true, obj, nil
|
||||||
|
}
|
||||||
|
evictionFailed = true
|
||||||
|
return true, nil, fmt.Errorf("pod %q was unexpectedly evicted", eviction.Name)
|
||||||
|
}
|
||||||
|
return true, obj, nil
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
var nodes []*v1.Node
|
||||||
|
for _, node := range test.nodes {
|
||||||
|
nodes = append(nodes, node)
|
||||||
|
}
|
||||||
|
|
||||||
|
podEvictor := evictions.NewPodEvictor(
|
||||||
|
fakeClient,
|
||||||
|
"v1",
|
||||||
|
false,
|
||||||
|
test.maxPodsToEvictPerNode,
|
||||||
|
nodes,
|
||||||
|
false,
|
||||||
|
false,
|
||||||
|
false,
|
||||||
|
)
|
||||||
|
|
||||||
|
strategy := api.DeschedulerStrategy{
|
||||||
|
Enabled: true,
|
||||||
|
Params: &api.StrategyParameters{
|
||||||
|
NodeResourceUtilizationThresholds: &api.NodeResourceUtilizationThresholds{
|
||||||
|
Thresholds: test.thresholds,
|
||||||
|
},
|
||||||
|
NodeFit: true,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
HighNodeUtilization(ctx, fakeClient, strategy, nodes, podEvictor)
|
||||||
|
|
||||||
|
podsEvicted := podEvictor.TotalEvicted()
|
||||||
|
if test.expectedPodsEvicted != podsEvicted {
|
||||||
|
t.Errorf("Expected %#v pods to be evicted but %#v got evicted", test.expectedPodsEvicted, podsEvicted)
|
||||||
|
}
|
||||||
|
if evictionFailed {
|
||||||
|
t.Errorf("Pod evictions failed unexpectedly")
|
||||||
|
}
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestValidateHighNodeUtilizationStrategyConfig(t *testing.T) {
|
||||||
|
tests := []struct {
|
||||||
|
name string
|
||||||
|
thresholds api.ResourceThresholds
|
||||||
|
targetThresholds api.ResourceThresholds
|
||||||
|
errInfo error
|
||||||
|
}{
|
||||||
|
{
|
||||||
|
name: "passing target thresholds",
|
||||||
|
thresholds: api.ResourceThresholds{
|
||||||
|
v1.ResourceCPU: 20,
|
||||||
|
v1.ResourceMemory: 20,
|
||||||
|
},
|
||||||
|
targetThresholds: api.ResourceThresholds{
|
||||||
|
v1.ResourceCPU: 80,
|
||||||
|
v1.ResourceMemory: 80,
|
||||||
|
},
|
||||||
|
errInfo: fmt.Errorf("targetThresholds is not applicable for HighNodeUtilization"),
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "passing empty thresholds",
|
||||||
|
thresholds: api.ResourceThresholds{},
|
||||||
|
errInfo: fmt.Errorf("thresholds config is not valid: no resource threshold is configured"),
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "passing invalid thresholds",
|
||||||
|
thresholds: api.ResourceThresholds{
|
||||||
|
v1.ResourceCPU: 80,
|
||||||
|
v1.ResourceMemory: 120,
|
||||||
|
},
|
||||||
|
errInfo: fmt.Errorf("thresholds config is not valid: %v", fmt.Errorf(
|
||||||
|
"%v threshold not in [%v, %v] range", v1.ResourceMemory, MinResourcePercentage, MaxResourcePercentage)),
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "passing valid strategy config",
|
||||||
|
thresholds: api.ResourceThresholds{
|
||||||
|
v1.ResourceCPU: 80,
|
||||||
|
v1.ResourceMemory: 80,
|
||||||
|
},
|
||||||
|
errInfo: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "passing valid strategy config with extended resource",
|
||||||
|
thresholds: api.ResourceThresholds{
|
||||||
|
v1.ResourceCPU: 80,
|
||||||
|
v1.ResourceMemory: 80,
|
||||||
|
extendedResource: 80,
|
||||||
|
},
|
||||||
|
errInfo: nil,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, testCase := range tests {
|
||||||
|
validateErr := validateHighUtilizationStrategyConfig(testCase.thresholds, testCase.targetThresholds)
|
||||||
|
|
||||||
|
if validateErr == nil || testCase.errInfo == nil {
|
||||||
|
if validateErr != testCase.errInfo {
|
||||||
|
t.Errorf("expected validity of strategy config: thresholds %#v targetThresholds %#v to be %v but got %v instead",
|
||||||
|
testCase.thresholds, testCase.targetThresholds, testCase.errInfo, validateErr)
|
||||||
|
}
|
||||||
|
} else if validateErr.Error() != testCase.errInfo.Error() {
|
||||||
|
t.Errorf("expected validity of strategy config: thresholds %#v targetThresholds %#v to be %v but got %v instead",
|
||||||
|
testCase.thresholds, testCase.targetThresholds, testCase.errInfo, validateErr)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestHighNodeUtilizationWithTaints(t *testing.T) {
|
||||||
|
ctx := context.Background()
|
||||||
|
strategy := api.DeschedulerStrategy{
|
||||||
|
Enabled: true,
|
||||||
|
Params: &api.StrategyParameters{
|
||||||
|
NodeResourceUtilizationThresholds: &api.NodeResourceUtilizationThresholds{
|
||||||
|
Thresholds: api.ResourceThresholds{
|
||||||
|
v1.ResourceCPU: 40,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
n1 := test.BuildTestNode("n1", 1000, 3000, 10, nil)
|
||||||
|
n2 := test.BuildTestNode("n2", 1000, 3000, 10, nil)
|
||||||
|
n3 := test.BuildTestNode("n3", 1000, 3000, 10, nil)
|
||||||
|
n3withTaints := n3.DeepCopy()
|
||||||
|
n3withTaints.Spec.Taints = []v1.Taint{
|
||||||
|
{
|
||||||
|
Key: "key",
|
||||||
|
Value: "value",
|
||||||
|
Effect: v1.TaintEffectNoSchedule,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
podThatToleratesTaint := test.BuildTestPod("tolerate_pod", 200, 0, n1.Name, test.SetRSOwnerRef)
|
||||||
|
podThatToleratesTaint.Spec.Tolerations = []v1.Toleration{
|
||||||
|
{
|
||||||
|
Key: "key",
|
||||||
|
Value: "value",
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
tests := []struct {
|
||||||
|
name string
|
||||||
|
nodes []*v1.Node
|
||||||
|
pods []*v1.Pod
|
||||||
|
evictionsExpected int
|
||||||
|
}{
|
||||||
|
{
|
||||||
|
name: "No taints",
|
||||||
|
nodes: []*v1.Node{n1, n2, n3},
|
||||||
|
pods: []*v1.Pod{
|
||||||
|
//Node 1 pods
|
||||||
|
test.BuildTestPod(fmt.Sprintf("pod_1_%s", n1.Name), 200, 0, n1.Name, test.SetRSOwnerRef),
|
||||||
|
test.BuildTestPod(fmt.Sprintf("pod_2_%s", n1.Name), 200, 0, n1.Name, test.SetRSOwnerRef),
|
||||||
|
test.BuildTestPod(fmt.Sprintf("pod_3_%s", n1.Name), 200, 0, n1.Name, test.SetRSOwnerRef),
|
||||||
|
// Node 2 pods
|
||||||
|
test.BuildTestPod(fmt.Sprintf("pod_4_%s", n2.Name), 200, 0, n2.Name, test.SetRSOwnerRef),
|
||||||
|
},
|
||||||
|
evictionsExpected: 1,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "No pod tolerates node taint",
|
||||||
|
nodes: []*v1.Node{n1, n3withTaints},
|
||||||
|
pods: []*v1.Pod{
|
||||||
|
//Node 1 pods
|
||||||
|
test.BuildTestPod(fmt.Sprintf("pod_1_%s", n1.Name), 200, 0, n1.Name, test.SetRSOwnerRef),
|
||||||
|
// Node 3 pods
|
||||||
|
test.BuildTestPod(fmt.Sprintf("pod_2_%s", n3withTaints.Name), 200, 0, n3withTaints.Name, test.SetRSOwnerRef),
|
||||||
|
},
|
||||||
|
evictionsExpected: 0,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "Pod which tolerates node taint",
|
||||||
|
nodes: []*v1.Node{n1, n3withTaints},
|
||||||
|
pods: []*v1.Pod{
|
||||||
|
//Node 1 pods
|
||||||
|
test.BuildTestPod(fmt.Sprintf("pod_1_%s", n1.Name), 100, 0, n1.Name, test.SetRSOwnerRef),
|
||||||
|
podThatToleratesTaint,
|
||||||
|
// Node 3 pods
|
||||||
|
test.BuildTestPod(fmt.Sprintf("pod_9_%s", n3withTaints.Name), 500, 0, n3withTaints.Name, test.SetRSOwnerRef),
|
||||||
|
},
|
||||||
|
evictionsExpected: 1,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, item := range tests {
|
||||||
|
t.Run(item.name, func(t *testing.T) {
|
||||||
|
var objs []runtime.Object
|
||||||
|
for _, node := range item.nodes {
|
||||||
|
objs = append(objs, node)
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, pod := range item.pods {
|
||||||
|
objs = append(objs, pod)
|
||||||
|
}
|
||||||
|
|
||||||
|
fakeClient := fake.NewSimpleClientset(objs...)
|
||||||
|
|
||||||
|
podEvictor := evictions.NewPodEvictor(
|
||||||
|
fakeClient,
|
||||||
|
"policy/v1",
|
||||||
|
false,
|
||||||
|
item.evictionsExpected,
|
||||||
|
item.nodes,
|
||||||
|
false,
|
||||||
|
false,
|
||||||
|
false,
|
||||||
|
)
|
||||||
|
|
||||||
|
HighNodeUtilization(ctx, fakeClient, strategy, item.nodes, podEvictor)
|
||||||
|
|
||||||
|
if item.evictionsExpected != podEvictor.TotalEvicted() {
|
||||||
|
t.Errorf("Expected %v evictions, got %v", item.evictionsExpected, podEvictor.TotalEvicted())
|
||||||
|
}
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
188
pkg/descheduler/strategies/nodeutilization/lownodeutilization.go
Normal file
188
pkg/descheduler/strategies/nodeutilization/lownodeutilization.go
Normal file
@@ -0,0 +1,188 @@
|
|||||||
|
/*
|
||||||
|
Copyright 2017 The Kubernetes Authors.
|
||||||
|
|
||||||
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
you may not use this file except in compliance with the License.
|
||||||
|
You may obtain a copy of the License at
|
||||||
|
|
||||||
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
|
Unless required by applicable law or agreed to in writing, software
|
||||||
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
See the License for the specific language governing permissions and
|
||||||
|
limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package nodeutilization
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"fmt"
|
||||||
|
|
||||||
|
v1 "k8s.io/api/core/v1"
|
||||||
|
"k8s.io/apimachinery/pkg/api/resource"
|
||||||
|
clientset "k8s.io/client-go/kubernetes"
|
||||||
|
"k8s.io/klog/v2"
|
||||||
|
|
||||||
|
"sigs.k8s.io/descheduler/pkg/api"
|
||||||
|
"sigs.k8s.io/descheduler/pkg/descheduler/evictions"
|
||||||
|
nodeutil "sigs.k8s.io/descheduler/pkg/descheduler/node"
|
||||||
|
"sigs.k8s.io/descheduler/pkg/utils"
|
||||||
|
)
|
||||||
|
|
||||||
|
// LowNodeUtilization evicts pods from overutilized nodes to underutilized nodes. Note that CPU/Memory requests are used
|
||||||
|
// to calculate nodes' utilization and not the actual resource usage.
|
||||||
|
func LowNodeUtilization(ctx context.Context, client clientset.Interface, strategy api.DeschedulerStrategy, nodes []*v1.Node, podEvictor *evictions.PodEvictor) {
|
||||||
|
// TODO: May be create a struct for the strategy as well, so that we don't have to pass along the all the params?
|
||||||
|
if err := validateNodeUtilizationParams(strategy.Params); err != nil {
|
||||||
|
klog.ErrorS(err, "Invalid LowNodeUtilization parameters")
|
||||||
|
return
|
||||||
|
}
|
||||||
|
thresholdPriority, err := utils.GetPriorityFromStrategyParams(ctx, client, strategy.Params)
|
||||||
|
if err != nil {
|
||||||
|
klog.ErrorS(err, "Failed to get threshold priority from strategy's params")
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
nodeFit := false
|
||||||
|
if strategy.Params != nil {
|
||||||
|
nodeFit = strategy.Params.NodeFit
|
||||||
|
}
|
||||||
|
|
||||||
|
thresholds := strategy.Params.NodeResourceUtilizationThresholds.Thresholds
|
||||||
|
targetThresholds := strategy.Params.NodeResourceUtilizationThresholds.TargetThresholds
|
||||||
|
if err := validateLowUtilizationStrategyConfig(thresholds, targetThresholds); err != nil {
|
||||||
|
klog.ErrorS(err, "LowNodeUtilization config is not valid")
|
||||||
|
return
|
||||||
|
}
|
||||||
|
// check if Pods/CPU/Mem are set, if not, set them to 100
|
||||||
|
if _, ok := thresholds[v1.ResourcePods]; !ok {
|
||||||
|
thresholds[v1.ResourcePods] = MaxResourcePercentage
|
||||||
|
targetThresholds[v1.ResourcePods] = MaxResourcePercentage
|
||||||
|
}
|
||||||
|
if _, ok := thresholds[v1.ResourceCPU]; !ok {
|
||||||
|
thresholds[v1.ResourceCPU] = MaxResourcePercentage
|
||||||
|
targetThresholds[v1.ResourceCPU] = MaxResourcePercentage
|
||||||
|
}
|
||||||
|
if _, ok := thresholds[v1.ResourceMemory]; !ok {
|
||||||
|
thresholds[v1.ResourceMemory] = MaxResourcePercentage
|
||||||
|
targetThresholds[v1.ResourceMemory] = MaxResourcePercentage
|
||||||
|
}
|
||||||
|
resourceNames := getResourceNames(thresholds)
|
||||||
|
|
||||||
|
lowNodes, sourceNodes := classifyNodes(
|
||||||
|
getNodeUsage(ctx, client, nodes, thresholds, targetThresholds, resourceNames),
|
||||||
|
// The node has to be schedulable (to be able to move workload there)
|
||||||
|
func(node *v1.Node, usage NodeUsage) bool {
|
||||||
|
if nodeutil.IsNodeUnschedulable(node) {
|
||||||
|
klog.V(2).InfoS("Node is unschedulable, thus not considered as underutilized", "node", klog.KObj(node))
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
return isNodeWithLowUtilization(usage)
|
||||||
|
},
|
||||||
|
func(node *v1.Node, usage NodeUsage) bool {
|
||||||
|
return isNodeAboveTargetUtilization(usage)
|
||||||
|
},
|
||||||
|
)
|
||||||
|
|
||||||
|
// log message in one line
|
||||||
|
keysAndValues := []interface{}{
|
||||||
|
"CPU", thresholds[v1.ResourceCPU],
|
||||||
|
"Mem", thresholds[v1.ResourceMemory],
|
||||||
|
"Pods", thresholds[v1.ResourcePods],
|
||||||
|
}
|
||||||
|
for name := range thresholds {
|
||||||
|
if !isBasicResource(name) {
|
||||||
|
keysAndValues = append(keysAndValues, string(name), int64(thresholds[name]))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
klog.V(1).InfoS("Criteria for a node under utilization", keysAndValues...)
|
||||||
|
klog.V(1).InfoS("Number of underutilized nodes", "totalNumber", len(lowNodes))
|
||||||
|
|
||||||
|
// log message in one line
|
||||||
|
keysAndValues = []interface{}{
|
||||||
|
"CPU", targetThresholds[v1.ResourceCPU],
|
||||||
|
"Mem", targetThresholds[v1.ResourceMemory],
|
||||||
|
"Pods", targetThresholds[v1.ResourcePods],
|
||||||
|
}
|
||||||
|
for name := range targetThresholds {
|
||||||
|
if !isBasicResource(name) {
|
||||||
|
keysAndValues = append(keysAndValues, string(name), int64(targetThresholds[name]))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
klog.V(1).InfoS("Criteria for a node above target utilization", keysAndValues...)
|
||||||
|
klog.V(1).InfoS("Number of overutilized nodes", "totalNumber", len(sourceNodes))
|
||||||
|
|
||||||
|
if len(lowNodes) == 0 {
|
||||||
|
klog.V(1).InfoS("No node is underutilized, nothing to do here, you might tune your thresholds further")
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(lowNodes) <= strategy.Params.NodeResourceUtilizationThresholds.NumberOfNodes {
|
||||||
|
klog.V(1).InfoS("Number of nodes underutilized is less or equal than NumberOfNodes, nothing to do here", "underutilizedNodes", len(lowNodes), "numberOfNodes", strategy.Params.NodeResourceUtilizationThresholds.NumberOfNodes)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(lowNodes) == len(nodes) {
|
||||||
|
klog.V(1).InfoS("All nodes are underutilized, nothing to do here")
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(sourceNodes) == 0 {
|
||||||
|
klog.V(1).InfoS("All nodes are under target utilization, nothing to do here")
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
evictable := podEvictor.Evictable(evictions.WithPriorityThreshold(thresholdPriority), evictions.WithNodeFit(nodeFit))
|
||||||
|
|
||||||
|
// stop if node utilization drops below target threshold or any of required capacity (cpu, memory, pods) is moved
|
||||||
|
continueEvictionCond := func(nodeUsage NodeUsage, totalAvailableUsage map[v1.ResourceName]*resource.Quantity) bool {
|
||||||
|
if !isNodeAboveTargetUtilization(nodeUsage) {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
for name := range totalAvailableUsage {
|
||||||
|
if totalAvailableUsage[name].CmpInt64(0) < 1 {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
|
||||||
|
evictPodsFromSourceNodes(
|
||||||
|
ctx,
|
||||||
|
sourceNodes,
|
||||||
|
lowNodes,
|
||||||
|
podEvictor,
|
||||||
|
evictable.IsEvictable,
|
||||||
|
resourceNames,
|
||||||
|
"LowNodeUtilization",
|
||||||
|
continueEvictionCond)
|
||||||
|
|
||||||
|
klog.V(1).InfoS("Total number of pods evicted", "evictedPods", podEvictor.TotalEvicted())
|
||||||
|
}
|
||||||
|
|
||||||
|
// validateLowUtilizationStrategyConfig checks if the strategy's config is valid
|
||||||
|
func validateLowUtilizationStrategyConfig(thresholds, targetThresholds api.ResourceThresholds) error {
|
||||||
|
// validate thresholds and targetThresholds config
|
||||||
|
if err := validateThresholds(thresholds); err != nil {
|
||||||
|
return fmt.Errorf("thresholds config is not valid: %v", err)
|
||||||
|
}
|
||||||
|
if err := validateThresholds(targetThresholds); err != nil {
|
||||||
|
return fmt.Errorf("targetThresholds config is not valid: %v", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
// validate if thresholds and targetThresholds have same resources configured
|
||||||
|
if len(thresholds) != len(targetThresholds) {
|
||||||
|
return fmt.Errorf("thresholds and targetThresholds configured different resources")
|
||||||
|
}
|
||||||
|
for resourceName, value := range thresholds {
|
||||||
|
if targetValue, ok := targetThresholds[resourceName]; !ok {
|
||||||
|
return fmt.Errorf("thresholds and targetThresholds configured different resources")
|
||||||
|
} else if value > targetValue {
|
||||||
|
return fmt.Errorf("thresholds' %v percentage is greater than targetThresholds'", resourceName)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
File diff suppressed because it is too large
Load Diff
@@ -18,18 +18,16 @@ package nodeutilization
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
"sigs.k8s.io/descheduler/pkg/api"
|
"fmt"
|
||||||
"sort"
|
|
||||||
|
|
||||||
v1 "k8s.io/api/core/v1"
|
v1 "k8s.io/api/core/v1"
|
||||||
"k8s.io/apimachinery/pkg/api/resource"
|
"k8s.io/apimachinery/pkg/api/resource"
|
||||||
|
clientset "k8s.io/client-go/kubernetes"
|
||||||
"k8s.io/klog/v2"
|
"k8s.io/klog/v2"
|
||||||
|
"sigs.k8s.io/descheduler/pkg/api"
|
||||||
"sigs.k8s.io/descheduler/pkg/descheduler/evictions"
|
"sigs.k8s.io/descheduler/pkg/descheduler/evictions"
|
||||||
"sigs.k8s.io/descheduler/pkg/descheduler/node"
|
|
||||||
nodeutil "sigs.k8s.io/descheduler/pkg/descheduler/node"
|
|
||||||
podutil "sigs.k8s.io/descheduler/pkg/descheduler/pod"
|
podutil "sigs.k8s.io/descheduler/pkg/descheduler/pod"
|
||||||
"sigs.k8s.io/descheduler/pkg/framework"
|
|
||||||
"sigs.k8s.io/descheduler/pkg/utils"
|
"sigs.k8s.io/descheduler/pkg/utils"
|
||||||
|
"sort"
|
||||||
)
|
)
|
||||||
|
|
||||||
// NodeUsage stores a node's info, pods on it, thresholds and its resource usage
|
// NodeUsage stores a node's info, pods on it, thresholds and its resource usage
|
||||||
@@ -37,19 +35,12 @@ type NodeUsage struct {
|
|||||||
node *v1.Node
|
node *v1.Node
|
||||||
usage map[v1.ResourceName]*resource.Quantity
|
usage map[v1.ResourceName]*resource.Quantity
|
||||||
allPods []*v1.Pod
|
allPods []*v1.Pod
|
||||||
}
|
|
||||||
|
|
||||||
type NodeThresholds struct {
|
|
||||||
lowResourceThreshold map[v1.ResourceName]*resource.Quantity
|
lowResourceThreshold map[v1.ResourceName]*resource.Quantity
|
||||||
highResourceThreshold map[v1.ResourceName]*resource.Quantity
|
highResourceThreshold map[v1.ResourceName]*resource.Quantity
|
||||||
}
|
}
|
||||||
|
|
||||||
type NodeInfo struct {
|
type continueEvictionCond func(nodeUsage NodeUsage, totalAvailableUsage map[v1.ResourceName]*resource.Quantity) bool
|
||||||
NodeUsage
|
|
||||||
thresholds NodeThresholds
|
|
||||||
}
|
|
||||||
|
|
||||||
type continueEvictionCond func(nodeInfo NodeInfo, totalAvailableUsage map[v1.ResourceName]*resource.Quantity) bool
|
|
||||||
|
|
||||||
// NodePodsMap is a set of (node, pods) pairs
|
// NodePodsMap is a set of (node, pods) pairs
|
||||||
type NodePodsMap map[*v1.Node][]*v1.Pod
|
type NodePodsMap map[*v1.Node][]*v1.Pod
|
||||||
@@ -61,106 +52,88 @@ const (
|
|||||||
MaxResourcePercentage = 100
|
MaxResourcePercentage = 100
|
||||||
)
|
)
|
||||||
|
|
||||||
func normalizePercentage(percent api.Percentage) api.Percentage {
|
func validateNodeUtilizationParams(params *api.StrategyParameters) error {
|
||||||
if percent > MaxResourcePercentage {
|
if params == nil || params.NodeResourceUtilizationThresholds == nil {
|
||||||
return MaxResourcePercentage
|
return fmt.Errorf("NodeResourceUtilizationThresholds not set")
|
||||||
}
|
}
|
||||||
if percent < MinResourcePercentage {
|
if params.ThresholdPriority != nil && params.ThresholdPriorityClassName != "" {
|
||||||
return MinResourcePercentage
|
return fmt.Errorf("only one of thresholdPriority and thresholdPriorityClassName can be set")
|
||||||
}
|
}
|
||||||
return percent
|
|
||||||
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func getNodeThresholds(
|
// validateThresholds checks if thresholds have valid resource name and resource percentage configured
|
||||||
nodes []*v1.Node,
|
func validateThresholds(thresholds api.ResourceThresholds) error {
|
||||||
lowThreshold, highThreshold api.ResourceThresholds,
|
if thresholds == nil || len(thresholds) == 0 {
|
||||||
resourceNames []v1.ResourceName,
|
return fmt.Errorf("no resource threshold is configured")
|
||||||
getPodsAssignedToNode podutil.GetPodsAssignedToNodeFunc,
|
|
||||||
useDeviationThresholds bool,
|
|
||||||
) map[string]NodeThresholds {
|
|
||||||
nodeThresholdsMap := map[string]NodeThresholds{}
|
|
||||||
|
|
||||||
averageResourceUsagePercent := api.ResourceThresholds{}
|
|
||||||
if useDeviationThresholds {
|
|
||||||
averageResourceUsagePercent = averageNodeBasicresources(nodes, getPodsAssignedToNode, resourceNames)
|
|
||||||
}
|
}
|
||||||
|
for name, percent := range thresholds {
|
||||||
for _, node := range nodes {
|
if percent < MinResourcePercentage || percent > MaxResourcePercentage {
|
||||||
nodeCapacity := node.Status.Capacity
|
return fmt.Errorf("%v threshold not in [%v, %v] range", name, MinResourcePercentage, MaxResourcePercentage)
|
||||||
if len(node.Status.Allocatable) > 0 {
|
|
||||||
nodeCapacity = node.Status.Allocatable
|
|
||||||
}
|
}
|
||||||
|
|
||||||
nodeThresholdsMap[node.Name] = NodeThresholds{
|
|
||||||
lowResourceThreshold: map[v1.ResourceName]*resource.Quantity{},
|
|
||||||
highResourceThreshold: map[v1.ResourceName]*resource.Quantity{},
|
|
||||||
}
|
|
||||||
|
|
||||||
for _, resourceName := range resourceNames {
|
|
||||||
if useDeviationThresholds {
|
|
||||||
cap := nodeCapacity[resourceName]
|
|
||||||
if lowThreshold[resourceName] == MinResourcePercentage {
|
|
||||||
nodeThresholdsMap[node.Name].lowResourceThreshold[resourceName] = &cap
|
|
||||||
nodeThresholdsMap[node.Name].highResourceThreshold[resourceName] = &cap
|
|
||||||
} else {
|
|
||||||
nodeThresholdsMap[node.Name].lowResourceThreshold[resourceName] = resourceThreshold(nodeCapacity, resourceName, normalizePercentage(averageResourceUsagePercent[resourceName]-lowThreshold[resourceName]))
|
|
||||||
nodeThresholdsMap[node.Name].highResourceThreshold[resourceName] = resourceThreshold(nodeCapacity, resourceName, normalizePercentage(averageResourceUsagePercent[resourceName]+highThreshold[resourceName]))
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
nodeThresholdsMap[node.Name].lowResourceThreshold[resourceName] = resourceThreshold(nodeCapacity, resourceName, lowThreshold[resourceName])
|
|
||||||
nodeThresholdsMap[node.Name].highResourceThreshold[resourceName] = resourceThreshold(nodeCapacity, resourceName, highThreshold[resourceName])
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
}
|
}
|
||||||
return nodeThresholdsMap
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func getNodeUsage(
|
func getNodeUsage(
|
||||||
|
ctx context.Context,
|
||||||
|
client clientset.Interface,
|
||||||
nodes []*v1.Node,
|
nodes []*v1.Node,
|
||||||
|
lowThreshold, highThreshold api.ResourceThresholds,
|
||||||
resourceNames []v1.ResourceName,
|
resourceNames []v1.ResourceName,
|
||||||
getPodsAssignedToNode podutil.GetPodsAssignedToNodeFunc,
|
|
||||||
) []NodeUsage {
|
) []NodeUsage {
|
||||||
var nodeUsageList []NodeUsage
|
var nodeUsageList []NodeUsage
|
||||||
|
|
||||||
for _, node := range nodes {
|
for _, node := range nodes {
|
||||||
pods, err := podutil.ListPodsOnANode(node.Name, getPodsAssignedToNode, nil)
|
pods, err := podutil.ListPodsOnANode(ctx, client, node)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
klog.V(2).InfoS("Node will not be processed, error accessing its pods", "node", klog.KObj(node), "err", err)
|
klog.V(2).InfoS("Node will not be processed, error accessing its pods", "node", klog.KObj(node), "err", err)
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// A threshold is in percentages but in <0;100> interval.
|
||||||
|
// Performing `threshold * 0.01` will convert <0;100> interval into <0;1>.
|
||||||
|
// Multiplying it with capacity will give fraction of the capacity corresponding to the given high/low resource threshold in Quantity units.
|
||||||
|
nodeCapacity := node.Status.Capacity
|
||||||
|
if len(node.Status.Allocatable) > 0 {
|
||||||
|
nodeCapacity = node.Status.Allocatable
|
||||||
|
}
|
||||||
|
lowResourceThreshold := map[v1.ResourceName]*resource.Quantity{
|
||||||
|
v1.ResourceCPU: resource.NewMilliQuantity(int64(float64(lowThreshold[v1.ResourceCPU])*float64(nodeCapacity.Cpu().MilliValue())*0.01), resource.DecimalSI),
|
||||||
|
v1.ResourceMemory: resource.NewQuantity(int64(float64(lowThreshold[v1.ResourceMemory])*float64(nodeCapacity.Memory().Value())*0.01), resource.BinarySI),
|
||||||
|
v1.ResourcePods: resource.NewQuantity(int64(float64(lowThreshold[v1.ResourcePods])*float64(nodeCapacity.Pods().Value())*0.01), resource.DecimalSI),
|
||||||
|
}
|
||||||
|
for _, name := range resourceNames {
|
||||||
|
if !isBasicResource(name) {
|
||||||
|
cap := nodeCapacity[name]
|
||||||
|
lowResourceThreshold[name] = resource.NewQuantity(int64(float64(lowThreshold[name])*float64(cap.Value())*0.01), resource.DecimalSI)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
highResourceThreshold := map[v1.ResourceName]*resource.Quantity{
|
||||||
|
v1.ResourceCPU: resource.NewMilliQuantity(int64(float64(highThreshold[v1.ResourceCPU])*float64(nodeCapacity.Cpu().MilliValue())*0.01), resource.DecimalSI),
|
||||||
|
v1.ResourceMemory: resource.NewQuantity(int64(float64(highThreshold[v1.ResourceMemory])*float64(nodeCapacity.Memory().Value())*0.01), resource.BinarySI),
|
||||||
|
v1.ResourcePods: resource.NewQuantity(int64(float64(highThreshold[v1.ResourcePods])*float64(nodeCapacity.Pods().Value())*0.01), resource.DecimalSI),
|
||||||
|
}
|
||||||
|
for _, name := range resourceNames {
|
||||||
|
if !isBasicResource(name) {
|
||||||
|
cap := nodeCapacity[name]
|
||||||
|
highResourceThreshold[name] = resource.NewQuantity(int64(float64(highThreshold[name])*float64(cap.Value())*0.01), resource.DecimalSI)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
nodeUsageList = append(nodeUsageList, NodeUsage{
|
nodeUsageList = append(nodeUsageList, NodeUsage{
|
||||||
node: node,
|
node: node,
|
||||||
usage: nodeutil.NodeUtilization(pods, resourceNames),
|
usage: nodeUtilization(node, pods, resourceNames),
|
||||||
allPods: pods,
|
allPods: pods,
|
||||||
|
lowResourceThreshold: lowResourceThreshold,
|
||||||
|
highResourceThreshold: highResourceThreshold,
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
return nodeUsageList
|
return nodeUsageList
|
||||||
}
|
}
|
||||||
|
|
||||||
func resourceThreshold(nodeCapacity v1.ResourceList, resourceName v1.ResourceName, threshold api.Percentage) *resource.Quantity {
|
|
||||||
defaultFormat := resource.DecimalSI
|
|
||||||
if resourceName == v1.ResourceMemory {
|
|
||||||
defaultFormat = resource.BinarySI
|
|
||||||
}
|
|
||||||
|
|
||||||
resourceCapacityFraction := func(resourceNodeCapacity int64) int64 {
|
|
||||||
// A threshold is in percentages but in <0;100> interval.
|
|
||||||
// Performing `threshold * 0.01` will convert <0;100> interval into <0;1>.
|
|
||||||
// Multiplying it with capacity will give fraction of the capacity corresponding to the given resource threshold in Quantity units.
|
|
||||||
return int64(float64(threshold) * 0.01 * float64(resourceNodeCapacity))
|
|
||||||
}
|
|
||||||
|
|
||||||
resourceCapacityQuantity := nodeCapacity.Name(resourceName, defaultFormat)
|
|
||||||
|
|
||||||
if resourceName == v1.ResourceCPU {
|
|
||||||
return resource.NewMilliQuantity(resourceCapacityFraction(resourceCapacityQuantity.MilliValue()), defaultFormat)
|
|
||||||
}
|
|
||||||
return resource.NewQuantity(resourceCapacityFraction(resourceCapacityQuantity.Value()), defaultFormat)
|
|
||||||
}
|
|
||||||
|
|
||||||
func resourceUsagePercentages(nodeUsage NodeUsage) map[v1.ResourceName]float64 {
|
func resourceUsagePercentages(nodeUsage NodeUsage) map[v1.ResourceName]float64 {
|
||||||
nodeCapacity := nodeUsage.node.Status.Capacity
|
nodeCapacity := nodeUsage.node.Status.Capacity
|
||||||
if len(nodeUsage.node.Status.Allocatable) > 0 {
|
if len(nodeUsage.node.Status.Allocatable) > 0 {
|
||||||
@@ -182,24 +155,19 @@ func resourceUsagePercentages(nodeUsage NodeUsage) map[v1.ResourceName]float64 {
|
|||||||
// low and high thresholds, it is simply ignored.
|
// low and high thresholds, it is simply ignored.
|
||||||
func classifyNodes(
|
func classifyNodes(
|
||||||
nodeUsages []NodeUsage,
|
nodeUsages []NodeUsage,
|
||||||
nodeThresholds map[string]NodeThresholds,
|
lowThresholdFilter, highThresholdFilter func(node *v1.Node, usage NodeUsage) bool,
|
||||||
lowThresholdFilter, highThresholdFilter func(node *v1.Node, usage NodeUsage, threshold NodeThresholds) bool,
|
) ([]NodeUsage, []NodeUsage) {
|
||||||
) ([]NodeInfo, []NodeInfo) {
|
lowNodes, highNodes := []NodeUsage{}, []NodeUsage{}
|
||||||
lowNodes, highNodes := []NodeInfo{}, []NodeInfo{}
|
|
||||||
|
|
||||||
for _, nodeUsage := range nodeUsages {
|
for _, nodeUsage := range nodeUsages {
|
||||||
nodeInfo := NodeInfo{
|
if lowThresholdFilter(nodeUsage.node, nodeUsage) {
|
||||||
NodeUsage: nodeUsage,
|
klog.V(2).InfoS("Node is underutilized", "node", klog.KObj(nodeUsage.node), "usage", nodeUsage.usage, "usagePercentage", resourceUsagePercentages(nodeUsage))
|
||||||
thresholds: nodeThresholds[nodeUsage.node.Name],
|
lowNodes = append(lowNodes, nodeUsage)
|
||||||
}
|
} else if highThresholdFilter(nodeUsage.node, nodeUsage) {
|
||||||
if lowThresholdFilter(nodeUsage.node, nodeUsage, nodeThresholds[nodeUsage.node.Name]) {
|
klog.V(2).InfoS("Node is overutilized", "node", klog.KObj(nodeUsage.node), "usage", nodeUsage.usage, "usagePercentage", resourceUsagePercentages(nodeUsage))
|
||||||
klog.InfoS("Node is underutilized", "node", klog.KObj(nodeUsage.node), "usage", nodeUsage.usage, "usagePercentage", resourceUsagePercentages(nodeUsage))
|
highNodes = append(highNodes, nodeUsage)
|
||||||
lowNodes = append(lowNodes, nodeInfo)
|
|
||||||
} else if highThresholdFilter(nodeUsage.node, nodeUsage, nodeThresholds[nodeUsage.node.Name]) {
|
|
||||||
klog.InfoS("Node is overutilized", "node", klog.KObj(nodeUsage.node), "usage", nodeUsage.usage, "usagePercentage", resourceUsagePercentages(nodeUsage))
|
|
||||||
highNodes = append(highNodes, nodeInfo)
|
|
||||||
} else {
|
} else {
|
||||||
klog.InfoS("Node is appropriately utilized", "node", klog.KObj(nodeUsage.node), "usage", nodeUsage.usage, "usagePercentage", resourceUsagePercentages(nodeUsage))
|
klog.V(2).InfoS("Node is appropriately utilized", "node", klog.KObj(nodeUsage.node), "usage", nodeUsage.usage, "usagePercentage", resourceUsagePercentages(nodeUsage))
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -211,12 +179,16 @@ func classifyNodes(
|
|||||||
// TODO: @ravig Break this function into smaller functions.
|
// TODO: @ravig Break this function into smaller functions.
|
||||||
func evictPodsFromSourceNodes(
|
func evictPodsFromSourceNodes(
|
||||||
ctx context.Context,
|
ctx context.Context,
|
||||||
sourceNodes, destinationNodes []NodeInfo,
|
sourceNodes, destinationNodes []NodeUsage,
|
||||||
podEvictor framework.Evictor,
|
podEvictor *evictions.PodEvictor,
|
||||||
podFilter func(pod *v1.Pod) bool,
|
podFilter func(pod *v1.Pod) bool,
|
||||||
resourceNames []v1.ResourceName,
|
resourceNames []v1.ResourceName,
|
||||||
|
strategy string,
|
||||||
continueEviction continueEvictionCond,
|
continueEviction continueEvictionCond,
|
||||||
) {
|
) {
|
||||||
|
|
||||||
|
sortNodesByUsage(sourceNodes)
|
||||||
|
|
||||||
// upper bound on total number of pods/cpu/memory and optional extended resources to be moved
|
// upper bound on total number of pods/cpu/memory and optional extended resources to be moved
|
||||||
totalAvailableUsage := map[v1.ResourceName]*resource.Quantity{
|
totalAvailableUsage := map[v1.ResourceName]*resource.Quantity{
|
||||||
v1.ResourcePods: {},
|
v1.ResourcePods: {},
|
||||||
@@ -232,7 +204,7 @@ func evictPodsFromSourceNodes(
|
|||||||
if _, ok := totalAvailableUsage[name]; !ok {
|
if _, ok := totalAvailableUsage[name]; !ok {
|
||||||
totalAvailableUsage[name] = resource.NewQuantity(0, resource.DecimalSI)
|
totalAvailableUsage[name] = resource.NewQuantity(0, resource.DecimalSI)
|
||||||
}
|
}
|
||||||
totalAvailableUsage[name].Add(*node.thresholds.highResourceThreshold[name])
|
totalAvailableUsage[name].Add(*node.highResourceThreshold[name])
|
||||||
totalAvailableUsage[name].Sub(*node.usage[name])
|
totalAvailableUsage[name].Sub(*node.usage[name])
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -244,7 +216,7 @@ func evictPodsFromSourceNodes(
|
|||||||
"Pods", totalAvailableUsage[v1.ResourcePods].Value(),
|
"Pods", totalAvailableUsage[v1.ResourcePods].Value(),
|
||||||
}
|
}
|
||||||
for name := range totalAvailableUsage {
|
for name := range totalAvailableUsage {
|
||||||
if !node.IsBasicResource(name) {
|
if !isBasicResource(name) {
|
||||||
keysAndValues = append(keysAndValues, string(name), totalAvailableUsage[name].Value())
|
keysAndValues = append(keysAndValues, string(name), totalAvailableUsage[name].Value())
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -264,97 +236,96 @@ func evictPodsFromSourceNodes(
|
|||||||
klog.V(1).InfoS("Evicting pods based on priority, if they have same priority, they'll be evicted based on QoS tiers")
|
klog.V(1).InfoS("Evicting pods based on priority, if they have same priority, they'll be evicted based on QoS tiers")
|
||||||
// sort the evictable Pods based on priority. This also sorts them based on QoS. If there are multiple pods with same priority, they are sorted based on QoS tiers.
|
// sort the evictable Pods based on priority. This also sorts them based on QoS. If there are multiple pods with same priority, they are sorted based on QoS tiers.
|
||||||
podutil.SortPodsBasedOnPriorityLowToHigh(removablePods)
|
podutil.SortPodsBasedOnPriorityLowToHigh(removablePods)
|
||||||
evictPods(ctx, removablePods, node, totalAvailableUsage, taintsOfDestinationNodes, podEvictor, continueEviction)
|
evictPods(ctx, removablePods, node, totalAvailableUsage, taintsOfDestinationNodes, podEvictor, strategy, continueEviction)
|
||||||
|
klog.V(1).InfoS("Evicted pods from node", "node", klog.KObj(node.node), "evictedPods", podEvictor.NodeEvicted(node.node), "usage", node.usage)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func evictPods(
|
func evictPods(
|
||||||
ctx context.Context,
|
ctx context.Context,
|
||||||
inputPods []*v1.Pod,
|
inputPods []*v1.Pod,
|
||||||
nodeInfo NodeInfo,
|
nodeUsage NodeUsage,
|
||||||
totalAvailableUsage map[v1.ResourceName]*resource.Quantity,
|
totalAvailableUsage map[v1.ResourceName]*resource.Quantity,
|
||||||
taintsOfLowNodes map[string][]v1.Taint,
|
taintsOfLowNodes map[string][]v1.Taint,
|
||||||
podEvictor framework.Evictor,
|
podEvictor *evictions.PodEvictor,
|
||||||
|
strategy string,
|
||||||
continueEviction continueEvictionCond,
|
continueEviction continueEvictionCond,
|
||||||
) {
|
) {
|
||||||
|
|
||||||
if continueEviction(nodeInfo, totalAvailableUsage) {
|
if continueEviction(nodeUsage, totalAvailableUsage) {
|
||||||
for _, pod := range inputPods {
|
for _, pod := range inputPods {
|
||||||
if !utils.PodToleratesTaints(pod, taintsOfLowNodes) {
|
if !utils.PodToleratesTaints(pod, taintsOfLowNodes) {
|
||||||
klog.V(3).InfoS("Skipping eviction for pod, doesn't tolerate node taint", "pod", klog.KObj(pod))
|
klog.V(3).InfoS("Skipping eviction for pod, doesn't tolerate node taint", "pod", klog.KObj(pod))
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
|
|
||||||
if podEvictor.Evict(ctx, pod, evictions.EvictOptions{}) {
|
success, err := podEvictor.EvictPod(ctx, pod, nodeUsage.node, strategy)
|
||||||
klog.V(3).InfoS("Evicted pods", "pod", klog.KObj(pod))
|
if err != nil {
|
||||||
|
klog.ErrorS(err, "Error evicting pod", "pod", klog.KObj(pod))
|
||||||
|
break
|
||||||
|
}
|
||||||
|
|
||||||
|
if success {
|
||||||
|
klog.V(3).InfoS("Evicted pods", "pod", klog.KObj(pod), "err", err)
|
||||||
|
|
||||||
for name := range totalAvailableUsage {
|
for name := range totalAvailableUsage {
|
||||||
if name == v1.ResourcePods {
|
if name == v1.ResourcePods {
|
||||||
nodeInfo.usage[name].Sub(*resource.NewQuantity(1, resource.DecimalSI))
|
nodeUsage.usage[name].Sub(*resource.NewQuantity(1, resource.DecimalSI))
|
||||||
totalAvailableUsage[name].Sub(*resource.NewQuantity(1, resource.DecimalSI))
|
totalAvailableUsage[name].Sub(*resource.NewQuantity(1, resource.DecimalSI))
|
||||||
} else {
|
} else {
|
||||||
quantity := utils.GetResourceRequestQuantity(pod, name)
|
quantity := utils.GetResourceRequestQuantity(pod, name)
|
||||||
nodeInfo.usage[name].Sub(quantity)
|
nodeUsage.usage[name].Sub(quantity)
|
||||||
totalAvailableUsage[name].Sub(quantity)
|
totalAvailableUsage[name].Sub(quantity)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
keysAndValues := []interface{}{
|
keysAndValues := []interface{}{
|
||||||
"node", nodeInfo.node.Name,
|
"node", nodeUsage.node.Name,
|
||||||
"CPU", nodeInfo.usage[v1.ResourceCPU].MilliValue(),
|
"CPU", nodeUsage.usage[v1.ResourceCPU].MilliValue(),
|
||||||
"Mem", nodeInfo.usage[v1.ResourceMemory].Value(),
|
"Mem", nodeUsage.usage[v1.ResourceMemory].Value(),
|
||||||
"Pods", nodeInfo.usage[v1.ResourcePods].Value(),
|
"Pods", nodeUsage.usage[v1.ResourcePods].Value(),
|
||||||
}
|
}
|
||||||
for name := range totalAvailableUsage {
|
for name := range totalAvailableUsage {
|
||||||
if !nodeutil.IsBasicResource(name) {
|
if !isBasicResource(name) {
|
||||||
keysAndValues = append(keysAndValues, string(name), totalAvailableUsage[name].Value())
|
keysAndValues = append(keysAndValues, string(name), totalAvailableUsage[name].Value())
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
klog.V(3).InfoS("Updated node usage", keysAndValues...)
|
klog.V(3).InfoS("Updated node usage", keysAndValues...)
|
||||||
// check if pods can be still evicted
|
// check if pods can be still evicted
|
||||||
if !continueEviction(nodeInfo, totalAvailableUsage) {
|
if !continueEviction(nodeUsage, totalAvailableUsage) {
|
||||||
break
|
break
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if podEvictor.NodeLimitExceeded(nodeInfo.node) {
|
|
||||||
return
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// sortNodesByUsage sorts nodes based on usage according to the given plugin.
|
// sortNodesByUsage sorts nodes based on usage in descending order
|
||||||
func sortNodesByUsage(nodes []NodeInfo, ascending bool) {
|
func sortNodesByUsage(nodes []NodeUsage) {
|
||||||
sort.Slice(nodes, func(i, j int) bool {
|
sort.Slice(nodes, func(i, j int) bool {
|
||||||
ti := nodes[i].usage[v1.ResourceMemory].Value() + nodes[i].usage[v1.ResourceCPU].MilliValue() + nodes[i].usage[v1.ResourcePods].Value()
|
ti := nodes[i].usage[v1.ResourceMemory].Value() + nodes[i].usage[v1.ResourceCPU].MilliValue() + nodes[i].usage[v1.ResourcePods].Value()
|
||||||
tj := nodes[j].usage[v1.ResourceMemory].Value() + nodes[j].usage[v1.ResourceCPU].MilliValue() + nodes[j].usage[v1.ResourcePods].Value()
|
tj := nodes[j].usage[v1.ResourceMemory].Value() + nodes[j].usage[v1.ResourceCPU].MilliValue() + nodes[j].usage[v1.ResourcePods].Value()
|
||||||
|
|
||||||
// extended resources
|
// extended resources
|
||||||
for name := range nodes[i].usage {
|
for name := range nodes[i].usage {
|
||||||
if !nodeutil.IsBasicResource(name) {
|
if !isBasicResource(name) {
|
||||||
ti = ti + nodes[i].usage[name].Value()
|
ti = ti + nodes[i].usage[name].Value()
|
||||||
tj = tj + nodes[j].usage[name].Value()
|
tj = tj + nodes[j].usage[name].Value()
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Return ascending order for HighNodeUtilization plugin
|
// To return sorted in descending order
|
||||||
if ascending {
|
|
||||||
return ti < tj
|
|
||||||
}
|
|
||||||
|
|
||||||
// Return descending order for LowNodeUtilization plugin
|
|
||||||
return ti > tj
|
return ti > tj
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
// isNodeAboveTargetUtilization checks if a node is overutilized
|
// isNodeAboveTargetUtilization checks if a node is overutilized
|
||||||
// At least one resource has to be above the high threshold
|
// At least one resource has to be above the high threshold
|
||||||
func isNodeAboveTargetUtilization(usage NodeUsage, threshold map[v1.ResourceName]*resource.Quantity) bool {
|
func isNodeAboveTargetUtilization(usage NodeUsage) bool {
|
||||||
for name, nodeValue := range usage.usage {
|
for name, nodeValue := range usage.usage {
|
||||||
// usage.highResourceThreshold[name] < nodeValue
|
// usage.highResourceThreshold[name] < nodeValue
|
||||||
if threshold[name].Cmp(*nodeValue) == -1 {
|
if usage.highResourceThreshold[name].Cmp(*nodeValue) == -1 {
|
||||||
return true
|
return true
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -363,10 +334,10 @@ func isNodeAboveTargetUtilization(usage NodeUsage, threshold map[v1.ResourceName
|
|||||||
|
|
||||||
// isNodeWithLowUtilization checks if a node is underutilized
|
// isNodeWithLowUtilization checks if a node is underutilized
|
||||||
// All resources have to be below the low threshold
|
// All resources have to be below the low threshold
|
||||||
func isNodeWithLowUtilization(usage NodeUsage, threshold map[v1.ResourceName]*resource.Quantity) bool {
|
func isNodeWithLowUtilization(usage NodeUsage) bool {
|
||||||
for name, nodeValue := range usage.usage {
|
for name, nodeValue := range usage.usage {
|
||||||
// usage.lowResourceThreshold[name] < nodeValue
|
// usage.lowResourceThreshold[name] < nodeValue
|
||||||
if threshold[name].Cmp(*nodeValue) == -1 {
|
if usage.lowResourceThreshold[name].Cmp(*nodeValue) == -1 {
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -383,6 +354,43 @@ func getResourceNames(thresholds api.ResourceThresholds) []v1.ResourceName {
|
|||||||
return resourceNames
|
return resourceNames
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// isBasicResource checks if resource is basic native.
|
||||||
|
func isBasicResource(name v1.ResourceName) bool {
|
||||||
|
switch name {
|
||||||
|
case v1.ResourceCPU, v1.ResourceMemory, v1.ResourcePods:
|
||||||
|
return true
|
||||||
|
default:
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func nodeUtilization(node *v1.Node, pods []*v1.Pod, resourceNames []v1.ResourceName) map[v1.ResourceName]*resource.Quantity {
|
||||||
|
totalReqs := map[v1.ResourceName]*resource.Quantity{
|
||||||
|
v1.ResourceCPU: resource.NewMilliQuantity(0, resource.DecimalSI),
|
||||||
|
v1.ResourceMemory: resource.NewQuantity(0, resource.BinarySI),
|
||||||
|
v1.ResourcePods: resource.NewQuantity(int64(len(pods)), resource.DecimalSI),
|
||||||
|
}
|
||||||
|
for _, name := range resourceNames {
|
||||||
|
if !isBasicResource(name) {
|
||||||
|
totalReqs[name] = resource.NewQuantity(0, resource.DecimalSI)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, pod := range pods {
|
||||||
|
req, _ := utils.PodRequestsAndLimits(pod)
|
||||||
|
for _, name := range resourceNames {
|
||||||
|
quantity, ok := req[name]
|
||||||
|
if ok && name != v1.ResourcePods {
|
||||||
|
// As Quantity.Add says: Add adds the provided y quantity to the current value. If the current value is zero,
|
||||||
|
// the format of the quantity will be updated to the format of y.
|
||||||
|
totalReqs[name].Add(quantity)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return totalReqs
|
||||||
|
}
|
||||||
|
|
||||||
func classifyPods(pods []*v1.Pod, filter func(pod *v1.Pod) bool) ([]*v1.Pod, []*v1.Pod) {
|
func classifyPods(pods []*v1.Pod, filter func(pod *v1.Pod) bool) ([]*v1.Pod, []*v1.Pod) {
|
||||||
var nonRemovablePods, removablePods []*v1.Pod
|
var nonRemovablePods, removablePods []*v1.Pod
|
||||||
|
|
||||||
@@ -396,34 +404,3 @@ func classifyPods(pods []*v1.Pod, filter func(pod *v1.Pod) bool) ([]*v1.Pod, []*
|
|||||||
|
|
||||||
return nonRemovablePods, removablePods
|
return nonRemovablePods, removablePods
|
||||||
}
|
}
|
||||||
|
|
||||||
func averageNodeBasicresources(nodes []*v1.Node, getPodsAssignedToNode podutil.GetPodsAssignedToNodeFunc, resourceNames []v1.ResourceName) api.ResourceThresholds {
|
|
||||||
total := api.ResourceThresholds{}
|
|
||||||
average := api.ResourceThresholds{}
|
|
||||||
numberOfNodes := len(nodes)
|
|
||||||
for _, node := range nodes {
|
|
||||||
pods, err := podutil.ListPodsOnANode(node.Name, getPodsAssignedToNode, nil)
|
|
||||||
if err != nil {
|
|
||||||
numberOfNodes--
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
usage := nodeutil.NodeUtilization(pods, resourceNames)
|
|
||||||
nodeCapacity := node.Status.Capacity
|
|
||||||
if len(node.Status.Allocatable) > 0 {
|
|
||||||
nodeCapacity = node.Status.Allocatable
|
|
||||||
}
|
|
||||||
for resource, value := range usage {
|
|
||||||
nodeCapacityValue := nodeCapacity[resource]
|
|
||||||
if resource == v1.ResourceCPU {
|
|
||||||
total[resource] += api.Percentage(value.MilliValue()) / api.Percentage(nodeCapacityValue.MilliValue()) * 100.0
|
|
||||||
} else {
|
|
||||||
total[resource] += api.Percentage(value.Value()) / api.Percentage(nodeCapacityValue.Value()) * 100.0
|
|
||||||
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
for resource, value := range total {
|
|
||||||
average[resource] = value / api.Percentage(numberOfNodes)
|
|
||||||
}
|
|
||||||
return average
|
|
||||||
}
|
|
||||||
@@ -0,0 +1,158 @@
|
|||||||
|
/*
|
||||||
|
Copyright 2021 The Kubernetes Authors.
|
||||||
|
|
||||||
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
you may not use this file except in compliance with the License.
|
||||||
|
You may obtain a copy of the License at
|
||||||
|
|
||||||
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
|
Unless required by applicable law or agreed to in writing, software
|
||||||
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
See the License for the specific language governing permissions and
|
||||||
|
limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package nodeutilization
|
||||||
|
|
||||||
|
import (
|
||||||
|
"fmt"
|
||||||
|
v1 "k8s.io/api/core/v1"
|
||||||
|
"k8s.io/apimachinery/pkg/api/resource"
|
||||||
|
"math"
|
||||||
|
"sigs.k8s.io/descheduler/pkg/api"
|
||||||
|
"testing"
|
||||||
|
)
|
||||||
|
|
||||||
|
var (
|
||||||
|
lowPriority = int32(0)
|
||||||
|
highPriority = int32(10000)
|
||||||
|
extendedResource = v1.ResourceName("example.com/foo")
|
||||||
|
)
|
||||||
|
|
||||||
|
func TestValidateThresholds(t *testing.T) {
|
||||||
|
tests := []struct {
|
||||||
|
name string
|
||||||
|
input api.ResourceThresholds
|
||||||
|
errInfo error
|
||||||
|
}{
|
||||||
|
{
|
||||||
|
name: "passing nil map for threshold",
|
||||||
|
input: nil,
|
||||||
|
errInfo: fmt.Errorf("no resource threshold is configured"),
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "passing no threshold",
|
||||||
|
input: api.ResourceThresholds{},
|
||||||
|
errInfo: fmt.Errorf("no resource threshold is configured"),
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "passing extended resource name other than cpu/memory/pods",
|
||||||
|
input: api.ResourceThresholds{
|
||||||
|
v1.ResourceCPU: 40,
|
||||||
|
extendedResource: 50,
|
||||||
|
},
|
||||||
|
errInfo: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "passing invalid resource value",
|
||||||
|
input: api.ResourceThresholds{
|
||||||
|
v1.ResourceCPU: 110,
|
||||||
|
v1.ResourceMemory: 80,
|
||||||
|
},
|
||||||
|
errInfo: fmt.Errorf("%v threshold not in [%v, %v] range", v1.ResourceCPU, MinResourcePercentage, MaxResourcePercentage),
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "passing a valid threshold with max and min resource value",
|
||||||
|
input: api.ResourceThresholds{
|
||||||
|
v1.ResourceCPU: 100,
|
||||||
|
v1.ResourceMemory: 0,
|
||||||
|
},
|
||||||
|
errInfo: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "passing a valid threshold with only cpu",
|
||||||
|
input: api.ResourceThresholds{
|
||||||
|
v1.ResourceCPU: 80,
|
||||||
|
},
|
||||||
|
errInfo: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "passing a valid threshold with cpu, memory and pods",
|
||||||
|
input: api.ResourceThresholds{
|
||||||
|
v1.ResourceCPU: 20,
|
||||||
|
v1.ResourceMemory: 30,
|
||||||
|
v1.ResourcePods: 40,
|
||||||
|
},
|
||||||
|
errInfo: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "passing a valid threshold with only extended resource",
|
||||||
|
input: api.ResourceThresholds{
|
||||||
|
extendedResource: 80,
|
||||||
|
},
|
||||||
|
errInfo: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "passing a valid threshold with cpu, memory, pods and extended resource",
|
||||||
|
input: api.ResourceThresholds{
|
||||||
|
v1.ResourceCPU: 20,
|
||||||
|
v1.ResourceMemory: 30,
|
||||||
|
v1.ResourcePods: 40,
|
||||||
|
extendedResource: 50,
|
||||||
|
},
|
||||||
|
errInfo: nil,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, test := range tests {
|
||||||
|
validateErr := validateThresholds(test.input)
|
||||||
|
|
||||||
|
if validateErr == nil || test.errInfo == nil {
|
||||||
|
if validateErr != test.errInfo {
|
||||||
|
t.Errorf("expected validity of threshold: %#v to be %v but got %v instead", test.input, test.errInfo, validateErr)
|
||||||
|
}
|
||||||
|
} else if validateErr.Error() != test.errInfo.Error() {
|
||||||
|
t.Errorf("expected validity of threshold: %#v to be %v but got %v instead", test.input, test.errInfo, validateErr)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestResourceUsagePercentages(t *testing.T) {
|
||||||
|
resourceUsagePercentage := resourceUsagePercentages(NodeUsage{
|
||||||
|
node: &v1.Node{
|
||||||
|
Status: v1.NodeStatus{
|
||||||
|
Capacity: v1.ResourceList{
|
||||||
|
v1.ResourceCPU: *resource.NewMilliQuantity(2000, resource.DecimalSI),
|
||||||
|
v1.ResourceMemory: *resource.NewQuantity(3977868*1024, resource.BinarySI),
|
||||||
|
v1.ResourcePods: *resource.NewQuantity(29, resource.BinarySI),
|
||||||
|
},
|
||||||
|
Allocatable: v1.ResourceList{
|
||||||
|
v1.ResourceCPU: *resource.NewMilliQuantity(1930, resource.DecimalSI),
|
||||||
|
v1.ResourceMemory: *resource.NewQuantity(3287692*1024, resource.BinarySI),
|
||||||
|
v1.ResourcePods: *resource.NewQuantity(29, resource.BinarySI),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
usage: map[v1.ResourceName]*resource.Quantity{
|
||||||
|
v1.ResourceCPU: resource.NewMilliQuantity(1220, resource.DecimalSI),
|
||||||
|
v1.ResourceMemory: resource.NewQuantity(3038982964, resource.BinarySI),
|
||||||
|
v1.ResourcePods: resource.NewQuantity(11, resource.BinarySI),
|
||||||
|
},
|
||||||
|
})
|
||||||
|
|
||||||
|
expectedUsageInIntPercentage := map[v1.ResourceName]float64{
|
||||||
|
v1.ResourceCPU: 63,
|
||||||
|
v1.ResourceMemory: 90,
|
||||||
|
v1.ResourcePods: 37,
|
||||||
|
}
|
||||||
|
|
||||||
|
for resourceName, percentage := range expectedUsageInIntPercentage {
|
||||||
|
if math.Floor(resourceUsagePercentage[resourceName]) != percentage {
|
||||||
|
t.Errorf("Incorrect percentange computation, expected %v, got math.Floor(%v) instead", percentage, resourceUsagePercentage[resourceName])
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
t.Logf("resourceUsagePercentage: %#v\n", resourceUsagePercentage)
|
||||||
|
}
|
||||||
@@ -14,86 +14,94 @@ See the License for the specific language governing permissions and
|
|||||||
limitations under the License.
|
limitations under the License.
|
||||||
*/
|
*/
|
||||||
|
|
||||||
package removepodsviolatinginterpodantiaffinity
|
package strategies
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
"fmt"
|
"fmt"
|
||||||
|
|
||||||
"k8s.io/apimachinery/pkg/runtime"
|
"sigs.k8s.io/descheduler/pkg/api"
|
||||||
"k8s.io/apimachinery/pkg/util/sets"
|
|
||||||
"sigs.k8s.io/descheduler/pkg/apis/componentconfig"
|
|
||||||
"sigs.k8s.io/descheduler/pkg/descheduler/evictions"
|
"sigs.k8s.io/descheduler/pkg/descheduler/evictions"
|
||||||
podutil "sigs.k8s.io/descheduler/pkg/descheduler/pod"
|
podutil "sigs.k8s.io/descheduler/pkg/descheduler/pod"
|
||||||
"sigs.k8s.io/descheduler/pkg/framework"
|
|
||||||
"sigs.k8s.io/descheduler/pkg/utils"
|
"sigs.k8s.io/descheduler/pkg/utils"
|
||||||
|
|
||||||
v1 "k8s.io/api/core/v1"
|
v1 "k8s.io/api/core/v1"
|
||||||
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
||||||
|
clientset "k8s.io/client-go/kubernetes"
|
||||||
"k8s.io/klog/v2"
|
"k8s.io/klog/v2"
|
||||||
)
|
)
|
||||||
|
|
||||||
const PluginName = "RemovePodsViolatingInterPodAntiAffinity"
|
func validateRemovePodsViolatingInterPodAntiAffinityParams(params *api.StrategyParameters) error {
|
||||||
|
if params == nil {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
// RemovePodsViolatingInterPodAntiAffinity evicts pods on the node which violate inter pod anti affinity
|
// At most one of include/exclude can be set
|
||||||
type RemovePodsViolatingInterPodAntiAffinity struct {
|
if params.Namespaces != nil && len(params.Namespaces.Include) > 0 && len(params.Namespaces.Exclude) > 0 {
|
||||||
handle framework.Handle
|
return fmt.Errorf("only one of Include/Exclude namespaces can be set")
|
||||||
args *componentconfig.RemovePodsViolatingInterPodAntiAffinityArgs
|
}
|
||||||
podFilter podutil.FilterFunc
|
if params.ThresholdPriority != nil && params.ThresholdPriorityClassName != "" {
|
||||||
|
return fmt.Errorf("only one of thresholdPriority and thresholdPriorityClassName can be set")
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
var _ framework.DeschedulePlugin = &RemovePodsViolatingInterPodAntiAffinity{}
|
// RemovePodsViolatingInterPodAntiAffinity evicts pods on the node which are having a pod affinity rules.
|
||||||
|
func RemovePodsViolatingInterPodAntiAffinity(ctx context.Context, client clientset.Interface, strategy api.DeschedulerStrategy, nodes []*v1.Node, podEvictor *evictions.PodEvictor) {
|
||||||
// New builds plugin from its arguments while passing a handle
|
if err := validateRemovePodsViolatingInterPodAntiAffinityParams(strategy.Params); err != nil {
|
||||||
func New(args runtime.Object, handle framework.Handle) (framework.Plugin, error) {
|
klog.ErrorS(err, "Invalid RemovePodsViolatingInterPodAntiAffinity parameters")
|
||||||
interPodAntiAffinityArgs, ok := args.(*componentconfig.RemovePodsViolatingInterPodAntiAffinityArgs)
|
return
|
||||||
if !ok {
|
|
||||||
return nil, fmt.Errorf("want args to be of type RemovePodsViolatingInterPodAntiAffinityArgs, got %T", args)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
var includedNamespaces, excludedNamespaces sets.String
|
var includedNamespaces, excludedNamespaces []string
|
||||||
if interPodAntiAffinityArgs.Namespaces != nil {
|
var labelSelector *metav1.LabelSelector
|
||||||
includedNamespaces = sets.NewString(interPodAntiAffinityArgs.Namespaces.Include...)
|
if strategy.Params != nil {
|
||||||
excludedNamespaces = sets.NewString(interPodAntiAffinityArgs.Namespaces.Exclude...)
|
if strategy.Params.Namespaces != nil {
|
||||||
|
includedNamespaces = strategy.Params.Namespaces.Include
|
||||||
|
excludedNamespaces = strategy.Params.Namespaces.Exclude
|
||||||
|
}
|
||||||
|
labelSelector = strategy.Params.LabelSelector
|
||||||
}
|
}
|
||||||
|
|
||||||
podFilter, err := podutil.NewOptions().
|
thresholdPriority, err := utils.GetPriorityFromStrategyParams(ctx, client, strategy.Params)
|
||||||
WithNamespaces(includedNamespaces).
|
|
||||||
WithoutNamespaces(excludedNamespaces).
|
|
||||||
WithLabelSelector(interPodAntiAffinityArgs.LabelSelector).
|
|
||||||
BuildFilterFunc()
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, fmt.Errorf("error initializing pod filter function: %v", err)
|
klog.ErrorS(err, "Failed to get threshold priority from strategy's params")
|
||||||
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
return &RemovePodsViolatingInterPodAntiAffinity{
|
nodeFit := false
|
||||||
handle: handle,
|
if strategy.Params != nil {
|
||||||
podFilter: podFilter,
|
nodeFit = strategy.Params.NodeFit
|
||||||
args: interPodAntiAffinityArgs,
|
}
|
||||||
}, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// Name retrieves the plugin name
|
evictable := podEvictor.Evictable(evictions.WithPriorityThreshold(thresholdPriority), evictions.WithNodeFit(nodeFit))
|
||||||
func (d *RemovePodsViolatingInterPodAntiAffinity) Name() string {
|
|
||||||
return PluginName
|
|
||||||
}
|
|
||||||
|
|
||||||
func (d *RemovePodsViolatingInterPodAntiAffinity) Deschedule(ctx context.Context, nodes []*v1.Node) *framework.Status {
|
|
||||||
loop:
|
|
||||||
for _, node := range nodes {
|
for _, node := range nodes {
|
||||||
klog.V(1).InfoS("Processing node", "node", klog.KObj(node))
|
klog.V(1).InfoS("Processing node", "node", klog.KObj(node))
|
||||||
pods, err := podutil.ListPodsOnANode(node.Name, d.handle.GetPodsAssignedToNodeFunc(), d.podFilter)
|
pods, err := podutil.ListPodsOnANode(
|
||||||
|
ctx,
|
||||||
|
client,
|
||||||
|
node,
|
||||||
|
podutil.WithNamespaces(includedNamespaces),
|
||||||
|
podutil.WithoutNamespaces(excludedNamespaces),
|
||||||
|
podutil.WithLabelSelector(labelSelector),
|
||||||
|
)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return &framework.Status{
|
return
|
||||||
Err: fmt.Errorf("error listing pods: %v", err),
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
// sort the evictable Pods based on priority, if there are multiple pods with same priority, they are sorted based on QoS tiers.
|
// sort the evictable Pods based on priority, if there are multiple pods with same priority, they are sorted based on QoS tiers.
|
||||||
podutil.SortPodsBasedOnPriorityLowToHigh(pods)
|
podutil.SortPodsBasedOnPriorityLowToHigh(pods)
|
||||||
totalPods := len(pods)
|
totalPods := len(pods)
|
||||||
for i := 0; i < totalPods; i++ {
|
for i := 0; i < totalPods; i++ {
|
||||||
if checkPodsWithAntiAffinityExist(pods[i], pods) && d.handle.Evictor().Filter(pods[i]) {
|
if checkPodsWithAntiAffinityExist(pods[i], pods) && evictable.IsEvictable(pods[i]) {
|
||||||
if d.handle.Evictor().Evict(ctx, pods[i], evictions.EvictOptions{}) {
|
success, err := podEvictor.EvictPod(ctx, pods[i], node, "InterPodAntiAffinity")
|
||||||
|
if err != nil {
|
||||||
|
klog.ErrorS(err, "Error evicting pod")
|
||||||
|
break
|
||||||
|
}
|
||||||
|
|
||||||
|
if success {
|
||||||
// Since the current pod is evicted all other pods which have anti-affinity with this
|
// Since the current pod is evicted all other pods which have anti-affinity with this
|
||||||
// pod need not be evicted.
|
// pod need not be evicted.
|
||||||
// Update pods.
|
// Update pods.
|
||||||
@@ -102,12 +110,8 @@ loop:
|
|||||||
totalPods--
|
totalPods--
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if d.handle.Evictor().NodeLimitExceeded(node) {
|
|
||||||
continue loop
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
return nil
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// checkPodsWithAntiAffinityExist checks if there are other pods on the node that the current pod cannot tolerate.
|
// checkPodsWithAntiAffinityExist checks if there are other pods on the node that the current pod cannot tolerate.
|
||||||
208
pkg/descheduler/strategies/pod_antiaffinity_test.go
Normal file
208
pkg/descheduler/strategies/pod_antiaffinity_test.go
Normal file
@@ -0,0 +1,208 @@
|
|||||||
|
/*
|
||||||
|
Copyright 2017 The Kubernetes Authors.
|
||||||
|
|
||||||
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
you may not use this file except in compliance with the License.
|
||||||
|
You may obtain a copy of the License at
|
||||||
|
|
||||||
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
|
Unless required by applicable law or agreed to in writing, software
|
||||||
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
See the License for the specific language governing permissions and
|
||||||
|
limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package strategies
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"testing"
|
||||||
|
|
||||||
|
v1 "k8s.io/api/core/v1"
|
||||||
|
policyv1 "k8s.io/api/policy/v1"
|
||||||
|
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
||||||
|
"k8s.io/apimachinery/pkg/runtime"
|
||||||
|
"k8s.io/client-go/kubernetes/fake"
|
||||||
|
core "k8s.io/client-go/testing"
|
||||||
|
"sigs.k8s.io/descheduler/pkg/api"
|
||||||
|
"sigs.k8s.io/descheduler/pkg/descheduler/evictions"
|
||||||
|
"sigs.k8s.io/descheduler/pkg/utils"
|
||||||
|
"sigs.k8s.io/descheduler/test"
|
||||||
|
)
|
||||||
|
|
||||||
|
func TestPodAntiAffinity(t *testing.T) {
|
||||||
|
ctx := context.Background()
|
||||||
|
node1 := test.BuildTestNode("n1", 2000, 3000, 10, nil)
|
||||||
|
node2 := test.BuildTestNode("n2", 2000, 3000, 10, func(node *v1.Node) {
|
||||||
|
node.ObjectMeta.Labels = map[string]string{
|
||||||
|
"datacenter": "east",
|
||||||
|
}
|
||||||
|
})
|
||||||
|
|
||||||
|
node3 := test.BuildTestNode("n3", 2000, 3000, 10, func(node *v1.Node) {
|
||||||
|
node.Spec = v1.NodeSpec{
|
||||||
|
Unschedulable: true,
|
||||||
|
}
|
||||||
|
})
|
||||||
|
|
||||||
|
p1 := test.BuildTestPod("p1", 100, 0, node1.Name, nil)
|
||||||
|
p2 := test.BuildTestPod("p2", 100, 0, node1.Name, nil)
|
||||||
|
p3 := test.BuildTestPod("p3", 100, 0, node1.Name, nil)
|
||||||
|
p4 := test.BuildTestPod("p4", 100, 0, node1.Name, nil)
|
||||||
|
p5 := test.BuildTestPod("p5", 100, 0, node1.Name, nil)
|
||||||
|
p6 := test.BuildTestPod("p6", 100, 0, node1.Name, nil)
|
||||||
|
p7 := test.BuildTestPod("p7", 100, 0, node1.Name, nil)
|
||||||
|
p8 := test.BuildTestPod("p8", 100, 0, node1.Name, nil)
|
||||||
|
|
||||||
|
criticalPriority := utils.SystemCriticalPriority
|
||||||
|
nonEvictablePod := test.BuildTestPod("non-evict", 100, 0, node1.Name, func(pod *v1.Pod) {
|
||||||
|
pod.Spec.Priority = &criticalPriority
|
||||||
|
})
|
||||||
|
p2.Labels = map[string]string{"foo": "bar"}
|
||||||
|
p5.Labels = map[string]string{"foo": "bar"}
|
||||||
|
p6.Labels = map[string]string{"foo": "bar"}
|
||||||
|
p7.Labels = map[string]string{"foo1": "bar1"}
|
||||||
|
nonEvictablePod.Labels = map[string]string{"foo": "bar"}
|
||||||
|
test.SetNormalOwnerRef(p1)
|
||||||
|
test.SetNormalOwnerRef(p2)
|
||||||
|
test.SetNormalOwnerRef(p3)
|
||||||
|
test.SetNormalOwnerRef(p4)
|
||||||
|
test.SetNormalOwnerRef(p5)
|
||||||
|
test.SetNormalOwnerRef(p6)
|
||||||
|
test.SetNormalOwnerRef(p7)
|
||||||
|
|
||||||
|
// set pod anti affinity
|
||||||
|
setPodAntiAffinity(p1, "foo", "bar")
|
||||||
|
setPodAntiAffinity(p3, "foo", "bar")
|
||||||
|
setPodAntiAffinity(p4, "foo", "bar")
|
||||||
|
setPodAntiAffinity(p5, "foo1", "bar1")
|
||||||
|
setPodAntiAffinity(p6, "foo1", "bar1")
|
||||||
|
setPodAntiAffinity(p7, "foo", "bar")
|
||||||
|
|
||||||
|
// set pod priority
|
||||||
|
test.SetPodPriority(p5, 100)
|
||||||
|
test.SetPodPriority(p6, 50)
|
||||||
|
test.SetPodPriority(p7, 0)
|
||||||
|
|
||||||
|
// Set pod node selectors
|
||||||
|
p8.Spec.NodeSelector = map[string]string{
|
||||||
|
"datacenter": "west",
|
||||||
|
}
|
||||||
|
|
||||||
|
tests := []struct {
|
||||||
|
description string
|
||||||
|
maxPodsToEvictPerNode int
|
||||||
|
pods []v1.Pod
|
||||||
|
expectedEvictedPodCount int
|
||||||
|
nodeFit bool
|
||||||
|
nodes []*v1.Node
|
||||||
|
}{
|
||||||
|
{
|
||||||
|
description: "Maximum pods to evict - 0",
|
||||||
|
maxPodsToEvictPerNode: 0,
|
||||||
|
pods: []v1.Pod{*p1, *p2, *p3, *p4},
|
||||||
|
nodes: []*v1.Node{node1},
|
||||||
|
expectedEvictedPodCount: 3,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
description: "Maximum pods to evict - 3",
|
||||||
|
maxPodsToEvictPerNode: 3,
|
||||||
|
pods: []v1.Pod{*p1, *p2, *p3, *p4},
|
||||||
|
nodes: []*v1.Node{node1},
|
||||||
|
expectedEvictedPodCount: 3,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
description: "Evict only 1 pod after sorting",
|
||||||
|
maxPodsToEvictPerNode: 0,
|
||||||
|
pods: []v1.Pod{*p5, *p6, *p7},
|
||||||
|
nodes: []*v1.Node{node1},
|
||||||
|
expectedEvictedPodCount: 1,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
description: "Evicts pod that conflicts with critical pod (but does not evict critical pod)",
|
||||||
|
maxPodsToEvictPerNode: 1,
|
||||||
|
pods: []v1.Pod{*p1, *nonEvictablePod},
|
||||||
|
nodes: []*v1.Node{node1},
|
||||||
|
expectedEvictedPodCount: 1,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
description: "Evicts pod that conflicts with critical pod (but does not evict critical pod)",
|
||||||
|
maxPodsToEvictPerNode: 1,
|
||||||
|
pods: []v1.Pod{*p1, *nonEvictablePod},
|
||||||
|
nodes: []*v1.Node{node1},
|
||||||
|
expectedEvictedPodCount: 1,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
description: "Won't evict pods because node selectors don't match available nodes",
|
||||||
|
maxPodsToEvictPerNode: 1,
|
||||||
|
pods: []v1.Pod{*p8, *nonEvictablePod},
|
||||||
|
nodes: []*v1.Node{node1, node2},
|
||||||
|
expectedEvictedPodCount: 0,
|
||||||
|
nodeFit: true,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
description: "Won't evict pods because only other node is not schedulable",
|
||||||
|
maxPodsToEvictPerNode: 1,
|
||||||
|
pods: []v1.Pod{*p8, *nonEvictablePod},
|
||||||
|
nodes: []*v1.Node{node1, node3},
|
||||||
|
expectedEvictedPodCount: 0,
|
||||||
|
nodeFit: true,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, test := range tests {
|
||||||
|
// create fake client
|
||||||
|
fakeClient := &fake.Clientset{}
|
||||||
|
fakeClient.Fake.AddReactor("list", "pods", func(action core.Action) (bool, runtime.Object, error) {
|
||||||
|
return true, &v1.PodList{Items: test.pods}, nil
|
||||||
|
})
|
||||||
|
fakeClient.Fake.AddReactor("get", "nodes", func(action core.Action) (bool, runtime.Object, error) {
|
||||||
|
return true, node1, nil
|
||||||
|
})
|
||||||
|
|
||||||
|
podEvictor := evictions.NewPodEvictor(
|
||||||
|
fakeClient,
|
||||||
|
policyv1.SchemeGroupVersion.String(),
|
||||||
|
false,
|
||||||
|
test.maxPodsToEvictPerNode,
|
||||||
|
test.nodes,
|
||||||
|
false,
|
||||||
|
false,
|
||||||
|
false,
|
||||||
|
)
|
||||||
|
strategy := api.DeschedulerStrategy{
|
||||||
|
Params: &api.StrategyParameters{
|
||||||
|
NodeFit: test.nodeFit,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
RemovePodsViolatingInterPodAntiAffinity(ctx, fakeClient, strategy, test.nodes, podEvictor)
|
||||||
|
podsEvicted := podEvictor.TotalEvicted()
|
||||||
|
if podsEvicted != test.expectedEvictedPodCount {
|
||||||
|
t.Errorf("Unexpected no of pods evicted: pods evicted: %d, expected: %d", podsEvicted, test.expectedEvictedPodCount)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func setPodAntiAffinity(inputPod *v1.Pod, labelKey, labelValue string) {
|
||||||
|
inputPod.Spec.Affinity = &v1.Affinity{
|
||||||
|
PodAntiAffinity: &v1.PodAntiAffinity{
|
||||||
|
RequiredDuringSchedulingIgnoredDuringExecution: []v1.PodAffinityTerm{
|
||||||
|
{
|
||||||
|
LabelSelector: &metav1.LabelSelector{
|
||||||
|
MatchExpressions: []metav1.LabelSelectorRequirement{
|
||||||
|
{
|
||||||
|
Key: labelKey,
|
||||||
|
Operator: metav1.LabelSelectorOpIn,
|
||||||
|
Values: []string{labelValue},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
TopologyKey: "region",
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
}
|
||||||
141
pkg/descheduler/strategies/pod_lifetime.go
Normal file
141
pkg/descheduler/strategies/pod_lifetime.go
Normal file
@@ -0,0 +1,141 @@
|
|||||||
|
/*
|
||||||
|
Copyright 2020 The Kubernetes Authors.
|
||||||
|
|
||||||
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
you may not use this file except in compliance with the License.
|
||||||
|
You may obtain a copy of the License at
|
||||||
|
|
||||||
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
|
Unless required by applicable law or agreed to in writing, software
|
||||||
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
See the License for the specific language governing permissions and
|
||||||
|
limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package strategies
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"fmt"
|
||||||
|
|
||||||
|
v1 "k8s.io/api/core/v1"
|
||||||
|
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
||||||
|
clientset "k8s.io/client-go/kubernetes"
|
||||||
|
"k8s.io/klog/v2"
|
||||||
|
|
||||||
|
"sigs.k8s.io/descheduler/pkg/api"
|
||||||
|
"sigs.k8s.io/descheduler/pkg/descheduler/evictions"
|
||||||
|
podutil "sigs.k8s.io/descheduler/pkg/descheduler/pod"
|
||||||
|
"sigs.k8s.io/descheduler/pkg/utils"
|
||||||
|
)
|
||||||
|
|
||||||
|
func validatePodLifeTimeParams(params *api.StrategyParameters) error {
|
||||||
|
if params == nil || params.PodLifeTime == nil || params.PodLifeTime.MaxPodLifeTimeSeconds == nil {
|
||||||
|
return fmt.Errorf("MaxPodLifeTimeSeconds not set")
|
||||||
|
}
|
||||||
|
|
||||||
|
if params.PodLifeTime.PodStatusPhases != nil {
|
||||||
|
for _, phase := range params.PodLifeTime.PodStatusPhases {
|
||||||
|
if phase != string(v1.PodPending) && phase != string(v1.PodRunning) {
|
||||||
|
return fmt.Errorf("only Pending and Running phases are supported in PodLifeTime")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// At most one of include/exclude can be set
|
||||||
|
if params.Namespaces != nil && len(params.Namespaces.Include) > 0 && len(params.Namespaces.Exclude) > 0 {
|
||||||
|
return fmt.Errorf("only one of Include/Exclude namespaces can be set")
|
||||||
|
}
|
||||||
|
if params.ThresholdPriority != nil && params.ThresholdPriorityClassName != "" {
|
||||||
|
return fmt.Errorf("only one of thresholdPriority and thresholdPriorityClassName can be set")
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// PodLifeTime evicts pods on nodes that were created more than strategy.Params.MaxPodLifeTimeSeconds seconds ago.
|
||||||
|
func PodLifeTime(ctx context.Context, client clientset.Interface, strategy api.DeschedulerStrategy, nodes []*v1.Node, podEvictor *evictions.PodEvictor) {
|
||||||
|
if err := validatePodLifeTimeParams(strategy.Params); err != nil {
|
||||||
|
klog.ErrorS(err, "Invalid PodLifeTime parameters")
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
thresholdPriority, err := utils.GetPriorityFromStrategyParams(ctx, client, strategy.Params)
|
||||||
|
if err != nil {
|
||||||
|
klog.ErrorS(err, "Failed to get threshold priority from strategy's params")
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
var includedNamespaces, excludedNamespaces []string
|
||||||
|
if strategy.Params.Namespaces != nil {
|
||||||
|
includedNamespaces = strategy.Params.Namespaces.Include
|
||||||
|
excludedNamespaces = strategy.Params.Namespaces.Exclude
|
||||||
|
}
|
||||||
|
|
||||||
|
evictable := podEvictor.Evictable(evictions.WithPriorityThreshold(thresholdPriority))
|
||||||
|
|
||||||
|
filter := evictable.IsEvictable
|
||||||
|
if strategy.Params.PodLifeTime.PodStatusPhases != nil {
|
||||||
|
filter = func(pod *v1.Pod) bool {
|
||||||
|
for _, phase := range strategy.Params.PodLifeTime.PodStatusPhases {
|
||||||
|
if string(pod.Status.Phase) == phase {
|
||||||
|
return evictable.IsEvictable(pod)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, node := range nodes {
|
||||||
|
klog.V(1).InfoS("Processing node", "node", klog.KObj(node))
|
||||||
|
|
||||||
|
pods := listOldPodsOnNode(ctx, client, node, includedNamespaces, excludedNamespaces, strategy.Params.LabelSelector, *strategy.Params.PodLifeTime.MaxPodLifeTimeSeconds, filter)
|
||||||
|
for _, pod := range pods {
|
||||||
|
success, err := podEvictor.EvictPod(ctx, pod, node, "PodLifeTime")
|
||||||
|
if success {
|
||||||
|
klog.V(1).InfoS("Evicted pod because it exceeded its lifetime", "pod", klog.KObj(pod), "maxPodLifeTime", *strategy.Params.PodLifeTime.MaxPodLifeTimeSeconds)
|
||||||
|
}
|
||||||
|
|
||||||
|
if err != nil {
|
||||||
|
klog.ErrorS(err, "Error evicting pod", "pod", klog.KObj(pod))
|
||||||
|
break
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func listOldPodsOnNode(
|
||||||
|
ctx context.Context,
|
||||||
|
client clientset.Interface,
|
||||||
|
node *v1.Node,
|
||||||
|
includedNamespaces, excludedNamespaces []string,
|
||||||
|
labelSelector *metav1.LabelSelector,
|
||||||
|
maxPodLifeTimeSeconds uint,
|
||||||
|
filter func(pod *v1.Pod) bool,
|
||||||
|
) []*v1.Pod {
|
||||||
|
pods, err := podutil.ListPodsOnANode(
|
||||||
|
ctx,
|
||||||
|
client,
|
||||||
|
node,
|
||||||
|
podutil.WithFilter(filter),
|
||||||
|
podutil.WithNamespaces(includedNamespaces),
|
||||||
|
podutil.WithoutNamespaces(excludedNamespaces),
|
||||||
|
podutil.WithLabelSelector(labelSelector),
|
||||||
|
)
|
||||||
|
if err != nil {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
var oldPods []*v1.Pod
|
||||||
|
for _, pod := range pods {
|
||||||
|
podAgeSeconds := uint(metav1.Now().Sub(pod.GetCreationTimestamp().Local()).Seconds())
|
||||||
|
if podAgeSeconds > maxPodLifeTimeSeconds {
|
||||||
|
oldPods = append(oldPods, pod)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return oldPods
|
||||||
|
}
|
||||||
276
pkg/descheduler/strategies/pod_lifetime_test.go
Normal file
276
pkg/descheduler/strategies/pod_lifetime_test.go
Normal file
@@ -0,0 +1,276 @@
|
|||||||
|
/*
|
||||||
|
Copyright 2020 The Kubernetes Authors.
|
||||||
|
|
||||||
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
you may not use this file except in compliance with the License.
|
||||||
|
You may obtain a copy of the License at
|
||||||
|
|
||||||
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
|
Unless required by applicable law or agreed to in writing, software
|
||||||
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
See the License for the specific language governing permissions and
|
||||||
|
limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package strategies
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"testing"
|
||||||
|
"time"
|
||||||
|
|
||||||
|
v1 "k8s.io/api/core/v1"
|
||||||
|
policyv1 "k8s.io/api/policy/v1"
|
||||||
|
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
||||||
|
"k8s.io/apimachinery/pkg/runtime"
|
||||||
|
"k8s.io/client-go/kubernetes/fake"
|
||||||
|
core "k8s.io/client-go/testing"
|
||||||
|
"sigs.k8s.io/descheduler/pkg/api"
|
||||||
|
"sigs.k8s.io/descheduler/pkg/descheduler/evictions"
|
||||||
|
"sigs.k8s.io/descheduler/test"
|
||||||
|
)
|
||||||
|
|
||||||
|
func TestPodLifeTime(t *testing.T) {
|
||||||
|
ctx := context.Background()
|
||||||
|
node1 := test.BuildTestNode("n1", 2000, 3000, 10, nil)
|
||||||
|
olderPodCreationTime := metav1.NewTime(time.Date(2009, time.November, 10, 23, 0, 0, 0, time.UTC))
|
||||||
|
newerPodCreationTime := metav1.NewTime(time.Now())
|
||||||
|
|
||||||
|
// Setup pods, one should be evicted
|
||||||
|
p1 := test.BuildTestPod("p1", 100, 0, node1.Name, nil)
|
||||||
|
p1.Namespace = "dev"
|
||||||
|
p1.ObjectMeta.CreationTimestamp = newerPodCreationTime
|
||||||
|
p2 := test.BuildTestPod("p2", 100, 0, node1.Name, nil)
|
||||||
|
p2.Namespace = "dev"
|
||||||
|
p2.ObjectMeta.CreationTimestamp = olderPodCreationTime
|
||||||
|
|
||||||
|
ownerRef1 := test.GetReplicaSetOwnerRefList()
|
||||||
|
p1.ObjectMeta.OwnerReferences = ownerRef1
|
||||||
|
p2.ObjectMeta.OwnerReferences = ownerRef1
|
||||||
|
|
||||||
|
// Setup pods, zero should be evicted
|
||||||
|
p3 := test.BuildTestPod("p3", 100, 0, node1.Name, nil)
|
||||||
|
p3.Namespace = "dev"
|
||||||
|
p3.ObjectMeta.CreationTimestamp = newerPodCreationTime
|
||||||
|
p4 := test.BuildTestPod("p4", 100, 0, node1.Name, nil)
|
||||||
|
p4.Namespace = "dev"
|
||||||
|
p4.ObjectMeta.CreationTimestamp = newerPodCreationTime
|
||||||
|
|
||||||
|
ownerRef2 := test.GetReplicaSetOwnerRefList()
|
||||||
|
p3.ObjectMeta.OwnerReferences = ownerRef2
|
||||||
|
p4.ObjectMeta.OwnerReferences = ownerRef2
|
||||||
|
|
||||||
|
// Setup pods, one should be evicted
|
||||||
|
p5 := test.BuildTestPod("p5", 100, 0, node1.Name, nil)
|
||||||
|
p5.Namespace = "dev"
|
||||||
|
p5.ObjectMeta.CreationTimestamp = newerPodCreationTime
|
||||||
|
p6 := test.BuildTestPod("p6", 100, 0, node1.Name, nil)
|
||||||
|
p6.Namespace = "dev"
|
||||||
|
p6.ObjectMeta.CreationTimestamp = metav1.NewTime(time.Now().Add(time.Second * 605))
|
||||||
|
|
||||||
|
ownerRef3 := test.GetReplicaSetOwnerRefList()
|
||||||
|
p5.ObjectMeta.OwnerReferences = ownerRef3
|
||||||
|
p6.ObjectMeta.OwnerReferences = ownerRef3
|
||||||
|
|
||||||
|
// Setup pods, zero should be evicted
|
||||||
|
p7 := test.BuildTestPod("p7", 100, 0, node1.Name, nil)
|
||||||
|
p7.Namespace = "dev"
|
||||||
|
p7.ObjectMeta.CreationTimestamp = newerPodCreationTime
|
||||||
|
p8 := test.BuildTestPod("p8", 100, 0, node1.Name, nil)
|
||||||
|
p8.Namespace = "dev"
|
||||||
|
p8.ObjectMeta.CreationTimestamp = metav1.NewTime(time.Now().Add(time.Second * 595))
|
||||||
|
|
||||||
|
ownerRef4 := test.GetReplicaSetOwnerRefList()
|
||||||
|
p5.ObjectMeta.OwnerReferences = ownerRef4
|
||||||
|
p6.ObjectMeta.OwnerReferences = ownerRef4
|
||||||
|
|
||||||
|
// Setup two old pods with different status phases
|
||||||
|
p9 := test.BuildTestPod("p9", 100, 0, node1.Name, nil)
|
||||||
|
p9.Namespace = "dev"
|
||||||
|
p9.ObjectMeta.CreationTimestamp = olderPodCreationTime
|
||||||
|
p10 := test.BuildTestPod("p10", 100, 0, node1.Name, nil)
|
||||||
|
p10.Namespace = "dev"
|
||||||
|
p10.ObjectMeta.CreationTimestamp = olderPodCreationTime
|
||||||
|
|
||||||
|
p9.Status.Phase = "Pending"
|
||||||
|
p10.Status.Phase = "Running"
|
||||||
|
p9.ObjectMeta.OwnerReferences = ownerRef1
|
||||||
|
p10.ObjectMeta.OwnerReferences = ownerRef1
|
||||||
|
|
||||||
|
p11 := test.BuildTestPod("p11", 100, 0, node1.Name, func(pod *v1.Pod) {
|
||||||
|
pod.Spec.Volumes = []v1.Volume{
|
||||||
|
{
|
||||||
|
Name: "pvc", VolumeSource: v1.VolumeSource{
|
||||||
|
PersistentVolumeClaim: &v1.PersistentVolumeClaimVolumeSource{ClaimName: "foo"},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
pod.Namespace = "dev"
|
||||||
|
pod.ObjectMeta.CreationTimestamp = olderPodCreationTime
|
||||||
|
pod.ObjectMeta.OwnerReferences = ownerRef1
|
||||||
|
})
|
||||||
|
|
||||||
|
// Setup two old pods with different labels
|
||||||
|
p12 := test.BuildTestPod("p12", 100, 0, node1.Name, nil)
|
||||||
|
p12.Namespace = "dev"
|
||||||
|
p12.ObjectMeta.CreationTimestamp = olderPodCreationTime
|
||||||
|
p13 := test.BuildTestPod("p13", 100, 0, node1.Name, nil)
|
||||||
|
p13.Namespace = "dev"
|
||||||
|
p13.ObjectMeta.CreationTimestamp = olderPodCreationTime
|
||||||
|
|
||||||
|
p12.ObjectMeta.Labels = map[string]string{"foo": "bar"}
|
||||||
|
p13.ObjectMeta.Labels = map[string]string{"foo": "bar1"}
|
||||||
|
p12.ObjectMeta.OwnerReferences = ownerRef1
|
||||||
|
p13.ObjectMeta.OwnerReferences = ownerRef1
|
||||||
|
|
||||||
|
var maxLifeTime uint = 600
|
||||||
|
testCases := []struct {
|
||||||
|
description string
|
||||||
|
strategy api.DeschedulerStrategy
|
||||||
|
maxPodsToEvictPerNode int
|
||||||
|
pods []v1.Pod
|
||||||
|
nodes []*v1.Node
|
||||||
|
expectedEvictedPodCount int
|
||||||
|
ignorePvcPods bool
|
||||||
|
}{
|
||||||
|
{
|
||||||
|
description: "Two pods in the `dev` Namespace, 1 is new and 1 very is old. 1 should be evicted.",
|
||||||
|
strategy: api.DeschedulerStrategy{
|
||||||
|
Enabled: true,
|
||||||
|
Params: &api.StrategyParameters{
|
||||||
|
PodLifeTime: &api.PodLifeTime{MaxPodLifeTimeSeconds: &maxLifeTime},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
maxPodsToEvictPerNode: 5,
|
||||||
|
pods: []v1.Pod{*p1, *p2},
|
||||||
|
nodes: []*v1.Node{node1},
|
||||||
|
expectedEvictedPodCount: 1,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
description: "Two pods in the `dev` Namespace, 2 are new and 0 are old. 0 should be evicted.",
|
||||||
|
strategy: api.DeschedulerStrategy{
|
||||||
|
Enabled: true,
|
||||||
|
Params: &api.StrategyParameters{
|
||||||
|
PodLifeTime: &api.PodLifeTime{MaxPodLifeTimeSeconds: &maxLifeTime},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
maxPodsToEvictPerNode: 5,
|
||||||
|
pods: []v1.Pod{*p3, *p4},
|
||||||
|
nodes: []*v1.Node{node1},
|
||||||
|
expectedEvictedPodCount: 0,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
description: "Two pods in the `dev` Namespace, 1 created 605 seconds ago. 1 should be evicted.",
|
||||||
|
strategy: api.DeschedulerStrategy{
|
||||||
|
Enabled: true,
|
||||||
|
Params: &api.StrategyParameters{
|
||||||
|
PodLifeTime: &api.PodLifeTime{MaxPodLifeTimeSeconds: &maxLifeTime},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
maxPodsToEvictPerNode: 5,
|
||||||
|
pods: []v1.Pod{*p5, *p6},
|
||||||
|
nodes: []*v1.Node{node1},
|
||||||
|
expectedEvictedPodCount: 1,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
description: "Two pods in the `dev` Namespace, 1 created 595 seconds ago. 0 should be evicted.",
|
||||||
|
strategy: api.DeschedulerStrategy{
|
||||||
|
Enabled: true,
|
||||||
|
Params: &api.StrategyParameters{
|
||||||
|
PodLifeTime: &api.PodLifeTime{MaxPodLifeTimeSeconds: &maxLifeTime},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
maxPodsToEvictPerNode: 5,
|
||||||
|
pods: []v1.Pod{*p7, *p8},
|
||||||
|
nodes: []*v1.Node{node1},
|
||||||
|
expectedEvictedPodCount: 0,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
description: "Two old pods with different status phases. 1 should be evicted.",
|
||||||
|
strategy: api.DeschedulerStrategy{
|
||||||
|
Enabled: true,
|
||||||
|
Params: &api.StrategyParameters{
|
||||||
|
PodLifeTime: &api.PodLifeTime{
|
||||||
|
MaxPodLifeTimeSeconds: &maxLifeTime,
|
||||||
|
PodStatusPhases: []string{"Pending"},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
maxPodsToEvictPerNode: 5,
|
||||||
|
pods: []v1.Pod{*p9, *p10},
|
||||||
|
nodes: []*v1.Node{node1},
|
||||||
|
expectedEvictedPodCount: 1,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
description: "does not evict pvc pods with ignorePvcPods set to true",
|
||||||
|
strategy: api.DeschedulerStrategy{
|
||||||
|
Enabled: true,
|
||||||
|
Params: &api.StrategyParameters{
|
||||||
|
PodLifeTime: &api.PodLifeTime{MaxPodLifeTimeSeconds: &maxLifeTime},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
maxPodsToEvictPerNode: 5,
|
||||||
|
pods: []v1.Pod{*p11},
|
||||||
|
nodes: []*v1.Node{node1},
|
||||||
|
expectedEvictedPodCount: 0,
|
||||||
|
ignorePvcPods: true,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
description: "evicts pvc pods with ignorePvcPods set to false (or unset)",
|
||||||
|
strategy: api.DeschedulerStrategy{
|
||||||
|
Enabled: true,
|
||||||
|
Params: &api.StrategyParameters{
|
||||||
|
PodLifeTime: &api.PodLifeTime{MaxPodLifeTimeSeconds: &maxLifeTime},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
maxPodsToEvictPerNode: 5,
|
||||||
|
pods: []v1.Pod{*p11},
|
||||||
|
nodes: []*v1.Node{node1},
|
||||||
|
expectedEvictedPodCount: 1,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
description: "Two old pods with different labels, 1 selected by labelSelector",
|
||||||
|
strategy: api.DeschedulerStrategy{
|
||||||
|
Enabled: true,
|
||||||
|
Params: &api.StrategyParameters{
|
||||||
|
PodLifeTime: &api.PodLifeTime{MaxPodLifeTimeSeconds: &maxLifeTime},
|
||||||
|
LabelSelector: &metav1.LabelSelector{
|
||||||
|
MatchLabels: map[string]string{"foo": "bar"},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
maxPodsToEvictPerNode: 5,
|
||||||
|
pods: []v1.Pod{*p12, *p13},
|
||||||
|
nodes: []*v1.Node{node1},
|
||||||
|
expectedEvictedPodCount: 1,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, tc := range testCases {
|
||||||
|
fakeClient := &fake.Clientset{}
|
||||||
|
|
||||||
|
fakeClient.Fake.AddReactor("list", "pods", func(action core.Action) (bool, runtime.Object, error) {
|
||||||
|
return true, &v1.PodList{Items: tc.pods}, nil
|
||||||
|
})
|
||||||
|
|
||||||
|
podEvictor := evictions.NewPodEvictor(
|
||||||
|
fakeClient,
|
||||||
|
policyv1.SchemeGroupVersion.String(),
|
||||||
|
false,
|
||||||
|
tc.maxPodsToEvictPerNode,
|
||||||
|
tc.nodes,
|
||||||
|
false,
|
||||||
|
false,
|
||||||
|
tc.ignorePvcPods,
|
||||||
|
)
|
||||||
|
|
||||||
|
PodLifeTime(ctx, fakeClient, tc.strategy, tc.nodes, podEvictor)
|
||||||
|
podsEvicted := podEvictor.TotalEvicted()
|
||||||
|
if podsEvicted != tc.expectedEvictedPodCount {
|
||||||
|
t.Errorf("Test error for description: %s. Expected evicted pods count %v, got %v", tc.description, tc.expectedEvictedPodCount, podsEvicted)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
123
pkg/descheduler/strategies/toomanyrestarts.go
Normal file
123
pkg/descheduler/strategies/toomanyrestarts.go
Normal file
@@ -0,0 +1,123 @@
|
|||||||
|
/*
|
||||||
|
Copyright 2018 The Kubernetes Authors.
|
||||||
|
|
||||||
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
you may not use this file except in compliance with the License.
|
||||||
|
You may obtain a copy of the License at
|
||||||
|
|
||||||
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
|
Unless required by applicable law or agreed to in writing, software
|
||||||
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
See the License for the specific language governing permissions and
|
||||||
|
limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package strategies
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"fmt"
|
||||||
|
|
||||||
|
v1 "k8s.io/api/core/v1"
|
||||||
|
clientset "k8s.io/client-go/kubernetes"
|
||||||
|
"k8s.io/klog/v2"
|
||||||
|
|
||||||
|
"sigs.k8s.io/descheduler/pkg/api"
|
||||||
|
"sigs.k8s.io/descheduler/pkg/descheduler/evictions"
|
||||||
|
podutil "sigs.k8s.io/descheduler/pkg/descheduler/pod"
|
||||||
|
"sigs.k8s.io/descheduler/pkg/utils"
|
||||||
|
)
|
||||||
|
|
||||||
|
func validateRemovePodsHavingTooManyRestartsParams(params *api.StrategyParameters) error {
|
||||||
|
if params == nil || params.PodsHavingTooManyRestarts == nil || params.PodsHavingTooManyRestarts.PodRestartThreshold < 1 {
|
||||||
|
return fmt.Errorf("PodsHavingTooManyRestarts threshold not set")
|
||||||
|
}
|
||||||
|
|
||||||
|
// At most one of include/exclude can be set
|
||||||
|
if params.Namespaces != nil && len(params.Namespaces.Include) > 0 && len(params.Namespaces.Exclude) > 0 {
|
||||||
|
return fmt.Errorf("only one of Include/Exclude namespaces can be set")
|
||||||
|
}
|
||||||
|
if params.ThresholdPriority != nil && params.ThresholdPriorityClassName != "" {
|
||||||
|
return fmt.Errorf("only one of thresholdPriority and thresholdPriorityClassName can be set")
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// RemovePodsHavingTooManyRestarts removes the pods that have too many restarts on node.
|
||||||
|
// There are too many cases leading this issue: Volume mount failed, app error due to nodes' different settings.
|
||||||
|
// As of now, this strategy won't evict daemonsets, mirror pods, critical pods and pods with local storages.
|
||||||
|
func RemovePodsHavingTooManyRestarts(ctx context.Context, client clientset.Interface, strategy api.DeschedulerStrategy, nodes []*v1.Node, podEvictor *evictions.PodEvictor) {
|
||||||
|
if err := validateRemovePodsHavingTooManyRestartsParams(strategy.Params); err != nil {
|
||||||
|
klog.ErrorS(err, "Invalid RemovePodsHavingTooManyRestarts parameters")
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
thresholdPriority, err := utils.GetPriorityFromStrategyParams(ctx, client, strategy.Params)
|
||||||
|
if err != nil {
|
||||||
|
klog.ErrorS(err, "Failed to get threshold priority from strategy's params")
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
var includedNamespaces, excludedNamespaces []string
|
||||||
|
if strategy.Params.Namespaces != nil {
|
||||||
|
includedNamespaces = strategy.Params.Namespaces.Include
|
||||||
|
excludedNamespaces = strategy.Params.Namespaces.Exclude
|
||||||
|
}
|
||||||
|
|
||||||
|
nodeFit := false
|
||||||
|
if strategy.Params != nil {
|
||||||
|
nodeFit = strategy.Params.NodeFit
|
||||||
|
}
|
||||||
|
|
||||||
|
evictable := podEvictor.Evictable(evictions.WithPriorityThreshold(thresholdPriority), evictions.WithNodeFit(nodeFit))
|
||||||
|
|
||||||
|
for _, node := range nodes {
|
||||||
|
klog.V(1).InfoS("Processing node", "node", klog.KObj(node))
|
||||||
|
pods, err := podutil.ListPodsOnANode(
|
||||||
|
ctx,
|
||||||
|
client,
|
||||||
|
node,
|
||||||
|
podutil.WithFilter(evictable.IsEvictable),
|
||||||
|
podutil.WithNamespaces(includedNamespaces),
|
||||||
|
podutil.WithoutNamespaces(excludedNamespaces),
|
||||||
|
podutil.WithLabelSelector(strategy.Params.LabelSelector),
|
||||||
|
)
|
||||||
|
if err != nil {
|
||||||
|
klog.ErrorS(err, "Error listing a nodes pods", "node", klog.KObj(node))
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
|
for i, pod := range pods {
|
||||||
|
restarts, initRestarts := calcContainerRestarts(pod)
|
||||||
|
if strategy.Params.PodsHavingTooManyRestarts.IncludingInitContainers {
|
||||||
|
if restarts+initRestarts < strategy.Params.PodsHavingTooManyRestarts.PodRestartThreshold {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
} else if restarts < strategy.Params.PodsHavingTooManyRestarts.PodRestartThreshold {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
if _, err := podEvictor.EvictPod(ctx, pods[i], node, "TooManyRestarts"); err != nil {
|
||||||
|
klog.ErrorS(err, "Error evicting pod", "pod", klog.KObj(pod))
|
||||||
|
break
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// calcContainerRestarts get container restarts and init container restarts.
|
||||||
|
func calcContainerRestarts(pod *v1.Pod) (int32, int32) {
|
||||||
|
var restarts, initRestarts int32
|
||||||
|
|
||||||
|
for _, cs := range pod.Status.ContainerStatuses {
|
||||||
|
restarts += cs.RestartCount
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, cs := range pod.Status.InitContainerStatuses {
|
||||||
|
initRestarts += cs.RestartCount
|
||||||
|
}
|
||||||
|
|
||||||
|
return restarts, initRestarts
|
||||||
|
}
|
||||||
228
pkg/descheduler/strategies/toomanyrestarts_test.go
Normal file
228
pkg/descheduler/strategies/toomanyrestarts_test.go
Normal file
@@ -0,0 +1,228 @@
|
|||||||
|
/*
|
||||||
|
Copyright 2018 The Kubernetes Authors.
|
||||||
|
|
||||||
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
you may not use this file except in compliance with the License.
|
||||||
|
You may obtain a copy of the License at
|
||||||
|
|
||||||
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
|
Unless required by applicable law or agreed to in writing, software
|
||||||
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
See the License for the specific language governing permissions and
|
||||||
|
limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package strategies
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"testing"
|
||||||
|
|
||||||
|
"fmt"
|
||||||
|
|
||||||
|
v1 "k8s.io/api/core/v1"
|
||||||
|
policyv1 "k8s.io/api/policy/v1"
|
||||||
|
"k8s.io/apimachinery/pkg/api/resource"
|
||||||
|
"k8s.io/apimachinery/pkg/runtime"
|
||||||
|
"k8s.io/client-go/kubernetes/fake"
|
||||||
|
core "k8s.io/client-go/testing"
|
||||||
|
"sigs.k8s.io/descheduler/pkg/api"
|
||||||
|
"sigs.k8s.io/descheduler/pkg/descheduler/evictions"
|
||||||
|
"sigs.k8s.io/descheduler/test"
|
||||||
|
)
|
||||||
|
|
||||||
|
func initPods(node *v1.Node) []v1.Pod {
|
||||||
|
pods := make([]v1.Pod, 0)
|
||||||
|
|
||||||
|
for i := int32(0); i <= 9; i++ {
|
||||||
|
pod := test.BuildTestPod(fmt.Sprintf("pod-%d", i), 100, 0, node.Name, nil)
|
||||||
|
pod.ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
||||||
|
|
||||||
|
// pod at index i will have 25 * i restarts.
|
||||||
|
pod.Status = v1.PodStatus{
|
||||||
|
InitContainerStatuses: []v1.ContainerStatus{
|
||||||
|
{
|
||||||
|
RestartCount: 5 * i,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
ContainerStatuses: []v1.ContainerStatus{
|
||||||
|
{
|
||||||
|
RestartCount: 10 * i,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
RestartCount: 10 * i,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
pods = append(pods, *pod)
|
||||||
|
}
|
||||||
|
|
||||||
|
// The following 3 pods won't get evicted.
|
||||||
|
// A daemonset.
|
||||||
|
pods[6].ObjectMeta.OwnerReferences = test.GetDaemonSetOwnerRefList()
|
||||||
|
// A pod with local storage.
|
||||||
|
pods[7].ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
||||||
|
pods[7].Spec.Volumes = []v1.Volume{
|
||||||
|
{
|
||||||
|
Name: "sample",
|
||||||
|
VolumeSource: v1.VolumeSource{
|
||||||
|
HostPath: &v1.HostPathVolumeSource{Path: "somePath"},
|
||||||
|
EmptyDir: &v1.EmptyDirVolumeSource{
|
||||||
|
SizeLimit: resource.NewQuantity(int64(10), resource.BinarySI)},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
// A Mirror Pod.
|
||||||
|
pods[8].Annotations = test.GetMirrorPodAnnotation()
|
||||||
|
|
||||||
|
return pods
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestRemovePodsHavingTooManyRestarts(t *testing.T) {
|
||||||
|
ctx := context.Background()
|
||||||
|
|
||||||
|
node1 := test.BuildTestNode("node1", 2000, 3000, 10, nil)
|
||||||
|
node2 := test.BuildTestNode("node2", 2000, 3000, 10, func(node *v1.Node) {
|
||||||
|
node.Spec.Taints = []v1.Taint{
|
||||||
|
{
|
||||||
|
Key: "hardware",
|
||||||
|
Value: "gpu",
|
||||||
|
Effect: v1.TaintEffectNoSchedule,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
})
|
||||||
|
node3 := test.BuildTestNode("node3", 2000, 3000, 10, func(node *v1.Node) {
|
||||||
|
node.Spec = v1.NodeSpec{
|
||||||
|
Unschedulable: true,
|
||||||
|
}
|
||||||
|
})
|
||||||
|
|
||||||
|
pods := initPods(node1)
|
||||||
|
|
||||||
|
createStrategy := func(enabled, includingInitContainers bool, restartThresholds int32, nodeFit bool) api.DeschedulerStrategy {
|
||||||
|
return api.DeschedulerStrategy{
|
||||||
|
Enabled: enabled,
|
||||||
|
Params: &api.StrategyParameters{
|
||||||
|
PodsHavingTooManyRestarts: &api.PodsHavingTooManyRestarts{
|
||||||
|
PodRestartThreshold: restartThresholds,
|
||||||
|
IncludingInitContainers: includingInitContainers,
|
||||||
|
},
|
||||||
|
NodeFit: nodeFit,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
tests := []struct {
|
||||||
|
description string
|
||||||
|
nodes []*v1.Node
|
||||||
|
strategy api.DeschedulerStrategy
|
||||||
|
expectedEvictedPodCount int
|
||||||
|
maxPodsToEvictPerNode int
|
||||||
|
}{
|
||||||
|
{
|
||||||
|
description: "All pods have total restarts under threshold, no pod evictions",
|
||||||
|
strategy: createStrategy(true, true, 10000, false),
|
||||||
|
nodes: []*v1.Node{node1},
|
||||||
|
expectedEvictedPodCount: 0,
|
||||||
|
maxPodsToEvictPerNode: 0,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
description: "Some pods have total restarts bigger than threshold",
|
||||||
|
strategy: createStrategy(true, true, 1, false),
|
||||||
|
nodes: []*v1.Node{node1},
|
||||||
|
expectedEvictedPodCount: 6,
|
||||||
|
maxPodsToEvictPerNode: 0,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
description: "Nine pods have total restarts equals threshold(includingInitContainers=true), 6 pod evictions",
|
||||||
|
strategy: createStrategy(true, true, 1*25, false),
|
||||||
|
nodes: []*v1.Node{node1},
|
||||||
|
expectedEvictedPodCount: 6,
|
||||||
|
maxPodsToEvictPerNode: 0,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
description: "Nine pods have total restarts equals threshold(includingInitContainers=false), 5 pod evictions",
|
||||||
|
strategy: createStrategy(true, false, 1*25, false),
|
||||||
|
nodes: []*v1.Node{node1},
|
||||||
|
expectedEvictedPodCount: 5,
|
||||||
|
maxPodsToEvictPerNode: 0,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
description: "All pods have total restarts equals threshold(includingInitContainers=true), 6 pod evictions",
|
||||||
|
strategy: createStrategy(true, true, 1*20, false),
|
||||||
|
nodes: []*v1.Node{node1},
|
||||||
|
expectedEvictedPodCount: 6,
|
||||||
|
maxPodsToEvictPerNode: 0,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
description: "Nine pods have total restarts equals threshold(includingInitContainers=false), 6 pod evictions",
|
||||||
|
strategy: createStrategy(true, false, 1*20, false),
|
||||||
|
nodes: []*v1.Node{node1},
|
||||||
|
expectedEvictedPodCount: 6,
|
||||||
|
maxPodsToEvictPerNode: 0,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
description: "Five pods have total restarts bigger than threshold(includingInitContainers=true), but only 1 pod eviction",
|
||||||
|
strategy: createStrategy(true, true, 5*25+1, false),
|
||||||
|
nodes: []*v1.Node{node1},
|
||||||
|
expectedEvictedPodCount: 1,
|
||||||
|
maxPodsToEvictPerNode: 0,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
description: "Five pods have total restarts bigger than threshold(includingInitContainers=false), but only 1 pod eviction",
|
||||||
|
strategy: createStrategy(true, false, 5*20+1, false),
|
||||||
|
nodes: []*v1.Node{node1},
|
||||||
|
expectedEvictedPodCount: 1,
|
||||||
|
maxPodsToEvictPerNode: 0,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
description: "All pods have total restarts equals threshold(maxPodsToEvictPerNode=3), 3 pod evictions",
|
||||||
|
strategy: createStrategy(true, true, 1, false),
|
||||||
|
nodes: []*v1.Node{node1},
|
||||||
|
expectedEvictedPodCount: 3,
|
||||||
|
maxPodsToEvictPerNode: 3,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
description: "All pods have total restarts equals threshold(maxPodsToEvictPerNode=3) but the only other node is tained, 0 pod evictions",
|
||||||
|
strategy: createStrategy(true, true, 1, true),
|
||||||
|
nodes: []*v1.Node{node1, node2},
|
||||||
|
expectedEvictedPodCount: 0,
|
||||||
|
maxPodsToEvictPerNode: 3,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
description: "All pods have total restarts equals threshold(maxPodsToEvictPerNode=3) but the only other node is not schedulable, 0 pod evictions",
|
||||||
|
strategy: createStrategy(true, true, 1, true),
|
||||||
|
nodes: []*v1.Node{node1, node3},
|
||||||
|
expectedEvictedPodCount: 0,
|
||||||
|
maxPodsToEvictPerNode: 3,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, tc := range tests {
|
||||||
|
|
||||||
|
fakeClient := &fake.Clientset{}
|
||||||
|
fakeClient.Fake.AddReactor("list", "pods", func(action core.Action) (bool, runtime.Object, error) {
|
||||||
|
return true, &v1.PodList{Items: pods}, nil
|
||||||
|
})
|
||||||
|
|
||||||
|
podEvictor := evictions.NewPodEvictor(
|
||||||
|
fakeClient,
|
||||||
|
policyv1.SchemeGroupVersion.String(),
|
||||||
|
false,
|
||||||
|
tc.maxPodsToEvictPerNode,
|
||||||
|
tc.nodes,
|
||||||
|
false,
|
||||||
|
false,
|
||||||
|
false,
|
||||||
|
)
|
||||||
|
|
||||||
|
RemovePodsHavingTooManyRestarts(ctx, fakeClient, tc.strategy, tc.nodes, podEvictor)
|
||||||
|
actualEvictedPodCount := podEvictor.TotalEvicted()
|
||||||
|
if actualEvictedPodCount != tc.expectedEvictedPodCount {
|
||||||
|
t.Errorf("Test %#v failed, expected %v pod evictions, but got %v pod evictions\n", tc.description, tc.expectedEvictedPodCount, actualEvictedPodCount)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
@@ -1,9 +1,12 @@
|
|||||||
/*
|
/*
|
||||||
Copyright 2020 The Kubernetes Authors.
|
Copyright 2020 The Kubernetes Authors.
|
||||||
|
|
||||||
Licensed under the Apache License, Version 2.0 (the "License");
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
you may not use this file except in compliance with the License.
|
you may not use this file except in compliance with the License.
|
||||||
You may obtain a copy of the License at
|
You may obtain a copy of the License at
|
||||||
|
|
||||||
http://www.apache.org/licenses/LICENSE-2.0
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
Unless required by applicable law or agreed to in writing, software
|
Unless required by applicable law or agreed to in writing, software
|
||||||
distributed under the License is distributed on an "AS IS" BASIS,
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
@@ -11,32 +14,29 @@ See the License for the specific language governing permissions and
|
|||||||
limitations under the License.
|
limitations under the License.
|
||||||
*/
|
*/
|
||||||
|
|
||||||
package removepodsviolatingtopologyspreadconstraint
|
package strategies
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
"fmt"
|
"fmt"
|
||||||
"math"
|
"math"
|
||||||
"reflect"
|
|
||||||
"sort"
|
"sort"
|
||||||
|
|
||||||
|
utilerrors "k8s.io/apimachinery/pkg/util/errors"
|
||||||
|
|
||||||
v1 "k8s.io/api/core/v1"
|
v1 "k8s.io/api/core/v1"
|
||||||
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
||||||
"k8s.io/apimachinery/pkg/labels"
|
"k8s.io/apimachinery/pkg/labels"
|
||||||
"k8s.io/apimachinery/pkg/runtime"
|
clientset "k8s.io/client-go/kubernetes"
|
||||||
"k8s.io/apimachinery/pkg/util/sets"
|
|
||||||
"k8s.io/klog/v2"
|
"k8s.io/klog/v2"
|
||||||
|
|
||||||
"sigs.k8s.io/descheduler/pkg/apis/componentconfig"
|
"sigs.k8s.io/descheduler/pkg/api"
|
||||||
"sigs.k8s.io/descheduler/pkg/descheduler/evictions"
|
"sigs.k8s.io/descheduler/pkg/descheduler/evictions"
|
||||||
"sigs.k8s.io/descheduler/pkg/descheduler/node"
|
nodeutil "sigs.k8s.io/descheduler/pkg/descheduler/node"
|
||||||
podutil "sigs.k8s.io/descheduler/pkg/descheduler/pod"
|
"sigs.k8s.io/descheduler/pkg/descheduler/strategies/validation"
|
||||||
"sigs.k8s.io/descheduler/pkg/framework"
|
|
||||||
"sigs.k8s.io/descheduler/pkg/utils"
|
"sigs.k8s.io/descheduler/pkg/utils"
|
||||||
)
|
)
|
||||||
|
|
||||||
const PluginName = "RemovePodsViolatingTopologySpreadConstraint"
|
|
||||||
|
|
||||||
// AntiAffinityTerm's topology key value used in predicate metadata
|
// AntiAffinityTerm's topology key value used in predicate metadata
|
||||||
type topologyPair struct {
|
type topologyPair struct {
|
||||||
key string
|
key string
|
||||||
@@ -48,44 +48,25 @@ type topology struct {
|
|||||||
pods []*v1.Pod
|
pods []*v1.Pod
|
||||||
}
|
}
|
||||||
|
|
||||||
// RemovePodsViolatingTopologySpreadConstraint evicts pods which violate their topology spread constraints
|
func RemovePodsViolatingTopologySpreadConstraint(
|
||||||
type RemovePodsViolatingTopologySpreadConstraint struct {
|
ctx context.Context,
|
||||||
handle framework.Handle
|
client clientset.Interface,
|
||||||
args *componentconfig.RemovePodsViolatingTopologySpreadConstraintArgs
|
strategy api.DeschedulerStrategy,
|
||||||
podFilter podutil.FilterFunc
|
nodes []*v1.Node,
|
||||||
}
|
podEvictor *evictions.PodEvictor,
|
||||||
|
) {
|
||||||
var _ framework.BalancePlugin = &RemovePodsViolatingTopologySpreadConstraint{}
|
strategyParams, err := validation.ValidateAndParseStrategyParams(ctx, client, strategy.Params)
|
||||||
|
|
||||||
// New builds plugin from its arguments while passing a handle
|
|
||||||
func New(args runtime.Object, handle framework.Handle) (framework.Plugin, error) {
|
|
||||||
pluginArgs, ok := args.(*componentconfig.RemovePodsViolatingTopologySpreadConstraintArgs)
|
|
||||||
if !ok {
|
|
||||||
return nil, fmt.Errorf("want args to be of type RemovePodsViolatingTopologySpreadConstraintArgs, got %T", args)
|
|
||||||
}
|
|
||||||
|
|
||||||
podFilter, err := podutil.NewOptions().
|
|
||||||
WithFilter(handle.Evictor().Filter).
|
|
||||||
WithLabelSelector(pluginArgs.LabelSelector).
|
|
||||||
BuildFilterFunc()
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, fmt.Errorf("error initializing pod filter function: %v", err)
|
klog.ErrorS(err, "Invalid RemovePodsViolatingTopologySpreadConstraint parameters")
|
||||||
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
return &RemovePodsViolatingTopologySpreadConstraint{
|
evictable := podEvictor.Evictable(
|
||||||
handle: handle,
|
evictions.WithPriorityThreshold(strategyParams.ThresholdPriority),
|
||||||
podFilter: podFilter,
|
evictions.WithNodeFit(strategyParams.NodeFit),
|
||||||
args: pluginArgs,
|
evictions.WithLabelSelector(strategyParams.LabelSelector),
|
||||||
}, nil
|
)
|
||||||
}
|
|
||||||
|
|
||||||
// Name retrieves the plugin name
|
|
||||||
func (d *RemovePodsViolatingTopologySpreadConstraint) Name() string {
|
|
||||||
return PluginName
|
|
||||||
}
|
|
||||||
|
|
||||||
// nolint: gocyclo
|
|
||||||
func (d *RemovePodsViolatingTopologySpreadConstraint) Balance(ctx context.Context, nodes []*v1.Node) *framework.Status {
|
|
||||||
nodeMap := make(map[string]*v1.Node, len(nodes))
|
nodeMap := make(map[string]*v1.Node, len(nodes))
|
||||||
for _, node := range nodes {
|
for _, node := range nodes {
|
||||||
nodeMap[node.Name] = node
|
nodeMap[node.Name] = node
|
||||||
@@ -104,26 +85,17 @@ func (d *RemovePodsViolatingTopologySpreadConstraint) Balance(ctx context.Contex
|
|||||||
// if diff > maxSkew, add this pod in the current bucket for eviction
|
// if diff > maxSkew, add this pod in the current bucket for eviction
|
||||||
|
|
||||||
// First record all of the constraints by namespace
|
// First record all of the constraints by namespace
|
||||||
client := d.handle.ClientSet()
|
|
||||||
namespaces, err := client.CoreV1().Namespaces().List(ctx, metav1.ListOptions{})
|
namespaces, err := client.CoreV1().Namespaces().List(ctx, metav1.ListOptions{})
|
||||||
if err != nil {
|
if err != nil {
|
||||||
klog.ErrorS(err, "Couldn't list namespaces")
|
klog.ErrorS(err, "Couldn't list namespaces")
|
||||||
return &framework.Status{
|
return
|
||||||
Err: fmt.Errorf("list namespace: %w", err),
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
klog.V(1).InfoS("Processing namespaces for topology spread constraints")
|
klog.V(1).InfoS("Processing namespaces for topology spread constraints")
|
||||||
podsForEviction := make(map[*v1.Pod]struct{})
|
podsForEviction := make(map[*v1.Pod]struct{})
|
||||||
var includedNamespaces, excludedNamespaces sets.String
|
|
||||||
if d.args.Namespaces != nil {
|
|
||||||
includedNamespaces = sets.NewString(d.args.Namespaces.Include...)
|
|
||||||
excludedNamespaces = sets.NewString(d.args.Namespaces.Exclude...)
|
|
||||||
}
|
|
||||||
|
|
||||||
// 1. for each namespace...
|
// 1. for each namespace...
|
||||||
for _, namespace := range namespaces.Items {
|
for _, namespace := range namespaces.Items {
|
||||||
if (len(includedNamespaces) > 0 && !includedNamespaces.Has(namespace.Name)) ||
|
if (len(strategyParams.IncludedNamespaces) > 0 && !strategyParams.IncludedNamespaces.Has(namespace.Name)) ||
|
||||||
(len(excludedNamespaces) > 0 && excludedNamespaces.Has(namespace.Name)) {
|
(len(strategyParams.ExcludedNamespaces) > 0 && strategyParams.ExcludedNamespaces.Has(namespace.Name)) {
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
namespacePods, err := client.CoreV1().Pods(namespace.Name).List(ctx, metav1.ListOptions{})
|
namespacePods, err := client.CoreV1().Pods(namespace.Name).List(ctx, metav1.ListOptions{})
|
||||||
@@ -133,20 +105,14 @@ func (d *RemovePodsViolatingTopologySpreadConstraint) Balance(ctx context.Contex
|
|||||||
}
|
}
|
||||||
|
|
||||||
// ...where there is a topology constraint
|
// ...where there is a topology constraint
|
||||||
namespaceTopologySpreadConstraints := []v1.TopologySpreadConstraint{}
|
namespaceTopologySpreadConstraints := make(map[v1.TopologySpreadConstraint]struct{})
|
||||||
for _, pod := range namespacePods.Items {
|
for _, pod := range namespacePods.Items {
|
||||||
for _, constraint := range pod.Spec.TopologySpreadConstraints {
|
for _, constraint := range pod.Spec.TopologySpreadConstraints {
|
||||||
// Ignore soft topology constraints if they are not included
|
// Ignore soft topology constraints if they are not included
|
||||||
if constraint.WhenUnsatisfiable == v1.ScheduleAnyway && (d.args == nil || !d.args.IncludeSoftConstraints) {
|
if constraint.WhenUnsatisfiable == v1.ScheduleAnyway && (strategy.Params == nil || !strategy.Params.IncludeSoftConstraints) {
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
// Need to check v1.TopologySpreadConstraint deepEquality because
|
namespaceTopologySpreadConstraints[constraint] = struct{}{}
|
||||||
// v1.TopologySpreadConstraint has pointer fields
|
|
||||||
// and we don't need to go over duplicated constraints later on
|
|
||||||
if hasIdenticalConstraints(constraint, namespaceTopologySpreadConstraints) {
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
namespaceTopologySpreadConstraints = append(namespaceTopologySpreadConstraints, constraint)
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if len(namespaceTopologySpreadConstraints) == 0 {
|
if len(namespaceTopologySpreadConstraints) == 0 {
|
||||||
@@ -154,7 +120,7 @@ func (d *RemovePodsViolatingTopologySpreadConstraint) Balance(ctx context.Contex
|
|||||||
}
|
}
|
||||||
|
|
||||||
// 2. for each topologySpreadConstraint in that namespace
|
// 2. for each topologySpreadConstraint in that namespace
|
||||||
for _, constraint := range namespaceTopologySpreadConstraints {
|
for constraint := range namespaceTopologySpreadConstraints {
|
||||||
constraintTopologies := make(map[topologyPair][]*v1.Pod)
|
constraintTopologies := make(map[topologyPair][]*v1.Pod)
|
||||||
// pre-populate the topologyPair map with all the topologies available from the nodeMap
|
// pre-populate the topologyPair map with all the topologies available from the nodeMap
|
||||||
// (we can't just build it from existing pods' nodes because a topology may have 0 pods)
|
// (we can't just build it from existing pods' nodes because a topology may have 0 pods)
|
||||||
@@ -174,16 +140,12 @@ func (d *RemovePodsViolatingTopologySpreadConstraint) Balance(ctx context.Contex
|
|||||||
// (this loop is where we count the number of pods per topologyValue that match this constraint's selector)
|
// (this loop is where we count the number of pods per topologyValue that match this constraint's selector)
|
||||||
var sumPods float64
|
var sumPods float64
|
||||||
for i := range namespacePods.Items {
|
for i := range namespacePods.Items {
|
||||||
// skip pods that are being deleted.
|
|
||||||
if utils.IsPodTerminating(&namespacePods.Items[i]) {
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
// 4. if the pod matches this TopologySpreadConstraint LabelSelector
|
// 4. if the pod matches this TopologySpreadConstraint LabelSelector
|
||||||
if !selector.Matches(labels.Set(namespacePods.Items[i].Labels)) {
|
if !selector.Matches(labels.Set(namespacePods.Items[i].Labels)) {
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
|
|
||||||
// 5. If the pod's node matches this constraint's topologyKey, create a topoPair and add the pod
|
// 5. If the pod's node matches this constraint'selector topologyKey, create a topoPair and add the pod
|
||||||
node, ok := nodeMap[namespacePods.Items[i].Spec.NodeName]
|
node, ok := nodeMap[namespacePods.Items[i].Spec.NodeName]
|
||||||
if !ok {
|
if !ok {
|
||||||
// If ok is false, node is nil in which case node.Labels will panic. In which case a pod is yet to be scheduled. So it's safe to just continue here.
|
// If ok is false, node is nil in which case node.Labels will panic. In which case a pod is yet to be scheduled. So it's safe to just continue here.
|
||||||
@@ -203,35 +165,19 @@ func (d *RemovePodsViolatingTopologySpreadConstraint) Balance(ctx context.Contex
|
|||||||
klog.V(2).InfoS("Skipping topology constraint because it is already balanced", "constraint", constraint)
|
klog.V(2).InfoS("Skipping topology constraint because it is already balanced", "constraint", constraint)
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
balanceDomains(d.handle.GetPodsAssignedToNodeFunc(), podsForEviction, constraint, constraintTopologies, sumPods, d.handle.Evictor().Filter, nodes)
|
balanceDomains(podsForEviction, constraint, constraintTopologies, sumPods, evictable.IsEvictable, nodeMap)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
nodeLimitExceeded := map[string]bool{}
|
|
||||||
for pod := range podsForEviction {
|
for pod := range podsForEviction {
|
||||||
if nodeLimitExceeded[pod.Spec.NodeName] {
|
if !evictable.IsEvictable(pod) {
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
if !d.podFilter(pod) {
|
if _, err := podEvictor.EvictPod(ctx, pod, nodeMap[pod.Spec.NodeName], "PodTopologySpread"); err != nil {
|
||||||
continue
|
klog.ErrorS(err, "Error evicting pod", "pod", klog.KObj(pod))
|
||||||
}
|
break
|
||||||
d.handle.Evictor().Evict(ctx, pod, evictions.EvictOptions{})
|
|
||||||
if d.handle.Evictor().NodeLimitExceeded(nodeMap[pod.Spec.NodeName]) {
|
|
||||||
nodeLimitExceeded[pod.Spec.NodeName] = true
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// hasIdenticalConstraints checks if we already had an identical TopologySpreadConstraint in namespaceTopologySpreadConstraints slice
|
|
||||||
func hasIdenticalConstraints(newConstraint v1.TopologySpreadConstraint, namespaceTopologySpreadConstraints []v1.TopologySpreadConstraint) bool {
|
|
||||||
for _, constraint := range namespaceTopologySpreadConstraints {
|
|
||||||
if reflect.DeepEqual(newConstraint, constraint) {
|
|
||||||
return true
|
|
||||||
}
|
|
||||||
}
|
|
||||||
return false
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// topologyIsBalanced checks if any domains in the topology differ by more than the MaxSkew
|
// topologyIsBalanced checks if any domains in the topology differ by more than the MaxSkew
|
||||||
@@ -266,7 +212,7 @@ func topologyIsBalanced(topology map[topologyPair][]*v1.Pod, constraint v1.Topol
|
|||||||
// whichever number is less.
|
// whichever number is less.
|
||||||
//
|
//
|
||||||
// (Note, we will only move as many pods from a domain as possible without bringing it below the ideal average,
|
// (Note, we will only move as many pods from a domain as possible without bringing it below the ideal average,
|
||||||
// and we will not bring any smaller domain above the average)
|
// and we will not bring any smaller domain above the average)
|
||||||
// If the diff is within the skew, we move to the next highest domain.
|
// If the diff is within the skew, we move to the next highest domain.
|
||||||
// If the higher domain can't give any more without falling below the average, we move to the next lowest "high" domain
|
// If the higher domain can't give any more without falling below the average, we move to the next lowest "high" domain
|
||||||
//
|
//
|
||||||
@@ -274,19 +220,15 @@ func topologyIsBalanced(topology map[topologyPair][]*v1.Pod, constraint v1.Topol
|
|||||||
// [5, 5, 5, 5, 5, 5]
|
// [5, 5, 5, 5, 5, 5]
|
||||||
// (assuming even distribution by the scheduler of the evicted pods)
|
// (assuming even distribution by the scheduler of the evicted pods)
|
||||||
func balanceDomains(
|
func balanceDomains(
|
||||||
getPodsAssignedToNode podutil.GetPodsAssignedToNodeFunc,
|
|
||||||
podsForEviction map[*v1.Pod]struct{},
|
podsForEviction map[*v1.Pod]struct{},
|
||||||
constraint v1.TopologySpreadConstraint,
|
constraint v1.TopologySpreadConstraint,
|
||||||
constraintTopologies map[topologyPair][]*v1.Pod,
|
constraintTopologies map[topologyPair][]*v1.Pod,
|
||||||
sumPods float64,
|
sumPods float64,
|
||||||
isEvictable func(pod *v1.Pod) bool,
|
isEvictable func(*v1.Pod) bool,
|
||||||
nodes []*v1.Node) {
|
nodeMap map[string]*v1.Node) {
|
||||||
|
|
||||||
idealAvg := sumPods / float64(len(constraintTopologies))
|
idealAvg := sumPods / float64(len(constraintTopologies))
|
||||||
sortedDomains := sortDomains(constraintTopologies, isEvictable)
|
sortedDomains := sortDomains(constraintTopologies, isEvictable)
|
||||||
|
|
||||||
nodesBelowIdealAvg := filterNodesBelowIdealAvg(nodes, sortedDomains, constraint.TopologyKey, idealAvg)
|
|
||||||
|
|
||||||
// i is the index for belowOrEqualAvg
|
// i is the index for belowOrEqualAvg
|
||||||
// j is the index for aboveAvg
|
// j is the index for aboveAvg
|
||||||
i := 0
|
i := 0
|
||||||
@@ -326,20 +268,8 @@ func balanceDomains(
|
|||||||
// also (just for tracking), add them to the list of pods in the lower topology
|
// also (just for tracking), add them to the list of pods in the lower topology
|
||||||
aboveToEvict := sortedDomains[j].pods[len(sortedDomains[j].pods)-movePods:]
|
aboveToEvict := sortedDomains[j].pods[len(sortedDomains[j].pods)-movePods:]
|
||||||
for k := range aboveToEvict {
|
for k := range aboveToEvict {
|
||||||
// PodFitsAnyOtherNode excludes the current node because, for the sake of domain balancing only, we care about if there is any other
|
if err := validatePodFitsOnOtherNodes(aboveToEvict[k], nodeMap); err != nil {
|
||||||
// place it could theoretically fit.
|
klog.V(2).InfoS(fmt.Sprintf("ignoring pod for eviction due to: %s", err.Error()), "pod", klog.KObj(aboveToEvict[k]))
|
||||||
// If the pod doesn't fit on its current node, that is a job for RemovePodsViolatingNodeAffinity, and irrelevant to Topology Spreading
|
|
||||||
// Also, if the pod has a hard nodeAffinity/nodeSelector/toleration that only matches this node,
|
|
||||||
// don't bother evicting it as it will just end up back on the same node
|
|
||||||
// however we still account for it "being evicted" so the algorithm can complete
|
|
||||||
// TODO(@damemi): Since we don't order pods wrt their affinities, we should refactor this to skip the current pod
|
|
||||||
// but still try to get the required # of movePods (instead of just chopping that value off the slice above).
|
|
||||||
// In other words, PTS can perform suboptimally if some of its chosen pods don't fit on other nodes.
|
|
||||||
// This is because the chosen pods aren't sorted, but immovable pods still count as "evicted" toward the PTS algorithm.
|
|
||||||
// So, a better selection heuristic could improve performance.
|
|
||||||
|
|
||||||
if !node.PodFitsAnyOtherNode(getPodsAssignedToNode, aboveToEvict[k], nodesBelowIdealAvg) {
|
|
||||||
klog.V(2).InfoS("ignoring pod for eviction as it does not fit on any other node", "pod", klog.KObj(aboveToEvict[k]))
|
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -350,23 +280,54 @@ func balanceDomains(
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// filterNodesBelowIdealAvg will return nodes that have fewer pods matching topology domain than the idealAvg count.
|
// validatePodFitsOnOtherNodes performs validation based on scheduling predicates for affinity and toleration.
|
||||||
// the desired behavior is to not consider nodes in a given topology domain that are already packed.
|
// It excludes the current node because, for the sake of domain balancing only, we care about if there is any other
|
||||||
func filterNodesBelowIdealAvg(nodes []*v1.Node, sortedDomains []topology, topologyKey string, idealAvg float64) []*v1.Node {
|
// place it could theoretically fit.
|
||||||
topologyNodesMap := make(map[string][]*v1.Node, len(sortedDomains))
|
// If the pod doesn't fit on its current node, that is a job for RemovePodsViolatingNodeAffinity, and irrelevant to Topology Spreading
|
||||||
for _, node := range nodes {
|
func validatePodFitsOnOtherNodes(pod *v1.Pod, nodeMap map[string]*v1.Node) error {
|
||||||
if topologyDomain, ok := node.Labels[topologyKey]; ok {
|
// if the pod has a hard nodeAffinity/nodeSelector/toleration that only matches this node,
|
||||||
topologyNodesMap[topologyDomain] = append(topologyNodesMap[topologyDomain], node)
|
// don't bother evicting it as it will just end up back on the same node
|
||||||
}
|
// however we still account for it "being evicted" so the algorithm can complete
|
||||||
|
// TODO(@damemi): Since we don't order pods wrt their affinities, we should refactor this to skip the current pod
|
||||||
|
// but still try to get the required # of movePods (instead of just chopping that value off the slice above)
|
||||||
|
isRequiredDuringSchedulingIgnoredDuringExecution := pod.Spec.Affinity != nil &&
|
||||||
|
pod.Spec.Affinity.NodeAffinity != nil &&
|
||||||
|
pod.Spec.Affinity.NodeAffinity.RequiredDuringSchedulingIgnoredDuringExecution != nil
|
||||||
|
|
||||||
|
hardTaintsFilter := func(taint *v1.Taint) bool {
|
||||||
|
return taint.Effect == v1.TaintEffectNoSchedule || taint.Effect == v1.TaintEffectNoExecute
|
||||||
}
|
}
|
||||||
|
|
||||||
var nodesBelowIdealAvg []*v1.Node
|
var eligibleNodesCount, ineligibleAffinityNodesCount, ineligibleTaintedNodesCount int
|
||||||
for _, domain := range sortedDomains {
|
for _, node := range nodeMap {
|
||||||
if float64(len(domain.pods)) < idealAvg {
|
if node == nodeMap[pod.Spec.NodeName] {
|
||||||
nodesBelowIdealAvg = append(nodesBelowIdealAvg, topologyNodesMap[domain.pair.value]...)
|
continue
|
||||||
}
|
}
|
||||||
|
if pod.Spec.NodeSelector != nil || isRequiredDuringSchedulingIgnoredDuringExecution {
|
||||||
|
if !nodeutil.PodFitsCurrentNode(pod, node) {
|
||||||
|
ineligibleAffinityNodesCount++
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if !utils.TolerationsTolerateTaintsWithFilter(pod.Spec.Tolerations, node.Spec.Taints, hardTaintsFilter) {
|
||||||
|
ineligibleTaintedNodesCount++
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
|
eligibleNodesCount++
|
||||||
}
|
}
|
||||||
return nodesBelowIdealAvg
|
|
||||||
|
if eligibleNodesCount == 0 {
|
||||||
|
var errs []error
|
||||||
|
if ineligibleAffinityNodesCount > 0 {
|
||||||
|
errs = append(errs, fmt.Errorf("%d nodes with ineligible selector/affinity", ineligibleAffinityNodesCount))
|
||||||
|
}
|
||||||
|
if ineligibleTaintedNodesCount > 0 {
|
||||||
|
errs = append(errs, fmt.Errorf("%d nodes with taints that are not tolerated", ineligibleTaintedNodesCount))
|
||||||
|
}
|
||||||
|
return utilerrors.NewAggregate(errs)
|
||||||
|
}
|
||||||
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// sortDomains sorts and splits the list of topology domains based on their size
|
// sortDomains sorts and splits the list of topology domains based on their size
|
||||||
@@ -376,7 +337,7 @@ func filterNodesBelowIdealAvg(nodes []*v1.Node, sortedDomains []topology, topolo
|
|||||||
// 3. pods in descending priority
|
// 3. pods in descending priority
|
||||||
// 4. all other pods
|
// 4. all other pods
|
||||||
// We then pop pods off the back of the list for eviction
|
// We then pop pods off the back of the list for eviction
|
||||||
func sortDomains(constraintTopologyPairs map[topologyPair][]*v1.Pod, isEvictable func(pod *v1.Pod) bool) []topology {
|
func sortDomains(constraintTopologyPairs map[topologyPair][]*v1.Pod, isEvictable func(*v1.Pod) bool) []topology {
|
||||||
sortedTopologies := make([]topology, 0, len(constraintTopologyPairs))
|
sortedTopologies := make([]topology, 0, len(constraintTopologyPairs))
|
||||||
// sort the topologies and return 2 lists: those <= the average and those > the average (> list inverted)
|
// sort the topologies and return 2 lists: those <= the average and those > the average (> list inverted)
|
||||||
for pair, list := range constraintTopologyPairs {
|
for pair, list := range constraintTopologyPairs {
|
||||||
@@ -1,40 +1,30 @@
|
|||||||
package removepodsviolatingtopologyspreadconstraint
|
package strategies
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
"fmt"
|
"fmt"
|
||||||
"testing"
|
"testing"
|
||||||
|
|
||||||
|
"sigs.k8s.io/descheduler/pkg/api"
|
||||||
|
|
||||||
v1 "k8s.io/api/core/v1"
|
v1 "k8s.io/api/core/v1"
|
||||||
policy "k8s.io/api/policy/v1"
|
|
||||||
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
||||||
"k8s.io/apimachinery/pkg/runtime"
|
"k8s.io/apimachinery/pkg/runtime"
|
||||||
"k8s.io/apimachinery/pkg/util/sets"
|
|
||||||
"k8s.io/client-go/informers"
|
|
||||||
"k8s.io/client-go/kubernetes/fake"
|
"k8s.io/client-go/kubernetes/fake"
|
||||||
core "k8s.io/client-go/testing"
|
core "k8s.io/client-go/testing"
|
||||||
"k8s.io/client-go/tools/events"
|
|
||||||
|
|
||||||
"sigs.k8s.io/descheduler/pkg/api"
|
|
||||||
"sigs.k8s.io/descheduler/pkg/apis/componentconfig"
|
|
||||||
"sigs.k8s.io/descheduler/pkg/descheduler/evictions"
|
"sigs.k8s.io/descheduler/pkg/descheduler/evictions"
|
||||||
podutil "sigs.k8s.io/descheduler/pkg/descheduler/pod"
|
|
||||||
"sigs.k8s.io/descheduler/pkg/framework"
|
|
||||||
frameworkfake "sigs.k8s.io/descheduler/pkg/framework/fake"
|
|
||||||
"sigs.k8s.io/descheduler/pkg/framework/plugins/defaultevictor"
|
|
||||||
"sigs.k8s.io/descheduler/test"
|
"sigs.k8s.io/descheduler/test"
|
||||||
)
|
)
|
||||||
|
|
||||||
func TestTopologySpreadConstraint(t *testing.T) {
|
func TestTopologySpreadConstraint(t *testing.T) {
|
||||||
|
ctx := context.Background()
|
||||||
testCases := []struct {
|
testCases := []struct {
|
||||||
name string
|
name string
|
||||||
pods []*v1.Pod
|
pods []*v1.Pod
|
||||||
expectedEvictedCount uint
|
expectedEvictedCount int
|
||||||
expectedEvictedPods []string // if specified, will assert specific pods were evicted
|
|
||||||
nodes []*v1.Node
|
nodes []*v1.Node
|
||||||
|
strategy api.DeschedulerStrategy
|
||||||
namespaces []string
|
namespaces []string
|
||||||
args componentconfig.RemovePodsViolatingTopologySpreadConstraintArgs
|
|
||||||
nodeFit bool
|
|
||||||
}{
|
}{
|
||||||
{
|
{
|
||||||
name: "2 domains, sizes [2,1], maxSkew=1, move 0 pods",
|
name: "2 domains, sizes [2,1], maxSkew=1, move 0 pods",
|
||||||
@@ -68,8 +58,12 @@ func TestTopologySpreadConstraint(t *testing.T) {
|
|||||||
},
|
},
|
||||||
}),
|
}),
|
||||||
expectedEvictedCount: 0,
|
expectedEvictedCount: 0,
|
||||||
namespaces: []string{"ns1"},
|
strategy: api.DeschedulerStrategy{
|
||||||
args: componentconfig.RemovePodsViolatingTopologySpreadConstraintArgs{},
|
Params: &api.StrategyParameters{
|
||||||
|
NodeFit: false,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
namespaces: []string{"ns1"},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "2 domains, sizes [3,1], maxSkew=1, move 1 pod to achieve [2,2]",
|
name: "2 domains, sizes [3,1], maxSkew=1, move 1 pod to achieve [2,2]",
|
||||||
@@ -96,8 +90,12 @@ func TestTopologySpreadConstraint(t *testing.T) {
|
|||||||
},
|
},
|
||||||
}),
|
}),
|
||||||
expectedEvictedCount: 1,
|
expectedEvictedCount: 1,
|
||||||
namespaces: []string{"ns1"},
|
strategy: api.DeschedulerStrategy{
|
||||||
args: componentconfig.RemovePodsViolatingTopologySpreadConstraintArgs{},
|
Params: &api.StrategyParameters{
|
||||||
|
NodeFit: false,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
namespaces: []string{"ns1"},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "2 domains, sizes [3,1], maxSkew=1, move 1 pod to achieve [2,2] (soft constraints)",
|
name: "2 domains, sizes [3,1], maxSkew=1, move 1 pod to achieve [2,2] (soft constraints)",
|
||||||
@@ -131,8 +129,8 @@ func TestTopologySpreadConstraint(t *testing.T) {
|
|||||||
},
|
},
|
||||||
}),
|
}),
|
||||||
expectedEvictedCount: 1,
|
expectedEvictedCount: 1,
|
||||||
|
strategy: api.DeschedulerStrategy{Params: &api.StrategyParameters{IncludeSoftConstraints: true}},
|
||||||
namespaces: []string{"ns1"},
|
namespaces: []string{"ns1"},
|
||||||
args: componentconfig.RemovePodsViolatingTopologySpreadConstraintArgs{IncludeSoftConstraints: true},
|
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "2 domains, sizes [3,1], maxSkew=1, no pods eligible, move 0 pods",
|
name: "2 domains, sizes [3,1], maxSkew=1, no pods eligible, move 0 pods",
|
||||||
@@ -162,8 +160,12 @@ func TestTopologySpreadConstraint(t *testing.T) {
|
|||||||
},
|
},
|
||||||
}),
|
}),
|
||||||
expectedEvictedCount: 0,
|
expectedEvictedCount: 0,
|
||||||
namespaces: []string{"ns1"},
|
strategy: api.DeschedulerStrategy{
|
||||||
args: componentconfig.RemovePodsViolatingTopologySpreadConstraintArgs{},
|
Params: &api.StrategyParameters{
|
||||||
|
NodeFit: false,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
namespaces: []string{"ns1"},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "2 domains, sizes [3,1], maxSkew=1, move 1 pod to achieve [2,2], exclude kube-system namespace",
|
name: "2 domains, sizes [3,1], maxSkew=1, move 1 pod to achieve [2,2], exclude kube-system namespace",
|
||||||
@@ -190,9 +192,8 @@ func TestTopologySpreadConstraint(t *testing.T) {
|
|||||||
},
|
},
|
||||||
}),
|
}),
|
||||||
expectedEvictedCount: 1,
|
expectedEvictedCount: 1,
|
||||||
|
strategy: api.DeschedulerStrategy{Enabled: true, Params: &api.StrategyParameters{NodeFit: true, Namespaces: &api.Namespaces{Exclude: []string{"kube-system"}}}},
|
||||||
namespaces: []string{"ns1"},
|
namespaces: []string{"ns1"},
|
||||||
args: componentconfig.RemovePodsViolatingTopologySpreadConstraintArgs{Namespaces: &api.Namespaces{Exclude: []string{"kube-system"}}},
|
|
||||||
nodeFit: true,
|
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "2 domains, sizes [5,2], maxSkew=1, move 1 pod to achieve [4,3]",
|
name: "2 domains, sizes [5,2], maxSkew=1, move 1 pod to achieve [4,3]",
|
||||||
@@ -219,8 +220,12 @@ func TestTopologySpreadConstraint(t *testing.T) {
|
|||||||
},
|
},
|
||||||
}),
|
}),
|
||||||
expectedEvictedCount: 1,
|
expectedEvictedCount: 1,
|
||||||
namespaces: []string{"ns1"},
|
strategy: api.DeschedulerStrategy{
|
||||||
args: componentconfig.RemovePodsViolatingTopologySpreadConstraintArgs{},
|
Params: &api.StrategyParameters{
|
||||||
|
NodeFit: false,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
namespaces: []string{"ns1"},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "2 domains, sizes [4,0], maxSkew=1, move 2 pods to achieve [2,2]",
|
name: "2 domains, sizes [4,0], maxSkew=1, move 2 pods to achieve [2,2]",
|
||||||
@@ -242,8 +247,12 @@ func TestTopologySpreadConstraint(t *testing.T) {
|
|||||||
},
|
},
|
||||||
}),
|
}),
|
||||||
expectedEvictedCount: 2,
|
expectedEvictedCount: 2,
|
||||||
namespaces: []string{"ns1"},
|
strategy: api.DeschedulerStrategy{
|
||||||
args: componentconfig.RemovePodsViolatingTopologySpreadConstraintArgs{},
|
Params: &api.StrategyParameters{
|
||||||
|
NodeFit: false,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
namespaces: []string{"ns1"},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "2 domains, sizes [4,0], maxSkew=1, only move 1 pod since pods with nodeSelector and nodeAffinity aren't evicted",
|
name: "2 domains, sizes [4,0], maxSkew=1, only move 1 pod since pods with nodeSelector and nodeAffinity aren't evicted",
|
||||||
@@ -282,9 +291,12 @@ func TestTopologySpreadConstraint(t *testing.T) {
|
|||||||
},
|
},
|
||||||
}),
|
}),
|
||||||
expectedEvictedCount: 1,
|
expectedEvictedCount: 1,
|
||||||
namespaces: []string{"ns1"},
|
strategy: api.DeschedulerStrategy{
|
||||||
args: componentconfig.RemovePodsViolatingTopologySpreadConstraintArgs{},
|
Params: &api.StrategyParameters{
|
||||||
nodeFit: true,
|
NodeFit: true,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
namespaces: []string{"ns1"},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "2 domains, sizes [4,0], maxSkew=1, move 2 pods since selector matches multiple nodes",
|
name: "2 domains, sizes [4,0], maxSkew=1, move 2 pods since selector matches multiple nodes",
|
||||||
@@ -326,8 +338,12 @@ func TestTopologySpreadConstraint(t *testing.T) {
|
|||||||
},
|
},
|
||||||
}),
|
}),
|
||||||
expectedEvictedCount: 2,
|
expectedEvictedCount: 2,
|
||||||
namespaces: []string{"ns1"},
|
strategy: api.DeschedulerStrategy{
|
||||||
args: componentconfig.RemovePodsViolatingTopologySpreadConstraintArgs{},
|
Params: &api.StrategyParameters{
|
||||||
|
NodeFit: false,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
namespaces: []string{"ns1"},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "3 domains, sizes [0, 1, 100], maxSkew=1, move 66 pods to get [34, 33, 34]",
|
name: "3 domains, sizes [0, 1, 100], maxSkew=1, move 66 pods to get [34, 33, 34]",
|
||||||
@@ -350,8 +366,8 @@ func TestTopologySpreadConstraint(t *testing.T) {
|
|||||||
},
|
},
|
||||||
}),
|
}),
|
||||||
expectedEvictedCount: 66,
|
expectedEvictedCount: 66,
|
||||||
|
strategy: api.DeschedulerStrategy{},
|
||||||
namespaces: []string{"ns1"},
|
namespaces: []string{"ns1"},
|
||||||
args: componentconfig.RemovePodsViolatingTopologySpreadConstraintArgs{},
|
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "4 domains, sizes [0, 1, 3, 5], should move 3 to get [2, 2, 3, 2]",
|
name: "4 domains, sizes [0, 1, 3, 5], should move 3 to get [2, 2, 3, 2]",
|
||||||
@@ -380,8 +396,12 @@ func TestTopologySpreadConstraint(t *testing.T) {
|
|||||||
},
|
},
|
||||||
}),
|
}),
|
||||||
expectedEvictedCount: 3,
|
expectedEvictedCount: 3,
|
||||||
namespaces: []string{"ns1"},
|
strategy: api.DeschedulerStrategy{
|
||||||
args: componentconfig.RemovePodsViolatingTopologySpreadConstraintArgs{},
|
Params: &api.StrategyParameters{
|
||||||
|
NodeFit: false,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
namespaces: []string{"ns1"},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "2 domains size [2 6], maxSkew=2, should move 1 to get [3 5]",
|
name: "2 domains size [2 6], maxSkew=2, should move 1 to get [3 5]",
|
||||||
@@ -408,8 +428,12 @@ func TestTopologySpreadConstraint(t *testing.T) {
|
|||||||
},
|
},
|
||||||
}),
|
}),
|
||||||
expectedEvictedCount: 1,
|
expectedEvictedCount: 1,
|
||||||
namespaces: []string{"ns1"},
|
strategy: api.DeschedulerStrategy{
|
||||||
args: componentconfig.RemovePodsViolatingTopologySpreadConstraintArgs{},
|
Params: &api.StrategyParameters{
|
||||||
|
NodeFit: false,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
namespaces: []string{"ns1"},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "2 domains size [2 6], maxSkew=2, can't move any because of node taints",
|
name: "2 domains size [2 6], maxSkew=2, can't move any because of node taints",
|
||||||
@@ -452,38 +476,12 @@ func TestTopologySpreadConstraint(t *testing.T) {
|
|||||||
},
|
},
|
||||||
}),
|
}),
|
||||||
expectedEvictedCount: 0,
|
expectedEvictedCount: 0,
|
||||||
namespaces: []string{"ns1"},
|
strategy: api.DeschedulerStrategy{
|
||||||
args: componentconfig.RemovePodsViolatingTopologySpreadConstraintArgs{},
|
Params: &api.StrategyParameters{
|
||||||
nodeFit: true,
|
NodeFit: true,
|
||||||
},
|
},
|
||||||
{
|
|
||||||
name: "2 domains size [2 6], maxSkew=2, can't move any because node1 does not have enough CPU",
|
|
||||||
nodes: []*v1.Node{
|
|
||||||
test.BuildTestNode("n1", 200, 3000, 10, func(n *v1.Node) { n.Labels["zone"] = "zoneA" }),
|
|
||||||
test.BuildTestNode("n2", 2000, 3000, 10, func(n *v1.Node) { n.Labels["zone"] = "zoneB" }),
|
|
||||||
},
|
},
|
||||||
pods: createTestPods([]testPodList{
|
namespaces: []string{"ns1"},
|
||||||
{
|
|
||||||
count: 1,
|
|
||||||
node: "n1",
|
|
||||||
labels: map[string]string{"foo": "bar"},
|
|
||||||
constraints: getDefaultTopologyConstraints(2),
|
|
||||||
},
|
|
||||||
{
|
|
||||||
count: 1,
|
|
||||||
node: "n1",
|
|
||||||
labels: map[string]string{"foo": "bar"},
|
|
||||||
},
|
|
||||||
{
|
|
||||||
count: 6,
|
|
||||||
node: "n2",
|
|
||||||
labels: map[string]string{"foo": "bar"},
|
|
||||||
},
|
|
||||||
}),
|
|
||||||
expectedEvictedCount: 0,
|
|
||||||
namespaces: []string{"ns1"},
|
|
||||||
args: componentconfig.RemovePodsViolatingTopologySpreadConstraintArgs{},
|
|
||||||
nodeFit: true,
|
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
// see https://github.com/kubernetes-sigs/descheduler/issues/564
|
// see https://github.com/kubernetes-sigs/descheduler/issues/564
|
||||||
@@ -587,8 +585,8 @@ func TestTopologySpreadConstraint(t *testing.T) {
|
|||||||
},
|
},
|
||||||
}),
|
}),
|
||||||
expectedEvictedCount: 1,
|
expectedEvictedCount: 1,
|
||||||
|
strategy: api.DeschedulerStrategy{Params: &api.StrategyParameters{IncludeSoftConstraints: true}},
|
||||||
namespaces: []string{"ns1"},
|
namespaces: []string{"ns1"},
|
||||||
args: componentconfig.RemovePodsViolatingTopologySpreadConstraintArgs{IncludeSoftConstraints: true},
|
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "3 domains size [8 7 0], maxSkew=1, should move 5 to get [5 5 5]",
|
name: "3 domains size [8 7 0], maxSkew=1, should move 5 to get [5 5 5]",
|
||||||
@@ -612,9 +610,8 @@ func TestTopologySpreadConstraint(t *testing.T) {
|
|||||||
},
|
},
|
||||||
}),
|
}),
|
||||||
expectedEvictedCount: 5,
|
expectedEvictedCount: 5,
|
||||||
expectedEvictedPods: []string{"pod-5", "pod-6", "pod-7", "pod-8"},
|
strategy: api.DeschedulerStrategy{},
|
||||||
namespaces: []string{"ns1"},
|
namespaces: []string{"ns1"},
|
||||||
args: componentconfig.RemovePodsViolatingTopologySpreadConstraintArgs{},
|
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "3 domains size [5 5 5], maxSkew=1, should move 0 to retain [5 5 5]",
|
name: "3 domains size [5 5 5], maxSkew=1, should move 0 to retain [5 5 5]",
|
||||||
@@ -644,8 +641,8 @@ func TestTopologySpreadConstraint(t *testing.T) {
|
|||||||
},
|
},
|
||||||
}),
|
}),
|
||||||
expectedEvictedCount: 0,
|
expectedEvictedCount: 0,
|
||||||
|
strategy: api.DeschedulerStrategy{},
|
||||||
namespaces: []string{"ns1"},
|
namespaces: []string{"ns1"},
|
||||||
args: componentconfig.RemovePodsViolatingTopologySpreadConstraintArgs{},
|
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "2 domains, sizes [2,0], maxSkew=1, move 1 pod since pod tolerates the node with taint",
|
name: "2 domains, sizes [2,0], maxSkew=1, move 1 pod since pod tolerates the node with taint",
|
||||||
@@ -685,9 +682,8 @@ func TestTopologySpreadConstraint(t *testing.T) {
|
|||||||
},
|
},
|
||||||
}),
|
}),
|
||||||
expectedEvictedCount: 1,
|
expectedEvictedCount: 1,
|
||||||
expectedEvictedPods: []string{"pod-0"},
|
strategy: api.DeschedulerStrategy{},
|
||||||
namespaces: []string{"ns1"},
|
namespaces: []string{"ns1"},
|
||||||
args: componentconfig.RemovePodsViolatingTopologySpreadConstraintArgs{},
|
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "2 domains, sizes [2,0], maxSkew=1, move 0 pods since pod does not tolerate the tainted node",
|
name: "2 domains, sizes [2,0], maxSkew=1, move 0 pods since pod does not tolerate the tainted node",
|
||||||
@@ -719,42 +715,8 @@ func TestTopologySpreadConstraint(t *testing.T) {
|
|||||||
},
|
},
|
||||||
}),
|
}),
|
||||||
expectedEvictedCount: 0,
|
expectedEvictedCount: 0,
|
||||||
|
strategy: api.DeschedulerStrategy{},
|
||||||
namespaces: []string{"ns1"},
|
namespaces: []string{"ns1"},
|
||||||
args: componentconfig.RemovePodsViolatingTopologySpreadConstraintArgs{},
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "2 domains, sizes [2,0], maxSkew=1, move 0 pods since pod does not tolerate the tainted node, and NodeFit is enabled",
|
|
||||||
nodes: []*v1.Node{
|
|
||||||
test.BuildTestNode("n1", 2000, 3000, 10, func(n *v1.Node) { n.Labels["zone"] = "zoneA" }),
|
|
||||||
test.BuildTestNode("n2", 2000, 3000, 10, func(n *v1.Node) {
|
|
||||||
n.Labels["zone"] = "zoneB"
|
|
||||||
n.Spec.Taints = []v1.Taint{
|
|
||||||
{
|
|
||||||
Key: "taint-test",
|
|
||||||
Value: "test",
|
|
||||||
Effect: v1.TaintEffectNoSchedule,
|
|
||||||
},
|
|
||||||
}
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
pods: createTestPods([]testPodList{
|
|
||||||
{
|
|
||||||
count: 1,
|
|
||||||
node: "n1",
|
|
||||||
labels: map[string]string{"foo": "bar"},
|
|
||||||
constraints: getDefaultTopologyConstraints(1),
|
|
||||||
},
|
|
||||||
{
|
|
||||||
count: 1,
|
|
||||||
node: "n1",
|
|
||||||
labels: map[string]string{"foo": "bar"},
|
|
||||||
nodeSelector: map[string]string{"zone": "zoneA"},
|
|
||||||
},
|
|
||||||
}),
|
|
||||||
expectedEvictedCount: 0,
|
|
||||||
namespaces: []string{"ns1"},
|
|
||||||
args: componentconfig.RemovePodsViolatingTopologySpreadConstraintArgs{},
|
|
||||||
nodeFit: true,
|
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "2 domains, sizes [2,0], maxSkew=1, move 1 pod for node with PreferNoSchedule Taint",
|
name: "2 domains, sizes [2,0], maxSkew=1, move 1 pod for node with PreferNoSchedule Taint",
|
||||||
@@ -786,9 +748,8 @@ func TestTopologySpreadConstraint(t *testing.T) {
|
|||||||
},
|
},
|
||||||
}),
|
}),
|
||||||
expectedEvictedCount: 1,
|
expectedEvictedCount: 1,
|
||||||
expectedEvictedPods: []string{"pod-0"},
|
strategy: api.DeschedulerStrategy{},
|
||||||
namespaces: []string{"ns1"},
|
namespaces: []string{"ns1"},
|
||||||
args: componentconfig.RemovePodsViolatingTopologySpreadConstraintArgs{},
|
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "2 domains, sizes [2,0], maxSkew=1, move 0 pod for node with unmatched label filtering",
|
name: "2 domains, sizes [2,0], maxSkew=1, move 0 pod for node with unmatched label filtering",
|
||||||
@@ -810,8 +771,12 @@ func TestTopologySpreadConstraint(t *testing.T) {
|
|||||||
},
|
},
|
||||||
}),
|
}),
|
||||||
expectedEvictedCount: 0,
|
expectedEvictedCount: 0,
|
||||||
namespaces: []string{"ns1"},
|
strategy: api.DeschedulerStrategy{
|
||||||
args: componentconfig.RemovePodsViolatingTopologySpreadConstraintArgs{LabelSelector: getLabelSelector("foo", []string{"baz"}, metav1.LabelSelectorOpIn)},
|
Params: &api.StrategyParameters{
|
||||||
|
LabelSelector: getLabelSelector("foo", []string{"baz"}, metav1.LabelSelectorOpIn),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
namespaces: []string{"ns1"},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "2 domains, sizes [2,0], maxSkew=1, move 1 pod for node with matched label filtering",
|
name: "2 domains, sizes [2,0], maxSkew=1, move 1 pod for node with matched label filtering",
|
||||||
@@ -833,9 +798,12 @@ func TestTopologySpreadConstraint(t *testing.T) {
|
|||||||
},
|
},
|
||||||
}),
|
}),
|
||||||
expectedEvictedCount: 1,
|
expectedEvictedCount: 1,
|
||||||
expectedEvictedPods: []string{"pod-1"},
|
strategy: api.DeschedulerStrategy{
|
||||||
namespaces: []string{"ns1"},
|
Params: &api.StrategyParameters{
|
||||||
args: componentconfig.RemovePodsViolatingTopologySpreadConstraintArgs{LabelSelector: getLabelSelector("foo", []string{"bar"}, metav1.LabelSelectorOpIn)},
|
LabelSelector: getLabelSelector("foo", []string{"bar"}, metav1.LabelSelectorOpIn),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
namespaces: []string{"ns1"},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "2 domains, sizes [2,0], maxSkew=1, move 1 pod for node with matched label filtering (NotIn op)",
|
name: "2 domains, sizes [2,0], maxSkew=1, move 1 pod for node with matched label filtering (NotIn op)",
|
||||||
@@ -857,390 +825,57 @@ func TestTopologySpreadConstraint(t *testing.T) {
|
|||||||
},
|
},
|
||||||
}),
|
}),
|
||||||
expectedEvictedCount: 1,
|
expectedEvictedCount: 1,
|
||||||
expectedEvictedPods: []string{"pod-1"},
|
strategy: api.DeschedulerStrategy{
|
||||||
namespaces: []string{"ns1"},
|
Params: &api.StrategyParameters{
|
||||||
args: componentconfig.RemovePodsViolatingTopologySpreadConstraintArgs{LabelSelector: getLabelSelector("foo", []string{"baz"}, metav1.LabelSelectorOpNotIn)},
|
LabelSelector: getLabelSelector("foo", []string{"baz"}, metav1.LabelSelectorOpNotIn),
|
||||||
},
|
},
|
||||||
{
|
|
||||||
name: "2 domains, sizes [4,2], maxSkew=1, 2 pods in termination; nothing should be moved",
|
|
||||||
nodes: []*v1.Node{
|
|
||||||
test.BuildTestNode("n1", 2000, 3000, 10, func(n *v1.Node) { n.Labels["zone"] = "zoneA" }),
|
|
||||||
test.BuildTestNode("n2", 2000, 3000, 10, func(n *v1.Node) { n.Labels["zone"] = "zoneB" }),
|
|
||||||
},
|
},
|
||||||
pods: createTestPods([]testPodList{
|
namespaces: []string{"ns1"},
|
||||||
{
|
|
||||||
count: 2,
|
|
||||||
node: "n1",
|
|
||||||
labels: map[string]string{"foo": "bar"},
|
|
||||||
constraints: getDefaultTopologyConstraints(1),
|
|
||||||
},
|
|
||||||
{
|
|
||||||
count: 2,
|
|
||||||
node: "n1",
|
|
||||||
labels: map[string]string{"foo": "bar"},
|
|
||||||
constraints: getDefaultTopologyConstraints(1),
|
|
||||||
deletionTimestamp: &metav1.Time{},
|
|
||||||
},
|
|
||||||
{
|
|
||||||
count: 2,
|
|
||||||
node: "n2",
|
|
||||||
labels: map[string]string{"foo": "bar"},
|
|
||||||
constraints: getDefaultTopologyConstraints(1),
|
|
||||||
},
|
|
||||||
}),
|
|
||||||
expectedEvictedCount: 0,
|
|
||||||
namespaces: []string{"ns1"},
|
|
||||||
args: componentconfig.RemovePodsViolatingTopologySpreadConstraintArgs{LabelSelector: getLabelSelector("foo", []string{"bar"}, metav1.LabelSelectorOpIn)},
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "3 domains, sizes [2,3,4], maxSkew=1, NodeFit is enabled, and not enough cpu on zoneA; nothing should be moved",
|
|
||||||
nodes: []*v1.Node{
|
|
||||||
test.BuildTestNode("n1", 250, 2000, 9, func(n *v1.Node) { n.Labels["zone"] = "zoneA" }),
|
|
||||||
test.BuildTestNode("n2", 1000, 2000, 9, func(n *v1.Node) { n.Labels["zone"] = "zoneB" }),
|
|
||||||
test.BuildTestNode("n3", 1000, 2000, 9, func(n *v1.Node) { n.Labels["zone"] = "zoneC" }),
|
|
||||||
},
|
|
||||||
pods: createTestPods([]testPodList{
|
|
||||||
{
|
|
||||||
count: 2,
|
|
||||||
node: "n1",
|
|
||||||
labels: map[string]string{"foo": "bar"},
|
|
||||||
constraints: getDefaultTopologyConstraints(1),
|
|
||||||
},
|
|
||||||
{
|
|
||||||
count: 3,
|
|
||||||
node: "n2",
|
|
||||||
labels: map[string]string{"foo": "bar"},
|
|
||||||
constraints: getDefaultTopologyConstraints(1),
|
|
||||||
},
|
|
||||||
{
|
|
||||||
count: 4,
|
|
||||||
node: "n3",
|
|
||||||
labels: map[string]string{"foo": "bar"},
|
|
||||||
constraints: getDefaultTopologyConstraints(1),
|
|
||||||
},
|
|
||||||
}),
|
|
||||||
expectedEvictedCount: 0,
|
|
||||||
namespaces: []string{"ns1"},
|
|
||||||
args: componentconfig.RemovePodsViolatingTopologySpreadConstraintArgs{},
|
|
||||||
nodeFit: true,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "3 domains, sizes [[1,0], [1,1], [2,1]], maxSkew=1, NodeFit is enabled, and not enough cpu on ZoneA; nothing should be moved",
|
|
||||||
nodes: []*v1.Node{
|
|
||||||
test.BuildTestNode("A1", 150, 2000, 9, func(n *v1.Node) { n.Labels["zone"] = "zoneA" }),
|
|
||||||
test.BuildTestNode("B1", 1000, 2000, 9, func(n *v1.Node) { n.Labels["zone"] = "zoneB" }),
|
|
||||||
test.BuildTestNode("C1", 1000, 2000, 9, func(n *v1.Node) { n.Labels["zone"] = "zoneC" }),
|
|
||||||
test.BuildTestNode("A2", 50, 2000, 9, func(n *v1.Node) { n.Labels["zone"] = "zoneA" }),
|
|
||||||
test.BuildTestNode("B2", 1000, 2000, 9, func(n *v1.Node) { n.Labels["zone"] = "zoneB" }),
|
|
||||||
test.BuildTestNode("C2", 1000, 2000, 9, func(n *v1.Node) { n.Labels["zone"] = "zoneC" }),
|
|
||||||
},
|
|
||||||
pods: createTestPods([]testPodList{
|
|
||||||
{
|
|
||||||
count: 1,
|
|
||||||
node: "A1",
|
|
||||||
labels: map[string]string{"foo": "bar"},
|
|
||||||
constraints: getDefaultTopologyConstraints(1),
|
|
||||||
},
|
|
||||||
{
|
|
||||||
count: 1,
|
|
||||||
node: "B1",
|
|
||||||
labels: map[string]string{"foo": "bar"},
|
|
||||||
constraints: getDefaultTopologyConstraints(1),
|
|
||||||
},
|
|
||||||
{
|
|
||||||
count: 1,
|
|
||||||
node: "B2",
|
|
||||||
labels: map[string]string{"foo": "bar"},
|
|
||||||
constraints: getDefaultTopologyConstraints(1),
|
|
||||||
},
|
|
||||||
{
|
|
||||||
count: 2,
|
|
||||||
node: "C1",
|
|
||||||
labels: map[string]string{"foo": "bar"},
|
|
||||||
constraints: getDefaultTopologyConstraints(1),
|
|
||||||
},
|
|
||||||
{
|
|
||||||
count: 1,
|
|
||||||
node: "C2",
|
|
||||||
labels: map[string]string{"foo": "bar"},
|
|
||||||
constraints: getDefaultTopologyConstraints(1),
|
|
||||||
},
|
|
||||||
}),
|
|
||||||
expectedEvictedCount: 0,
|
|
||||||
namespaces: []string{"ns1"},
|
|
||||||
args: componentconfig.RemovePodsViolatingTopologySpreadConstraintArgs{},
|
|
||||||
nodeFit: true,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "2 domains, sizes [[1,4], [2,1]], maxSkew=1, NodeFit is enabled; should move 1",
|
|
||||||
nodes: []*v1.Node{
|
|
||||||
test.BuildTestNode("A1", 1000, 2000, 9, func(n *v1.Node) { n.Labels["zone"] = "zoneA" }),
|
|
||||||
test.BuildTestNode("A2", 1000, 2000, 9, func(n *v1.Node) { n.Labels["zone"] = "zoneA" }),
|
|
||||||
test.BuildTestNode("B1", 1000, 2000, 9, func(n *v1.Node) { n.Labels["zone"] = "zoneB" }),
|
|
||||||
test.BuildTestNode("B2", 1000, 2000, 9, func(n *v1.Node) { n.Labels["zone"] = "zoneB" }),
|
|
||||||
},
|
|
||||||
pods: createTestPods([]testPodList{
|
|
||||||
{
|
|
||||||
count: 1,
|
|
||||||
node: "A1",
|
|
||||||
labels: map[string]string{"foo": "bar"},
|
|
||||||
constraints: getDefaultTopologyConstraints(1),
|
|
||||||
},
|
|
||||||
{
|
|
||||||
count: 4,
|
|
||||||
node: "A2",
|
|
||||||
labels: map[string]string{"foo": "bar"},
|
|
||||||
constraints: getDefaultTopologyConstraints(1),
|
|
||||||
},
|
|
||||||
{
|
|
||||||
count: 2,
|
|
||||||
node: "B1",
|
|
||||||
labels: map[string]string{"foo": "bar"},
|
|
||||||
constraints: getDefaultTopologyConstraints(1),
|
|
||||||
},
|
|
||||||
{
|
|
||||||
count: 1,
|
|
||||||
node: "B2",
|
|
||||||
labels: map[string]string{"foo": "bar"},
|
|
||||||
constraints: getDefaultTopologyConstraints(1),
|
|
||||||
},
|
|
||||||
}),
|
|
||||||
expectedEvictedCount: 1,
|
|
||||||
expectedEvictedPods: []string{"pod-4"},
|
|
||||||
namespaces: []string{"ns1"},
|
|
||||||
args: componentconfig.RemovePodsViolatingTopologySpreadConstraintArgs{},
|
|
||||||
nodeFit: true,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
// https://github.com/kubernetes-sigs/descheduler/issues/839
|
|
||||||
name: "2 domains, sizes [3, 1], maxSkew=1, Tainted nodes; nothing should be moved",
|
|
||||||
nodes: []*v1.Node{
|
|
||||||
test.BuildTestNode("controlplane1", 2000, 3000, 10, func(n *v1.Node) {
|
|
||||||
n.Labels["zone"] = "zoneA"
|
|
||||||
n.Labels["kubernetes.io/hostname"] = "cp-1"
|
|
||||||
n.Spec.Taints = []v1.Taint{
|
|
||||||
{
|
|
||||||
Effect: v1.TaintEffectNoSchedule,
|
|
||||||
Key: "node-role.kubernetes.io/controlplane",
|
|
||||||
Value: "true",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Effect: v1.TaintEffectNoSchedule,
|
|
||||||
Key: "node-role.kubernetes.io/etcd",
|
|
||||||
Value: "true",
|
|
||||||
},
|
|
||||||
}
|
|
||||||
}),
|
|
||||||
test.BuildTestNode("infraservices1", 2000, 3000, 10, func(n *v1.Node) {
|
|
||||||
n.Labels["zone"] = "zoneA"
|
|
||||||
n.Labels["infra_service"] = "true"
|
|
||||||
n.Labels["kubernetes.io/hostname"] = "infra-1"
|
|
||||||
n.Spec.Taints = []v1.Taint{
|
|
||||||
{
|
|
||||||
Effect: v1.TaintEffectNoSchedule,
|
|
||||||
Key: "infra_service",
|
|
||||||
},
|
|
||||||
}
|
|
||||||
}),
|
|
||||||
test.BuildTestNode("app1", 2000, 3000, 10, func(n *v1.Node) {
|
|
||||||
n.Labels["zone"] = "zoneA"
|
|
||||||
n.Labels["app_service"] = "true"
|
|
||||||
n.Labels["kubernetes.io/hostname"] = "app-1"
|
|
||||||
}),
|
|
||||||
test.BuildTestNode("app2", 2000, 3000, 10, func(n *v1.Node) {
|
|
||||||
n.Labels["zone"] = "zoneB"
|
|
||||||
n.Labels["app_service"] = "true"
|
|
||||||
n.Labels["kubernetes.io/hostname"] = "app-2"
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
pods: createTestPods([]testPodList{
|
|
||||||
{
|
|
||||||
count: 2,
|
|
||||||
node: "app1",
|
|
||||||
labels: map[string]string{"app.kubernetes.io/name": "service"},
|
|
||||||
constraints: []v1.TopologySpreadConstraint{
|
|
||||||
{
|
|
||||||
MaxSkew: 1,
|
|
||||||
TopologyKey: "kubernetes.io/hostname",
|
|
||||||
WhenUnsatisfiable: v1.DoNotSchedule,
|
|
||||||
LabelSelector: &metav1.LabelSelector{MatchLabels: map[string]string{"app.kubernetes.io/name": "service"}},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
nodeAffinity: &v1.Affinity{
|
|
||||||
NodeAffinity: &v1.NodeAffinity{
|
|
||||||
RequiredDuringSchedulingIgnoredDuringExecution: &v1.NodeSelector{
|
|
||||||
NodeSelectorTerms: []v1.NodeSelectorTerm{
|
|
||||||
{
|
|
||||||
MatchExpressions: []v1.NodeSelectorRequirement{
|
|
||||||
{
|
|
||||||
Key: "app_service",
|
|
||||||
Operator: v1.NodeSelectorOpExists,
|
|
||||||
},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
{
|
|
||||||
count: 2,
|
|
||||||
node: "app2",
|
|
||||||
labels: map[string]string{"app.kubernetes.io/name": "service"},
|
|
||||||
constraints: []v1.TopologySpreadConstraint{
|
|
||||||
{
|
|
||||||
MaxSkew: 1,
|
|
||||||
TopologyKey: "kubernetes.io/hostname",
|
|
||||||
WhenUnsatisfiable: v1.DoNotSchedule,
|
|
||||||
LabelSelector: &metav1.LabelSelector{MatchLabels: map[string]string{"app.kubernetes.io/name": "service"}},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
nodeAffinity: &v1.Affinity{
|
|
||||||
NodeAffinity: &v1.NodeAffinity{
|
|
||||||
RequiredDuringSchedulingIgnoredDuringExecution: &v1.NodeSelector{
|
|
||||||
NodeSelectorTerms: []v1.NodeSelectorTerm{
|
|
||||||
{
|
|
||||||
MatchExpressions: []v1.NodeSelectorRequirement{
|
|
||||||
{
|
|
||||||
Key: "app_service",
|
|
||||||
Operator: v1.NodeSelectorOpExists,
|
|
||||||
},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
}),
|
|
||||||
expectedEvictedCount: 0,
|
|
||||||
namespaces: []string{"ns1"},
|
|
||||||
args: componentconfig.RemovePodsViolatingTopologySpreadConstraintArgs{},
|
|
||||||
nodeFit: true,
|
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
for _, tc := range testCases {
|
for _, tc := range testCases {
|
||||||
t.Run(tc.name, func(t *testing.T) {
|
t.Run(tc.name, func(t *testing.T) {
|
||||||
ctx, cancel := context.WithCancel(context.Background())
|
fakeClient := &fake.Clientset{}
|
||||||
defer cancel()
|
fakeClient.Fake.AddReactor("list", "pods", func(action core.Action) (bool, runtime.Object, error) {
|
||||||
|
podList := make([]v1.Pod, 0, len(tc.pods))
|
||||||
var objs []runtime.Object
|
for _, pod := range tc.pods {
|
||||||
for _, node := range tc.nodes {
|
podList = append(podList, *pod)
|
||||||
objs = append(objs, node)
|
|
||||||
}
|
|
||||||
for _, pod := range tc.pods {
|
|
||||||
objs = append(objs, pod)
|
|
||||||
}
|
|
||||||
objs = append(objs, &v1.Namespace{ObjectMeta: metav1.ObjectMeta{Name: "ns1"}})
|
|
||||||
fakeClient := fake.NewSimpleClientset(objs...)
|
|
||||||
|
|
||||||
sharedInformerFactory := informers.NewSharedInformerFactory(fakeClient, 0)
|
|
||||||
podInformer := sharedInformerFactory.Core().V1().Pods()
|
|
||||||
|
|
||||||
getPodsAssignedToNode, err := podutil.BuildGetPodsAssignedToNodeFunc(podInformer)
|
|
||||||
if err != nil {
|
|
||||||
t.Errorf("Build get pods assigned to node function error: %v", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
sharedInformerFactory.Start(ctx.Done())
|
|
||||||
sharedInformerFactory.WaitForCacheSync(ctx.Done())
|
|
||||||
|
|
||||||
var evictedPods []string
|
|
||||||
fakeClient.PrependReactor("create", "pods", func(action core.Action) (bool, runtime.Object, error) {
|
|
||||||
if action.GetSubresource() == "eviction" {
|
|
||||||
createAct, matched := action.(core.CreateActionImpl)
|
|
||||||
if !matched {
|
|
||||||
return false, nil, fmt.Errorf("unable to convert action to core.CreateActionImpl")
|
|
||||||
}
|
|
||||||
if eviction, matched := createAct.Object.(*policy.Eviction); matched {
|
|
||||||
evictedPods = append(evictedPods, eviction.GetName())
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
return false, nil, nil // fallback to the default reactor
|
return true, &v1.PodList{Items: podList}, nil
|
||||||
|
})
|
||||||
|
fakeClient.Fake.AddReactor("list", "namespaces", func(action core.Action) (bool, runtime.Object, error) {
|
||||||
|
return true, &v1.NamespaceList{Items: []v1.Namespace{{ObjectMeta: metav1.ObjectMeta{Name: "ns1", Namespace: "ns1"}}}}, nil
|
||||||
})
|
})
|
||||||
|
|
||||||
eventRecorder := &events.FakeRecorder{}
|
|
||||||
|
|
||||||
podEvictor := evictions.NewPodEvictor(
|
podEvictor := evictions.NewPodEvictor(
|
||||||
fakeClient,
|
fakeClient,
|
||||||
"v1",
|
"v1",
|
||||||
false,
|
false,
|
||||||
nil,
|
100,
|
||||||
nil,
|
|
||||||
tc.nodes,
|
tc.nodes,
|
||||||
false,
|
false,
|
||||||
eventRecorder,
|
false,
|
||||||
|
false,
|
||||||
)
|
)
|
||||||
|
RemovePodsViolatingTopologySpreadConstraint(ctx, fakeClient, tc.strategy, tc.nodes, podEvictor)
|
||||||
defaultevictorArgs := &defaultevictor.DefaultEvictorArgs{
|
|
||||||
EvictLocalStoragePods: false,
|
|
||||||
EvictSystemCriticalPods: false,
|
|
||||||
IgnorePvcPods: false,
|
|
||||||
EvictFailedBarePods: false,
|
|
||||||
NodeFit: tc.nodeFit,
|
|
||||||
}
|
|
||||||
|
|
||||||
evictorFilter, err := defaultevictor.New(
|
|
||||||
defaultevictorArgs,
|
|
||||||
&frameworkfake.HandleImpl{
|
|
||||||
ClientsetImpl: fakeClient,
|
|
||||||
GetPodsAssignedToNodeFuncImpl: getPodsAssignedToNode,
|
|
||||||
SharedInformerFactoryImpl: sharedInformerFactory,
|
|
||||||
},
|
|
||||||
)
|
|
||||||
|
|
||||||
if err != nil {
|
|
||||||
t.Fatalf("Unable to initialize the plugin: %v", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
handle := &frameworkfake.HandleImpl{
|
|
||||||
ClientsetImpl: fakeClient,
|
|
||||||
GetPodsAssignedToNodeFuncImpl: getPodsAssignedToNode,
|
|
||||||
PodEvictorImpl: podEvictor,
|
|
||||||
EvictorFilterImpl: evictorFilter.(framework.EvictorPlugin),
|
|
||||||
SharedInformerFactoryImpl: sharedInformerFactory,
|
|
||||||
}
|
|
||||||
|
|
||||||
plugin, err := New(
|
|
||||||
&tc.args,
|
|
||||||
handle,
|
|
||||||
)
|
|
||||||
if err != nil {
|
|
||||||
t.Fatalf("Unable to initialize the plugin: %v", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
plugin.(framework.BalancePlugin).Balance(ctx, tc.nodes)
|
|
||||||
podsEvicted := podEvictor.TotalEvicted()
|
podsEvicted := podEvictor.TotalEvicted()
|
||||||
if podsEvicted != tc.expectedEvictedCount {
|
if podsEvicted != tc.expectedEvictedCount {
|
||||||
t.Errorf("Test error for description: %s. Expected evicted pods count %v, got %v", tc.name, tc.expectedEvictedCount, podsEvicted)
|
t.Errorf("Test error for description: %s. Expected evicted pods count %v, got %v", tc.name, tc.expectedEvictedCount, podsEvicted)
|
||||||
}
|
}
|
||||||
|
|
||||||
if tc.expectedEvictedPods != nil {
|
|
||||||
diff := sets.NewString(tc.expectedEvictedPods...).Difference(sets.NewString(evictedPods...))
|
|
||||||
if diff.Len() > 0 {
|
|
||||||
t.Errorf(
|
|
||||||
"Expected pods %v to be evicted but %v were not evicted. Actual pods evicted: %v",
|
|
||||||
tc.expectedEvictedPods,
|
|
||||||
diff.List(),
|
|
||||||
evictedPods,
|
|
||||||
)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
type testPodList struct {
|
type testPodList struct {
|
||||||
count int
|
count int
|
||||||
node string
|
node string
|
||||||
labels map[string]string
|
labels map[string]string
|
||||||
constraints []v1.TopologySpreadConstraint
|
constraints []v1.TopologySpreadConstraint
|
||||||
nodeSelector map[string]string
|
nodeSelector map[string]string
|
||||||
nodeAffinity *v1.Affinity
|
nodeAffinity *v1.Affinity
|
||||||
noOwners bool
|
noOwners bool
|
||||||
tolerations []v1.Toleration
|
tolerations []v1.Toleration
|
||||||
deletionTimestamp *metav1.Time
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func createTestPods(testPods []testPodList) []*v1.Pod {
|
func createTestPods(testPods []testPodList) []*v1.Pod {
|
||||||
@@ -1261,7 +896,6 @@ func createTestPods(testPods []testPodList) []*v1.Pod {
|
|||||||
p.Spec.NodeSelector = tp.nodeSelector
|
p.Spec.NodeSelector = tp.nodeSelector
|
||||||
p.Spec.Affinity = tp.nodeAffinity
|
p.Spec.Affinity = tp.nodeAffinity
|
||||||
p.Spec.Tolerations = tp.tolerations
|
p.Spec.Tolerations = tp.tolerations
|
||||||
p.ObjectMeta.DeletionTimestamp = tp.deletionTimestamp
|
|
||||||
}))
|
}))
|
||||||
podNum++
|
podNum++
|
||||||
}
|
}
|
||||||
@@ -1285,54 +919,3 @@ func getDefaultTopologyConstraints(maxSkew int32) []v1.TopologySpreadConstraint
|
|||||||
},
|
},
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestCheckIdenticalConstraints(t *testing.T) {
|
|
||||||
newConstraintSame := v1.TopologySpreadConstraint{
|
|
||||||
MaxSkew: 2,
|
|
||||||
TopologyKey: "zone",
|
|
||||||
WhenUnsatisfiable: v1.DoNotSchedule,
|
|
||||||
LabelSelector: &metav1.LabelSelector{MatchLabels: map[string]string{"foo": "bar"}},
|
|
||||||
}
|
|
||||||
newConstraintDifferent := v1.TopologySpreadConstraint{
|
|
||||||
MaxSkew: 3,
|
|
||||||
TopologyKey: "node",
|
|
||||||
WhenUnsatisfiable: v1.DoNotSchedule,
|
|
||||||
LabelSelector: &metav1.LabelSelector{MatchLabels: map[string]string{"foo": "bar"}},
|
|
||||||
}
|
|
||||||
namespaceTopologySpreadConstraint := []v1.TopologySpreadConstraint{}
|
|
||||||
namespaceTopologySpreadConstraint = []v1.TopologySpreadConstraint{
|
|
||||||
{
|
|
||||||
MaxSkew: 2,
|
|
||||||
TopologyKey: "zone",
|
|
||||||
WhenUnsatisfiable: v1.DoNotSchedule,
|
|
||||||
LabelSelector: &metav1.LabelSelector{MatchLabels: map[string]string{"foo": "bar"}},
|
|
||||||
},
|
|
||||||
}
|
|
||||||
testCases := []struct {
|
|
||||||
name string
|
|
||||||
namespaceTopologySpreadConstraints []v1.TopologySpreadConstraint
|
|
||||||
newConstraint v1.TopologySpreadConstraint
|
|
||||||
expectedResult bool
|
|
||||||
}{
|
|
||||||
{
|
|
||||||
name: "new constraint is identical",
|
|
||||||
namespaceTopologySpreadConstraints: namespaceTopologySpreadConstraint,
|
|
||||||
newConstraint: newConstraintSame,
|
|
||||||
expectedResult: true,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "new constraint is different",
|
|
||||||
namespaceTopologySpreadConstraints: namespaceTopologySpreadConstraint,
|
|
||||||
newConstraint: newConstraintDifferent,
|
|
||||||
expectedResult: false,
|
|
||||||
},
|
|
||||||
}
|
|
||||||
for _, tc := range testCases {
|
|
||||||
t.Run(tc.name, func(t *testing.T) {
|
|
||||||
isIdentical := hasIdenticalConstraints(tc.newConstraint, tc.namespaceTopologySpreadConstraints)
|
|
||||||
if isIdentical != tc.expectedResult {
|
|
||||||
t.Errorf("Test error for description: %s. Expected result %v, got %v", tc.name, tc.expectedResult, isIdentical)
|
|
||||||
}
|
|
||||||
})
|
|
||||||
}
|
|
||||||
}
|
|
||||||
@@ -22,22 +22,20 @@ type ValidatedStrategyParams struct {
|
|||||||
NodeFit bool
|
NodeFit bool
|
||||||
}
|
}
|
||||||
|
|
||||||
func DefaultValidatedStrategyParams() ValidatedStrategyParams {
|
|
||||||
return ValidatedStrategyParams{ThresholdPriority: utils.SystemCriticalPriority}
|
|
||||||
}
|
|
||||||
|
|
||||||
func ValidateAndParseStrategyParams(
|
func ValidateAndParseStrategyParams(
|
||||||
ctx context.Context,
|
ctx context.Context,
|
||||||
client clientset.Interface,
|
client clientset.Interface,
|
||||||
params *api.StrategyParameters,
|
params *api.StrategyParameters,
|
||||||
) (*ValidatedStrategyParams, error) {
|
) (*ValidatedStrategyParams, error) {
|
||||||
|
var includedNamespaces, excludedNamespaces sets.String
|
||||||
if params == nil {
|
if params == nil {
|
||||||
defaultValidatedStrategyParams := DefaultValidatedStrategyParams()
|
return &ValidatedStrategyParams{
|
||||||
return &defaultValidatedStrategyParams, nil
|
IncludedNamespaces: includedNamespaces,
|
||||||
|
ExcludedNamespaces: excludedNamespaces,
|
||||||
|
}, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// At most one of include/exclude can be set
|
// At most one of include/exclude can be set
|
||||||
var includedNamespaces, excludedNamespaces sets.String
|
|
||||||
if params.Namespaces != nil && len(params.Namespaces.Include) > 0 && len(params.Namespaces.Exclude) > 0 {
|
if params.Namespaces != nil && len(params.Namespaces.Include) > 0 && len(params.Namespaces.Exclude) > 0 {
|
||||||
return nil, fmt.Errorf("only one of Include/Exclude namespaces can be set")
|
return nil, fmt.Errorf("only one of Include/Exclude namespaces can be set")
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,273 +0,0 @@
|
|||||||
/*
|
|
||||||
Copyright 2022 The Kubernetes Authors.
|
|
||||||
|
|
||||||
Licensed under the Apache License, Version 2.0 (the "License");
|
|
||||||
you may not use this file except in compliance with the License.
|
|
||||||
You may obtain a copy of the License at
|
|
||||||
|
|
||||||
http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
|
|
||||||
Unless required by applicable law or agreed to in writing, software
|
|
||||||
distributed under the License is distributed on an "AS IS" BASIS,
|
|
||||||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
||||||
See the License for the specific language governing permissions and
|
|
||||||
limitations under the License.
|
|
||||||
*/
|
|
||||||
|
|
||||||
package descheduler
|
|
||||||
|
|
||||||
import (
|
|
||||||
"context"
|
|
||||||
|
|
||||||
v1 "k8s.io/api/core/v1"
|
|
||||||
"k8s.io/klog/v2"
|
|
||||||
"sigs.k8s.io/descheduler/pkg/api"
|
|
||||||
"sigs.k8s.io/descheduler/pkg/apis/componentconfig"
|
|
||||||
"sigs.k8s.io/descheduler/pkg/apis/componentconfig/validation"
|
|
||||||
"sigs.k8s.io/descheduler/pkg/framework"
|
|
||||||
"sigs.k8s.io/descheduler/pkg/framework/plugins/nodeutilization"
|
|
||||||
"sigs.k8s.io/descheduler/pkg/framework/plugins/podlifetime"
|
|
||||||
"sigs.k8s.io/descheduler/pkg/framework/plugins/removeduplicates"
|
|
||||||
"sigs.k8s.io/descheduler/pkg/framework/plugins/removefailedpods"
|
|
||||||
"sigs.k8s.io/descheduler/pkg/framework/plugins/removepodshavingtoomanyrestarts"
|
|
||||||
"sigs.k8s.io/descheduler/pkg/framework/plugins/removepodsviolatinginterpodantiaffinity"
|
|
||||||
"sigs.k8s.io/descheduler/pkg/framework/plugins/removepodsviolatingnodeaffinity"
|
|
||||||
"sigs.k8s.io/descheduler/pkg/framework/plugins/removepodsviolatingnodetaints"
|
|
||||||
"sigs.k8s.io/descheduler/pkg/framework/plugins/removepodsviolatingtopologyspreadconstraint"
|
|
||||||
)
|
|
||||||
|
|
||||||
// Once all strategies are migrated the arguments get read from the configuration file
|
|
||||||
// without any wiring. Keeping the wiring here so the descheduler can still use
|
|
||||||
// the v1alpha1 configuration during the strategy migration to plugins.
|
|
||||||
|
|
||||||
var pluginsMap = map[string]func(ctx context.Context, nodes []*v1.Node, params *api.StrategyParameters, handle *handleImpl){
|
|
||||||
"RemovePodsViolatingNodeTaints": func(ctx context.Context, nodes []*v1.Node, params *api.StrategyParameters, handle *handleImpl) {
|
|
||||||
args := &componentconfig.RemovePodsViolatingNodeTaintsArgs{
|
|
||||||
Namespaces: params.Namespaces,
|
|
||||||
LabelSelector: params.LabelSelector,
|
|
||||||
IncludePreferNoSchedule: params.IncludePreferNoSchedule,
|
|
||||||
ExcludedTaints: params.ExcludedTaints,
|
|
||||||
}
|
|
||||||
if err := validation.ValidateRemovePodsViolatingNodeTaintsArgs(args); err != nil {
|
|
||||||
klog.V(1).ErrorS(err, "unable to validate plugin arguments", "pluginName", removepodsviolatingnodetaints.PluginName)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
pg, err := removepodsviolatingnodetaints.New(args, handle)
|
|
||||||
if err != nil {
|
|
||||||
klog.V(1).ErrorS(err, "unable to initialize a plugin", "pluginName", removepodsviolatingnodetaints.PluginName)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
status := pg.(framework.DeschedulePlugin).Deschedule(ctx, nodes)
|
|
||||||
if status != nil && status.Err != nil {
|
|
||||||
klog.V(1).ErrorS(err, "plugin finished with error", "pluginName", removepodsviolatingnodetaints.PluginName)
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"RemoveFailedPods": func(ctx context.Context, nodes []*v1.Node, params *api.StrategyParameters, handle *handleImpl) {
|
|
||||||
failedPodsParams := params.FailedPods
|
|
||||||
if failedPodsParams == nil {
|
|
||||||
failedPodsParams = &api.FailedPods{}
|
|
||||||
}
|
|
||||||
args := &componentconfig.RemoveFailedPodsArgs{
|
|
||||||
Namespaces: params.Namespaces,
|
|
||||||
LabelSelector: params.LabelSelector,
|
|
||||||
IncludingInitContainers: failedPodsParams.IncludingInitContainers,
|
|
||||||
MinPodLifetimeSeconds: failedPodsParams.MinPodLifetimeSeconds,
|
|
||||||
ExcludeOwnerKinds: failedPodsParams.ExcludeOwnerKinds,
|
|
||||||
Reasons: failedPodsParams.Reasons,
|
|
||||||
}
|
|
||||||
if err := validation.ValidateRemoveFailedPodsArgs(args); err != nil {
|
|
||||||
klog.V(1).ErrorS(err, "unable to validate plugin arguments", "pluginName", removefailedpods.PluginName)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
pg, err := removefailedpods.New(args, handle)
|
|
||||||
if err != nil {
|
|
||||||
klog.V(1).ErrorS(err, "unable to initialize a plugin", "pluginName", removefailedpods.PluginName)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
status := pg.(framework.DeschedulePlugin).Deschedule(ctx, nodes)
|
|
||||||
if status != nil && status.Err != nil {
|
|
||||||
klog.V(1).ErrorS(err, "plugin finished with error", "pluginName", removefailedpods.PluginName)
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"RemovePodsViolatingNodeAffinity": func(ctx context.Context, nodes []*v1.Node, params *api.StrategyParameters, handle *handleImpl) {
|
|
||||||
args := &componentconfig.RemovePodsViolatingNodeAffinityArgs{
|
|
||||||
Namespaces: params.Namespaces,
|
|
||||||
LabelSelector: params.LabelSelector,
|
|
||||||
NodeAffinityType: params.NodeAffinityType,
|
|
||||||
}
|
|
||||||
if err := validation.ValidateRemovePodsViolatingNodeAffinityArgs(args); err != nil {
|
|
||||||
klog.V(1).ErrorS(err, "unable to validate plugin arguments", "pluginName", removepodsviolatingnodeaffinity.PluginName)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
pg, err := removepodsviolatingnodeaffinity.New(args, handle)
|
|
||||||
if err != nil {
|
|
||||||
klog.V(1).ErrorS(err, "unable to initialize a plugin", "pluginName", removepodsviolatingnodeaffinity.PluginName)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
status := pg.(framework.DeschedulePlugin).Deschedule(ctx, nodes)
|
|
||||||
if status != nil && status.Err != nil {
|
|
||||||
klog.V(1).ErrorS(err, "plugin finished with error", "pluginName", removepodsviolatingnodeaffinity.PluginName)
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"RemovePodsViolatingInterPodAntiAffinity": func(ctx context.Context, nodes []*v1.Node, params *api.StrategyParameters, handle *handleImpl) {
|
|
||||||
args := &componentconfig.RemovePodsViolatingInterPodAntiAffinityArgs{
|
|
||||||
Namespaces: params.Namespaces,
|
|
||||||
LabelSelector: params.LabelSelector,
|
|
||||||
}
|
|
||||||
if err := validation.ValidateRemovePodsViolatingInterPodAntiAffinityArgs(args); err != nil {
|
|
||||||
klog.V(1).ErrorS(err, "unable to validate plugin arguments", "pluginName", removepodsviolatinginterpodantiaffinity.PluginName)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
pg, err := removepodsviolatinginterpodantiaffinity.New(args, handle)
|
|
||||||
if err != nil {
|
|
||||||
klog.V(1).ErrorS(err, "unable to initialize a plugin", "pluginName", removepodsviolatinginterpodantiaffinity.PluginName)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
status := pg.(framework.DeschedulePlugin).Deschedule(ctx, nodes)
|
|
||||||
if status != nil && status.Err != nil {
|
|
||||||
klog.V(1).ErrorS(err, "plugin finished with error", "pluginName", removepodsviolatinginterpodantiaffinity.PluginName)
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"RemovePodsHavingTooManyRestarts": func(ctx context.Context, nodes []*v1.Node, params *api.StrategyParameters, handle *handleImpl) {
|
|
||||||
tooManyRestartsParams := params.PodsHavingTooManyRestarts
|
|
||||||
if tooManyRestartsParams == nil {
|
|
||||||
tooManyRestartsParams = &api.PodsHavingTooManyRestarts{}
|
|
||||||
}
|
|
||||||
args := &componentconfig.RemovePodsHavingTooManyRestartsArgs{
|
|
||||||
Namespaces: params.Namespaces,
|
|
||||||
LabelSelector: params.LabelSelector,
|
|
||||||
PodRestartThreshold: tooManyRestartsParams.PodRestartThreshold,
|
|
||||||
IncludingInitContainers: tooManyRestartsParams.IncludingInitContainers,
|
|
||||||
}
|
|
||||||
if err := validation.ValidateRemovePodsHavingTooManyRestartsArgs(args); err != nil {
|
|
||||||
klog.V(1).ErrorS(err, "unable to validate plugin arguments", "pluginName", removepodshavingtoomanyrestarts.PluginName)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
pg, err := removepodshavingtoomanyrestarts.New(args, handle)
|
|
||||||
if err != nil {
|
|
||||||
klog.V(1).ErrorS(err, "unable to initialize a plugin", "pluginName", removepodshavingtoomanyrestarts.PluginName)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
status := pg.(framework.DeschedulePlugin).Deschedule(ctx, nodes)
|
|
||||||
if status != nil && status.Err != nil {
|
|
||||||
klog.V(1).ErrorS(err, "plugin finished with error", "pluginName", removepodshavingtoomanyrestarts.PluginName)
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"PodLifeTime": func(ctx context.Context, nodes []*v1.Node, params *api.StrategyParameters, handle *handleImpl) {
|
|
||||||
podLifeTimeParams := params.PodLifeTime
|
|
||||||
if podLifeTimeParams == nil {
|
|
||||||
podLifeTimeParams = &api.PodLifeTime{}
|
|
||||||
}
|
|
||||||
|
|
||||||
var states []string
|
|
||||||
if podLifeTimeParams.PodStatusPhases != nil {
|
|
||||||
states = append(states, podLifeTimeParams.PodStatusPhases...)
|
|
||||||
}
|
|
||||||
if podLifeTimeParams.States != nil {
|
|
||||||
states = append(states, podLifeTimeParams.States...)
|
|
||||||
}
|
|
||||||
|
|
||||||
args := &componentconfig.PodLifeTimeArgs{
|
|
||||||
Namespaces: params.Namespaces,
|
|
||||||
LabelSelector: params.LabelSelector,
|
|
||||||
MaxPodLifeTimeSeconds: podLifeTimeParams.MaxPodLifeTimeSeconds,
|
|
||||||
States: states,
|
|
||||||
}
|
|
||||||
if err := validation.ValidatePodLifeTimeArgs(args); err != nil {
|
|
||||||
klog.V(1).ErrorS(err, "unable to validate plugin arguments", "pluginName", podlifetime.PluginName)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
pg, err := podlifetime.New(args, handle)
|
|
||||||
if err != nil {
|
|
||||||
klog.V(1).ErrorS(err, "unable to initialize a plugin", "pluginName", podlifetime.PluginName)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
status := pg.(framework.DeschedulePlugin).Deschedule(ctx, nodes)
|
|
||||||
if status != nil && status.Err != nil {
|
|
||||||
klog.V(1).ErrorS(err, "plugin finished with error", "pluginName", podlifetime.PluginName)
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"RemoveDuplicates": func(ctx context.Context, nodes []*v1.Node, params *api.StrategyParameters, handle *handleImpl) {
|
|
||||||
args := &componentconfig.RemoveDuplicatesArgs{
|
|
||||||
Namespaces: params.Namespaces,
|
|
||||||
}
|
|
||||||
if params.RemoveDuplicates != nil {
|
|
||||||
args.ExcludeOwnerKinds = params.RemoveDuplicates.ExcludeOwnerKinds
|
|
||||||
}
|
|
||||||
if err := validation.ValidateRemoveDuplicatesArgs(args); err != nil {
|
|
||||||
klog.V(1).ErrorS(err, "unable to validate plugin arguments", "pluginName", removeduplicates.PluginName)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
pg, err := removeduplicates.New(args, handle)
|
|
||||||
if err != nil {
|
|
||||||
klog.V(1).ErrorS(err, "unable to initialize a plugin", "pluginName", removeduplicates.PluginName)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
status := pg.(framework.BalancePlugin).Balance(ctx, nodes)
|
|
||||||
if status != nil && status.Err != nil {
|
|
||||||
klog.V(1).ErrorS(err, "plugin finished with error", "pluginName", removeduplicates.PluginName)
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"RemovePodsViolatingTopologySpreadConstraint": func(ctx context.Context, nodes []*v1.Node, params *api.StrategyParameters, handle *handleImpl) {
|
|
||||||
args := &componentconfig.RemovePodsViolatingTopologySpreadConstraintArgs{
|
|
||||||
Namespaces: params.Namespaces,
|
|
||||||
LabelSelector: params.LabelSelector,
|
|
||||||
IncludeSoftConstraints: params.IncludeSoftConstraints,
|
|
||||||
}
|
|
||||||
if err := validation.ValidateRemovePodsViolatingTopologySpreadConstraintArgs(args); err != nil {
|
|
||||||
klog.V(1).ErrorS(err, "unable to validate plugin arguments", "pluginName", removepodsviolatingtopologyspreadconstraint.PluginName)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
pg, err := removepodsviolatingtopologyspreadconstraint.New(args, handle)
|
|
||||||
if err != nil {
|
|
||||||
klog.V(1).ErrorS(err, "unable to initialize a plugin", "pluginName", removepodsviolatingtopologyspreadconstraint.PluginName)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
status := pg.(framework.BalancePlugin).Balance(ctx, nodes)
|
|
||||||
if status != nil && status.Err != nil {
|
|
||||||
klog.V(1).ErrorS(err, "plugin finished with error", "pluginName", removepodsviolatingtopologyspreadconstraint.PluginName)
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"HighNodeUtilization": func(ctx context.Context, nodes []*v1.Node, params *api.StrategyParameters, handle *handleImpl) {
|
|
||||||
args := &componentconfig.HighNodeUtilizationArgs{
|
|
||||||
Thresholds: params.NodeResourceUtilizationThresholds.Thresholds,
|
|
||||||
NumberOfNodes: params.NodeResourceUtilizationThresholds.NumberOfNodes,
|
|
||||||
}
|
|
||||||
|
|
||||||
if err := validation.ValidateHighNodeUtilizationArgs(args); err != nil {
|
|
||||||
klog.V(1).ErrorS(err, "unable to validate plugin arguments", "pluginName", nodeutilization.HighNodeUtilizationPluginName)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
pg, err := nodeutilization.NewHighNodeUtilization(args, handle)
|
|
||||||
if err != nil {
|
|
||||||
klog.V(1).ErrorS(err, "unable to initialize a plugin", "pluginName", nodeutilization.HighNodeUtilizationPluginName)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
status := pg.(framework.BalancePlugin).Balance(ctx, nodes)
|
|
||||||
if status != nil && status.Err != nil {
|
|
||||||
klog.V(1).ErrorS(err, "plugin finished with error", "pluginName", nodeutilization.HighNodeUtilizationPluginName)
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"LowNodeUtilization": func(ctx context.Context, nodes []*v1.Node, params *api.StrategyParameters, handle *handleImpl) {
|
|
||||||
args := &componentconfig.LowNodeUtilizationArgs{
|
|
||||||
Thresholds: params.NodeResourceUtilizationThresholds.Thresholds,
|
|
||||||
TargetThresholds: params.NodeResourceUtilizationThresholds.TargetThresholds,
|
|
||||||
UseDeviationThresholds: params.NodeResourceUtilizationThresholds.UseDeviationThresholds,
|
|
||||||
NumberOfNodes: params.NodeResourceUtilizationThresholds.NumberOfNodes,
|
|
||||||
}
|
|
||||||
|
|
||||||
if err := validation.ValidateLowNodeUtilizationArgs(args); err != nil {
|
|
||||||
klog.V(1).ErrorS(err, "unable to validate plugin arguments", "pluginName", nodeutilization.LowNodeUtilizationPluginName)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
pg, err := nodeutilization.NewLowNodeUtilization(args, handle)
|
|
||||||
if err != nil {
|
|
||||||
klog.V(1).ErrorS(err, "unable to initialize a plugin", "pluginName", nodeutilization.LowNodeUtilizationPluginName)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
status := pg.(framework.BalancePlugin).Balance(ctx, nodes)
|
|
||||||
if status != nil && status.Err != nil {
|
|
||||||
klog.V(1).ErrorS(err, "plugin finished with error", "pluginName", nodeutilization.LowNodeUtilizationPluginName)
|
|
||||||
}
|
|
||||||
},
|
|
||||||
}
|
|
||||||
@@ -1,45 +0,0 @@
|
|||||||
package fake
|
|
||||||
|
|
||||||
import (
|
|
||||||
"context"
|
|
||||||
|
|
||||||
v1 "k8s.io/api/core/v1"
|
|
||||||
"k8s.io/client-go/informers"
|
|
||||||
clientset "k8s.io/client-go/kubernetes"
|
|
||||||
|
|
||||||
"sigs.k8s.io/descheduler/pkg/descheduler/evictions"
|
|
||||||
podutil "sigs.k8s.io/descheduler/pkg/descheduler/pod"
|
|
||||||
"sigs.k8s.io/descheduler/pkg/framework"
|
|
||||||
)
|
|
||||||
|
|
||||||
type HandleImpl struct {
|
|
||||||
ClientsetImpl clientset.Interface
|
|
||||||
GetPodsAssignedToNodeFuncImpl podutil.GetPodsAssignedToNodeFunc
|
|
||||||
SharedInformerFactoryImpl informers.SharedInformerFactory
|
|
||||||
EvictorFilterImpl framework.EvictorPlugin
|
|
||||||
PodEvictorImpl *evictions.PodEvictor
|
|
||||||
}
|
|
||||||
|
|
||||||
var _ framework.Handle = &HandleImpl{}
|
|
||||||
|
|
||||||
func (hi *HandleImpl) ClientSet() clientset.Interface {
|
|
||||||
return hi.ClientsetImpl
|
|
||||||
}
|
|
||||||
func (hi *HandleImpl) GetPodsAssignedToNodeFunc() podutil.GetPodsAssignedToNodeFunc {
|
|
||||||
return hi.GetPodsAssignedToNodeFuncImpl
|
|
||||||
}
|
|
||||||
func (hi *HandleImpl) SharedInformerFactory() informers.SharedInformerFactory {
|
|
||||||
return hi.SharedInformerFactoryImpl
|
|
||||||
}
|
|
||||||
func (hi *HandleImpl) Evictor() framework.Evictor {
|
|
||||||
return hi
|
|
||||||
}
|
|
||||||
func (hi *HandleImpl) Filter(pod *v1.Pod) bool {
|
|
||||||
return hi.EvictorFilterImpl.Filter(pod)
|
|
||||||
}
|
|
||||||
func (hi *HandleImpl) Evict(ctx context.Context, pod *v1.Pod, opts evictions.EvictOptions) bool {
|
|
||||||
return hi.PodEvictorImpl.EvictPod(ctx, pod, opts)
|
|
||||||
}
|
|
||||||
func (hi *HandleImpl) NodeLimitExceeded(node *v1.Node) bool {
|
|
||||||
return hi.PodEvictorImpl.NodeLimitExceeded(node)
|
|
||||||
}
|
|
||||||
@@ -1,193 +0,0 @@
|
|||||||
/*
|
|
||||||
Copyright 2022 The Kubernetes Authors.
|
|
||||||
Licensed under the Apache License, Version 2.0 (the "License");
|
|
||||||
you may not use this file except in compliance with the License.
|
|
||||||
You may obtain a copy of the License at
|
|
||||||
http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
Unless required by applicable law or agreed to in writing, software
|
|
||||||
distributed under the License is distributed on an "AS IS" BASIS,
|
|
||||||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
||||||
See the License for the specific language governing permissions and
|
|
||||||
limitations under the License.
|
|
||||||
*/
|
|
||||||
|
|
||||||
package defaultevictor
|
|
||||||
|
|
||||||
import (
|
|
||||||
// "context"
|
|
||||||
"context"
|
|
||||||
"fmt"
|
|
||||||
|
|
||||||
v1 "k8s.io/api/core/v1"
|
|
||||||
"k8s.io/apimachinery/pkg/labels"
|
|
||||||
"k8s.io/apimachinery/pkg/runtime"
|
|
||||||
"k8s.io/apimachinery/pkg/util/errors"
|
|
||||||
"k8s.io/klog/v2"
|
|
||||||
nodeutil "sigs.k8s.io/descheduler/pkg/descheduler/node"
|
|
||||||
podutil "sigs.k8s.io/descheduler/pkg/descheduler/pod"
|
|
||||||
"sigs.k8s.io/descheduler/pkg/framework"
|
|
||||||
"sigs.k8s.io/descheduler/pkg/utils"
|
|
||||||
)
|
|
||||||
|
|
||||||
const (
|
|
||||||
PluginName = "DefaultEvictor"
|
|
||||||
evictPodAnnotationKey = "descheduler.alpha.kubernetes.io/evict"
|
|
||||||
)
|
|
||||||
|
|
||||||
var _ framework.EvictorPlugin = &DefaultEvictor{}
|
|
||||||
|
|
||||||
type constraint func(pod *v1.Pod) error
|
|
||||||
|
|
||||||
// DefaultEvictor is the first EvictorPlugin, which defines the default extension points of the
|
|
||||||
// pre-baked evictor that is shipped.
|
|
||||||
// Even though we name this plugin DefaultEvictor, it does not actually evict anything,
|
|
||||||
// This plugin is only meant to customize other actions (extension points) of the evictor,
|
|
||||||
// like filtering, sorting, and other ones that might be relevant in the future
|
|
||||||
type DefaultEvictor struct {
|
|
||||||
constraints []constraint
|
|
||||||
}
|
|
||||||
|
|
||||||
// IsPodEvictableBasedOnPriority checks if the given pod is evictable based on priority resolved from pod Spec.
|
|
||||||
func IsPodEvictableBasedOnPriority(pod *v1.Pod, priority int32) bool {
|
|
||||||
return pod.Spec.Priority == nil || *pod.Spec.Priority < priority
|
|
||||||
}
|
|
||||||
|
|
||||||
// HaveEvictAnnotation checks if the pod have evict annotation
|
|
||||||
func HaveEvictAnnotation(pod *v1.Pod) bool {
|
|
||||||
_, found := pod.ObjectMeta.Annotations[evictPodAnnotationKey]
|
|
||||||
return found
|
|
||||||
}
|
|
||||||
|
|
||||||
// New builds plugin from its arguments while passing a handle
|
|
||||||
func New(args runtime.Object, handle framework.Handle) (framework.Plugin, error) {
|
|
||||||
defaultEvictorArgs, ok := args.(*DefaultEvictorArgs)
|
|
||||||
if !ok {
|
|
||||||
return nil, fmt.Errorf("want args to be of type defaultEvictorFilterArgs, got %T", args)
|
|
||||||
}
|
|
||||||
|
|
||||||
ev := &DefaultEvictor{}
|
|
||||||
|
|
||||||
if defaultEvictorArgs.EvictFailedBarePods {
|
|
||||||
klog.V(1).InfoS("Warning: EvictFailedBarePods is set to True. This could cause eviction of pods without ownerReferences.")
|
|
||||||
ev.constraints = append(ev.constraints, func(pod *v1.Pod) error {
|
|
||||||
ownerRefList := podutil.OwnerRef(pod)
|
|
||||||
// Enable evictFailedBarePods to evict bare pods in failed phase
|
|
||||||
if len(ownerRefList) == 0 && pod.Status.Phase != v1.PodFailed {
|
|
||||||
return fmt.Errorf("pod does not have any ownerRefs and is not in failed phase")
|
|
||||||
}
|
|
||||||
return nil
|
|
||||||
})
|
|
||||||
} else {
|
|
||||||
ev.constraints = append(ev.constraints, func(pod *v1.Pod) error {
|
|
||||||
ownerRefList := podutil.OwnerRef(pod)
|
|
||||||
if len(ownerRefList) == 0 {
|
|
||||||
return fmt.Errorf("pod does not have any ownerRefs")
|
|
||||||
}
|
|
||||||
return nil
|
|
||||||
})
|
|
||||||
}
|
|
||||||
if !defaultEvictorArgs.EvictSystemCriticalPods {
|
|
||||||
ev.constraints = append(ev.constraints, func(pod *v1.Pod) error {
|
|
||||||
if utils.IsCriticalPriorityPod(pod) {
|
|
||||||
return fmt.Errorf("pod has system critical priority")
|
|
||||||
}
|
|
||||||
return nil
|
|
||||||
})
|
|
||||||
|
|
||||||
if defaultEvictorArgs.PriorityThreshold != nil && (defaultEvictorArgs.PriorityThreshold.Value != nil || len(defaultEvictorArgs.PriorityThreshold.Name) > 0) {
|
|
||||||
thresholdPriority, err := utils.GetPriorityValueFromPriorityThreshold(context.TODO(), handle.ClientSet(), defaultEvictorArgs.PriorityThreshold)
|
|
||||||
if err != nil {
|
|
||||||
return nil, fmt.Errorf("failed to get priority threshold: %v", err)
|
|
||||||
}
|
|
||||||
ev.constraints = append(ev.constraints, func(pod *v1.Pod) error {
|
|
||||||
if IsPodEvictableBasedOnPriority(pod, thresholdPriority) {
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
return fmt.Errorf("pod has higher priority than specified priority class threshold")
|
|
||||||
})
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
klog.V(1).InfoS("Warning: EvictSystemCriticalPods is set to True. This could cause eviction of Kubernetes system pods.")
|
|
||||||
}
|
|
||||||
if !defaultEvictorArgs.EvictLocalStoragePods {
|
|
||||||
ev.constraints = append(ev.constraints, func(pod *v1.Pod) error {
|
|
||||||
if utils.IsPodWithLocalStorage(pod) {
|
|
||||||
return fmt.Errorf("pod has local storage and descheduler is not configured with evictLocalStoragePods")
|
|
||||||
}
|
|
||||||
return nil
|
|
||||||
})
|
|
||||||
}
|
|
||||||
if defaultEvictorArgs.IgnorePvcPods {
|
|
||||||
ev.constraints = append(ev.constraints, func(pod *v1.Pod) error {
|
|
||||||
if utils.IsPodWithPVC(pod) {
|
|
||||||
return fmt.Errorf("pod has a PVC and descheduler is configured to ignore PVC pods")
|
|
||||||
}
|
|
||||||
return nil
|
|
||||||
})
|
|
||||||
}
|
|
||||||
if defaultEvictorArgs.NodeFit {
|
|
||||||
ev.constraints = append(ev.constraints, func(pod *v1.Pod) error {
|
|
||||||
nodes, err := nodeutil.ReadyNodes(context.TODO(), handle.ClientSet(), handle.SharedInformerFactory().Core().V1().Nodes(), defaultEvictorArgs.NodeSelector)
|
|
||||||
if err != nil {
|
|
||||||
return fmt.Errorf("could not list nodes when processing NodeFit")
|
|
||||||
}
|
|
||||||
if !nodeutil.PodFitsAnyOtherNode(handle.GetPodsAssignedToNodeFunc(), pod, nodes) {
|
|
||||||
return fmt.Errorf("pod does not fit on any other node because of nodeSelector(s), Taint(s), or nodes marked as unschedulable")
|
|
||||||
}
|
|
||||||
return nil
|
|
||||||
})
|
|
||||||
}
|
|
||||||
if defaultEvictorArgs.LabelSelector != nil && !defaultEvictorArgs.LabelSelector.Empty() {
|
|
||||||
ev.constraints = append(ev.constraints, func(pod *v1.Pod) error {
|
|
||||||
if !defaultEvictorArgs.LabelSelector.Matches(labels.Set(pod.Labels)) {
|
|
||||||
return fmt.Errorf("pod labels do not match the labelSelector filter in the policy parameter")
|
|
||||||
}
|
|
||||||
return nil
|
|
||||||
})
|
|
||||||
}
|
|
||||||
|
|
||||||
return ev, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// Name retrieves the plugin name
|
|
||||||
func (d *DefaultEvictor) Name() string {
|
|
||||||
return PluginName
|
|
||||||
}
|
|
||||||
|
|
||||||
func (d *DefaultEvictor) Filter(pod *v1.Pod) bool {
|
|
||||||
checkErrs := []error{}
|
|
||||||
|
|
||||||
if HaveEvictAnnotation(pod) {
|
|
||||||
return true
|
|
||||||
}
|
|
||||||
|
|
||||||
ownerRefList := podutil.OwnerRef(pod)
|
|
||||||
if utils.IsDaemonsetPod(ownerRefList) {
|
|
||||||
checkErrs = append(checkErrs, fmt.Errorf("pod is a DaemonSet pod"))
|
|
||||||
}
|
|
||||||
|
|
||||||
if utils.IsMirrorPod(pod) {
|
|
||||||
checkErrs = append(checkErrs, fmt.Errorf("pod is a mirror pod"))
|
|
||||||
}
|
|
||||||
|
|
||||||
if utils.IsStaticPod(pod) {
|
|
||||||
checkErrs = append(checkErrs, fmt.Errorf("pod is a static pod"))
|
|
||||||
}
|
|
||||||
|
|
||||||
if utils.IsPodTerminating(pod) {
|
|
||||||
checkErrs = append(checkErrs, fmt.Errorf("pod is terminating"))
|
|
||||||
}
|
|
||||||
|
|
||||||
for _, c := range d.constraints {
|
|
||||||
if err := c(pod); err != nil {
|
|
||||||
checkErrs = append(checkErrs, err)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
if len(checkErrs) > 0 {
|
|
||||||
klog.V(4).InfoS("Pod fails the following checks", "pod", klog.KObj(pod), "checks", errors.NewAggregate(checkErrs).Error())
|
|
||||||
return false
|
|
||||||
}
|
|
||||||
|
|
||||||
return true
|
|
||||||
}
|
|
||||||
@@ -1,636 +0,0 @@
|
|||||||
/*
|
|
||||||
Copyright 2022 The Kubernetes Authors.
|
|
||||||
Licensed under the Apache License, Version 2.0 (the "License");
|
|
||||||
you may not use this file except in compliance with the License.
|
|
||||||
You may obtain a copy of the License at
|
|
||||||
http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
Unless required by applicable law or agreed to in writing, software
|
|
||||||
distributed under the License is distributed on an "AS IS" BASIS,
|
|
||||||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
||||||
See the License for the specific language governing permissions and
|
|
||||||
limitations under the License.
|
|
||||||
*/
|
|
||||||
|
|
||||||
package defaultevictor
|
|
||||||
|
|
||||||
import (
|
|
||||||
"context"
|
|
||||||
"testing"
|
|
||||||
|
|
||||||
v1 "k8s.io/api/core/v1"
|
|
||||||
"k8s.io/apimachinery/pkg/api/resource"
|
|
||||||
"k8s.io/apimachinery/pkg/runtime"
|
|
||||||
"k8s.io/client-go/informers"
|
|
||||||
"k8s.io/client-go/kubernetes/fake"
|
|
||||||
"sigs.k8s.io/descheduler/pkg/api"
|
|
||||||
podutil "sigs.k8s.io/descheduler/pkg/descheduler/pod"
|
|
||||||
"sigs.k8s.io/descheduler/pkg/framework"
|
|
||||||
frameworkfake "sigs.k8s.io/descheduler/pkg/framework/fake"
|
|
||||||
"sigs.k8s.io/descheduler/pkg/utils"
|
|
||||||
"sigs.k8s.io/descheduler/test"
|
|
||||||
)
|
|
||||||
|
|
||||||
func TestDefaultEvictorFilter(t *testing.T) {
|
|
||||||
n1 := test.BuildTestNode("node1", 1000, 2000, 13, nil)
|
|
||||||
lowPriority := int32(800)
|
|
||||||
highPriority := int32(900)
|
|
||||||
|
|
||||||
nodeTaintKey := "hardware"
|
|
||||||
nodeTaintValue := "gpu"
|
|
||||||
|
|
||||||
nodeLabelKey := "datacenter"
|
|
||||||
nodeLabelValue := "east"
|
|
||||||
type testCase struct {
|
|
||||||
description string
|
|
||||||
pods []*v1.Pod
|
|
||||||
nodes []*v1.Node
|
|
||||||
evictFailedBarePods bool
|
|
||||||
evictLocalStoragePods bool
|
|
||||||
evictSystemCriticalPods bool
|
|
||||||
priorityThreshold *int32
|
|
||||||
nodeFit bool
|
|
||||||
result bool
|
|
||||||
}
|
|
||||||
|
|
||||||
testCases := []testCase{
|
|
||||||
{
|
|
||||||
description: "Failed pod eviction with no ownerRefs",
|
|
||||||
pods: []*v1.Pod{
|
|
||||||
test.BuildTestPod("bare_pod_failed", 400, 0, n1.Name, func(pod *v1.Pod) {
|
|
||||||
pod.Status.Phase = v1.PodFailed
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
evictFailedBarePods: false,
|
|
||||||
result: false,
|
|
||||||
}, {
|
|
||||||
description: "Normal pod eviction with no ownerRefs and evictFailedBarePods enabled",
|
|
||||||
pods: []*v1.Pod{test.BuildTestPod("bare_pod", 400, 0, n1.Name, nil)},
|
|
||||||
evictFailedBarePods: true,
|
|
||||||
result: false,
|
|
||||||
}, {
|
|
||||||
description: "Failed pod eviction with no ownerRefs",
|
|
||||||
pods: []*v1.Pod{
|
|
||||||
test.BuildTestPod("bare_pod_failed_but_can_be_evicted", 400, 0, n1.Name, func(pod *v1.Pod) {
|
|
||||||
pod.Status.Phase = v1.PodFailed
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
evictFailedBarePods: true,
|
|
||||||
result: true,
|
|
||||||
}, {
|
|
||||||
description: "Normal pod eviction with normal ownerRefs",
|
|
||||||
pods: []*v1.Pod{
|
|
||||||
test.BuildTestPod("p1", 400, 0, n1.Name, func(pod *v1.Pod) {
|
|
||||||
pod.ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
evictLocalStoragePods: false,
|
|
||||||
evictSystemCriticalPods: false,
|
|
||||||
result: true,
|
|
||||||
}, {
|
|
||||||
description: "Normal pod eviction with normal ownerRefs and descheduler.alpha.kubernetes.io/evict annotation",
|
|
||||||
pods: []*v1.Pod{
|
|
||||||
test.BuildTestPod("p2", 400, 0, n1.Name, func(pod *v1.Pod) {
|
|
||||||
pod.Annotations = map[string]string{"descheduler.alpha.kubernetes.io/evict": "true"}
|
|
||||||
pod.ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
evictLocalStoragePods: false,
|
|
||||||
evictSystemCriticalPods: false,
|
|
||||||
result: true,
|
|
||||||
}, {
|
|
||||||
description: "Normal pod eviction with replicaSet ownerRefs",
|
|
||||||
pods: []*v1.Pod{
|
|
||||||
test.BuildTestPod("p3", 400, 0, n1.Name, func(pod *v1.Pod) {
|
|
||||||
pod.ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
evictLocalStoragePods: false,
|
|
||||||
evictSystemCriticalPods: false,
|
|
||||||
result: true,
|
|
||||||
}, {
|
|
||||||
description: "Normal pod eviction with replicaSet ownerRefs and descheduler.alpha.kubernetes.io/evict annotation",
|
|
||||||
pods: []*v1.Pod{
|
|
||||||
test.BuildTestPod("p4", 400, 0, n1.Name, func(pod *v1.Pod) {
|
|
||||||
pod.Annotations = map[string]string{"descheduler.alpha.kubernetes.io/evict": "true"}
|
|
||||||
pod.ObjectMeta.OwnerReferences = test.GetReplicaSetOwnerRefList()
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
evictLocalStoragePods: false,
|
|
||||||
evictSystemCriticalPods: false,
|
|
||||||
result: true,
|
|
||||||
}, {
|
|
||||||
description: "Normal pod eviction with statefulSet ownerRefs",
|
|
||||||
pods: []*v1.Pod{
|
|
||||||
test.BuildTestPod("p18", 400, 0, n1.Name, func(pod *v1.Pod) {
|
|
||||||
pod.ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
evictLocalStoragePods: false,
|
|
||||||
evictSystemCriticalPods: false,
|
|
||||||
result: true,
|
|
||||||
}, {
|
|
||||||
description: "Normal pod eviction with statefulSet ownerRefs and descheduler.alpha.kubernetes.io/evict annotation",
|
|
||||||
pods: []*v1.Pod{
|
|
||||||
test.BuildTestPod("p19", 400, 0, n1.Name, func(pod *v1.Pod) {
|
|
||||||
pod.Annotations = map[string]string{"descheduler.alpha.kubernetes.io/evict": "true"}
|
|
||||||
pod.ObjectMeta.OwnerReferences = test.GetStatefulSetOwnerRefList()
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
evictLocalStoragePods: false,
|
|
||||||
evictSystemCriticalPods: false,
|
|
||||||
result: true,
|
|
||||||
}, {
|
|
||||||
description: "Pod not evicted because it is bound to a PV and evictLocalStoragePods = false",
|
|
||||||
pods: []*v1.Pod{
|
|
||||||
test.BuildTestPod("p5", 400, 0, n1.Name, func(pod *v1.Pod) {
|
|
||||||
pod.ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
|
||||||
pod.Spec.Volumes = []v1.Volume{
|
|
||||||
{
|
|
||||||
Name: "sample",
|
|
||||||
VolumeSource: v1.VolumeSource{
|
|
||||||
HostPath: &v1.HostPathVolumeSource{Path: "somePath"},
|
|
||||||
EmptyDir: &v1.EmptyDirVolumeSource{
|
|
||||||
SizeLimit: resource.NewQuantity(int64(10), resource.BinarySI)},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
}
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
evictLocalStoragePods: false,
|
|
||||||
evictSystemCriticalPods: false,
|
|
||||||
result: false,
|
|
||||||
}, {
|
|
||||||
description: "Pod is evicted because it is bound to a PV and evictLocalStoragePods = true",
|
|
||||||
pods: []*v1.Pod{
|
|
||||||
test.BuildTestPod("p6", 400, 0, n1.Name, func(pod *v1.Pod) {
|
|
||||||
pod.ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
|
||||||
pod.Spec.Volumes = []v1.Volume{
|
|
||||||
{
|
|
||||||
Name: "sample",
|
|
||||||
VolumeSource: v1.VolumeSource{
|
|
||||||
HostPath: &v1.HostPathVolumeSource{Path: "somePath"},
|
|
||||||
EmptyDir: &v1.EmptyDirVolumeSource{
|
|
||||||
SizeLimit: resource.NewQuantity(int64(10), resource.BinarySI)},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
}
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
evictLocalStoragePods: true,
|
|
||||||
evictSystemCriticalPods: false,
|
|
||||||
result: true,
|
|
||||||
}, {
|
|
||||||
description: "Pod is evicted because it is bound to a PV and evictLocalStoragePods = false, but it has scheduler.alpha.kubernetes.io/evict annotation",
|
|
||||||
pods: []*v1.Pod{
|
|
||||||
test.BuildTestPod("p7", 400, 0, n1.Name, func(pod *v1.Pod) {
|
|
||||||
pod.Annotations = map[string]string{"descheduler.alpha.kubernetes.io/evict": "true"}
|
|
||||||
pod.ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
|
||||||
pod.Spec.Volumes = []v1.Volume{
|
|
||||||
{
|
|
||||||
Name: "sample",
|
|
||||||
VolumeSource: v1.VolumeSource{
|
|
||||||
HostPath: &v1.HostPathVolumeSource{Path: "somePath"},
|
|
||||||
EmptyDir: &v1.EmptyDirVolumeSource{
|
|
||||||
SizeLimit: resource.NewQuantity(int64(10), resource.BinarySI)},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
}
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
evictLocalStoragePods: false,
|
|
||||||
evictSystemCriticalPods: false,
|
|
||||||
result: true,
|
|
||||||
}, {
|
|
||||||
description: "Pod not evicted becasuse it is part of a daemonSet",
|
|
||||||
pods: []*v1.Pod{
|
|
||||||
test.BuildTestPod("p8", 400, 0, n1.Name, func(pod *v1.Pod) {
|
|
||||||
pod.ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
|
||||||
pod.ObjectMeta.OwnerReferences = test.GetDaemonSetOwnerRefList()
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
evictLocalStoragePods: false,
|
|
||||||
evictSystemCriticalPods: false,
|
|
||||||
result: false,
|
|
||||||
}, {
|
|
||||||
description: "Pod is evicted becasuse it is part of a daemonSet, but it has scheduler.alpha.kubernetes.io/evict annotation",
|
|
||||||
pods: []*v1.Pod{
|
|
||||||
test.BuildTestPod("p9", 400, 0, n1.Name, func(pod *v1.Pod) {
|
|
||||||
pod.Annotations = map[string]string{"descheduler.alpha.kubernetes.io/evict": "true"}
|
|
||||||
pod.ObjectMeta.OwnerReferences = test.GetDaemonSetOwnerRefList()
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
evictLocalStoragePods: false,
|
|
||||||
evictSystemCriticalPods: false,
|
|
||||||
result: true,
|
|
||||||
}, {
|
|
||||||
description: "Pod not evicted becasuse it is a mirror poddsa",
|
|
||||||
pods: []*v1.Pod{
|
|
||||||
test.BuildTestPod("p10", 400, 0, n1.Name, func(pod *v1.Pod) {
|
|
||||||
pod.ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
|
||||||
pod.Annotations = test.GetMirrorPodAnnotation()
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
evictLocalStoragePods: false,
|
|
||||||
evictSystemCriticalPods: false,
|
|
||||||
result: false,
|
|
||||||
}, {
|
|
||||||
description: "Pod is evicted becasuse it is a mirror pod, but it has scheduler.alpha.kubernetes.io/evict annotation",
|
|
||||||
pods: []*v1.Pod{
|
|
||||||
test.BuildTestPod("p11", 400, 0, n1.Name, func(pod *v1.Pod) {
|
|
||||||
pod.ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
|
||||||
pod.Annotations = test.GetMirrorPodAnnotation()
|
|
||||||
pod.Annotations["descheduler.alpha.kubernetes.io/evict"] = "true"
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
evictLocalStoragePods: false,
|
|
||||||
evictSystemCriticalPods: false,
|
|
||||||
result: true,
|
|
||||||
}, {
|
|
||||||
description: "Pod not evicted becasuse it has system critical priority",
|
|
||||||
pods: []*v1.Pod{
|
|
||||||
test.BuildTestPod("p12", 400, 0, n1.Name, func(pod *v1.Pod) {
|
|
||||||
pod.ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
|
||||||
priority := utils.SystemCriticalPriority
|
|
||||||
pod.Spec.Priority = &priority
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
evictLocalStoragePods: false,
|
|
||||||
evictSystemCriticalPods: false,
|
|
||||||
result: false,
|
|
||||||
}, {
|
|
||||||
description: "Pod is evicted becasuse it has system critical priority, but it has scheduler.alpha.kubernetes.io/evict annotation",
|
|
||||||
pods: []*v1.Pod{
|
|
||||||
test.BuildTestPod("p13", 400, 0, n1.Name, func(pod *v1.Pod) {
|
|
||||||
pod.ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
|
||||||
priority := utils.SystemCriticalPriority
|
|
||||||
pod.Spec.Priority = &priority
|
|
||||||
pod.Annotations = map[string]string{
|
|
||||||
"descheduler.alpha.kubernetes.io/evict": "true",
|
|
||||||
}
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
evictLocalStoragePods: false,
|
|
||||||
evictSystemCriticalPods: false,
|
|
||||||
result: true,
|
|
||||||
}, {
|
|
||||||
description: "Pod not evicted becasuse it has a priority higher than the configured priority threshold",
|
|
||||||
pods: []*v1.Pod{
|
|
||||||
test.BuildTestPod("p14", 400, 0, n1.Name, func(pod *v1.Pod) {
|
|
||||||
pod.ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
|
||||||
pod.Spec.Priority = &highPriority
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
evictLocalStoragePods: false,
|
|
||||||
evictSystemCriticalPods: false,
|
|
||||||
priorityThreshold: &lowPriority,
|
|
||||||
result: false,
|
|
||||||
}, {
|
|
||||||
description: "Pod is evicted becasuse it has a priority higher than the configured priority threshold, but it has scheduler.alpha.kubernetes.io/evict annotation",
|
|
||||||
pods: []*v1.Pod{
|
|
||||||
test.BuildTestPod("p15", 400, 0, n1.Name, func(pod *v1.Pod) {
|
|
||||||
pod.ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
|
||||||
pod.Annotations = map[string]string{"descheduler.alpha.kubernetes.io/evict": "true"}
|
|
||||||
pod.Spec.Priority = &highPriority
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
evictLocalStoragePods: false,
|
|
||||||
evictSystemCriticalPods: false,
|
|
||||||
priorityThreshold: &lowPriority,
|
|
||||||
result: true,
|
|
||||||
}, {
|
|
||||||
description: "Pod is evicted becasuse it has system critical priority, but evictSystemCriticalPods = true",
|
|
||||||
pods: []*v1.Pod{
|
|
||||||
test.BuildTestPod("p16", 400, 0, n1.Name, func(pod *v1.Pod) {
|
|
||||||
pod.ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
|
||||||
priority := utils.SystemCriticalPriority
|
|
||||||
pod.Spec.Priority = &priority
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
evictLocalStoragePods: false,
|
|
||||||
evictSystemCriticalPods: true,
|
|
||||||
result: true,
|
|
||||||
}, {
|
|
||||||
description: "Pod is evicted becasuse it has system critical priority, but evictSystemCriticalPods = true and it has scheduler.alpha.kubernetes.io/evict annotation",
|
|
||||||
pods: []*v1.Pod{
|
|
||||||
test.BuildTestPod("p16", 400, 0, n1.Name, func(pod *v1.Pod) {
|
|
||||||
pod.ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
|
||||||
pod.Annotations = map[string]string{"descheduler.alpha.kubernetes.io/evict": "true"}
|
|
||||||
priority := utils.SystemCriticalPriority
|
|
||||||
pod.Spec.Priority = &priority
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
evictLocalStoragePods: false,
|
|
||||||
evictSystemCriticalPods: true,
|
|
||||||
result: true,
|
|
||||||
}, {
|
|
||||||
description: "Pod is evicted becasuse it has a priority higher than the configured priority threshold, but evictSystemCriticalPods = true",
|
|
||||||
pods: []*v1.Pod{
|
|
||||||
test.BuildTestPod("p17", 400, 0, n1.Name, func(pod *v1.Pod) {
|
|
||||||
pod.ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
|
||||||
pod.Spec.Priority = &highPriority
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
evictLocalStoragePods: false,
|
|
||||||
evictSystemCriticalPods: true,
|
|
||||||
priorityThreshold: &lowPriority,
|
|
||||||
result: true,
|
|
||||||
}, {
|
|
||||||
description: "Pod is evicted becasuse it has a priority higher than the configured priority threshold, but evictSystemCriticalPods = true and it has scheduler.alpha.kubernetes.io/evict annotation",
|
|
||||||
pods: []*v1.Pod{
|
|
||||||
test.BuildTestPod("p17", 400, 0, n1.Name, func(pod *v1.Pod) {
|
|
||||||
pod.ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
|
||||||
pod.Annotations = map[string]string{"descheduler.alpha.kubernetes.io/evict": "true"}
|
|
||||||
pod.Spec.Priority = &highPriority
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
evictLocalStoragePods: false,
|
|
||||||
evictSystemCriticalPods: true,
|
|
||||||
priorityThreshold: &lowPriority,
|
|
||||||
result: true,
|
|
||||||
}, {
|
|
||||||
description: "Pod with no tolerations running on normal node, all other nodes tainted",
|
|
||||||
pods: []*v1.Pod{
|
|
||||||
test.BuildTestPod("p1", 400, 0, n1.Name, func(pod *v1.Pod) {
|
|
||||||
pod.ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
nodes: []*v1.Node{
|
|
||||||
test.BuildTestNode("node2", 1000, 2000, 13, func(node *v1.Node) {
|
|
||||||
node.Spec.Taints = []v1.Taint{
|
|
||||||
{
|
|
||||||
Key: nodeTaintKey,
|
|
||||||
Value: nodeTaintValue,
|
|
||||||
Effect: v1.TaintEffectNoSchedule,
|
|
||||||
},
|
|
||||||
}
|
|
||||||
}),
|
|
||||||
test.BuildTestNode("node3", 1000, 2000, 13, func(node *v1.Node) {
|
|
||||||
node.Spec.Taints = []v1.Taint{
|
|
||||||
{
|
|
||||||
Key: nodeTaintKey,
|
|
||||||
Value: nodeTaintValue,
|
|
||||||
Effect: v1.TaintEffectNoSchedule,
|
|
||||||
},
|
|
||||||
}
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
evictLocalStoragePods: false,
|
|
||||||
evictSystemCriticalPods: false,
|
|
||||||
nodeFit: true,
|
|
||||||
result: false,
|
|
||||||
}, {
|
|
||||||
description: "Pod with correct tolerations running on normal node, all other nodes tainted",
|
|
||||||
pods: []*v1.Pod{
|
|
||||||
test.BuildTestPod("p1", 400, 0, n1.Name, func(pod *v1.Pod) {
|
|
||||||
pod.ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
|
||||||
pod.Spec.Tolerations = []v1.Toleration{
|
|
||||||
{
|
|
||||||
Key: nodeTaintKey,
|
|
||||||
Value: nodeTaintValue,
|
|
||||||
Effect: v1.TaintEffectNoSchedule,
|
|
||||||
},
|
|
||||||
}
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
nodes: []*v1.Node{
|
|
||||||
test.BuildTestNode("node2", 1000, 2000, 13, func(node *v1.Node) {
|
|
||||||
node.Spec.Taints = []v1.Taint{
|
|
||||||
{
|
|
||||||
Key: nodeTaintKey,
|
|
||||||
Value: nodeTaintValue,
|
|
||||||
Effect: v1.TaintEffectNoSchedule,
|
|
||||||
},
|
|
||||||
}
|
|
||||||
}),
|
|
||||||
test.BuildTestNode("node3", 1000, 2000, 13, func(node *v1.Node) {
|
|
||||||
node.Spec.Taints = []v1.Taint{
|
|
||||||
{
|
|
||||||
Key: nodeTaintKey,
|
|
||||||
Value: nodeTaintValue,
|
|
||||||
Effect: v1.TaintEffectNoSchedule,
|
|
||||||
},
|
|
||||||
}
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
evictLocalStoragePods: false,
|
|
||||||
evictSystemCriticalPods: false,
|
|
||||||
nodeFit: true,
|
|
||||||
result: true,
|
|
||||||
}, {
|
|
||||||
description: "Pod with incorrect node selector",
|
|
||||||
pods: []*v1.Pod{
|
|
||||||
test.BuildTestPod("p1", 400, 0, n1.Name, func(pod *v1.Pod) {
|
|
||||||
pod.ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
|
||||||
pod.Spec.NodeSelector = map[string]string{
|
|
||||||
nodeLabelKey: "fail",
|
|
||||||
}
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
nodes: []*v1.Node{
|
|
||||||
test.BuildTestNode("node2", 1000, 2000, 13, func(node *v1.Node) {
|
|
||||||
node.ObjectMeta.Labels = map[string]string{
|
|
||||||
nodeLabelKey: nodeLabelValue,
|
|
||||||
}
|
|
||||||
}),
|
|
||||||
test.BuildTestNode("node3", 1000, 2000, 13, func(node *v1.Node) {
|
|
||||||
node.ObjectMeta.Labels = map[string]string{
|
|
||||||
nodeLabelKey: nodeLabelValue,
|
|
||||||
}
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
evictLocalStoragePods: false,
|
|
||||||
evictSystemCriticalPods: false,
|
|
||||||
nodeFit: true,
|
|
||||||
result: false,
|
|
||||||
}, {
|
|
||||||
description: "Pod with correct node selector",
|
|
||||||
pods: []*v1.Pod{
|
|
||||||
test.BuildTestPod("p1", 400, 0, n1.Name, func(pod *v1.Pod) {
|
|
||||||
pod.ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
|
||||||
pod.Spec.NodeSelector = map[string]string{
|
|
||||||
nodeLabelKey: nodeLabelValue,
|
|
||||||
}
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
nodes: []*v1.Node{
|
|
||||||
test.BuildTestNode("node2", 1000, 2000, 13, func(node *v1.Node) {
|
|
||||||
node.ObjectMeta.Labels = map[string]string{
|
|
||||||
nodeLabelKey: nodeLabelValue,
|
|
||||||
}
|
|
||||||
}),
|
|
||||||
test.BuildTestNode("node3", 1000, 2000, 13, func(node *v1.Node) {
|
|
||||||
node.ObjectMeta.Labels = map[string]string{
|
|
||||||
nodeLabelKey: nodeLabelValue,
|
|
||||||
}
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
evictLocalStoragePods: false,
|
|
||||||
evictSystemCriticalPods: false,
|
|
||||||
nodeFit: true,
|
|
||||||
result: true,
|
|
||||||
}, {
|
|
||||||
description: "Pod with correct node selector, but only available node doesn't have enough CPU",
|
|
||||||
pods: []*v1.Pod{
|
|
||||||
test.BuildTestPod("p1", 12, 8, n1.Name, func(pod *v1.Pod) {
|
|
||||||
pod.ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
|
||||||
pod.Spec.NodeSelector = map[string]string{
|
|
||||||
nodeLabelKey: nodeLabelValue,
|
|
||||||
}
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
nodes: []*v1.Node{
|
|
||||||
test.BuildTestNode("node2-TEST", 10, 16, 10, func(node *v1.Node) {
|
|
||||||
node.ObjectMeta.Labels = map[string]string{
|
|
||||||
nodeLabelKey: nodeLabelValue,
|
|
||||||
}
|
|
||||||
}),
|
|
||||||
test.BuildTestNode("node3-TEST", 10, 16, 10, func(node *v1.Node) {
|
|
||||||
node.ObjectMeta.Labels = map[string]string{
|
|
||||||
nodeLabelKey: nodeLabelValue,
|
|
||||||
}
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
evictLocalStoragePods: false,
|
|
||||||
evictSystemCriticalPods: false,
|
|
||||||
nodeFit: true,
|
|
||||||
result: false,
|
|
||||||
}, {
|
|
||||||
description: "Pod with correct node selector, and one node has enough memory",
|
|
||||||
pods: []*v1.Pod{
|
|
||||||
test.BuildTestPod("p1", 12, 8, n1.Name, func(pod *v1.Pod) {
|
|
||||||
pod.ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
|
||||||
pod.Spec.NodeSelector = map[string]string{
|
|
||||||
nodeLabelKey: nodeLabelValue,
|
|
||||||
}
|
|
||||||
}),
|
|
||||||
test.BuildTestPod("node2-pod-10GB-mem", 20, 10, "node2", func(pod *v1.Pod) {
|
|
||||||
pod.ObjectMeta.Labels = map[string]string{
|
|
||||||
"test": "true",
|
|
||||||
}
|
|
||||||
}),
|
|
||||||
test.BuildTestPod("node3-pod-10GB-mem", 20, 10, "node3", func(pod *v1.Pod) {
|
|
||||||
pod.ObjectMeta.Labels = map[string]string{
|
|
||||||
"test": "true",
|
|
||||||
}
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
nodes: []*v1.Node{
|
|
||||||
test.BuildTestNode("node2", 100, 16, 10, func(node *v1.Node) {
|
|
||||||
node.ObjectMeta.Labels = map[string]string{
|
|
||||||
nodeLabelKey: nodeLabelValue,
|
|
||||||
}
|
|
||||||
}),
|
|
||||||
test.BuildTestNode("node3", 100, 20, 10, func(node *v1.Node) {
|
|
||||||
node.ObjectMeta.Labels = map[string]string{
|
|
||||||
nodeLabelKey: nodeLabelValue,
|
|
||||||
}
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
evictLocalStoragePods: false,
|
|
||||||
evictSystemCriticalPods: false,
|
|
||||||
nodeFit: true,
|
|
||||||
result: true,
|
|
||||||
}, {
|
|
||||||
description: "Pod with correct node selector, but both nodes don't have enough memory",
|
|
||||||
pods: []*v1.Pod{
|
|
||||||
test.BuildTestPod("p1", 12, 8, n1.Name, func(pod *v1.Pod) {
|
|
||||||
pod.ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
|
||||||
pod.Spec.NodeSelector = map[string]string{
|
|
||||||
nodeLabelKey: nodeLabelValue,
|
|
||||||
}
|
|
||||||
}),
|
|
||||||
test.BuildTestPod("node2-pod-10GB-mem", 10, 10, "node2", func(pod *v1.Pod) {
|
|
||||||
pod.ObjectMeta.Labels = map[string]string{
|
|
||||||
"test": "true",
|
|
||||||
}
|
|
||||||
}),
|
|
||||||
test.BuildTestPod("node3-pod-10GB-mem", 10, 10, "node3", func(pod *v1.Pod) {
|
|
||||||
pod.ObjectMeta.Labels = map[string]string{
|
|
||||||
"test": "true",
|
|
||||||
}
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
nodes: []*v1.Node{
|
|
||||||
test.BuildTestNode("node2", 100, 16, 10, func(node *v1.Node) {
|
|
||||||
node.ObjectMeta.Labels = map[string]string{
|
|
||||||
nodeLabelKey: nodeLabelValue,
|
|
||||||
}
|
|
||||||
}),
|
|
||||||
test.BuildTestNode("node3", 100, 16, 10, func(node *v1.Node) {
|
|
||||||
node.ObjectMeta.Labels = map[string]string{
|
|
||||||
nodeLabelKey: nodeLabelValue,
|
|
||||||
}
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
evictLocalStoragePods: false,
|
|
||||||
evictSystemCriticalPods: false,
|
|
||||||
nodeFit: true,
|
|
||||||
result: false,
|
|
||||||
},
|
|
||||||
}
|
|
||||||
|
|
||||||
for _, test := range testCases {
|
|
||||||
|
|
||||||
t.Run(test.description, func(t *testing.T) {
|
|
||||||
ctx, cancel := context.WithCancel(context.Background())
|
|
||||||
defer cancel()
|
|
||||||
|
|
||||||
var objs []runtime.Object
|
|
||||||
for _, node := range test.nodes {
|
|
||||||
objs = append(objs, node)
|
|
||||||
}
|
|
||||||
for _, pod := range test.pods {
|
|
||||||
objs = append(objs, pod)
|
|
||||||
}
|
|
||||||
|
|
||||||
fakeClient := fake.NewSimpleClientset(objs...)
|
|
||||||
|
|
||||||
sharedInformerFactory := informers.NewSharedInformerFactory(fakeClient, 0)
|
|
||||||
podInformer := sharedInformerFactory.Core().V1().Pods()
|
|
||||||
|
|
||||||
getPodsAssignedToNode, err := podutil.BuildGetPodsAssignedToNodeFunc(podInformer)
|
|
||||||
if err != nil {
|
|
||||||
t.Errorf("Build get pods assigned to node function error: %v", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
sharedInformerFactory.Start(ctx.Done())
|
|
||||||
sharedInformerFactory.WaitForCacheSync(ctx.Done())
|
|
||||||
|
|
||||||
// var opts []func(opts *FilterOptions)
|
|
||||||
// if test.priorityThreshold != nil {
|
|
||||||
// opts = append(opts, WithPriorityThreshold(*test.priorityThreshold))
|
|
||||||
// }
|
|
||||||
// if test.nodeFit {
|
|
||||||
// opts = append(opts, WithNodeFit(true))
|
|
||||||
// }
|
|
||||||
|
|
||||||
defaultEvictorArgs := &DefaultEvictorArgs{
|
|
||||||
EvictLocalStoragePods: test.evictLocalStoragePods,
|
|
||||||
EvictSystemCriticalPods: test.evictSystemCriticalPods,
|
|
||||||
IgnorePvcPods: false,
|
|
||||||
EvictFailedBarePods: test.evictFailedBarePods,
|
|
||||||
PriorityThreshold: &api.PriorityThreshold{
|
|
||||||
Value: test.priorityThreshold,
|
|
||||||
},
|
|
||||||
NodeFit: test.nodeFit,
|
|
||||||
}
|
|
||||||
|
|
||||||
evictorPlugin, err := New(
|
|
||||||
defaultEvictorArgs,
|
|
||||||
&frameworkfake.HandleImpl{
|
|
||||||
ClientsetImpl: fakeClient,
|
|
||||||
GetPodsAssignedToNodeFuncImpl: getPodsAssignedToNode,
|
|
||||||
SharedInformerFactoryImpl: sharedInformerFactory,
|
|
||||||
})
|
|
||||||
if err != nil {
|
|
||||||
t.Fatalf("Unable to initialize the plugin: %v", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
result := evictorPlugin.(framework.EvictorPlugin).Filter(test.pods[0])
|
|
||||||
if result != test.result {
|
|
||||||
t.Errorf("Filter should return for pod %s %t, but it returns %t", test.pods[0].Name, test.result, result)
|
|
||||||
}
|
|
||||||
|
|
||||||
})
|
|
||||||
}
|
|
||||||
}
|
|
||||||
@@ -1,37 +0,0 @@
|
|||||||
/*
|
|
||||||
Copyright 2022 The Kubernetes Authors.
|
|
||||||
Licensed under the Apache License, Version 2.0 (the "License");
|
|
||||||
you may not use this file except in compliance with the License.
|
|
||||||
You may obtain a copy of the License at
|
|
||||||
http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
Unless required by applicable law or agreed to in writing, software
|
|
||||||
distributed under the License is distributed on an "AS IS" BASIS,
|
|
||||||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
||||||
See the License for the specific language governing permissions and
|
|
||||||
limitations under the License.
|
|
||||||
*/
|
|
||||||
|
|
||||||
package defaultevictor
|
|
||||||
|
|
||||||
import (
|
|
||||||
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
|
||||||
"k8s.io/apimachinery/pkg/labels"
|
|
||||||
"sigs.k8s.io/descheduler/pkg/api"
|
|
||||||
)
|
|
||||||
|
|
||||||
// +k8s:deepcopy-gen=true
|
|
||||||
// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
|
|
||||||
|
|
||||||
// DefaultEvictorArgs holds arguments used to configure DefaultEvictor plugin.
|
|
||||||
type DefaultEvictorArgs struct {
|
|
||||||
metav1.TypeMeta
|
|
||||||
|
|
||||||
NodeSelector string
|
|
||||||
EvictLocalStoragePods bool
|
|
||||||
EvictSystemCriticalPods bool
|
|
||||||
IgnorePvcPods bool
|
|
||||||
EvictFailedBarePods bool
|
|
||||||
LabelSelector labels.Selector
|
|
||||||
PriorityThreshold *api.PriorityThreshold
|
|
||||||
NodeFit bool
|
|
||||||
}
|
|
||||||
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user