preemptor missing when preempting in same job
lowang-bh opened this issue · 4 comments
This case should preempt two pods in same job with low priority
create(){
kubectl apply -f - <<EOF
apiVersion: batch/v1
kind: Job
metadata:
annotations:
useKubeBatch: "true"
name: preeptee
namespace: default
spec:
backoffLimit: 2
completions: 2
parallelism: 2
ttlSecondsAfterFinished: 600
template:
metadata:
annotations:
scheduling.k8s.io/group-name: group1
spec:
containers:
- image: busybox
imagePullPolicy: IfNotPresent
name: busybox
command: ['/bin/sh']
args: ['-c', 'sleep 160']
resources:
requests:
cpu: 1000m
#limits:
# nvidia.com/gpu: 3
restartPolicy: Never
terminationGracePeriodSeconds: 5
schedulerName: kube-batch
---
apiVersion: scheduling.incubator.k8s.io/v1alpha1
kind: PodGroup
metadata:
name: group1
namespace: default
spec:
minMember: 1
queue: default
EOF
sleep 10
kubectl apply -f - <<EOF
apiVersion: batch/v1
kind: Job
metadata:
name: preeptor
namespace: default
spec:
backoffLimit: 2
completions: 2
parallelism: 2
ttlSecondsAfterFinished: 600
template:
metadata:
annotations:
scheduling.k8s.io/group-name: group1
spec:
containers:
- image: busybox
imagePullPolicy: IfNotPresent
name: busybox
command: ['/bin/sh']
args: ['-c', 'sleep 120']
resources:
requests:
cpu: 1000m
restartPolicy: Never
terminationGracePeriodSeconds: 5
schedulerName: kube-batch
priorityClassName: high-priority
EOF
}
Is this a BUG REPORT or FEATURE REQUEST?:
Bug report
Uncomment only one, leave it on its own line:
kind bug
/kind feature
What happened:
What you expected to happen:
How to reproduce it (as minimally and precisely as possible):
Anything else we need to know?:
Environment:
- Kubernetes version (use
kubectl version
): - Cloud provider or hardware configuration:
- OS (e.g. from /etc/os-release):
- Kernel (e.g.
uname -a
): - Install tools:
- Others:
The Kubernetes project currently lacks enough contributors to adequately respond to all issues and PRs.
This bot triages issues and PRs according to the following rules:
- After 90d of inactivity,
lifecycle/stale
is applied - After 30d of inactivity since
lifecycle/stale
was applied,lifecycle/rotten
is applied - After 30d of inactivity since
lifecycle/rotten
was applied, the issue is closed
You can:
- Mark this issue or PR as fresh with
/remove-lifecycle stale
- Mark this issue or PR as rotten with
/lifecycle rotten
- Close this issue or PR with
/close
- Offer to help out with Issue Triage
Please send feedback to sig-contributor-experience at kubernetes/community.
/lifecycle stale
The Kubernetes project currently lacks enough active contributors to adequately respond to all issues and PRs.
This bot triages issues and PRs according to the following rules:
- After 90d of inactivity,
lifecycle/stale
is applied - After 30d of inactivity since
lifecycle/stale
was applied,lifecycle/rotten
is applied - After 30d of inactivity since
lifecycle/rotten
was applied, the issue is closed
You can:
- Mark this issue or PR as fresh with
/remove-lifecycle rotten
- Close this issue or PR with
/close
- Offer to help out with Issue Triage
Please send feedback to sig-contributor-experience at kubernetes/community.
/lifecycle rotten
The Kubernetes project currently lacks enough active contributors to adequately respond to all issues and PRs.
This bot triages issues and PRs according to the following rules:
- After 90d of inactivity,
lifecycle/stale
is applied - After 30d of inactivity since
lifecycle/stale
was applied,lifecycle/rotten
is applied - After 30d of inactivity since
lifecycle/rotten
was applied, the issue is closed
You can:
- Reopen this issue or PR with
/reopen
- Mark this issue or PR as fresh with
/remove-lifecycle rotten
- Offer to help out with Issue Triage
Please send feedback to sig-contributor-experience at kubernetes/community.
/close
@k8s-triage-robot: Closing this issue.
In response to this:
The Kubernetes project currently lacks enough active contributors to adequately respond to all issues and PRs.
This bot triages issues and PRs according to the following rules:
- After 90d of inactivity,
lifecycle/stale
is applied- After 30d of inactivity since
lifecycle/stale
was applied,lifecycle/rotten
is applied- After 30d of inactivity since
lifecycle/rotten
was applied, the issue is closedYou can:
- Reopen this issue or PR with
/reopen
- Mark this issue or PR as fresh with
/remove-lifecycle rotten
- Offer to help out with Issue Triage
Please send feedback to sig-contributor-experience at kubernetes/community.
/close
Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository.