kubernetes-retired / kube-batch

A batch scheduler of kubernetes for high performance workload, e.g. AI/ML, BigData, HPC

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

how to deploy kube-batch on arm64

magnate3 opened this issue · comments

Is this a BUG REPORT or FEATURE REQUEST?:

Uncomment only one, leave it on its own line:

/kind bug
/kind feature

What happened:
root@ubuntu:# helm install /kube-batch/deployment/kube-batch --namespace kube-system --generate-name
NAME: kube-batch-1625472656
LAST DEPLOYED: Mon Jul 5 16:10:56 2021
NAMESPACE: kube-system
STATUS: deployed
REVISION: 1
TEST SUITE: None
NOTES:
The batch scheduler of Kubernetes.
root@ubuntu:
# kubectl logs kube-batch-7bfddfb67c-9l8pw -n kube-system
standard_init_linux.go:228: exec user process caused: exec format error
root@ubuntu:
#
root@ubuntu:# kubectl get pods --all-namespaces -o wide | grep kube-batch
kube-system kube-batch-7bfddfb67c-9l8pw 0/1 Error 3 57s 10.244.29.10 bogon
root@ubuntu:
#

root@ubuntu:~# kubectl describe pods kube-batch-7bfddfb67c-9l8pw -n kube-system
Name: kube-batch-7bfddfb67c-9l8pw
Namespace: kube-system
Priority: 0
Node: bogon/10.10.16.81
Start Time: Mon, 05 Jul 2021 16:11:00 +0800
Labels: app=kube-batch
pod-template-hash=7bfddfb67c
Annotations: cni.projectcalico.org/podIP: 10.244.29.10/32
cni.projectcalico.org/podIPs: 10.244.29.10/32
Status: Running
IP: 10.244.29.10
IPs:
IP: 10.244.29.10
Controlled By: ReplicaSet/kube-batch-7bfddfb67c
Containers:
kube-batch:
Container ID: docker://25062fd9fa1f32fab79bb09653802239105dbb76f3d05381088ed8073fbec0af
Image: kubesigs/kube-batch:v0.5
Image ID: docker-pullable://kubesigs/kube-batch@sha256:4e4c2ebf19a341fd450efdd5d4cffd1addb5abbc7163fd9b08f7f7b1bc42e0ab
Port:
Host Port:
Args:
--logtostderr
--v
3
State: Terminated
Reason: Error
Exit Code: 1
Started: Mon, 05 Jul 2021 16:12:38 +0800
Finished: Mon, 05 Jul 2021 16:12:38 +0800
Last State: Terminated
Reason: Error
Exit Code: 1
Started: Mon, 05 Jul 2021 16:11:44 +0800
Finished: Mon, 05 Jul 2021 16:11:44 +0800
Ready: False
Restart Count: 4
Limits:
cpu: 2
memory: 2Gi
Requests:
cpu: 2
memory: 2Gi
Environment:
Mounts:
/var/run/secrets/kubernetes.io/serviceaccount from default-token-5w25p (ro)
Conditions:
Type Status
Initialized True
Ready False
ContainersReady False
PodScheduled True
Volumes:
default-token-5w25p:
Type: Secret (a volume populated by a Secret)
SecretName: default-token-5w25p
Optional: false
QoS Class: Guaranteed
Node-Selectors:
Tolerations: node.kubernetes.io/not-ready:NoExecute for 300s
node.kubernetes.io/unreachable:NoExecute for 300s
Events:
Type Reason Age From Message


Normal Scheduled default-scheduler Successfully assigned kube-system/kube-batch-7bfddfb67c-9l8pw to bogon
Normal Pulled 9s (x5 over 104s) kubelet, bogon Container image "kubesigs/kube-batch:v0.5" already present on machine
Normal Created 9s (x5 over 104s) kubelet, bogon Created container kube-batch
Normal Started 9s (x5 over 104s) kubelet, bogon Started container kube-batch
Warning BackOff 8s (x9 over 102s) kubelet, bogon Back-off restarting failed container
root@ubuntu:~#
What you expected to happen:

How to reproduce it (as minimally and precisely as possible):

Anything else we need to know?:

Environment:

root@ubuntu:# uname -a
Linux ubuntu 5.0.0-23-generic #24
18.04.1-Ubuntu SMP Mon Jul 29 16:10:24 UTC 2019 aarch64 aarch64 aarch64 GNU/Linux
root@ubuntu:# kubeadm version
kubeadm version: &version.Info{Major:"1", Minor:"18", GitVersion:"v1.18.1", GitCommit:"7879fc12a63337efff607952a323df90cdc7a335", GitTreeState:"clean", BuildDate:"2020-04-08T17:36:32Z", GoVersion:"go1.13.9", Compiler:"gc", Platform:"linux/arm64"}
root@ubuntu:
#

The Kubernetes project currently lacks enough contributors to adequately respond to all issues and PRs.

This bot triages issues and PRs according to the following rules:

  • After 90d of inactivity, lifecycle/stale is applied
  • After 30d of inactivity since lifecycle/stale was applied, lifecycle/rotten is applied
  • After 30d of inactivity since lifecycle/rotten was applied, the issue is closed

You can:

  • Mark this issue or PR as fresh with /remove-lifecycle stale
  • Mark this issue or PR as rotten with /lifecycle rotten
  • Close this issue or PR with /close
  • Offer to help out with Issue Triage

Please send feedback to sig-contributor-experience at kubernetes/community.

/lifecycle stale

The Kubernetes project currently lacks enough active contributors to adequately respond to all issues and PRs.

This bot triages issues and PRs according to the following rules:

  • After 90d of inactivity, lifecycle/stale is applied
  • After 30d of inactivity since lifecycle/stale was applied, lifecycle/rotten is applied
  • After 30d of inactivity since lifecycle/rotten was applied, the issue is closed

You can:

  • Mark this issue or PR as fresh with /remove-lifecycle rotten
  • Close this issue or PR with /close
  • Offer to help out with Issue Triage

Please send feedback to sig-contributor-experience at kubernetes/community.

/lifecycle rotten

/retitle how to deploy kube-batch on arm64

The Kubernetes project currently lacks enough active contributors to adequately respond to all issues and PRs.

This bot triages issues and PRs according to the following rules:

  • After 90d of inactivity, lifecycle/stale is applied
  • After 30d of inactivity since lifecycle/stale was applied, lifecycle/rotten is applied
  • After 30d of inactivity since lifecycle/rotten was applied, the issue is closed

You can:

  • Reopen this issue or PR with /reopen
  • Mark this issue or PR as fresh with /remove-lifecycle rotten
  • Offer to help out with Issue Triage

Please send feedback to sig-contributor-experience at kubernetes/community.

/close

@k8s-triage-robot: Closing this issue.

In response to this:

The Kubernetes project currently lacks enough active contributors to adequately respond to all issues and PRs.

This bot triages issues and PRs according to the following rules:

  • After 90d of inactivity, lifecycle/stale is applied
  • After 30d of inactivity since lifecycle/stale was applied, lifecycle/rotten is applied
  • After 30d of inactivity since lifecycle/rotten was applied, the issue is closed

You can:

  • Reopen this issue or PR with /reopen
  • Mark this issue or PR as fresh with /remove-lifecycle rotten
  • Offer to help out with Issue Triage

Please send feedback to sig-contributor-experience at kubernetes/community.

/close

Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository.