Skip to content
This repository has been archived by the owner on Apr 24, 2023. It is now read-only.

executor pod schedule stucked with enough resource #152

Open
askeySnip opened this issue Oct 28, 2020 · 2 comments
Open

executor pod schedule stucked with enough resource #152

askeySnip opened this issue Oct 28, 2020 · 2 comments

Comments

@askeySnip
Copy link

when i submit a batch of spark jobs, it runs doesn't like the expection.
Some executor pods stucking although there are enough resources in each node for it to run.
It annoyed me, and I wonder if there is something that I don't considered.
ps. I run these spark jobs like the example and it works ok for running a single job

@onursatici
Copy link
Contributor

@askeySnip can you describe the stuck driver pods and share the scheduling errors?

@chia7712
Copy link
Contributor

I encountered similar issue.

There are two nodes in my cluster. Also, there are two spark jobs and the total resources they required are larger than the cluster (i.e the k8s cluster can't run both jobs concurrently). If I submit second job after all executors of first job are running, it works well. However, some pods get hang (see following screenshot) if I submit two jobs at the same time. I traced the log and it seems that scheduler predicate the node (assign the resource) for both jobs at the same time. Hence, some pods can't get enough resources.

截圖 2021-07-10 下午7 57 28

截圖 2021-07-10 下午8 00 05

Is it the expected behavior? Can it be configured that scheduler predicates second job only if first job has been scheduled successfully? Or we should NOT submit jobs at the same time?

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants