job status in pipes
#17857
Replies: 2 comments 1 reply
-
I think this is worthy of a bug report, filing one here #17868 |
Beta Was this translation helpful? Give feedback.
1 reply
-
The thinking so far has been
So definitely agree we need some opt-in "strict mode" , will focus #17868 on that. We may shift around what the right default behavior is as we continue to get feedback. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hi all!
we at Staffbase have been porting our Spark jobs to pipes (running in k8s) with some good overall results. I have observed something that could become a problem down the line, which I'd like to raise here to see if this intended behavior and how to potentially mitigate it.
For instance, when a Spark driver pod OOMs, typically the dagster context is not aware of this and will show the job as having completed successfully. I'm investigating how to report this back up the chain from the
spark-submit
process, but noticed in dagster logs that, if pipes is never fully initialized, dagster seems to think the job was successful too.I can see the following message in the logs:
From the user perspective, it would seem that if no back-and-forth communication channel was established, dagster should consider the job as failed, as this could otherwise lead to things failing silently.
What is the rationale behind the current semantics? Is there a "strict mode" where we can enforce this? Would it be possible to extend this even further to the notion that a job should be considered failed if no materialization was reported in the end?
Thanks!!
Beta Was this translation helpful? Give feedback.
All reactions