You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I set the restartPolicy to type: Always, but it never restarts on error, OOM Kill.
And when I edit the driver configuration in Kubernetes, I see that the policy applied is: NEVER
Whereas in the SparkApplication, it's Always!
Tip: To restart the application, sometimes go to the shell in the pod, do an exit 1 and it restarts, if it remains in error, delete the pod and it restarts... Otherwise, delete the SparkAplication and Upgrade the chart if necessary / re-apply the yaml.
On spark application : restartPolicy type: Always
On driver config : restartPolicy type: Never (automatic definition by spark operator bug ?)
This issue has been automatically marked as stale because it has not had recent activity. It will be closed if no further activity occurs. Thank you for your contributions.
Description
We are using spark operator v1beta2-1.6.2-3.5.0 in production. We have a spark application with the below policy.
When driver pod failed for some reason, operator has the below logs
driver pod is in error state and the sparkapplication state is
only after deleting the sparkapplication manually, operator has the logs and started the spark application
How to make sure that my sparkapplication gets restarted when the driver failed, this is happening regularly.
Reproduction Code [Required]
Submit the sparkApplication to spark operator with
when the driver pod fails, sparkapplication fails and doesn't submit again
Expected behavior
Spark application should be restarted
Actual behavior
SparkApplication has finish time and is in Failing State.
Terminal Output Screenshot(s)
Environment & Versions
Additional context
The text was updated successfully, but these errors were encountered: