diff --git a/docs/modules/spark-k8s/examples/getting_started/getting_started.sh b/docs/modules/spark-k8s/examples/getting_started/getting_started.sh index 0c012c5a..36e5f126 100755 --- a/docs/modules/spark-k8s/examples/getting_started/getting_started.sh +++ b/docs/modules/spark-k8s/examples/getting_started/getting_started.sh @@ -45,6 +45,25 @@ exit 1 ;; esac +# TODO: Remove once https://github.com/stackabletech/issues/issues/828 has been +# implemented (see that issue for details). +until kubectl get crd sparkapplications.spark.stackable.tech >/dev/null 2>&1; do + echo "Waiting for CRDs to be installed" + sleep 1 +done +until kubectl get crd sparkhistoryservers.spark.stackable.tech >/dev/null 2>&1; do + echo "Waiting for CRDs to be installed" + sleep 1 +done +until kubectl get crd sparkconnectservers.spark.stackable.tech >/dev/null 2>&1; do + echo "Waiting for CRDs to be installed" + sleep 1 +done +until kubectl get crd sparkapptemplates.spark.stackable.tech >/dev/null 2>&1; do + echo "Waiting for CRDs to be installed" + sleep 1 +done + echo "Creating a Spark Application..." # tag::install-sparkapp[] kubectl apply -f application.yaml @@ -54,16 +73,12 @@ sleep 15 echo "Waiting for job to complete ..." # tag::wait-for-job[] -kubectl wait pods -l 'job-name=pyspark-pi' \ +if kubectl wait pods -l 'job-name=pyspark-pi' \ --for jsonpath='{.status.phase}'=Succeeded \ - --timeout 300s -# end::wait-for-job[] - -result=$(kubectl logs -l 'spark-role=driver' --tail=-1 | grep "Pi is roughly") - -if [ "$result" == "" ]; then - echo "Log result was not found!" - exit 1 + --timeout 300s; then + echo "job succeeded" else - echo "Job result:" "$result" + echo "job failed" + exit 1 fi +# end::wait-for-job[] diff --git a/docs/modules/spark-k8s/examples/getting_started/getting_started.sh.j2 b/docs/modules/spark-k8s/examples/getting_started/getting_started.sh.j2 index 179357f4..358ad5dc 100755 --- a/docs/modules/spark-k8s/examples/getting_started/getting_started.sh.j2 +++ b/docs/modules/spark-k8s/examples/getting_started/getting_started.sh.j2 @@ -45,6 +45,25 @@ exit 1 ;; esac +# TODO: Remove once https://github.com/stackabletech/issues/issues/828 has been +# implemented (see that issue for details). +until kubectl get crd sparkapplications.spark.stackable.tech >/dev/null 2>&1; do + echo "Waiting for CRDs to be installed" + sleep 1 +done +until kubectl get crd sparkhistoryservers.spark.stackable.tech >/dev/null 2>&1; do + echo "Waiting for CRDs to be installed" + sleep 1 +done +until kubectl get crd sparkconnectservers.spark.stackable.tech >/dev/null 2>&1; do + echo "Waiting for CRDs to be installed" + sleep 1 +done +until kubectl get crd sparkapptemplates.spark.stackable.tech >/dev/null 2>&1; do + echo "Waiting for CRDs to be installed" + sleep 1 +done + echo "Creating a Spark Application..." # tag::install-sparkapp[] kubectl apply -f application.yaml @@ -54,16 +73,12 @@ sleep 15 echo "Waiting for job to complete ..." # tag::wait-for-job[] -kubectl wait pods -l 'job-name=pyspark-pi' \ +if kubectl wait pods -l 'job-name=pyspark-pi' \ --for jsonpath='{.status.phase}'=Succeeded \ - --timeout 300s -# end::wait-for-job[] - -result=$(kubectl logs -l 'spark-role=driver' --tail=-1 | grep "Pi is roughly") - -if [ "$result" == "" ]; then - echo "Log result was not found!" - exit 1 + --timeout 300s; then + echo "job succeeded" else - echo "Job result:" "$result" + echo "job failed" + exit 1 fi +# end::wait-for-job[] diff --git a/docs/modules/spark-k8s/images/getting_started/spark_log.png b/docs/modules/spark-k8s/images/getting_started/spark_log.png deleted file mode 100644 index 2e77479e..00000000 Binary files a/docs/modules/spark-k8s/images/getting_started/spark_log.png and /dev/null differ diff --git a/docs/modules/spark-k8s/pages/getting_started/first_steps.adoc b/docs/modules/spark-k8s/pages/getting_started/first_steps.adoc index 93e10415..e59a2e0a 100644 --- a/docs/modules/spark-k8s/pages/getting_started/first_steps.adoc +++ b/docs/modules/spark-k8s/pages/getting_started/first_steps.adoc @@ -56,8 +56,3 @@ The initial job is persisted for several minutes before being removed. The completed state looks like this: image::getting_started/spark_complete.png[Completed job] - -The driver logs can be inspected for more information about the results of the job. -In this case the result of our (approximate!) pi calculation can be found: - -image::getting_started/spark_log.png[Driver log]