Bluemix Spark: spark-submit failing when downloading stderr and stdout? -


i using spark service in ibm bluemix. trying launch java piece of code executing spark process using spark-submit.sh script.

my command line is:

./spark-submit.sh --vcap ./vcap.json --deploy-mode cluster --class org.apache.spark.examples.javasparkpi \ --master https://169.54.219.20 ~/documents/spark/javasparkpi.jar  

i using latest spark-submit.sh version (as of yesterday).

./spark-submit.sh --version spark-submit.sh  version : '1.0.0.0.20160330.1' 

this worked fine couple of weeks ago (with old spark-submit.sh) getting following error:

downloading stdout_1461024849908170118   % total    % received % xferd  average speed   time    time     time  current                                  dload  upload   total   spent    left  speed   0    89    0    89    0     0     56      0 --:--:--  0:00:01 --:--:--   108 failed download workdir/driver-20160418191414-0020-5e7fb175-6856-4980-97bc-8e8aa0d1f137/stdout     stdout_1461024849908170118  downloading stderr_1461024849908170118   % total    % received % xferd  average speed   time    time     time  current                                  dload  upload   total   spent    left  speed   0    89    0    89    0     0     50      0 --:--:--  0:00:01 --:--:--   108 failed download workdir/driver-20160418191414-0020-5e7fb175-6856-4980-97bc-8e8aa0d1f137/stderr     stderr_1461024849908170118 

any ideas on doing wrong? in advance.

edit:

by looking @ log file have found problem not while downloading stdout , stderr when submitting job.

{   "action" : "submissionstatusresponse",   "driverstate" : "failed",   "message" : "exception cluster: org.apache.spark.sparkexception: failed change container cwd org.apache.spark.deploy.master.egoapplicationmanager.egodriverexitcallback(egoapplicationmanager.scala:168) org.apache.spark.deploy.master.masterscheduledelegatordriver.oncontainerexit(masterscheduledelegatordriver.scala:144) org.apache.spark.deploy.master.resourcemanager.resourcemanageregoslot.handleactivityfinish(resourcemanageregoslot.scala:555) org.apache.spark.deploy.master.resourcemanager.resourcemanageregoslot.callbackcontainerstatechg(resourcemanageregoslot.scala:525) org.apache.spark.deploy.master.resourcemanager.resourcecallbackmanager$$anonfun$callbackcontainerstatechg$1.apply(resourcemanager.scala:158) org.apache.spark.deploy.master.resourcemanager.resourcecallbackmanager$$anonfun$callbackcontainerstatechg$1.apply(resourcemanager.scala:157) scala.option.foreach(option.scala:236) org.apache.spark.deploy.master.resourcemanager.resourcecallbackmanager$.callbackcontainerstatechg(resourcemanager.scala:157)",   "serversparkversion" : "1.6.0",   "submissionid" : "driver-20160420043532-0027-6e579720-2c9d-428f-b2c7-6613f4845146",   "success" : true } driverstatus failed 

edit2:

finally problem had when submitting job has been solved creating brand new instance of spark service. job executes , finishes after few seconds.

but still receive error when trying download stdout , stderr files.

downloading stdout_1461156506108609180 % total    % received % xferd  average speed   time    time     time  current dload  upload   total   spent    left  speed   0    90    0    90    0     0     61      0 --:--:--  0:00:01 --:--:--   125 failed download workdir2/driver-20160420074922-0008-1400fc20-95c1-442d-9c37-32de3a7d1f0a/stdout stdout_1461156506108609180  downloading stderr_1461156506108609180 % total    % received % xferd  average speed   time    time     time  current dload  upload   total   spent    left  speed   0    90    0    90    0     0     56      0 --:--:--  0:00:01 --:--:--   109 failed download workdir2/driver-20160420074922-0008-1400fc20-95c1-442d-9c37-32de3a7d1f0a/stderr stderr_1461156506108609180 

any ideas?

i found old spark-submit trying retrieve stdout , stderr workdir folder ...

failed download workdir/driver-20160418191414-0020-5e7fb175-6856-4980-97bc-8e8aa0d1f137/stdout     stdout_1461024849908170118 

while new (downloaded yesterday) spark-submit trying download them workdir2 folder ...

failed download workdir2/driver-20160420074922-0008-1400fc20-95c1-442d-9c37-32de3a7d1f0a/stdout stdout_1461156506108609180 

the folder in use fixed variable ss_spark_work_dir initialized in spark-submit

if [ -z ${ss_spark_work_dir} ];  ss_spark_work_dir="workdir2"; fi # work directory on spark cluster 

i changed value workdir , works now. have downloaded new (today) spark-submit bluemix site , problem has been fixed. variable points workdir.

so, if fails, sure got last spark-submit script bluemix.


Comments

Popular posts from this blog

Ansible - ERROR! the field 'hosts' is required but was not set -

customize file_field button ruby on rails -

SoapUI on windows 10 - high DPI/4K scaling issue -