I configure Spark 0.9 on AWS and find that when I launch the Pyspark interactive shell, my artists / remote workers first register:
14/07/08 22:48:05 INFO cluster.SparkDeploySchedulerBackend: Registered executor: Actor[akka.tcp:// sparkExecutor@ip-xx-xx-xxx-xxx.ec2.internal :54110/user/ Executor
and then disassociated almost immediately before I have a chance to run something:
14/07/08 22:48:05 INFO cluster.SparkDeploySchedulerBackend: Executor 0 disconnected, so removing it 14/07/08 22:48:05 ERROR scheduler.TaskSchedulerImpl: Lost an executor 0 (already removed): remote Akka client disassociated
Any idea what could be wrong? I tried setting the JVM parameters spark.akka.frameSize and spark.akka.timeout, but I am sure that this is not a problem, because (1) I do not start anything to start, and (2) my artists disconnect a few seconds after the start. which corresponds to a timeout of 100 seconds by default.
Thanks!
Jack
source share