2015-10-16 73 views
2

雖然開始工作節點我得到以下錯誤:星火工人無法連接到主

Spark Command: /usr/lib/jvm/default-java/bin/java -cp /home/ubuntu/spark-1.5.1-bin-hadoop2.6/sbin/../conf/:/home/ubuntu/spark-1.5.1-bin-hadoop2.6/lib/spark-assembly-1.5.1-hadoop2.6.0.jar:/home/ubuntu/spark-1.5.1-bin-hadoop2.6/lib/datanucleus-core-3.2.10.jar:/home/ubuntu/spark-1.5.1-bin-hadoop2.6/lib/datanucleus-rdbms-3.2.9.jar:/home/ubuntu/spark-1.5.1-bin-hadoop2.6/lib/datanucleus-api-jdo-3.2.6.jar -Xms1g -Xmx1g -XX:MaxPermSize=256m org.apache.spark.deploy.worker.Worker --webui-port 8081 spark://ip-1-70-44-5:7077 
Using Spark's default log4j profile: org/apache/spark/log4j-defaults.properties 
15/10/16 19:19:10 INFO Worker: Registered signal handlers for [TERM, HUP, INT] 
15/10/16 19:19:11 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 
15/10/16 19:19:11 INFO SecurityManager: Changing view acls to: ubuntu 
15/10/16 19:19:11 INFO SecurityManager: Changing modify acls to: ubuntu 
15/10/16 19:19:11 INFO SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(ubuntu); users with modify permissions: Set(ubuntu) 
15/10/16 19:19:12 INFO Slf4jLogger: Slf4jLogger started 
15/10/16 19:19:12 INFO Remoting: Starting remoting 
15/10/16 19:19:12 INFO Remoting: Remoting started; listening on addresses :[akka.tcp://[email protected]:55126] 
15/10/16 19:19:12 INFO Utils: Successfully started service 'sparkWorker' on port 55126. 
15/10/16 19:19:12 INFO Worker: Starting Spark worker 1.70.44.4:55126 with 2 cores, 2.9 GB RAM 
15/10/16 19:19:12 INFO Worker: Running Spark version 1.5.1 
15/10/16 19:19:12 INFO Worker: Spark home: /home/ubuntu/spark-1.5.1-bin-hadoop2.6 
15/10/16 19:19:12 INFO Utils: Successfully started service 'WorkerUI' on port 8081. 
15/10/16 19:19:12 INFO WorkerWebUI: Started WorkerWebUI at http://1.70.44.4:8081 
15/10/16 19:19:12 INFO Worker: Connecting to master ip-1-70-44-5:7077... 
15/10/16 19:19:24 INFO Worker: Retrying connection to master (attempt # 1) 
15/10/16 19:19:24 ERROR SparkUncaughtExceptionHandler: Uncaught exception in thread Thread[sparkWorker-akka.actor.default-dispatcher-5,5,main] 
java.util.concurrent.RejectedExecutionException: Task [email protected] rejected from [email protected][Running, pool size = 1, active threads = 0, queued tasks = 0, completed tasks = 0] 
     at java.util.concurrent.ThreadPoolExecutor$AbortPolicy.rejectedExecution(ThreadPoolExecutor.java:2048) 
     at java.util.concurrent.ThreadPoolExecutor.reject(ThreadPoolExecutor.java:821) 
     at java.util.concurrent.ThreadPoolExecutor.execute(ThreadPoolExecutor.java:1372) 
     at java.util.concurrent.AbstractExecutorService.submit(AbstractExecutorService.java:110) 
     at org.apache.spark.deploy.worker.Worker$$anonfun$org$apache$spark$deploy$worker$Worker$$tryRegisterAllMasters$1.apply(Worker.scala:211) 
     at org.apache.spark.deploy.worker.Worker$$anonfun$org$apache$spark$deploy$worker$Worker$$tryRegisterAllMasters$1.apply(Worker.scala:210) 
     at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:244) 
     at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:244) 
     at scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:33) 
     at scala.collection.mutable.ArrayOps$ofRef.foreach(ArrayOps.scala:108) 
     at scala.collection.TraversableLike$class.map(TraversableLike.scala:244) 
     at scala.collection.mutable.ArrayOps$ofRef.map(ArrayOps.scala:108) 
     at org.apache.spark.deploy.worker.Worker.org$apache$spark$deploy$worker$Worker$$tryRegisterAllMasters(Worker.scala:210) 
     at org.apache.spark.deploy.worker.Worker$$anonfun$org$apache$spark$deploy$worker$Worker$$reregisterWithMaster$1.apply$mcV$sp(Worker.scala:288) 
     at org.apache.spark.util.Utils$.tryOrExit(Utils.scala:1119) 
     at org.apache.spark.deploy.worker.Worker.org$apache$spark$deploy$worker$Worker$$reregisterWithMaster(Worker.scala:234) 
     at org.apache.spark.deploy.worker.Worker$$anonfun$receive$1.applyOrElse(Worker.scala:521) 
     at org.apache.spark.rpc.akka.AkkaRpcEnv.org$apache$spark$rpc$akka$AkkaRpcEnv$$processMessage(AkkaRpcEnv.scala:177) 
     at org.apache.spark.rpc.akka.AkkaRpcEnv$$anonfun$actorRef$lzycompute$1$1$$anon$1$$anonfun$receiveWithLogging$1$$anonfun$applyOrElse$4.apply$mcV$sp(AkkaRpcEnv.scala:126) 
     at org.apache.spark.rpc.akka.AkkaRpcEnv.org$apache$spark$rpc$akka$AkkaRpcEnv$$safelyCall(AkkaRpcEnv.scala:197) 
     at org.apache.spark.rpc.akka.AkkaRpcEnv$$anonfun$actorRef$lzycompute$1$1$$anon$1$$anonfun$receiveWithLogging$1.applyOrElse(AkkaRpcEnv.scala:125) 
     at scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33) 
     at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33) 
     at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25) 
     at org.apache.spark.util.ActorLogReceive$$anon$1.apply(ActorLogReceive.scala:59) 
     at org.apache.spark.util.ActorLogReceive$$anon$1.apply(ActorLogReceive.scala:42) 
     at scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118) 
     at org.apache.spark.util.ActorLogReceive$$anon$1.applyOrElse(ActorLogReceive.scala:42) 
     at akka.actor.Actor$class.aroundReceive(Actor.scala:467) 
     at org.apache.spark.rpc.akka.AkkaRpcEnv$$anonfun$actorRef$lzycompute$1$1$$anon$1.aroundReceive(AkkaRpcEnv.scala:92) 
     at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516) 
     at akka.actor.ActorCell.invoke(ActorCell.scala:487) 
     at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:238) 
     at akka.dispatch.Mailbox.run(Mailbox.scala:220) 
     at akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:397) 
     at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) 
     at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339) 
     at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) 
     at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) 
15/10/16 19:19:24 INFO ShutdownHookManager: Shutdown hook called 

我已經加入了主機名對conf /奴隸文件。我不知道要在spark-env.sh中設置哪些環境變量,所以沒有使用它。

任何指向解決方案的指針? 另外,如果我應該使用spark-env.sh,那麼應該運行哪個環境變量?

安裝細節: 2個ubuntu14機器,每個機器有兩個核心。

請指教。

感謝

+0

顯示我的'火花defaults.conf'內容請。 –

回答

0

所以,一些修修補補後,各地,我發現從沒有能夠在給定端口上與師父溝通。我更改了安全訪問規則並啓用了所有端口上的所有TCP通信。這解決了這個問題。

要檢查端口是開放的:

telnet master.ip master.port

的默認端口是7077.

我spark-env.sh:

export SPARK_WORKER_INSTANCES=2 export SPARK_MASTER_IP=<ip address>

+0

你是如何「啓用所有端口上的所有TCP通信」? –

+0

我在vm中使用這個,我改變了入站規則。方向:Ingress,協議:TCP,端口:1-6535,ip前綴:0.0.0.0/0 – Soumitra

0

我害怕你的主機名可能對Spark無效,並且你希望改變你的spark-env.sh。

您可以將變量SPARK_MASTER_IP設置爲master的真實ip,而不是其主機名。 例如

export SPARK_MASTER_IP=1.70.44.5 

INSTEAD OF

export SPARK_MASTER_IP=ip-1-70-44-5