Sparks give Null pointer exception during InputSplit for Hbase

I am using Spark 1.2.1, Hbase 0.98.10 and Hadoop 2.6.0. I got a zero point exception while retrieving the hbase data form. Find the stack trace below.

[sparkDriver-akka.actor.default-dispatcher-2] DEBUG NewHadoopRDD - Failed to use InputSplit # getLocationInfo. java.lang.NullPointerException: null on scala.collection.mutable.ArrayOps $ ofRef $ .length $ extension (ArrayOps.scala: 114) ~ [scala -library-2.10.4.jar: na] at scala.collection.mutable. ArrayOps $ ofRef.length (ArrayOps.scala: 114) ~ [scala -library-2.10.4.jar: na] at scala.collection.IndexedSeqOptimized $ class.foreach (IndexedSeqOptimized.scala: 32) ~ [scala -library-2.10 .4.jar: na] at scala.collection.mutable.ArrayOps $ ofRef.foreach (ArrayOps.scala: 108) ~ [scala -library-2.10.4.jar: na] at org.apache.spark.rdd.HadoopRDD $ .convertSplitLocationInfo (HadoopRDD.scala: 401) ~ [spark-core_2.10-1.2.1.jar: 1.2.1] at org.apache.spark.rdd.NewHadoopRDD.getPreferredLocations (NewHadoopRDD.scala: 215) ~ [spark -core_2.10-1.2.1.jar: 1.2.1] at org.apache.spark.rdd.RDD $$ anonfun $ preferredLocations $ 2.Apply (RDD.scala: 234) [spark-core_2.10-1.2.1.jar: 1.2.1] at org.apache.spark.rdd.RDD $ $ anonfun $ preferredLocations $ 2.Apply (RDD.scala: 234) [spark-core_2.10-1.2.1.jar: 1.2.1] at scala.Option.getOrElse (Option.scala: 120) [scala -library-2.10 .4.jar: na] at org.apache.spark.rdd.RDD.preferredLocations (RDD.scala: 233) [spark-core_2.10-1.2.1.jar: 1.2.1] at org.apache.spark. scheduler.DAGScheduler.org $ apache $ spark $ scheduler $ DAGScheduler $$ getPreferredLocsInternal (DAGScheduler.scala: 1326) [spark-core_2.10-1.2.1.jar: 1.2.1] at org.apache.spark.scheduler.DAGScheduler $$ anonfun $ org $ apache $ spark $ scheduler $ DAGScheduler $$ getPreferredLocsInternal $ 2 $$ anonfun $ apply $ 2.Apply $ mcVI ​​$ sp (DAGScheduler.scala: 1336) [spark-core_2.10-1.2.1.jar : 1.2.1] at org.apache.spark.scheduler.DAGScheduler $$ anonfun $ org $ Apache $ spark $ scheduler $ DAGScheduler $$ getPreferredLocsInternal $ 2 $$ anonfun $ apply $ 2.Apply (DAGScheduler.scala: 1335) [spark-core_2.10-1.2.1.jar: 1.2.1] at org.apache.spark.scheduler.DAGScheduler $$ anonfun $ org $ Apache $ spark $ scheduler $ DAGScheduler $$ getPreferredLocsInternal $ 2 $$ anonfun $ apply $ 2.Apply (DAGScheduler.scala: 1335) [spark-core_2.10-1.2 .1.jar: 1.2.1] at scala.collection.immutable.List.foreach (List.scala: 318) [scala -library-2.10.4.jar: na] at org.apache.spark.scheduler.DAGScheduler $ $ anonfun $ org $ apache $ spark $ scheduler $ DAGScheduler $$ getPreferredLocsInternal $ 2.Apply (DAGScheduler.scala: 1335) [spark-core_2.10-1.2.1.jar: 1.2.1] at org.apache.spark.scheduler .DAGScheduler $$ anonfun $ org $ Apache $ spark $ scheduler $ DAGScheduler $$ getPreferredLocsInternal $ 2.Apply (DAGScheduler.scala: 1333) [spark-core_2.10-1.2.1.jar: 1.2.1] at scala.collection.immutable.List.foreach (List.scala: 318) [scala -library-2.10.4.jar: na ] at org.apache.spark.scheduler.DAGScheduler.org $ apache $ spark $ scheduler $ DAGScheduler $$ getPreferredLocsInternal (DAGScheduler.scala: 1333) [spark-core_2.10-1.2.1.jar: 1.2.1] at org .apache.spark.scheduler.DAGScheduler $$ anonfun $ org $ Apache $ spark $ planner $ DAGScheduler $$ getPreferredLocsInternal $ 2 $$ anonfun $ apply $ 2.Apply $ mcVI ​​$ sp (DAGScheduler.scala: 1336) [spark-core_ .10-1.2.1.jar: 1.2.1] at org.apache.spark.scheduler.DAGScheduler $$ anonfun $ org $ Apache $ spark $ scheduler $ DAGScheduler $$ getPreferredLocsInternal $ 2 $$ anonfun $ apply $ 2.Apply (DAGScheduler .scala: 1335) [spark-core_2.10-1.2.1.jar: 1.2.1] at org.apache.spark.scheduler.DAGScheduler $$ anonfun $ org $ Apache $ spark $ scheduler $ DAGScheduler $$ getPreferredLocsInternal $ 2 $$ anonfun $ apply $ 2.Apply (DAGScheduler.scala: 1335) [spark-core_2.10-1.2.1.jar: 1.2.1] at scala.collection.immutable.List.foreach (List.scala: 318) [scala -library-2.10.4.jar: na] at org.apache.spark.scheduler.DAGScheduler $$ anonfun $ org $ apache $ spark $ scheduler $ DAGScheduler $$ getPreferredLocsInternal $ 2.Apply (DAGScheduler.scala: 1335) [spark-core_2.10-1.2.1.jar: 1.2.1] at org.apache.spark.scheduler.DAGScheduler $$ anonfun $ org $ apache $ spark $ planner $ DAGScheduler $$ getPreferredLocsInternal $ 2.Apply (DAGScheduler.scala: 1333) [spark-core_2.10-1.2.1.jar: 1.2.1] at scala.collection.immutable.List.foreach (List.scala : 318) [scala -library-2.10.4.jar: na] at org.apache.spark.scheduler.DAGScheduler.org $ apache $ spark $ scheduler $ DAGScheduler $$ getPreferredLocsInternal (DAGScheduler.scala: 1333) [spark-core_2.10-1.2.1.jar: 1.2.1] at org.apache.spark.scheduler.DAGScheduler.getPreferredLocs (DAGScheduler .scala: 1304) [spark-core_2.10-1.2.1.jar: 1.2.1] at org.apache.spark.scheduler.DAGScheduler $$ anonfun $ 17.apply (DAGScheduler.scala: 862) [Spark core_2.10 -1.2.1.jar: 1.2.1] on org.apache.spark.scheduler.DAGScheduler $$ anonfun $ 17.apply (DAGScheduler.scala: 859) [spark-core_2.10-1.2.1.jar: 1.2.1 ] at scala.collection.TraversableLike $$ anonfun $ Map $ 1.Apply (TraversableLike.scala: 244) [scala -library-2.10.4.jar: na] at scala.collection.TraversableLike $$ anonfun $ Map $ 1.Apply ( TraversableLike.scala: 244) [scala -library-2.10.4.jar: na] at scala.collection.Iterator $ class.foreach (Iterator.scala: 727) [scala -library-2.10.4.jar: na] at scala.collection.AbstractIterator.foreach (Iterator.scala: 1157) [scala -library-2.10.4.jar: na] at scala.collection.IterableLike $ class.foreach (IterableLike.scala: 72) [scala -library-2.10.4 .jar: na] at scala.collection.AbstractIterable.foreach (Iterable.scala: 54) [scala -library-2.10.4.jar: na] at scala.collection.TraversableLike $ class.map (TraversableLike.scala: 244) [scala -library-2.10.4.jar: na] at scala.collection.AbstractTraversable.map (Traversable.scala: 105) [scala -library-2.10.4.jar: na] at org.apache.spark.scheduler. DAGScheduler.org $ Apache $ spark $ scheduler $ DAGScheduler $$ submitMissingTasks (DAGScheduler.scala: 859) [spark-core_2.10-1.2.1.jar: 1.2.1] at org.apache.spark.scheduler.DAGScheduler.org $ apache $ sparks $ planner $ DAGScheduler $$ submitStage (DAGScheduler.scala: 778) [spark-core_2.10-1.2.1.jar: 1.2.1] at org.apache.spark.scheduler.DAGScheduler.handleJobSubmitted (DAGScheduler.scala: 762) [spark-core_2.10-1.2.1.jar: 1.2.1] at org.apache.spark.scheduler.DAGSchedulerEventProcessActor $$ anonfun $ get $ 2.applyOrElse (DAGScheduler. scala: 1389) [spark-core_2.10-1.2.1.jar: 1.2.1] at akka.actor.Actor $ class.aroundReceive (Actor.scala: 465) [akka-actor_2.10-2.3.4-spark .jar: na] at org.apache.spark.scheduler.DAGSchedulerEventProcessActor.aroundReceive (DAGScheduler.scala: 1375) [spark-core_2.10-1.2.1.jar: 1.2.1] at akka.actor.ActorCell.receiveMessage ( ActorCell.scala: 516) [akka-actor_2.10-2.3.4-spark.jar: na] at akka.actor.ActorCell.invoke (ActorCell.scala: 487) [akka-actor_2.10-2.3.4-spark .jar: na] at akka.dispatch.Mailbox.processMailbox (Mailbox.scala: 238) [akka-actor_2.10-2.3.4-spark.jar: na] at akka.dispatch.Mailbox.run (Mailbox.scala: 220) [akka-actor_2.10-2.3.4-spark.jar: na] at akka.dispatch.ForkJoinExecutorConfigurator $ AkkaForkJoinTask.exec (AbstractDispatcher.scala: 393) [akka-actor_2.10-2.3.4-spark.jar: na] at scala.concurrent.forkjoin.ForkJoinTask.doExecececoecececoececec .java: 260) [scala -library-2.10.4.jar: na] at scala.concurrent.forkjoin.ForkJoinPool $ WorkQueue.runTask (ForkJoinPool.java:1339) [scala -library-2.10.4.jar: na] at scala.concurrent.forkjoin.ForkJoinPool.runWorker (ForkJoinPool.java:1979) [scala -library-2.10.4.jar: na] at scala.concurrent.forkjoin.ForkJoinWorkerThread.run (ForkJoinWorkerThread.java10 library-2.10.4.jar: on]forkjoin.ForkJoinPool $ WorkQueue.runTask (ForkJoinPool.java:1339) [scala -library-2.10.4.jar: na] at scala.concurrent.forkjoin.ForkJoinPool.runWorker (ForkJoinPool.java:1979) [scala -l.10l .4.jar: na] at scala.concurrent.forkjoin.ForkJoinWorkerThread.run (ForkJoinWorkerThread.java:107) [scala -library-2.10.4.jar: on]forkjoin.ForkJoinPool $ WorkQueue.runTask (ForkJoinPool.java:1339) [scala -library-2.10.4.jar: na] at scala.concurrent.forkjoin.ForkJoinPool.runWorker (ForkJoinPool.java:1979) [scala -l.10l .4.jar: na] at scala.concurrent.forkjoin.ForkJoinWorkerThread.run (ForkJoinWorkerThread.java:107) [scala -library-2.10.4.jar: on]

, .

+4
1

getPreferredLocations, hbase ConfigurationI hbase.table.name hbase.master( , , HMaster) ,

+1

All Articles