My build.sbt file has the following:
scalaVersion := "2.10.3" libraryDependencies += "com.databricks" % "spark-csv_2.10" % "1.1.0"
I run Spark in an offline cluster mode, and SparkConf is SparkConf().setMaster("spark://ec2-[ip].compute-1.amazonaws.com:7077").setAppName("Simple Application") (I don't using the setJars method, not sure if I need this).
I will pack the jar using the sbt package command. The command I use to start the application is ./bin/spark-submit --master spark://ec2-[ip].compute-1.amazonaws.com:7077 --class "[classname]" target/scala-2.10/[jarname]_2.10-1.0.jar .
When doing this, I get this error:
java.lang.RuntimeException: Failed to load class for data source: com.databricks.spark.csv
What is the problem?
source share