I installed Spark using the AWS EC2 manual, and I can run the program using the bin/pyspark script to get to the spark hint and also successfully complete the quick start task.
However, I cannot understand for life how to stop all the detailed INFO entries after each command.
I used almost all the possible scripts in the code below (commenting, setting to OFF) in my log4j.properties file in the conf folder, where I run the application, and also on each node and does nothing. I still get logging of INFO statements after each statement is executed.
I am very confused about how this should work.
Here is my full classpath when I use SPARK_PRINT_LAUNCH_COMMAND :
Spark command: / Library / Java / JavaVirtualMachines / jdk 1.8.0_05.jdk / Contents / Home / bin / java -cp: /root/spark-1.0.1-bin-hadoop2/conf: /root/spark-1.0.1 -bin-hadoop2 / conf: /root/spark-1.0.1-bin-hadoop2/lib/spark- assembly-1.0.1-hadoop2.2.0.jar: /root/spark-1.0.1-bin-hadoop2/lib /datanucleus-api-jdo-3.2.1.jar:/root/spark-1.0.1-bin- hadoop2 / Lib / DataNucleus kernel-3.2.2.jar: /root/spark-1.0.1-bin-hadoop2/ lib / datanucleus-rdbms-3.2.1.jar -XX: MaxPermSize = 128m -Djava.library.path = -Xms512m -Xmx512m org.apache.spark.deploy.SparkSubmit spark-shell -class org.apache.spark.repl. Main
spark-env.sh content:
#!/usr/bin/env bash
python scala hadoop yarn apache-spark pyspark
horatio1701d Aug 07 '14 at 22:48 2014-08-07 22:48
source share