python - What to set `SPARK_HOME` to? -


installed apache-maven-3.3.3, scala 2.11.6, ran:

$ git clone git://github.com/apache/spark.git -b branch-1.4 $ cd spark $ build/mvn -dskiptests clean package 

finally:

$ git clone https://github.com/apache/incubator-zeppelin $ cd incubator-zeppelin/ $ mvn install -dskiptests 

then ran server:

$ bin/zeppelin-daemon.sh start 

running simple notebook beginning %pyspark, got error py4j not being found. did pip install py4j (ref).

now i'm getting error:

pyspark not responding traceback (most recent call last):   file "/tmp/zeppelin_pyspark.py", line 22, in <module>     pyspark.conf import sparkconf importerror: no module named pyspark.conf 

i've tried setting spark_home to: /spark/python:/spark/python/lib. no change.

two environment variables required:

spark_home=/spark pythonpath=$spark_home/python:$spark_home/python/lib/py4j-version-src.zip:$pythonpath 

Comments

Popular posts from this blog

c# - Validate object ID from GET to POST -

node.js - Custom Model Validator SailsJS -

php - Find a regex to take part of Email -