python - What to set `SPARK_HOME` to? -
installed apache-maven-3.3.3, scala 2.11.6, ran:
$ git clone git://github.com/apache/spark.git -b branch-1.4 $ cd spark $ build/mvn -dskiptests clean package
finally:
$ git clone https://github.com/apache/incubator-zeppelin $ cd incubator-zeppelin/ $ mvn install -dskiptests
then ran server:
$ bin/zeppelin-daemon.sh start
running simple notebook beginning %pyspark
, got error py4j
not being found. did pip install py4j
(ref).
now i'm getting error:
pyspark not responding traceback (most recent call last): file "/tmp/zeppelin_pyspark.py", line 22, in <module> pyspark.conf import sparkconf importerror: no module named pyspark.conf
i've tried setting spark_home
to: /spark/python:/spark/python/lib
. no change.
two environment variables required:
spark_home=/spark pythonpath=$spark_home/python:$spark_home/python/lib/py4j-version-src.zip:$pythonpath
Comments
Post a Comment