python - What to set `SPARK_HOME` to? -
installed apache-maven-3.3.3, scala 2.11.6, ran:
$ git clone git://github.com/apache/spark.git -b branch-1.4 $ cd spark $ build/mvn -dskiptests clean package finally:
$ git clone https://github.com/apache/incubator-zeppelin $ cd incubator-zeppelin/ $ mvn install -dskiptests then ran server:
$ bin/zeppelin-daemon.sh start running simple notebook beginning %pyspark, got error py4j not being found. did pip install py4j (ref).
now i'm getting error:
pyspark not responding traceback (most recent call last): file "/tmp/zeppelin_pyspark.py", line 22, in <module> pyspark.conf import sparkconf importerror: no module named pyspark.conf i've tried setting spark_home to: /spark/python:/spark/python/lib. no change.
two environment variables required:
spark_home=/spark pythonpath=$spark_home/python:$spark_home/python/lib/py4j-version-src.zip:$pythonpath
Comments
Post a Comment