/
How to install Spark for Jupyter

How to install Spark for Jupyter

In order to configure GeoMesa Jupyter Kernels, we need to install Spark on the server running Jupyter.

Step-by-step guide

  1. Install Python 2.7 or 3.x for you Linux distribution.
  2. Download Spark.
  3. Unpack Spark and set up environment variables.

Note: Install Python before proceeding. Jupyter requires 2.7 or 3.x. Instructions for upgrading Python

Download and extract the binary files
wget http://d3kbcqa49mib13.cloudfront.net/spark-1.6.2-bin-hadoop2.6.tgz
tar xvf spark-1.6.2.tgz

You can select a different version here.

Configure environment variables, edit the paths as needed.

Setup environment variables
export SPARK_HADOOP_VERSION=2.6.0
export SPARK_YARN=true sbt/sbt assembly
export SPARK_HOME=~/spark-1.6.2-bin-hadoop2.6
PYTHONPATH=$SPARK_HOME/python:$SPARK_HOME/python/lib/py4j-0.9-src.zip:$PYTHONPATH

Add these to .bashrc as well.

Related content

How to install the Scala Spark (Apache Toree) Jupyter kernel with GeoMesa support
How to install the Scala Spark (Apache Toree) Jupyter kernel with GeoMesa support
More like this
Building and installing Toree
Building and installing Toree
More like this
How to build GeoMesa 1.2.x for Scala 2.10 / Spark 1.x
How to build GeoMesa 1.2.x for Scala 2.10 / Spark 1.x
More like this
Your first JIRA ticket
Your first JIRA ticket
More like this
GeoMesa 2.3.1 Release Notes
GeoMesa 2.3.1 Release Notes
More like this
How to install PIP for Jupyter
How to install PIP for Jupyter
More like this