- Spark should be locally set up.
- Be able to start spark shell in local mode.
- Be able to start spark shell in yarn mode.
- java 1.8 is installed (a tip: manage ur java versions using https://albertattard.github.io/java-boot-camp/docs/primer/sdkman/)
- scala is installed
brew install scala
- Hadoop local setup should completed to ensure its( daemons like Namenode, Resources Manager etc) are running.
- For yarn mode : in .bash_profile or .zshrc
export HADOOP_CONF_DIR=/Users/<YOUR USER>/hadoop-2.7.7/etc/hadoop
ORYARN_CONF_DIR=/Users/<YOUR USER>/hadoop-2.7.7/etc/hadoop
spark-shell --master local
spark-shell --master yarn
Official docs : https://spark.apache.org/docs/2.2.1/running-on-yarn.html