NJT145 - Overview

1

In order to use Apache Spark on windows, you need to have Java 8+.

2

  (If you have multiple java versions and if you want to chose one of them, set JAVA_HOME to your preferred one's directory.)

3

You also need python or scala to start coding...

4

You also need to have winutils files for hadoop, and set its directory path as HADOOP_HOME.

5

Of course, you need spark prebuilt package for the hadoop version that you will use.