NJT145 - Overview
1
In order to use Apache Spark on windows, you need to have Java 8+.
2
(If you have multiple java versions and if you want to chose one of them, set JAVA_HOME to your preferred one's directory.)
3
You also need python or scala to start coding...
4
You also need to have winutils files for hadoop, and set its directory path as HADOOP_HOME.
5
Of course, you need spark prebuilt package for the hadoop version that you will use.