For some reason javac errors when I import any of the Apache-Spark classes. I've gone through a number of setup guides to see if there's some environment variable I'm missing, but most don't have a step where an environment variable is set. Anyone have a clue what I'm missing? Spark Shell works and I'm able to run example scripts. Do I not compile with javac? Anyone have a clue what I'm missing?
Sounds like the jar is not in your classpath. You should have CLASSPATH environment variable, and one of the entries in it should be where the apache jar is stored.
If not, either add the relevant dir to the classpath, or move the jar to somewhere that's already in the classpath
It's been setup on an unix machine that I ssh to, I can start the Spark Shell and run some example scala code. The issue is I don't know if they did a crappy job setting it up or just a case of my user profile not having the class paths needed.