spark-shell error on Windows - can it be ignored if not using hadoop?

后端 未结 1 728
半阙折子戏
半阙折子戏 2020-12-01 12:02

I got the following error when starting the spark-shell. I\'m going to use Spark to process data in SQL Server. Can I ignore the errors?

java.io.IOExcept

1条回答
  •  [愿得一人]
    2020-12-01 12:16

    tl;dr You'd rather not.

    Well, it may be possible, but given you've just started your journey to Spark's land the efforts would not pay off.


    Windows has never been a developer-friendly OS to me and whenever I teach people Spark and they use Windows I just take it as granted that we'll have to go through the winutils.exe setup but many times also how to work on command line.


    Please install winutils.exe as follows:

    1. Run cmd as administrator
    2. Download winutils.exe binary from https://github.com/steveloughran/winutils repository (use hadoop-2.7.1 for Spark 2)
    3. Save winutils.exe binary to a directory of your choice, e.g. c:\hadoop\bin
    4. Set HADOOP_HOME to reflect the directory with winutils.exe (without bin), e.g. set HADOOP_HOME=c:\hadoop
    5. Set PATH environment variable to include %HADOOP_HOME%\bin
    6. Create c:\tmp\hive directory
    7. Execute winutils.exe chmod -R 777 \tmp\hive
    8. Open spark-shell and run spark.range(1).show to see a one-row dataset.

    0 讨论(0)
提交回复
热议问题