Whenever I am trying to execute a simple processing in pyspark, it fails to open the socket.
>>> myRDD = sc.parallelize(range(6), 3) >>> sc
Its not the ideal solution but now I am aware of the cause. Pyspark is unable to create jvm socket with JDK 1.8 (64-bit) version, so I just set my java path to jdk 1.7 and it worked.