How to share data from Spark RDD between two applications
问题 What is the best way to share spark RDD data between two spark jobs. I have a case where job 1: Spark Sliding window Streaming App, will be consuming data at regular intervals and creating RDD. This we do not want to persist to storage. Job 2: Query job that will access the same RDD created in job 1 and generate reports. I have seen few queries where they were suggesting SPARK Job Server, but as it is a open source not sure if it a possible solution, but any pointers will be of great help.