What is RDD in spark

后端 未结 9 1515
傲寒
傲寒 2020-12-12 19:20

Definition says:

RDD is immutable distributed collection of objects

I don\'t quite understand what does it mean. Is it like da

9条回答
  •  抹茶落季
    2020-12-12 20:04

    To compare RDD with scala collection, below are few differences

    1. Same but runs on a cluster
    2. Lazy in nature where scala collections are strict
    3. RDD is always Immutable i.e., you can not change the state of the data in the collection
    4. RDD are self recovered i.e., fault-tolerant

提交回复
热议问题