Hadoop Reducer Values in Memory?

前端 未结 3 2080
故里飘歌
故里飘歌 2020-12-30 11:31

I\'m writing a MapReduce job that may end up with a huge number of values in the reducer. I am concerned about all of these values being loaded into memory at once.

3条回答
  •  爱一瞬间的悲伤
    2020-12-30 12:05

    It's not entirely in memory, some of it comes from the disk, looking at the code seems like the framework breaks the Iterable into segments, and load them form disk into memory 1 by one.

    org.apache.hadoop.mapreduce.task.ReduceContextImpl org.apache.hadoop.mapred.BackupStore

提交回复
热议问题