Override hadoop's mapreduce.fileoutputcommitter.marksuccessfuljobs in oozie

后端 未结 3 997
盖世英雄少女心
盖世英雄少女心 2020-12-19 06:49

mapreduce.fileoutputcommitter.marksuccessfuljobs
false

I wan

相关标签:
3条回答
  • 2020-12-19 06:54

    I ran into the same issue and ended up using a shell action to create the flag.

    Here's a full example: http://nathan.vertile.com/blog/2014/09/02/oozie-data-pipeline-done-flag/

    0 讨论(0)
  • 2020-12-19 06:58

    Hive unfortunately overrides this capability by setting it's own NullOutputComitter:

    conf.setOutputCommitter(NullOutputCommitter.class);
    

    see

    src/shims/src/0.20/java/org/apache/hadoop/hive/shims/Hadoop20Shims.java
    src/shims/src/common-secure/java/org/apache/hadoop/hive/shims/HadoopShimsSecure.java
    

    Looks like you'll have to create the flag manually. We have filed HIVE-3700 for this.

    0 讨论(0)
  • 2020-12-19 07:08

    You can add 'dfs' command to your hive script, like

    dfs -touchz '$table_base_path'/dt='${partition}'/_SUCCESS

    https://archive.cloudera.com/cdh4/cdh/4/hive/language_manual/cli.html

    0 讨论(0)
提交回复
热议问题