hadoop2

Hadoop2.4.1分布式安装

喜夏-厌秋 提交于 2019-12-02 04:44:29
1、做好下文中的所有配置: Hadoop1.2.1分布式安装-1-准备篇 。 2、Hadoop2.x的发行版中有个小问题:libhadoop.so.1.0.0在64位OS中存在问题,因为它是32位的,在64位OS中 hadoop启动时会报一个WARN的日志 。这个包的作用是调用native的api,可以提高hadoop的性能,如果这个包失效,那就是使用jvm做压缩等工作,效率就会很低。处理方法就是重新编译Hadoop,见xxx(link article)。 3、 在打算做namenode的机器上,wget或其他方式下载hadoop的压缩包,并解压到本地指定目录。 下载解压命令参考 Linux常用命令 。 4、各种配置文件和hadoop1会有所不同,共有七个文件,以下分别描述。 /hadoop-2.4.1/etc/hadoop/hadoop-env.sh # The java implementation to use. export JAVA_HOME=${JAVA_HOME} /hadoop-2.4.1/etc/hadoop/yarn-env.sh # some Java parameters # export JAVA_HOME=/home/y/libexec/jdk1.6.0/ if [ "$JAVA_HOME" != "" ]; then #echo "run java

NULL column names in Hive query result

陌路散爱 提交于 2019-12-02 00:48:37
I have downloaded the weather .txt files from NOAA , which looks like: WBAN,Date,Time,StationType,SkyCondition,SkyConditionFlag,Visibility,VisibilityFlag,WeatherType,WeatherTypeFlag,DryBulbFarenheit,DryBulbFarenheitFlag,DryBulbCelsius,DryBulbCelsiusFlag,WetBulbFarenheit,WetBulbFarenheitFlag,WetBulbCelsius,WetBulbCelsiusFlag,DewPointFarenheit,DewPointFarenheitFlag,DewPointCelsius,DewPointCelsiusFlag,RelativeHumidity,RelativeHumidityFlag,WindSpeed,WindSpeedFlag,WindDirection,WindDirectionFlag,ValueForWindCharacter,ValueForWindCharacterFlag,StationPressure,StationPressureFlag,PressureTendency

How to check whether the file exist in HDFS location, using oozie?

醉酒当歌 提交于 2019-12-01 22:24:36
问题 How to check whether a file in HDFS location is exist or not, using Oozie? In my HDFS location I will get a file like this test_08_01_2016.csv at 11PM , on a daily basis. I want check whether this file exist after 11.15 PM. I can schedule the batch using a Oozie coordinator job. But how can I validate if the file exists in HDFS? 回答1: you can use EL expression in oozie like: <decision name="CheckFile"> <switch> <case to="nextOozieTask"> ${fs:exists('/path/test_08_01_2016.csv')} <!--do note the

How do the hive sql queries are submitted as mr job from hive cli

余生颓废 提交于 2019-12-01 21:12:36
I have deployed a CDH-5.9 cluster with MR as hive execution engine. I have a hive table named "users" with 50 rows. Whenever I execute the query select * from users works fine as follows : hive> select * from users; OK Adam 1 38 ATK093 CHEF Benjamin 2 24 ATK032 SERVANT Charles 3 45 ATK107 CASHIER Ivy 4 30 ATK384 SERVANT Linda 5 23 ATK132 ASSISTANT . . . Time taken: 0.059 seconds, Fetched: 50 row(s) But issuing select max(age) from users failed after submitting as mr job. The container log also doesn't have any information to figure it out why its getting failed. hive> select max(age) from

Applications not shown in yarn UI when running mapreduce hadoop job?

感情迁移 提交于 2019-12-01 17:18:46
I am using Hadoop2.2 . I see that my jobs are completed with success. I can browse the filesystem to find the output. However, when I browse http://NNode:8088/cluster/apps , I am unable to see any applications that have been completed so far ( I ran 3 wordcount jobs, but none of it is seen here). Are there any configurations that need to be taken into account? Here is the yarn-site.xml <property> <name>yarn.resourcemanager.hostname</name> <value>NNode</value> </property> <property> <name>yarn.nodemanager.aux-services</name> <value>mapreduce_shuffle</value> </property> <!-- <property> <name

Applications not shown in yarn UI when running mapreduce hadoop job?

喜欢而已 提交于 2019-12-01 16:48:51
问题 I am using Hadoop2.2 . I see that my jobs are completed with success. I can browse the filesystem to find the output. However, when I browse http://NNode:8088/cluster/apps , I am unable to see any applications that have been completed so far ( I ran 3 wordcount jobs, but none of it is seen here). Are there any configurations that need to be taken into account? Here is the yarn-site.xml <property> <name>yarn.resourcemanager.hostname</name> <value>NNode</value> </property> <property> <name>yarn

MapReduce in Hadoop 2.2.0 not working

与世无争的帅哥 提交于 2019-12-01 13:31:13
After installing and configuring my Hadoop 2.2.0 in pseudo-distributed mode everything is running, as you can see in the jps : $ jps 2287 JobHistoryServer 1926 ResourceManager 2162 NodeManager 1834 DataNode 1756 NameNode 3013 Jps Then I ran the wordcount example with hadoop jar $HADOOP_HOME/share/hadoop/mapreduce/hadoop-mapreduce-examples-2.2.0.jar wordcount /user/hdfs/file /output And the execution frezees (?) as follows: $ hadoop jar $HADOOP_HOME/share/hadoop/mapreduce/hadoop-mapreduce-examples-2.2.0.jar wordcount /user/hdfs/file /output OpenJDK 64-Bit Server VM warning: You have loaded

MapReduce in Hadoop 2.2.0 not working

时光毁灭记忆、已成空白 提交于 2019-12-01 11:32:03
问题 After installing and configuring my Hadoop 2.2.0 in pseudo-distributed mode everything is running, as you can see in the jps : $ jps 2287 JobHistoryServer 1926 ResourceManager 2162 NodeManager 1834 DataNode 1756 NameNode 3013 Jps Then I ran the wordcount example with hadoop jar $HADOOP_HOME/share/hadoop/mapreduce/hadoop-mapreduce-examples-2.2.0.jar wordcount /user/hdfs/file /output And the execution frezees (?) as follows: $ hadoop jar $HADOOP_HOME/share/hadoop/mapreduce/hadoop-mapreduce

Put file on HDFS with spaces in name

梦想的初衷 提交于 2019-12-01 09:35:41
I have a file named file name(1).zip (with the space and parentheses in it) and I want to put this file on the HDFS. But everytime I try to put it via hadoop fs -put ... , I get a an exception. I have even tried to add quotes around the file and even tried to escape the space and parentheses but it doesn't work. hduser@localhost:/tmp$ hadoop fs -put file\ name\(1\).zip /tmp/one 15/06/05 15:57:46 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable put: unexpected URISyntaxException hduser@localhost:/tmp$ hadoop fs

Put file on HDFS with spaces in name

廉价感情. 提交于 2019-12-01 09:26:19
问题 I have a file named file name(1).zip (with the space and parentheses in it) and I want to put this file on the HDFS. But everytime I try to put it via hadoop fs -put ... , I get a an exception. I have even tried to add quotes around the file and even tried to escape the space and parentheses but it doesn't work. hduser@localhost:/tmp$ hadoop fs -put file\ name\(1\).zip /tmp/one 15/06/05 15:57:46 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using