hbase

unable to load data in hbase table from hive

一世执手 提交于 2019-12-13 07:16:54
问题 I am using hadoop version 2.7.0, hive version 1.1.0, HBase version hbase-0.98.14-hadoop2. I have created a hbase table from hive successfully. hive (Koushik)> CREATE TABLE hive_hbase_emp_test(eid int, ename string, esal double) > STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' > WITH SERDEPROPERTIES > ("hbase.columns.mapping" = ":key,cfstr:enm,cfsal:esl") > TBLPROPERTIES ("hbase.table.name" = "hive_hbase_emp_test"); OK Time taken: 0.874 seconds hbase(main):004:0> describe 'hive

How insert overwrite table in hive with diffrent where clauses?

半腔热情 提交于 2019-12-13 06:36:50
问题 I want to read a .tsv file from Hbase into hive. The file has a columnfamily, which has 3 columns inside: news, social and all. The aim is to store these columns in an table in hbase which has the columns news, social and all. CREATE EXTERNAL TABLE IF NOT EXISTS topwords_logs (key String, columnfamily String , wort String , col String, occurance int)ROW FORMAT DELIMITED FIELDS TERMINATED BY '\t'STORED AS TEXTFILE LOCATION '/home /hfu/Testdaten'; load data local inpath '/home/hfu/Testdaten

HBase shell shows error when listing tables

ぐ巨炮叔叔 提交于 2019-12-13 05:42:43
问题 hbase(main):001:0> list ... ava:75) at org.jruby.internal.runtime.methods.InterpretedMethod.call(InterpretedMethod.java:147) at org.jruby.internal.runtime.methods.DefaultMethod.call(DefaultMethod.java:163) at org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:292) at org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:135) at usr.local.hbase.bin.$_dot_dot_.bin.hirb.block_2$RUBY$start(/usr/local/hbase/bin/../bin/hirb.rb:183) at usr$local$hbase$bin$$

No FileSystem for scheme: hdfs

跟風遠走 提交于 2019-12-13 05:25:05
问题 i have the following error when i'm running my Storm Topology that contains an Hbase Bolt. java.io.IOException: No FileSystem for scheme: hdfs at org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSystem.java:2298) ~[hadoop-common-2.0.0-cdh4.7.0.jar:na] at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2305) ~[hadoop-common-2.0.0-cdh4.7.0.jar:na] at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:89) ~[hadoop-common-2.0.0-cdh4.7.0.jar:na] at org.apache.hadoop

MySQL and HBase Connectivity

早过忘川 提交于 2019-12-13 05:19:46
问题 I am new to Hadoop. I want to get data from mysql to Hbase continuously(Continuous Integration) and vice versa as well. Is there any way to do this type of stuff ?? 回答1: You should look at Apache Sqoop: a tool designed for transferring bulk data between Hadoop and datastores such as relational databases. 来源: https://stackoverflow.com/questions/33436794/mysql-and-hbase-connectivity

How to store and retrieve video files and image files in hbase table

流过昼夜 提交于 2019-12-13 05:01:13
问题 I have searched in documentation ,but there is no information about storing video file in hbase.I want implement in my project 回答1: Please dont use HBase to store 1GB of video file. Thats not a good use case for HBase. If your file is bigger than few(0-10) MB's then dont store it in HBase. 来源: https://stackoverflow.com/questions/31399843/how-to-store-and-retrieve-video-files-and-image-files-in-hbase-table

Error while inserting from Hive to Hbase

本小妞迷上赌 提交于 2019-12-13 04:54:41
问题 I am using CDH 4.7.1 cluster. Map seems completed 100% and failing the reduce part. I have added the below part to hive-site.xml. Actual error message is pasted in the last part of this post. Thanks. Any help is appreciated. <property> <name>hive.aux.jars.path</name> <value> file:///opt/cloudera/parcels/CDH/lib/hbase/hbase.jar, file:///opt/cloudera/parcels/CDH-4.7.1-1.cdh4.7.1.p0.47/lib/hive/lib/hive-hbase-handler-0.10.0-cdh4.7.1.jar, file:///opt/cloudera/parcels/CDH-4.7.1-1.cdh4.7.1.p0.47

HBase get returns old values even with max versions = 1

被刻印的时光 ゝ 提交于 2019-12-13 04:37:33
问题 I have the desire to find the columns that have not been updated for more than a specific time period. So I want to do a scan against the columns with a timerange. The normal behaviour of HBase is that you then get the latest value in that time range (which is not what I want). As far as I understand the way HBase should work is that if you set the maximum number of versions for the values in a column family to '1' it should retain only the last value that was put into the cell. What I found

HBase :HBase高级shell管理命令

≡放荡痞女 提交于 2019-12-13 04:37:13
Shell命令 status: 例如:显示服务器状态 hbase(main):058:0> status ‘node01’ whoami 显示HBase当前用户,例如: hbase> whoami list 显示当前所有的表 count 统计指定表的记录数,例如: hbase> count ‘user’ describe 展示表结构信息 exists 检查表是否存在,适用于表量特别多的情况 is_enabled、is_disabled 检查表是否启用或禁用 alter 该命令可以改变表和列族的模式,例如: 为当前表增加列族: hbase> alter ‘user’, NAME => ‘CF2’, VERSIONS => 2 为当前表删除列族: hbase(main):002:0> alter ‘user’, ‘delete’ => ‘CF2’ disable/enable 禁用一张表/启用一张表 drop 删除一张表,记得在删除表之前必须先禁用 truncate 禁用表-删除表-创建表 来源: CSDN 作者: 真情流露哦呦 链接: https://blog.csdn.net/weixin_43563705/article/details/103510273

Get an “IOException: Broken pipe” during submiting a spark job which is connecting hbase by pyspark code

感情迁移 提交于 2019-12-13 04:34:16
问题 I submit a spark job to do some easy stuff by pyspark newAPIHadoopRDD, which will connecting hbase during the job running. Our CHD enable the kerberos, But I think I have pass the authentication. I will show my code, shell, exception and some CM config. > "19/01/16 10:55:42 INFO client.ConnectionManager$HConnectionImplementation: Closing zookeeper sessionid=0x36850456cea05e5 19/01/16 10:55:42 INFO zookeeper.ZooKeeper: Session: 0x36850456cea05e5 closed Traceback (most recent call last): File "