hbase

Using Hbase with C#

末鹿安然 提交于 2019-12-11 02:09:35
问题 How can I use a Hbase database with C#/VB.NET ? (use=connect, query, get the result, insert, update, delete) I don't find useful answers with google. 回答1: I've just published HBase C# Thrift binding as nuget package. Alternatively you can grab code/binaries from bitbucket: https://bitbucket.org/vadim/hbase-sharp/downloads 回答2: From the description: A REST-ful Web service gateway that supports XML, Protobuf, and binary data encoding options There's a protobuf port for .NET and there are many

How to rename a column family in HBase using HBase Shell

ⅰ亾dé卋堺 提交于 2019-12-11 01:24:45
问题 In the HBase shell, it's possible to delete a column family using: alter 'tablename', NAME => 'familyname', METHOD => 'delete' It's also possible to create a new column family using: alter 'tablename', NAME => 'newfamilyname' So it's possible to delete a column family and create a new one, but if I have existing data, deleting the column family deletes the data within that family as well. How do I rename an HBase column family in an existing table without losing the existing data? 来源: https:/

HBase with Kerberos - keep a HTable instance open more than 10 hours

巧了我就是萌 提交于 2019-12-11 00:58:52
问题 Here is the situation: we have a secured (Kerberos) HBase cluster. I have an object that creates an instance of HTable at startup and hang on to it. It calls: UserGroupInformation.setConfiguration(configuration); UserGroupInformation.loginUserFromKeytab(user, keytab); to login to the Kerberized cluster. This object then hangs around unused for many hours. After more than 10 hours (the timeout on a ticket from our Kerberos cluster), the next call to scan the table results in this: 16/12/01 18

HBase: /hbase/meta-region-server node does not exist

旧巷老猫 提交于 2019-12-11 00:38:23
问题 I have installed cloudera and hdfs, mapreduce, zookeper, hbase on it. 4 nodes with these services (3 zookeeper). All are installed by cloudera wizard and have no configuration issues in cloudera. On connect from java I have got an error: 9:32:23.020 [main-SendThread()] INFO org.apache.zookeeper.ClientCnxn - Opening socket connection to server /172.20.7.6:2181 09:32:23.020 [main] INFO org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper - Process identifier=hconnection-0x301abf87 connecting

With Hadoop, can I create a tasktracker on a machine that isn't running a datanode?

依然范特西╮ 提交于 2019-12-10 23:22:16
问题 So here's my situation : I have a mapreduce job that uses HBase. My mapper takes one line of text input and updates HBase. I have no reducer, and I'm not writing any output to the disc. I would like the ability to add more processing power to my cluster when I'm expecting a burst of utilization, and then scale back down when utilization decreases. Let's assume for the moment that I can't use Amazon or any other cloud provider; I'm running in a private cluster. One solution would be to add new

Issue on Spark Streaming data put data into HBase

岁酱吖の 提交于 2019-12-10 23:13:17
问题 I am a beginner in this field, so I can not get a sense of it... HBase ver: 0.98.24-hadoop2 Spark ver: 2.1.0 The following code try to put receiving data from Spark Streming-Kafka producer into HBase. Kafka input data format is like this : Line1,TAG1,123 Line1,TAG2,134 Spark-streaming process split the receiving line by delimiter ',' then put the data into HBase. However, my application met an error when it call the htable.put() method. Can any one help why the below code is throwing error?

How To Create Partitions In Hbase Table As Like Hive Table Partitions

天大地大妈咪最大 提交于 2019-12-10 21:37:36
问题 We are planning to migrate from CDH3 to CDH4, as part of this migration we also planning to bring HBASE into out system because it also updates to the data, in CDH3 we are using Hive as warehouse. Here we are having the major problem in migration, Hive supports partitions to tables. And our system has many tables in different schemas and some tables has partitions base on date, we have the history of data from last 5 years (365 * 5 partitions exists in some tables). We want to achieve the

How to renew expiring Kerberos ticket in HBase?

前提是你 提交于 2019-12-10 21:16:08
问题 I have a small spring service, which provides basic functionality like put/delete/get from hbase table. Everything seems to work, but there is one problem. After 10 hours after starting my tomcat server, my kerberos ticket expires, so I should renew it. I tried to use java api for hbase and in my code in every method, which connects to hbase I added this line: UserGroupInformation.getLoginUser().checkTGTAndReloginFromKeytab(); I tried also with: UserGroupInformation.getLoginUser()

NodeJS Hbase thrift weirdness

 ̄綄美尐妖づ 提交于 2019-12-10 20:11:47
问题 I'm running into a very odd situation. I've been trying to get a nodeJS client, using thrift, for Hbase running, and I have seemingly had success most of the way except getting any return data. I'm running hbase.0.94.8 (currently the stable version), and it is definitely running, I'm running thrift 0.9.0, it is built and running as well. With both of them running I'm able to query Hbase with the shell and also get to the thrift webpage, but when I run the following code, nothing happens: var

I want to scan lots of data (Range based queries), what all optimizations I can do while writing the data so that scan becomes faster?

▼魔方 西西 提交于 2019-12-10 19:44:52
问题 I have billion of rows in hbase I want to scan million rows at a time. what are the best optimization techniques which I can do to make this scan as fast as possible. 回答1: We have similar problem, we need to scan million rows by keys and we used the map reduce techniques for this. There is no standard solution for this, so we write a custom input format that extends InputFormat<ImmutableBytesWritable, Result> . There is a shot description how we done this. First you need to create a splits so