hbase

can HBase , MapReduce and HDFS can work on a single machine having Hadoop installed and running on it?

邮差的信 提交于 2019-12-12 10:18:01
问题 I am working on a search engine design, which is to be run on cloud. We have just started, and have not much idea about Hdoop. Can anyone tell if HBase , MapReduce and HDFS can work on a single machine having Hdoop installed and running on it ? 回答1: Yes you can. You can even create a Virtual Machine and run it on there on a single "computer" (which is what I have :) ). The key is to simply install Hadoop in "Pseudo Distributed Mode" which is even described in the Hadoop Quickstart. If you use

Latest compatible version of hadoop and hbase

时光总嘲笑我的痴心妄想 提交于 2019-12-12 09:07:22
问题 I have to install hadoop and hbase on 4 machines. I have found latest version of hadoop-2.6.0 and hbase-0.98.9-hadoop2-bin . 1)Are they compatible? 2)How I should install hbase and hadoop on 4 machines i.e. how many masters and slaves etc.? 回答1: Answering your questions : See the Hbase-Hadoop compatibility link . hbase-0.98.9 should support hadoop-2.6.0. You have two options 1 master node and 3 slave nodes. (Since you have only 4 nodes this configuration would be suffice) If you wanted to

Hbase: How to specify hostname for Hbase master

China☆狼群 提交于 2019-12-12 07:34:23
问题 I'm struggling to setup a Hbase distributed cluster with 2 nodes, one is my machine and one is the VM, using the "host-only" Adapter in VirtualBox. My problem is that the region server (from VM machine) can't connect to Hbase master running on host machine. Although in Hbase shell I can list, create table, ..., in regionserver on VM machine ('slave'), the log always show org.apache.hadoop.hbase.regionserver.HRegionServer: Unable to connect to master. Retrying. Error was: java.net

python3.5 thrift cybin.ProtocolError: No protocol version header

狂风中的少年 提交于 2019-12-12 06:48:12
问题 Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/hskj/anaconda3/lib/python3.5/site-packages/happybase/connection.py", line 242, in tables names = self.client.getTableNames() File "/hskj/anaconda3/lib/python3.5/site-packages/thriftpy/thrift.py", line 198, in _req return self._recv(_api) File "/hskj/anaconda3/lib/python3.5/site-packages/thriftpy/thrift.py", line 210, in _recv fname, mtype, rseqid = self._iprot.read_message_begin() File "thriftpy/protocol/cybin/cybin

Will the hbase kerberos token expired

喜夏-厌秋 提交于 2019-12-12 06:24:13
问题 I have a spark streaming application, for every batch, I need to insert it to the hbase which is protected by kerberos. I found a solution, that is in the driver side I create a connection and obtain a token from that conn and then pass it to the executor. In the executor side, I decode it and get the token, in this way I can insert data to hbase successfully. This seems good, but my concern is that will the token expired? If so , how to solve it please? My code snippet is val ugi

error while executing select query in hive

99封情书 提交于 2019-12-12 06:20:01
问题 I'm using hadoop 1.1.2 , hbase 0.94.8 and hive 0.14 . I'am trying to create a table in hbase using hive and load data in it later by insert overwrite . for the moment I was able to create the table: CREATE TABLE hbase_table_emp(id int, name string, role string) STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' WITH SERDEPROPERTIES ("hbase.columns.mapping" = ":key,cf1:name,cf1:role") TBLPROPERTIES ("hbase.table.name" = "emp"); and load data into another table that I will overwrite

How to load a flat file(not delimited file) into HBase?

老子叫甜甜 提交于 2019-12-12 05:29:09
问题 I am new to hbase and I have a flat file(not delimited file) that I would like to load into a single hbase table. Here is a preview of a row in my file: 0107E07201512310015071C11100747012015123100 I know fo example that from position 1 to 7 it's an id and from position 7 to 15 it's a date.... The problem is how to build a schema that correspond to my file or if there is a way to convert it to a delimited file or read such file using jaql because I'm working with Infosphere BigInsights. Any

Using HBase importtsv tool to bulk load data from Java code

与世无争的帅哥 提交于 2019-12-12 05:11:45
问题 I am trying to bulk load csv file to hbase using importtsv and LoadIncrementalHFiles tools that ship with Apache HBase. We can find the tutorials at these pages: cloudera, apache I am using Apache hadoop and hbase. Both sources explains how to use these tools through command prompt. However I want to get this done from Java code. I know I can write custom map reduce as explained on cloudera page. However I want know if I can use classes corresponding to these tools directly in my Java code.

Flink thowing serialization error when reading from hbase

非 Y 不嫁゛ 提交于 2019-12-12 05:08:58
问题 When I read from hbase using richfatMapFunction inside a map I am getting serialization error. What I am trying to do is if a datastream equals to a particular string read from hbase else ignore. Below is the sample program and error I am getting. package com.abb.Flinktest import java.text.SimpleDateFormat import java.util.Properties import scala.collection.concurrent.TrieMap import org.apache.flink.addons.hbase.TableInputFormat import org.apache.flink.api.common.functions.RichFlatMapFunction

Spark Hbase connection issue

99封情书 提交于 2019-12-12 05:06:42
问题 Hitting with followiong error while i am trying to connect the hbase through spark(using newhadoopAPIRDD ) in HDP 2.4.2.Already tried increasing the RPC time in hbase site xml file,still getting the same. any idea how to fix ? Exception in thread "main" org.apache.hadoop.hbase.client.RetriesExhaustedException: Failed after attempts=36, exceptions: Wed Nov 16 14:59:36 IST 2016, null, java.net.SocketTimeoutException: callTimeout=60000, callDuration=71216: row 'scores,,00000000000000' on table