hbase

Getting null pointer exception when running saveAsNewAPIHadoopDataset in scala spark2 to hbase

两盒软妹~` 提交于 2020-07-30 08:01:26
问题 I am saving a puts RDD to Hbase using saveAsNewAPIHadoopDataset. Below is my job creation and submition. val outputTableName = "test3" val conf2 = HBaseConfiguration.create() conf2.set("hbase.zookeeper.quorum", "xx.xx.xx.xx") conf2.set("hbase.mapred.outputtable", outputTableName) conf2.set("mapreduce.outputformat.class", "org.apache.hadoop.hbase.mapreduce.TableOutputFormat") val job = createJob(outputTableName, conf2) val outputTable = sc.broadcast(outputTableName) val hbasePuts = simpleRdd

Getting null pointer exception when running saveAsNewAPIHadoopDataset in scala spark2 to hbase

跟風遠走 提交于 2020-07-30 08:00:23
问题 I am saving a puts RDD to Hbase using saveAsNewAPIHadoopDataset. Below is my job creation and submition. val outputTableName = "test3" val conf2 = HBaseConfiguration.create() conf2.set("hbase.zookeeper.quorum", "xx.xx.xx.xx") conf2.set("hbase.mapred.outputtable", outputTableName) conf2.set("mapreduce.outputformat.class", "org.apache.hadoop.hbase.mapreduce.TableOutputFormat") val job = createJob(outputTableName, conf2) val outputTable = sc.broadcast(outputTableName) val hbasePuts = simpleRdd

HBase Shell Logging

白昼怎懂夜的黑 提交于 2020-06-27 08:22:09
问题 When using the HBase shell, I'm getting a great deal of logging, including INFO and DEBUG messages. While this is interesting in terms of learning HBase internals, it is quite verbose and can bury the output. I've tried changing the logging levels in a number of different ways, including as described here, and while some of the warnings do disappear, I continue to get a large number of INFO and DEBUG messages, i.e.: 18:50:49.500 [main] INFO org.apache.zookeeper.ZooKeeper - Client environment

HBase connection pooling for very frequent scanning of row

不羁岁月 提交于 2020-06-13 05:39:21
问题 I have to scan the table for row very frequent(~million times) in an hour. I have the information about rowid(which is a byte array). I am creating rowid for creating startrow and endrow which are essentially the same in my case. public String someMethod(byte[] rowid){ if (aTable == null) { aTable = new HTable(Config.getHadoopConfig(), Config.getATable()); } byte[] endRow = new byte[rowId.length]; endrow = System.copyArray(rowId, 0, endRow, 0, rowId.length) Scan scan = new Scan(rowId , endRow

Bulk Insert Data in HBase using Structured Spark Streaming

淺唱寂寞╮ 提交于 2020-06-09 19:01:29
问题 I'm reading data coming from a Kafka (100.000 line per second) using Structured Spark Streaming, and i'm trying to insert all the data in HBase. I'm in Cloudera Hadoop 2.6 and I'm using Spark 2.3 I tried something like I've seen here. eventhubs.writeStream .foreach(new MyHBaseWriter[Row]) .option("checkpointLocation", checkpointDir) .start() .awaitTermination() MyHBaseWriter looks like this : class AtomeHBaseWriter[RECORD] extends HBaseForeachWriter[Row] { override def toPut(record: Row): Put

Bulk Insert Data in HBase using Structured Spark Streaming

独自空忆成欢 提交于 2020-06-09 19:01:12
问题 I'm reading data coming from a Kafka (100.000 line per second) using Structured Spark Streaming, and i'm trying to insert all the data in HBase. I'm in Cloudera Hadoop 2.6 and I'm using Spark 2.3 I tried something like I've seen here. eventhubs.writeStream .foreach(new MyHBaseWriter[Row]) .option("checkpointLocation", checkpointDir) .start() .awaitTermination() MyHBaseWriter looks like this : class AtomeHBaseWriter[RECORD] extends HBaseForeachWriter[Row] { override def toPut(record: Row): Put

Why am I getting negative allocated mappers in Tez job? Vertex failure?

流过昼夜 提交于 2020-05-16 05:13:10
问题 I'm trying to use the PhoenixStorageHandler as documented here, and populate it with the following query in beeline shell: insert into table pheonix_table select * from hive_table; I get the following breakdown of the mappers in the Tez session: ... INFO : Map 1: 0(+50)/50 INFO : Map 1: 0(+50)/50 INFO : Map 1: 0(+50,-2)/50 INFO : Map 1: 0(+50,-3)/50 ... before the session crashes with a very long error message (422 lines) about vertex failure: Error: Error while processing statement: FAILED:

Hbase Error “ERROR: KeeperErrorCode = NoNode for /hbase/master”

别等时光非礼了梦想. 提交于 2020-05-09 06:49:47
问题 While executing any command in hbase shell, I am receiving the following error "ERROR: KeeperErrorCode = NoNode for /hbase/master" in hbase shell. Started HBASE : HOSTCHND:hbase-2.0.0 gvm$ ./bin/start-hbase.sh localhost: running zookeeper, logging to /usr/local/Cellar/hbase-2.0.0/bin/../logs/hbase-gvm-zookeeper-HOSTCHND.local.out running master, logging to /usr/local/Cellar/hbase-2.0.0/logs/hbase-gvm-master-HOSTCHND.local.out : running regionserver, logging to /usr/local/Cellar/hbase-2.0.0

FAILED: Execution Error, return code -101 from org.apache.hadoop.hive.ql.exec.mr.MapRedTask. com/yammer/metrics/core/MetricsRegistry

妖精的绣舞 提交于 2020-04-17 22:12:08
问题 We facing some issue in beeline while we connecting via beeline to hbase table. We have two hiveserver2, one of the node we got this error like: INFO : Query ID = hive_20190719154444_babd2ce5-4d41-400b-9be5-313acaffc9bf INFO : Total jobs = 1 INFO : Launching Job 1 out of 1 INFO : Starting task [Stage-0:MAPRED] in serial mode INFO : Number of reduce tasks is set to 0 since there's no reduce operator ERROR : FAILED: Execution Error, return code -101 from org.apache.hadoop.hive.ql.exec.mr

FAILED: Execution Error, return code -101 from org.apache.hadoop.hive.ql.exec.mr.MapRedTask. com/yammer/metrics/core/MetricsRegistry

随声附和 提交于 2020-04-17 22:11:50
问题 We facing some issue in beeline while we connecting via beeline to hbase table. We have two hiveserver2, one of the node we got this error like: INFO : Query ID = hive_20190719154444_babd2ce5-4d41-400b-9be5-313acaffc9bf INFO : Total jobs = 1 INFO : Launching Job 1 out of 1 INFO : Starting task [Stage-0:MAPRED] in serial mode INFO : Number of reduce tasks is set to 0 since there's no reduce operator ERROR : FAILED: Execution Error, return code -101 from org.apache.hadoop.hive.ql.exec.mr