Have you changed your hbase config recently? If so, please make sure that
you've copied the latest config to Impala.
Thanks,
Alan
On Tue, Mar 11, 2014 at 4:46 PM, Marcel Kornacker wrote:
Please provide the impalad and catalogd log files.
org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.prefetchRegionCache(HConnectionManager.java:1141)
org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.locateRegionInMeta(HConnectionManager.java:1201)
org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.locateRegion(HConnectionManager.java:1093)
org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.locateRegion(HConnectionManager.java:1050)
com.cloudera.impala.catalog.CatalogObjectCache$1.call(CatalogObjectCache.java:229)
com.cloudera.impala.catalog.CatalogObjectCache$1.call(CatalogObjectCache.java:226)
com.google.common.cache.LocalCache$LocalManualCache$1.load(LocalCache.java:4767)
com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3568)
com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2350)
com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2313)
com.google.common.cache.LocalCache$LocalManualCache.get(LocalCache.java:4764)
com.cloudera.impala.catalog.CatalogObjectCache.get(CatalogObjectCache.java:226)
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
org.apache.hadoop.hbase.client.MetaScanner.metaScan(MetaScanner.java:152)
org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.prefetchRegionCache(HConnectionManager.java:1141)
org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.locateRegionInMeta(HConnectionManager.java:1201)
org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.locateRegion(HConnectionManager.java:1093)
org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.locateRegion(HConnectionManager.java:1050)
com.cloudera.impala.catalog.CatalogObjectCache$1.call(CatalogObjectCache.java:229)
com.cloudera.impala.catalog.CatalogObjectCache$1.call(CatalogObjectCache.java:226)
com.google.common.cache.LocalCache$LocalManualCache$1.load(LocalCache.java:4767)
com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3568)
com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2350)
com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2313)
com.google.common.cache.LocalCache$LocalManualCache.get(LocalCache.java:4764)
com.cloudera.impala.catalog.CatalogObjectCache.get(CatalogObjectCache.java:226)
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
STARTKEY
value=1394470128013
value=1394470127965
=> '',
value=1394470120098
STARTKEY
value=1394470127884
STARTKEY
value=1394470128046
value=1394470127865
might be
error and
any other
HBase
verify
discussion
http://www.cloudera.com/content/cloudera-content/cloudera-docs/Impala/latest/Installing-and-Using-Impala/ciiu_impala_hbase.html
:
'1',
'true'},
'NONE',
'ints',
'false',
'NONE',
COMPRESSION
KEEP_DELETED_CELLS =>
'true'}
lastAccessTime:0,
type:int,
comment:null),
outputFormat:org.apache.hadoop.hive.ql.io.HivePassThroughOutputFormat,
hbase.columns.mapping=:key,bools:bool_col,ints:tinyint_col,ints:smallint_col,ints:int_col,ints:bigint_col,floats:float_col,floats:double_col,strings:date_string_col,strings:string_col,strings:timestamp_col}),
transient_lastDdlTime=1394217917,
tableType:EXTERNAL_TABLE)
when
statement
send
To unsubscribe from this group and stop receiving emails from it, send
an email to impala-user...@cloudera.org.To unsubscribe from this group and stop receiving emails from it, send an
email to impala-user+unsubscribe@cloudera.org.To unsubscribe from this group and stop receiving emails from it, send an
email to impala-user+unsubscribe@cloudera.org.
To unsubscribe from this group and stop receiving emails from it, send an email to impala-user+unsubscribe@cloudera.org.Please provide the impalad and catalogd log files.
On Tue, Mar 11, 2014 at 4:45 PM, Percy Zhao wrote:
Any update on this? Also I can see from netstat that catalogd is
connecting
to the right region server.
thanks,
percy
org.apache.hadoop.hbase.client.MetaScanner.metaScan(MetaScanner.java:152)Any update on this? Also I can see from netstat that catalogd is
connecting
to the right region server.
thanks,
percy
On Monday, March 10, 2014 11:31:57 AM UTC-7, Percy Zhao wrote:
I restarted impalad and catalogd, then catalogd log file keep throwing
error message like this over and over again
W0311 02:14:39.846324 10258 HConnectionManager.java:1144] Encountered
problems when prefetch hbase:meta table:
Java exception follows:
java.io.IOException: HRegionInfo was null or empty in Meta for ssp,
row=ssp,,99999999999999
at
I restarted impalad and catalogd, then catalogd log file keep throwing
error message like this over and over again
W0311 02:14:39.846324 10258 HConnectionManager.java:1144] Encountered
problems when prefetch hbase:meta table:
Java exception follows:
java.io.IOException: HRegionInfo was null or empty in Meta for ssp,
row=ssp,,99999999999999
at
at
at
at
at
at
org.apache.hadoop.hbase.client.HTable.finishSetup(HTable.java:330)
at org.apache.hadoop.hbase.client.HTable.<init>(HTable.java:314)
at org.apache.hadoop.hbase.client.HTable.<init>(HTable.java:186)
at org.apache.hadoop.hbase.client.HTable.<init>(HTable.java:154)
at
com.cloudera.impala.catalog.HBaseTable.load(HBaseTable.java:235)
at com.cloudera.impala.catalog.Table.load(Table.java:176)
at com.cloudera.impala.catalog.Db.loadTable(Db.java:95)
at com.cloudera.impala.catalog.Db.access$000(Db.java:52)
at com.cloudera.impala.catalog.Db$1.load(Db.java:69)
at com.cloudera.impala.catalog.Db$1.load(Db.java:65)
at
org.apache.hadoop.hbase.client.HTable.finishSetup(HTable.java:330)
at org.apache.hadoop.hbase.client.HTable.<init>(HTable.java:314)
at org.apache.hadoop.hbase.client.HTable.<init>(HTable.java:186)
at org.apache.hadoop.hbase.client.HTable.<init>(HTable.java:154)
at
com.cloudera.impala.catalog.HBaseTable.load(HBaseTable.java:235)
at com.cloudera.impala.catalog.Table.load(Table.java:176)
at com.cloudera.impala.catalog.Db.loadTable(Db.java:95)
at com.cloudera.impala.catalog.Db.access$000(Db.java:52)
at com.cloudera.impala.catalog.Db$1.load(Db.java:69)
at com.cloudera.impala.catalog.Db$1.load(Db.java:65)
at
at
at
at
at
at
at
com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2228)
at com.google.common.cache.LocalCache.get(LocalCache.java:3965)
at
com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2228)
at com.google.common.cache.LocalCache.get(LocalCache.java:3965)
at
at
at com.cloudera.impala.catalog.Db.getTable(Db.java:189)
at com.cloudera.impala.catalog.Catalog$2.run(Catalog.java:513)
at
at com.cloudera.impala.catalog.Catalog$2.run(Catalog.java:513)
at
at
at java.lang.Thread.run(Thread.java:744)
W0311 02:14:39.854532 10261 HConnectionManager.java:1144] Encountered
problems when prefetch hbase:meta table:
Java exception follows:
java.io.IOException: HRegionInfo was null or empty in Meta for
hbasealltypessmall, row=hbasealltypessmall,,99999999999999
at
W0311 02:14:39.854532 10261 HConnectionManager.java:1144] Encountered
problems when prefetch hbase:meta table:
Java exception follows:
java.io.IOException: HRegionInfo was null or empty in Meta for
hbasealltypessmall, row=hbasealltypessmall,,99999999999999
at
at
at
at
at
at
org.apache.hadoop.hbase.client.HTable.finishSetup(HTable.java:330)
at org.apache.hadoop.hbase.client.HTable.<init>(HTable.java:314)
at org.apache.hadoop.hbase.client.HTable.<init>(HTable.java:186)
at org.apache.hadoop.hbase.client.HTable.<init>(HTable.java:154)
at
com.cloudera.impala.catalog.HBaseTable.load(HBaseTable.java:235)
at com.cloudera.impala.catalog.Table.load(Table.java:176)
at com.cloudera.impala.catalog.Db.loadTable(Db.java:95)
at com.cloudera.impala.catalog.Db.access$000(Db.java:52)
at com.cloudera.impala.catalog.Db$1.load(Db.java:69)
at com.cloudera.impala.catalog.Db$1.load(Db.java:65)
at
org.apache.hadoop.hbase.client.HTable.finishSetup(HTable.java:330)
at org.apache.hadoop.hbase.client.HTable.<init>(HTable.java:314)
at org.apache.hadoop.hbase.client.HTable.<init>(HTable.java:186)
at org.apache.hadoop.hbase.client.HTable.<init>(HTable.java:154)
at
com.cloudera.impala.catalog.HBaseTable.load(HBaseTable.java:235)
at com.cloudera.impala.catalog.Table.load(Table.java:176)
at com.cloudera.impala.catalog.Db.loadTable(Db.java:95)
at com.cloudera.impala.catalog.Db.access$000(Db.java:52)
at com.cloudera.impala.catalog.Db$1.load(Db.java:69)
at com.cloudera.impala.catalog.Db$1.load(Db.java:65)
at
at
at
at
at
at
at
com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2228)
at com.google.common.cache.LocalCache.get(LocalCache.java:3965)
at
com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2228)
at com.google.common.cache.LocalCache.get(LocalCache.java:3965)
at
at
at com.cloudera.impala.catalog.Db.getTable(Db.java:189)
at com.cloudera.impala.catalog.Catalog$2.run(Catalog.java:513)
at
at com.cloudera.impala.catalog.Catalog$2.run(Catalog.java:513)
at
at
at java.lang.Thread.run(Thread.java:744)
and here's the table scan of hbase meta table:
hbase(main):009:0* scan 'hbase:meta'
ROW
COLUMN+CELL
hbase:namespace,,1387263658257.16638622f2b9f25a825aab6867baaab5.
column=info:regioninfo, timestamp=1387263658870, value={ENCODED =>
16638622f2b9f25a825aab6867baaab5, NAME =>
'hbase:namespace,,1387263658257.16638622f2b9f25a825aab6867baaab5.',
and here's the table scan of hbase meta table:
hbase(main):009:0* scan 'hbase:meta'
ROW
COLUMN+CELL
hbase:namespace,,1387263658257.16638622f2b9f25a825aab6867baaab5.
column=info:regioninfo, timestamp=1387263658870, value={ENCODED =>
16638622f2b9f25a825aab6867baaab5, NAME =>
'hbase:namespace,,1387263658257.16638622f2b9f25a825aab6867baaab5.',
=> '
',
ENDKEY => ''}
hbase:namespace,,1387263658257.16638622f2b9f25a825aab6867baaab5.
column=info:seqnumDuringOpen, timestamp=1394470137408,
value=\x00\x00\x00\x00\x00\x00\x00\x05
hbase:namespace,,1387263658257.16638622f2b9f25a825aab6867baaab5.
column=info:server, timestamp=1394470137408,
value=dumbo010-cdh5.dumbo.allyes.com:60020
hbase:namespace,,1387263658257.16638622f2b9f25a825aab6867baaab5.
column=info:serverstartcode, timestamp=1394470137408,
',
ENDKEY => ''}
hbase:namespace,,1387263658257.16638622f2b9f25a825aab6867baaab5.
column=info:seqnumDuringOpen, timestamp=1394470137408,
value=\x00\x00\x00\x00\x00\x00\x00\x05
hbase:namespace,,1387263658257.16638622f2b9f25a825aab6867baaab5.
column=info:server, timestamp=1394470137408,
value=dumbo010-cdh5.dumbo.allyes.com:60020
hbase:namespace,,1387263658257.16638622f2b9f25a825aab6867baaab5.
column=info:serverstartcode, timestamp=1394470137408,
hbasealltypessmall,,1394217597873.badd2977a76007d2b086a4ebc6cd85
column=info:regioninfo, timestamp=1394217598196, value={ENCODED =>
badd2977a76007d2b086a4ebc6cd858a, NAME =>
'hbasealltypessmall,,1394217597873.badd2977a76007d2b086a4ebc6cd858a.',
STARTKEY =
8a. > '',
ENDKEY => ''}
hbasealltypessmall,,1394217597873.badd2977a76007d2b086a4ebc6cd85
column=info:seqnumDuringOpen, timestamp=1394470137309,
value=\x00\x00\x00\x00\x00\x00\x00\x01
8a.
hbasealltypessmall,,1394217597873.badd2977a76007d2b086a4ebc6cd85
column=info:server, timestamp=1394470137309,
value=dumbo003-cdh5.dumbo.allyes.com:60020
8a.
hbasealltypessmall,,1394217597873.badd2977a76007d2b086a4ebc6cd85
column=info:serverstartcode, timestamp=1394470137309,
column=info:regioninfo, timestamp=1394217598196, value={ENCODED =>
badd2977a76007d2b086a4ebc6cd858a, NAME =>
'hbasealltypessmall,,1394217597873.badd2977a76007d2b086a4ebc6cd858a.',
STARTKEY =
8a. > '',
ENDKEY => ''}
hbasealltypessmall,,1394217597873.badd2977a76007d2b086a4ebc6cd85
column=info:seqnumDuringOpen, timestamp=1394470137309,
value=\x00\x00\x00\x00\x00\x00\x00\x01
8a.
hbasealltypessmall,,1394217597873.badd2977a76007d2b086a4ebc6cd85
column=info:server, timestamp=1394470137309,
value=dumbo003-cdh5.dumbo.allyes.com:60020
8a.
hbasealltypessmall,,1394217597873.badd2977a76007d2b086a4ebc6cd85
column=info:serverstartcode, timestamp=1394470137309,
8a.
job_info,,1392968629874.25d3984c294f7cacb9d43b209d71e693.
column=info:regioninfo, timestamp=1392968630664, value={ENCODED =>
25d3984c294f7cacb9d43b209d71e693, NAME =>
'job_info,,1392968629874.25d3984c294f7cacb9d43b209d71e693.', STARTKEY
job_info,,1392968629874.25d3984c294f7cacb9d43b209d71e693.
column=info:regioninfo, timestamp=1392968630664, value={ENCODED =>
25d3984c294f7cacb9d43b209d71e693, NAME =>
'job_info,,1392968629874.25d3984c294f7cacb9d43b209d71e693.', STARTKEY
ENDK
EY =>
''}
job_info,,1392968629874.25d3984c294f7cacb9d43b209d71e693.
column=info:seqnumDuringOpen, timestamp=1394470438617,
value=\x00\x00\x00\x00\x00\x00\x00;
job_info,,1392968629874.25d3984c294f7cacb9d43b209d71e693.
column=info:server, timestamp=1394470438617,
value=dumbo006-cdh5.dumbo.allyes.com:60020
job_info,,1392968629874.25d3984c294f7cacb9d43b209d71e693.
column=info:serverstartcode, timestamp=1394470438617,
EY =>
''}
job_info,,1392968629874.25d3984c294f7cacb9d43b209d71e693.
column=info:seqnumDuringOpen, timestamp=1394470438617,
value=\x00\x00\x00\x00\x00\x00\x00;
job_info,,1392968629874.25d3984c294f7cacb9d43b209d71e693.
column=info:server, timestamp=1394470438617,
value=dumbo006-cdh5.dumbo.allyes.com:60020
job_info,,1392968629874.25d3984c294f7cacb9d43b209d71e693.
column=info:serverstartcode, timestamp=1394470438617,
job_info_record,,1392968665194.770fde6b862cf1adb82d4da2ed2ba37e.
column=info:regioninfo, timestamp=1392968665575, value={ENCODED =>
770fde6b862cf1adb82d4da2ed2ba37e, NAME =>
'job_info_record,,1392968665194.770fde6b862cf1adb82d4da2ed2ba37e.',
column=info:regioninfo, timestamp=1392968665575, value={ENCODED =>
770fde6b862cf1adb82d4da2ed2ba37e, NAME =>
'job_info_record,,1392968665194.770fde6b862cf1adb82d4da2ed2ba37e.',
=> '
',
ENDKEY => ''}
job_info_record,,1392968665194.770fde6b862cf1adb82d4da2ed2ba37e.
column=info:seqnumDuringOpen, timestamp=1394470137269,
value=\x00\x00\x00\x00\x00\x00\x00\x01
job_info_record,,1392968665194.770fde6b862cf1adb82d4da2ed2ba37e.
column=info:server, timestamp=1394470137269,
value=dumbo009-cdh5.dumbo.allyes.com:60020
job_info_record,,1392968665194.770fde6b862cf1adb82d4da2ed2ba37e.
column=info:serverstartcode, timestamp=1394470137269,
',
ENDKEY => ''}
job_info_record,,1392968665194.770fde6b862cf1adb82d4da2ed2ba37e.
column=info:seqnumDuringOpen, timestamp=1394470137269,
value=\x00\x00\x00\x00\x00\x00\x00\x01
job_info_record,,1392968665194.770fde6b862cf1adb82d4da2ed2ba37e.
column=info:server, timestamp=1394470137269,
value=dumbo009-cdh5.dumbo.allyes.com:60020
job_info_record,,1392968665194.770fde6b862cf1adb82d4da2ed2ba37e.
column=info:serverstartcode, timestamp=1394470137269,
job_info_report,,1392968686718.33ba7ad3617955c8afa64cc48b1ace35.
column=info:regioninfo, timestamp=1392968687098, value={ENCODED =>
33ba7ad3617955c8afa64cc48b1ace35, NAME =>
'job_info_report,,1392968686718.33ba7ad3617955c8afa64cc48b1ace35.',
column=info:regioninfo, timestamp=1392968687098, value={ENCODED =>
33ba7ad3617955c8afa64cc48b1ace35, NAME =>
'job_info_report,,1392968686718.33ba7ad3617955c8afa64cc48b1ace35.',
=> '
',
ENDKEY => ''}
job_info_report,,1392968686718.33ba7ad3617955c8afa64cc48b1ace35.
column=info:seqnumDuringOpen, timestamp=1394470137295,
value=\x00\x00\x00\x00\x00\x00\x00\x01
job_info_report,,1392968686718.33ba7ad3617955c8afa64cc48b1ace35.
column=info:server, timestamp=1394470137295,
value=dumbo008-cdh5.dumbo.allyes.com:60020
job_info_report,,1392968686718.33ba7ad3617955c8afa64cc48b1ace35.
column=info:serverstartcode, timestamp=1394470137295,
',
ENDKEY => ''}
job_info_report,,1392968686718.33ba7ad3617955c8afa64cc48b1ace35.
column=info:seqnumDuringOpen, timestamp=1394470137295,
value=\x00\x00\x00\x00\x00\x00\x00\x01
job_info_report,,1392968686718.33ba7ad3617955c8afa64cc48b1ace35.
column=info:server, timestamp=1394470137295,
value=dumbo008-cdh5.dumbo.allyes.com:60020
job_info_report,,1392968686718.33ba7ad3617955c8afa64cc48b1ace35.
column=info:serverstartcode, timestamp=1394470137295,
ssp,,1394422999709.3b4a0a51b92d219ab3e42be693f1d001.
column=info:regioninfo, timestamp=1394422999998, value={ENCODED =>
3b4a0a51b92d219ab3e42be693f1d001, NAME =>
'ssp,,1394422999709.3b4a0a51b92d219ab3e42be693f1d001.', STARTKEY => '',
ENDKEY =>
''}
ssp,,1394422999709.3b4a0a51b92d219ab3e42be693f1d001.
column=info:seqnumDuringOpen, timestamp=1394470137489,
value=\x00\x00\x00\x00\x00\x00\x00E
ssp,,1394422999709.3b4a0a51b92d219ab3e42be693f1d001.
column=info:server, timestamp=1394470137489,
value=dumbo005-cdh5.dumbo.allyes.com:60020
ssp,,1394422999709.3b4a0a51b92d219ab3e42be693f1d001.
column=info:serverstartcode, timestamp=1394470137489,
column=info:regioninfo, timestamp=1394422999998, value={ENCODED =>
3b4a0a51b92d219ab3e42be693f1d001, NAME =>
'ssp,,1394422999709.3b4a0a51b92d219ab3e42be693f1d001.', STARTKEY => '',
ENDKEY =>
''}
ssp,,1394422999709.3b4a0a51b92d219ab3e42be693f1d001.
column=info:seqnumDuringOpen, timestamp=1394470137489,
value=\x00\x00\x00\x00\x00\x00\x00E
ssp,,1394422999709.3b4a0a51b92d219ab3e42be693f1d001.
column=info:server, timestamp=1394470137489,
value=dumbo005-cdh5.dumbo.allyes.com:60020
ssp,,1394422999709.3b4a0a51b92d219ab3e42be693f1d001.
column=info:serverstartcode, timestamp=1394470137489,
6 row(s) in 0.1070 seconds
I also read somewhere else on a similar error message, and it was
suggesting to create the hbase table with region splits. Hope this
I also read somewhere else on a similar error message, and it was
suggesting to create the hbase table with region splits. Hope this
helpful.
Thanks again for your help!
Percy
Thanks again for your help!
Percy
On Monday, March 10, 2014 10:09:04 AM UTC-7, lskuff wrote:
Please provide the impalad and catalogd log files.
Thanks,
Lenni
Please provide the impalad and catalogd log files.
Thanks,
Lenni
On Mon, Mar 10, 2014 at 10:00 AM, Percy Zhao wrote:
Hi Lenni,
Thanks much for the help. Yes I'm able to query this hbase table from
hive. I checked hbase logs, there are a few zk connection related
Hi Lenni,
Thanks much for the help. Yes I'm able to query this hbase table from
hive. I checked hbase logs, there are a few zk connection related
I assume hbase recovered from them. Please let me know if there is
information I can provide to help you troubleshoot.
Thanks,
Percy
Thanks,
Percy
On Monday, March 10, 2014 9:12:31 AM UTC-7, lskuff wrote:
Hi Percy,
This looks like it might be due an HBase issue. Can you check the
Hi Percy,
This looks like it might be due an HBase issue. Can you check the
logs to verify HBase is running properly? It also would be good to
whether you are able to query the table (not just describe) via Hive.
If you see errors in your HBase log you may want to move the
If you see errors in your HBase log you may want to move the
to cdh-...@cloudera.com.
Thanks,
Lenni
Thanks,
Lenni
On Fri, Mar 7, 2014 at 2:16 PM, Percy Zhao wrote:
Environment:
Hadoop CDH5 Beta 2
Hive: 0.12.0-cdh5.0.0-beta-2
HBase: 0.96.1.1-cdh5.0.0-beta-2
Impala : v1.2.3-cdh5.0.0-beta-2 (8e266e0)
Follow the instruction on
Environment:
Hadoop CDH5 Beta 2
Hive: 0.12.0-cdh5.0.0-beta-2
HBase: 0.96.1.1-cdh5.0.0-beta-2
Impala : v1.2.3-cdh5.0.0-beta-2 (8e266e0)
Follow the instruction on
:
1. create hbase table successfully
hbase(main):003:0> describe 'hbasealltypessmall'
'hbasealltypessmall', {NAME => 'bools', DATA_BLOCK_ENCODING =>
'NONE', BLOOMFILTER => 'ROW', REPLICATION_SCOPE => '0', VERSIONS =>
hbase(main):003:0> describe 'hbasealltypessmall'
'hbasealltypessmall', {NAME => 'bools', DATA_BLOCK_ENCODING =>
'NONE', BLOOMFILTER => 'ROW', REPLICATION_SCOPE => '0', VERSIONS =>
COMPRESSION => 'NONE',
MIN_VERSIONS => '0', TTL => '2147483647', KEEP_DELETED_CELLS =>
'false', BLOCKSIZE => '65536', IN_MEMORY => 'false', BLOCKCACHE =>
MIN_VERSIONS => '0', TTL => '2147483647', KEEP_DELETED_CELLS =>
'false', BLOCKSIZE => '65536', IN_MEMORY => 'false', BLOCKCACHE =>
{NAME => 'floats', DATA_BLOCK_ENCODING => 'NONE', BLOOMFILTER =>
'ROW', REPLICATION_SCOPE => '0', VERSIONS => '1', COMPRESSION =>
'ROW', REPLICATION_SCOPE => '0', VERSIONS => '1', COMPRESSION =>
MIN_VERSIONS => '0',
TTL => '2147483647',KEEP_DELETED_CELLS => 'false', BLOCKSIZE =>
'65536', IN_MEMORY => 'false', BLOCKCACHE => 'true'}, {NAME =>
TTL => '2147483647',KEEP_DELETED_CELLS => 'false', BLOCKSIZE =>
'65536', IN_MEMORY => 'false', BLOCKCACHE => 'true'}, {NAME =>
DATA_BLOCK_ENCODING => 'NONE',
BLOOMFILTER => 'ROW', REPLICATION_SCOPE => '0', VERSIONS => '1',
COMPRESSION => 'NONE', MIN_VERSIONS => '0', TTL => '2147483647',
KEEP_DELETED_CELLS => 'false', BLOCKSIZE => '65536', IN_MEMORY =>
BLOOMFILTER => 'ROW', REPLICATION_SCOPE => '0', VERSIONS => '1',
COMPRESSION => 'NONE', MIN_VERSIONS => '0', TTL => '2147483647',
KEEP_DELETED_CELLS => 'false', BLOCKSIZE => '65536', IN_MEMORY =>
BLOCKCACHE => 'true'}, {NAME => 'strings', DATA_BLOCK_ENCODING =>
BLOOMFILTER => 'ROW', REPLICATION_SCOPE => '0', VERSIONS => '1',
=> NONE', MIN_VERSIONS => '0', TTL => '2147483647',
'false', BLOCKSIZE => '65536', IN_MEMORY => 'false', BLOCKCACHE =>
2. create hive table successfully
hive> desc extended hbasealltypessmall;
OK
id int from deserializer
bool_col boolean from deserializer
tinyint_col tinyint from deserializer
smallint_col smallint from deserializer
int_col int from deserializer
bigint_col bigint from deserializer
float_col float from deserializer
double_col double from deserializer
date_string_col string from deserializer
string_col string from deserializer
timestamp_col timestamp from deserializer
Detailed Table Information Table(tableName:hbasealltypessmall,
dbName:default, owner:hadoop, createTime:1394217917,
hive> desc extended hbasealltypessmall;
OK
id int from deserializer
bool_col boolean from deserializer
tinyint_col tinyint from deserializer
smallint_col smallint from deserializer
int_col int from deserializer
bigint_col bigint from deserializer
float_col float from deserializer
double_col double from deserializer
date_string_col string from deserializer
string_col string from deserializer
timestamp_col timestamp from deserializer
Detailed Table Information Table(tableName:hbasealltypessmall,
dbName:default, owner:hadoop, createTime:1394217917,
retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:id,
comment:null), FieldSchema(name:bool_col, type:boolean,
FieldSchema(name:tinyint_col, type:tinyint, comment:null),
FieldSchema(name:smallint_col, type:smallint, comment:null),
FieldSchema(name:int_col, type:int, comment:null),
FieldSchema(name:bigint_col, type:bigint, comment:null),
FieldSchema(name:float_col, type:float, comment:null),
FieldSchema(name:double_col, type:double, comment:null),
FieldSchema(name:date_string_col, type:string, comment:null),
FieldSchema(name:string_col, type:string, comment:null),
FieldSchema(name:timestamp_col, type:timestamp, comment:null)],
location:hdfs://cdh5/user/hive/warehouse/hbasealltypessmall,
inputFormat:org.apache.hadoop.hive.hbase.HiveHBaseTableInputFormat,
FieldSchema(name:smallint_col, type:smallint, comment:null),
FieldSchema(name:int_col, type:int, comment:null),
FieldSchema(name:bigint_col, type:bigint, comment:null),
FieldSchema(name:float_col, type:float, comment:null),
FieldSchema(name:double_col, type:double, comment:null),
FieldSchema(name:date_string_col, type:string, comment:null),
FieldSchema(name:string_col, type:string, comment:null),
FieldSchema(name:timestamp_col, type:timestamp, comment:null)],
location:hdfs://cdh5/user/hive/warehouse/hbasealltypessmall,
inputFormat:org.apache.hadoop.hive.hbase.HiveHBaseTableInputFormat,
compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null,
serializationLib:org.apache.hadoop.hive.hbase.HBaseSerDe,
parameters:{serialization.format=1,
serializationLib:org.apache.hadoop.hive.hbase.HBaseSerDe,
parameters:{serialization.format=1,
bucketCols:[], sortCols:[], parameters:{},
skewedInfo:SkewedInfo(skewedColNames:[], skewedColValues:[],
skewedColValueLocationMaps:{}), storedAsSubDirectories:false),
partitionKeys:[], parameters:{EXTERNAL=TRUE,
hbase.table.name=hbasealltypessmall,
skewedInfo:SkewedInfo(skewedColNames:[], skewedColValues:[],
skewedColValueLocationMaps:{}), storedAsSubDirectories:false),
partitionKeys:[], parameters:{EXTERNAL=TRUE,
hbase.table.name=hbasealltypessmall,
storage_handler=org.apache.hadoop.hive.hbase.HBaseStorageHandler},
viewOriginalText:null, viewExpandedText:null,
viewOriginalText:null, viewExpandedText:null,
Time taken: 5.197 seconds, Fetched: 13 row(s)
And running select statement against this table was successful as
well.
3. run impala shell, invalidate metadata. I can see the table but
And running select statement against this table was successful as
well.
3. run impala shell, invalidate metadata. I can see the table but
I try to access this table, i got this following error. A select
against this table generates the same error:
[dumbo002-cdh5:21000] > desc hbasealltypessmall;
Query: describe hbasealltypessmall
ERROR: AnalysisException: Failed to load metadata for table:
default.hbasealltypessmall
CAUSED BY: TableLoadingException: TableLoadingException: Failed to
load metadata for HBase table: hbasealltypessmall
CAUSED BY: IOException: HRegionInfo was null or empty in hbase:meta,
row=keyvalues=NONE
BTW, I was able to perform this integration in CDH5-beta1.
Thanks,
Percy
To unsubscribe from this group and stop receiving emails from it,
[dumbo002-cdh5:21000] > desc hbasealltypessmall;
Query: describe hbasealltypessmall
ERROR: AnalysisException: Failed to load metadata for table:
default.hbasealltypessmall
CAUSED BY: TableLoadingException: TableLoadingException: Failed to
load metadata for HBase table: hbasealltypessmall
CAUSED BY: IOException: HRegionInfo was null or empty in hbase:meta,
row=keyvalues=NONE
BTW, I was able to perform this integration in CDH5-beta1.
Thanks,
Percy
To unsubscribe from this group and stop receiving emails from it,
an email to impala-user...@cloudera.org.
an email to impala-user...@cloudera.org.
email to impala-user+unsubscribe@cloudera.org.
email to impala-user+unsubscribe@cloudera.org.