FAQ
Manisha,

Could you try running 'refresh' at the impala shell?

Thanks,
Udai

On Wed, Jan 30, 2013 at 1:35 AM, Manisha Agrawal wrote:

yes.
Data is accessible through hive

But Impala cannot read the table names even.
On issuing command: show databases, it jus returns a single db named
default.
On Tuesday, January 29, 2013 12:50:53 AM UTC+5:30, Udai wrote:

Manisha,

Are you able to read the data via Hive on the same table?

Thanks,
Udai

On Thu, Jan 24, 2013 at 11:10 PM, Manisha Agrawal wrote:

Hi Udai,

Thanks.

I resolved root issue by adding it to dfs.block.local-path-**access.**user
in hdfs-site.xml and also changed file permissions to 777.
Now I am starting all services as root along with impala-shell.

It worked once, but now metadata is unreadbale with root user.

File format being queried is a row delimited file.
Are there any checks need to be made before startng impala?

Please help.
Thanks
Mansiha Agrawal

On Friday, January 25, 2013 8:15:36 AM UTC+5:30, Udai wrote:

Hello Manisha,

Appreciate the Configs. What is the file format that you are querying?

Thanks,
Udai


On Wed, Jan 23, 2013 at 9:22 PM, Manisha Agrawal <agrawalm...@gmail.com
wrote:
Hi all,

I have CDH 4.1.2 running in pseudo mode on a RHEL 64 bit machine.
Mysql is configured as Hive's remote datasore.
Hive succsessfully reads metadata and fetches correct results.

Impala is running on top of it. I am unable to excute queries on
impala.

I have a external table named sample and permissions on file are as
follows:
[root@ip-10-100-255-207 conf]# sudo -u hdfs hadoop fs -ls -R
/user/impala
-rw-r--r-- 2 impala hdfs 31 2013-01-23 23:10
/user/impala/test_ext.out


This is how I start statestored and impalad services

GLOG_v=1 /usr/bin/statestored -state_store_port=24000 &
GLOG_v=1 /usr/bin/impalad -state_store_host=localhost -nn=localhost
-nn_port=8020 -hostname=localhost -ipaddress=10.100.255.207 &

When I start services as *root user* , hdfs is unreadable.
and impala-shell throws below error:

readDirect: FSDataInputStream#read error:
org.apache.hadoop.hdfs.**BlockMi**ssingException: Could not obtain
block: BP-344648192-10.100.255.207-**13**58752982703:blk_**
18336384570160**8725_1522 file=/user/impala/test_ext.out
at org.apache.hadoop.hdfs.**DFSInpu**tStream.chooseDataNode(**
DFSInpu**tStream.java:734)
at org.apache.hadoop.hdfs.**DFSInpu**tStream.blockSeekTo(**
DFSInputSt**ream.java:448)
at org.apache.hadoop.hdfs.**DFSInpu**tStream.**
readWithStrategy(**DFSIn**putStream.java:645)
at org.apache.hadoop.hdfs.**DFSInpu**tStream.read(**
DFSInputStream.**java:696)
at org.apache.hadoop.fs.**FSDataInp**utStream.read(**
FSDataInputStrea**m.java:123)
hdfsOpenFile(hdfs://localhost:****8020/user/impala/test_ext.out)****:
WARN: Unexpected error 255 when testing for direct read compatibility


On starting services as *impala user,* metadata is unreadable
error in impala-shell is as below:

Query: select * from sample
ERROR: com.cloudera.impala.common.**Ana**lysisException: Analysis
exception (in select * from sample)
at com.cloudera.impala.analysis.**A**nalysisContext.analyze(**
Analysi**sContext.java:133)
at com.cloudera.impala.service.**Fr**ontend.createExecRequest(
**Front**end.java:216)
at com.cloudera.impala.service.**Jn**
iFrontend.createExecRequest(**Jn**iFrontend.java:86)
Caused by: com.cloudera.impala.common.**Ana**lysisException: Unknown
table: 'sample'
at com.cloudera.impala.analysis.**A**
nalyzer.registerBaseTableRef(**A**nalyzer.java:178)
at com.cloudera.impala.analysis.**B**aseTableRef.analyze(**
BaseTableR**ef.java:51)
at com.cloudera.impala.analysis.**S**
electStmt.analyze(SelectStmt.**j**ava:115)
at com.cloudera.impala.analysis.**A**nalysisContext.analyze(**
Analysi**sContext.java:130)
... 2 more

Please provide suggestions to resolve the issue.

Thanks in advance
Manisha
==============================****==============================****
==============================
*core-site.xml*

<configuration>
<property>
<name>fs.default.name</name>
<value>hdfs://localhost:8020</****value>
</property>
<!-- OOZIE proxy user setting -->
<property>
<name>hadoop.proxyuser.oozie.**h**osts</name>
<value>*</value>
</property>
<property>
<name>hadoop.proxyuser.oozie.**g**roups</name>
<value>*</value>
</property>
<!-- HTTPFS proxy user setting -->
<property>
<name>hadoop.proxyuser.httpfs.****hosts</name>
<value>*</value>
</property>
<property>
<name>hadoop.proxyuser.httpfs.****groups</name>
<value>*</value>
</property>
<property>
<name>dfs.client.read.**shortcir**cuit</name>
<value>true</value>
</property>
</configuration>

==============================****==============================****
==============================

*hdfs-site.xml*

<configuration>
<property>
<name>dfs.replication</name>
<value>1</value>
</property>
<!-- Immediately exit safemode as soon as one DataNode checks in.
On a multi-node cluster, these configurations must be removed.
-->
<property>
<name>dfs.safemode.extension</****name>
<value>0</value>
</property>
<property>
<name>dfs.safemode.min.**datanod**es</name>
<value>1</value>
</property>
<property>
<name>hadoop.tmp.dir</name>
<value>/var/lib/hadoop-hdfs/**ca**che/${user.name}</value>
</property>
<property>
<name>dfs.namenode.name.dir</**n**ame>
<value>/var/lib/hadoop-hdfs/**ca**che/${user.name}/dfs/name</**
val**ue>
</property>
<property>
<name>dfs.namenode.checkpoint.****dir</name>
<value>/var/lib/hadoop-hdfs/**ca**che/${user.name}/dfs/**
namesecon**dary</value>
</property>
<property>
<name>dfs.datanode.data.dir</**n**ame>
<value>/var/lib/hadoop-hdfs/**ca**che/${user.name}/dfs/data</**
val**ue>
</property>
<property>
<name>dfs.datanode.data.dir.**pe**rm</name>
<value>750</value>
</property>
<property>
<name>dfs.block.local-path-**acc**ess.user</name>
<value>impala,hdfs,mapred</**val**ue>
</property>
<property>
<name>dfs.datanode.hdfs-**blocks**-metadata.enabled</name>
<value>true</value>
</property>
</configuration>
==============================****==============================****
==============================

*hive-site.xml*

<configuration>
<property>
<name>javax.jdo.option.**Connect**ionURL</name>
<value>jdbc:mysql://localhost/****metastore</value>
<description>JDBC connect string for a JDBC metastore</description>
</property>
<property>
<name>javax.jdo.option.**Connect**ionDriverName</name>
<value>com.mysql.jdbc.Driver</****value>
<description>Driver class name for a JDBC metastore</description>
</property>
<property>
<name>javax.jdo.option.**Connect**ionUserName</name>
<value>hive</value>
</property>

<property>
<name>javax.jdo.option.**Connect**ionPassword</name>
<value>hive</value>
</property>

<property>
<name>hive.metastore.local</**na**me>
<value>true</value>
</property>
</configuration>
=============================****==============================****
==============================****=============




--

Search Discussions

Discussion Posts

Previous

Related Discussions

Discussion Navigation
viewthread | post
posts ‹ prev | 5 of 5 | next ›
Discussion Overview
groupimpala-user @
categorieshadoop
postedJan 25, '13 at 2:45a
activeJan 30, '13 at 4:36p
posts5
users2
websitecloudera.com
irc#hadoop

People

Translate

site design / logo © 2022 Grokbase