FAQ
Hi!
I have installed cdh4.1,Hive0.9, Impala0.5 without Cloudera Manager as
explained in the documentation and I can run queries in Hive without any
errors, But i run a query of "select ... from ..." in impala-shell, The
impalad gave me errors as following:
......
13/03/02 11:47:14 WARN hdfs.DFSClient: BlockReaderLocal: Removing
BP-1467891591-127.0.0.1-1361530651473:blk_7749180477903214186_1004 from
cache because local file
/tmp/hadoop-hdfs/dfs/data/current/BP-1467891591-127.0.0.1-1361530651473/current/finalized/blk_7749180477903214186
could not be opened.
13/03/02 11:47:14 WARN hdfs.DFSClient: Failed to connect to
/192.168.1.55:50010 for block, add to deadNodes and continue.
java.io.FileNotFoundException:
/tmp/hadoop-hdfs/dfs/data/current/BP-1467891591-127.0.0.1-1361530651473/current/finalized/blk_7749180477903214186
(Permission denied)
java.io.FileNotFoundException:
/tmp/hadoop-hdfs/dfs/data/current/BP-1467891591-127.0.0.1-1361530651473/current/finalized/blk_7749180477903214186
(Permission denied)
at java.io.FileInputStream.open(Native Method)
......
13/03/02 11:47:14 INFO hdfs.DFSClient: Could not obtain block
BP-1467891591-127.0.0.1-1361530651473:blk_7749180477903214186_1004 from any
node: java.io.IOException: No live nodes contain current block. Will get
new block locations from namenode and retry...
13/03/02 11:47:14 WARN hdfs.DFSClient: DFS chooseDataNode: got # 2
IOException, will wait for 7959.810222203046 msec.
13/03/02 11:47:22 WARN hdfs.DFSClient: BlockReaderLocal: Removing
BP-1467891591-127.0.0.1-1361530651473:blk_7749180477903214186_1004 from
cache because local file
/tmp/hadoop-hdfs/dfs/data/current/BP-1467891591-127.0.0.1-1361530651473/current/finalized/blk_7749180477903214186
could not be opened.
13/03/02 11:47:22 WARN hdfs.DFSClient: Failed to connect to
/192.168.1.55:50010 for block, add to deadNodes and continue.
java.io.FileNotFoundException:
/tmp/hadoop-hdfs/dfs/data/current/BP-1467891591-127.0.0.1-1361530651473/current/finalized/blk_7749180477903214186
(Permission denied)
java.io.FileNotFoundException:
/tmp/hadoop-hdfs/dfs/data/current/BP-1467891591-127.0.0.1-1361530651473/current/finalized/blk_7749180477903214186
(Permission denied)
at java.io.FileInputStream.open(Native Method)
......
13/03/02 11:47:32 WARN hdfs.DFSClient: DFS Read
org.apache.hadoop.hdfs.BlockMissingException: Could not obtain block:
BP-1467891591-127.0.0.1-1361530651473:blk_7749180477903214186_1004
file=/user/hive/warehouse/yyfx.db/yysj/Exaiu-mm-cdr-201301180540-00002.dat
at
org.apache.hadoop.hdfs.DFSInputStream.chooseDataNode(DFSInputStream.java:734)
......

I have done the following settings or operations as explained in the
documentation:
in the hadoop/conf/core-site.xml, i added:
<property>
<name>dfs.client.read.shortcircuit</name>
<value>true</value>
</property>

in the hadoop/conf/hdfs-site.xml, i added:
<property>
<name>dfs.datanode.data.dir.perm</name>
<value>750</value>
</property>
<property>
<name>dfs.block.local-path-access.user</name>
<value>impala</value>
</property>

And done the following commands:
$ sudo -u hdfs hadoop fs -mkdir /tmp
$ sudo -u hdfs hadoop fs -chmod -R 1777 /tmp

I'm starting impalad with command line:
$ sudo -u impala GLOG_v=1 impalad -use_statestore=false -nn=cloudera-host1
-nn_port=8020 hostname=cloundera-host2 -ipaddress=192.168.1.55

Why did impalad give me the errors?

Thanks!

Search Discussions

  • Zjp... at Mar 4, 2013 at 3:38 am
    Hi, everybody!
    I have found the solution.

    Thank you for your attention!

    在 2013年3月2日星期六UTC上午6时08分05秒,zjp...写道:
    Hi!
    I have installed cdh4.1,Hive0.9, Impala0.5 without Cloudera Manager as
    explained in the documentation and I can run queries in Hive without any
    errors, But i run a query of "select ... from ..." in impala-shell, The
    impalad gave me errors as following:
    ......
    13/03/02 11:47:14 WARN hdfs.DFSClient: BlockReaderLocal: Removing
    BP-1467891591-127.0.0.1-1361530651473:blk_7749180477903214186_1004 from
    cache because local file
    /tmp/hadoop-hdfs/dfs/data/current/BP-1467891591-127.0.0.1-1361530651473/current/finalized/blk_7749180477903214186
    could not be opened.
    13/03/02 11:47:14 WARN hdfs.DFSClient: Failed to connect to /
    192.168.1.55:50010 for block, add to deadNodes and continue.
    java.io.FileNotFoundException:
    /tmp/hadoop-hdfs/dfs/data/current/BP-1467891591-127.0.0.1-1361530651473/current/finalized/blk_7749180477903214186
    (Permission denied)
    java.io.FileNotFoundException:
    /tmp/hadoop-hdfs/dfs/data/current/BP-1467891591-127.0.0.1-1361530651473/current/finalized/blk_7749180477903214186
    (Permission denied)
    at java.io.FileInputStream.open(Native Method)
    ......
    13/03/02 11:47:14 INFO hdfs.DFSClient: Could not obtain block
    BP-1467891591-127.0.0.1-1361530651473:blk_7749180477903214186_1004 from any
    node: java.io.IOException: No live nodes contain current block. Will get
    new block locations from namenode and retry...
    13/03/02 11:47:14 WARN hdfs.DFSClient: DFS chooseDataNode: got # 2
    IOException, will wait for 7959.810222203046 msec.
    13/03/02 11:47:22 WARN hdfs.DFSClient: BlockReaderLocal: Removing
    BP-1467891591-127.0.0.1-1361530651473:blk_7749180477903214186_1004 from
    cache because local file
    /tmp/hadoop-hdfs/dfs/data/current/BP-1467891591-127.0.0.1-1361530651473/current/finalized/blk_7749180477903214186
    could not be opened.
    13/03/02 11:47:22 WARN hdfs.DFSClient: Failed to connect to /
    192.168.1.55:50010 for block, add to deadNodes and continue.
    java.io.FileNotFoundException:
    /tmp/hadoop-hdfs/dfs/data/current/BP-1467891591-127.0.0.1-1361530651473/current/finalized/blk_7749180477903214186
    (Permission denied)
    java.io.FileNotFoundException:
    /tmp/hadoop-hdfs/dfs/data/current/BP-1467891591-127.0.0.1-1361530651473/current/finalized/blk_7749180477903214186
    (Permission denied)
    at java.io.FileInputStream.open(Native Method)
    ......
    13/03/02 11:47:32 WARN hdfs.DFSClient: DFS Read
    org.apache.hadoop.hdfs.BlockMissingException: Could not obtain block:
    BP-1467891591-127.0.0.1-1361530651473:blk_7749180477903214186_1004
    file=/user/hive/warehouse/yyfx.db/yysj/Exaiu-mm-cdr-201301180540-00002.dat
    at
    org.apache.hadoop.hdfs.DFSInputStream.chooseDataNode(DFSInputStream.java:734)
    ......

    I have done the following settings or operations as explained in the
    documentation:
    in the hadoop/conf/core-site.xml, i added:
    <property>
    <name>dfs.client.read.shortcircuit</name>
    <value>true</value>
    </property>

    in the hadoop/conf/hdfs-site.xml, i added:
    <property>
    <name>dfs.datanode.data.dir.perm</name>
    <value>750</value>
    </property>
    <property>
    <name>dfs.block.local-path-access.user</name>
    <value>impala</value>
    </property>

    And done the following commands:
    $ sudo -u hdfs hadoop fs -mkdir /tmp
    $ sudo -u hdfs hadoop fs -chmod -R 1777 /tmp

    I'm starting impalad with command line:
    $ sudo -u impala GLOG_v=1 impalad -use_statestore=false -nn=cloudera-host1
    -nn_port=8020 hostname=cloundera-host2 -ipaddress=192.168.1.55

    Why did impalad give me the errors?

    Thanks!
  • Harsh J at Mar 4, 2013 at 5:03 am
    Hi zjp…,

    I believe the issue may have been improper short-circuit read
    configuration at the server/client (specifically
    dfs.datanode.data.dir.perm permission configs). In any case, please do
    share what solved your specific case in spirit of
    http://xkcd.com/979/.

    Note that with CDH 4.2 onwards,the short-circuit reads config has
    changed a little to cover secure clients and remove need for setting
    up dfs.datanode.data.dir.perm to expose blocks as o+r. This is
    covered at https://ccp.cloudera.com/display/CDH4DOC/Tips+and+Guidelines#TipsandGuidelines-ImprovePerformanceforLocalReads
    and also in Impala's guide at
    https://ccp.cloudera.com/display/IMPALA10BETADOC/Configuring+Impala+for+Performance.
    On Mon, Mar 4, 2013 at 9:08 AM, zjp... wrote:
    Hi, everybody!
    I have found the solution.

    Thank you for your attention!

    在 2013年3月2日星期六UTC上午6时08分05秒,zjp...写道:
    Hi!
    I have installed cdh4.1,Hive0.9, Impala0.5 without Cloudera Manager as
    explained in the documentation and I can run queries in Hive without any
    errors, But i run a query of "select ... from ..." in impala-shell, The
    impalad gave me errors as following:
    ......
    13/03/02 11:47:14 WARN hdfs.DFSClient: BlockReaderLocal: Removing
    BP-1467891591-127.0.0.1-1361530651473:blk_7749180477903214186_1004 from
    cache because local file
    /tmp/hadoop-hdfs/dfs/data/current/BP-1467891591-127.0.0.1-1361530651473/current/finalized/blk_7749180477903214186
    could not be opened.
    13/03/02 11:47:14 WARN hdfs.DFSClient: Failed to connect to
    /192.168.1.55:50010 for block, add to deadNodes and continue.
    java.io.FileNotFoundException:
    /tmp/hadoop-hdfs/dfs/data/current/BP-1467891591-127.0.0.1-1361530651473/current/finalized/blk_7749180477903214186
    (Permission denied)
    java.io.FileNotFoundException:
    /tmp/hadoop-hdfs/dfs/data/current/BP-1467891591-127.0.0.1-1361530651473/current/finalized/blk_7749180477903214186
    (Permission denied)
    at java.io.FileInputStream.open(Native Method)
    ......
    13/03/02 11:47:14 INFO hdfs.DFSClient: Could not obtain block
    BP-1467891591-127.0.0.1-1361530651473:blk_7749180477903214186_1004 from any
    node: java.io.IOException: No live nodes contain current block. Will get new
    block locations from namenode and retry...
    13/03/02 11:47:14 WARN hdfs.DFSClient: DFS chooseDataNode: got # 2
    IOException, will wait for 7959.810222203046 msec.
    13/03/02 11:47:22 WARN hdfs.DFSClient: BlockReaderLocal: Removing
    BP-1467891591-127.0.0.1-1361530651473:blk_7749180477903214186_1004 from
    cache because local file
    /tmp/hadoop-hdfs/dfs/data/current/BP-1467891591-127.0.0.1-1361530651473/current/finalized/blk_7749180477903214186
    could not be opened.
    13/03/02 11:47:22 WARN hdfs.DFSClient: Failed to connect to
    /192.168.1.55:50010 for block, add to deadNodes and continue.
    java.io.FileNotFoundException:
    /tmp/hadoop-hdfs/dfs/data/current/BP-1467891591-127.0.0.1-1361530651473/current/finalized/blk_7749180477903214186
    (Permission denied)
    java.io.FileNotFoundException:
    /tmp/hadoop-hdfs/dfs/data/current/BP-1467891591-127.0.0.1-1361530651473/current/finalized/blk_7749180477903214186
    (Permission denied)
    at java.io.FileInputStream.open(Native Method)
    ......
    13/03/02 11:47:32 WARN hdfs.DFSClient: DFS Read
    org.apache.hadoop.hdfs.BlockMissingException: Could not obtain block:
    BP-1467891591-127.0.0.1-1361530651473:blk_7749180477903214186_1004
    file=/user/hive/warehouse/yyfx.db/yysj/Exaiu-mm-cdr-201301180540-00002.dat
    at
    org.apache.hadoop.hdfs.DFSInputStream.chooseDataNode(DFSInputStream.java:734)
    ......

    I have done the following settings or operations as explained in the
    documentation:
    in the hadoop/conf/core-site.xml, i added:
    <property>
    <name>dfs.client.read.shortcircuit</name>
    <value>true</value>
    </property>

    in the hadoop/conf/hdfs-site.xml, i added:
    <property>
    <name>dfs.datanode.data.dir.perm</name>
    <value>750</value>
    </property>
    <property>
    <name>dfs.block.local-path-access.user</name>
    <value>impala</value>
    </property>

    And done the following commands:
    $ sudo -u hdfs hadoop fs -mkdir /tmp
    $ sudo -u hdfs hadoop fs -chmod -R 1777 /tmp

    I'm starting impalad with command line:
    $ sudo -u impala GLOG_v=1 impalad -use_statestore=false -nn=cloudera-host1
    -nn_port=8020 hostname=cloundera-host2 -ipaddress=192.168.1.55

    Why did impalad give me the errors?

    Thanks!


    --
    Harsh J

Related Discussions

Discussion Navigation
viewthread | post
Discussion Overview
groupimpala-user @
categorieshadoop
postedMar 2, '13 at 6:08a
activeMar 4, '13 at 5:03a
posts3
users2
websitecloudera.com
irc#hadoop

2 users in discussion

Zjp...: 2 posts Harsh J: 1 post

People

Translate

site design / logo © 2021 Grokbase