FAQ
hi,
my job runs once ervey day. but it failed sometimes.
i checked the log in job tracker. It seems a hdfs error?
thanks a lot!

2011-08-16 21:07:13,247 INFO org.apache.hadoop.mapred.TaskInProgress:
Error from attempt_201106021431_1719_r_000498_1: org.ap
ache.hadoop.ipc.RemoteException: java.io.IOException: Could not
complete write to file /xxx/yyy/110816_210016/_tem
porary/_attempt_201106021431_1719_r_000498_1/part-r-00498 by
DFSClient_attempt_201106021431_1719_r_000498_1
at org.apache.hadoop.hdfs.server.namenode.NameNode.complete(NameNode.java:449)
at sun.reflect.GeneratedMethodAccessor12.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:508)
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:959)
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:955)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:396)
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:953)

at org.apache.hadoop.ipc.Client.call(Client.java:740)
at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:220)
at $Proxy1.complete(Unknown Source)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:82)
at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:59)
at $Proxy1.complete(Unknown Source)
at org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.closeInternal(DFSClient.java:3264)
at org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.close(DFSClient.java:3188)
at org.apache.hadoop.fs.FSDataOutputStream$PositionCache.close(FSDataOutputStream.java:61)
at org.apache.hadoop.fs.FSDataOutputStream.close(FSDataOutputStream.java:86)
at org.apache.hadoop.io.SequenceFile$Writer.close(SequenceFile.java:959)
at org.apache.hadoop.io.SequenceFile$BlockCompressWriter.close(SequenceFile.java:1290)
at org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat$1.close(SequenceFileOutputFormat.java:78)
at org.apache.hadoop.mapred.ReduceTask.runNewReducer(ReduceTask.java:567)
at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:408)
at org.apache.hadoop.mapred.Child.main(Child.java:170)

Search Discussions

  • Harsh J at Aug 17, 2011 at 4:38 am
    Do you notice anything related in the NameNode logs? One reason for
    this is that the NameNode may be in safe mode for some reason, but
    there are many other reasons so the NameNode's log would be the best
    place to look for exactly why the complete()-op fails.
    On Wed, Aug 17, 2011 at 8:20 AM, Jianxin Wang wrote:
    hi,
    my job runs once ervey day. but it failed sometimes.
    i checked the log in job tracker. It seems a hdfs error?
    thanks a lot!

    2011-08-16 21:07:13,247 INFO org.apache.hadoop.mapred.TaskInProgress:
    Error from attempt_201106021431_1719_r_000498_1: org.ap
    ache.hadoop.ipc.RemoteException: java.io.IOException: Could not
    complete write to file /xxx/yyy/110816_210016/_tem
    porary/_attempt_201106021431_1719_r_000498_1/part-r-00498 by
    DFSClient_attempt_201106021431_1719_r_000498_1
    at org.apache.hadoop.hdfs.server.namenode.NameNode.complete(NameNode.java:449)
    at sun.reflect.GeneratedMethodAccessor12.invoke(Unknown Source)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
    at java.lang.reflect.Method.invoke(Method.java:597)
    at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:508)
    at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:959)
    at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:955)
    at java.security.AccessController.doPrivileged(Native Method)
    at javax.security.auth.Subject.doAs(Subject.java:396)
    at org.apache.hadoop.ipc.Server$Handler.run(Server.java:953)

    at org.apache.hadoop.ipc.Client.call(Client.java:740)
    at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:220)
    at $Proxy1.complete(Unknown Source)
    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
    at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
    at java.lang.reflect.Method.invoke(Method.java:597)
    at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:82)
    at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:59)
    at $Proxy1.complete(Unknown Source)
    at org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.closeInternal(DFSClient.java:3264)
    at org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.close(DFSClient.java:3188)
    at org.apache.hadoop.fs.FSDataOutputStream$PositionCache.close(FSDataOutputStream.java:61)
    at org.apache.hadoop.fs.FSDataOutputStream.close(FSDataOutputStream.java:86)
    at org.apache.hadoop.io.SequenceFile$Writer.close(SequenceFile.java:959)
    at org.apache.hadoop.io.SequenceFile$BlockCompressWriter.close(SequenceFile.java:1290)
    at org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat$1.close(SequenceFileOutputFormat.java:78)
    at org.apache.hadoop.mapred.ReduceTask.runNewReducer(ReduceTask.java:567)
    at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:408)
    at org.apache.hadoop.mapred.Child.main(Child.java:170)


    --
    Harsh J
  • Jianxin Wang at Aug 17, 2011 at 6:22 am
    Thanks Harsh :)
    the hadoop system started 3 months ago. so i think it is not in safe mode.
    i found some old tasks started 10 days ago, the tasks seem blocked by
    some unknown reason. I killed these tasks now. but i don't know why a
    task can be blocked and exists so long.

    I found another type of exception like this in namenode's node.
    2011-08-16 11:25:44,352 INFO org.apache.hadoop.mapred.TaskInProgress:
    Error from attempt_201106021431_1704_r_000506_1: org.ap
    ache.hadoop.ipc.RemoteException:
    org.apache.hadoop.hdfs.server.namenode.LeaseExpiredException: No lease
    on /walter/send_album
    s/110816_111850/_temporary/_attempt_201106021431_1704_r_000506_1/part-r-00506
    File does not exist. Holder DFSClient_attempt_2
    01106021431_1704_r_000506_1 does not have any open files.
    at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkLease(FSNamesystem.java:1332)
    at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkLease(FSNamesystem.java:1323)
    at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1251)
    at org.apache.hadoop.hdfs.server.namenode.NameNode.addBlock(NameNode.java:422)
    at sun.reflect.GeneratedMethodAccessor17.invoke(Unknown Source)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
    at java.lang.reflect.Method.invoke(Method.java:597)
    at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:508)
    at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:959)
    at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:955)
    at java.security.AccessController.doPrivileged(Native Method)
    at javax.security.auth.Subject.doAs(Subject.java:396)
    at org.apache.hadoop.ipc.Server$Handler.run(Server.java:953)

    at org.apache.hadoop.ipc.Client.call(Client.java:740)
    at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:220)
    at $Proxy1.addBlock(Unknown Source)
    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
    at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)

    2011/8/17, Harsh J <harsh@cloudera.com>:
    Do you notice anything related in the NameNode logs? One reason for
    this is that the NameNode may be in safe mode for some reason, but
    there are many other reasons so the NameNode's log would be the best
    place to look for exactly why the complete()-op fails.
    On Wed, Aug 17, 2011 at 8:20 AM, Jianxin Wang wrote:
    hi,
    my job runs once ervey day. but it failed sometimes.
    i checked the log in job tracker. It seems a hdfs error?
    thanks a lot!

    2011-08-16 21:07:13,247 INFO org.apache.hadoop.mapred.TaskInProgress:
    Error from attempt_201106021431_1719_r_000498_1: org.ap
    ache.hadoop.ipc.RemoteException: java.io.IOException: Could not
    complete write to file /xxx/yyy/110816_210016/_tem
    porary/_attempt_201106021431_1719_r_000498_1/part-r-00498 by
    DFSClient_attempt_201106021431_1719_r_000498_1
    at
    org.apache.hadoop.hdfs.server.namenode.NameNode.complete(NameNode.java:449)
    at sun.reflect.GeneratedMethodAccessor12.invoke(Unknown Source)
    at
    sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
    at java.lang.reflect.Method.invoke(Method.java:597)
    at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:508)
    at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:959)
    at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:955)
    at java.security.AccessController.doPrivileged(Native Method)
    at javax.security.auth.Subject.doAs(Subject.java:396)
    at org.apache.hadoop.ipc.Server$Handler.run(Server.java:953)

    at org.apache.hadoop.ipc.Client.call(Client.java:740)
    at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:220)
    at $Proxy1.complete(Unknown Source)
    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
    at
    sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
    at
    sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
    at java.lang.reflect.Method.invoke(Method.java:597)
    at
    org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:82)
    at
    org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:59)
    at $Proxy1.complete(Unknown Source)
    at
    org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.closeInternal(DFSClient.java:3264)
    at
    org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.close(DFSClient.java:3188)
    at
    org.apache.hadoop.fs.FSDataOutputStream$PositionCache.close(FSDataOutputStream.java:61)
    at
    org.apache.hadoop.fs.FSDataOutputStream.close(FSDataOutputStream.java:86)
    at org.apache.hadoop.io.SequenceFile$Writer.close(SequenceFile.java:959)
    at
    org.apache.hadoop.io.SequenceFile$BlockCompressWriter.close(SequenceFile.java:1290)
    at
    org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat$1.close(SequenceFileOutputFormat.java:78)
    at org.apache.hadoop.mapred.ReduceTask.runNewReducer(ReduceTask.java:567)
    at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:408)
    at org.apache.hadoop.mapred.Child.main(Child.java:170)


    --
    Harsh J
  • Jianxin Wang at Aug 23, 2011 at 3:31 am
    hi,harsh, i wrote the namenode's log here. The problem occurs occasionally
    Thanks a lot!

    2011-08-22 14:41:05,939 INFO
    org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
    ugi=tpic,users,dialout ip=/172.28.1.
    29 cmd=delete src=/walter/send_albums/110822_143455/_temporary
    dst=null perm=null
    2011-08-22 14:41:06,025 INFO
    org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
    ugi=tpic,users,dialout ip=/172.24.31
    .148 cmd=create
    src=/walter/send_albums/110822_143455/_temporary/_attempt_201106021431_1838_r_000509_1/part-r-00509
    dst=null
    perm=tpic:supergroup:rw-r--r--
    2011-08-22 14:41:06,227 INFO org.apache.hadoop.ipc.Server: IPC Server
    handler 6 on 9000, call addBlock(/walter/send_albums/11
    0822_143455/_temporary/_attempt_201106021431_1838_r_000506_1/part-r-00506,
    DFSClient_attempt_201106021431_1838_r_000506_1) fr
    om 10.163.138.122:42866: error:
    org.apache.hadoop.hdfs.server.namenode.LeaseExpiredException: No lease
    on /walter/send_albums
    /110822_143455/_temporary/_attempt_201106021431_1838_r_000506_1/part-r-00506
    File does not exist. Holder DFSClient_attempt_20
    1106021431_1838_r_000506_1 does not have any open files.
    org.apache.hadoop.hdfs.server.namenode.LeaseExpiredException: No lease
    on /walter/send_albums/110822_143455/_temporary/_attem
    pt_201106021431_1838_r_000506_1/part-r-00506 File does not exist.
    Holder DFSClient_attempt_201106021431_1838_r_000506_1 does
    not have any open files.
    at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkLease(FSNamesystem.java:1332)
    at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkLease(FSNamesystem.java:1323)
    at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1251)
    at org.apache.hadoop.hdfs.server.namenode.NameNode.addBlock(NameNode.java:422)
    at sun.reflect.GeneratedMethodAccessor17.invoke(Unknown Source)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
    at java.lang.reflect.Method.invoke(Method.java:597)
    at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:508)
    at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:959)
    at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:955)
    at java.security.AccessController.doPrivileged(Native Method)
    at javax.security.auth.Subject.doAs(Subject.java:396)
    at org.apache.hadoop.ipc.Server$Handler.run(Server.java:953)
    2011-08-22 14:41:06,297 INFO org.apache.hadoop.hdfs.StateChange:
    BLOCK* NameSystem.addStoredBlock: addStoredBlock request rec
    eived for blk_521784094891641529_825020 on 172.24.32.27:50010 size
    627045 But it does not belong to any file.
    2011-08-22 14:41:06,299 INFO org.apache.hadoop.hdfs.StateChange:
    BLOCK* NameSystem.addStoredBlock: addStoredBlock request rec
    eived for blk_521784094891641529_825020 on 10.163.128.50:50010 size
    627045 But it does not belong to any file.
    2011-08-22 14:41:06,299 WARN org.apache.hadoop.hdfs.StateChange: DIR*
    NameSystem.completeFile: failed to complete /walter/sen
    d_albums/110822_143455/_temporary/_attempt_201106021431_1838_r_000501_1/part-r-00501
    because dir.getFileBlocks() is null and
    pendingFile is null
    2011-08-22 14:41:06,299 INFO org.apache.hadoop.ipc.Server: IPC Server
    handler 8 on 9000, call complete(/walter/send_albums/11
    0822_143455/_temporary/_attempt_201106021431_1838_r_000501_1/part-r-00501,
    DFSClient_attempt_201106021431_1838_r_000501_1) fr
    om 10.163.128.50:38026: error: java.io.IOException: Could not complete
    write to file /walter/send_albums/110822_143455/_tempo
    rary/_attempt_201106021431_1838_r_000501_1/part-r-00501 by
    DFSClient_attempt_201106021431_1838_r_000501_1
    java.io.IOException: Could not complete write to file
    /walter/send_albums/110822_143455/_temporary/_attempt_201106021431_1838
    _r_000501_1/part-r-00501 by DFSClient_attempt_201106021431_1838_r_000501_1
    at org.apache.hadoop.hdfs.server.namenode.NameNode.complete(NameNode.java:449)
    at sun.reflect.GeneratedMethodAccessor12.invoke(Unknown Source)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)

    2011/8/17, Jianxin Wang <wangjx798@gmail.com>:
    Thanks Harsh :)
    the hadoop system started 3 months ago. so i think it is not in safe mode.
    i found some old tasks started 10 days ago, the tasks seem blocked by
    some unknown reason. I killed these tasks now. but i don't know why a
    task can be blocked and exists so long.

    I found another type of exception like this in namenode's node.
    2011-08-16 11:25:44,352 INFO org.apache.hadoop.mapred.TaskInProgress:
    Error from attempt_201106021431_1704_r_000506_1: org.ap
    ache.hadoop.ipc.RemoteException:
    org.apache.hadoop.hdfs.server.namenode.LeaseExpiredException: No lease
    on /walter/send_album
    s/110816_111850/_temporary/_attempt_201106021431_1704_r_000506_1/part-r-00506
    File does not exist. Holder DFSClient_attempt_2
    01106021431_1704_r_000506_1 does not have any open files.
    at
    org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkLease(FSNamesystem.java:1332)
    at
    org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkLease(FSNamesystem.java:1323)
    at
    org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1251)
    at
    org.apache.hadoop.hdfs.server.namenode.NameNode.addBlock(NameNode.java:422)
    at sun.reflect.GeneratedMethodAccessor17.invoke(Unknown Source)
    at
    sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
    at java.lang.reflect.Method.invoke(Method.java:597)
    at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:508)
    at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:959)
    at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:955)
    at java.security.AccessController.doPrivileged(Native Method)
    at javax.security.auth.Subject.doAs(Subject.java:396)
    at org.apache.hadoop.ipc.Server$Handler.run(Server.java:953)

    at org.apache.hadoop.ipc.Client.call(Client.java:740)
    at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:220)
    at $Proxy1.addBlock(Unknown Source)
    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
    at
    sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
    at
    sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)

    2011/8/17, Harsh J <harsh@cloudera.com>:
    Do you notice anything related in the NameNode logs? One reason for
    this is that the NameNode may be in safe mode for some reason, but
    there are many other reasons so the NameNode's log would be the best
    place to look for exactly why the complete()-op fails.

    On Wed, Aug 17, 2011 at 8:20 AM, Jianxin Wang <wangjx798@gmail.com>
    wrote:
    hi,
    my job runs once ervey day. but it failed sometimes.
    i checked the log in job tracker. It seems a hdfs error?
    thanks a lot!

    2011-08-16 21:07:13,247 INFO org.apache.hadoop.mapred.TaskInProgress:
    Error from attempt_201106021431_1719_r_000498_1: org.ap
    ache.hadoop.ipc.RemoteException: java.io.IOException: Could not
    complete write to file /xxx/yyy/110816_210016/_tem
    porary/_attempt_201106021431_1719_r_000498_1/part-r-00498 by
    DFSClient_attempt_201106021431_1719_r_000498_1
    at
    org.apache.hadoop.hdfs.server.namenode.NameNode.complete(NameNode.java:449)
    at sun.reflect.GeneratedMethodAccessor12.invoke(Unknown Source)
    at
    sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
    at java.lang.reflect.Method.invoke(Method.java:597)
    at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:508)
    at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:959)
    at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:955)
    at java.security.AccessController.doPrivileged(Native Method)
    at javax.security.auth.Subject.doAs(Subject.java:396)
    at org.apache.hadoop.ipc.Server$Handler.run(Server.java:953)

    at org.apache.hadoop.ipc.Client.call(Client.java:740)
    at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:220)
    at $Proxy1.complete(Unknown Source)
    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
    at
    sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
    at
    sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
    at java.lang.reflect.Method.invoke(Method.java:597)
    at
    org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:82)
    at
    org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:59)
    at $Proxy1.complete(Unknown Source)
    at
    org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.closeInternal(DFSClient.java:3264)
    at
    org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.close(DFSClient.java:3188)
    at
    org.apache.hadoop.fs.FSDataOutputStream$PositionCache.close(FSDataOutputStream.java:61)
    at
    org.apache.hadoop.fs.FSDataOutputStream.close(FSDataOutputStream.java:86)
    at
    org.apache.hadoop.io.SequenceFile$Writer.close(SequenceFile.java:959)
    at
    org.apache.hadoop.io.SequenceFile$BlockCompressWriter.close(SequenceFile.java:1290)
    at
    org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat$1.close(SequenceFileOutputFormat.java:78)
    at
    org.apache.hadoop.mapred.ReduceTask.runNewReducer(ReduceTask.java:567)
    at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:408)
    at org.apache.hadoop.mapred.Child.main(Child.java:170)


    --
    Harsh J

Related Discussions

Discussion Navigation
viewthread | post
Discussion Overview
groupcommon-user @
categorieshadoop
postedAug 17, '11 at 2:50a
activeAug 23, '11 at 3:31a
posts4
users2
websitehadoop.apache.org...
irc#hadoop

2 users in discussion

Jianxin Wang: 3 posts Harsh J: 1 post

People

Translate

site design / logo © 2022 Grokbase