Search Discussions

23 discussions - 100 posts

  • Hello, My company will be building a small but quickly growing Hadoop deployment, and I had a question regarding best practice for configuring the storage for the datanodes. Cloudera has a page where ...
    John BuchananJohn Buchanan
    Feb 7, 2011 at 8:25 pm
    Feb 22, 2011 at 12:34 pm
  • Currently I am using the default block size of 64MB. I would like to change it for my cluster to 256 megabytes since I deal with large files (over 2GB). What is the best way to do this? What file do ...
    Feb 3, 2011 at 12:35 pm
    Feb 8, 2011 at 12:15 am
  • I'm a newbie to hadoop and HDFS. I'm seeing odd behavior in HDFS that I hope somebody can clear up for me. I'm running hadoop version 0.20.1+169.127 from the cloudera distro on 4 identical nodes, ...
    Jeffrey BuellJeffrey Buell
    Feb 25, 2011 at 10:53 pm
    Feb 28, 2011 at 5:02 am
  • Hi , My hadoop version is basic on hadoop 0.20.2 realase, patched HADOOP-4675,5745,MAPREDUCE-1070,551,1089 (support ganglia31,fairscheduler preemption,hdfs append), and patched ...
    Jameson LiJameson Li
    Feb 14, 2011 at 9:04 am
    Feb 16, 2011 at 10:02 am
  • Hi All, I have run the hadoop 0.20 append branch . Can someone please clarify the following behavior? A writer writing a file but he has not flushed the data and not closed the file. Could a parallel ...
    Gokulakannan MGokulakannan M
    Feb 10, 2011 at 3:12 pm
    Feb 15, 2011 at 4:21 am
  • Is there a programmatic way to determine if a datanode is down? -- --- Get your facts first, then you can distort them as you please.--
    Feb 15, 2011 at 1:22 pm
    Feb 25, 2011 at 7:19 pm
  • How to copy a file from a HDS to local file system with a JAVA API ? where i can find a documentation and example about it? thanks
    Alessandro BinharaAlessandro Binhara
    Feb 25, 2011 at 1:09 pm
    Feb 25, 2011 at 6:25 pm
  • http://answers.oreilly.com/topic/459-anatomy-of-a-mapreduce-job-run-with-hadoop/ "Computes the input splits for the job. If the splits cannot be computed, because the input paths don’t exist, for ...
    Sean BigdatafunSean Bigdatafun
    Feb 9, 2011 at 9:09 pm
    Feb 9, 2011 at 9:53 pm
  • hello to all i'm new user of hadoop hdfs i configure it like indicated in site on 2 virtual machines when i want to start it, datanode start but the namenode fail to start with the secondary namenode ...
    Khaled Ben BahriKhaled Ben Bahri
    Feb 24, 2011 at 6:53 pm
    Feb 24, 2011 at 10:02 pm
  • I've hit a data curroption problem in a system we were rapidly loading up, and I could really use some pointers on where to look for the root of the problem as well as any possible solutions. I'm ...
    Chris TarnasChris Tarnas
    Feb 19, 2011 at 1:43 am
    Feb 19, 2011 at 8:38 am
  • Hi, My new adding datanodes are working well. But I just use start-dfs.sh in the namenode and hadoop-daemons.sh start datanode in the namenode can start the new adding datanode, I can't use ...
    Jameson LiJameson Li
    Feb 16, 2011 at 10:21 am
    Feb 18, 2011 at 3:38 am
  • While decommissioning nodes I am seeing the following in my namenode logs: 2011-02-11 21:05:16,290 WARN org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Not able to place enough replicas, still ...
    James LittonJames Litton
    Feb 11, 2011 at 9:10 pm
    Feb 17, 2011 at 6:44 pm
  • Anyone have any experience with using the HDFS-RAID module recently introduced by Facebook? I was wondering if anyone has measured the actual storage savings achieved in practice.
    Nathan RutmanNathan Rutman
    Feb 22, 2011 at 12:30 am
    Feb 22, 2011 at 5:14 am
  • Hy Guys, I have 2 TB of data to process on my MSC work, but I share resources with others students and don't have all that space. So I gzipped my files in splits with sizes similars with the block ...
    Charles GonçalvesCharles Gonçalves
    Feb 10, 2011 at 8:58 pm
    Feb 10, 2011 at 11:56 pm
  • In my existing 48-node cluster, the engineer who originally designed it (no longer here) did not specify logical racks in the HDFS configuration, instead leaving everything in "default-rack". Now I ...
    Jonathan DisherJonathan Disher
    Feb 5, 2011 at 1:00 am
    Feb 8, 2011 at 6:47 pm
  • Hi there, I've been having a problem with a pig script, and when I googled the error message, most of the hits I got were for messages/threads on other hadoop components, so I've got a wee hunch that ...
    Kris CowardKris Coward
    Feb 28, 2011 at 7:12 pm
    Feb 28, 2011 at 7:12 pm
  • Hi- We're working in an environment where we'd like to add a new DataNode and quickly begin leveraging it, usually after detection of high load at the rest of our DataNodes. However, most of the ...
    Ben ClayBen Clay
    Feb 24, 2011 at 3:22 pm
    Feb 24, 2011 at 3:22 pm
  • Not sure if this is possible in hdfs dfsadmin tools but is it possible to preserve date/time of the file system attributes when copying to hdfs? -- --- Get your facts first, then you can distort them ...
    Feb 21, 2011 at 8:04 pm
    Feb 21, 2011 at 8:04 pm
  • Hi, I have to upload some terabytes of data that is text files. What would be good option to do so: i) using hadoop fs -put to copy text files directly on hdfs. ii) copying text files as sequence ...
    Mapred LearnMapred Learn
    Feb 17, 2011 at 12:37 am
    Feb 17, 2011 at 12:37 am
  • Hi , Our cluster sometimes is busy, and some of the slave nodes(DT, TT, regionserver and zookeeper.HQuorumPeer on every node) is in high-load state. Today when I see the NN browser ...
    Jameson LiJameson Li
    Feb 15, 2011 at 9:37 am
    Feb 15, 2011 at 9:37 am
  • Hi all, we are using hadoop 0.20.2 with hbase 0.20.6 on about 30 nodes. Our cluster is under heavy write load, causing hbase to do a lot of compactions, which in turn causes many files with many new ...
    Jan LukavskýJan Lukavský
    Feb 11, 2011 at 9:24 am
    Feb 11, 2011 at 9:24 am
  • I am not sure why it can't find the block, If I go to the master I am able to view the file just fine. 11/02/05 10:41:27 INFO hdfs.DFSClient: Could not obtain block blk_-6466695353427601952_10797526 ...
    Felix gaoFelix gao
    Feb 5, 2011 at 7:15 pm
    Feb 5, 2011 at 7:15 pm
  • Are there some undocumented dependencies for this JIRA (https://issues.apache.org/jira/browse/HADOOP-4584) that it was not added to a release until 0.21.0? Could the patches be applied to a 0.20.2 ...
    Feb 2, 2011 at 3:25 am
    Feb 2, 2011 at 3:25 am
Group Navigation
period‹ prev | Feb 2011 | next ›
Group Overview
grouphdfs-user @

41 users for February 2011

Jameson Li: 7 posts Rita: 7 posts Allen Wittenauer: 6 posts Ayon Sinha: 6 posts Bharath Mundlapudi: 5 posts Gokulakannan M: 5 posts Todd Lipcon: 5 posts Jeffrey Buell: 3 posts John Buchanan: 3 posts Nathan Rutman: 3 posts Patrick Angeles: 3 posts Sean Bigdatafun: 3 posts Stuart Smith: 3 posts Adam Phelps: 2 posts Charles Gonçalves: 2 posts Chris Tarnas: 2 posts Dhodapkar, Chinmay: 2 posts Dhruba Borthakur: 2 posts Felix gao: 2 posts Harsh J: 2 posts
show more