FAQ

Search Discussions

285 discussions - 824 posts

  • Hi, we like to monitor our cluster with ganglia. On a vanilla installation it's a easy job, - just put the hadoop-metrics.properties into the conf folder, start the gmond and gmetad and you are ...
    Christian SchneiderChristian Schneider
    Mar 6, 2013 at 3:58 pm
    Jun 24, 2013 at 4:51 pm
  • Hi, I'm trying to parallelize actions in oozie using fork. I'm able to run upto 3 parallel jobs and anything beyond that, there seems to be a deadlock. I'm using CDH 4.1.2 and it's a 2 node cluster ...
    CDHUserCDHUser
    Mar 28, 2013 at 4:40 pm
    Dec 14, 2013 at 3:12 pm
  • Hi, We have a flume-ng(hadoop pointed to cdh4) setup which writes to hdfs sink using .gz compression. There is a different oozie setup(hadoop pointed to cdh3u2) which tries read these .gz files and ...
    SundiSundi
    Mar 22, 2013 at 5:13 pm
    Mar 27, 2013 at 4:59 pm
  • I've got a recently setup Hadoop cluster with CDH4.2. I'm trying to use Hiveserver2 with beeline, and I'm running into issues where it won't run in distributed mode. Snippet from yarn-site.xml ...
    Matt GettyMatt Getty
    Mar 4, 2013 at 11:47 pm
    Mar 7, 2013 at 5:03 pm
  • We have a number of data nodes that are reporting about 10% higher disk utilization for HDFS than the rest of our nodes. These nodes do have one thing in common, they all recently had bad disks ...
    M shirleyM shirley
    Mar 25, 2013 at 4:08 pm
    Oct 26, 2013 at 7:44 am
  • I am attempting to run a MapReduce job, and after a period of time it fails with the above message. As you can see it doesn't get anywhere with either the map or the reduce. According to the manager, ...
    Chuong NguyenChuong Nguyen
    Mar 16, 2013 at 2:17 am
    Mar 20, 2013 at 6:46 pm
  • Hi, I am trying to run a simple pig script that uses HbaseStorage class to load data from a hbase table. The pig script runs perfectly fine when run standalone in mapreduce mode. But when i submit it ...
    Praveen BysaniPraveen Bysani
    Mar 14, 2013 at 9:30 am
    Mar 20, 2013 at 6:10 am
  • Please see details http://stackoverflow.com/questions/13388595/org-apache-hadoop-mapred-shufflehandler-not-found-on-starting-nodemanager-cdh4 , hadoop-mapreduce is not installed automatically when ...
    Hui HuHui Hu
    Mar 26, 2013 at 6:51 am
    Oct 22, 2013 at 9:58 pm
  • I'm trying to build a workflow that will do incremental imports from a MySQL database on a scheduled basis. I've successfully run the sqoop job from the command line several times, but it ...
    Jc ListmailJc Listmail
    Mar 25, 2013 at 5:49 pm
    Sep 12, 2013 at 6:36 am
  • We've been using Snappy for almost all of our compression needs but I recently read that snappy isn't good to store permanent files because of its lack of splittability. Which codec should be used in ...
    StaticVoidStaticVoid
    Mar 25, 2013 at 7:41 pm
    Apr 24, 2013 at 5:55 pm
  • Hi, I am working on CDH4 install. We do not have a MySQL or a PostGreSQL to work with Hive. How can I allow multiple users to share the same space in derby db? I have tried giving 777 permission ...
    Sumit GhoshSumit Ghosh
    Mar 7, 2013 at 7:18 am
    Mar 8, 2013 at 12:36 am
  • Hey all, We're exploring Hadoop as a replacement for an existing log storage (and processing) solution and there's one sticky point we've come up against: the amount of additional storage capacity ...
    Some GuySome Guy
    Mar 18, 2013 at 5:18 pm
    Apr 5, 2013 at 10:57 pm
  • I am seeing something peculiar using hive-server2 in CDH4.2 managed by CM4.5 when it comes to external tables and partitions. I created an external hive table based on the avro data format. I would ...
    Benjamin KimBenjamin Kim
    Mar 21, 2013 at 3:56 am
    Mar 26, 2013 at 8:20 pm
  • When i run oozie example, ./oozie job -oozie http://localhost:11000/oozie -config ../examples/apps/map-reduce/job.properties -run I am getting the error as Error: E0501 : E0501: Could not perform ...
    Sandeep KurdagiSandeep Kurdagi
    Mar 12, 2013 at 6:34 pm
    Sep 18, 2013 at 10:59 am
  • I have a Pig workflow that works perfectly but when I try to add it as a subworkflow I am coming across a few problems (bugs?) First off. In the subworkflow page I can never seem to uncheck ...
    StaticVoidStaticVoid
    Mar 26, 2013 at 6:48 pm
    Apr 1, 2013 at 5:59 pm
  • Hi we are facing issues with CDH4.1.2 version while configuring namenode HA using QJM, I just want to confirm with you folks if the following issue is the root cause of the problem, ...
    Raghu DoppalapudiRaghu Doppalapudi
    Mar 29, 2013 at 5:51 pm
    Mar 30, 2013 at 3:49 am
  • I write to flume directly as my data comes in at a rate of 1000's per second. It in turn is responsible for aggregating and writing to HDFS via the event sink. What happens if it is not able to reach ...
    Gary MaloufGary Malouf
    Mar 26, 2013 at 8:02 pm
    Mar 27, 2013 at 9:08 pm
  • Hi, I'm running a fresh install of CentOS 6.2 with CDH4.2 installed via Cloudera Manager 4.5. I'm getting the following errors when I try to compile. Is there something simple I'm overlooking?? ...
    Sameer FarooquiSameer Farooqui
    Mar 27, 2013 at 7:26 am
    Mar 27, 2013 at 10:40 am
  • Of course I discovered this after it became a problem, and now am not sure what to do next without losing data/metadata. Background: I just added two nodes to a CDH 4.1.3 cluster. The new nodes got ...
    Dan BongertDan Bongert
    Mar 20, 2013 at 7:03 pm
    Mar 26, 2013 at 5:14 pm
  • Hello. I'm Using Teiid 8.1 Jboss 7.1.1 Windows 7 And Clouderra VM https://ccp.cloudera.com/display/SUPPORT/Demo+VMs I installed driver hive-jdbc-0.10.0-cdh4.2.0.jar I copied it from Clouderra Hive ...
    Eugene KislyEugene Kisly
    Mar 15, 2013 at 1:55 pm
    Mar 18, 2013 at 4:44 am
  • Hi, We are using CDH 4.1.2 HDFS 2.0.0 and were previously using the FileSystem API. Calls like FileSystem.listStatus(new Path("hdfs://tmp/")) were successful. We are looking to use FIleContext to ...
    Rahul RavindranRahul Ravindran
    Mar 1, 2013 at 11:04 pm
    Mar 7, 2013 at 11:21 pm
  • Hi All, I am planing to use flume in one of the POC project . I am new to flume Do you have any supported doc/link/example from where i will get all the context ASAP. Regards, samir. --
    Samir das mohapatraSamir das mohapatra
    Mar 1, 2013 at 7:16 pm
    Mar 4, 2013 at 6:43 pm
  • Hi. What are approaches for Hive unit testing? I would like to: 1. Test generated SQL: is it valid? 2. Execute generated SQL using small subset of sample data: does query output matches with etalon ...
    Serega SheypakSerega Sheypak
    Mar 12, 2013 at 3:12 pm
    Oct 14, 2013 at 10:01 pm
  • We recently upgraded to CDH4.2 from 3u5. We are running all of the same jobs and the configs all remained the same between versions. However, we are seeing memory issues now that we never saw ...
    Bryan BeaudreaultBryan Beaudreault
    Mar 11, 2013 at 5:23 pm
    Jun 21, 2013 at 9:17 pm
  • SEVERE: The web application [/oozie] created a ThreadLocal with key of type [java.lang.ThreadLocal] (value [java.lang.ThreadLocal@7102be12]) and a value of type ...
    Vishal rajanVishal rajan
    Mar 25, 2013 at 8:34 am
    May 23, 2013 at 12:42 am
  • I currently have Flume configured to roll files when they reach the minimum hdfs block size. The timestamps on the files represent the time the file was created. I was wondering if anyone has found a ...
    Gary MaloufGary Malouf
    Mar 26, 2013 at 12:46 am
    May 8, 2013 at 1:52 pm
  • Hi java.lang.IllegalStateException: Socket Socket[addr=/10.86.203.112,port=1004,localport=35170] does not have a channel at com.google.common.base.Preconditions.checkState(Preconditions.java:172) at ...
    Subroto SanyalSubroto Sanyal
    Mar 5, 2013 at 2:54 pm
    Apr 30, 2013 at 9:31 pm
  • Hi All I am running a query through Hive and looks like when my task count is job never completes There are total 2310 map task in my job and every time I run it fail after 150 jobs makes me think ...
    DKDK
    Mar 8, 2013 at 8:26 am
    Apr 16, 2013 at 2:53 am
  • Hi, we are using Flume 1.3 We did decide to use such conf for our HDFS sink: http_landing.sinks.http_landing_hdfs_sink2.hdfs.fileType = SequenceFile ...
    Serega SheypakSerega Sheypak
    Mar 28, 2013 at 7:46 am
    Apr 8, 2013 at 6:07 pm
  • Hi, Can we get HDFS-4581 into cdh3u6 update? The issue fixes Datanode not to do check disk when the exceptions are network related. We would like get into cdh3 updates, as we have many clusters ...
    Amareshwari sriramdasuAmareshwari sriramdasu
    Mar 19, 2013 at 4:54 am
    Apr 4, 2013 at 7:27 pm
  • Hi, My HDFS default block size in hdfs-site.xml is 128 MB, but while keepings those defaults intact, I want to write a new file and override the default block size for just this one file to something ...
    Sameer FarooquiSameer Farooqui
    Mar 31, 2013 at 9:36 pm
    Apr 1, 2013 at 7:16 pm
  • How can i fix thios issue. 2013-03-22 16:39:11,571 INFO ActionStartCommand:525 - USER[tes-test-org] GROUP[users] TOKEN[] APP[org-ingestion-wf] JOB[0000379-130322160345207-oozie-oozi-W] ...
    Vishal rajanVishal rajan
    Mar 23, 2013 at 7:18 pm
    Mar 24, 2013 at 5:44 am
  • Hi guys, In my cluster, I removed (via sudo -u hdfs hadoop fs -rm) a ton of files from HDFS, but afterwards, my DFS Used% remained unchanged. Is this normal? Is there a periodic "cleanup" routine ...
    Chris NealChris Neal
    Mar 21, 2013 at 9:08 pm
    Mar 23, 2013 at 5:46 am
  • Hi all, I get an exception when I try to connect to a remote cluster from my client using: hadoop --config <downloaded-client-config-files fs -put <file <hdfs-path Exception in ...
    Björn JónssonBjörn Jónsson
    Mar 21, 2013 at 4:07 am
    Mar 23, 2013 at 4:42 am
  • Hi all, I'm trying to copy data from a 3u5 cluster to a new 4.2 cluster for our HBase but I'm having serious difficulty trying to run the destination. Right now broadly I'm: Shutting down src HBase, ...
    Steph GoslingSteph Gosling
    Mar 16, 2013 at 8:30 pm
    Mar 19, 2013 at 4:14 pm
  • Hi All, How necessary to have HA for MapReduceV1? Since the M/R job/JobTracker is re-startable. Thanks in advance --
    Alicia LeongAlicia Leong
    Mar 18, 2013 at 11:49 pm
    Mar 19, 2013 at 1:05 am
  • I am trying to install oracle jdbc drivers for sqoop. I understand for the sqoop CLI to work the driver jar has to be placed in /usr/lib/sqoop/lib/. For the oozie sqoop action this may not suffice as ...
    CDHUserCDHUser
    Mar 12, 2013 at 12:03 am
    Mar 16, 2013 at 2:14 am
  • When we restart our node HBase region server does not start the logs says log4j:ERROR setFile(null,true) call failed. java.io.FileNotFoundException ...
    BencBenc
    Mar 14, 2013 at 9:27 am
    Mar 14, 2013 at 3:44 pm
  • Hi , I’m building Hadoop environment and everything is OK. I’m stuck to find formal document/reference that explains some parameters in flume: TwitterAgent.sources.Twitter.consumerKey = <required ...
    Ammar FayoumiAmmar Fayoumi
    Mar 10, 2013 at 3:19 pm
    Mar 11, 2013 at 4:14 pm
  • I keep running into this error on some new nodes that I added to the cluster. It's odd because the machines all share their config files from nfs shares so they are all using the exact same configs ...
    Jason KingJason King
    Mar 4, 2013 at 9:10 pm
    Mar 8, 2013 at 5:39 pm
  • Hi, I'm having an issue using the "--options-file" option in Oozie Sqoop action. I'm able to use this when I run sqoop from command line. The same doesn't work from oozie and I get the following ...
    CDHUserCDHUser
    Mar 4, 2013 at 7:29 pm
    Mar 5, 2013 at 12:57 am
  • Hi, after our upgrade to CDH4.2 the Beeswax Server does not start anymore. The HUE Webinterface says: Could not connect to michaela:8002 If i try to restart the server by CM4.5 it says: Command ...
    Christian SchneiderChristian Schneider
    Mar 1, 2013 at 2:08 pm
    Mar 4, 2013 at 8:32 am
  • Hi Guys, We have configured ulimit value as 65536. When I check with hdfs user and mapred user process not at all taking this value why? It's taking Max open files 32768 How to change this value to ...
    Dhanasekaran AnbalaganDhanasekaran Anbalagan
    Mar 17, 2013 at 12:27 pm
    Jun 28, 2013 at 6:52 am
  • Dear all, I want to move my virtual cluster to another building where the IP address of network is different. I would be install and configure again the CDH or there is any other easy way. Thanks ...
    Mukhtaj KhanMukhtaj Khan
    Mar 17, 2013 at 1:55 pm
    Jun 19, 2013 at 5:01 pm
  • Hi All Currently we have 200 nodes hadoop cluster (CDH3U4)using 2 replication factor and two racks( rack 0 and rack 1) with rack awareness . we would like to implement another rack(rack 2) on ...
    MohamedMohamed
    Mar 27, 2013 at 4:21 pm
    Apr 4, 2013 at 8:46 pm
  • Hi, I was wondering if someone has any recommendation or available third party products for securing data at rest in Hadoop. I see it can be secured through encryption compression codec but ...
    Deepak TiwariDeepak Tiwari
    Mar 31, 2013 at 6:00 am
    Apr 4, 2013 at 12:31 am
  • Hello, Sorry for the novice question, but reading the documentations, I was not able to understand the difference between Beeswax and Cloudera Impala. They both seem to be doing the same thing. Then ...
    Panshul WhisperPanshul Whisper
    Mar 28, 2013 at 10:58 am
    Mar 31, 2013 at 9:47 pm
  • We are running two namenodes in HA Configuration, with 3 journalnodes. Recently one of the namenodes died, and failover worked fine. But when we tried to bring the dead namenode back up, we ran into ...
    Bryan BeaudreaultBryan Beaudreault
    Mar 27, 2013 at 4:08 pm
    Mar 31, 2013 at 8:26 pm
  • Hi, I downloaded CDH 4.2.0 tarball and its missing container-executor is missing in the bin folder, can we use the one from 4.1.2 without any problem, please comment. --
    Raghu HbaseRaghu Hbase
    Mar 29, 2013 at 11:42 pm
    Mar 30, 2013 at 3:51 am
  • After installing a cluster via the Cloudera Manager is it necessary to bump up system, user or group limits in /etc/security/limits.conf or will these processes take care of themselves? If not, which ...
    StaticVoidStaticVoid
    Mar 28, 2013 at 4:10 pm
    Mar 29, 2013 at 4:25 pm
Group Navigation
period‹ prev | Mar 2013 | next ›
Group Overview
groupcdh-user @
categorieshadoop
discussions285
posts824
users209
websitecloudera.com
irc#hadoop

209 users for March 2013

Harsh J: 71 posts Samir das mohapatra: 24 posts Dhanasekaran Anbalagan: 20 posts Serega Sheypak: 19 posts Jarek Jarcec Cecho: 18 posts Benjamin Kim: 16 posts Mike Percy: 16 posts StaticVoid: 16 posts Alejandro Abdelnur: 15 posts Praveen Bysani: 15 posts Srikraj8341: 14 posts Sumit Ghosh: 14 posts Sundi: 14 posts CDHUser: 13 posts Robert Kanter: 13 posts Sandy Ryza: 12 posts Brock Noland: 10 posts Matt Getty: 10 posts Aaron T. Myers: 9 posts Ayman shalaby: 9 posts
show more