" CM auto-configured your heaps to 62Mb when the default is 1 gig, then you
are trying to run too many things on one host"
yes, I have one host where all services are running on it and the
auto-configured memory heap was 62Mb before I changed it to 1gig
"The socket timeout exception doesn't really indicate the real problem.
There's got to be something more interesting in the role logs, stderr, or
stdout, but it's likely related to memory issues."
The log is too long , however it keeps repeating the following lines :
Error getting HDFS summary for hdfs: org.apache.avro.AvroRemoteException:
java.net.SocketTimeoutException: connect timed out
WARN [767509864@scm-web-40092:tsquery.TimeSeriesQueryService@503]
com.cloudera.server.cmf.tsquery.TimeSeriesQueryService@2fb39e9b failed on
nozzle HOST_MONITORING
java.util.concurrent.TimeoutException
at java.util.concurrent.FutureTask.get(FutureTask.java:201)
at
com.cloudera.server.cmf.tsquery.NozzleRequest.getResponse(NozzleRequest.java:70)
at
com.cloudera.server.cmf.tsquery.TimeSeriesQueryService.queryTimeSeries(TimeSeriesQueryService.java:307)
at
com.cloudera.server.web.cmf.charts.TimeSeriesQueryController.queryTimeSeriesHelper(TimeSeriesQueryController.java:310)
at
com.cloudera.server.web.cmf.charts.TimeSeriesQueryController.queryTimeSeries(TimeSeriesQueryController.java:271)
at sun.reflect.GeneratedMethodAccessor1159.invoke(Unknown Source)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
"If you don't have any data on your cluster yet anyway, it's probably
easiest to delete your cluster (and management service) and then re-create
it with a smaller set of services, allowing CM to auto-configure memory to
be a bit better. Alternatively you could delete services / roles and
manually modify heaps to be more reasonable."
Since I have a single host , Can I delete the cluster without deleting CM ?
Can you please point me to how to safely delete it?
"I'm a little confused by your email since you are talking about HDFS roles
and "mapred.java.child.opts", which only applies to MapReduce. It's
possible that you didn't correctly revert everything to where you had it
before if you're getting these mixed up. Still, removing services is likely
the best option for you."
I think changing "mapred.java.child.opts" propperty of mapred-site.xml is
by the configuration panel as I already did rather than
modify the file at conf folder , Am I right or they are different ? correct
me please If I am not right.
On Thu, May 8, 2014 at 6:31 PM, Darren Lo wrote:Hi Ghadeer,
If CM auto-configured your heaps to 62Mb when the default is 1 gig, then
you are trying to run too many things on one host. You should probably not
select the "All Services" option on such a small machine, especially if
it's a single-host cluster.
The socket timeout exception doesn't really indicate the real problem.
There's got to be something more interesting in the role logs, stderr, or
stdout, but it's likely related to memory issues.
If you don't have any data on your cluster yet anyway, it's probably
easiest to delete your cluster (and management service) and then re-create
it with a smaller set of services, allowing CM to auto-configure memory to
be a bit better. Alternatively you could delete services / roles and
manually modify heaps to be more reasonable.
I'm a little confused by your email since you are talking about HDFS roles
and "mapred.java.child.opts", which only applies to MapReduce. It's
possible that you didn't correctly revert everything to where you had it
before if you're getting these mixed up. Still, removing services is likely
the best option for you.
Thanks,
Darren
On Thu, May 8, 2014 at 2:18 AM, Ghadeer wrote:
Hello everybody ,
I need your help please to solve the following problem I encountered :
I was trying a Hadoop process on CDH 4 , when I got " OutOfMemoryErorr"
: java heap space exception . I have 8 G of RAM on my machine , so I
assumed I should use them to increase the heap space allocated for java on
namenode and datanode " mapred.java.child.opts" property using cloudera
manager .
I did modify the "Java heap size of data node in bytes" at configuration
window of HDFS and I did the same for name node and secondary node. I
modify it to the default value 1 GB where it was about 60MB before.
The HDFS and MAPREDUCE services have unexpectedly stopped when I was
trying to run that Hadoop process. Thus I try to change the java heap size
on nodes back to 62 Mb and tried to start the services. unfortunately they
did not start where the details error was : failed to start its roles
(datanode, namenode ) , "command a time out after 150 seconds"
And for the whole machine I have "memory overcommit" , so why did the
changes do not take place when I change it back to 62 MB and how can I
start the services as it was before?
For the log it has the following :
Error getting HDFS summary for hdfs: org.apache.avro.AvroRemoteException:
java.net.SocketTimeoutException: Read timed out
thanks in advance,
To unsubscribe from this group and stop receiving emails from it, send an
email to scm-users+unsubscribe@cloudera.org.
To unsubscribe from this group and stop receiving emails from it, send an email to scm-users+unsubscribe@cloudera.org.