hadoop/contrib/index，but I found that when the input files become big（such
as one file is 16G）,the OOM exception will be throwed .The cause is that: in
combiner ,"writer.addIndexNoOptimize()",this use much memory cause to OOM,
it's the Lucene OOM insead of the MapReduce OOM, but I hope create a new
method like the "spill" to solve this problem ,how can I do? My English is
View this message in context: http://lucene.472066.n3.nabble.com/mapreduce-combiner-tp3612513p3612513.html
Sent from the Hadoop lucene-dev mailing list archive at Nabble.com.