Hi guys :
I have a map/r job that has always worked fine, but which fails due to a
heap space error on my local machine during unit tests.
It runs in hadoop's default mode, and just fails durring the constructor of
the MapOutputBuffer.... Any thoughts on why ?
I dont do any custom memory settings in by unit tests, because they aren't
really needed --- So I assume this is related to /tmp files
or something ... but cant track down the issue.
Any thoughts would be very much appreciated ..
12/05/01 19:15:53 WARN mapred.LocalJobRunner: job_local_0002
java.lang.OutOfMemoryError: Java heap space
Harsh J 2012-05-02, 04:36