Another Out of Heap Space Thread
I am trying to analyze about 400,000 records w/ 5 attributes on Decision Tree learner. I end up running out of heap space. I am currently using the xms/xmx2048m option. My system has 4 GB of memory, but running on a 10 year old hard drive so it really slows down if I have to use the swap space (Running Ubuntu Linux 10.04). Aside from upgrading my RAM and killing some processes to free up 3 GB for the heap, is there anything else that I can do besides upgrading my system memory? Is there a way to make the operator conserve memory?