[khmer] exceeding defined RAM limits?

C. Titus Brown ctb at msu.edu
Tue Dec 17 17:14:09 PST 2013


On Tue, Dec 17, 2013 at 04:36:34PM -0800, C. Titus Brown wrote:
> On Tue, Dec 17, 2013 at 07:53:18PM +0000, Oh, Julia (NIH/NHGRI) [F] wrote:
> Now, on to your real question :)
> 
> > $python2.7 /home/ohjs/khmer/scripts/normalize-by-median.py -C 5 -k 20 -N 4 -x 16e9 round2.unaligned.keep.abundfilt; 
> > 
> > I thought I would be maxing out at 64 GB ram for the hash table (I?ve also used 32e9), but I get the following RAM usage report of 
> > 
> > 4986693.biobos elapsed time:        23358 seconds
> > 4986693.biobos walltime:         06:28:36 hh:mm:ss
> > 4986693.biobos memory limit:       249.00 GB
> > 4986693.biobos memory used:        249.76 GB
> > 4986693.biobos cpupercent used:     98.00 %
> 
> What the heck!? That's not supposed to happen!
> 
> This is either a bug, or (most likely) is being caused by an overabundance of
> high-abundance k-mers.  The latter is easy to fix -- I've filed a bug report to
> fix the latter in the software overall [0] -- but would require you to modify
> the script at the moment.  If you're up for that, put
> 
> 	ht.set_use_bigcount(False)
> 
> at line 186 of normalize-by-median:

Darn it, that can't be the problem; I just wrote a test against this
behavior and we actually did things right in the script and ignored
high abundance k-mers.

So, this must be a bug of some sort.  Umm... Michael, any ideas?!

cheers,
--titus
-- 
C. Titus Brown, ctb at msu.edu




More information about the khmer mailing list