<div dir="ltr"><div>Hi<br><br></div>Yes, results with --enable-munmap are similar but some preliminary benchmarks for the kind of allocations we do, seem to indicate degradation in performance as compared to unmap disabled. While you are right and ideally it would be great to unmap all the memory as and when it is not needed, if there is a performance tradeoff then it may not be feasible. Our aim as of now is to ensure that we cause minimal memory fragmentation (thereby reducing our virtual memory footprint) and still get as good a performance as we can. Selectively unmapping only huge allocations seemed to be a compromise between the two extremes.<br>
<br>--<br>Abhishek<br><div class="gmail_extra"><br><br><div class="gmail_quote">On Sat, May 4, 2013 at 10:14 PM, Jason Evans <span dir="ltr"><<a href="mailto:jasone@canonware.com" target="_blank">jasone@canonware.com</a>></span> wrote:<br>
<blockquote class="gmail_quote" style="margin:0px 0px 0px 0.8ex;border-left:1px solid rgb(204,204,204);padding-left:1ex"><div class=""><div class="h5">On Apr 28, 2013, at 9:03 PM, Abhishek Singh <<a href="mailto:abhishek@abhishek-singh.com">abhishek@abhishek-singh.com</a>> wrote:<br>
> We are trying to replace glibc malloc with jemalloc because we have several concurrent allocations and in all our benchmarks jemalloc is consistently better than glibc malloc and many others.<br>
><br>
> Our setups start typically with 96 GB of RAM and up. We have observed that using jemalloc the virtual memory usage of our process rises up to around 75GB. While the resident memory stays low and it is not a problem as such, when we try to fork a process from within here, it fails as the kernel assumes there is not enough memory to copy the VM space. Perhaps a vfork would be better but we can't use that for now.<br>
><br>
> So we have made some modifications to jemalloc so that all huge memory allocations are forced to unmap their memory on freeing up. Non huge memory allocations and freeing up remain the same. This seems to help us. I have attached the patch here which is against jemalloc-3.3.1. Please review and suggest if there is a better way to handle this.<br>
<br>
</div></div>Does --enable-munmap give you similar results? Ideally, munmap() would always be enabled, but Linux has some unfortunate VM map fragmentation issues (it doesn't consistently reuse holes left by munmap()). I don't think there's much benefit to using munmap() only selectively, because avoiding the VM map fragmentation issue is an all-or-nothing proposition.<br>
<br>
At times I've considered adding pre-fork() code that unmaps everything possible, and purges unused dirty pages. The problem with this is that it's expensive, and it doesn't pay off if the process does exec() right after fork() -- the common case.<br>
<span class=""><font color="#888888"><br>
Jason</font></span></blockquote></div></div></div>