[systemd-devel] Problems with systemd-coredump

Manuel Reimer Manuel.Spam at nurfuerspam.de
Sat Mar 1 00:42:41 PST 2014


On 02/18/2014 11:05 AM, Thomas Bächler wrote:
> Am 17.02.2014 21:27, schrieb Manuel Reimer:
>> As soon as a bigger coredump (about 500 MB) is to be stored, the whole
>> system slows down significantly. Seems like storing such big amounts of
>> data takes pretty long and is a very CPU hungry process...
>
> I completely agree. Since the kernel ignores the maximum coredump size
> when core_pattern is used, a significant amount of time passes whenever
> a larger process crashes, with no benefit (since the dump never gets
> saved anywhere).
>
> This is extremely annoying if processes with sizes in the tens or
> hundreds of gigabytes crash, which sadly happened to me quite a few
> times recently.

If this feature is broken by design, why is it still enabled by default 
on Arch Linux? systemd-coredump makes it nearly impossible to debug 
bigger processes and it took me quite some time to figure out how to get 
coredumps placed to /var/tmp so I can use them to find out why my 
process has crashed.

Yours

Manuel



More information about the systemd-devel mailing list