You can subscribe to this list here.
2003 |
Jan
|
Feb
|
Mar
(58) |
Apr
(261) |
May
(169) |
Jun
(214) |
Jul
(201) |
Aug
(219) |
Sep
(198) |
Oct
(203) |
Nov
(241) |
Dec
(94) |
---|---|---|---|---|---|---|---|---|---|---|---|---|
2004 |
Jan
(137) |
Feb
(149) |
Mar
(150) |
Apr
(193) |
May
(95) |
Jun
(173) |
Jul
(137) |
Aug
(236) |
Sep
(157) |
Oct
(150) |
Nov
(136) |
Dec
(90) |
2005 |
Jan
(139) |
Feb
(130) |
Mar
(274) |
Apr
(138) |
May
(184) |
Jun
(152) |
Jul
(261) |
Aug
(409) |
Sep
(239) |
Oct
(241) |
Nov
(260) |
Dec
(137) |
2006 |
Jan
(191) |
Feb
(142) |
Mar
(169) |
Apr
(75) |
May
(141) |
Jun
(169) |
Jul
(131) |
Aug
(141) |
Sep
(192) |
Oct
(176) |
Nov
(142) |
Dec
(95) |
2007 |
Jan
(98) |
Feb
(120) |
Mar
(93) |
Apr
(96) |
May
(95) |
Jun
(65) |
Jul
(62) |
Aug
(56) |
Sep
(53) |
Oct
(95) |
Nov
(106) |
Dec
(87) |
2008 |
Jan
(58) |
Feb
(149) |
Mar
(175) |
Apr
(110) |
May
(106) |
Jun
(72) |
Jul
(55) |
Aug
(89) |
Sep
(26) |
Oct
(96) |
Nov
(83) |
Dec
(93) |
2009 |
Jan
(97) |
Feb
(106) |
Mar
(74) |
Apr
(64) |
May
(115) |
Jun
(83) |
Jul
(137) |
Aug
(103) |
Sep
(56) |
Oct
(59) |
Nov
(61) |
Dec
(37) |
2010 |
Jan
(94) |
Feb
(71) |
Mar
(53) |
Apr
(105) |
May
(79) |
Jun
(111) |
Jul
(110) |
Aug
(81) |
Sep
(50) |
Oct
(82) |
Nov
(49) |
Dec
(21) |
2011 |
Jan
(87) |
Feb
(105) |
Mar
(108) |
Apr
(99) |
May
(91) |
Jun
(94) |
Jul
(114) |
Aug
(77) |
Sep
(58) |
Oct
(58) |
Nov
(131) |
Dec
(62) |
2012 |
Jan
(76) |
Feb
(93) |
Mar
(68) |
Apr
(95) |
May
(62) |
Jun
(109) |
Jul
(90) |
Aug
(87) |
Sep
(49) |
Oct
(54) |
Nov
(66) |
Dec
(84) |
2013 |
Jan
(67) |
Feb
(52) |
Mar
(93) |
Apr
(65) |
May
(33) |
Jun
(34) |
Jul
(52) |
Aug
(42) |
Sep
(52) |
Oct
(48) |
Nov
(66) |
Dec
(14) |
2014 |
Jan
(66) |
Feb
(51) |
Mar
(34) |
Apr
(47) |
May
(58) |
Jun
(27) |
Jul
(52) |
Aug
(41) |
Sep
(78) |
Oct
(30) |
Nov
(28) |
Dec
(26) |
2015 |
Jan
(41) |
Feb
(42) |
Mar
(20) |
Apr
(73) |
May
(31) |
Jun
(48) |
Jul
(23) |
Aug
(55) |
Sep
(36) |
Oct
(47) |
Nov
(48) |
Dec
(41) |
2016 |
Jan
(32) |
Feb
(34) |
Mar
(33) |
Apr
(22) |
May
(14) |
Jun
(31) |
Jul
(29) |
Aug
(41) |
Sep
(17) |
Oct
(27) |
Nov
(38) |
Dec
(28) |
2017 |
Jan
(28) |
Feb
(30) |
Mar
(16) |
Apr
(9) |
May
(27) |
Jun
(57) |
Jul
(28) |
Aug
(43) |
Sep
(31) |
Oct
(20) |
Nov
(24) |
Dec
(18) |
2018 |
Jan
(34) |
Feb
(50) |
Mar
(18) |
Apr
(26) |
May
(13) |
Jun
(31) |
Jul
(13) |
Aug
(11) |
Sep
(15) |
Oct
(12) |
Nov
(18) |
Dec
(13) |
2019 |
Jan
(12) |
Feb
(29) |
Mar
(51) |
Apr
(22) |
May
(13) |
Jun
(20) |
Jul
(13) |
Aug
(12) |
Sep
(21) |
Oct
(6) |
Nov
(9) |
Dec
(5) |
2020 |
Jan
(13) |
Feb
(5) |
Mar
(25) |
Apr
(4) |
May
(40) |
Jun
(27) |
Jul
(5) |
Aug
(17) |
Sep
(21) |
Oct
(1) |
Nov
(5) |
Dec
(15) |
2021 |
Jan
(28) |
Feb
(6) |
Mar
(11) |
Apr
(5) |
May
(7) |
Jun
(8) |
Jul
(5) |
Aug
(5) |
Sep
(11) |
Oct
(9) |
Nov
(10) |
Dec
(12) |
2022 |
Jan
(7) |
Feb
(13) |
Mar
(8) |
Apr
(7) |
May
(12) |
Jun
(27) |
Jul
(14) |
Aug
(27) |
Sep
(27) |
Oct
(17) |
Nov
(17) |
Dec
|
2023 |
Jan
(10) |
Feb
(18) |
Mar
(9) |
Apr
(26) |
May
|
Jun
(13) |
Jul
(18) |
Aug
(5) |
Sep
(12) |
Oct
(16) |
Nov
(1) |
Dec
|
2024 |
Jan
(4) |
Feb
(3) |
Mar
(6) |
Apr
(17) |
May
(2) |
Jun
(33) |
Jul
(13) |
Aug
(1) |
Sep
(6) |
Oct
(8) |
Nov
(6) |
Dec
(15) |
2025 |
Jan
(5) |
Feb
(11) |
Mar
(8) |
Apr
(20) |
May
(1) |
Jun
|
Jul
|
Aug
(2) |
Sep
|
Oct
|
Nov
|
Dec
|
From: Philippe W. <phi...@sk...> - 2022-08-31 22:14:27
|
On Wed, 2022-08-31 at 17:42 +0000, Bresalier, Rob (Nokia - US/Murray Hill) wrote: > > When running memcheck on a massive monolith embedded executable > > (237MB stripped, 1.8GiB unstripped), after I stop the executable under > > valgrind I see the "HEAP SUMMARY" but then valgrind dies before any leak > > reports are printed. The parent process sees that the return status of > > memcheck is that it was SIGKILLed (status returned in waitpid call is '9'). > > We found that removing a call to _exit(0) made it so that valgrind is no longer > SIGKILLED. > > Any ideas why using _exit(0) may get rid of valgrind getting SIGKILLed? > > Previously exit(0) was called, without the leading underscore, but changed it to > _exit(0) to really make sure no memory was being deallocated. This worked well on a > different process, so we carried it over to this one, that is why we did it. > > Even with exit(0) (no underscore), in this process there is not much deallocation going > on in exit handlers, so have lots of doubts that valgrind/memcheck was using too much > memory and invoking the OOM killer. > > Using strace and dmesg while we had _exit(0) in use didn't show that OOM killer was > SIGKILLing valgrind. > > I also tried reducing number of callers from 12 to 6 when using _exit(0), still got the > SIGKILL. > > Also tried using a system that had an additional 4GByte of memory, and also got the > SIGKILL there. > > So I have many doubts that Valgrind was getting SIGKILLed due to too much memory usage. > > Don't know why removing _exit(0) got rid of the SIGKILL. Was wondering if anyone had any > ideas? Normally, if it is the OOM that kills a process, you should find a trace of this in the system logs. I do not understand what you mean by reducing the nr of callers from 12 to 6. What are these callers ? Is that some threads of the process you are running under valgrind ? And just in case: are you using the last version of Valgrind ? You might use "strace" on valgrind to see what is going on at the time _exit(0) is called. You might also start valgrind with some debug trace e.g. -d -d -d -d -v -v -v -v Philippe |
From: Bresalier, R. (N. - US/M. Hill) <rob...@no...> - 2022-08-31 19:16:21
|
> When running memcheck on a massive monolith embedded executable > (237MB stripped, 1.8GiB unstripped), after I stop the executable under > valgrind I see the "HEAP SUMMARY" but then valgrind dies before any leak > reports are printed. The parent process sees that the return status of > memcheck is that it was SIGKILLed (status returned in waitpid call is '9'). We found that removing a call to _exit(0) made it so that valgrind is no longer SIGKILLED. Any ideas why using _exit(0) may get rid of valgrind getting SIGKILLed? Previously exit(0) was called, without the leading underscore, but changed it to _exit(0) to really make sure no memory was being deallocated. This worked well on a different process, so we carried it over to this one, that is why we did it. Even with exit(0) (no underscore), in this process there is not much deallocation going on in exit handlers, so have lots of doubts that valgrind/memcheck was using too much memory and invoking the OOM killer. Using strace and dmesg while we had _exit(0) in use didn't show that OOM killer was SIGKILLing valgrind. I also tried reducing number of callers from 12 to 6 when using _exit(0), still got the SIGKILL. Also tried using a system that had an additional 4GByte of memory, and also got the SIGKILL there. So I have many doubts that Valgrind was getting SIGKILLed due to too much memory usage. Don't know why removing _exit(0) got rid of the SIGKILL. Was wondering if anyone had any ideas? |
From: Philippe W. <phi...@sk...> - 2022-08-06 08:35:01
|
> > > Is there anything that can be done with memcheck to make it consume less memory? > > No. In fact, Yes :). Or more precisely, yes, memory can be somewhat reduced :). See my other mail. Philippe |
From: Philippe W. <phi...@sk...> - 2022-08-06 08:32:56
|
On Fri, 2022-08-05 at 15:34 +0000, Bresalier, Rob (Nokia - US/Murray Hill) wrote: > > If finding memory leaks is the only goal (for instance, if you are satisfied that > > memcheck has found all the overrun blocks, uninitialized reads, etc.) then > > https://github.com/KDE/heaptrack is the best tool. > > Thanks! I didn't know about heaptrack. I will look definitely into that. Does heaptrack > also show the 'still reachable' types of leaks that memcheck does? > > Any chance that the 'massif' tool would survive the OOM killer? This may be easier for > me to get going as I already have valgrind built. > > Is there anything that can be done with memcheck to make it consume less memory? You might be interested in looking at the slides of the FOSDEM presentation 'Tuning Valgrind for your workload' https://archive.fosdem.org/2015/schedule/event/valgrind_tuning/attachments/slides/743/export/events/attachments/valgrind_tuning/slides/743/tuning_V_for_your_workload.pdf There are several things you can do to reduce memcheck memory usage. Note also that you can also run leak search while your program runs, either via memcheck client requests or from the shell, using vgdb. Philippe |
From: Julian S. <jse...@gm...> - 2022-08-06 06:43:36
|
> Is there anything that can be done with memcheck to make it consume less memory? First of all, figure out whether memcheck got sigkilled because the machine ran out of space, or because you hit some shell limit/ulimit. In the former case, you can then try adding swap space to the machine. In the latter case you'll need to mess with the shell's ulimit settings. You could also try reducing the (data) size of the workload. Massif and Memcheck are different tools and do largely different things. Whether or not you can use one or the other depends a lot on the specifics of what problem you're trying to solve. J |
From: Eliot M. <mo...@cs...> - 2022-08-06 02:05:51
|
On 8/5/2022 8:47 PM, G N Srinivasa Prasanna wrote: > Thanks for this information. > > We are doing a memory system simulation, and need the address stream. At this point of time, we > don't care if we need a Terabyte even, we can delete the files later. > > Is there anything we can use from Valgrind? The lackey tool does just that - output a trace of memory references. -- Eliot Moss |
From: G N S. P. <gns...@ii...> - 2022-08-06 01:49:20
|
Thanks, will check it out. Best ________________________________ From: Eliot Moss <mo...@cs...> Sent: 06 August 2022 07:10 To: G N Srinivasa Prasanna <gns...@ii...>; John Reiser <jr...@bi...>; val...@li... <val...@li...> Subject: Re: [Valgrind-users] Valgrind trace Memory Addresses while running? On 8/5/2022 8:47 PM, G N Srinivasa Prasanna wrote: > Thanks for this information. > > We are doing a memory system simulation, and need the address stream. At this point of time, we > don't care if we need a Terabyte even, we can delete the files later. > > Is there anything we can use from Valgrind? The lackey tool does just that - output a trace of memory references. -- Eliot Moss |
From: G N S. P. <gns...@ii...> - 2022-08-06 00:47:29
|
Thanks for this information. We are doing a memory system simulation, and need the address stream. At this point of time, we don't care if we need a Terabyte even, we can delete the files later. Is there anything we can use from Valgrind? Best ________________________________ From: John Reiser <jr...@bi...> Sent: 06 August 2022 01:18 To: val...@li... <val...@li...> Subject: Re: [Valgrind-users] Valgrind trace Memory Addresses while running? >> if we can get a list of all the physical addresses the program used, in the order the program accessed them, and whether read/write. > For any real world application the size of the log would be overwhelmingly huge ... (unless you only want unique addresses). Of course this is the purpose of data compression (such as gzip, etc). You get some/much/most of the benefit of restricting to unique addresses while still capturing the entire stream of references. But as Paul noted, valgrind works in virtual addresses. Getting all the actual physical addresses is close to impossible. If you are working in an embedded device environment and care only about a small handful of memory-mapped device registers, then you can (must) process the mapping yourself. _______________________________________________ Valgrind-users mailing list Val...@li... https://lists.sourceforge.net/lists/listinfo/valgrind-users |
From: Bresalier, R. (N. - US/M. Hill) <rob...@no...> - 2022-08-05 23:37:06
|
I tried 'massif' on a simple program shown below where there are "definitely lost" leaks. massif doesn't seem to find "definitely lost" leaks, is this correct? I'm tried with both 3.19.0 and 3.15.0 versions of valgrind/massif, same result, "definitely lost" leaks are not found. I launch massif via: valgrind --tool=massif --sigill-diagnostics=no --error-limit=no --massif-out-file=definitely.%p.massif definitely.elf When I use memcheck it does find these definite leaks as below: ==29917== 60 bytes in 3 blocks are definitely lost in loss record 1 of 1 ==29917== at 0x402F67C: malloc (vg_replace_malloc.c:381) ==29917== by 0x80491D1: f2() (definitely.cpp:11) ==29917== by 0x804920F: f1() (definitely.cpp:17) ==29917== by 0x8049262: main (definitely.cpp:25) But massif doesn't find them at all? Is this correct? When I use massif on a program with "still reachable" it does find the still reachable, but it isn't finding definite leaks. Shouldn't massif also find definite leaks? The C code for "definitely.elf" is below: #include <stdlib.h> #include <stdio.h> #include <string.h> void* f2() { return malloc(20); } void f1() { f2(); } int main() { for (int i = 1; i <= 3; i++) { f1(); } return 0; } Thanks, Rob |
From: John R. <jr...@bi...> - 2022-08-05 19:49:06
|
>> if we can get a list of all the physical addresses the program used, in the order the program accessed them, and whether read/write. > For any real world application the size of the log would be overwhelmingly huge ... (unless you only want unique addresses). Of course this is the purpose of data compression (such as gzip, etc). You get some/much/most of the benefit of restricting to unique addresses while still capturing the entire stream of references. But as Paul noted, valgrind works in virtual addresses. Getting all the actual physical addresses is close to impossible. If you are working in an embedded device environment and care only about a small handful of memory-mapped device registers, then you can (must) process the mapping yourself. |
From: John R. <jr...@bi...> - 2022-08-05 19:39:20
|
>> Is there anything that can be done with memcheck to make it consume less memory? > > No. Well, you can use the command-line argument "--num-callers=<number>" to reduce the length of tracebacks that are stored in the "red zones" just before and after an allocated block. This might help enough if you have zillions of "still reachable" blocks. But you get shorter tracebacks, which might not give enough information to find and fix the leak quickly. If you do not have zillions of "still reachable" blocks, then --num-callers will not help so much; but probably would not be needed anyway. |
From: Paul F. <pj...@wa...> - 2022-08-05 19:30:44
|
> On 5 Aug 2022, at 20:53, G N Srinivasa Prasanna <gns...@ii...> wrote: > > > This is the first time we are using Valgrind, and we want to know if we can get a list of all the physical addresses the program used, in the order the program accessed them, and whether read/write. > > Please let us know if we can get this from Valgrind - the webpage information is not clear. > Hi Why do you need this? I’m not sure how to translate from virtual to physical address. Do you really mean physical address? For any real world application the size of the log would be overwhelmingly huge and I suspect would very rapidly fill most disks (unless you only want unique addresses). A+ Paul |
From: John R. <jr...@bi...> - 2022-08-05 19:28:04
|
> Does heaptrack also show the 'still reachable' types of leaks that memcheck does? Heaptrack intercepts malloc+free+etc, then logs the parameters, result, and traceback; but otherwise lets the progcess-original malloc+free+etc do the work. Heaptrack does not notice, and does not care, what you do with the result of malloc(), except whether or not the pointer returned by malloc() ever gets passed as an argument to free(). When heaptrack performs analysis, then any result from malloc() that has not been free()d is a "leak" as far as heaptrack is concerned. So that includes what memcheck calls "still reachable" but not (yet) a leak. > Any chance that the 'massif' tool would survive the OOM killer? This may be easier for me to get going as I already have valgrind built. Worth a try if you have a day or so to spend. Like all valgrind tools, massif relies on emulating the instruction stream, so the basic ~10X run-time slowdown applies. > Is there anything that can be done with memcheck to make it consume less memory? No. |
From: Bresalier, R. (N. - US/M. Hill) <rob...@no...> - 2022-08-05 19:16:05
|
> > If you want to know for sure who killed it then strace it while it > > runs and it should show you who sends the signel but my bet is that > > it's the kernel. > I tried strace -p <pid> on my process before I triggered its exit. The strace output ends saying with: "+++ killed by SIGKILL +++", but I don't find anything about who sent it. > Or possibly watch `dmesg -w` running in another shell. > I tried 'dmesg -w' but it didn't say anything about the SIGKILL. Is there something that has to be configured for dmesg to say the source of the SIGKILL? |
From: G N S. P. <gns...@ii...> - 2022-08-05 18:52:33
|
This is the first time we are using Valgrind, and we want to know if we can get a list of all the physical addresses the program used, in the order the program accessed them, and whether read/write. Please let us know if we can get this from Valgrind - the webpage information is not clear. Thanks Prasanna |
From: Bresalier, R. (N. - US/M. Hill) <rob...@no...> - 2022-08-05 18:49:19
|
Thanks Tom. Do you think I'd have better luck using the "massif" tool? Would "massif" be able to avoid the OOM killer? Or is there a way to reduce the amount of memory that memcheck will use? -----Original Message----- From: Tom Hughes <to...@co...> Sent: Friday, August 5, 2022 10:08 AM To: Bresalier, Rob (Nokia - US/Murray Hill) <rob...@no...>; val...@li... Subject: Re: memcheck is getting SIGKILLed before leak report is output On 05/08/2022 14:09, Bresalier, Rob (Nokia - US/Murray Hill) wrote: > When running memcheck on a massive monolith embedded executable (237MB > stripped, 1.8GiB unstripped), after I stop the executable under > valgrind I see the “HEAP SUMMARY” but then valgrind dies before any > leak reports are printed. The parent process sees that the return > status of memcheck is that it was SIGKILLed (status returned in > waitpid call is ‘9’). I am 99.9% sure that the parent process is not the one sending the SIGKILL. > Is it possible that valgrind SIGKILLs itself? Is there a reason that > the linux kernel (Wind River Linux) could be sending a SIGKILL to > valgrind/memcheck? I do not see any messages about Out of Memory/OOM > killer killing valgrind. Previous experience with this executable is > that there are almost 3 million leak reports (most of them are “still > reachable”), could that be occupying too much memory. Any ideas/advice > to figure out what is going on? Almost certainly the kernel OOM kiied it. If you want to know for sure who killed it then strace it while it runs and it should show you who sends the signel but my bet is that it's the kernel. > One thing I see in the logs is about “unhandled ioctl 0xa5 with no > size/direction hints”. Could this be a trigger for this crash/sigkill? Not really, no. Tom -- Tom Hughes (to...@co...) http://compton.nu/ |
From: Bresalier, R. (N. - US/M. Hill) <rob...@no...> - 2022-08-05 15:34:58
|
> If finding memory leaks is the only goal (for instance, if you are satisfied that > memcheck has found all the overrun blocks, uninitialized reads, etc.) then > https://github.com/KDE/heaptrack is the best tool. Thanks! I didn't know about heaptrack. I will look definitely into that. Does heaptrack also show the 'still reachable' types of leaks that memcheck does? Any chance that the 'massif' tool would survive the OOM killer? This may be easier for me to get going as I already have valgrind built. Is there anything that can be done with memcheck to make it consume less memory? |
From: Julian S. <jse...@gm...> - 2022-08-05 15:22:04
|
On 05/08/2022 16:08, Tom Hughes via Valgrind-users wrote: > If you want to know for sure who killed it then strace it while > it runs and it should show you who sends the signel but my bet is > that it's the kernel. Or possibly watch `dmesg -w` running in another shell. J |
From: John R. <jr...@bi...> - 2022-08-05 15:11:27
|
> When running memcheck on a massive monolith embedded executable (237MB stripped, 1.8GiB unstripped), after I stop the executable under valgrind I see the “HEAP SUMMARY” but then valgrind dies before any leak reports are printed. If finding memory leaks is the only goal (for instance, if you are satisfied that memcheck has found all the overrun blocks, uninitialized reads, etc.) then https://github.com/KDE/heaptrack is the best tool. The data-gathering phase runs in any Linux process using LD_PRELOAD and libunwind. The analysis phase runs a GUI under KDE, and/or generates *useful* text reports: leaks by individual size, leaks by total size for a given traceback, allocations (leaked or not) by frequency or total size, etc. I like the text-only analysis, which avoids the requirement for KDE. Heaptrack CPU overhead tends to be around 20% or less, so it does not take forever. Heaptrack does require disk space to record data (sequential access only), so you may need several gigabytes (locally or via network.) |
From: Tom H. <to...@co...> - 2022-08-05 14:08:25
|
On 05/08/2022 14:09, Bresalier, Rob (Nokia - US/Murray Hill) wrote: > When running memcheck on a massive monolith embedded executable (237MB > stripped, 1.8GiB unstripped), after I stop the executable under valgrind > I see the “HEAP SUMMARY” but then valgrind dies before any leak reports > are printed. The parent process sees that the return status of memcheck > is that it was SIGKILLed (status returned in waitpid call is ‘9’). I am > 99.9% sure that the parent process is not the one sending the SIGKILL. > Is it possible that valgrind SIGKILLs itself? Is there a reason that the > linux kernel (Wind River Linux) could be sending a SIGKILL to > valgrind/memcheck? I do not see any messages about Out of Memory/OOM > killer killing valgrind. Previous experience with this executable is > that there are almost 3 million leak reports (most of them are “still > reachable”), could that be occupying too much memory. Any ideas/advice > to figure out what is going on? Almost certainly the kernel OOM kiied it. If you want to know for sure who killed it then strace it while it runs and it should show you who sends the signel but my bet is that it's the kernel. > One thing I see in the logs is about “unhandled ioctl 0xa5 with no > size/direction hints”. Could this be a trigger for this crash/sigkill? Not really, no. Tom -- Tom Hughes (to...@co...) http://compton.nu/ |
From: Bresalier, R. (N. - US/M. Hill) <rob...@no...> - 2022-08-05 13:42:43
|
When running memcheck on a massive monolith embedded executable (237MB stripped, 1.8GiB unstripped), after I stop the executable under valgrind I see the "HEAP SUMMARY" but then valgrind dies before any leak reports are printed. The parent process sees that the return status of memcheck is that it was SIGKILLed (status returned in waitpid call is '9'). I am 99.9% sure that the parent process is not the one sending the SIGKILL. Is it possible that valgrind SIGKILLs itself? Is there a reason that the linux kernel (Wind River Linux) could be sending a SIGKILL to valgrind/memcheck? I do not see any messages about Out of Memory/OOM killer killing valgrind. Previous experience with this executable is that there are almost 3 million leak reports (most of them are "still reachable"), could that be occupying too much memory. Any ideas/advice to figure out what is going on? We don't seem to get the sigkill if valgrind/memcheck is stopped earlier in the life of this executable. But to find the leak I need it to run past that point. I've tried many different versions of valgrind that have worked to find leaks on this executable in the past (3.16.1, 3.18.1, 3.19.0) but they all have this same issue of being sigkilled before any leaks get printed. One thing I see in the logs is about "unhandled ioctl 0xa5 with no size/direction hints". Could this be a trigger for this crash/sigkill? Would appreciate any ideas/advice. Thanks, Rob |
From: Paul F. <pj...@wa...> - 2022-08-04 06:50:58
|
Hi Sgcheck never got beyond experimental and was removed from Valgrind a few versions ago. My advice is simply to not use it. A+ Paul > On 4 Aug 2022, at 07:45, Pahome Chen via Valgrind-users <val...@li...> wrote: > > > Dear all, > > I read the sgcheck’s doc and know it’s a experimental tool, but it seems found no error even a very simple program. > Does this still work or need to wait for another version? > > Below is from my script and experiment, and fail in Valgrind-3.8.1/3.9.0/3.10.0/3.11.0/3.12.0 > E.g. > >$ cat test_valgrind.c > #include<stdio.h> > #include<stdlib.h> > int main() > { > int val[10] = {0}; > int tmp = val[1], i = 0; > tmp += val[15]; // array overrun > tmp *= val[20]; // array overrun > for (i=0; i<20; ++i) { int tmp = val[i]; } // array overrun > return 0; > } > > When I run above version of Valgrind mentioned, it always comes out following message. > > ==11673== exp-sgcheck, a stack and global array overrun detector > ==11673== NOTE: This is an Experimental-Class Valgrind Tool > ==11673== Copyright (C) 2003-2015, and GNU GPL'd, by OpenWorks Ltd et al. > ==11673== Using Valgrind-3.11.0 and LibVEX; rerun with -h for copyright info > ==11673== Command: ./run > ==11673== > > exp-sgcheck: sg_main.c:2332 (sg_instrument_IRStmt): the 'impossible' happened. > > host stacktrace: > ==11673== at 0x3800CC09: show_sched_status_wrk (m_libcassert.c:343) > ==11673== by 0x3800CEF4: report_and_quit (m_libcassert.c:415) > ==11673== by 0x3800D127: vgPlain_assert_fail (m_libcassert.c:481) > ==11673== by 0x38004A03: sg_instrument_IRStmt (sg_main.c:2332) > ==11673== by 0x380003B3: h_instrument (h_main.c:683) > ==11673== by 0x3802968D: tool_instrument_then_gdbserver_if_needed (m_translate.c:238) > ==11673== by 0x380D3290: LibVEX_Translate (main_main.c:934) > ==11673== by 0x380271BF: vgPlain_translate (m_translate.c:1765) > ==11673== by 0x3805F857: vgPlain_scheduler (scheduler.c:1048) > ==11673== by 0x38090445: run_a_thread_NORETURN (syswrap-linux.c:102) > > sched status: > running_tid=1 > > Thread 1: status = VgTs_Runnable (lwpid 11673) > ==11673== at 0x40169EA: _dl_runtime_resolve_xsave (in /usr/lib64/ld-2.17.so) > ==11673== by 0x1B: ??? > ==11673== by 0x40057F: ??? (in /PATH/peihung/test/run) > ==11673== by 0xFFEFFF517: ??? > > > Note: see also the FAQ in the source distribution. > It contains workarounds to several common problems. > In particular, if Valgrind aborted or crashed after > identifying problems in your program, there's a good chance > that fixing those problems will prevent Valgrind aborting or > crashing, especially if it happened in m_mallocfree.c. > > My machine environment is Centos7 and x86_64 > > > > > Thanks. > Best regards, > Pahome > > _______________________________________________ > Valgrind-users mailing list > Val...@li... > https://lists.sourceforge.net/lists/listinfo/valgrind-users |
From: Pahome C. <Pei...@sy...> - 2022-08-04 05:44:46
|
Dear all, I read the sgcheck's doc and know it's a experimental tool, but it seems found no error even a very simple program. Does this still work or need to wait for another version? Below is from my script and experiment, and fail in Valgrind-3.8.1/3.9.0/3.10.0/3.11.0/3.12.0 E.g. >$ cat test_valgrind.c #include<stdio.h> #include<stdlib.h> int main() { int val[10] = {0}; int tmp = val[1], i = 0; tmp += val[15]; // array overrun tmp *= val[20]; // array overrun for (i=0; i<20; ++i) { int tmp = val[i]; } // array overrun return 0; } When I run above version of Valgrind mentioned, it always comes out following message. ==11673== exp-sgcheck, a stack and global array overrun detector ==11673== NOTE: This is an Experimental-Class Valgrind Tool ==11673== Copyright (C) 2003-2015, and GNU GPL'd, by OpenWorks Ltd et al. ==11673== Using Valgrind-3.11.0 and LibVEX; rerun with -h for copyright info ==11673== Command: ./run ==11673== exp-sgcheck: sg_main.c:2332 (sg_instrument_IRStmt): the 'impossible' happened. host stacktrace: ==11673== at 0x3800CC09: show_sched_status_wrk (m_libcassert.c:343) ==11673== by 0x3800CEF4: report_and_quit (m_libcassert.c:415) ==11673== by 0x3800D127: vgPlain_assert_fail (m_libcassert.c:481) ==11673== by 0x38004A03: sg_instrument_IRStmt (sg_main.c:2332) ==11673== by 0x380003B3: h_instrument (h_main.c:683) ==11673== by 0x3802968D: tool_instrument_then_gdbserver_if_needed (m_translate.c:238) ==11673== by 0x380D3290: LibVEX_Translate (main_main.c:934) ==11673== by 0x380271BF: vgPlain_translate (m_translate.c:1765) ==11673== by 0x3805F857: vgPlain_scheduler (scheduler.c:1048) ==11673== by 0x38090445: run_a_thread_NORETURN (syswrap-linux.c:102) sched status: running_tid=1 Thread 1: status = VgTs_Runnable (lwpid 11673) ==11673== at 0x40169EA: _dl_runtime_resolve_xsave (in /usr/lib64/ld-2.17.so) ==11673== by 0x1B: ??? ==11673== by 0x40057F: ??? (in /PATH/peihung/test/run) ==11673== by 0xFFEFFF517: ??? Note: see also the FAQ in the source distribution. It contains workarounds to several common problems. In particular, if Valgrind aborted or crashed after identifying problems in your program, there's a good chance that fixing those problems will prevent Valgrind aborting or crashing, especially if it happened in m_mallocfree.c. My machine environment is Centos7 and x86_64 Thanks. Best regards, Pahome |
From: Mark W. <ma...@kl...> - 2022-08-03 17:00:55
|
Hi, On Wed, Aug 03, 2022 at 05:32:56PM +0100, Tom Hughes wrote: > No, it's clone3: > > https://bugs.kde.org/show_bug.cgi?id=420906 So please upgrade to valgrind 3.18.0 or higher (latest is 3.19.0). Cheers, Mark |
From: Mark R. <ma...@cs...> - 2022-08-03 16:58:33
|
Just tried running the Valgrind test suite on WSL2 (win 10, Ubuntu 22.04). I'm not surprised that there were lots of failures. But the majority were: WARNING: unhandled amd64-linux syscall: 435 I suspect WSL is not a platform you care much about, but looking at syswrap for Darwin I see this might be pid hibernate? Would it be difficult to add support for this? Thank you, Mark -----Original Message----- From: Mark Wielaard [mailto:ma...@kl...] Sent: Monday, August 1, 2022 4:56 AM To: Tom Hughes <to...@co...>; Mark Roberts <ma...@cs...>; val...@li... Subject: Re: [Valgrind-users] new error message from Valgrind On Thu, 2022-07-28 at 22:22 +0100, Tom Hughes via Valgrind-users wrote: > On 28/07/2022 21:39, Mark Roberts wrote: > > I recently upgraded from Ubutu 20.04 to 22.04 and am now getting a > > new error message from Valgrind: > > > > --915-- WARNING: unhandled amd64-linux syscall: 334 > > > > --915-- You may be able to write your own handler. > > > > --915-- Read the file README_MISSING_SYSCALL_OR_IOCTL. > > > > --915-- Nevertheless we consider this a bug. Please report > > > > --915-- it at http://valgrind.org/support/bug_reports.html > > <http://valgrind.org/support/bug_reports.html>. > > > > Using same version of Valgrind as before (3.17). > > > > Any ideas as to what’s happening? > > Yes, your libc has started trying to use rseq. > > It's harmless - the next version of valgrind will silently reject it > with ENOSYS which is what is happening now anyway just with a warning. Where the next version of valgrind is 3.19.0 which is already released (in April). So you might just want to upgrade your valgrind. If you want to backport to older versions then the commit that got rid of the warning was: commit 1024237358f01009fe233cb1294f3b8211304eaa Author: Mark Wielaard <ma...@kl...> Date: Fri Dec 10 17:41:59 2021 +0100 Implement linux rseq syscall as ENOSYS This implements rseq for amd64, arm, arm64, ppc32, ppc64, s390x and x86 linux as ENOSYS (without warning). glibc will start using rseq to accelerate sched_getcpu, if available. This would cause a warning from valgrind every time a new thread is started. Real rseq (restartable sequences) support is pretty hard, so for now just explicitly return ENOSYS (just like we do for clone3). https://sourceware.org/pipermail/libc-alpha/2021-December/133656.html Cheers, Mark |