You can subscribe to this list here.
2005 |
Jan
|
Feb
(53) |
Mar
(62) |
Apr
(88) |
May
(55) |
Jun
(204) |
Jul
(52) |
Aug
|
Sep
(1) |
Oct
(94) |
Nov
(15) |
Dec
(68) |
---|---|---|---|---|---|---|---|---|---|---|---|---|
2006 |
Jan
(130) |
Feb
(105) |
Mar
(34) |
Apr
(61) |
May
(41) |
Jun
(92) |
Jul
(176) |
Aug
(102) |
Sep
(247) |
Oct
(69) |
Nov
(32) |
Dec
(140) |
2007 |
Jan
(58) |
Feb
(51) |
Mar
(11) |
Apr
(20) |
May
(34) |
Jun
(37) |
Jul
(18) |
Aug
(60) |
Sep
(41) |
Oct
(105) |
Nov
(19) |
Dec
(14) |
2008 |
Jan
(3) |
Feb
|
Mar
(7) |
Apr
(5) |
May
(123) |
Jun
(5) |
Jul
(1) |
Aug
(29) |
Sep
(15) |
Oct
(21) |
Nov
(51) |
Dec
(3) |
2009 |
Jan
|
Feb
(36) |
Mar
(29) |
Apr
|
May
|
Jun
(7) |
Jul
(4) |
Aug
|
Sep
(4) |
Oct
|
Nov
(13) |
Dec
|
2010 |
Jan
|
Feb
|
Mar
(9) |
Apr
(11) |
May
(16) |
Jun
|
Jul
|
Aug
|
Sep
(1) |
Oct
|
Nov
|
Dec
|
2011 |
Jan
|
Feb
|
Mar
|
Apr
|
May
(1) |
Jun
|
Jul
|
Aug
|
Sep
|
Oct
|
Nov
|
Dec
|
2012 |
Jan
(7) |
Feb
(3) |
Mar
|
Apr
|
May
|
Jun
(3) |
Jul
|
Aug
|
Sep
|
Oct
(92) |
Nov
(28) |
Dec
(16) |
2013 |
Jan
(9) |
Feb
(2) |
Mar
|
Apr
(4) |
May
(4) |
Jun
(6) |
Jul
(14) |
Aug
(12) |
Sep
(4) |
Oct
(13) |
Nov
(1) |
Dec
(6) |
2014 |
Jan
(23) |
Feb
(19) |
Mar
(10) |
Apr
(14) |
May
(11) |
Jun
(6) |
Jul
(11) |
Aug
(15) |
Sep
(41) |
Oct
(95) |
Nov
(23) |
Dec
(11) |
2015 |
Jan
(3) |
Feb
(9) |
Mar
(19) |
Apr
(3) |
May
(1) |
Jun
(3) |
Jul
(11) |
Aug
(1) |
Sep
(15) |
Oct
(5) |
Nov
(2) |
Dec
|
2016 |
Jan
(7) |
Feb
(11) |
Mar
(8) |
Apr
(1) |
May
(3) |
Jun
(17) |
Jul
(12) |
Aug
(3) |
Sep
(5) |
Oct
(19) |
Nov
(12) |
Dec
(6) |
2017 |
Jan
(30) |
Feb
(23) |
Mar
(12) |
Apr
(32) |
May
(27) |
Jun
(7) |
Jul
(13) |
Aug
(16) |
Sep
(6) |
Oct
(11) |
Nov
|
Dec
(12) |
2018 |
Jan
(1) |
Feb
(5) |
Mar
(6) |
Apr
(7) |
May
(23) |
Jun
(3) |
Jul
(2) |
Aug
(1) |
Sep
(6) |
Oct
(6) |
Nov
(10) |
Dec
(3) |
2019 |
Jan
(26) |
Feb
(15) |
Mar
(9) |
Apr
|
May
(8) |
Jun
(14) |
Jul
(10) |
Aug
(10) |
Sep
(4) |
Oct
(2) |
Nov
(20) |
Dec
(10) |
2020 |
Jan
(10) |
Feb
(14) |
Mar
(29) |
Apr
(11) |
May
(25) |
Jun
(21) |
Jul
(23) |
Aug
(12) |
Sep
(19) |
Oct
(6) |
Nov
(8) |
Dec
(12) |
2021 |
Jan
(29) |
Feb
(9) |
Mar
(8) |
Apr
(8) |
May
(2) |
Jun
(2) |
Jul
(9) |
Aug
(9) |
Sep
(3) |
Oct
(4) |
Nov
(12) |
Dec
(13) |
2022 |
Jan
(4) |
Feb
|
Mar
(4) |
Apr
(12) |
May
(15) |
Jun
(7) |
Jul
(10) |
Aug
(2) |
Sep
|
Oct
(1) |
Nov
(8) |
Dec
|
2023 |
Jan
(15) |
Feb
|
Mar
(23) |
Apr
(1) |
May
(2) |
Jun
(10) |
Jul
|
Aug
(22) |
Sep
(19) |
Oct
(2) |
Nov
(20) |
Dec
|
2024 |
Jan
(1) |
Feb
|
Mar
(16) |
Apr
(15) |
May
(6) |
Jun
(4) |
Jul
(1) |
Aug
(1) |
Sep
|
Oct
(13) |
Nov
(18) |
Dec
(6) |
2025 |
Jan
(12) |
Feb
|
Mar
(2) |
Apr
(1) |
May
(11) |
Jun
(5) |
Jul
(4) |
Aug
(1) |
Sep
|
Oct
|
Nov
|
Dec
|
From: Stephen D. <sd...@gm...> - 2006-02-04 07:51:11
|
On 2/3/06, Vlad Seryakov <vl...@cr...> wrote: > > Here is the test http://www.crystalballinc.com/vlad/tmp/memtest.c > > It give very strange results, it works for Linux only because it uses > mmap only and it looks like brk uses mmap internally according to Linux > 2.6.13 kernel and it allows unlimited mmap-ed regions (or as i > understand up to vm.max_map_count =3D 65536). > > According to this test, when i use random sizes from 0-128k, Tcl > allocator gives worse results than Linux malloc. On small amounts > ckalloc faster but once over 64k, Linux malloc is faster. And, my small > malloc implementaion which is based on first version of Lea's malloc and > uses mmap only and supports per-thread memory only beats all mallocs, > especially on bigger sizes. It does not crash, even on 5Mil loops, but i > am not sure why it is so simple and so effective. Have you taken fragmentation into account? There's some memory related links in this blog post I read recently: http://primates.ximian.com/~federico/news-2005-12.html#14 Federico makes a good point: this has all been done before... |
From: Vlad S. <vl...@cr...> - 2006-02-04 06:13:05
|
Here is the test http://www.crystalballinc.com/vlad/tmp/memtest.c It give very strange results, it works for Linux only because it uses mmap only and it looks like brk uses mmap internally according to Linux 2.6.13 kernel and it allows unlimited mmap-ed regions (or as i understand up to vm.max_map_count = 65536). According to this test, when i use random sizes from 0-128k, Tcl allocator gives worse results than Linux malloc. On small amounts ckalloc faster but once over 64k, Linux malloc is faster. And, my small malloc implementaion which is based on first version of Lea's malloc and uses mmap only and supports per-thread memory only beats all mallocs, especially on bigger sizes. It does not crash, even on 5Mil loops, but i am not sure why it is so simple and so effective. -- Vlad Seryakov 571 262-8608 office vl...@cr... http://www.crystalballinc.com/vlad/ |
From: Vlad S. <vl...@cr...> - 2006-02-03 18:41:21
|
Let's start working on it Zoran Vasiljevic wrote: > Hi ! > > As Bernd broke the ice (thanks Bernd!) with submitting > the first doctools based manpage, I believe we should > slowly start to add some content there... > > The doctools format is rather simple and by looking at > one source page, you can grasp very easily how it's done. > If I look at the: > > llength [info comm ns_*] > 194 > > that would mean considerable work. Not counting the C-api. > But to have Tcl-API documentend, plus some nice-looking > bunch of useful config files for 2-3 types of common uses, > would really allow us to go 5.0. > The C-API can be documented in the code itself and we can > use some existing utilities to extract the docs directly > out of the C-files. > > What do you think? > > Cheers > Zoran > > > > > ------------------------------------------------------- > This SF.net email is sponsored by: Splunk Inc. Do you grep through log > files > for problems? Stop! Download the new AJAX search engine that makes > searching your log files as easy as surfing the web. DOWNLOAD SPLUNK! > http://sel.as-us.falkag.net/sel?cmd=lnk&kid=103432&bid=230486&dat=121642 > _______________________________________________ > naviserver-devel mailing list > nav...@li... > https://lists.sourceforge.net/lists/listinfo/naviserver-devel > -- Vlad Seryakov 571 262-8608 office vl...@cr... http://www.crystalballinc.com/vlad/ |
From: Zoran V. <zv...@ar...> - 2006-02-03 18:28:09
|
Hi ! As Bernd broke the ice (thanks Bernd!) with submitting the first doctools based manpage, I believe we should slowly start to add some content there... The doctools format is rather simple and by looking at one source page, you can grasp very easily how it's done. If I look at the: llength [info comm ns_*] 194 that would mean considerable work. Not counting the C-api. But to have Tcl-API documentend, plus some nice-looking bunch of useful config files for 2-3 types of common uses, would really allow us to go 5.0. The C-API can be documented in the code itself and we can use some existing utilities to extract the docs directly out of the C-files. What do you think? Cheers Zoran |
From: Zoran V. <zv...@ar...> - 2006-02-03 08:16:45
|
Am 02.02.2006 um 20:54 schrieb Andrew Piskorski: > > Google's TCMalloc thing is also relevent, but it's not the malloc > replacement I was thinking of. Ah yes, I was thinking of "Hoard": > Oh, apparently most of such tools are really tuned to multi-cpu systems. The problem I see is that when running MT-software on a single CPU, the things start to be different. As far as I could see, the Tcl MT allocator is pretty good overall (with a few exceptions) and it is just "there". But I'm curious to see other players as well. As soon as we get our internal release out of the door I will get some of those alternate allocatos under test. Cheer Zoran |
From: Zoran V. <zv...@ar...> - 2006-02-03 08:11:46
|
Am 02.02.2006 um 21:21 schrieb Gustaf Neumann: > how comes, that ckalloc is so much faster? It avoids the malloc's lock and builds its own malloc tables per-thread. So when lots of threads start to attack the malloc, there is quite a lot of contention on the mutex, so they go sleep for a while. > how do malloc/ckmalloc relate to ns_malloc? malloc is the bottom layer as provides with the OS (or the malloc library). ckalloc is a macro defined in Tcl so you can declare TCL_MEM_DEBUG and it will add some additional layer so the Tcl memory debugging tools can be used. If no TCL_MEM_DEBUG is declared, it defaults to Tcl_Alloc. The Tcl_Alloc is different for non-thread and thread-builds. This is controlled by USE_THREAD_ALLOC when compiling the Tcl library and is default for threaded builds. This activates special MT-optimized allocator. It handles all memory allocations <16284 bytes in per-thread tables, instead of shared tables thus avioding lock contention. This is what AS used before and it got into Tcl as it was/is pretty efficient overall. This is not always the case for 1cpu boxes, as I've seen in my tests. ns_malloc is just a wrapper arround the ckalloc. Cheers Zoran |
From: Zoran V. <zv...@ar...> - 2006-02-03 08:02:54
|
Am 03.02.2006 um 02:49 schrieb Vlad Seryakov: > Just as an idea, Zoran, when you are ready to release 4.99.1, it > could be a good idea to make a release on freshmeat.net as well. > Also, can we make modules subdir as a separate .tar.gz as well? > I see no reason why not. For naviserver, we have a "make dist". Is there anything like that for the rest of the modules? I believe not. So, something like "make dist" for all modules you made would be a good thing. Cheers Zoran |
From: Vlad S. <vl...@cr...> - 2006-02-03 01:49:29
|
Just as an idea, Zoran, when you are ready to release 4.99.1, it could be a good idea to make a release on freshmeat.net as well. Also, can we make modules subdir as a separate .tar.gz as well? -- Vlad Seryakov 571 262-8608 office vl...@cr... http://www.crystalballinc.com/vlad/ |
From: Gustaf N. <ne...@wu...> - 2006-02-02 20:21:23
|
Zoran Vasiljevic schrieb: > >> could not resist to try this on our p5 production system under >> modest load >> (64bit, linux, lpar with 25 processors, 8 dual-core with ibms >> version of hyperthreading) >> processor : 25 >> cpu : POWER5 (gr) >> clock : 1904.448000MHz >> revision : 2.3 > > > Urgs!?? What is this for a monster-machine??? Cool, isn't it? its an ibm p570 running Red Hat Enterprise Linux AS release 4 (Nahant Update 2), everything compiled 64 bit (this was more effort than expected). > > The timings what you get are what I expected on a multi-cpu box. > However all our single-cpu boxes are WAY slower with ckalloc > then with the regular malloc. > Do you happen to have a single-cpu box where you can try this out? not with this processor. if you are interested, i could get access to a 4-cpu box with an p5 processor. > > What I'm trying to understand is: is this pattern regular or not? with vlad's version, i get: Tcl: 8.4.11 starting 16 malloc threads...waiting....done: 20000 mallocs, 0 seconds, 392396 usec starting 16 ckalloc threads...waiting....done: 20000 mallocs, 0 seconds, 85702 usec modifying the #of threads, i get: threads malloc ckalloc ratio 8 0,121891 0,033443 3,644738809 16 0,392396 0,085702 4,578609601 24 0,813853 0,13622 5,974548524 32 1,122965 0,144308 7,781723813 40 1,564372 0,17262 9,062518827 48 2,490847 0,184043 13,53404911 56 3,299245 0,209622 15,73902071 100 8,139274 0,374034 21,76078645 how comes, that ckalloc is so much faster? how do malloc/ckmalloc relate to ns_malloc? even the ratio goes up by the # of threads: for 100 threads it is 21 times faster, while for 8 threads, the ratio is "only" 3.6. See the excel file for the graphics... -gustaf |
From: Andrew P. <at...@pi...> - 2006-02-02 19:54:55
|
On Thu, Feb 02, 2006 at 07:28:17PM +0000, Neophytos Demetriou wrote: > Andrew Piskorski wrote: > >If anyone is really interested in this, the best thing to do would be > >to try various more sophisticated high-performance multi-threaded > >malloc replacements, rather than just ns_malloc. This was discussed a > >bit on the AOLserver list a year or three ago, if anyone cares to > >search. > > Here's the message I had in mind: > http://www.opensubscriber.com/message/aol...@li.../1027255.html Google's TCMalloc thing is also relevent, but it's not the malloc replacement I was thinking of. Ah yes, I was thinking of "Hoard": http://www.hoard.org/ http://www.cs.umass.edu/~emery/hoard/ http://www.mail-archive.com/cgi-bin/htsearch?config=aolserver_listserv_aol_com&restrict=&exclude=&words=hoard -- Andrew Piskorski <at...@pi...> http://www.piskorski.com/ |
From: Neophytos D. <k2...@ph...> - 2006-02-02 19:28:31
|
Andrew Piskorski wrote: > If anyone is really interested in this, the best thing to do would be > to try various more sophisticated high-performance multi-threaded > malloc replacements, rather than just ns_malloc. This was discussed a > bit on the AOLserver list a year or three ago, if anyone cares to > search. Here's the message I had in mind: http://www.opensubscriber.com/message/aol...@li.../1027255.html Best, Neophytos |
From: Neophytos D. <k2...@ph...> - 2006-02-02 19:26:41
|
Andrew Piskorski wrote: > If anyone is really interested in this, the best thing to do would be > to try various more sophisticated high-performance multi-threaded > malloc replacements, rather than just ns_malloc. This was discussed a > bit on the AOLserver list a year or three ago, if anyone cares to > search. Andrew, are you talking about Google's PerfTools project? Best, Neophytos |
From: Andrew P. <at...@pi...> - 2006-02-02 19:22:37
|
On Thu, Feb 02, 2006 at 07:14:14PM +0100, Zoran Vasiljevic wrote: > For us: bottom line is: we will stay with ns_malloc as-is as the > speed penalty vs. malloc on Solaris/Mac 1cpu is not worth the > effort. If anyone is really interested in this, the best thing to do would be to try various more sophisticated high-performance multi-threaded malloc replacements, rather than just ns_malloc. This was discussed a bit on the AOLserver list a year or three ago, if anyone cares to search. -- Andrew Piskorski <at...@pi...> http://www.piskorski.com/ |
From: Zoran V. <zv...@ar...> - 2006-02-02 18:15:31
|
Am 02.02.2006 um 15:52 schrieb Vlad Seryakov: > I modified memtest to exclude thread related timings. > > It is at the http://www.crystalballinc.com/vlad/tmp/memtest.c > > when i call it with memtest 100000 and +, mallocs gettting faster > than Tcl alloc on my single CPU box After poking and poking and poking arround, I came to a conclusion that everything is right and everything is wrong. Nice conclusion, hm? You can really build upon it :-/ I have modified the nsthread/nsthreadtest.c to make the memory test with random blocksizes (up to 16K) instead of the fixed 10 bytes (which was indeed not very representative). Still, I observe different results using different hardware and OS. Generally, for Linux, we get mostly ns_malloc faster, although I have bunch of 1cpu boxes here where this is NOT true. OTOH, on Solaris/Mac 1cpu boxes, the malloc is faster or close to the ns_malloc. On 2+ CPU boxes the ns_malloc is always faster. It is difficult to say more about that w/o gathering extensive statistics data from various people, which I find very time consuming. Hence, it is good to know that one has to pay attention when running the Tcl MT-app (or naviserver) on a single cpu box and that box is a Solaris or Mac. Anyways it has been quite revealing for me to learn all of this thanks to your support and patience! For us: bottom line is: we will stay with ns_malloc as-is as the speed penalty vs. malloc on Solaris/Mac 1cpu is not worth the effort. Thanks again for your support! Zoran |
From: Vlad S. <vl...@cr...> - 2006-02-02 14:48:59
|
I modified memtest to exclude thread related timings. It is at the http://www.crystalballinc.com/vlad/tmp/memtest.c when i call it with memtest 100000 and +, mallocs gettting faster than Tcl alloc on my single CPU box Zoran Vasiljevic wrote: > > Am 02.02.2006 um 13:59 schrieb Andrew Piskorski: > >> >> Zoran, the boxes where you're tests show unexpectedly slow ns_malloc >> were all Mac PowerPC boxes running OS-X, is that right? If so, then >> the common thread here is OS-X, no? OS-X is known to be significantly >> less efficient than Linux in some areas, this is probably one of them? >> > > Oh no... > > I had tested Linux and Solaris boxes as well. > I have 2 Solaris boxes: one with 1 cpu and one with 2 cpu's. > The same goes for Mac OSX. > Unfortunaltey, I have only 1cpu linux boxes to test... > > What I saw is: 1cpu malloc better than ckalloc > 2cpu ckalloc better than malloc > > regardless which OS. > > Cheers > Zoran > > > ------------------------------------------------------- > This SF.net email is sponsored by: Splunk Inc. Do you grep through log > files > for problems? Stop! Download the new AJAX search engine that makes > searching your log files as easy as surfing the web. DOWNLOAD SPLUNK! > http://sel.as-us.falkag.net/sel?cmd=lnk&kid=103432&bid=230486&dat=121642 > _______________________________________________ > naviserver-devel mailing list > nav...@li... > https://lists.sourceforge.net/lists/listinfo/naviserver-devel > -- Vlad Seryakov 571 262-8608 office vl...@cr... http://www.crystalballinc.com/vlad/ |
From: Bernd E. <eid...@we...> - 2006-02-02 14:41:25
|
> Am 02.02.2006 um 12:01 schrieb Bernd Eidenschink: > > Or is it supposed to run veeeery long on my machine? :-) > > No. I believe the Tcl library you linked with is > not compiled with --enable-threads. Can you check that? Thanks for your hint with the LD_LIBRARY_PATH... oh boy! So, here: ------------------------- (A) Tcl: 8.4.12 starting 16 malloc threads...waiting....done: 0 seconds, 101180 usec starting 16 ckalloc threads...waiting....done: 0 seconds, 38316 usec A single CPU, SuSE, Kernel 2.6.13-15.7-default: cat /proc/cpuinfo model name : Intel(R) Pentium(R) 4 CPU 2.40GHz cpu MHz : 2390.523 cache size : 512 KB ------------------------- (B) Tcl: 8.4.12 starting 16 malloc threads...waiting....done: 0 seconds, 38916 usec starting 16 ckalloc threads...waiting....done: 0 seconds, 29926 usec A SMP-Kernel, SuSE, 2.6.13-15.7-smp, with Hyperthreading "2" Processors: model name : Intel(R) Pentium(R) 4 CPU 3.00GHz cpu MHz : 2993.229 cache size : 1024 KB ------------------------- (C) Tcl: 8.4.12 starting 16 malloc threads...waiting....done: 0 seconds, 25762 usec starting 16 ckalloc threads...waiting....done: 0 seconds, 17170 usec A SMP-Kernel 2.6.5-7.108-smp, SuSE, with 2 Processors and Hyperthreading (so 4 Processors are counted): model name : Intel(R) Xeon(TM) CPU 3.00GHz cpu MHz : 3002.125 cache size : 1024 KB |
From: Neophytos D. <k2...@ph...> - 2006-02-02 14:32:28
|
Zoran Vasiljevic wrote: >> this is my laptop, an IBM thinkpad X40: 1.4-GHz Pentium M >> > > But this one.... this one is killing me! This one is a single-cpu, right? Right. |
From: Vlad S. <vl...@cr...> - 2006-02-02 14:29:38
|
I tried on single CPU box 3.2Ghz with 1Gb of RAM, on 2CPU box i got the same result, ns_malloc is fatser Zoran Vasiljevic wrote: > > Am 01.02.2006 um 17:15 schrieb Vlad Seryakov: > >> On my machine with tcl 8.4.12 >> >> starting 10 malloc threads...waiting....done: 0 seconds, 16003 usec >> starting 10 ns_malloc threads...waiting....done: 0 seconds, 13207 usec > > > I've been trying to see why I'm getting worse values with ns_malloc > as with malloc and it turned out to be that only in 2+CPU box I was > able to get ns_malloc outperform the malloc. On all single-cpu boxes > the times were 2 up to 4 times better with plain malloc! > > Does anybody have a single AND multi-cpu box to try out? > > Cheers > Zoran > > > ------------------------------------------------------- > This SF.net email is sponsored by: Splunk Inc. Do you grep through log > files > for problems? Stop! Download the new AJAX search engine that makes > searching your log files as easy as surfing the web. DOWNLOAD SPLUNK! > http://sel.as-us.falkag.net/sel?cmd=lnk&kid=103432&bid=230486&dat=121642 > _______________________________________________ > naviserver-devel mailing list > nav...@li... > https://lists.sourceforge.net/lists/listinfo/naviserver-devel > -- Vlad Seryakov 571 262-8608 office vl...@cr... http://www.crystalballinc.com/vlad/ |
From: Zoran V. <zv...@ar...> - 2006-02-02 14:23:50
|
Am 02.02.2006 um 14:48 schrieb Neophytos Demetriou: > Zoran Vasiljevic wrote: >> Am 02.02.2006 um 14:10 schrieb Neophytos Demetriou: >>> Intel Pentium 4 3GHz [Hyperthreading] >>> Tcl: 8.4.11 >>> starting 16 malloc threads...waiting....done: 0 seconds, 58726 usec >>> starting 16 ckalloc threads...waiting....done: 0 seconds, 41410 >>> usec >> mult-icpu, right? > > single-cpu [with hyperthreading] This counts as 1+cpu allright. > >>> IBM X40 >>> Tcl: 8.4.11 >>> starting 16 malloc threads...waiting....done: 0 seconds, 101392 >>> usec >>> starting 16 ckalloc threads...waiting....done: 0 seconds, 45640 >>> usec >>> >> Is this also multicpu or a single-cpu? > > this is my laptop, an IBM thinkpad X40: 1.4-GHz Pentium M > But this one.... this one is killing me! This one is a single-cpu, right? Cheers Zoran |
From: Neophytos D. <k2...@ph...> - 2006-02-02 13:48:26
|
Zoran Vasiljevic wrote: > > Am 02.02.2006 um 14:10 schrieb Neophytos Demetriou: > >> Intel Pentium 4 3GHz [Hyperthreading] >> Tcl: 8.4.11 >> starting 16 malloc threads...waiting....done: 0 seconds, 58726 usec >> starting 16 ckalloc threads...waiting....done: 0 seconds, 41410 usec > > > mult-icpu, right? single-cpu [with hyperthreading] >> IBM X40 >> Tcl: 8.4.11 >> starting 16 malloc threads...waiting....done: 0 seconds, 101392 usec >> starting 16 ckalloc threads...waiting....done: 0 seconds, 45640 usec >> > > Is this also multicpu or a single-cpu? this is my laptop, an IBM thinkpad X40: 1.4-GHz Pentium M best, neophytos PS. I'm using Gentoo linux (2.6.12-gentoo-r4) on both machines: gcc 3.3.5-20050130 (Gentoo 3.3.5.20050130-r1, ssp-3.3.5.20050130-1, pie-8.7.7.1) |
From: Zoran V. <zv...@ar...> - 2006-02-02 13:26:52
|
Am 02.02.2006 um 14:10 schrieb Neophytos Demetriou: > Intel Pentium 4 3GHz [Hyperthreading] > Tcl: 8.4.11 > starting 16 malloc threads...waiting....done: 0 seconds, 58726 usec > starting 16 ckalloc threads...waiting....done: 0 seconds, 41410 usec mult-icpu, right? > > IBM X40 > Tcl: 8.4.11 > starting 16 malloc threads...waiting....done: 0 seconds, 101392 usec > starting 16 ckalloc threads...waiting....done: 0 seconds, 45640 usec > Is this also multicpu or a single-cpu? Do you have some single-cpu machine to test on? Cheers Zoran |
From: Zoran V. <zv...@ar...> - 2006-02-02 13:25:50
|
Am 02.02.2006 um 13:59 schrieb Andrew Piskorski: > > Zoran, the boxes where you're tests show unexpectedly slow ns_malloc > were all Mac PowerPC boxes running OS-X, is that right? If so, then > the common thread here is OS-X, no? OS-X is known to be significantly > less efficient than Linux in some areas, this is probably one of them? > Oh no... I had tested Linux and Solaris boxes as well. I have 2 Solaris boxes: one with 1 cpu and one with 2 cpu's. The same goes for Mac OSX. Unfortunaltey, I have only 1cpu linux boxes to test... What I saw is: 1cpu malloc better than ckalloc 2cpu ckalloc better than malloc regardless which OS. Cheers Zoran |
From: Neophytos D. <k2...@ph...> - 2006-02-02 13:10:55
|
Intel Pentium 4 3GHz [Hyperthreading] Tcl: 8.4.11 starting 16 malloc threads...waiting....done: 0 seconds, 58726 usec starting 16 ckalloc threads...waiting....done: 0 seconds, 41410 usec IBM X40 Tcl: 8.4.11 starting 16 malloc threads...waiting....done: 0 seconds, 101392 usec starting 16 ckalloc threads...waiting....done: 0 seconds, 45640 usec |
From: Andrew P. <at...@pi...> - 2006-02-02 13:00:17
|
On Thu, Feb 02, 2006 at 01:44:47PM +0100, Zoran Vasiljevic wrote: > The timings what you get are what I expected on a multi-cpu box. > However all our single-cpu boxes are WAY slower with ckalloc > then with the regular malloc. Zoran, the boxes where you're tests show unexpectedly slow ns_malloc were all Mac PowerPC boxes running OS-X, is that right? If so, then the common thread here is OS-X, no? OS-X is known to be significantly less efficient than Linux in some areas, this is probably one of them? -- Andrew Piskorski <at...@pi...> http://www.piskorski.com/ |
From: Zoran V. <zv...@ar...> - 2006-02-02 12:46:04
|
Am 02.02.2006 um 13:23 schrieb Gustaf Neumann: > could not resist to try this on our p5 production system under > modest load > (64bit, linux, lpar with 25 processors, 8 dual-core with ibms > version of hyperthreading) > processor : 25 > cpu : POWER5 (gr) > clock : 1904.448000MHz > revision : 2.3 Urgs!?? What is this for a monster-machine??? The timings what you get are what I expected on a multi-cpu box. However all our single-cpu boxes are WAY slower with ckalloc then with the regular malloc. Do you happen to have a single-cpu box where you can try this out? What I'm trying to understand is: is this pattern regular or not? If yes, then the default tcl allocator used for threading builds is sub-optimal for general builds and has to be runtime-dependent (use it for multi-cpu boxes and not on single-cpu). But before I go the tough way of getting this done in Tcl (I will definitely experience fierce opposition from "works-for-me" people...) I'd better collect some very hard ammunition... If not, then something is wrong with ALL our single-box machines which is very, very hard to believe... Thanks, Zoran |