|
From: didier <dga...@ma...> - 2001-12-13 15:08:55
|
Hi everybody Greg Lepore wrote: > > I am willing to enable profiling on my recent install of the 12/9 build of > 3.2 beta 4 to figure out some bottlenecks. I just need to know how and > what to look for in the results. I have 10 or so collections ranging from > 200 - 500,000 documents. I tried to search the site but the search has > been down for a couple of days (Internal Server Error). The system specs > are: RedHat 7.2, dual 1GHz processors, 1GB of RAM, 5 18GB SCSI hard > drives, four are RAID 5 (with /opt/ on that partition). If you guys can > tell me how to re-compile and what to look for I can do it. My initial > digs (over an NFS mount) were taking from 8-10 times as long as 3.1.3. > My rough re-design of the site is still available at: > http://rhobard.com/htdig/ > I still like it.... > I'm afraid profiling won't help a lot. First look at CPU usage with top, vmstat, time, whatever. If you have a cpu load around 10, 15% then the bottleneck is IO subsystem not the CPU. I didn't use 3.2 for a long time (I had 1, 2% CPU load), but if db size > RAM then you're toasted. Anyway for profiling (don't know if it's the right way but should work). make distclean export CFLAGS=-pg export CXXFLAGS=-pg ./configure --enable-shared=false and so on Don't use rundig (you won't get profile info for htdig) rather use htdig -i and other options after a long time. gprof htdig (you should have a gmon.out in your current dir). Geoff what about for -i: going the old way, ascii --> sort --> htmerge. or htdig uses a tempory db (half RAM size?) and bulk insert in the background, if there's such a thing with berkeley db. Didier |