Re: [SSI-users] OpenSSI 1.9.2 on Fedora Core 3 instability
Brought to you by:
brucewalker,
rogertsang
From: <fx...@ma...> - 2006-09-01 08:44:58
|
Hello Roger, here something strange that I have notice: during a file extraction with tar, I see that a numer of cfs_async (5-6 process) is running and, the load with 'localview top' is around 8.75 (consider that now I'm running with only 1 node and no other users are accessing the cluster). What do you think? Is normal? Regards Pete On Thursday, August 31, 2006, at 07:28AM, Roger Tsang <rog...@gm...> wrote: >Hi, > >Yup I am using dedicated network segment for ICS. At the moment mine >are directly connected and gigE full-duplex and MTU 6800 though I >think I will get similar numbers when connected to my gigE switch and >MTU 1500. Both my ICS interfaces are using the same latest Yukon >Marvell drive from syskonnect. > >1. /etc/clustertab tells you which interface is on the ICS. >2. Have you done ttcp raw network speed tests on the ICS? Does your >network support jumbo frames? Check for health of your network with >netstat. Try connecting two nodes directly without your switch. Are >you getting same results? >3. You are copying from node1 to node1 on the same filesystem in your >test below and getting 3MB/sec correct? > >Roger > > >On 8/30/06, fx...@ma... <fx...@ma...> wrote: >> Hello Roger, >> >> your test seems that your cfs works well; does your cluster use a dedicated network segment for the interconnect? If yes, what's the link speed? >> >> All the 2nd (eth1) network card of all my servers are connected to a dedicated switch 10/100/1000; all the card are e1000, and the link that I see in my logs is 1000 MB/s full duples but, when I try to write from any nodes to /home, for example, I got an average speed of 700/800 Kb/s... (the sar -N dev show me that the interconnect network segment is not busy at all) >> >> Where I can find the problem? >> >> Now the entire cluster is installed from scratch, and node1 and node2 are directly connected to a SAN (before the SSI installation, hdparm told me that the transfer rate is about 120 MB/s). >> >> This is my fstab: >> >> # This file is edited by fstab-sync - see 'man fstab-sync' for details >> UUID=df880e73-72d8-4ffe-a397-bb4ab4aa95a5 / ext3 chard,defaults,node=1:2 1 1 >> LABEL=/boot /boot ext3 defaults,node=1 1 2 >> LABEL=/home1 /home1 ext3 chard,defaults,node=1:2 1 2 >> LABEL=/home2 /home2 ext3 chard,defaults,node=1:2 1 2 >> LABEL=/shared /shared ext3 chard,defaults,node=1:2 1 2 >> none /dev/pts devpts gid=5,mode=620,node=* 0 0 >> #none /dev/shm tmpfs defaults 0 0 >> none /proc proc defaults,node=* 0 0 >> none /sys sysfs defaults,node=* 0 0 >> /dev/sdd2 swap swap defaults,node=1 0 0 >> /dev/sdd2 swap swap defaults,node=2 0 0 >> /dev/sda1 swap swap defaults,node=3 0 0 >> /dev/sda1 swap swap defaults,node=4 0 0 >> /dev/sda8 swap swap defaults,node=5 0 0 >> /dev/hda /media/cdrom auto pamconsole,ro,exec,noauto,managed 0 0 >> /dev/fd0 /media/floppy auto pamconsole,exec,noauto,managed 0 0 >> >> /, /home1, /home2 and /shared and /boot on the SAN >> node2 have it's own boot device (and boot properly). >> >> I've added another node to the cluster, with the same configuration as the other ones (so now, it's 5 nodes) >> >> I've just tried from node1 (but the same result on all the others node and, consider that there are no other users connected) to copy a dir named "test" (about 4.6 GB) to "test.1". He took 26 minutes! (as you can see from my cut and paste below) >> >> [root@node1-public work1]# time cp -r test test1.1 >> >> real 26m5.864s >> user 0m0.491s >> sys 0m23.288s >> [root@node1-public work1]# >> >> During the transfer, I notice that the interconnect segment is not busy at all. Why this happen? CFS works via interconnect, right? >> >> The behaviour is very strange; the copy start with a good speed, then, begin, without no reason to slow; often, the copy is completly stalled... >> >> Any idea? >> >> Thanks in advance for your kindly answer >> Pete >> >> On Tuesday, August 29, 2006, at 01:13AM, Roger Tsang <rog...@gm...> wrote: >> >> >About your CFS performance problem I don't run into the same problem >> >on my 2 node cluster and my cluster is not half as powerful as yours - >> >just SATA 7200rpm disks and UP's. When copying whole directories >> >about 1GB each into another directory on the same filesystem (chard >> >mount), I get the following. I know it's kinda crude test, but it >> >clearly doens't slow down to 2-3MB/sec on my cluster. >> > >> >File I/O on hard mount is slower than soft mount because hard mounts >> >guarantee data has been written to support filesystem failover. >> > >> >Copy operation on just one node: >> >real 0m37.734s >> >user 0m0.093s >> >sys 0m3.515s >> > >> >Copy operation when there is another copy operation on the 2nd node at >> >the same time: >> >real 0m57.153s >> >user 0m0.092s >> >sys 0m3.448s >> > >> >It doesn't slow down to 2-3MB/sec. >> > >> >I also have QoS (HTB+SFQ) on the ICS network interfaces putting things >> >like ICMP and UDP ICS related traffic at highest priority. Maybe that >> >helps. >> > >> >Roger >> > >> > >> >On 8/28/06, fx...@ma... <fx...@ma...> wrote: >> >> Hello guys >> >> >> >> (First of all sorry for my bad english, I will try to do the best I can) >> >> >> >> I've installed, 4 weeks ago, openSSI 1.9.2 on Fedora Core3, following all the instruction of the various README.* >> >> >> >> Initially I've a lot of difficulties due to the documentation not update for FC3 and SSI1.9.2 (especially regarding the DRBD) but, to the end I've just installed a 4 node SSI cluster and I have some problems that I would submitted to you folks. >> >> >> >> The cluster is a 4 node: >> >> >> >> node1 (init): 4-way 3,16 GHz Xeon with 1 MB L2 cache and 8 GB RAM >> >> node2 (init): 2x2 core 2.80 GHz 2 MB L2 cache and 12 GB RAM >> >> node3: 2-way 3.00 GHz 1 MB L2 cache and 8 GB RAM >> >> node4: 2-way 3.00 GHz 1 MB L2 cache and 8 GB RAM >> >> >> >> Each node have 2 NIC, one connected to the "public" network and the other one connected to the "interconnect" network segment. >> >> >> >> HA-CVIP is configured on 2 the primary init node, and all the cluster is seen from the public network with only 1 IP address and the connection load balancing is working well. >> >> >> >> Each network segment are full duplex 1 GB/s, and, clearly, the interconnect network is a dedicated segment connected to a private switch (a 3Com switch 1 GB/s) >> >> >> >> On the inits node, will be connected a 4.0 TB SAN in order to have root and home failover (at the moment, the cluster is configured with root failover but without the SAN attached so, il failover occour all the cluster will going down). >> >> >> >> I've notice some strange behaviour and I'm wondering if some of you folks can help me: >> >> >> >> 1) The I/O of the entire cluster is quite slow; if more than 1 user try to do some massive I/O, in read or write (for example, a cvs checkout of 3 GB module) the entire cluster performance will be affected; I've done a lots of tests, but the results seems that the I/O throught CFS is quite slow (for example, if I try an scp copy from the public network, my scp copy will be load-balanced from CVIP to one of the 4 nodes and I have a throughput of 40-50 MB/s; if another user try to do the same, concurrent scp, the network transfer go down to 2-3 MB/s.... in order to exclude a network problem, I log into the cluster, and try a CP from a directory to another; the transfer rate is about 25-30 MB/s, if I try to add another cp (or, whaever I/O) during the cp, the throughput go down to 2/3 MB/s and the entire cluster is completely in stuck. Of course, the disks on the inits node are Ultra320, 15000 rpm disks (so I expect better performance). >> >> >> >> 2) If I use a clusternode_shutdown -t0 -h -N2 now (for example, but the beahviour is the same on all the node), the kernel node panic. No problem with a clusternode_shutdown -t0 -r -Nxxx >> >> >> >> 3) The process load-balancing seems that doesn't balance the load in equal part on all the nodes; on my cluster, the node1 (the init node) is always much loaded than the other node >> >> >> >> 4) Randomly, one of the 2 normal (not init) node joining the cluster, panic. >> >> >> >> 5) Java and cvs pserver processes is not migrating at all (into dmesg I see a message like "the process has exited" or something like that); CVS migrate but is not working anymore (socket migration problem?) >> >> >> >> 6) Why the kernel is not compiled with BIG memory support? The are some technical reason? >> >> >> >> >> >> Thanks to anyone would help me. >> >> >> >> Regards Pete >> >> >> >> >> >> ------------------------------------------------------------------------- >> >> Using Tomcat but need to do more? Need to support web services, security? >> >> Get stuff done quickly with pre-integrated technology to make your job easier >> >> Download IBM WebSphere Application Server v.1.0.1 based on Apache Geronimo >> >> http://sel.as-us.falkag.net/sel?cmd=lnk&kid=120709&bid=263057&dat=121642 >> >> _______________________________________________ >> >> Ssic-linux-users mailing list >> >> Ssi...@li... >> >> https://lists.sourceforge.net/lists/listinfo/ssic-linux-users >> >> >> > >> > >> > > |