From: Michal B. <mic...@ge...> - 2011-02-08 07:47:04
|
Hi! This is very interesting question. Our experience shows that when you have a big file to write (eg. 100 GB) it is good to logically divide it into blocks of one chunk size (64MB) and write these blocks to different files and later merge these files into one by using mfsappendchunks tool. This allows us to optimally use the throughput of the network. Mind that the last block may be filled with "zeros" up to 64MB. Another option is to also divide a big file into chunks and write them to the target file simultaneously do different positions of the file (using seek). So each "writer" saves a portion of data of [ chunksize*i, chunksize*(i+1)] range to different position. Unfortunately I am not sure if this could be implemented for optimal writing of vm images. Regards Michal From: Thomas S Hatch [mailto:tha...@gm...] Sent: Thursday, January 27, 2011 9:41 PM To: moosefs-users Subject: [Moosefs-users] Optimal writes for MooseFS We already know that vm image writes are not the optimal way to write to MooseFS, and randomIO is obviously not the best, so let me ask, what it the optimal way to write to a MooseFS filesystem? just sequentially? Is an append to an existing file faster? Thanks! -Thomas S Hatch |