Prev: [2.6.33.4 PATCH] V4L/uvcvideo: Add support for Suyin Corp. Lenovo Webcam
Next: [PATCH] scripts: checkpatch.pl
From: Christoph Hellwig on 23 Jun 2010 05:30 On Tue, Jun 22, 2010 at 05:34:59PM -0400, Jeff Moyer wrote: > Hi, > > Running iozone with the fsync flag, or fs_mark, the performance of CFQ is > far worse than that of deadline for enterprise class storage when dealing > with file sizes of 8MB or less. I used the following command line as a > representative test case: > > fs_mark -S 1 -D 10000 -N 100000 -d /mnt/test/fs_mark -s 65536 -t 1 -w 4096 -F > > When run using the deadline I/O scheduler, an average of the first 5 numbers > will give you 448.4 files / second. CFQ will yield only 106.7. With > this patch series applied (and the two patches I sent yesterday), CFQ now > achieves 462.5 files / second. > > This patch set is still an RFC. I'd like to make it perform better when > there is a competing sequential reader present. For now, I've addressed > the concerns voiced about the previous posting. What happened to the initial idea of just using the BIO_RW_META flag for log writes? In the end log writes are the most important writes you have in a journaled filesystem, and they should not be effect to any kind of queue idling logic or other interruption. Log I/O is usually very little (unless you use old XFS code with a worst-case directory manipulation workload), and very latency sensitive. -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo(a)vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/
From: Tao Ma on 23 Jun 2010 05:40 Hi Jeff, On 06/23/2010 05:34 AM, Jeff Moyer wrote: > Hi, > > Running iozone with the fsync flag, or fs_mark, the performance of CFQ is > far worse than that of deadline for enterprise class storage when dealing > with file sizes of 8MB or less. I used the following command line as a > representative test case: > > fs_mark -S 1 -D 10000 -N 100000 -d /mnt/test/fs_mark -s 65536 -t 1 -w 4096 -F > > When run using the deadline I/O scheduler, an average of the first 5 numbers > will give you 448.4 files / second. CFQ will yield only 106.7. With > this patch series applied (and the two patches I sent yesterday), CFQ now > achieves 462.5 files / second. which 2 patches? Could you paste the link or the subject? Just want to make my test env like yours. ;) As Joel mentioned in another mail, ocfs2 also use jbd/jbd2, so I'd like to give it a try and give you some feedback about the test. Regards, Tao > > This patch set is still an RFC. I'd like to make it perform better when > there is a competing sequential reader present. For now, I've addressed > the concerns voiced about the previous posting. > > Review and testing would be greatly appreciated. > > Thanks! > Jeff > > --- > > New from the last round: > > - removed the think time calculation I added for the sync-noidle service tree > - replaced above with a suggestion from Vivek to only guard against currently > active sequential readers when determining if we can preempt the sync-noidle > service tree. > - bug fixes > > Over all, I think it's simpler now thanks to the suggestions from Jens and > Vivek. > > [PATCH 1/3] block: Implement a blk_yield function to voluntarily give up the I/O scheduler. > [PATCH 2/3] jbd: yield the device queue when waiting for commits > [PATCH 3/3] jbd2: yield the device queue when waiting for journal commits > -- > To unsubscribe from this list: send the line "unsubscribe linux-kernel" in > the body of a message to majordomo(a)vger.kernel.org > More majordomo info at http://vger.kernel.org/majordomo-info.html > Please read the FAQ at http://www.tux.org/lkml/ > -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo(a)vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/
From: Tao Ma on 24 Jun 2010 02:00 Hi Jeff, On 06/23/2010 05:30 PM, Tao Ma wrote: > Hi Jeff, > > On 06/23/2010 05:34 AM, Jeff Moyer wrote: >> Hi, >> >> Running iozone with the fsync flag, or fs_mark, the performance of CFQ is >> far worse than that of deadline for enterprise class storage when dealing >> with file sizes of 8MB or less. I used the following command line as a >> representative test case: >> >> fs_mark -S 1 -D 10000 -N 100000 -d /mnt/test/fs_mark -s 65536 -t 1 -w >> 4096 -F >> >> When run using the deadline I/O scheduler, an average of the first 5 >> numbers >> will give you 448.4 files / second. CFQ will yield only 106.7. With >> this patch series applied (and the two patches I sent yesterday), CFQ now >> achieves 462.5 files / second. > which 2 patches? Could you paste the link or the subject? Just want to > make my test env like yours. ;) > As Joel mentioned in another mail, ocfs2 also use jbd/jbd2, so I'd like > to give it a try and give you some feedback about the test. I am sorry to say that the patch make jbd2 locked up when I tested fs_mark using ocfs2. I have attached the log from my netconsole server. After I reverted the patch [3/3], the box works again. Regards, Tao
From: Tao Ma on 28 Jun 2010 02:50 Hi Jeff, On 06/27/2010 09:48 PM, Jeff Moyer wrote: > Tao Ma<tao.ma(a)oracle.com> writes: > >> Hi Jeff, >> >> On 06/23/2010 05:30 PM, Tao Ma wrote: >>> Hi Jeff, >>> >>> On 06/23/2010 05:34 AM, Jeff Moyer wrote: >>>> Hi, >>>> >>>> Running iozone with the fsync flag, or fs_mark, the performance of CFQ is >>>> far worse than that of deadline for enterprise class storage when dealing >>>> with file sizes of 8MB or less. I used the following command line as a >>>> representative test case: >>>> >>>> fs_mark -S 1 -D 10000 -N 100000 -d /mnt/test/fs_mark -s 65536 -t 1 -w >>>> 4096 -F >>>> >>>> When run using the deadline I/O scheduler, an average of the first 5 >>>> numbers >>>> will give you 448.4 files / second. CFQ will yield only 106.7. With >>>> this patch series applied (and the two patches I sent yesterday), CFQ now >>>> achieves 462.5 files / second. >>> which 2 patches? Could you paste the link or the subject? Just want to >>> make my test env like yours. ;) >>> As Joel mentioned in another mail, ocfs2 also use jbd/jbd2, so I'd like >>> to give it a try and give you some feedback about the test. >> I am sorry to say that the patch make jbd2 locked up when I tested >> fs_mark using ocfs2. >> I have attached the log from my netconsole server. After I reverted >> the patch [3/3], the box works again. > > I can't reproduce this, unfortunately. Also, when building with the > .config you sent me, the disassembly doesn't line up with the stack > trace you posted. > > I'm not sure why yielding the queue would cause a deadlock. The only > explanation I can come up with is that I/O is not being issued. I'm > assuming that no other I/O will be completed to the file system in > question. Is that right? Could you send along the output from sysrq-t? yes, I just mounted it and begin the test, so there should be no outstanding I/O. So do you need me to setup another disk for test? I have attached the sysrq output in sysrq.log. please check. btw, I also met with a NULL pointer deference in cfq_yield. I have attached the null.log also. This seems to be related to the previous deadlock and happens when I try to remount the same volume after reboot and ocfs2 try to do some recovery. Regards, Tao
From: Tao Ma on 28 Jun 2010 19:20 Jeff Moyer wrote: > Tao Ma <tao.ma(a)oracle.com> writes: > > >> btw, I also met with a NULL pointer deference in cfq_yield. I have >> attached the null.log also. This seems to be related to the previous >> deadlock and happens when I try to remount the same volume after >> reboot and ocfs2 try to do some recovery. >> > > Since I can't reproduce your binary kernel even with your .config, could > you send me the disassembly of the cfq_yield function from your vmlinux > binary? > no problem. I have attached it. btw, if you have any debug patch, I am happy to run it to make the problem more clear to you. Regards, Tao
|
Next
|
Last
Pages: 1 2 Prev: [2.6.33.4 PATCH] V4L/uvcvideo: Add support for Suyin Corp. Lenovo Webcam Next: [PATCH] scripts: checkpatch.pl |