Prev: KVM: SVM: Handle MCE intercepts always on host level
Next: core: workque: workqueue recursion when unplugging usb WCDMA modem on 2.6.32 kernel
From: Vivek Goyal on 23 Apr 2010 10:50 On Thu, Apr 22, 2010 at 05:17:51PM -0700, Paul E. McKenney wrote: > On Thu, Apr 22, 2010 at 07:55:55PM -0400, Vivek Goyal wrote: > > On Thu, Apr 22, 2010 at 04:15:56PM -0700, Paul E. McKenney wrote: > > > On Thu, Apr 22, 2010 at 11:54:52AM -0400, Vivek Goyal wrote: > > > > With RCU correctness on, We see following warning. This patch fixes it. > > > > > > This is in initialization code, so that there cannot be any concurrent > > > updates, correct? If so, looks good. > > > > > > > I think theoritically two instances of cfq_init_queue() can be running > > in parallel (for two different devices), and they both can call > > blkiocg_add_blkio_group(). But then we use a spin lock to protect > > blkio_cgroup. > > > > spin_lock_irqsave(&blkcg->lock, flags); > > > > So I guess two parallel updates should be fine. > > OK, in that case, would it be possible add this spinlock to the condition > checked by css_id()'s rcu_dereference_check()? Hi Paul, I think adding these spinlock to condition checked might become little messy. And the reason being that this lock is subsystem (controller) specific and maintained by controller. Now if any controller implements a lock and we add that lock in css_id() rcu_dereference_check(), it will look ugly. So probably a better way is to make sure that css_id() is always called under rcu read lock so that we don't hit this warning? > At first glance, css_id() > needs to gain access to the blkio_cgroup structure that references > the cgroup_subsys_state structure passed to css_id(). > > This means that there is only one blkio_cgroup structure referencing > a given cgroup_subsys_state structure, right? Otherwise, we could still > have concurrent access. Yes. In fact css object is embedded in blkio_cgroup structure. So we take a rcu_read_lock() so that data structures associated with cgroup subsystem don't go away and then take controller specific blkio_cgroup spin lock to make sure multiple writers don't end up modifying a list at the same time. Am I missing something. Thanks Vivek > > > (Just wanting to make sure that we are not papering over a real error!) > > > > > > Thanx, Paul > > > > > > > [ 103.790505] =================================================== > > > > [ 103.790509] [ INFO: suspicious rcu_dereference_check() usage. ] > > > > [ 103.790511] --------------------------------------------------- > > > > [ 103.790514] kernel/cgroup.c:4432 invoked rcu_dereference_check() without protection! > > > > [ 103.790517] > > > > [ 103.790517] other info that might help us debug this: > > > > [ 103.790519] > > > > [ 103.790521] > > > > [ 103.790521] rcu_scheduler_active = 1, debug_locks = 1 > > > > [ 103.790524] 4 locks held by bash/4422: > > > > [ 103.790526] #0: (&buffer->mutex){+.+.+.}, at: [<ffffffff8114befa>] sysfs_write_file+0x3c/0x144 > > > > [ 103.790537] #1: (s_active#102){.+.+.+}, at: [<ffffffff8114bfa5>] sysfs_write_file+0xe7/0x144 > > > > [ 103.790544] #2: (&q->sysfs_lock){+.+.+.}, at: [<ffffffff812263b1>] queue_attr_store+0x49/0x8f > > > > [ 103.790552] #3: (&(&blkcg->lock)->rlock){......}, at: [<ffffffff8122e4db>] blkiocg_add_blkio_group+0x2b/0xad > > > > [ 103.790560] > > > > [ 103.790561] stack backtrace: > > > > [ 103.790564] Pid: 4422, comm: bash Not tainted 2.6.34-rc4-blkio-second-crash #81 > > > > [ 103.790567] Call Trace: > > > > [ 103.790572] [<ffffffff81068f57>] lockdep_rcu_dereference+0x9d/0xa5 > > > > [ 103.790577] [<ffffffff8107fac1>] css_id+0x44/0x57 > > > > [ 103.790581] [<ffffffff8122e503>] blkiocg_add_blkio_group+0x53/0xad > > > > [ 103.790586] [<ffffffff81231936>] cfq_init_queue+0x139/0x32c > > > > [ 103.790591] [<ffffffff8121f2d0>] elv_iosched_store+0xbf/0x1bf > > > > [ 103.790595] [<ffffffff812263d8>] queue_attr_store+0x70/0x8f > > > > [ 103.790599] [<ffffffff8114bfa5>] ? sysfs_write_file+0xe7/0x144 > > > > [ 103.790603] [<ffffffff8114bfc6>] sysfs_write_file+0x108/0x144 > > > > [ 103.790609] [<ffffffff810f527f>] vfs_write+0xae/0x10b > > > > [ 103.790612] [<ffffffff81069863>] ? trace_hardirqs_on_caller+0x10c/0x130 > > > > [ 103.790616] [<ffffffff810f539c>] sys_write+0x4a/0x6e > > > > [ 103.790622] [<ffffffff81002b5b>] system_call_fastpath+0x16/0x1b > > > > [ 103.790625] > > > > > > > > Signed-off-by: Vivek Goyal <vgoyal(a)redhat.com> > > > > --- > > > > block/cfq-iosched.c | 2 ++ > > > > 1 files changed, 2 insertions(+), 0 deletions(-) > > > > > > > > diff --git a/block/cfq-iosched.c b/block/cfq-iosched.c > > > > index 002a5b6..9386bf8 100644 > > > > --- a/block/cfq-iosched.c > > > > +++ b/block/cfq-iosched.c > > > > @@ -3741,8 +3741,10 @@ static void *cfq_init_queue(struct request_queue *q) > > > > * to make sure that cfq_put_cfqg() does not try to kfree root group > > > > */ > > > > atomic_set(&cfqg->ref, 1); > > > > + rcu_read_lock(); > > > > blkiocg_add_blkio_group(&blkio_root_cgroup, &cfqg->blkg, (void *)cfqd, > > > > 0); > > > > + rcu_read_unlock(); > > > > #endif > > > > /* > > > > * Not strictly needed (since RB_ROOT just clears the node and we > > > > -- > > > > 1.6.2.5 > > > > -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo(a)vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/
From: Vivek Goyal on 22 Apr 2010 20:00 On Thu, Apr 22, 2010 at 04:15:56PM -0700, Paul E. McKenney wrote: > On Thu, Apr 22, 2010 at 11:54:52AM -0400, Vivek Goyal wrote: > > With RCU correctness on, We see following warning. This patch fixes it. > > This is in initialization code, so that there cannot be any concurrent > updates, correct? If so, looks good. > I think theoritically two instances of cfq_init_queue() can be running in parallel (for two different devices), and they both can call blkiocg_add_blkio_group(). But then we use a spin lock to protect blkio_cgroup. spin_lock_irqsave(&blkcg->lock, flags); So I guess two parallel updates should be fine. Thanks Vivek > (Just wanting to make sure that we are not papering over a real error!) > > Thanx, Paul > > > [ 103.790505] =================================================== > > [ 103.790509] [ INFO: suspicious rcu_dereference_check() usage. ] > > [ 103.790511] --------------------------------------------------- > > [ 103.790514] kernel/cgroup.c:4432 invoked rcu_dereference_check() without protection! > > [ 103.790517] > > [ 103.790517] other info that might help us debug this: > > [ 103.790519] > > [ 103.790521] > > [ 103.790521] rcu_scheduler_active = 1, debug_locks = 1 > > [ 103.790524] 4 locks held by bash/4422: > > [ 103.790526] #0: (&buffer->mutex){+.+.+.}, at: [<ffffffff8114befa>] sysfs_write_file+0x3c/0x144 > > [ 103.790537] #1: (s_active#102){.+.+.+}, at: [<ffffffff8114bfa5>] sysfs_write_file+0xe7/0x144 > > [ 103.790544] #2: (&q->sysfs_lock){+.+.+.}, at: [<ffffffff812263b1>] queue_attr_store+0x49/0x8f > > [ 103.790552] #3: (&(&blkcg->lock)->rlock){......}, at: [<ffffffff8122e4db>] blkiocg_add_blkio_group+0x2b/0xad > > [ 103.790560] > > [ 103.790561] stack backtrace: > > [ 103.790564] Pid: 4422, comm: bash Not tainted 2.6.34-rc4-blkio-second-crash #81 > > [ 103.790567] Call Trace: > > [ 103.790572] [<ffffffff81068f57>] lockdep_rcu_dereference+0x9d/0xa5 > > [ 103.790577] [<ffffffff8107fac1>] css_id+0x44/0x57 > > [ 103.790581] [<ffffffff8122e503>] blkiocg_add_blkio_group+0x53/0xad > > [ 103.790586] [<ffffffff81231936>] cfq_init_queue+0x139/0x32c > > [ 103.790591] [<ffffffff8121f2d0>] elv_iosched_store+0xbf/0x1bf > > [ 103.790595] [<ffffffff812263d8>] queue_attr_store+0x70/0x8f > > [ 103.790599] [<ffffffff8114bfa5>] ? sysfs_write_file+0xe7/0x144 > > [ 103.790603] [<ffffffff8114bfc6>] sysfs_write_file+0x108/0x144 > > [ 103.790609] [<ffffffff810f527f>] vfs_write+0xae/0x10b > > [ 103.790612] [<ffffffff81069863>] ? trace_hardirqs_on_caller+0x10c/0x130 > > [ 103.790616] [<ffffffff810f539c>] sys_write+0x4a/0x6e > > [ 103.790622] [<ffffffff81002b5b>] system_call_fastpath+0x16/0x1b > > [ 103.790625] > > > > Signed-off-by: Vivek Goyal <vgoyal(a)redhat.com> > > --- > > block/cfq-iosched.c | 2 ++ > > 1 files changed, 2 insertions(+), 0 deletions(-) > > > > diff --git a/block/cfq-iosched.c b/block/cfq-iosched.c > > index 002a5b6..9386bf8 100644 > > --- a/block/cfq-iosched.c > > +++ b/block/cfq-iosched.c > > @@ -3741,8 +3741,10 @@ static void *cfq_init_queue(struct request_queue *q) > > * to make sure that cfq_put_cfqg() does not try to kfree root group > > */ > > atomic_set(&cfqg->ref, 1); > > + rcu_read_lock(); > > blkiocg_add_blkio_group(&blkio_root_cgroup, &cfqg->blkg, (void *)cfqd, > > 0); > > + rcu_read_unlock(); > > #endif > > /* > > * Not strictly needed (since RB_ROOT just clears the node and we > > -- > > 1.6.2.5 > > -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo(a)vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/
From: Vivek Goyal on 26 Apr 2010 09:40 On Sun, Apr 25, 2010 at 07:06:31PM -0700, Paul E. McKenney wrote: > On Mon, Apr 26, 2010 at 09:33:46AM +0800, Li Zefan wrote: > > >>>>>> With RCU correctness on, We see following warning. This patch fixes it. > > >>>>> This is in initialization code, so that there cannot be any concurrent > > >>>>> updates, correct? If so, looks good. > > >>>>> > > >>>> I think theoritically two instances of cfq_init_queue() can be running > > >>>> in parallel (for two different devices), and they both can call > > >>>> blkiocg_add_blkio_group(). But then we use a spin lock to protect > > >>>> blkio_cgroup. > > >>>> > > >>>> spin_lock_irqsave(&blkcg->lock, flags); > > >>>> > > >>>> So I guess two parallel updates should be fine. > > >>> OK, in that case, would it be possible add this spinlock to the condition > > >>> checked by css_id()'s rcu_dereference_check()? > > >> Hi Paul, > > >> > > >> I think adding these spinlock to condition checked might become little > > >> messy. And the reason being that this lock is subsystem (controller) > > >> specific and maintained by controller. Now if any controller implements > > >> a lock and we add that lock in css_id() rcu_dereference_check(), it will > > >> look ugly. > > >> > > >> So probably a better way is to make sure that css_id() is always called > > >> under rcu read lock so that we don't hit this warning? > > > > > > As long as holding rcu_read_lock() prevents css_id() from the usual > > > problems such as access memory that was concurrently freed, yes. > > > > blkiocg_add_blkio_group() also calls cgroup_path(), which also needs to > > be called within rcu_read_lock, so I think Vivek's patch is better than > > the one you posted in another mail thread. > > My apologies, Vivek! I lost track of your patch. I have now replaced > my patch with yours. Thanks Paul. I sent this patch to Jens also, thinking he will apply to his tree. Looks like he has not applied it yet though. Jens, is it ok if this patch gets merged through paul's tree or should it go through blk tree? Thanks Vivek -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo(a)vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/
From: Vivek Goyal on 26 Apr 2010 16:50 On Mon, Apr 26, 2010 at 07:45:42AM -0700, Paul E. McKenney wrote: > On Mon, Apr 26, 2010 at 09:39:20AM -0400, Vivek Goyal wrote: > > On Sun, Apr 25, 2010 at 07:06:31PM -0700, Paul E. McKenney wrote: > > > On Mon, Apr 26, 2010 at 09:33:46AM +0800, Li Zefan wrote: > > > > >>>>>> With RCU correctness on, We see following warning. This patch fixes it. > > > > >>>>> This is in initialization code, so that there cannot be any concurrent > > > > >>>>> updates, correct? If so, looks good. > > > > >>>>> > > > > >>>> I think theoritically two instances of cfq_init_queue() can be running > > > > >>>> in parallel (for two different devices), and they both can call > > > > >>>> blkiocg_add_blkio_group(). But then we use a spin lock to protect > > > > >>>> blkio_cgroup. > > > > >>>> > > > > >>>> spin_lock_irqsave(&blkcg->lock, flags); > > > > >>>> > > > > >>>> So I guess two parallel updates should be fine. > > > > >>> OK, in that case, would it be possible add this spinlock to the condition > > > > >>> checked by css_id()'s rcu_dereference_check()? > > > > >> Hi Paul, > > > > >> > > > > >> I think adding these spinlock to condition checked might become little > > > > >> messy. And the reason being that this lock is subsystem (controller) > > > > >> specific and maintained by controller. Now if any controller implements > > > > >> a lock and we add that lock in css_id() rcu_dereference_check(), it will > > > > >> look ugly. > > > > >> > > > > >> So probably a better way is to make sure that css_id() is always called > > > > >> under rcu read lock so that we don't hit this warning? > > > > > > > > > > As long as holding rcu_read_lock() prevents css_id() from the usual > > > > > problems such as access memory that was concurrently freed, yes. > > > > > > > > blkiocg_add_blkio_group() also calls cgroup_path(), which also needs to > > > > be called within rcu_read_lock, so I think Vivek's patch is better than > > > > the one you posted in another mail thread. > > > > > > My apologies, Vivek! I lost track of your patch. I have now replaced > > > my patch with yours. > > > > Thanks Paul. > > > > I sent this patch to Jens also, thinking he will apply to his tree. Looks > > like he has not applied it yet though. > > > > Jens, is it ok if this patch gets merged through paul's tree or should it > > go through blk tree? > > I am happy for it to go either way, so just let me know! I am also happy to go either way. I guess you can go ahead with pulling it in. Thanks Vivek -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo(a)vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/
From: Vivek Goyal on 26 Apr 2010 17:00
On Mon, Apr 26, 2010 at 01:47:04PM -0700, Paul E. McKenney wrote: > On Mon, Apr 26, 2010 at 04:42:08PM -0400, Vivek Goyal wrote: > > On Mon, Apr 26, 2010 at 07:45:42AM -0700, Paul E. McKenney wrote: > > > On Mon, Apr 26, 2010 at 09:39:20AM -0400, Vivek Goyal wrote: > > > > On Sun, Apr 25, 2010 at 07:06:31PM -0700, Paul E. McKenney wrote: > > > > > On Mon, Apr 26, 2010 at 09:33:46AM +0800, Li Zefan wrote: > > > > > > >>>>>> With RCU correctness on, We see following warning. This patch fixes it. > > > > > > >>>>> This is in initialization code, so that there cannot be any concurrent > > > > > > >>>>> updates, correct? If so, looks good. > > > > > > >>>>> > > > > > > >>>> I think theoritically two instances of cfq_init_queue() can be running > > > > > > >>>> in parallel (for two different devices), and they both can call > > > > > > >>>> blkiocg_add_blkio_group(). But then we use a spin lock to protect > > > > > > >>>> blkio_cgroup. > > > > > > >>>> > > > > > > >>>> spin_lock_irqsave(&blkcg->lock, flags); > > > > > > >>>> > > > > > > >>>> So I guess two parallel updates should be fine. > > > > > > >>> OK, in that case, would it be possible add this spinlock to the condition > > > > > > >>> checked by css_id()'s rcu_dereference_check()? > > > > > > >> Hi Paul, > > > > > > >> > > > > > > >> I think adding these spinlock to condition checked might become little > > > > > > >> messy. And the reason being that this lock is subsystem (controller) > > > > > > >> specific and maintained by controller. Now if any controller implements > > > > > > >> a lock and we add that lock in css_id() rcu_dereference_check(), it will > > > > > > >> look ugly. > > > > > > >> > > > > > > >> So probably a better way is to make sure that css_id() is always called > > > > > > >> under rcu read lock so that we don't hit this warning? > > > > > > > > > > > > > > As long as holding rcu_read_lock() prevents css_id() from the usual > > > > > > > problems such as access memory that was concurrently freed, yes. > > > > > > > > > > > > blkiocg_add_blkio_group() also calls cgroup_path(), which also needs to > > > > > > be called within rcu_read_lock, so I think Vivek's patch is better than > > > > > > the one you posted in another mail thread. > > > > > > > > > > My apologies, Vivek! I lost track of your patch. I have now replaced > > > > > my patch with yours. > > > > > > > > Thanks Paul. > > > > > > > > I sent this patch to Jens also, thinking he will apply to his tree. Looks > > > > like he has not applied it yet though. > > > > > > > > Jens, is it ok if this patch gets merged through paul's tree or should it > > > > go through blk tree? > > > > > > I am happy for it to go either way, so just let me know! > > > > I am also happy to go either way. I guess you can go ahead with pulling it in. > > I have it queued for 2.6.34. Thanks Paul. Where can I get your tree to clone from? I can test the changes. I can't find it on kernel.org. Thanks Vivek -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo(a)vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/ |