From: KAMEZAWA Hiroyuki on 1 Mar 2010 19:30 On Mon, 1 Mar 2010 22:23:40 +0100 Andrea Righi <arighi(a)develer.com> wrote: > Apply the cgroup dirty pages accounting and limiting infrastructure to > the opportune kernel functions. > > Signed-off-by: Andrea Righi <arighi(a)develer.com> Seems nice. Hmm. the last problem is moving account between memcg. Right ? Thanks, -Kame > --- > fs/fuse/file.c | 5 +++ > fs/nfs/write.c | 4 ++ > fs/nilfs2/segment.c | 10 +++++- > mm/filemap.c | 1 + > mm/page-writeback.c | 84 ++++++++++++++++++++++++++++++++------------------ > mm/rmap.c | 4 +- > mm/truncate.c | 2 + > 7 files changed, 76 insertions(+), 34 deletions(-) > > diff --git a/fs/fuse/file.c b/fs/fuse/file.c > index a9f5e13..dbbdd53 100644 > --- a/fs/fuse/file.c > +++ b/fs/fuse/file.c > @@ -11,6 +11,7 @@ > #include <linux/pagemap.h> > #include <linux/slab.h> > #include <linux/kernel.h> > +#include <linux/memcontrol.h> > #include <linux/sched.h> > #include <linux/module.h> > > @@ -1129,6 +1130,8 @@ static void fuse_writepage_finish(struct fuse_conn *fc, struct fuse_req *req) > > list_del(&req->writepages_entry); > dec_bdi_stat(bdi, BDI_WRITEBACK); > + mem_cgroup_update_stat(req->pages[0], > + MEM_CGROUP_STAT_WRITEBACK_TEMP, -1); > dec_zone_page_state(req->pages[0], NR_WRITEBACK_TEMP); > bdi_writeout_inc(bdi); > wake_up(&fi->page_waitq); > @@ -1240,6 +1243,8 @@ static int fuse_writepage_locked(struct page *page) > req->inode = inode; > > inc_bdi_stat(mapping->backing_dev_info, BDI_WRITEBACK); > + mem_cgroup_update_stat(tmp_page, > + MEM_CGROUP_STAT_WRITEBACK_TEMP, 1); > inc_zone_page_state(tmp_page, NR_WRITEBACK_TEMP); > end_page_writeback(page); > > diff --git a/fs/nfs/write.c b/fs/nfs/write.c > index b753242..7316f7a 100644 > --- a/fs/nfs/write.c > +++ b/fs/nfs/write.c > @@ -439,6 +439,7 @@ nfs_mark_request_commit(struct nfs_page *req) > req->wb_index, > NFS_PAGE_TAG_COMMIT); > spin_unlock(&inode->i_lock); > + mem_cgroup_update_stat(req->wb_page, MEM_CGROUP_STAT_UNSTABLE_NFS, 1); > inc_zone_page_state(req->wb_page, NR_UNSTABLE_NFS); > inc_bdi_stat(req->wb_page->mapping->backing_dev_info, BDI_UNSTABLE); > __mark_inode_dirty(inode, I_DIRTY_DATASYNC); > @@ -450,6 +451,7 @@ nfs_clear_request_commit(struct nfs_page *req) > struct page *page = req->wb_page; > > if (test_and_clear_bit(PG_CLEAN, &(req)->wb_flags)) { > + mem_cgroup_update_stat(page, MEM_CGROUP_STAT_UNSTABLE_NFS, -1); > dec_zone_page_state(page, NR_UNSTABLE_NFS); > dec_bdi_stat(page->mapping->backing_dev_info, BDI_UNSTABLE); > return 1; > @@ -1273,6 +1275,8 @@ nfs_commit_list(struct inode *inode, struct list_head *head, int how) > req = nfs_list_entry(head->next); > nfs_list_remove_request(req); > nfs_mark_request_commit(req); > + mem_cgroup_update_stat(req->wb_page, > + MEM_CGROUP_STAT_UNSTABLE_NFS, -1); > dec_zone_page_state(req->wb_page, NR_UNSTABLE_NFS); > dec_bdi_stat(req->wb_page->mapping->backing_dev_info, > BDI_UNSTABLE); > diff --git a/fs/nilfs2/segment.c b/fs/nilfs2/segment.c > index ada2f1b..aef6d13 100644 > --- a/fs/nilfs2/segment.c > +++ b/fs/nilfs2/segment.c > @@ -1660,8 +1660,11 @@ nilfs_copy_replace_page_buffers(struct page *page, struct list_head *out) > } while (bh = bh->b_this_page, bh2 = bh2->b_this_page, bh != head); > kunmap_atomic(kaddr, KM_USER0); > > - if (!TestSetPageWriteback(clone_page)) > + if (!TestSetPageWriteback(clone_page)) { > + mem_cgroup_update_stat(clone_page, > + MEM_CGROUP_STAT_WRITEBACK, 1); > inc_zone_page_state(clone_page, NR_WRITEBACK); > + } > unlock_page(clone_page); > > return 0; > @@ -1783,8 +1786,11 @@ static void __nilfs_end_page_io(struct page *page, int err) > } > > if (buffer_nilfs_allocated(page_buffers(page))) { > - if (TestClearPageWriteback(page)) > + if (TestClearPageWriteback(page)) { > + mem_cgroup_update_stat(clone_page, > + MEM_CGROUP_STAT_WRITEBACK, -1); > dec_zone_page_state(page, NR_WRITEBACK); > + } > } else > end_page_writeback(page); > } > diff --git a/mm/filemap.c b/mm/filemap.c > index fe09e51..f85acae 100644 > --- a/mm/filemap.c > +++ b/mm/filemap.c > @@ -135,6 +135,7 @@ void __remove_from_page_cache(struct page *page) > * having removed the page entirely. > */ > if (PageDirty(page) && mapping_cap_account_dirty(mapping)) { > + mem_cgroup_update_stat(page, MEM_CGROUP_STAT_FILE_DIRTY, -1); > dec_zone_page_state(page, NR_FILE_DIRTY); > dec_bdi_stat(mapping->backing_dev_info, BDI_DIRTY); > } > diff --git a/mm/page-writeback.c b/mm/page-writeback.c > index 5a0f8f3..d83f41c 100644 > --- a/mm/page-writeback.c > +++ b/mm/page-writeback.c > @@ -137,13 +137,14 @@ static struct prop_descriptor vm_dirties; > */ > static int calc_period_shift(void) > { > - unsigned long dirty_total; > + unsigned long dirty_total, dirty_bytes; > > - if (vm_dirty_bytes) > - dirty_total = vm_dirty_bytes / PAGE_SIZE; > + dirty_bytes = mem_cgroup_dirty_bytes(); > + if (dirty_bytes) > + dirty_total = dirty_bytes / PAGE_SIZE; > else > - dirty_total = (vm_dirty_ratio * determine_dirtyable_memory()) / > - 100; > + dirty_total = (mem_cgroup_dirty_ratio() * > + determine_dirtyable_memory()) / 100; > return 2 + ilog2(dirty_total - 1); > } > > @@ -408,14 +409,16 @@ static unsigned long highmem_dirtyable_memory(unsigned long total) > */ > unsigned long determine_dirtyable_memory(void) > { > - unsigned long x; > - > - x = global_page_state(NR_FREE_PAGES) + global_reclaimable_pages(); > + unsigned long memory; > + s64 memcg_memory; > > + memory = global_page_state(NR_FREE_PAGES) + global_reclaimable_pages(); > if (!vm_highmem_is_dirtyable) > - x -= highmem_dirtyable_memory(x); > - > - return x + 1; /* Ensure that we never return 0 */ > + memory -= highmem_dirtyable_memory(memory); > + memcg_memory = mem_cgroup_page_stat(MEMCG_NR_DIRTYABLE_PAGES); > + if (memcg_memory < 0) > + return memory + 1; > + return min((unsigned long)memcg_memory, memory + 1); > } > > void > @@ -423,26 +426,28 @@ get_dirty_limits(unsigned long *pbackground, unsigned long *pdirty, > unsigned long *pbdi_dirty, struct backing_dev_info *bdi) > { > unsigned long background; > - unsigned long dirty; > + unsigned long dirty, dirty_bytes, dirty_background; > unsigned long available_memory = determine_dirtyable_memory(); > struct task_struct *tsk; > > - if (vm_dirty_bytes) > - dirty = DIV_ROUND_UP(vm_dirty_bytes, PAGE_SIZE); > + dirty_bytes = mem_cgroup_dirty_bytes(); > + if (dirty_bytes) > + dirty = DIV_ROUND_UP(dirty_bytes, PAGE_SIZE); > else { > int dirty_ratio; > > - dirty_ratio = vm_dirty_ratio; > + dirty_ratio = mem_cgroup_dirty_ratio(); > if (dirty_ratio < 5) > dirty_ratio = 5; > dirty = (dirty_ratio * available_memory) / 100; > } > > - if (dirty_background_bytes) > - background = DIV_ROUND_UP(dirty_background_bytes, PAGE_SIZE); > + dirty_background = mem_cgroup_dirty_background_bytes(); > + if (dirty_background) > + background = DIV_ROUND_UP(dirty_background, PAGE_SIZE); > else > - background = (dirty_background_ratio * available_memory) / 100; > - > + background = (mem_cgroup_dirty_background_ratio() * > + available_memory) / 100; > if (background >= dirty) > background = dirty / 2; > tsk = current; > @@ -508,9 +513,13 @@ static void balance_dirty_pages(struct address_space *mapping, > get_dirty_limits(&background_thresh, &dirty_thresh, > &bdi_thresh, bdi); > > - nr_reclaimable = global_page_state(NR_FILE_DIRTY) + > + nr_reclaimable = mem_cgroup_page_stat(MEMCG_NR_RECLAIM_PAGES); > + nr_writeback = mem_cgroup_page_stat(MEMCG_NR_WRITEBACK); > + if ((nr_reclaimable < 0) || (nr_writeback < 0)) { > + nr_reclaimable = global_page_state(NR_FILE_DIRTY) + > global_page_state(NR_UNSTABLE_NFS); > - nr_writeback = global_page_state(NR_WRITEBACK); > + nr_writeback = global_page_state(NR_WRITEBACK); > + } > > bdi_nr_reclaimable = bdi_stat(bdi, BDI_DIRTY); > if (bdi_cap_account_unstable(bdi)) { > @@ -611,10 +620,12 @@ static void balance_dirty_pages(struct address_space *mapping, > * In normal mode, we start background writeout at the lower > * background_thresh, to keep the amount of dirty memory low. > */ > + nr_reclaimable = mem_cgroup_page_stat(MEMCG_NR_RECLAIM_PAGES); > + if (nr_reclaimable < 0) > + nr_reclaimable = global_page_state(NR_FILE_DIRTY) + > + global_page_state(NR_UNSTABLE_NFS); > if ((laptop_mode && pages_written) || > - (!laptop_mode && ((global_page_state(NR_FILE_DIRTY) > - + global_page_state(NR_UNSTABLE_NFS)) > - > background_thresh))) > + (!laptop_mode && (nr_reclaimable > background_thresh))) > bdi_start_writeback(bdi, NULL, 0); > } > > @@ -678,6 +689,8 @@ void throttle_vm_writeout(gfp_t gfp_mask) > unsigned long dirty_thresh; > > for ( ; ; ) { > + unsigned long dirty; > + > get_dirty_limits(&background_thresh, &dirty_thresh, NULL, NULL); > > /* > @@ -686,10 +699,14 @@ void throttle_vm_writeout(gfp_t gfp_mask) > */ > dirty_thresh += dirty_thresh / 10; /* wheeee... */ > > - if (global_page_state(NR_UNSTABLE_NFS) + > - global_page_state(NR_WRITEBACK) <= dirty_thresh) > - break; > - congestion_wait(BLK_RW_ASYNC, HZ/10); > + > + dirty = mem_cgroup_page_stat(MEMCG_NR_DIRTY_WRITEBACK_PAGES); > + if (dirty < 0) > + dirty = global_page_state(NR_UNSTABLE_NFS) + > + global_page_state(NR_WRITEBACK); > + if (dirty <= dirty_thresh) > + break; > + congestion_wait(BLK_RW_ASYNC, HZ/10); > > /* > * The caller might hold locks which can prevent IO completion > @@ -1096,6 +1113,7 @@ int __set_page_dirty_no_writeback(struct page *page) > void account_page_dirtied(struct page *page, struct address_space *mapping) > { > if (mapping_cap_account_dirty(mapping)) { > + mem_cgroup_update_stat(page, MEM_CGROUP_STAT_FILE_DIRTY, 1); > __inc_zone_page_state(page, NR_FILE_DIRTY); > __inc_bdi_stat(mapping->backing_dev_info, BDI_DIRTY); > task_dirty_inc(current); > @@ -1297,6 +1315,8 @@ int clear_page_dirty_for_io(struct page *page) > * for more comments. > */ > if (TestClearPageDirty(page)) { > + mem_cgroup_update_stat(page, > + MEM_CGROUP_STAT_FILE_DIRTY, -1); > dec_zone_page_state(page, NR_FILE_DIRTY); > dec_bdi_stat(mapping->backing_dev_info, > BDI_DIRTY); > @@ -1332,8 +1352,10 @@ int test_clear_page_writeback(struct page *page) > } else { > ret = TestClearPageWriteback(page); > } > - if (ret) > + if (ret) { > + mem_cgroup_update_stat(page, MEM_CGROUP_STAT_WRITEBACK, -1); > dec_zone_page_state(page, NR_WRITEBACK); > + } > return ret; > } > > @@ -1363,8 +1385,10 @@ int test_set_page_writeback(struct page *page) > } else { > ret = TestSetPageWriteback(page); > } > - if (!ret) > + if (!ret) { > + mem_cgroup_update_stat(page, MEM_CGROUP_STAT_WRITEBACK, 1); > inc_zone_page_state(page, NR_WRITEBACK); > + } > return ret; > > } > diff --git a/mm/rmap.c b/mm/rmap.c > index 4d2fb93..8d74335 100644 > --- a/mm/rmap.c > +++ b/mm/rmap.c > @@ -832,7 +832,7 @@ void page_add_file_rmap(struct page *page) > { > if (atomic_inc_and_test(&page->_mapcount)) { > __inc_zone_page_state(page, NR_FILE_MAPPED); > - mem_cgroup_update_file_mapped(page, 1); > + mem_cgroup_update_stat(page, MEM_CGROUP_STAT_FILE_MAPPED, 1); > } > } > > @@ -864,7 +864,7 @@ void page_remove_rmap(struct page *page) > __dec_zone_page_state(page, NR_ANON_PAGES); > } else { > __dec_zone_page_state(page, NR_FILE_MAPPED); > - mem_cgroup_update_file_mapped(page, -1); > + mem_cgroup_update_stat(page, MEM_CGROUP_STAT_FILE_MAPPED, -1); > } > /* > * It would be tidy to reset the PageAnon mapping here, > diff --git a/mm/truncate.c b/mm/truncate.c > index 2466e0c..5f437e7 100644 > --- a/mm/truncate.c > +++ b/mm/truncate.c > @@ -73,6 +73,8 @@ void cancel_dirty_page(struct page *page, unsigned int account_size) > if (TestClearPageDirty(page)) { > struct address_space *mapping = page->mapping; > if (mapping && mapping_cap_account_dirty(mapping)) { > + mem_cgroup_update_stat(page, > + MEM_CGROUP_STAT_FILE_DIRTY, -1); > dec_zone_page_state(page, NR_FILE_DIRTY); > dec_bdi_stat(mapping->backing_dev_info, > BDI_DIRTY); > -- > 1.6.3.3 > > -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo(a)vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/
From: Andrea Righi on 2 Mar 2010 03:10 On Tue, Mar 02, 2010 at 09:23:09AM +0900, KAMEZAWA Hiroyuki wrote: > On Mon, 1 Mar 2010 22:23:40 +0100 > Andrea Righi <arighi(a)develer.com> wrote: > > > Apply the cgroup dirty pages accounting and limiting infrastructure to > > the opportune kernel functions. > > > > Signed-off-by: Andrea Righi <arighi(a)develer.com> > > Seems nice. > > Hmm. the last problem is moving account between memcg. > > Right ? Correct. This was actually the last item of the TODO list. Anyway, I'm still considering if it's correct to move dirty pages when a task is migrated from a cgroup to another. Currently, dirty pages just remain in the original cgroup and are flushed depending on the original cgroup settings. That is not totally wrong... at least moving the dirty pages between memcgs should be optional (move_charge_at_immigrate?). Thanks for your ack and the detailed review! -Andrea -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo(a)vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/
From: Daisuke Nishimura on 2 Mar 2010 03:20 On Tue, 2 Mar 2010 09:01:58 +0100, Andrea Righi <arighi(a)develer.com> wrote: > On Tue, Mar 02, 2010 at 09:23:09AM +0900, KAMEZAWA Hiroyuki wrote: > > On Mon, 1 Mar 2010 22:23:40 +0100 > > Andrea Righi <arighi(a)develer.com> wrote: > > > > > Apply the cgroup dirty pages accounting and limiting infrastructure to > > > the opportune kernel functions. > > > > > > Signed-off-by: Andrea Righi <arighi(a)develer.com> > > > > Seems nice. > > > > Hmm. the last problem is moving account between memcg. > > > > Right ? > > Correct. This was actually the last item of the TODO list. Anyway, I'm > still considering if it's correct to move dirty pages when a task is > migrated from a cgroup to another. Currently, dirty pages just remain in > the original cgroup and are flushed depending on the original cgroup > settings. That is not totally wrong... at least moving the dirty pages > between memcgs should be optional (move_charge_at_immigrate?). > FYI, I'm planning to add file-cache and shmem/tmpfs support for move_charge feature for 2.6.35. But, hmm, it would be complicated if we try to move dirty account too. Thanks, Daisuke Nishimura. -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo(a)vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/
From: KAMEZAWA Hiroyuki on 2 Mar 2010 03:30 On Tue, 2 Mar 2010 09:01:58 +0100 Andrea Righi <arighi(a)develer.com> wrote: > On Tue, Mar 02, 2010 at 09:23:09AM +0900, KAMEZAWA Hiroyuki wrote: > > On Mon, 1 Mar 2010 22:23:40 +0100 > > Andrea Righi <arighi(a)develer.com> wrote: > > > > > Apply the cgroup dirty pages accounting and limiting infrastructure to > > > the opportune kernel functions. > > > > > > Signed-off-by: Andrea Righi <arighi(a)develer.com> > > > > Seems nice. > > > > Hmm. the last problem is moving account between memcg. > > > > Right ? > > Correct. This was actually the last item of the TODO list. Anyway, I'm > still considering if it's correct to move dirty pages when a task is > migrated from a cgroup to another. Currently, dirty pages just remain in > the original cgroup and are flushed depending on the original cgroup > settings. That is not totally wrong... at least moving the dirty pages > between memcgs should be optional (move_charge_at_immigrate?). > My concern is - migration between memcg is already suppoted - at task move - at rmdir Then, if you leave DIRTY_PAGE accounting to original cgroup, the new cgroup (migration target)'s Dirty page accounting may goes to be negative, or incorrect value. Please check FILE_MAPPED implementation in __mem_cgroup_move_account() As if (page_mapped(page) && !PageAnon(page)) { /* Update mapped_file data for mem_cgroup */ preempt_disable(); __this_cpu_dec(from->stat->count[MEM_CGROUP_STAT_FILE_MAPPED]); __this_cpu_inc(to->stat->count[MEM_CGROUP_STAT_FILE_MAPPED]); preempt_enable(); } then, FILE_MAPPED never goes negative. Thanks, -Kame > Thanks for your ack and the detailed review! > > -Andrea > > -- > To unsubscribe, send a message with 'unsubscribe linux-mm' in > the body to majordomo(a)kvack.org. For more info on Linux MM, > see: http://www.linux-mm.org/ . > Don't email: <a href=mailto:"dont(a)kvack.org"> email(a)kvack.org </a> > -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo(a)vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/
From: Kirill A. Shutemov on 2 Mar 2010 05:20
On Mon, Mar 1, 2010 at 11:23 PM, Andrea Righi <arighi(a)develer.com> wrote: > Apply the cgroup dirty pages accounting and limiting infrastructure to > the opportune kernel functions. > > Signed-off-by: Andrea Righi <arighi(a)develer.com> > --- >  fs/fuse/file.c    |   5 +++ >  fs/nfs/write.c    |   4 ++ >  fs/nilfs2/segment.c |  10 +++++- >  mm/filemap.c     |   1 + >  mm/page-writeback.c |  84 ++++++++++++++++++++++++++++++++------------------ >  mm/rmap.c      |   4 +- >  mm/truncate.c    |   2 + >  7 files changed, 76 insertions(+), 34 deletions(-) > > diff --git a/fs/fuse/file.c b/fs/fuse/file.c > index a9f5e13..dbbdd53 100644 > --- a/fs/fuse/file.c > +++ b/fs/fuse/file.c > @@ -11,6 +11,7 @@ >  #include <linux/pagemap.h> >  #include <linux/slab.h> >  #include <linux/kernel.h> > +#include <linux/memcontrol.h> >  #include <linux/sched.h> >  #include <linux/module.h> > > @@ -1129,6 +1130,8 @@ static void fuse_writepage_finish(struct fuse_conn *fc, struct fuse_req *req) > >     list_del(&req->writepages_entry); >     dec_bdi_stat(bdi, BDI_WRITEBACK); > +    mem_cgroup_update_stat(req->pages[0], > +            MEM_CGROUP_STAT_WRITEBACK_TEMP, -1); >     dec_zone_page_state(req->pages[0], NR_WRITEBACK_TEMP); >     bdi_writeout_inc(bdi); >     wake_up(&fi->page_waitq); > @@ -1240,6 +1243,8 @@ static int fuse_writepage_locked(struct page *page) >     req->inode = inode; > >     inc_bdi_stat(mapping->backing_dev_info, BDI_WRITEBACK); > +    mem_cgroup_update_stat(tmp_page, > +            MEM_CGROUP_STAT_WRITEBACK_TEMP, 1); >     inc_zone_page_state(tmp_page, NR_WRITEBACK_TEMP); >     end_page_writeback(page); > > diff --git a/fs/nfs/write.c b/fs/nfs/write.c > index b753242..7316f7a 100644 > --- a/fs/nfs/write.c > +++ b/fs/nfs/write.c > @@ -439,6 +439,7 @@ nfs_mark_request_commit(struct nfs_page *req) >             req->wb_index, >             NFS_PAGE_TAG_COMMIT); >     spin_unlock(&inode->i_lock); > +    mem_cgroup_update_stat(req->wb_page, MEM_CGROUP_STAT_UNSTABLE_NFS, 1); >     inc_zone_page_state(req->wb_page, NR_UNSTABLE_NFS); >     inc_bdi_stat(req->wb_page->mapping->backing_dev_info, BDI_UNSTABLE); >     __mark_inode_dirty(inode, I_DIRTY_DATASYNC); > @@ -450,6 +451,7 @@ nfs_clear_request_commit(struct nfs_page *req) >     struct page *page = req->wb_page; > >     if (test_and_clear_bit(PG_CLEAN, &(req)->wb_flags)) { > +        mem_cgroup_update_stat(page, MEM_CGROUP_STAT_UNSTABLE_NFS, -1); >         dec_zone_page_state(page, NR_UNSTABLE_NFS); >         dec_bdi_stat(page->mapping->backing_dev_info, BDI_UNSTABLE); >         return 1; > @@ -1273,6 +1275,8 @@ nfs_commit_list(struct inode *inode, struct list_head *head, int how) >         req = nfs_list_entry(head->next); >         nfs_list_remove_request(req); >         nfs_mark_request_commit(req); > +        mem_cgroup_update_stat(req->wb_page, > +                MEM_CGROUP_STAT_UNSTABLE_NFS, -1); >         dec_zone_page_state(req->wb_page, NR_UNSTABLE_NFS); >         dec_bdi_stat(req->wb_page->mapping->backing_dev_info, >                 BDI_UNSTABLE); > diff --git a/fs/nilfs2/segment.c b/fs/nilfs2/segment.c > index ada2f1b..aef6d13 100644 > --- a/fs/nilfs2/segment.c > +++ b/fs/nilfs2/segment.c > @@ -1660,8 +1660,11 @@ nilfs_copy_replace_page_buffers(struct page *page, struct list_head *out) >     } while (bh = bh->b_this_page, bh2 = bh2->b_this_page, bh != head); >     kunmap_atomic(kaddr, KM_USER0); > > -    if (!TestSetPageWriteback(clone_page)) > +    if (!TestSetPageWriteback(clone_page)) { > +        mem_cgroup_update_stat(clone_page, s/clone_page/page/ And #include <linux/memcontrol.h> is missed. > +                MEM_CGROUP_STAT_WRITEBACK, 1); >         inc_zone_page_state(clone_page, NR_WRITEBACK); > +    } >     unlock_page(clone_page); > >     return 0; > @@ -1783,8 +1786,11 @@ static void __nilfs_end_page_io(struct page *page, int err) >     } > >     if (buffer_nilfs_allocated(page_buffers(page))) { > -        if (TestClearPageWriteback(page)) > +        if (TestClearPageWriteback(page)) { > +            mem_cgroup_update_stat(clone_page, > +                    MEM_CGROUP_STAT_WRITEBACK, -1); >             dec_zone_page_state(page, NR_WRITEBACK); > +        } >     } else >         end_page_writeback(page); >  } > diff --git a/mm/filemap.c b/mm/filemap.c > index fe09e51..f85acae 100644 > --- a/mm/filemap.c > +++ b/mm/filemap.c > @@ -135,6 +135,7 @@ void __remove_from_page_cache(struct page *page) >     * having removed the page entirely. >     */ >     if (PageDirty(page) && mapping_cap_account_dirty(mapping)) { > +        mem_cgroup_update_stat(page, MEM_CGROUP_STAT_FILE_DIRTY, -1); >         dec_zone_page_state(page, NR_FILE_DIRTY); >         dec_bdi_stat(mapping->backing_dev_info, BDI_DIRTY); >     } > diff --git a/mm/page-writeback.c b/mm/page-writeback.c > index 5a0f8f3..d83f41c 100644 > --- a/mm/page-writeback.c > +++ b/mm/page-writeback.c > @@ -137,13 +137,14 @@ static struct prop_descriptor vm_dirties; >  */ >  static int calc_period_shift(void) >  { > -    unsigned long dirty_total; > +    unsigned long dirty_total, dirty_bytes; > > -    if (vm_dirty_bytes) > -        dirty_total = vm_dirty_bytes / PAGE_SIZE; > +    dirty_bytes = mem_cgroup_dirty_bytes(); > +    if (dirty_bytes) > +        dirty_total = dirty_bytes / PAGE_SIZE; >     else > -        dirty_total = (vm_dirty_ratio * determine_dirtyable_memory()) / > -                100; > +        dirty_total = (mem_cgroup_dirty_ratio() * > +                determine_dirtyable_memory()) / 100; >     return 2 + ilog2(dirty_total - 1); >  } > > @@ -408,14 +409,16 @@ static unsigned long highmem_dirtyable_memory(unsigned long total) >  */ >  unsigned long determine_dirtyable_memory(void) >  { > -    unsigned long x; > - > -    x = global_page_state(NR_FREE_PAGES) + global_reclaimable_pages(); > +    unsigned long memory; > +    s64 memcg_memory; > > +    memory = global_page_state(NR_FREE_PAGES) + global_reclaimable_pages(); >     if (!vm_highmem_is_dirtyable) > -        x -= highmem_dirtyable_memory(x); > - > -    return x + 1;  /* Ensure that we never return 0 */ > +        memory -= highmem_dirtyable_memory(memory); > +    memcg_memory = mem_cgroup_page_stat(MEMCG_NR_DIRTYABLE_PAGES); > +    if (memcg_memory < 0) > +        return memory + 1; > +    return min((unsigned long)memcg_memory, memory + 1); >  } > >  void > @@ -423,26 +426,28 @@ get_dirty_limits(unsigned long *pbackground, unsigned long *pdirty, >         unsigned long *pbdi_dirty, struct backing_dev_info *bdi) >  { >     unsigned long background; > -    unsigned long dirty; > +    unsigned long dirty, dirty_bytes, dirty_background; >     unsigned long available_memory = determine_dirtyable_memory(); >     struct task_struct *tsk; > > -    if (vm_dirty_bytes) > -        dirty = DIV_ROUND_UP(vm_dirty_bytes, PAGE_SIZE); > +    dirty_bytes = mem_cgroup_dirty_bytes(); > +    if (dirty_bytes) > +        dirty = DIV_ROUND_UP(dirty_bytes, PAGE_SIZE); >     else { >         int dirty_ratio; > > -        dirty_ratio = vm_dirty_ratio; > +        dirty_ratio = mem_cgroup_dirty_ratio(); >         if (dirty_ratio < 5) >             dirty_ratio = 5; >         dirty = (dirty_ratio * available_memory) / 100; >     } > > -    if (dirty_background_bytes) > -        background = DIV_ROUND_UP(dirty_background_bytes, PAGE_SIZE); > +    dirty_background = mem_cgroup_dirty_background_bytes(); > +    if (dirty_background) > +        background = DIV_ROUND_UP(dirty_background, PAGE_SIZE); >     else > -        background = (dirty_background_ratio * available_memory) / 100; > - > +        background = (mem_cgroup_dirty_background_ratio() * > +                    available_memory) / 100; >     if (background >= dirty) >         background = dirty / 2; >     tsk = current; > @@ -508,9 +513,13 @@ static void balance_dirty_pages(struct address_space *mapping, >         get_dirty_limits(&background_thresh, &dirty_thresh, >                 &bdi_thresh, bdi); > > -        nr_reclaimable = global_page_state(NR_FILE_DIRTY) + > +        nr_reclaimable = mem_cgroup_page_stat(MEMCG_NR_RECLAIM_PAGES); > +        nr_writeback = mem_cgroup_page_stat(MEMCG_NR_WRITEBACK); > +        if ((nr_reclaimable < 0) || (nr_writeback < 0)) { > +            nr_reclaimable = global_page_state(NR_FILE_DIRTY) + >                     global_page_state(NR_UNSTABLE_NFS); > -        nr_writeback = global_page_state(NR_WRITEBACK); > +            nr_writeback = global_page_state(NR_WRITEBACK); > +        } > >         bdi_nr_reclaimable = bdi_stat(bdi, BDI_DIRTY); >         if (bdi_cap_account_unstable(bdi)) { > @@ -611,10 +620,12 @@ static void balance_dirty_pages(struct address_space *mapping, >     * In normal mode, we start background writeout at the lower >     * background_thresh, to keep the amount of dirty memory low. >     */ > +    nr_reclaimable = mem_cgroup_page_stat(MEMCG_NR_RECLAIM_PAGES); > +    if (nr_reclaimable < 0) > +        nr_reclaimable = global_page_state(NR_FILE_DIRTY) + > +                global_page_state(NR_UNSTABLE_NFS); >     if ((laptop_mode && pages_written) || > -      (!laptop_mode && ((global_page_state(NR_FILE_DIRTY) > -                + global_page_state(NR_UNSTABLE_NFS)) > -                     > background_thresh))) > +      (!laptop_mode && (nr_reclaimable > background_thresh))) >         bdi_start_writeback(bdi, NULL, 0); >  } > > @@ -678,6 +689,8 @@ void throttle_vm_writeout(gfp_t gfp_mask) >     unsigned long dirty_thresh; > >     for ( ; ; ) { > +        unsigned long dirty; > + >         get_dirty_limits(&background_thresh, &dirty_thresh, NULL, NULL); > >         /* > @@ -686,10 +699,14 @@ void throttle_vm_writeout(gfp_t gfp_mask) >          */ >         dirty_thresh += dirty_thresh / 10;    /* wheeee... */ > > -         if (global_page_state(NR_UNSTABLE_NFS) + > -            global_page_state(NR_WRITEBACK) <= dirty_thresh) > -                break; > -         congestion_wait(BLK_RW_ASYNC, HZ/10); > + > +        dirty = mem_cgroup_page_stat(MEMCG_NR_DIRTY_WRITEBACK_PAGES); > +        if (dirty < 0) > +            dirty = global_page_state(NR_UNSTABLE_NFS) + > +                global_page_state(NR_WRITEBACK); > +        if (dirty <= dirty_thresh) > +            break; > +        congestion_wait(BLK_RW_ASYNC, HZ/10); > >         /* >         * The caller might hold locks which can prevent IO completion > @@ -1096,6 +1113,7 @@ int __set_page_dirty_no_writeback(struct page *page) >  void account_page_dirtied(struct page *page, struct address_space *mapping) >  { >     if (mapping_cap_account_dirty(mapping)) { > +        mem_cgroup_update_stat(page, MEM_CGROUP_STAT_FILE_DIRTY, 1); >         __inc_zone_page_state(page, NR_FILE_DIRTY); >         __inc_bdi_stat(mapping->backing_dev_info, BDI_DIRTY); >         task_dirty_inc(current); > @@ -1297,6 +1315,8 @@ int clear_page_dirty_for_io(struct page *page) >         * for more comments. >         */ >         if (TestClearPageDirty(page)) { > +            mem_cgroup_update_stat(page, > +                    MEM_CGROUP_STAT_FILE_DIRTY, -1); >             dec_zone_page_state(page, NR_FILE_DIRTY); >             dec_bdi_stat(mapping->backing_dev_info, >                     BDI_DIRTY); > @@ -1332,8 +1352,10 @@ int test_clear_page_writeback(struct page *page) >     } else { >         ret = TestClearPageWriteback(page); >     } > -    if (ret) > +    if (ret) { > +        mem_cgroup_update_stat(page, MEM_CGROUP_STAT_WRITEBACK, -1); >         dec_zone_page_state(page, NR_WRITEBACK); > +    } >     return ret; >  } > > @@ -1363,8 +1385,10 @@ int test_set_page_writeback(struct page *page) >     } else { >         ret = TestSetPageWriteback(page); >     } > -    if (!ret) > +    if (!ret) { > +        mem_cgroup_update_stat(page, MEM_CGROUP_STAT_WRITEBACK, 1); >         inc_zone_page_state(page, NR_WRITEBACK); > +    } >     return ret; > >  } > diff --git a/mm/rmap.c b/mm/rmap.c > index 4d2fb93..8d74335 100644 > --- a/mm/rmap.c > +++ b/mm/rmap.c > @@ -832,7 +832,7 @@ void page_add_file_rmap(struct page *page) >  { >     if (atomic_inc_and_test(&page->_mapcount)) { >         __inc_zone_page_state(page, NR_FILE_MAPPED); > -        mem_cgroup_update_file_mapped(page, 1); > +        mem_cgroup_update_stat(page, MEM_CGROUP_STAT_FILE_MAPPED, 1); >     } >  } > > @@ -864,7 +864,7 @@ void page_remove_rmap(struct page *page) >         __dec_zone_page_state(page, NR_ANON_PAGES); >     } else { >         __dec_zone_page_state(page, NR_FILE_MAPPED); > -        mem_cgroup_update_file_mapped(page, -1); > +        mem_cgroup_update_stat(page, MEM_CGROUP_STAT_FILE_MAPPED, -1); >     } >     /* >     * It would be tidy to reset the PageAnon mapping here, > diff --git a/mm/truncate.c b/mm/truncate.c > index 2466e0c..5f437e7 100644 > --- a/mm/truncate.c > +++ b/mm/truncate.c > @@ -73,6 +73,8 @@ void cancel_dirty_page(struct page *page, unsigned int account_size) >     if (TestClearPageDirty(page)) { >         struct address_space *mapping = page->mapping; >         if (mapping && mapping_cap_account_dirty(mapping)) { > +            mem_cgroup_update_stat(page, > +                    MEM_CGROUP_STAT_FILE_DIRTY, -1); >             dec_zone_page_state(page, NR_FILE_DIRTY); >             dec_bdi_stat(mapping->backing_dev_info, >                     BDI_DIRTY); > -- > 1.6.3.3 > > |