diff options
| author | Oleg Nesterov <oleg@redhat.com> | 2014-10-09 15:25:41 -0700 |
|---|---|---|
| committer | Moyster <oysterized@gmail.com> | 2016-09-28 15:15:13 +0200 |
| commit | 7a6a1025f836bfa0fdf7fef7240fdf295a46372d (patch) | |
| tree | 915aae24ea6471ee3a389180185e72fc741244de /fs/proc | |
| parent | 08b29b79c414d1122686d221bcfe9f5ae6c4ef84 (diff) | |
fs/proc/task_mmu.c: reintroduce m->version logic
Add the "last_addr" optimization back. Like before, every ->show()
method checks !seq_overflow() and sets m->version = vma->vm_start.
However, it also checks that m_next_vma(vma) != NULL, otherwise it
sets m->version = -1 for the lockless "EOF" fast-path in m_start().
m_start() can simply do find_vma() + m_next_vma() if last_addr is
not zero, the code looks clear and simple and this case is clearly
separated from "scan vmas" path.
Signed-off-by: Oleg Nesterov <oleg@redhat.com>
Cc: Kirill A. Shutemov <kirill.shutemov@linux.intel.com>
Cc: Cyrill Gorcunov <gorcunov@openvz.org>
Cc: "Eric W. Biederman" <ebiederm@xmission.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
Signed-off-by: W4TCH0UT <ateekujjawal@gmail.com>
Conflicts:
fs/proc/task_mmu.c
Diffstat (limited to 'fs/proc')
| -rw-r--r-- | fs/proc/task_mmu.c | 25 |
1 files changed, 19 insertions, 6 deletions
diff --git a/fs/proc/task_mmu.c b/fs/proc/task_mmu.c index f4a285ef3..6cc89fabb 100644 --- a/fs/proc/task_mmu.c +++ b/fs/proc/task_mmu.c @@ -1,3 +1,4 @@ + #include <linux/mm.h> #include <linux/hugetlb.h> #include <linux/huge_mm.h> @@ -201,19 +202,21 @@ m_next_vma(struct proc_maps_private *priv, struct vm_area_struct *vma) return vma->vm_next ?: priv->tail_vma; } +static void m_cache_vma(struct seq_file *m, struct vm_area_struct *vma) +{ + if (m->count < m->size) /* vma is copied successfully */ + m->version = m_next_vma(m->private, vma) ? vma->vm_start : -1UL; +} + static void *m_start(struct seq_file *m, loff_t *ppos) { struct proc_maps_private *priv = m->private; + unsigned long last_addr = m->version; struct mm_struct *mm; struct vm_area_struct *vma; unsigned int pos = *ppos; - /* - * We remember last_addr rather than next_addr to hit with - * mmap_cache most of the time. We have zero last_addr at - * the beginning and also after lseek. We will have -1 last_addr - * after the end of the vmas. - */ + /* See m_cache_vma(). Zero at the start or after lseek. */ if (last_addr == -1UL) return NULL; @@ -229,6 +232,13 @@ static void *m_start(struct seq_file *m, loff_t *ppos) hold_task_mempolicy(priv); priv->tail_vma = get_gate_vma(mm); + if (last_addr) { + vma = find_vma(mm, last_addr); + if (vma && (vma = m_next_vma(priv, vma))) + return vma; + } + + m->version = 0; if (pos < mm->map_count) { for (vma = mm->mmap; pos; pos--) vma = vma->vm_next; @@ -404,6 +414,7 @@ done: static int show_map(struct seq_file *m, void *v, int is_pid) { show_map_vma(m, v, is_pid); + m_cache_vma(m, v); return 0; } @@ -727,6 +738,7 @@ static int show_smap(struct seq_file *m, void *v, int is_pid) mss.nonlinear >> 10); show_smap_vma_flags(m, vma); + m_cache_vma(m, vma); if (vma_get_anon_name(vma)) { seq_puts(m, "Name: "); @@ -1444,6 +1456,7 @@ static int show_numa_map(struct seq_file *m, void *v, int is_pid) seq_printf(m, " N%d=%lu", n, md->node[n]); out: seq_putc(m, '\n'); + m_cache_vma(m, vma); return 0; } |
