Merge "vmstat: use this_cpu() to avoid irqon/off sequence in refresh_cpu_vm_stats"

This commit is contained in:
Linux Build Service Account 2015-03-14 00:59:29 -07:00 committed by Gerrit - the friendly Code Review server
commit a5ee86c509

View file

@ -440,33 +440,29 @@ static inline void fold_diff(int *diff)
* with the global counters. These could cause remote node cache line * with the global counters. These could cause remote node cache line
* bouncing and will have to be only done when necessary. * bouncing and will have to be only done when necessary.
*/ */
static void refresh_cpu_vm_stats(int cpu) static void refresh_cpu_vm_stats(void)
{ {
struct zone *zone; struct zone *zone;
int i; int i;
int global_diff[NR_VM_ZONE_STAT_ITEMS] = { 0, }; int global_diff[NR_VM_ZONE_STAT_ITEMS] = { 0, };
for_each_populated_zone(zone) { for_each_populated_zone(zone) {
struct per_cpu_pageset *p; struct per_cpu_pageset __percpu *p = zone->pageset;
p = per_cpu_ptr(zone->pageset, cpu); for (i = 0; i < NR_VM_ZONE_STAT_ITEMS; i++) {
for (i = 0; i < NR_VM_ZONE_STAT_ITEMS; i++)
if (p->vm_stat_diff[i]) {
unsigned long flags;
int v; int v;
local_irq_save(flags); v = this_cpu_xchg(p->vm_stat_diff[i], 0);
v = p->vm_stat_diff[i]; if (v) {
p->vm_stat_diff[i] = 0;
local_irq_restore(flags);
atomic_long_add(v, &zone->vm_stat[i]); atomic_long_add(v, &zone->vm_stat[i]);
global_diff[i] += v; global_diff[i] += v;
#ifdef CONFIG_NUMA #ifdef CONFIG_NUMA
/* 3 seconds idle till flush */ /* 3 seconds idle till flush */
p->expire = 3; __this_cpu_write(p->expire, 3);
#endif #endif
} }
}
cond_resched(); cond_resched();
#ifdef CONFIG_NUMA #ifdef CONFIG_NUMA
/* /*
@ -476,23 +472,24 @@ static void refresh_cpu_vm_stats(int cpu)
* Check if there are pages remaining in this pageset * Check if there are pages remaining in this pageset
* if not then there is nothing to expire. * if not then there is nothing to expire.
*/ */
if (!p->expire || !p->pcp.count) if (!__this_cpu_read(p->expire) ||
!__this_cpu_read(p->pcp.count))
continue; continue;
/* /*
* We never drain zones local to this processor. * We never drain zones local to this processor.
*/ */
if (zone_to_nid(zone) == numa_node_id()) { if (zone_to_nid(zone) == numa_node_id()) {
p->expire = 0; __this_cpu_write(p->expire, 0);
continue; continue;
} }
p->expire--;
if (p->expire) if (__this_cpu_dec_return(p->expire))
continue; continue;
if (p->pcp.count) if (__this_cpu_read(p->pcp.count))
drain_zone_pages(zone, &p->pcp); drain_zone_pages(zone, __this_cpu_ptr(&p->pcp));
#endif #endif
} }
fold_diff(global_diff); fold_diff(global_diff);
@ -1308,7 +1305,7 @@ int sysctl_stat_interval __read_mostly = HZ;
static void vmstat_update(struct work_struct *w) static void vmstat_update(struct work_struct *w)
{ {
refresh_cpu_vm_stats(smp_processor_id()); refresh_cpu_vm_stats();
schedule_delayed_work(&__get_cpu_var(vmstat_work), schedule_delayed_work(&__get_cpu_var(vmstat_work),
round_jiffies_relative(sysctl_stat_interval)); round_jiffies_relative(sysctl_stat_interval));
} }