forked from Minki/linux
[XFS] using a spinlock per cpu for superblock counter exclusion results in
a preēmpt counter overflow at 256p and above. Change the exclusion mechanism to use atomic bit operations and busy wait loops to emulate the spin lock exclusion mechanism but without the preempt count issues. SGI-PV: 950027 SGI-Modid: xfs-linux-melb:xfs-kern:25338a Signed-off-by: David Chinner <dgc@sgi.com> Signed-off-by: Nathan Scott <nathans@sgi.com>
This commit is contained in:
parent
87cbc49cd4
commit
01e1b69cfc
@ -75,6 +75,7 @@
|
|||||||
#include <linux/sort.h>
|
#include <linux/sort.h>
|
||||||
#include <linux/cpu.h>
|
#include <linux/cpu.h>
|
||||||
#include <linux/notifier.h>
|
#include <linux/notifier.h>
|
||||||
|
#include <linux/delay.h>
|
||||||
|
|
||||||
#include <asm/page.h>
|
#include <asm/page.h>
|
||||||
#include <asm/div64.h>
|
#include <asm/div64.h>
|
||||||
|
@ -1746,10 +1746,7 @@ xfs_icsb_cpu_notify(
|
|||||||
case CPU_UP_PREPARE:
|
case CPU_UP_PREPARE:
|
||||||
/* Easy Case - initialize the area and locks, and
|
/* Easy Case - initialize the area and locks, and
|
||||||
* then rebalance when online does everything else for us. */
|
* then rebalance when online does everything else for us. */
|
||||||
spin_lock_init(&cntp->icsb_lock);
|
memset(cntp, 0, sizeof(xfs_icsb_cnts_t));
|
||||||
cntp->icsb_icount = 0;
|
|
||||||
cntp->icsb_ifree = 0;
|
|
||||||
cntp->icsb_fdblocks = 0;
|
|
||||||
break;
|
break;
|
||||||
case CPU_ONLINE:
|
case CPU_ONLINE:
|
||||||
xfs_icsb_balance_counter(mp, XFS_SBS_ICOUNT, 0);
|
xfs_icsb_balance_counter(mp, XFS_SBS_ICOUNT, 0);
|
||||||
@ -1769,9 +1766,7 @@ xfs_icsb_cpu_notify(
|
|||||||
mp->m_sb.sb_ifree += cntp->icsb_ifree;
|
mp->m_sb.sb_ifree += cntp->icsb_ifree;
|
||||||
mp->m_sb.sb_fdblocks += cntp->icsb_fdblocks;
|
mp->m_sb.sb_fdblocks += cntp->icsb_fdblocks;
|
||||||
|
|
||||||
cntp->icsb_icount = 0;
|
memset(cntp, 0, sizeof(xfs_icsb_cnts_t));
|
||||||
cntp->icsb_ifree = 0;
|
|
||||||
cntp->icsb_fdblocks = 0;
|
|
||||||
|
|
||||||
xfs_icsb_balance_counter(mp, XFS_SBS_ICOUNT, XFS_ICSB_SB_LOCKED);
|
xfs_icsb_balance_counter(mp, XFS_SBS_ICOUNT, XFS_ICSB_SB_LOCKED);
|
||||||
xfs_icsb_balance_counter(mp, XFS_SBS_IFREE, XFS_ICSB_SB_LOCKED);
|
xfs_icsb_balance_counter(mp, XFS_SBS_IFREE, XFS_ICSB_SB_LOCKED);
|
||||||
@ -1800,7 +1795,7 @@ xfs_icsb_init_counters(
|
|||||||
|
|
||||||
for_each_online_cpu(i) {
|
for_each_online_cpu(i) {
|
||||||
cntp = (xfs_icsb_cnts_t *)per_cpu_ptr(mp->m_sb_cnts, i);
|
cntp = (xfs_icsb_cnts_t *)per_cpu_ptr(mp->m_sb_cnts, i);
|
||||||
spin_lock_init(&cntp->icsb_lock);
|
memset(cntp, 0, sizeof(xfs_icsb_cnts_t));
|
||||||
}
|
}
|
||||||
/*
|
/*
|
||||||
* start with all counters disabled so that the
|
* start with all counters disabled so that the
|
||||||
@ -1820,6 +1815,22 @@ xfs_icsb_destroy_counters(
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
STATIC inline void
|
||||||
|
xfs_icsb_lock_cntr(
|
||||||
|
xfs_icsb_cnts_t *icsbp)
|
||||||
|
{
|
||||||
|
while (test_and_set_bit(XFS_ICSB_FLAG_LOCK, &icsbp->icsb_flags)) {
|
||||||
|
ndelay(1000);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
STATIC inline void
|
||||||
|
xfs_icsb_unlock_cntr(
|
||||||
|
xfs_icsb_cnts_t *icsbp)
|
||||||
|
{
|
||||||
|
clear_bit(XFS_ICSB_FLAG_LOCK, &icsbp->icsb_flags);
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
STATIC inline void
|
STATIC inline void
|
||||||
xfs_icsb_lock_all_counters(
|
xfs_icsb_lock_all_counters(
|
||||||
@ -1830,7 +1841,7 @@ xfs_icsb_lock_all_counters(
|
|||||||
|
|
||||||
for_each_online_cpu(i) {
|
for_each_online_cpu(i) {
|
||||||
cntp = (xfs_icsb_cnts_t *)per_cpu_ptr(mp->m_sb_cnts, i);
|
cntp = (xfs_icsb_cnts_t *)per_cpu_ptr(mp->m_sb_cnts, i);
|
||||||
spin_lock(&cntp->icsb_lock);
|
xfs_icsb_lock_cntr(cntp);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -1843,7 +1854,7 @@ xfs_icsb_unlock_all_counters(
|
|||||||
|
|
||||||
for_each_online_cpu(i) {
|
for_each_online_cpu(i) {
|
||||||
cntp = (xfs_icsb_cnts_t *)per_cpu_ptr(mp->m_sb_cnts, i);
|
cntp = (xfs_icsb_cnts_t *)per_cpu_ptr(mp->m_sb_cnts, i);
|
||||||
spin_unlock(&cntp->icsb_lock);
|
xfs_icsb_unlock_cntr(cntp);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -2070,7 +2081,7 @@ xfs_icsb_modify_counters_int(
|
|||||||
again:
|
again:
|
||||||
cpu = get_cpu();
|
cpu = get_cpu();
|
||||||
icsbp = (xfs_icsb_cnts_t *)per_cpu_ptr(mp->m_sb_cnts, cpu),
|
icsbp = (xfs_icsb_cnts_t *)per_cpu_ptr(mp->m_sb_cnts, cpu),
|
||||||
spin_lock(&icsbp->icsb_lock);
|
xfs_icsb_lock_cntr(icsbp);
|
||||||
if (unlikely(xfs_icsb_counter_disabled(mp, field)))
|
if (unlikely(xfs_icsb_counter_disabled(mp, field)))
|
||||||
goto slow_path;
|
goto slow_path;
|
||||||
|
|
||||||
@ -2104,7 +2115,7 @@ again:
|
|||||||
BUG();
|
BUG();
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
spin_unlock(&icsbp->icsb_lock);
|
xfs_icsb_unlock_cntr(icsbp);
|
||||||
put_cpu();
|
put_cpu();
|
||||||
if (locked)
|
if (locked)
|
||||||
XFS_SB_UNLOCK(mp, s);
|
XFS_SB_UNLOCK(mp, s);
|
||||||
@ -2120,7 +2131,7 @@ again:
|
|||||||
* manner.
|
* manner.
|
||||||
*/
|
*/
|
||||||
slow_path:
|
slow_path:
|
||||||
spin_unlock(&icsbp->icsb_lock);
|
xfs_icsb_unlock_cntr(icsbp);
|
||||||
put_cpu();
|
put_cpu();
|
||||||
|
|
||||||
/* need to hold superblock incase we need
|
/* need to hold superblock incase we need
|
||||||
|
@ -280,9 +280,11 @@ typedef struct xfs_icsb_cnts {
|
|||||||
uint64_t icsb_fdblocks;
|
uint64_t icsb_fdblocks;
|
||||||
uint64_t icsb_ifree;
|
uint64_t icsb_ifree;
|
||||||
uint64_t icsb_icount;
|
uint64_t icsb_icount;
|
||||||
spinlock_t icsb_lock;
|
unsigned long icsb_flags;
|
||||||
} xfs_icsb_cnts_t;
|
} xfs_icsb_cnts_t;
|
||||||
|
|
||||||
|
#define XFS_ICSB_FLAG_LOCK (1 << 0) /* counter lock bit */
|
||||||
|
|
||||||
#define XFS_ICSB_SB_LOCKED (1 << 0) /* sb already locked */
|
#define XFS_ICSB_SB_LOCKED (1 << 0) /* sb already locked */
|
||||||
#define XFS_ICSB_LAZY_COUNT (1 << 1) /* accuracy not needed */
|
#define XFS_ICSB_LAZY_COUNT (1 << 1) /* accuracy not needed */
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user