2005-10-19 23:23:26 +00:00
|
|
|
/*
|
|
|
|
* Common time prototypes and such for all ppc machines.
|
|
|
|
*
|
|
|
|
* Written by Cort Dougan (cort@cs.nmt.edu) to merge
|
|
|
|
* Paul Mackerras' version and mine for PReP and Pmac.
|
|
|
|
*
|
|
|
|
* This program is free software; you can redistribute it and/or
|
|
|
|
* modify it under the terms of the GNU General Public License
|
|
|
|
* as published by the Free Software Foundation; either version
|
|
|
|
* 2 of the License, or (at your option) any later version.
|
|
|
|
*/
|
|
|
|
|
|
|
|
#ifndef __POWERPC_TIME_H
|
|
|
|
#define __POWERPC_TIME_H
|
|
|
|
|
|
|
|
#ifdef __KERNEL__
|
|
|
|
#include <linux/types.h>
|
|
|
|
#include <linux/percpu.h>
|
|
|
|
|
|
|
|
#include <asm/processor.h>
|
2006-06-28 01:51:41 +00:00
|
|
|
#ifdef CONFIG_PPC_ISERIES
|
2005-10-19 23:23:26 +00:00
|
|
|
#include <asm/paca.h>
|
2006-06-28 01:51:41 +00:00
|
|
|
#include <asm/firmware.h>
|
2005-11-01 05:59:20 +00:00
|
|
|
#include <asm/iseries/hv_call.h>
|
2005-10-19 23:23:26 +00:00
|
|
|
#endif
|
|
|
|
|
|
|
|
/* time.c */
|
|
|
|
extern unsigned long tb_ticks_per_jiffy;
|
|
|
|
extern unsigned long tb_ticks_per_usec;
|
|
|
|
extern unsigned long tb_ticks_per_sec;
|
|
|
|
extern u64 tb_to_xs;
|
|
|
|
extern unsigned tb_to_us;
|
|
|
|
|
|
|
|
struct rtc_time;
|
|
|
|
extern void to_tm(int tim, struct rtc_time * tm);
|
|
|
|
extern time_t last_rtc_update;
|
|
|
|
|
|
|
|
extern void generic_calibrate_decr(void);
|
|
|
|
extern void wakeup_decrementer(void);
|
powerpc: Implement accurate task and CPU time accounting
This implements accurate task and cpu time accounting for 64-bit
powerpc kernels. Instead of accounting a whole jiffy of time to a
task on a timer interrupt because that task happened to be running at
the time, we now account time in units of timebase ticks according to
the actual time spent by the task in user mode and kernel mode. We
also count the time spent processing hardware and software interrupts
accurately. This is conditional on CONFIG_VIRT_CPU_ACCOUNTING. If
that is not set, we do tick-based approximate accounting as before.
To get this accurate information, we read either the PURR (processor
utilization of resources register) on POWER5 machines, or the timebase
on other machines on
* each entry to the kernel from usermode
* each exit to usermode
* transitions between process context, hard irq context and soft irq
context in kernel mode
* context switches.
On POWER5 systems with shared-processor logical partitioning we also
read both the PURR and the timebase at each timer interrupt and
context switch in order to determine how much time has been taken by
the hypervisor to run other partitions ("steal" time). Unfortunately,
since we need values of the PURR on both threads at the same time to
accurately calculate the steal time, and since we can only calculate
steal time on a per-core basis, the apportioning of the steal time
between idle time (time which we ceded to the hypervisor in the idle
loop) and actual stolen time is somewhat approximate at the moment.
This is all based quite heavily on what s390 does, and it uses the
generic interfaces that were added by the s390 developers,
i.e. account_system_time(), account_user_time(), etc.
This patch doesn't add any new interfaces between the kernel and
userspace, and doesn't change the units in which time is reported to
userspace by things such as /proc/stat, /proc/<pid>/stat, getrusage(),
times(), etc. Internally the various task and cpu times are stored in
timebase units, but they are converted to USER_HZ units (1/100th of a
second) when reported to userspace. Some precision is therefore lost
but there should not be any accumulating error, since the internal
accumulation is at full precision.
Signed-off-by: Paul Mackerras <paulus@samba.org>
2006-02-23 23:06:59 +00:00
|
|
|
extern void snapshot_timebase(void);
|
2005-10-19 23:23:26 +00:00
|
|
|
|
|
|
|
/* Some sane defaults: 125 MHz timebase, 1GHz processor */
|
|
|
|
extern unsigned long ppc_proc_freq;
|
|
|
|
#define DEFAULT_PROC_FREQ (DEFAULT_TB_FREQ * 8)
|
|
|
|
extern unsigned long ppc_tb_freq;
|
|
|
|
#define DEFAULT_TB_FREQ 125000000UL
|
|
|
|
|
|
|
|
/*
|
|
|
|
* By putting all of this stuff into a single struct we
|
|
|
|
* reduce the number of cache lines touched by do_gettimeofday.
|
|
|
|
* Both by collecting all of the data in one cache line and
|
|
|
|
* by touching only one TOC entry on ppc64.
|
|
|
|
*/
|
|
|
|
struct gettimeofday_vars {
|
|
|
|
u64 tb_to_xs;
|
|
|
|
u64 stamp_xsec;
|
|
|
|
u64 tb_orig_stamp;
|
|
|
|
};
|
|
|
|
|
|
|
|
struct gettimeofday_struct {
|
|
|
|
unsigned long tb_ticks_per_sec;
|
|
|
|
struct gettimeofday_vars vars[2];
|
|
|
|
struct gettimeofday_vars * volatile varp;
|
|
|
|
unsigned var_idx;
|
|
|
|
unsigned tb_to_us;
|
|
|
|
};
|
|
|
|
|
|
|
|
struct div_result {
|
|
|
|
u64 result_high;
|
|
|
|
u64 result_low;
|
|
|
|
};
|
|
|
|
|
|
|
|
/* Accessor functions for the timebase (RTC on 601) registers. */
|
|
|
|
/* If one day CONFIG_POWER is added just define __USE_RTC as 1 */
|
|
|
|
#ifdef CONFIG_6xx
|
2005-10-20 12:33:06 +00:00
|
|
|
#define __USE_RTC() (!cpu_has_feature(CPU_FTR_USE_TB))
|
2005-10-19 23:23:26 +00:00
|
|
|
#else
|
|
|
|
#define __USE_RTC() 0
|
|
|
|
#endif
|
|
|
|
|
2006-10-20 04:37:05 +00:00
|
|
|
#ifdef CONFIG_PPC64
|
|
|
|
|
|
|
|
/* For compatibility, get_tbl() is defined as get_tb() on ppc64 */
|
|
|
|
#define get_tbl get_tb
|
|
|
|
|
|
|
|
#else
|
|
|
|
|
2005-10-19 23:23:26 +00:00
|
|
|
static inline unsigned long get_tbl(void)
|
|
|
|
{
|
|
|
|
#if defined(CONFIG_403GCX)
|
2006-10-20 04:37:05 +00:00
|
|
|
unsigned long tbl;
|
2005-10-19 23:23:26 +00:00
|
|
|
asm volatile("mfspr %0, 0x3dd" : "=r" (tbl));
|
2006-10-20 04:37:05 +00:00
|
|
|
return tbl;
|
2005-10-19 23:23:26 +00:00
|
|
|
#else
|
2006-10-20 04:37:05 +00:00
|
|
|
return mftbl();
|
2005-10-19 23:23:26 +00:00
|
|
|
#endif
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline unsigned int get_tbu(void)
|
|
|
|
{
|
2006-10-20 04:37:05 +00:00
|
|
|
#ifdef CONFIG_403GCX
|
2005-10-19 23:23:26 +00:00
|
|
|
unsigned int tbu;
|
|
|
|
asm volatile("mfspr %0, 0x3dc" : "=r" (tbu));
|
2006-10-20 04:37:05 +00:00
|
|
|
return tbu;
|
2005-10-19 23:23:26 +00:00
|
|
|
#else
|
2006-10-20 04:37:05 +00:00
|
|
|
return mftbu();
|
2005-10-19 23:23:26 +00:00
|
|
|
#endif
|
|
|
|
}
|
2006-10-20 04:37:05 +00:00
|
|
|
#endif /* !CONFIG_PPC64 */
|
2005-10-19 23:23:26 +00:00
|
|
|
|
|
|
|
static inline unsigned int get_rtcl(void)
|
|
|
|
{
|
|
|
|
unsigned int rtcl;
|
|
|
|
|
|
|
|
asm volatile("mfrtcl %0" : "=r" (rtcl));
|
|
|
|
return rtcl;
|
|
|
|
}
|
|
|
|
|
2005-10-23 07:14:56 +00:00
|
|
|
static inline u64 get_rtc(void)
|
|
|
|
{
|
|
|
|
unsigned int hi, lo, hi2;
|
|
|
|
|
|
|
|
do {
|
|
|
|
asm volatile("mfrtcu %0; mfrtcl %1; mfrtcu %2"
|
|
|
|
: "=r" (hi), "=r" (lo), "=r" (hi2));
|
|
|
|
} while (hi2 != hi);
|
|
|
|
return (u64)hi * 1000000000 + lo;
|
|
|
|
}
|
|
|
|
|
2005-10-19 23:23:26 +00:00
|
|
|
#ifdef CONFIG_PPC64
|
|
|
|
static inline u64 get_tb(void)
|
|
|
|
{
|
|
|
|
return mftb();
|
|
|
|
}
|
2006-10-20 04:37:05 +00:00
|
|
|
#else /* CONFIG_PPC64 */
|
2005-10-19 23:23:26 +00:00
|
|
|
static inline u64 get_tb(void)
|
|
|
|
{
|
|
|
|
unsigned int tbhi, tblo, tbhi2;
|
|
|
|
|
|
|
|
do {
|
|
|
|
tbhi = get_tbu();
|
|
|
|
tblo = get_tbl();
|
|
|
|
tbhi2 = get_tbu();
|
|
|
|
} while (tbhi != tbhi2);
|
|
|
|
|
|
|
|
return ((u64)tbhi << 32) | tblo;
|
|
|
|
}
|
2006-10-20 04:37:05 +00:00
|
|
|
#endif /* !CONFIG_PPC64 */
|
2005-10-19 23:23:26 +00:00
|
|
|
|
|
|
|
static inline void set_tb(unsigned int upper, unsigned int lower)
|
|
|
|
{
|
|
|
|
mtspr(SPRN_TBWL, 0);
|
|
|
|
mtspr(SPRN_TBWU, upper);
|
|
|
|
mtspr(SPRN_TBWL, lower);
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Accessor functions for the decrementer register.
|
|
|
|
* The 4xx doesn't even have a decrementer. I tried to use the
|
|
|
|
* generic timer interrupt code, which seems OK, with the 4xx PIT
|
|
|
|
* in auto-reload mode. The problem is PIT stops counting when it
|
|
|
|
* hits zero. If it would wrap, we could use it just like a decrementer.
|
|
|
|
*/
|
|
|
|
static inline unsigned int get_dec(void)
|
|
|
|
{
|
|
|
|
#if defined(CONFIG_40x)
|
|
|
|
return (mfspr(SPRN_PIT));
|
|
|
|
#else
|
|
|
|
return (mfspr(SPRN_DEC));
|
|
|
|
#endif
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void set_dec(int val)
|
|
|
|
{
|
|
|
|
#if defined(CONFIG_40x)
|
|
|
|
return; /* Have to let it auto-reload */
|
|
|
|
#elif defined(CONFIG_8xx_CPU6)
|
|
|
|
set_dec_cpu6(val);
|
|
|
|
#else
|
|
|
|
#ifdef CONFIG_PPC_ISERIES
|
|
|
|
int cur_dec;
|
|
|
|
|
2006-06-28 01:51:41 +00:00
|
|
|
if (firmware_has_feature(FW_FEATURE_ISERIES) &&
|
|
|
|
get_lppaca()->shared_proc) {
|
2006-01-12 23:26:42 +00:00
|
|
|
get_lppaca()->virtual_decr = val;
|
2005-10-19 23:23:26 +00:00
|
|
|
cur_dec = get_dec();
|
|
|
|
if (cur_dec > val)
|
|
|
|
HvCall_setVirtualDecr();
|
|
|
|
} else
|
|
|
|
#endif
|
|
|
|
mtspr(SPRN_DEC, val);
|
|
|
|
#endif /* not 40x or 8xx_CPU6 */
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline unsigned long tb_ticks_since(unsigned long tstamp)
|
|
|
|
{
|
|
|
|
if (__USE_RTC()) {
|
|
|
|
int delta = get_rtcl() - (unsigned int) tstamp;
|
|
|
|
return delta < 0 ? delta + 1000000000 : delta;
|
|
|
|
}
|
|
|
|
return get_tbl() - tstamp;
|
|
|
|
}
|
|
|
|
|
|
|
|
#define mulhwu(x,y) \
|
|
|
|
({unsigned z; asm ("mulhwu %0,%1,%2" : "=r" (z) : "r" (x), "r" (y)); z;})
|
|
|
|
|
|
|
|
#ifdef CONFIG_PPC64
|
|
|
|
#define mulhdu(x,y) \
|
|
|
|
({unsigned long z; asm ("mulhdu %0,%1,%2" : "=r" (z) : "r" (x), "r" (y)); z;})
|
|
|
|
#else
|
|
|
|
extern u64 mulhdu(u64, u64);
|
|
|
|
#endif
|
|
|
|
|
2005-10-22 04:55:23 +00:00
|
|
|
extern void smp_space_timers(unsigned int);
|
|
|
|
|
|
|
|
extern unsigned mulhwu_scale_factor(unsigned, unsigned);
|
|
|
|
extern void div128_by_32(u64 dividend_high, u64 dividend_low,
|
|
|
|
unsigned divisor, struct div_result *dr);
|
2005-10-19 23:23:26 +00:00
|
|
|
|
|
|
|
/* Used to store Processor Utilization register (purr) values */
|
|
|
|
|
|
|
|
struct cpu_usage {
|
|
|
|
u64 current_tb; /* Holds the current purr register values */
|
|
|
|
};
|
|
|
|
|
|
|
|
DECLARE_PER_CPU(struct cpu_usage, cpu_usage_array);
|
|
|
|
|
powerpc: Implement accurate task and CPU time accounting
This implements accurate task and cpu time accounting for 64-bit
powerpc kernels. Instead of accounting a whole jiffy of time to a
task on a timer interrupt because that task happened to be running at
the time, we now account time in units of timebase ticks according to
the actual time spent by the task in user mode and kernel mode. We
also count the time spent processing hardware and software interrupts
accurately. This is conditional on CONFIG_VIRT_CPU_ACCOUNTING. If
that is not set, we do tick-based approximate accounting as before.
To get this accurate information, we read either the PURR (processor
utilization of resources register) on POWER5 machines, or the timebase
on other machines on
* each entry to the kernel from usermode
* each exit to usermode
* transitions between process context, hard irq context and soft irq
context in kernel mode
* context switches.
On POWER5 systems with shared-processor logical partitioning we also
read both the PURR and the timebase at each timer interrupt and
context switch in order to determine how much time has been taken by
the hypervisor to run other partitions ("steal" time). Unfortunately,
since we need values of the PURR on both threads at the same time to
accurately calculate the steal time, and since we can only calculate
steal time on a per-core basis, the apportioning of the steal time
between idle time (time which we ceded to the hypervisor in the idle
loop) and actual stolen time is somewhat approximate at the moment.
This is all based quite heavily on what s390 does, and it uses the
generic interfaces that were added by the s390 developers,
i.e. account_system_time(), account_user_time(), etc.
This patch doesn't add any new interfaces between the kernel and
userspace, and doesn't change the units in which time is reported to
userspace by things such as /proc/stat, /proc/<pid>/stat, getrusage(),
times(), etc. Internally the various task and cpu times are stored in
timebase units, but they are converted to USER_HZ units (1/100th of a
second) when reported to userspace. Some precision is therefore lost
but there should not be any accumulating error, since the internal
accumulation is at full precision.
Signed-off-by: Paul Mackerras <paulus@samba.org>
2006-02-23 23:06:59 +00:00
|
|
|
#ifdef CONFIG_VIRT_CPU_ACCOUNTING
|
|
|
|
extern void account_process_vtime(struct task_struct *tsk);
|
|
|
|
#else
|
|
|
|
#define account_process_vtime(tsk) do { } while (0)
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#if defined(CONFIG_VIRT_CPU_ACCOUNTING) && defined(CONFIG_PPC_SPLPAR)
|
|
|
|
extern void calculate_steal_time(void);
|
|
|
|
extern void snapshot_timebases(void);
|
|
|
|
#else
|
|
|
|
#define calculate_steal_time() do { } while (0)
|
|
|
|
#define snapshot_timebases() do { } while (0)
|
|
|
|
#endif
|
|
|
|
|
2005-10-19 23:23:26 +00:00
|
|
|
#endif /* __KERNEL__ */
|
2006-09-26 22:46:37 +00:00
|
|
|
#endif /* __POWERPC_TIME_H */
|