mirror of
https://github.com/torvalds/linux.git
synced 2024-11-27 06:31:52 +00:00
6967037baf
The cpu array size calculation uses the NR_CPUS config option, which was recently increased from 64 to 256. With a value of 256, the cpu array will no longer fit into one APPLDATA record and loading the appldata_os module fails with the following error: could not insert 'appldata_os': Cannot allocate memory Use num_possible_cpus() instead of NR_CPUS. For z/VM, this will still result in a value of 64. This is not true for LPAR, but the appldata feature is not available for LPAR. Signed-off-by: Gerald Schaefer <gerald.schaefer@de.ibm.com> Signed-off-by: Martin Schwidefsky <schwidefsky@de.ibm.com>
219 lines
6.0 KiB
C
219 lines
6.0 KiB
C
/*
|
|
* Data gathering module for Linux-VM Monitor Stream, Stage 1.
|
|
* Collects misc. OS related data (CPU utilization, running processes).
|
|
*
|
|
* Copyright IBM Corp. 2003, 2006
|
|
*
|
|
* Author: Gerald Schaefer <gerald.schaefer@de.ibm.com>
|
|
*/
|
|
|
|
#define KMSG_COMPONENT "appldata"
|
|
#define pr_fmt(fmt) KMSG_COMPONENT ": " fmt
|
|
|
|
#include <linux/module.h>
|
|
#include <linux/init.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/errno.h>
|
|
#include <linux/kernel_stat.h>
|
|
#include <linux/netdevice.h>
|
|
#include <linux/sched.h>
|
|
#include <asm/appldata.h>
|
|
#include <asm/smp.h>
|
|
|
|
#include "appldata.h"
|
|
|
|
|
|
#define LOAD_INT(x) ((x) >> FSHIFT)
|
|
#define LOAD_FRAC(x) LOAD_INT(((x) & (FIXED_1-1)) * 100)
|
|
|
|
/*
|
|
* OS data
|
|
*
|
|
* This is accessed as binary data by z/VM. If changes to it can't be avoided,
|
|
* the structure version (product ID, see appldata_base.c) needs to be changed
|
|
* as well and all documentation and z/VM applications using it must be
|
|
* updated.
|
|
*
|
|
* The record layout is documented in the Linux for zSeries Device Drivers
|
|
* book:
|
|
* http://oss.software.ibm.com/developerworks/opensource/linux390/index.shtml
|
|
*/
|
|
struct appldata_os_per_cpu {
|
|
u32 per_cpu_user; /* timer ticks spent in user mode */
|
|
u32 per_cpu_nice; /* ... spent with modified priority */
|
|
u32 per_cpu_system; /* ... spent in kernel mode */
|
|
u32 per_cpu_idle; /* ... spent in idle mode */
|
|
|
|
/* New in 2.6 */
|
|
u32 per_cpu_irq; /* ... spent in interrupts */
|
|
u32 per_cpu_softirq; /* ... spent in softirqs */
|
|
u32 per_cpu_iowait; /* ... spent while waiting for I/O */
|
|
|
|
/* New in modification level 01 */
|
|
u32 per_cpu_steal; /* ... stolen by hypervisor */
|
|
u32 cpu_id; /* number of this CPU */
|
|
} __attribute__((packed));
|
|
|
|
struct appldata_os_data {
|
|
u64 timestamp;
|
|
u32 sync_count_1; /* after VM collected the record data, */
|
|
u32 sync_count_2; /* sync_count_1 and sync_count_2 should be the
|
|
same. If not, the record has been updated on
|
|
the Linux side while VM was collecting the
|
|
(possibly corrupt) data */
|
|
|
|
u32 nr_cpus; /* number of (virtual) CPUs */
|
|
u32 per_cpu_size; /* size of the per-cpu data struct */
|
|
u32 cpu_offset; /* offset of the first per-cpu data struct */
|
|
|
|
u32 nr_running; /* number of runnable threads */
|
|
u32 nr_threads; /* number of threads */
|
|
u32 avenrun[3]; /* average nr. of running processes during */
|
|
/* the last 1, 5 and 15 minutes */
|
|
|
|
/* New in 2.6 */
|
|
u32 nr_iowait; /* number of blocked threads
|
|
(waiting for I/O) */
|
|
|
|
/* per cpu data */
|
|
struct appldata_os_per_cpu os_cpu[0];
|
|
} __attribute__((packed));
|
|
|
|
static struct appldata_os_data *appldata_os_data;
|
|
|
|
static struct appldata_ops ops = {
|
|
.name = "os",
|
|
.record_nr = APPLDATA_RECORD_OS_ID,
|
|
.owner = THIS_MODULE,
|
|
.mod_lvl = {0xF0, 0xF1}, /* EBCDIC "01" */
|
|
};
|
|
|
|
|
|
/*
|
|
* appldata_get_os_data()
|
|
*
|
|
* gather OS data
|
|
*/
|
|
static void appldata_get_os_data(void *data)
|
|
{
|
|
int i, j, rc;
|
|
struct appldata_os_data *os_data;
|
|
unsigned int new_size;
|
|
|
|
os_data = data;
|
|
os_data->sync_count_1++;
|
|
|
|
os_data->nr_threads = nr_threads;
|
|
os_data->nr_running = nr_running();
|
|
os_data->nr_iowait = nr_iowait();
|
|
os_data->avenrun[0] = avenrun[0] + (FIXED_1/200);
|
|
os_data->avenrun[1] = avenrun[1] + (FIXED_1/200);
|
|
os_data->avenrun[2] = avenrun[2] + (FIXED_1/200);
|
|
|
|
j = 0;
|
|
for_each_online_cpu(i) {
|
|
os_data->os_cpu[j].per_cpu_user =
|
|
cputime_to_jiffies(kcpustat_cpu(i).cpustat[CPUTIME_USER]);
|
|
os_data->os_cpu[j].per_cpu_nice =
|
|
cputime_to_jiffies(kcpustat_cpu(i).cpustat[CPUTIME_NICE]);
|
|
os_data->os_cpu[j].per_cpu_system =
|
|
cputime_to_jiffies(kcpustat_cpu(i).cpustat[CPUTIME_SYSTEM]);
|
|
os_data->os_cpu[j].per_cpu_idle =
|
|
cputime_to_jiffies(kcpustat_cpu(i).cpustat[CPUTIME_IDLE]);
|
|
os_data->os_cpu[j].per_cpu_irq =
|
|
cputime_to_jiffies(kcpustat_cpu(i).cpustat[CPUTIME_IRQ]);
|
|
os_data->os_cpu[j].per_cpu_softirq =
|
|
cputime_to_jiffies(kcpustat_cpu(i).cpustat[CPUTIME_SOFTIRQ]);
|
|
os_data->os_cpu[j].per_cpu_iowait =
|
|
cputime_to_jiffies(kcpustat_cpu(i).cpustat[CPUTIME_IOWAIT]);
|
|
os_data->os_cpu[j].per_cpu_steal =
|
|
cputime_to_jiffies(kcpustat_cpu(i).cpustat[CPUTIME_STEAL]);
|
|
os_data->os_cpu[j].cpu_id = i;
|
|
j++;
|
|
}
|
|
|
|
os_data->nr_cpus = j;
|
|
|
|
new_size = sizeof(struct appldata_os_data) +
|
|
(os_data->nr_cpus * sizeof(struct appldata_os_per_cpu));
|
|
if (ops.size != new_size) {
|
|
if (ops.active) {
|
|
rc = appldata_diag(APPLDATA_RECORD_OS_ID,
|
|
APPLDATA_START_INTERVAL_REC,
|
|
(unsigned long) ops.data, new_size,
|
|
ops.mod_lvl);
|
|
if (rc != 0)
|
|
pr_err("Starting a new OS data collection "
|
|
"failed with rc=%d\n", rc);
|
|
|
|
rc = appldata_diag(APPLDATA_RECORD_OS_ID,
|
|
APPLDATA_STOP_REC,
|
|
(unsigned long) ops.data, ops.size,
|
|
ops.mod_lvl);
|
|
if (rc != 0)
|
|
pr_err("Stopping a faulty OS data "
|
|
"collection failed with rc=%d\n", rc);
|
|
}
|
|
ops.size = new_size;
|
|
}
|
|
os_data->timestamp = get_tod_clock();
|
|
os_data->sync_count_2++;
|
|
}
|
|
|
|
|
|
/*
|
|
* appldata_os_init()
|
|
*
|
|
* init data, register ops
|
|
*/
|
|
static int __init appldata_os_init(void)
|
|
{
|
|
int rc, max_size;
|
|
|
|
max_size = sizeof(struct appldata_os_data) +
|
|
(num_possible_cpus() * sizeof(struct appldata_os_per_cpu));
|
|
if (max_size > APPLDATA_MAX_REC_SIZE) {
|
|
pr_err("Maximum OS record size %i exceeds the maximum "
|
|
"record size %i\n", max_size, APPLDATA_MAX_REC_SIZE);
|
|
rc = -ENOMEM;
|
|
goto out;
|
|
}
|
|
|
|
appldata_os_data = kzalloc(max_size, GFP_KERNEL | GFP_DMA);
|
|
if (appldata_os_data == NULL) {
|
|
rc = -ENOMEM;
|
|
goto out;
|
|
}
|
|
|
|
appldata_os_data->per_cpu_size = sizeof(struct appldata_os_per_cpu);
|
|
appldata_os_data->cpu_offset = offsetof(struct appldata_os_data,
|
|
os_cpu);
|
|
|
|
ops.data = appldata_os_data;
|
|
ops.callback = &appldata_get_os_data;
|
|
rc = appldata_register_ops(&ops);
|
|
if (rc != 0)
|
|
kfree(appldata_os_data);
|
|
out:
|
|
return rc;
|
|
}
|
|
|
|
/*
|
|
* appldata_os_exit()
|
|
*
|
|
* unregister ops
|
|
*/
|
|
static void __exit appldata_os_exit(void)
|
|
{
|
|
appldata_unregister_ops(&ops);
|
|
kfree(appldata_os_data);
|
|
}
|
|
|
|
|
|
module_init(appldata_os_init);
|
|
module_exit(appldata_os_exit);
|
|
|
|
MODULE_LICENSE("GPL");
|
|
MODULE_AUTHOR("Gerald Schaefer");
|
|
MODULE_DESCRIPTION("Linux-VM Monitor Stream, OS statistics");
|