forked from Minki/linux
faf01aef05
At the moment KVM on PPC creates 4 types of entries under the kvm debugfs:
1) "%pid-%fd" per a KVM instance (for all platforms);
2) "vm%pid" (for PPC Book3s HV KVM);
3) "vm%u_vcpu%u_timing" (for PPC Book3e KVM);
4) "kvm-xive-%p" (for XIVE PPC Book3s KVM, the same for XICS);
The problem with this is that multiple VMs per process is not allowed for
2) and 3) which makes it possible for userspace to trigger errors when
creating duplicated debugfs entries.
This merges all these into 1).
This defines kvm_arch_create_kvm_debugfs() similar to
kvm_arch_create_vcpu_debugfs().
This defines 2 hooks in kvmppc_ops that allow specific KVM implementations
add necessary entries, this adds the _e500 suffix to
kvmppc_create_vcpu_debugfs_e500() to make it clear what platform it is for.
This makes use of already existing kvm_arch_create_vcpu_debugfs() on PPC.
This removes no more used debugfs_dir pointers from PPC kvm_arch structs.
This stops removing vcpu entries as once created vcpus stay around
for the entire life of a VM and removed when the KVM instance is closed,
see commit d56f5136b0
("KVM: let kvm_destroy_vm_debugfs clean up vCPU
debugfs directories").
Suggested-by: Fabiano Rosas <farosas@linux.ibm.com>
Signed-off-by: Alexey Kardashevskiy <aik@ozlabs.ru>
Reviewed-by: Cédric Le Goater <clg@kaod.org>
Signed-off-by: Michael Ellerman <mpe@ellerman.id.au>
Link: https://lore.kernel.org/r/20220111005404.162219-1-aik@ozlabs.ru
214 lines
6.0 KiB
C
214 lines
6.0 KiB
C
// SPDX-License-Identifier: GPL-2.0-only
|
|
/*
|
|
*
|
|
* Copyright IBM Corp. 2008
|
|
*
|
|
* Authors: Hollis Blanchard <hollisb@us.ibm.com>
|
|
* Christian Ehrhardt <ehrhardt@linux.vnet.ibm.com>
|
|
*/
|
|
|
|
#include <linux/kvm_host.h>
|
|
#include <linux/fs.h>
|
|
#include <linux/seq_file.h>
|
|
#include <linux/debugfs.h>
|
|
#include <linux/uaccess.h>
|
|
#include <linux/module.h>
|
|
|
|
#include <asm/time.h>
|
|
#include <asm-generic/div64.h>
|
|
|
|
#include "timing.h"
|
|
|
|
void kvmppc_init_timing_stats(struct kvm_vcpu *vcpu)
|
|
{
|
|
int i;
|
|
|
|
/* Take a lock to avoid concurrent updates */
|
|
mutex_lock(&vcpu->arch.exit_timing_lock);
|
|
|
|
vcpu->arch.last_exit_type = 0xDEAD;
|
|
for (i = 0; i < __NUMBER_OF_KVM_EXIT_TYPES; i++) {
|
|
vcpu->arch.timing_count_type[i] = 0;
|
|
vcpu->arch.timing_max_duration[i] = 0;
|
|
vcpu->arch.timing_min_duration[i] = 0xFFFFFFFF;
|
|
vcpu->arch.timing_sum_duration[i] = 0;
|
|
vcpu->arch.timing_sum_quad_duration[i] = 0;
|
|
}
|
|
vcpu->arch.timing_last_exit = 0;
|
|
vcpu->arch.timing_exit.tv64 = 0;
|
|
vcpu->arch.timing_last_enter.tv64 = 0;
|
|
|
|
mutex_unlock(&vcpu->arch.exit_timing_lock);
|
|
}
|
|
|
|
static void add_exit_timing(struct kvm_vcpu *vcpu, u64 duration, int type)
|
|
{
|
|
u64 old;
|
|
|
|
mutex_lock(&vcpu->arch.exit_timing_lock);
|
|
|
|
vcpu->arch.timing_count_type[type]++;
|
|
|
|
/* sum */
|
|
old = vcpu->arch.timing_sum_duration[type];
|
|
vcpu->arch.timing_sum_duration[type] += duration;
|
|
if (unlikely(old > vcpu->arch.timing_sum_duration[type])) {
|
|
printk(KERN_ERR"%s - wrap adding sum of durations"
|
|
" old %lld new %lld type %d exit # of type %d\n",
|
|
__func__, old, vcpu->arch.timing_sum_duration[type],
|
|
type, vcpu->arch.timing_count_type[type]);
|
|
}
|
|
|
|
/* square sum */
|
|
old = vcpu->arch.timing_sum_quad_duration[type];
|
|
vcpu->arch.timing_sum_quad_duration[type] += (duration*duration);
|
|
if (unlikely(old > vcpu->arch.timing_sum_quad_duration[type])) {
|
|
printk(KERN_ERR"%s - wrap adding sum of squared durations"
|
|
" old %lld new %lld type %d exit # of type %d\n",
|
|
__func__, old,
|
|
vcpu->arch.timing_sum_quad_duration[type],
|
|
type, vcpu->arch.timing_count_type[type]);
|
|
}
|
|
|
|
/* set min/max */
|
|
if (unlikely(duration < vcpu->arch.timing_min_duration[type]))
|
|
vcpu->arch.timing_min_duration[type] = duration;
|
|
if (unlikely(duration > vcpu->arch.timing_max_duration[type]))
|
|
vcpu->arch.timing_max_duration[type] = duration;
|
|
|
|
mutex_unlock(&vcpu->arch.exit_timing_lock);
|
|
}
|
|
|
|
void kvmppc_update_timing_stats(struct kvm_vcpu *vcpu)
|
|
{
|
|
u64 exit = vcpu->arch.timing_last_exit;
|
|
u64 enter = vcpu->arch.timing_last_enter.tv64;
|
|
|
|
/* save exit time, used next exit when the reenter time is known */
|
|
vcpu->arch.timing_last_exit = vcpu->arch.timing_exit.tv64;
|
|
|
|
if (unlikely(vcpu->arch.last_exit_type == 0xDEAD || exit == 0))
|
|
return; /* skip incomplete cycle (e.g. after reset) */
|
|
|
|
/* update statistics for average and standard deviation */
|
|
add_exit_timing(vcpu, (enter - exit), vcpu->arch.last_exit_type);
|
|
/* enter -> timing_last_exit is time spent in guest - log this too */
|
|
add_exit_timing(vcpu, (vcpu->arch.timing_last_exit - enter),
|
|
TIMEINGUEST);
|
|
}
|
|
|
|
static const char *kvm_exit_names[__NUMBER_OF_KVM_EXIT_TYPES] = {
|
|
[MMIO_EXITS] = "MMIO",
|
|
[SIGNAL_EXITS] = "SIGNAL",
|
|
[ITLB_REAL_MISS_EXITS] = "ITLBREAL",
|
|
[ITLB_VIRT_MISS_EXITS] = "ITLBVIRT",
|
|
[DTLB_REAL_MISS_EXITS] = "DTLBREAL",
|
|
[DTLB_VIRT_MISS_EXITS] = "DTLBVIRT",
|
|
[SYSCALL_EXITS] = "SYSCALL",
|
|
[ISI_EXITS] = "ISI",
|
|
[DSI_EXITS] = "DSI",
|
|
[EMULATED_INST_EXITS] = "EMULINST",
|
|
[EMULATED_MTMSRWE_EXITS] = "EMUL_WAIT",
|
|
[EMULATED_WRTEE_EXITS] = "EMUL_WRTEE",
|
|
[EMULATED_MTSPR_EXITS] = "EMUL_MTSPR",
|
|
[EMULATED_MFSPR_EXITS] = "EMUL_MFSPR",
|
|
[EMULATED_MTMSR_EXITS] = "EMUL_MTMSR",
|
|
[EMULATED_MFMSR_EXITS] = "EMUL_MFMSR",
|
|
[EMULATED_TLBSX_EXITS] = "EMUL_TLBSX",
|
|
[EMULATED_TLBWE_EXITS] = "EMUL_TLBWE",
|
|
[EMULATED_RFI_EXITS] = "EMUL_RFI",
|
|
[DEC_EXITS] = "DEC",
|
|
[EXT_INTR_EXITS] = "EXTINT",
|
|
[HALT_WAKEUP] = "HALT",
|
|
[USR_PR_INST] = "USR_PR_INST",
|
|
[FP_UNAVAIL] = "FP_UNAVAIL",
|
|
[DEBUG_EXITS] = "DEBUG",
|
|
[TIMEINGUEST] = "TIMEINGUEST"
|
|
};
|
|
|
|
static int kvmppc_exit_timing_show(struct seq_file *m, void *private)
|
|
{
|
|
struct kvm_vcpu *vcpu = m->private;
|
|
int i;
|
|
u64 min, max, sum, sum_quad;
|
|
|
|
seq_puts(m, "type count min max sum sum_squared\n");
|
|
|
|
for (i = 0; i < __NUMBER_OF_KVM_EXIT_TYPES; i++) {
|
|
|
|
min = vcpu->arch.timing_min_duration[i];
|
|
do_div(min, tb_ticks_per_usec);
|
|
max = vcpu->arch.timing_max_duration[i];
|
|
do_div(max, tb_ticks_per_usec);
|
|
sum = vcpu->arch.timing_sum_duration[i];
|
|
do_div(sum, tb_ticks_per_usec);
|
|
sum_quad = vcpu->arch.timing_sum_quad_duration[i];
|
|
do_div(sum_quad, tb_ticks_per_usec);
|
|
|
|
seq_printf(m, "%12s %10d %10lld %10lld %20lld %20lld\n",
|
|
kvm_exit_names[i],
|
|
vcpu->arch.timing_count_type[i],
|
|
min,
|
|
max,
|
|
sum,
|
|
sum_quad);
|
|
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
/* Write 'c' to clear the timing statistics. */
|
|
static ssize_t kvmppc_exit_timing_write(struct file *file,
|
|
const char __user *user_buf,
|
|
size_t count, loff_t *ppos)
|
|
{
|
|
int err = -EINVAL;
|
|
char c;
|
|
|
|
if (count > 1) {
|
|
goto done;
|
|
}
|
|
|
|
if (get_user(c, user_buf)) {
|
|
err = -EFAULT;
|
|
goto done;
|
|
}
|
|
|
|
if (c == 'c') {
|
|
struct seq_file *seqf = file->private_data;
|
|
struct kvm_vcpu *vcpu = seqf->private;
|
|
/* Write does not affect our buffers previously generated with
|
|
* show. seq_file is locked here to prevent races of init with
|
|
* a show call */
|
|
mutex_lock(&seqf->lock);
|
|
kvmppc_init_timing_stats(vcpu);
|
|
mutex_unlock(&seqf->lock);
|
|
err = count;
|
|
}
|
|
|
|
done:
|
|
return err;
|
|
}
|
|
|
|
static int kvmppc_exit_timing_open(struct inode *inode, struct file *file)
|
|
{
|
|
return single_open(file, kvmppc_exit_timing_show, inode->i_private);
|
|
}
|
|
|
|
static const struct file_operations kvmppc_exit_timing_fops = {
|
|
.owner = THIS_MODULE,
|
|
.open = kvmppc_exit_timing_open,
|
|
.read = seq_read,
|
|
.write = kvmppc_exit_timing_write,
|
|
.llseek = seq_lseek,
|
|
.release = single_release,
|
|
};
|
|
|
|
int kvmppc_create_vcpu_debugfs_e500(struct kvm_vcpu *vcpu,
|
|
struct dentry *debugfs_dentry)
|
|
{
|
|
debugfs_create_file("timing", 0666, debugfs_dentry,
|
|
vcpu, &kvmppc_exit_timing_fops);
|
|
return 0;
|
|
}
|