mirror of
https://github.com/torvalds/linux.git
synced 2024-11-25 13:41:51 +00:00
7d6e3fa87e
Core changes: - The usual set of small fixes and improvements all over the place, but nothing outstanding MSI changes: - Further consolidation of the PCI/MSI interrupt chip code - Make MSI sysfs code independent of PCI/MSI and expose the MSI interrupts of platform devices in the same way as PCI exposes them. Driver changes: - Support for ARM GICv3 EPPI partitions - Treewide conversion to generic_handle_domain_irq() for all chained interrupt controllers - Conversion to bitmap_zalloc() throughout the irq chip drivers - The usual set of small fixes and improvements -----BEGIN PGP SIGNATURE----- iQJHBAABCgAxFiEEQp8+kY+LLUocC4bMphj1TA10mKEFAmEsnpsTHHRnbHhAbGlu dXRyb25peC5kZQAKCRCmGPVMDXSYoS+/EACQdpRkzl3IDIYqThxVZ8KQzp2rKKVn qisAQiWg/6koNJx/yYy62KNAUyKjCIObNtRnWi7OAOx6OvNtQTD2WOLAwkh3Pgw1 8ePYYl55k+yCs8VoITsZM9jYeO+Tk878pU2A6R943zR+g6G7bskGJrxEyZ9TbzIe qKfusNKnRY9/jMQaRALUAAtA9VIVR867GqORX5X8hKz8yE2rqlpb4y+1CFba5BTV Vlxw7cIXvXBn7BKAom5diRqEGDNJEbX+56jJ7yDZshgLo7m11D7QLw72kmb6TNVC g7PchvFi4afpc1ifEAAp0tk4RiSIAQ91nS3n0+jLcLbodOjIkl14eY02ZCJGAP29 uslyzUbmy1wgejG6CA63JtZ4MYdrf/OSMGuoN78qnOKYcIsWFzOvlJmBWWNW34qW LCaUF9QdJ/slXu6B4vIx30GfN9q4myml8bFUobE5q9mBRrEk4R0B7iyBvPu1xKYr ZEan67prI5VEu+afJGpp4r294m4HNVkMLfl3nYmE5+y4MoLeMNKDY3IPTvI9iP4G kaFgoPvQo23WnuclNYpJ+CaA4aRASlB2nTY+oAXIYfehbey9EW5vq4/EK864ek6w oyUTepxxNhE81tG2jpQbf2tR4COsEHy986clxqPP4AvsZXcbypCw8O2FcflpQbHO 5DLEAfTmp7cziQ== =qyll -----END PGP SIGNATURE----- Merge tag 'irq-core-2021-08-30' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip Pull irq updates from Thomas Gleixner: "Updates to the interrupt core and driver subsystems: Core changes: - The usual set of small fixes and improvements all over the place, but nothing stands out MSI changes: - Further consolidation of the PCI/MSI interrupt chip code - Make MSI sysfs code independent of PCI/MSI and expose the MSI interrupts of platform devices in the same way as PCI exposes them. Driver changes: - Support for ARM GICv3 EPPI partitions - Treewide conversion to generic_handle_domain_irq() for all chained interrupt controllers - Conversion to bitmap_zalloc() throughout the irq chip drivers - The usual set of small fixes and improvements" * tag 'irq-core-2021-08-30' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip: (57 commits) platform-msi: Add ABI to show msi_irqs of platform devices genirq/msi: Move MSI sysfs handling from PCI to MSI core genirq/cpuhotplug: Demote debug printk to KERN_DEBUG irqchip/qcom-pdc: Trim unused levels of the interrupt hierarchy irqdomain: Export irq_domain_disconnect_hierarchy() irqchip/gic-v3: Fix priority comparison when non-secure priorities are used irqchip/apple-aic: Fix irq_disable from within irq handlers pinctrl/rockchip: drop the gpio related codes gpio/rockchip: drop irq_gc_lock/irq_gc_unlock for irq set type gpio/rockchip: support next version gpio controller gpio/rockchip: use struct rockchip_gpio_regs for gpio controller gpio/rockchip: add driver for rockchip gpio dt-bindings: gpio: change items restriction of clock for rockchip,gpio-bank pinctrl/rockchip: add pinctrl device to gpio bank struct pinctrl/rockchip: separate struct rockchip_pin_bank to a head file pinctrl/rockchip: always enable clock for gpio controller genirq: Fix kernel doc indentation EDAC/altera: Convert to generic_handle_domain_irq() powerpc: Bulk conversion to generic_handle_domain_irq() nios2: Bulk conversion to generic_handle_domain_irq() ...
959 lines
25 KiB
C
959 lines
25 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
// Copyright (C) 2016, Linaro Ltd - Daniel Lezcano <daniel.lezcano@linaro.org>
|
|
#define pr_fmt(fmt) "irq_timings: " fmt
|
|
|
|
#include <linux/kernel.h>
|
|
#include <linux/percpu.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/static_key.h>
|
|
#include <linux/init.h>
|
|
#include <linux/interrupt.h>
|
|
#include <linux/idr.h>
|
|
#include <linux/irq.h>
|
|
#include <linux/math64.h>
|
|
#include <linux/log2.h>
|
|
|
|
#include <trace/events/irq.h>
|
|
|
|
#include "internals.h"
|
|
|
|
DEFINE_STATIC_KEY_FALSE(irq_timing_enabled);
|
|
|
|
DEFINE_PER_CPU(struct irq_timings, irq_timings);
|
|
|
|
static DEFINE_IDR(irqt_stats);
|
|
|
|
void irq_timings_enable(void)
|
|
{
|
|
static_branch_enable(&irq_timing_enabled);
|
|
}
|
|
|
|
void irq_timings_disable(void)
|
|
{
|
|
static_branch_disable(&irq_timing_enabled);
|
|
}
|
|
|
|
/*
|
|
* The main goal of this algorithm is to predict the next interrupt
|
|
* occurrence on the current CPU.
|
|
*
|
|
* Currently, the interrupt timings are stored in a circular array
|
|
* buffer every time there is an interrupt, as a tuple: the interrupt
|
|
* number and the associated timestamp when the event occurred <irq,
|
|
* timestamp>.
|
|
*
|
|
* For every interrupt occurring in a short period of time, we can
|
|
* measure the elapsed time between the occurrences for the same
|
|
* interrupt and we end up with a suite of intervals. The experience
|
|
* showed the interrupts are often coming following a periodic
|
|
* pattern.
|
|
*
|
|
* The objective of the algorithm is to find out this periodic pattern
|
|
* in a fastest way and use its period to predict the next irq event.
|
|
*
|
|
* When the next interrupt event is requested, we are in the situation
|
|
* where the interrupts are disabled and the circular buffer
|
|
* containing the timings is filled with the events which happened
|
|
* after the previous next-interrupt-event request.
|
|
*
|
|
* At this point, we read the circular buffer and we fill the irq
|
|
* related statistics structure. After this step, the circular array
|
|
* containing the timings is empty because all the values are
|
|
* dispatched in their corresponding buffers.
|
|
*
|
|
* Now for each interrupt, we can predict the next event by using the
|
|
* suffix array, log interval and exponential moving average
|
|
*
|
|
* 1. Suffix array
|
|
*
|
|
* Suffix array is an array of all the suffixes of a string. It is
|
|
* widely used as a data structure for compression, text search, ...
|
|
* For instance for the word 'banana', the suffixes will be: 'banana'
|
|
* 'anana' 'nana' 'ana' 'na' 'a'
|
|
*
|
|
* Usually, the suffix array is sorted but for our purpose it is
|
|
* not necessary and won't provide any improvement in the context of
|
|
* the solved problem where we clearly define the boundaries of the
|
|
* search by a max period and min period.
|
|
*
|
|
* The suffix array will build a suite of intervals of different
|
|
* length and will look for the repetition of each suite. If the suite
|
|
* is repeating then we have the period because it is the length of
|
|
* the suite whatever its position in the buffer.
|
|
*
|
|
* 2. Log interval
|
|
*
|
|
* We saw the irq timings allow to compute the interval of the
|
|
* occurrences for a specific interrupt. We can reasonably assume the
|
|
* longer is the interval, the higher is the error for the next event
|
|
* and we can consider storing those interval values into an array
|
|
* where each slot in the array correspond to an interval at the power
|
|
* of 2 of the index. For example, index 12 will contain values
|
|
* between 2^11 and 2^12.
|
|
*
|
|
* At the end we have an array of values where at each index defines a
|
|
* [2^index - 1, 2 ^ index] interval values allowing to store a large
|
|
* number of values inside a small array.
|
|
*
|
|
* For example, if we have the value 1123, then we store it at
|
|
* ilog2(1123) = 10 index value.
|
|
*
|
|
* Storing those value at the specific index is done by computing an
|
|
* exponential moving average for this specific slot. For instance,
|
|
* for values 1800, 1123, 1453, ... fall under the same slot (10) and
|
|
* the exponential moving average is computed every time a new value
|
|
* is stored at this slot.
|
|
*
|
|
* 3. Exponential Moving Average
|
|
*
|
|
* The EMA is largely used to track a signal for stocks or as a low
|
|
* pass filter. The magic of the formula, is it is very simple and the
|
|
* reactivity of the average can be tuned with the factors called
|
|
* alpha.
|
|
*
|
|
* The higher the alphas are, the faster the average respond to the
|
|
* signal change. In our case, if a slot in the array is a big
|
|
* interval, we can have numbers with a big difference between
|
|
* them. The impact of those differences in the average computation
|
|
* can be tuned by changing the alpha value.
|
|
*
|
|
*
|
|
* -- The algorithm --
|
|
*
|
|
* We saw the different processing above, now let's see how they are
|
|
* used together.
|
|
*
|
|
* For each interrupt:
|
|
* For each interval:
|
|
* Compute the index = ilog2(interval)
|
|
* Compute a new_ema(buffer[index], interval)
|
|
* Store the index in a circular buffer
|
|
*
|
|
* Compute the suffix array of the indexes
|
|
*
|
|
* For each suffix:
|
|
* If the suffix is reverse-found 3 times
|
|
* Return suffix
|
|
*
|
|
* Return Not found
|
|
*
|
|
* However we can not have endless suffix array to be build, it won't
|
|
* make sense and it will add an extra overhead, so we can restrict
|
|
* this to a maximum suffix length of 5 and a minimum suffix length of
|
|
* 2. The experience showed 5 is the majority of the maximum pattern
|
|
* period found for different devices.
|
|
*
|
|
* The result is a pattern finding less than 1us for an interrupt.
|
|
*
|
|
* Example based on real values:
|
|
*
|
|
* Example 1 : MMC write/read interrupt interval:
|
|
*
|
|
* 223947, 1240, 1384, 1386, 1386,
|
|
* 217416, 1236, 1384, 1386, 1387,
|
|
* 214719, 1241, 1386, 1387, 1384,
|
|
* 213696, 1234, 1384, 1386, 1388,
|
|
* 219904, 1240, 1385, 1389, 1385,
|
|
* 212240, 1240, 1386, 1386, 1386,
|
|
* 214415, 1236, 1384, 1386, 1387,
|
|
* 214276, 1234, 1384, 1388, ?
|
|
*
|
|
* For each element, apply ilog2(value)
|
|
*
|
|
* 15, 8, 8, 8, 8,
|
|
* 15, 8, 8, 8, 8,
|
|
* 15, 8, 8, 8, 8,
|
|
* 15, 8, 8, 8, 8,
|
|
* 15, 8, 8, 8, 8,
|
|
* 15, 8, 8, 8, 8,
|
|
* 15, 8, 8, 8, 8,
|
|
* 15, 8, 8, 8, ?
|
|
*
|
|
* Max period of 5, we take the last (max_period * 3) 15 elements as
|
|
* we can be confident if the pattern repeats itself three times it is
|
|
* a repeating pattern.
|
|
*
|
|
* 8,
|
|
* 15, 8, 8, 8, 8,
|
|
* 15, 8, 8, 8, 8,
|
|
* 15, 8, 8, 8, ?
|
|
*
|
|
* Suffixes are:
|
|
*
|
|
* 1) 8, 15, 8, 8, 8 <- max period
|
|
* 2) 8, 15, 8, 8
|
|
* 3) 8, 15, 8
|
|
* 4) 8, 15 <- min period
|
|
*
|
|
* From there we search the repeating pattern for each suffix.
|
|
*
|
|
* buffer: 8, 15, 8, 8, 8, 8, 15, 8, 8, 8, 8, 15, 8, 8, 8
|
|
* | | | | | | | | | | | | | | |
|
|
* 8, 15, 8, 8, 8 | | | | | | | | | |
|
|
* 8, 15, 8, 8, 8 | | | | |
|
|
* 8, 15, 8, 8, 8
|
|
*
|
|
* When moving the suffix, we found exactly 3 matches.
|
|
*
|
|
* The first suffix with period 5 is repeating.
|
|
*
|
|
* The next event is (3 * max_period) % suffix_period
|
|
*
|
|
* In this example, the result 0, so the next event is suffix[0] => 8
|
|
*
|
|
* However, 8 is the index in the array of exponential moving average
|
|
* which was calculated on the fly when storing the values, so the
|
|
* interval is ema[8] = 1366
|
|
*
|
|
*
|
|
* Example 2:
|
|
*
|
|
* 4, 3, 5, 100,
|
|
* 3, 3, 5, 117,
|
|
* 4, 4, 5, 112,
|
|
* 4, 3, 4, 110,
|
|
* 3, 5, 3, 117,
|
|
* 4, 4, 5, 112,
|
|
* 4, 3, 4, 110,
|
|
* 3, 4, 5, 112,
|
|
* 4, 3, 4, 110
|
|
*
|
|
* ilog2
|
|
*
|
|
* 0, 0, 0, 4,
|
|
* 0, 0, 0, 4,
|
|
* 0, 0, 0, 4,
|
|
* 0, 0, 0, 4,
|
|
* 0, 0, 0, 4,
|
|
* 0, 0, 0, 4,
|
|
* 0, 0, 0, 4,
|
|
* 0, 0, 0, 4,
|
|
* 0, 0, 0, 4
|
|
*
|
|
* Max period 5:
|
|
* 0, 0, 4,
|
|
* 0, 0, 0, 4,
|
|
* 0, 0, 0, 4,
|
|
* 0, 0, 0, 4
|
|
*
|
|
* Suffixes:
|
|
*
|
|
* 1) 0, 0, 4, 0, 0
|
|
* 2) 0, 0, 4, 0
|
|
* 3) 0, 0, 4
|
|
* 4) 0, 0
|
|
*
|
|
* buffer: 0, 0, 4, 0, 0, 0, 4, 0, 0, 0, 4, 0, 0, 0, 4
|
|
* | | | | | | X
|
|
* 0, 0, 4, 0, 0, | X
|
|
* 0, 0
|
|
*
|
|
* buffer: 0, 0, 4, 0, 0, 0, 4, 0, 0, 0, 4, 0, 0, 0, 4
|
|
* | | | | | | | | | | | | | | |
|
|
* 0, 0, 4, 0, | | | | | | | | | | |
|
|
* 0, 0, 4, 0, | | | | | | |
|
|
* 0, 0, 4, 0, | | |
|
|
* 0 0 4
|
|
*
|
|
* Pattern is found 3 times, the remaining is 1 which results from
|
|
* (max_period * 3) % suffix_period. This value is the index in the
|
|
* suffix arrays. The suffix array for a period 4 has the value 4
|
|
* at index 1.
|
|
*/
|
|
#define EMA_ALPHA_VAL 64
|
|
#define EMA_ALPHA_SHIFT 7
|
|
|
|
#define PREDICTION_PERIOD_MIN 3
|
|
#define PREDICTION_PERIOD_MAX 5
|
|
#define PREDICTION_FACTOR 4
|
|
#define PREDICTION_MAX 10 /* 2 ^ PREDICTION_MAX useconds */
|
|
#define PREDICTION_BUFFER_SIZE 16 /* slots for EMAs, hardly more than 16 */
|
|
|
|
/*
|
|
* Number of elements in the circular buffer: If it happens it was
|
|
* flushed before, then the number of elements could be smaller than
|
|
* IRQ_TIMINGS_SIZE, so the count is used, otherwise the array size is
|
|
* used as we wrapped. The index begins from zero when we did not
|
|
* wrap. That could be done in a nicer way with the proper circular
|
|
* array structure type but with the cost of extra computation in the
|
|
* interrupt handler hot path. We choose efficiency.
|
|
*/
|
|
#define for_each_irqts(i, irqts) \
|
|
for (i = irqts->count < IRQ_TIMINGS_SIZE ? \
|
|
0 : irqts->count & IRQ_TIMINGS_MASK, \
|
|
irqts->count = min(IRQ_TIMINGS_SIZE, \
|
|
irqts->count); \
|
|
irqts->count > 0; irqts->count--, \
|
|
i = (i + 1) & IRQ_TIMINGS_MASK)
|
|
|
|
struct irqt_stat {
|
|
u64 last_ts;
|
|
u64 ema_time[PREDICTION_BUFFER_SIZE];
|
|
int timings[IRQ_TIMINGS_SIZE];
|
|
int circ_timings[IRQ_TIMINGS_SIZE];
|
|
int count;
|
|
};
|
|
|
|
/*
|
|
* Exponential moving average computation
|
|
*/
|
|
static u64 irq_timings_ema_new(u64 value, u64 ema_old)
|
|
{
|
|
s64 diff;
|
|
|
|
if (unlikely(!ema_old))
|
|
return value;
|
|
|
|
diff = (value - ema_old) * EMA_ALPHA_VAL;
|
|
/*
|
|
* We can use a s64 type variable to be added with the u64
|
|
* ema_old variable as this one will never have its topmost
|
|
* bit set, it will be always smaller than 2^63 nanosec
|
|
* interrupt interval (292 years).
|
|
*/
|
|
return ema_old + (diff >> EMA_ALPHA_SHIFT);
|
|
}
|
|
|
|
static int irq_timings_next_event_index(int *buffer, size_t len, int period_max)
|
|
{
|
|
int period;
|
|
|
|
/*
|
|
* Move the beginning pointer to the end minus the max period x 3.
|
|
* We are at the point we can begin searching the pattern
|
|
*/
|
|
buffer = &buffer[len - (period_max * 3)];
|
|
|
|
/* Adjust the length to the maximum allowed period x 3 */
|
|
len = period_max * 3;
|
|
|
|
/*
|
|
* The buffer contains the suite of intervals, in a ilog2
|
|
* basis, we are looking for a repetition. We point the
|
|
* beginning of the search three times the length of the
|
|
* period beginning at the end of the buffer. We do that for
|
|
* each suffix.
|
|
*/
|
|
for (period = period_max; period >= PREDICTION_PERIOD_MIN; period--) {
|
|
|
|
/*
|
|
* The first comparison always succeed because the
|
|
* suffix is deduced from the first n-period bytes of
|
|
* the buffer and we compare the initial suffix with
|
|
* itself, so we can skip the first iteration.
|
|
*/
|
|
int idx = period;
|
|
size_t size = period;
|
|
|
|
/*
|
|
* We look if the suite with period 'i' repeat
|
|
* itself. If it is truncated at the end, as it
|
|
* repeats we can use the period to find out the next
|
|
* element with the modulo.
|
|
*/
|
|
while (!memcmp(buffer, &buffer[idx], size * sizeof(int))) {
|
|
|
|
/*
|
|
* Move the index in a period basis
|
|
*/
|
|
idx += size;
|
|
|
|
/*
|
|
* If this condition is reached, all previous
|
|
* memcmp were successful, so the period is
|
|
* found.
|
|
*/
|
|
if (idx == len)
|
|
return buffer[len % period];
|
|
|
|
/*
|
|
* If the remaining elements to compare are
|
|
* smaller than the period, readjust the size
|
|
* of the comparison for the last iteration.
|
|
*/
|
|
if (len - idx < period)
|
|
size = len - idx;
|
|
}
|
|
}
|
|
|
|
return -1;
|
|
}
|
|
|
|
static u64 __irq_timings_next_event(struct irqt_stat *irqs, int irq, u64 now)
|
|
{
|
|
int index, i, period_max, count, start, min = INT_MAX;
|
|
|
|
if ((now - irqs->last_ts) >= NSEC_PER_SEC) {
|
|
irqs->count = irqs->last_ts = 0;
|
|
return U64_MAX;
|
|
}
|
|
|
|
/*
|
|
* As we want to find three times the repetition, we need a
|
|
* number of intervals greater or equal to three times the
|
|
* maximum period, otherwise we truncate the max period.
|
|
*/
|
|
period_max = irqs->count > (3 * PREDICTION_PERIOD_MAX) ?
|
|
PREDICTION_PERIOD_MAX : irqs->count / 3;
|
|
|
|
/*
|
|
* If we don't have enough irq timings for this prediction,
|
|
* just bail out.
|
|
*/
|
|
if (period_max <= PREDICTION_PERIOD_MIN)
|
|
return U64_MAX;
|
|
|
|
/*
|
|
* 'count' will depends if the circular buffer wrapped or not
|
|
*/
|
|
count = irqs->count < IRQ_TIMINGS_SIZE ?
|
|
irqs->count : IRQ_TIMINGS_SIZE;
|
|
|
|
start = irqs->count < IRQ_TIMINGS_SIZE ?
|
|
0 : (irqs->count & IRQ_TIMINGS_MASK);
|
|
|
|
/*
|
|
* Copy the content of the circular buffer into another buffer
|
|
* in order to linearize the buffer instead of dealing with
|
|
* wrapping indexes and shifted array which will be prone to
|
|
* error and extremely difficult to debug.
|
|
*/
|
|
for (i = 0; i < count; i++) {
|
|
int index = (start + i) & IRQ_TIMINGS_MASK;
|
|
|
|
irqs->timings[i] = irqs->circ_timings[index];
|
|
min = min_t(int, irqs->timings[i], min);
|
|
}
|
|
|
|
index = irq_timings_next_event_index(irqs->timings, count, period_max);
|
|
if (index < 0)
|
|
return irqs->last_ts + irqs->ema_time[min];
|
|
|
|
return irqs->last_ts + irqs->ema_time[index];
|
|
}
|
|
|
|
static __always_inline int irq_timings_interval_index(u64 interval)
|
|
{
|
|
/*
|
|
* The PREDICTION_FACTOR increase the interval size for the
|
|
* array of exponential average.
|
|
*/
|
|
u64 interval_us = (interval >> 10) / PREDICTION_FACTOR;
|
|
|
|
return likely(interval_us) ? ilog2(interval_us) : 0;
|
|
}
|
|
|
|
static __always_inline void __irq_timings_store(int irq, struct irqt_stat *irqs,
|
|
u64 interval)
|
|
{
|
|
int index;
|
|
|
|
/*
|
|
* Get the index in the ema table for this interrupt.
|
|
*/
|
|
index = irq_timings_interval_index(interval);
|
|
|
|
if (index > PREDICTION_BUFFER_SIZE - 1) {
|
|
irqs->count = 0;
|
|
return;
|
|
}
|
|
|
|
/*
|
|
* Store the index as an element of the pattern in another
|
|
* circular array.
|
|
*/
|
|
irqs->circ_timings[irqs->count & IRQ_TIMINGS_MASK] = index;
|
|
|
|
irqs->ema_time[index] = irq_timings_ema_new(interval,
|
|
irqs->ema_time[index]);
|
|
|
|
irqs->count++;
|
|
}
|
|
|
|
static inline void irq_timings_store(int irq, struct irqt_stat *irqs, u64 ts)
|
|
{
|
|
u64 old_ts = irqs->last_ts;
|
|
u64 interval;
|
|
|
|
/*
|
|
* The timestamps are absolute time values, we need to compute
|
|
* the timing interval between two interrupts.
|
|
*/
|
|
irqs->last_ts = ts;
|
|
|
|
/*
|
|
* The interval type is u64 in order to deal with the same
|
|
* type in our computation, that prevent mindfuck issues with
|
|
* overflow, sign and division.
|
|
*/
|
|
interval = ts - old_ts;
|
|
|
|
/*
|
|
* The interrupt triggered more than one second apart, that
|
|
* ends the sequence as predictable for our purpose. In this
|
|
* case, assume we have the beginning of a sequence and the
|
|
* timestamp is the first value. As it is impossible to
|
|
* predict anything at this point, return.
|
|
*
|
|
* Note the first timestamp of the sequence will always fall
|
|
* in this test because the old_ts is zero. That is what we
|
|
* want as we need another timestamp to compute an interval.
|
|
*/
|
|
if (interval >= NSEC_PER_SEC) {
|
|
irqs->count = 0;
|
|
return;
|
|
}
|
|
|
|
__irq_timings_store(irq, irqs, interval);
|
|
}
|
|
|
|
/**
|
|
* irq_timings_next_event - Return when the next event is supposed to arrive
|
|
*
|
|
* During the last busy cycle, the number of interrupts is incremented
|
|
* and stored in the irq_timings structure. This information is
|
|
* necessary to:
|
|
*
|
|
* - know if the index in the table wrapped up:
|
|
*
|
|
* If more than the array size interrupts happened during the
|
|
* last busy/idle cycle, the index wrapped up and we have to
|
|
* begin with the next element in the array which is the last one
|
|
* in the sequence, otherwise it is at the index 0.
|
|
*
|
|
* - have an indication of the interrupts activity on this CPU
|
|
* (eg. irq/sec)
|
|
*
|
|
* The values are 'consumed' after inserting in the statistical model,
|
|
* thus the count is reinitialized.
|
|
*
|
|
* The array of values **must** be browsed in the time direction, the
|
|
* timestamp must increase between an element and the next one.
|
|
*
|
|
* Returns a nanosec time based estimation of the earliest interrupt,
|
|
* U64_MAX otherwise.
|
|
*/
|
|
u64 irq_timings_next_event(u64 now)
|
|
{
|
|
struct irq_timings *irqts = this_cpu_ptr(&irq_timings);
|
|
struct irqt_stat *irqs;
|
|
struct irqt_stat __percpu *s;
|
|
u64 ts, next_evt = U64_MAX;
|
|
int i, irq = 0;
|
|
|
|
/*
|
|
* This function must be called with the local irq disabled in
|
|
* order to prevent the timings circular buffer to be updated
|
|
* while we are reading it.
|
|
*/
|
|
lockdep_assert_irqs_disabled();
|
|
|
|
if (!irqts->count)
|
|
return next_evt;
|
|
|
|
/*
|
|
* Number of elements in the circular buffer: If it happens it
|
|
* was flushed before, then the number of elements could be
|
|
* smaller than IRQ_TIMINGS_SIZE, so the count is used,
|
|
* otherwise the array size is used as we wrapped. The index
|
|
* begins from zero when we did not wrap. That could be done
|
|
* in a nicer way with the proper circular array structure
|
|
* type but with the cost of extra computation in the
|
|
* interrupt handler hot path. We choose efficiency.
|
|
*
|
|
* Inject measured irq/timestamp to the pattern prediction
|
|
* model while decrementing the counter because we consume the
|
|
* data from our circular buffer.
|
|
*/
|
|
for_each_irqts(i, irqts) {
|
|
irq = irq_timing_decode(irqts->values[i], &ts);
|
|
s = idr_find(&irqt_stats, irq);
|
|
if (s)
|
|
irq_timings_store(irq, this_cpu_ptr(s), ts);
|
|
}
|
|
|
|
/*
|
|
* Look in the list of interrupts' statistics, the earliest
|
|
* next event.
|
|
*/
|
|
idr_for_each_entry(&irqt_stats, s, i) {
|
|
|
|
irqs = this_cpu_ptr(s);
|
|
|
|
ts = __irq_timings_next_event(irqs, i, now);
|
|
if (ts <= now)
|
|
return now;
|
|
|
|
if (ts < next_evt)
|
|
next_evt = ts;
|
|
}
|
|
|
|
return next_evt;
|
|
}
|
|
|
|
void irq_timings_free(int irq)
|
|
{
|
|
struct irqt_stat __percpu *s;
|
|
|
|
s = idr_find(&irqt_stats, irq);
|
|
if (s) {
|
|
free_percpu(s);
|
|
idr_remove(&irqt_stats, irq);
|
|
}
|
|
}
|
|
|
|
int irq_timings_alloc(int irq)
|
|
{
|
|
struct irqt_stat __percpu *s;
|
|
int id;
|
|
|
|
/*
|
|
* Some platforms can have the same private interrupt per cpu,
|
|
* so this function may be called several times with the
|
|
* same interrupt number. Just bail out in case the per cpu
|
|
* stat structure is already allocated.
|
|
*/
|
|
s = idr_find(&irqt_stats, irq);
|
|
if (s)
|
|
return 0;
|
|
|
|
s = alloc_percpu(*s);
|
|
if (!s)
|
|
return -ENOMEM;
|
|
|
|
idr_preload(GFP_KERNEL);
|
|
id = idr_alloc(&irqt_stats, s, irq, irq + 1, GFP_NOWAIT);
|
|
idr_preload_end();
|
|
|
|
if (id < 0) {
|
|
free_percpu(s);
|
|
return id;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
#ifdef CONFIG_TEST_IRQ_TIMINGS
|
|
struct timings_intervals {
|
|
u64 *intervals;
|
|
size_t count;
|
|
};
|
|
|
|
/*
|
|
* Intervals are given in nanosecond base
|
|
*/
|
|
static u64 intervals0[] __initdata = {
|
|
10000, 50000, 200000, 500000,
|
|
10000, 50000, 200000, 500000,
|
|
10000, 50000, 200000, 500000,
|
|
10000, 50000, 200000, 500000,
|
|
10000, 50000, 200000, 500000,
|
|
10000, 50000, 200000, 500000,
|
|
10000, 50000, 200000, 500000,
|
|
10000, 50000, 200000, 500000,
|
|
10000, 50000, 200000,
|
|
};
|
|
|
|
static u64 intervals1[] __initdata = {
|
|
223947000, 1240000, 1384000, 1386000, 1386000,
|
|
217416000, 1236000, 1384000, 1386000, 1387000,
|
|
214719000, 1241000, 1386000, 1387000, 1384000,
|
|
213696000, 1234000, 1384000, 1386000, 1388000,
|
|
219904000, 1240000, 1385000, 1389000, 1385000,
|
|
212240000, 1240000, 1386000, 1386000, 1386000,
|
|
214415000, 1236000, 1384000, 1386000, 1387000,
|
|
214276000, 1234000,
|
|
};
|
|
|
|
static u64 intervals2[] __initdata = {
|
|
4000, 3000, 5000, 100000,
|
|
3000, 3000, 5000, 117000,
|
|
4000, 4000, 5000, 112000,
|
|
4000, 3000, 4000, 110000,
|
|
3000, 5000, 3000, 117000,
|
|
4000, 4000, 5000, 112000,
|
|
4000, 3000, 4000, 110000,
|
|
3000, 4000, 5000, 112000,
|
|
4000,
|
|
};
|
|
|
|
static u64 intervals3[] __initdata = {
|
|
1385000, 212240000, 1240000,
|
|
1386000, 214415000, 1236000,
|
|
1384000, 214276000, 1234000,
|
|
1386000, 214415000, 1236000,
|
|
1385000, 212240000, 1240000,
|
|
1386000, 214415000, 1236000,
|
|
1384000, 214276000, 1234000,
|
|
1386000, 214415000, 1236000,
|
|
1385000, 212240000, 1240000,
|
|
};
|
|
|
|
static u64 intervals4[] __initdata = {
|
|
10000, 50000, 10000, 50000,
|
|
10000, 50000, 10000, 50000,
|
|
10000, 50000, 10000, 50000,
|
|
10000, 50000, 10000, 50000,
|
|
10000, 50000, 10000, 50000,
|
|
10000, 50000, 10000, 50000,
|
|
10000, 50000, 10000, 50000,
|
|
10000, 50000, 10000, 50000,
|
|
10000,
|
|
};
|
|
|
|
static struct timings_intervals tis[] __initdata = {
|
|
{ intervals0, ARRAY_SIZE(intervals0) },
|
|
{ intervals1, ARRAY_SIZE(intervals1) },
|
|
{ intervals2, ARRAY_SIZE(intervals2) },
|
|
{ intervals3, ARRAY_SIZE(intervals3) },
|
|
{ intervals4, ARRAY_SIZE(intervals4) },
|
|
};
|
|
|
|
static int __init irq_timings_test_next_index(struct timings_intervals *ti)
|
|
{
|
|
int _buffer[IRQ_TIMINGS_SIZE];
|
|
int buffer[IRQ_TIMINGS_SIZE];
|
|
int index, start, i, count, period_max;
|
|
|
|
count = ti->count - 1;
|
|
|
|
period_max = count > (3 * PREDICTION_PERIOD_MAX) ?
|
|
PREDICTION_PERIOD_MAX : count / 3;
|
|
|
|
/*
|
|
* Inject all values except the last one which will be used
|
|
* to compare with the next index result.
|
|
*/
|
|
pr_debug("index suite: ");
|
|
|
|
for (i = 0; i < count; i++) {
|
|
index = irq_timings_interval_index(ti->intervals[i]);
|
|
_buffer[i & IRQ_TIMINGS_MASK] = index;
|
|
pr_cont("%d ", index);
|
|
}
|
|
|
|
start = count < IRQ_TIMINGS_SIZE ? 0 :
|
|
count & IRQ_TIMINGS_MASK;
|
|
|
|
count = min_t(int, count, IRQ_TIMINGS_SIZE);
|
|
|
|
for (i = 0; i < count; i++) {
|
|
int index = (start + i) & IRQ_TIMINGS_MASK;
|
|
buffer[i] = _buffer[index];
|
|
}
|
|
|
|
index = irq_timings_next_event_index(buffer, count, period_max);
|
|
i = irq_timings_interval_index(ti->intervals[ti->count - 1]);
|
|
|
|
if (index != i) {
|
|
pr_err("Expected (%d) and computed (%d) next indexes differ\n",
|
|
i, index);
|
|
return -EINVAL;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int __init irq_timings_next_index_selftest(void)
|
|
{
|
|
int i, ret;
|
|
|
|
for (i = 0; i < ARRAY_SIZE(tis); i++) {
|
|
|
|
pr_info("---> Injecting intervals number #%d (count=%zd)\n",
|
|
i, tis[i].count);
|
|
|
|
ret = irq_timings_test_next_index(&tis[i]);
|
|
if (ret)
|
|
break;
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int __init irq_timings_test_irqs(struct timings_intervals *ti)
|
|
{
|
|
struct irqt_stat __percpu *s;
|
|
struct irqt_stat *irqs;
|
|
int i, index, ret, irq = 0xACE5;
|
|
|
|
ret = irq_timings_alloc(irq);
|
|
if (ret) {
|
|
pr_err("Failed to allocate irq timings\n");
|
|
return ret;
|
|
}
|
|
|
|
s = idr_find(&irqt_stats, irq);
|
|
if (!s) {
|
|
ret = -EIDRM;
|
|
goto out;
|
|
}
|
|
|
|
irqs = this_cpu_ptr(s);
|
|
|
|
for (i = 0; i < ti->count; i++) {
|
|
|
|
index = irq_timings_interval_index(ti->intervals[i]);
|
|
pr_debug("%d: interval=%llu ema_index=%d\n",
|
|
i, ti->intervals[i], index);
|
|
|
|
__irq_timings_store(irq, irqs, ti->intervals[i]);
|
|
if (irqs->circ_timings[i & IRQ_TIMINGS_MASK] != index) {
|
|
ret = -EBADSLT;
|
|
pr_err("Failed to store in the circular buffer\n");
|
|
goto out;
|
|
}
|
|
}
|
|
|
|
if (irqs->count != ti->count) {
|
|
ret = -ERANGE;
|
|
pr_err("Count differs\n");
|
|
goto out;
|
|
}
|
|
|
|
ret = 0;
|
|
out:
|
|
irq_timings_free(irq);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int __init irq_timings_irqs_selftest(void)
|
|
{
|
|
int i, ret;
|
|
|
|
for (i = 0; i < ARRAY_SIZE(tis); i++) {
|
|
pr_info("---> Injecting intervals number #%d (count=%zd)\n",
|
|
i, tis[i].count);
|
|
ret = irq_timings_test_irqs(&tis[i]);
|
|
if (ret)
|
|
break;
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int __init irq_timings_test_irqts(struct irq_timings *irqts,
|
|
unsigned count)
|
|
{
|
|
int start = count >= IRQ_TIMINGS_SIZE ? count - IRQ_TIMINGS_SIZE : 0;
|
|
int i, irq, oirq = 0xBEEF;
|
|
u64 ots = 0xDEAD, ts;
|
|
|
|
/*
|
|
* Fill the circular buffer by using the dedicated function.
|
|
*/
|
|
for (i = 0; i < count; i++) {
|
|
pr_debug("%d: index=%d, ts=%llX irq=%X\n",
|
|
i, i & IRQ_TIMINGS_MASK, ots + i, oirq + i);
|
|
|
|
irq_timings_push(ots + i, oirq + i);
|
|
}
|
|
|
|
/*
|
|
* Compute the first elements values after the index wrapped
|
|
* up or not.
|
|
*/
|
|
ots += start;
|
|
oirq += start;
|
|
|
|
/*
|
|
* Test the circular buffer count is correct.
|
|
*/
|
|
pr_debug("---> Checking timings array count (%d) is right\n", count);
|
|
if (WARN_ON(irqts->count != count))
|
|
return -EINVAL;
|
|
|
|
/*
|
|
* Test the macro allowing to browse all the irqts.
|
|
*/
|
|
pr_debug("---> Checking the for_each_irqts() macro\n");
|
|
for_each_irqts(i, irqts) {
|
|
|
|
irq = irq_timing_decode(irqts->values[i], &ts);
|
|
|
|
pr_debug("index=%d, ts=%llX / %llX, irq=%X / %X\n",
|
|
i, ts, ots, irq, oirq);
|
|
|
|
if (WARN_ON(ts != ots || irq != oirq))
|
|
return -EINVAL;
|
|
|
|
ots++; oirq++;
|
|
}
|
|
|
|
/*
|
|
* The circular buffer should have be flushed when browsed
|
|
* with for_each_irqts
|
|
*/
|
|
pr_debug("---> Checking timings array is empty after browsing it\n");
|
|
if (WARN_ON(irqts->count))
|
|
return -EINVAL;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int __init irq_timings_irqts_selftest(void)
|
|
{
|
|
struct irq_timings *irqts = this_cpu_ptr(&irq_timings);
|
|
int i, ret;
|
|
|
|
/*
|
|
* Test the circular buffer with different number of
|
|
* elements. The purpose is to test at the limits (empty, half
|
|
* full, full, wrapped with the cursor at the boundaries,
|
|
* wrapped several times, etc ...
|
|
*/
|
|
int count[] = { 0,
|
|
IRQ_TIMINGS_SIZE >> 1,
|
|
IRQ_TIMINGS_SIZE,
|
|
IRQ_TIMINGS_SIZE + (IRQ_TIMINGS_SIZE >> 1),
|
|
2 * IRQ_TIMINGS_SIZE,
|
|
(2 * IRQ_TIMINGS_SIZE) + 3,
|
|
};
|
|
|
|
for (i = 0; i < ARRAY_SIZE(count); i++) {
|
|
|
|
pr_info("---> Checking the timings with %d/%d values\n",
|
|
count[i], IRQ_TIMINGS_SIZE);
|
|
|
|
ret = irq_timings_test_irqts(irqts, count[i]);
|
|
if (ret)
|
|
break;
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int __init irq_timings_selftest(void)
|
|
{
|
|
int ret;
|
|
|
|
pr_info("------------------- selftest start -----------------\n");
|
|
|
|
/*
|
|
* At this point, we don't except any subsystem to use the irq
|
|
* timings but us, so it should not be enabled.
|
|
*/
|
|
if (static_branch_unlikely(&irq_timing_enabled)) {
|
|
pr_warn("irq timings already initialized, skipping selftest\n");
|
|
return 0;
|
|
}
|
|
|
|
ret = irq_timings_irqts_selftest();
|
|
if (ret)
|
|
goto out;
|
|
|
|
ret = irq_timings_irqs_selftest();
|
|
if (ret)
|
|
goto out;
|
|
|
|
ret = irq_timings_next_index_selftest();
|
|
out:
|
|
pr_info("---------- selftest end with %s -----------\n",
|
|
ret ? "failure" : "success");
|
|
|
|
return ret;
|
|
}
|
|
early_initcall(irq_timings_selftest);
|
|
#endif
|